Brandito
Explorer
- Joined
- May 6, 2023
- Messages
- 72
Yeah, I have 91TB on the disks, I have 3 spares and some room on another nas but not a contiguous 91TB free anywhere. I would need a second disk shelf most likely.
The checksums have increased again
It's extra odd that it tells me to do a zpool status -v for a list but when I do that it doesn't actually list anything
With the swap change, I don't really recall adding or removing the swap. I believe it was enabled when I started on core, but I quickly migrated to scale after a few weeks. What I do recall doing was moving the system dataset from the boot-pool to another pool. I thought it was my home pool, but that can't be because I know I tried moving it there when this whole issue started. Does that affect the swap location? The two options are in the same menu, so I assume it does?
When I got the error that started this thread, I tried moving the system dataset to my home pool again thinking it would add the swap partition. Then I offlined a single drive in the new vdev and replaced it with itself assuming again that it would add in the swap partition. Is it possible that caused the corruption? Like I mentioned before, the resilver was strange, starting at a 1 day to complete and then actually completing within hours.
I am still using the "new" to me HBAs and brand new cables, but they're likely not as good as the original sff cables I had which were supermicro branded.
Should I shut down and at least swap the cables? I'm just worried to bring the machine offline now.
The checksums have increased again
Code:
root@truenas[~]# zpool status pool: Home state: ONLINE status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore the entire pool from backup. see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-8A scan: resilvered 1.32T in 02:06:45 with 0 errors on Sat Nov 11 13:45:36 2023 config: NAME STATE READ WRITE CKSUM Home ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 a7d78b0d-f891-11ed-a2f8-90e2baf17bf0 ONLINE 0 0 33 a7b00eef-f891-11ed-a2f8-90e2baf17bf0 ONLINE 0 0 33 a7d01f81-f891-11ed-a2f8-90e2baf17bf0 ONLINE 0 0 33 a7c951e3-f891-11ed-a2f8-90e2baf17bf0 ONLINE 0 0 0 a7bfef1b-f891-11ed-a2f8-90e2baf17bf0 ONLINE 0 0 0 e4f37ae1-f494-4baf-94e5-07db0c38cb0c ONLINE 0 0 0 raidz2-1 ONLINE 0 0 0 8cca2c8f-39ee-40a6-88e0-24ddf3485aa0 ONLINE 0 0 2 74f3cc23-1b32-4faf-89cc-ba0cd72ba308 ONLINE 0 0 46 4e5f5b16-6c2b-4e6b-a907-3e1b9b1c4886 ONLINE 0 0 46 cde58bb6-9d8e-4cdc-a1bf-847f459b459b ONLINE 0 0 46 58c22778-521b-4e8f-aadd-6d5ad17a8f68 ONLINE 0 0 2 33633f68-920b-4a40-bd4d-45e30b6872bc ONLINE 0 0 2 raidz2-2 ONLINE 0 0 0 2a2e5211-d4ea-4da9-8ea5-bdabdc542bdb ONLINE 0 0 0 56c07fd7-6cb6-4985-9a20-2b5ff9d42631 ONLINE 0 0 0 1147286d-8cd8-4025-8e5d-bbf06e2bd795 ONLINE 0 0 44 7e1fa408-7565-4913-b045-49447ef9253b ONLINE 0 0 44 3d56d2fa-d505-4bea-b9a2-80c121e4e559 ONLINE 0 0 44 a9906b32-2690-4f7b-8d8f-00ca915d8f3d ONLINE 0 0 0 raidz2-5 ONLINE 0 0 0 b8c63108-353b-4ed7-a927-ca3df817bd21 ONLINE 0 0 0 58782264-02f1-41c6-9b91-d07144cb0ccb ONLINE 0 0 32 03df98a5-a86d-4bc8-879a-5cf611d4306c ONLINE 0 0 32 022c7ffb-0a07-45cb-b3af-ad1730a08054 ONLINE 0 0 32 a5786a1f-a7ad-4a30-877a-88a03c94a774 ONLINE 0 0 0 4c59238e-5cbd-428e-8a72-a018d9dae9c2 ONLINE 0 0 0 logs mirror-6 ONLINE 0 0 0 5ba1f70b-be51-470f-94ed-777683425477 ONLINE 0 0 0 f2605776-46a9-4455-a4bc-322d4cf8a688 ONLINE 0 0 0 errors: 2 data errors, use '-v' for a list
It's extra odd that it tells me to do a zpool status -v for a list but when I do that it doesn't actually list anything
With the swap change, I don't really recall adding or removing the swap. I believe it was enabled when I started on core, but I quickly migrated to scale after a few weeks. What I do recall doing was moving the system dataset from the boot-pool to another pool. I thought it was my home pool, but that can't be because I know I tried moving it there when this whole issue started. Does that affect the swap location? The two options are in the same menu, so I assume it does?
When I got the error that started this thread, I tried moving the system dataset to my home pool again thinking it would add the swap partition. Then I offlined a single drive in the new vdev and replaced it with itself assuming again that it would add in the swap partition. Is it possible that caused the corruption? Like I mentioned before, the resilver was strange, starting at a 1 day to complete and then actually completing within hours.
I am still using the "new" to me HBAs and brand new cables, but they're likely not as good as the original sff cables I had which were supermicro branded.
Should I shut down and at least swap the cables? I'm just worried to bring the machine offline now.
Last edited: