# sg_scan -i # sg_format --format --size=512 /dev/sg0
Can you post the zpool status output?
Yes, I can do this. I have no issues with destroying the pool if that makes getting to a solution easier.Keep this card on server, for now. Where did you get the disks from, a datacenter? Can you destroy the pool, take all disks out and zero-format one with 512B sectors?
That means the vendor had them properly formatted, if they were used into datacenter. I thought maybe the disks came with the server as package.We just purchased the drives off of Amazon.
The HP drives did come with the server as part of the purchase. Was purchased off of Ebay. Not sure if it was used in datacenter or not.That means the vendor had them properly formatted, if they were used into datacenter. I thought maybe the disks came with the server as package.
I see, they are probably formatted with bad sectors, most HP OEM disks are. Let’s try the command I posted earlier on a HP drive that came with the server. Make sure you only have one disk installed.The HP drives did come with the server as part of the purchase. Was purchased off of Ebay. Not sure if it was used in datacenter or not.
Do I attempt to destroy the pool and then remove all the other disks? Or should I remove all but one of the HP disks and then perform that command?I see, they are probably formatted with bad sectors, most HP OEM disks are. Let’s try the command I posted earlier on a HP drive that came with the server. Make sure you only have one disk installed.
You will probably have a hard time destroying the pool, like I did.
Should I recreate the pool with z2 or z3? We went with z1 more for available storage in the pool. We should have purchased larger drives. We just assumed z1 would be good enough.That's an awfully wide Z1! I use a Z1 often, generally it's 3 drives per stripe though (better speeds for my uses). I wouldn't recommend 11 drives in a single Z1 stripe. I've seen this discussed on the ZFS subreddit (certain large drive systems), but it's generally freebsd. One theory is it's a driver bug. I'm not sure.
I just did a zpool clear. It looks alright now. Does this mean the drives are OK?So, degraded refers to checksum errors but the drives likely still work, I see the resilver reported zero errors. Have you ever done the zpool clear? The clear should change the status if all is well.
I'll keep an eye on things and see what happens. I'm probably going to end up re-creating this pool anyways.Might be. You may have solved the issue with all the previous steps. zfs would not say no errors on resilver if they were not ok (presuming this was the only resilver, if not it depends if the first resilver was also no errors corrected or not). If it doesn't reoccur, problem solved. When drives say degraded but it appears everything is working and resilver ok, you do the zpool clear to change the state back to ok. You do not just randomly do that though if you have real errors. Your OP shows 1 write error and the rest read. The various steps in this thread were good to try and resolve the issue, but were missing the clear to get rid of degraded.
Is there any way to change the RaidZ(x) of a pool once it's been created? I have an 8-drive pool that is also a Z1. Will this be an issue in the future?Before Truenas, I had a box with 10 drives, 3 stripes of 3 RaidZ (wanting the sequential read performance, lots of video access). Could have done 5 mirrors but too much space lost. All the drives were server pulls and used. I lost about 4 drives over 5 years, every resilver worked fine just as it should with ZFS. But If you truly had serious drive failures and 2 of them (more likely with your large 1 RaidZ1 set), it is much less fault tolerant. The general rule in ZFS that most people accept is largest set is 3 drives for a RaidZ1. Esp if you have larger drives.
It depends how many disks you use, for 6-12 disks I strongly recommend to use RaidZ2. You will not be happy when you lose all data just because you feel like gaining few terabytes. Is common to have another disk fail while resilvering another. It only happened once to me, to have another disk fail during resilvering another. It was pretty stressful until first disk finished. If another disk would’ve failed during that resilver process, bye bye all data. I always keep 3 spare disks in hand, for situations like that. I always buy disks from different vendors to avoid same batch.We went with z1 more for available storage in the pool.
No. Also, you cannot just remove disks, you need to destroy the pool first, unless you take a disk offline to replace it. Once you create the pool, it dies with that number of disks, until you destroy it. You can replace the disks to increase the pool storage size but you cannot expand/contract the number of disks. That’s why is common to create small vdevs of 6 disks or so. I always create 12 disk pools and vdevs, is the border line for CMR disks, when it comes time to resilver a failed disk.Is there any way to change the RaidZ(x) of a pool once it's been created?
Is there any way to change the RaidZ(x) of a pool once it's been created? I have an 8-drive pool that is also a Z1. Will this be an issue in the future?
After talking with my supervisor our plan is to create the pool from scratch with new drives of larger size (4TB) so that we can do a Z2 and get similar available space.It depends how many disks you use, for 6-12 disks I strongly recommend to use RaidZ2. You will not be happy when you lose all data just because you feel like gaining few terabytes. Is common to have another disk fail while resilvering another. It only happened once to me, to have another disk fail during resilvering another. It was pretty stressful until first disk finished. Of another disk would’ve failed, bye bye all data. I always keep 3 spare disks in hand, for situations like that. I always buy disks from different vendors to avoid same batch.
No. Also, you cannot just remove disks, you need to destroy the pool first, unless you take a disk offline to replace it. Once you create the pool, it dies with that number of disks, until you destroy it. You can replace the disks to increase the pool storage size but you cannot expand/contract the number of disks. That’s why is common to create small vdevs of 6 disks or so. I always create 12 disk pools and vdevs, is the border line for CMR disks, when it comes time to resilver a failed disk.
Back to our discussion, after you format the disk, when you create a new pool with one disk only, it will give you only one option, so you cannot make mistakes.
BTW, did I missed something, you solved the issue?
As stated in my previous reply we're going to rebuild the pool with larger drives. We're going to go with Z2 drives. What tests/scrub should I perform on these disks? Should these tests be done prior to building them into a pool?Any of the changes made could have fixed the issue in theory. But, it doesn't seem to me that this has been verified.
I would re-build the pool with Z2 and run tests/scrub to see if any of those errors come back before moving forward with leaning on the system.
And can I do this after the pool is created?I always do a long smartctl test for any drive before I use it in any manner.