darkwarrior
Patron
- Joined
- Mar 29, 2015
- Messages
- 336
Hi,
I'm with you on that one.
Much better IO performance and easier pool expansion
You guys do what you feel you need to do. If you are running a database or a bunch of VMs, there might be a usage where you can get higher performance from mirrors and I don't want you to think that I am arguing, but here is some food for though.Hi,
I'm with you on that one.
Much better IO performance and easier pool expansion![]()
Device: /dev/da11 [SAT], 8 Currently unreadable (pending) sectors Device: /dev/da11 [SAT], 8 Offline uncorrectable sectors Device: /dev/da11 [SAT], Self-Test Log error count increased from 0 to 1
and
[...snip ...]
Now, I can't leave work to go check on it or deal with it in any way until I get home, so, if my pool were made of mirrors, this would worry me. Because I am running RAID-z2, I have a second level of redundancy and I know that I don't have to worry because I can just replace that disk when I get home... It's all better now and I never broke a sweat.
[...snip...]
There is also the idea that mirrors rebuild faster. That would only be true if you were limited by read speed. I just don't see that in my pool, the limit is write speed. The speed of the rebuild is dependent on the amount of data that must be written to the replacement disk. With 700GB of data to transfer onto the drive, it doesn't matter if that drive is in a RAID-z2 or if it is in a mirror, 700GB is still 700GB and it takes time to transfer that much data onto the drive.
[... snip...]
I actually upgraded from a socket 1366 Xeon server to a socket 1155 Xeon just to get the IPMI. There were some other improvements, but the key feature was the IPMI.also, IPMI is gorgeous.
and
I had a disk fail on Friday (9 December 2016) and my NAS was helpful enough to e-mail me at 10:01 AM (while I am at work 60 miles away) just as follows:
Code:Device: /dev/da11 [SAT], 8 Currently unreadable (pending) sectors Device: /dev/da11 [SAT], 8 Offline uncorrectable sectors Device: /dev/da11 [SAT], Self-Test Log error count increased from 0 to 1
Now, I can't leave work to go check on it or deal with it in any way until I get home, so, if my pool were made of mirrors, this would worry me. Because I am running RAID-z2, I have a second level of redundancy and I know that I don't have to worry because I can just replace that disk when I get home... It's all better now and I never broke a sweat.
Disks can, and do, fail without warning. If you are running a mirror, you can only survive 1 failure. Unless you have been especially vigilant in your purchasing of drives, it is likely that the disks in your mirror are the same age and even from the same batch which means they stand a decent chance of failing about the same time. The theory being that when one disk of the mirror fails, the other disk in the mirror could fail before you can replace the first or while the rebuild is in progress.
# zpool create test mirror c0t0d0 c0t1d0 spare c0t2d0 # zpool status test pool: test state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 mirror ONLINE 0 0 0 c0t0d0 ONLINE 0 0 0 c0t1d0 ONLINE 0 0 0 spares c0t2d0 AVAIL errors: No known data errors
# zpool add <pool> spare <disk>
You can add spares/log/cache drives through the volume manager in the GUI.
I would not try to add a USB drive to your pool. USB is not a reliable drive interface. It is not and HBA.I have a 4x4Tb pool. If I buy a 4Tb USB3.0 drive, and keep it parked on the USB HBA... What if anything will happen if I run:
When you open the volume manager tab, you click the "Manual Setup" button.This actually brings up a really good point, that I haven't had a chance to poke at yet with FreeNAS. Way back in the day (2006-ish for ZFS purposes) I was an actual Sun employee. We configured our ZFS pools in Solaris 10 with hot-spares like this:
I would not try to add a USB drive to your pool. USB is not a reliable drive interface. It is not and HBA.
When you open the volume manager tab, you click the "Manual Setup" button.
What I did do, one time before, was attempt to attach a USB device to import data. The thing that happened to me was, it crashed the entire system. I wouldn't want to trust a USB device as part of my pool because of the possibility of it crashing the entire server and corrupting my storage pool.
I would rather buy another HBA to add more drives. I don't have that problem though because I have a 48 bay chassis with 12 empty bays.
I recognize that you have a background in this, but I do also. Those of us that offer assistance on the forum try to find out what the need of the user is before we make suggestions so that we can make suggestions that are applicable to the scenario that the user has and not simply based on what our personal preferences are. I understand that you like to have a hot-spare, but I would rather not (especially at home) and I will share the reasoning behind that if you are interested, just ask. In a business environment, hot spares can save the day and I have seen that happen personally, but I don't see the need at my house even though I have 36 drives in my system. I have a full, online, backup of everything in my storage pool in a backup pool on the same server and I run a backup server also. It would be inconvenient for sure if the main storage pool failed, but I would need three drives to fully fail in one vdev (not just bad sectors) in the main pool before I would loose it and even then I would still have all my data in the backup pool, which is online and shared in the same server, and I have a backup server with another copy that is synced hourly. The backup server is also shared to the network so if the whole primary server goes down for some reason, the data can still be accessed from the backup server. This is how I build for home. I have a primary server with 36 drives and a backup server with another 8 drives. I just don't think I need to go to the additional step of having hot spares. I think the odds of having three pools fail simultaneously are astronomical.As a former Storage QA engineer, believe me... If you have 36 spinning platters and no hot spares, you're on borrowed time!
Not sure what your point is here, but you can't utilize the network speed of 10Gb Ethernet only having 4 dives. The drives will not be fast enough to feed the network connection.My little $90 NAS has but 4 SATA ports, and for my needs, 10GbE is probably a higher priority than a chassis upgrade.
Please do, I would love to have reliable USB connectivity. To be completely honest, there have been two new versions of BSD since I tried it, so the problem could very well be fixed. Still, I have a hard time trusting that it will work reliably.If attaching a USB drive crashes the kernel, I want to reproduce that and catch it in the kernel debugger. Grab the stack trace and submit a bug... Or better yet, submit a bug with a patch!
I manage a system at work that was built with disk shelves. It has 16 drives in the head and 16 more in each of 4 shelves. It was setup two or three years before I started working here and the original configuration was for each set of 16 drives to be a 15 drive RAIDz2 vdev with the 16th drive in the shelf being a hot spare. So 5 hot spares in an 80 drive system. When I started working here, it had been ignored for a long time because I filled a position that had been vacant for over a year and nobody had been minding the servers. This server had already used all the hot spares because of drive failures and if one more drive had failed in the second vdev, the whole pool would have been lost. I recognize that hot spares can be the thing that saves a system, however, if a system is monitored it should never get to that point.But I remain committed to the use of hot spares. It may not make sense for a small NAS like mine, it may not make sense to use a poor attachment mechanism. But hot spares are a very good thing!
I check my server health every day and I have email alerts setup, so I will be notified of a drive fault immediately and I replace drives at the first reallocated sector. I am absolutely paranoid about any fault causing data to be corrupted or destroyed. I have burned-in and tested cold spares ready to replace a drive and, at worst, it would be 18 hours between the fault and the drive replacement.
I recognize that you have a background in this, but I do also.
(snip)
I think the odds of having three pools fail simultaneously are astronomical.
Not sure what your point is here, but you can't utilize the network speed of 10Gb Ethernet only having 4 dives. The drives will not be fast enough to feed the network connection.
I recognize that hot spares can be the thing that saves a system, however, if a system is monitored it should never get to that point.
(snip)
at worst, it would be 18 hours between the fault and the drive replacement.
For my purposes, my personal home use, I am not willing to spend the money for an offsite backup and my location is not likely to suffer a disaster that would destroy my systems. Even though I was without power for a month and had to have a new roof after Hurricane Katrina, I was back in operation as soon as power was restored. In my home, I am not paying the bill for a backup generator any more than I am paying a monthly charge to back my data up offsite. I am not made of money / it isn't worth that much to me. There would need to be a business case for it and my personal data doesn't make me any money. It is more like spending money on a boat. The more I spend on it, the more I spend on it...having all three pools in the same building is in and of itself a risk, and that's what your off-site backup is for.
I went there, and no, it isn't worth the hassle.The intermediate step of quad-1GbE is sitting here on my desk, I finally found the box the cards were hiding in, but I'm not sure it's worth the cabling & configuration hassle when 10GbE cards are so cheap.
A SAS card like this is the easy answer to add a few drives:Yes, I'm out of SATA ports, but I have a couple free PCIe slots...
That is what took three of the drives in the second vdev of that 80 drive system. The system had last logged the drive temperature at over 100 C and the two drives adjacent to it failed after that. I agree with the idea of hot-spares, when the circumstances call for it and I will likely build hot spares into the pool of the new 60 drive server that we are getting at work later this year. It isn't that I don't see the value of hot spares, I just don't see the value of them in my home system.Worn & failed drives get hot, and tend to cook adjacent drives,
I have had two drives start accumulating bad sectors within minutes of each other (two in one day) but most of the drive failures that cause me to replace drives, both at home and at work, are bad sectors and not catastrophic failures. I have had catastrophic failures, but it is not as common as it once was.What are the statistical odds of two drives failing in 18 hours?
We have a problem in my server room right now where one of the two cooling systems is not strong enough to keep it cool. We found out the hard way when the newer (stronger) system had a failure and the other system couldn't do the job. I am going to push management to get an additional cooling system. It is hard to get them to see the need for and approve spending for something that is only needed in the case that something else fails.Air Conditioning can fail.
Take a look at these scripts:Is there any good guidance out there about setting up a good alerting strategy? Does FreeNAS just do that out of the box if I've configured email alerts? Specifically what did you have to do, if anything, to receive alerts on reallocated sectors?
A SAS card like this is the easy answer to add a few drives
It is hard to get them to see the need for and approve spending for something that is only needed in the case that something else fails.