Chris Moore
Hall of Famer
- Joined
- May 2, 2015
- Messages
- 10,079
That is unfortunate.
Can you schedule down-time for this system?
I had asked what the purpose was, partly to find out how big the impact would be to taking it offline.
Are the enclosures you have situated so that you can see the drive serial numbers relatively easily?
I have a top-loaded system where I can slide the unit out, open the lid and look at all the serial numbers directly.
This is probably going to take a manual hunt for the missing serial number, but you have a list of all the working serial numbers, so when you find a serial number that isn't in your list, you know that is the bad drive.
This is a rough situation to be in. It sure looks like someone set you up for failure. Sorry.
The pool with the failed drive, named 'cRaid', from what you showed us, has 8TB drives and it should have a capacity around 599 TB, but you should keep usage below 475 TB if you go by the 80% rule, but you can push it closer to 90% without significant performance loss.
The real problem with both of these pools is the span of disks.
If it were possible, all the data needs to be moved from this system so the pools can be reconfigured. All of it could be in one large pool, but it needs to be broken into multiple vdevs (virtual devices) which is where the redundancy comes from.
You could also fill those 60 bay enclosures with more disks to add additional capacity.
Can you schedule down-time for this system?
I had asked what the purpose was, partly to find out how big the impact would be to taking it offline.
Are the enclosures you have situated so that you can see the drive serial numbers relatively easily?
I have a top-loaded system where I can slide the unit out, open the lid and look at all the serial numbers directly.
This is probably going to take a manual hunt for the missing serial number, but you have a list of all the working serial numbers, so when you find a serial number that isn't in your list, you know that is the bad drive.
This is a rough situation to be in. It sure looks like someone set you up for failure. Sorry.
The pool with the failed drive, named 'cRaid', from what you showed us, has 8TB drives and it should have a capacity around 599 TB, but you should keep usage below 475 TB if you go by the 80% rule, but you can push it closer to 90% without significant performance loss.
The real problem with both of these pools is the span of disks.
If it were possible, all the data needs to be moved from this system so the pools can be reconfigured. All of it could be in one large pool, but it needs to be broken into multiple vdevs (virtual devices) which is where the redundancy comes from.
You could also fill those 60 bay enclosures with more disks to add additional capacity.