Hello!
Before I get started, here's my system:
45 Drives Storinator
TrueNAS: TrueNAS-13.0-U5.3
MB: Supermicro X11DPL-i
HBA: LSI 9305 x 3
RAM: 128GB ECC
HD's: Mostly 8TB Seagate EXOS, some Seagate Ironwolf - all 8TB
Zpool is 4 vdevs of 10 drives as raidz3 with 5 drives as spares (I know, lol) - when I built this system, I was having multiple drive failures with the old system and decided to be overly cautious
Here's the problem:
I recently had a drive fail (da16), and one of my hot spares jumped in (da40). I got the notification, so I went to the UI, put the failed drive offline, and then shut down to replace da16. I put the new drive in, and started back up and let it resilver. When it completed without problems, and both drives were online, the spare did not return automatically to the spare pool. Instead of detaching the spare, I put it offline (doh!). It did not return to the spare set, so I tried to detach it, and the UI gives this error:
Even though the operation was unsuccessful, I do now see da40 back in the list of spares and ONLINE. However, it is also still in the pool where it originally jumped in as a spare:
I did some searching on the forums and the recommendation was to detach the drive from the CLI. Here's the pool:
The next thing I tried to was to online the drive with 'zpool online':
Got the same result if I try and detach:
So I'm a little stuck. da40 (gptid/0d93ec08-4c57-11eb-b083-3cecef6e76ba) is no longer part of the pool and is back in the list of spares, but zpool status still shows it in the output.
Do you guys have any ideas?
Thanks!
Before I get started, here's my system:
45 Drives Storinator
TrueNAS: TrueNAS-13.0-U5.3
MB: Supermicro X11DPL-i
HBA: LSI 9305 x 3
RAM: 128GB ECC
HD's: Mostly 8TB Seagate EXOS, some Seagate Ironwolf - all 8TB
Zpool is 4 vdevs of 10 drives as raidz3 with 5 drives as spares (I know, lol) - when I built this system, I was having multiple drive failures with the old system and decided to be overly cautious
Here's the problem:
I recently had a drive fail (da16), and one of my hot spares jumped in (da40). I got the notification, so I went to the UI, put the failed drive offline, and then shut down to replace da16. I put the new drive in, and started back up and let it resilver. When it completed without problems, and both drives were online, the spare did not return automatically to the spare pool. Instead of detaching the spare, I put it offline (doh!). It did not return to the spare set, so I tried to detach it, and the UI gives this error:
Even though the operation was unsuccessful, I do now see da40 back in the list of spares and ONLINE. However, it is also still in the pool where it originally jumped in as a spare:
I did some searching on the forums and the recommendation was to detach the drive from the CLI. Here's the pool:
Code:
root@goliath[~]# zpool status GOLIATH -f
cannot open '-f': name must begin with a letter
pool: GOLIATH
state: DEGRADED
status: One or more devices has been taken offline by the administrator.
Sufficient replicas exist for the pool to continue functioning in a
degraded state.
action: Online the device using 'zpool online' or replace the device with
'zpool replace'.
scan: resilvered 4.69T in 13:04:00 with 0 errors on Sat Aug 26 00:34:41 2023
config:
NAME STATE READ WRITE CKSUM
GOLIATH DEGRADED 0 0 0
raidz3-0 ONLINE 0 0 0
gptid/955cf876-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/a23ef856-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/96793d87-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/9cf24cd8-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/986b2ff2-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/a5bbaa71-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/7e3f0319-4375-11ee-aa23-3cecef6e76ba.eli ONLINE 0 0 0
gptid/add0e87d-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/b23ffb56-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/b1a2259c-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
raidz3-1 DEGRADED 0 0 0
gptid/9f55db7e-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/a3cb392f-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/a098b204-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/9ebdebc0-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/a4904562-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/5b63b7a9-4d49-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
spare-6 DEGRADED 0 0 0
gptid/f1d1d7af-d59c-11ed-a862-3cecef6e76ba.eli ONLINE 0 0 0
gptid/0d93ec08-4c57-11eb-b083-3cecef6e76ba.eli OFFLINE 0 0 0
gptid/d0f5e63f-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/d1f02e0f-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/dd660045-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
raidz3-2 ONLINE 0 0 0
gptid/d7875504-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/dc3d4213-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/e092bb98-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/e27dd2db-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/e456ac7b-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/da30a26a-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/de1d2773-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/e8b14a04-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/e8165f17-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/e084d613-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
raidz3-3 ONLINE 0 0 0
gptid/e68ae0f8-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/e7662f77-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/ed566e44-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/fa460ef2-4c56-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/05452e50-4c57-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/074b1868-4c57-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/0a02dac7-4c57-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/0c65792b-4c57-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/1349e22f-4c57-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
gptid/1445f15a-4c57-11eb-b083-3cecef6e76ba.eli ONLINE 0 0 0
spares
gptid/0d93ec08-4c57-11eb-b083-3cecef6e76ba.eli AVAIL
gptid/0ffba24a-4c57-11eb-b083-3cecef6e76ba.eli AVAIL
gptid/1430a462-4c57-11eb-b083-3cecef6e76ba.eli AVAIL
gptid/162bb3a8-4c57-11eb-b083-3cecef6e76ba.eli AVAIL
gptid/16a0022d-4c57-11eb-b083-3cecef6e76ba.eli AVAIL
errors: No known data errors
root@goliath[~]#
The next thing I tried to was to online the drive with 'zpool online':
Code:
root@goliath[~]# zpool online GOLIATH gptid/0d93ec08-4c57-11eb-b083-3cecef6e76ba cannot online gptid/0d93ec08-4c57-11eb-b083-3cecef6e76ba: no such device in pool root@goliath[~]#
Got the same result if I try and detach:
Code:
root@goliath[~]# zpool detach GOLIATH gptid/0d93ec08-4c57-11eb-b083-3cecef6e76ba cannot detach gptid/0d93ec08-4c57-11eb-b083-3cecef6e76ba: no such device in pool root@goliath[~]#
So I'm a little stuck. da40 (gptid/0d93ec08-4c57-11eb-b083-3cecef6e76ba) is no longer part of the pool and is back in the list of spares, but zpool status still shows it in the output.
Do you guys have any ideas?
Thanks!