Sprint
Explorer
- Joined
- Mar 30, 2019
- Messages
- 72
Morning all
So this is an odd one. Quick back ground, this is an all SSD pool, consisting of 6x1Tb Samsung Evo 860 Sata SSDs (in 2x 3 wide RaidZ1 vdevs) which has run flawlessly for a number of years. I also had a 7th brand new (when pool was built) 1Tb Samsung Evo 860 assigned as a hot spare. Pool also has slog/dedup/metadata vdevs provided by 2xOptane mirrored.
This morning I noticed that a drive had faulted, it LOOKS like the spare was invoked but that also has errors against it? If it was a HDD I wouldn't be surprised, but a brand new/never seen any work SSD to spit out errors, just doesn't seem right. I'd like to remove the spare SSD from the pool todo some quick test, then (assuming its ok) reconnect it and do a replace of the faulty SSD, but TrueNas isn't able to remove the drive, i get. I need to tread carefully here, as I really don't want to lose the pool, hence coming to you wonderful people for guidance.
Pool is encrypted, its primary function is as a iSCSI storage for Proxmox VMs, i have backups on a local and offsite backup servers, but I really want to avoid using them as my VMs will lose data.
Thanks in advance
Sprint
Edit: Running TrueNAS-13.0-U6.1
So this is an odd one. Quick back ground, this is an all SSD pool, consisting of 6x1Tb Samsung Evo 860 Sata SSDs (in 2x 3 wide RaidZ1 vdevs) which has run flawlessly for a number of years. I also had a 7th brand new (when pool was built) 1Tb Samsung Evo 860 assigned as a hot spare. Pool also has slog/dedup/metadata vdevs provided by 2xOptane mirrored.
This morning I noticed that a drive had faulted, it LOOKS like the spare was invoked but that also has errors against it? If it was a HDD I wouldn't be surprised, but a brand new/never seen any work SSD to spit out errors, just doesn't seem right. I'd like to remove the spare SSD from the pool todo some quick test, then (assuming its ok) reconnect it and do a replace of the faulty SSD, but TrueNas isn't able to remove the drive, i get
Code:
[EZFS_BUSY] Pool busy; removal may already be in progress
Pool is encrypted, its primary function is as a iSCSI storage for Proxmox VMs, i have backups on a local and offsite backup servers, but I really want to avoid using them as my VMs will lose data.
Code:
pool: SSD_Array state: DEGRADED status: One or more devices are faulted in response to persistent errors. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Replace the faulted device, or use 'zpool clear' to mark the device repaired. scan: resilvered 0B in 01:13:42 with 0 errors on Sun Mar 10 14:46:58 2024 config: NAME STATE READ WRITE CKSUM SSD_Array DEGRADED 0 0 0 raidz1-0 DEGRADED 0 0 0 spare-0 UNAVAIL 10 91 0 insufficient replicas gptid/64135c7d-5ec1-11ec-b267-ac1f6b781c6e FAULTED 0 146 0 too many errors gptid/0e2850c7-965d-11ec-9cae-ac1f6b781c6e FAULTED 6 111 0 too many errors gptid/6471b79d-5ec1-11ec-b267-ac1f6b781c6e ONLINE 0 0 0 gptid/64e1a36a-5ec1-11ec-b267-ac1f6b781c6e ONLINE 0 0 0 raidz1-1 ONLINE 0 0 0 gptid/6489cb08-5ec1-11ec-b267-ac1f6b781c6e ONLINE 0 0 0 gptid/64b14982-5ec1-11ec-b267-ac1f6b781c6e ONLINE 0 0 0 gptid/64cbc764-5ec1-11ec-b267-ac1f6b781c6e ONLINE 0 0 0 dedup mirror-2 ONLINE 0 0 0 nvd2p4 ONLINE 0 0 0 nvd3p4 ONLINE 0 0 0 special mirror-5 ONLINE 0 0 0 nvd2p3 ONLINE 0 0 0 nvd3p3 ONLINE 0 0 0 logs mirror-6 ONLINE 0 0 0 nvd2p2 ONLINE 0 0 0 nvd3p2 ONLINE 0 0 0 spares gptid/0e2850c7-965d-11ec-9cae-ac1f6b781c6e INUSE currently in use errors: No known data errors
Thanks in advance
Sprint
Edit: Running TrueNAS-13.0-U6.1
Last edited: