not much going right for me with TrueNAS-12.0-U4.1

tomkinsg

Cadet
Joined
Aug 11, 2021
Messages
4
I had an alert of a disk degraded two days ago. I ordered two replacement disks from amazon (figured i may as well have a spare).

They arrived and this afternoon, i had another disk report degraded. Oh no!

That's the 2nd nothing going right.

now when i go to the GUI and look at the pool, i see no drives. I see all drives under disks in the GUI.

1628726855451.png


When in the shell i issue a ZPOOL STATUS, this is returned:

Code:
root@TrueNAS:~ # zpool status
  pool: VOL001
state: DEGRADED
status: One or more devices has experienced an unrecoverable error.  An
        attempt was made to correct the error.  Applications are unaffected.
action: Determine if the device needs to be replaced, and clear the errors
        using 'zpool clear' or replace the device with 'zpool replace'.
   see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-9P
  scan: resilvered 444K in 00:00:02 with 0 errors on Wed Aug 11 15:31:24 2021
config:

        NAME                                            STATE     READ WRITE CKSUM
        VOL001                                          DEGRADED     0     0 0
          raidz1-0                                      DEGRADED     0     0 0
            gptid/af14e8bd-b359-11e6-ac8e-d05099760fd1  DEGRADED     0     0 0  too many errors
            gptid/b063d911-b359-11e6-ac8e-d05099760fd1  ONLINE       0     0 0
            gptid/b15f3b6b-b359-11e6-ac8e-d05099760fd1  DEGRADED     0     0 0  too many errors
            gptid/b279a1ed-b359-11e6-ac8e-d05099760fd1  ONLINE       0     0 0

errors: No known data errors

  pool: freenas-boot
state: ONLINE
status: Some supported features are not enabled on the pool. The pool can
        still be used, but some features are unavailable.
action: Enable all features using 'zpool upgrade'. Once this is done,
        the pool may no longer be accessible by software that does not support
        the features. See zpool-features(5) for details.
  scan: scrub repaired 0B in 00:00:37 with 0 errors on Tue Aug 10 03:45:37 2021
config:

        NAME                                            STATE     READ WRITE CKSUM
        freenas-boot                                    ONLINE       0     0 0
          mirror-0                                      ONLINE       0     0 0
  


i tried rebooting, removing the first degraded disk, putting a new one in and then issuing

r/truenas - nothing going right on TrueNAS
I rebooted, replaced the degraded original ada0 and tried again and got the same result.

The only good thing is that i appear to be able to access data on the drive thru SMB.

I really appreciate some help.

GT
 

tomkinsg

Cadet
Joined
Aug 11, 2021
Messages
4
ugh - i missed the step to get to the drive status. SMH,. what a numpty. Failed drive is now replaced. Trying the second now.
 
Joined
Jan 4, 2014
Messages
1,644
I hope you have a backup. If not, get your data off while you still can. RAIDZ1 can tolerate and recover from a loss of one drive, not two.

Are you using non-SMR drives? What's the output of sesutil show?

You haven't described the h/w TrueNAS is running on. It may provide some additional clues on why you're losing disks.
 
Last edited:

tomkinsg

Cadet
Joined
Aug 11, 2021
Messages
4
SESUTIL SHOW says no devices found.

I left it last night with the 2nd disk replaced and it was copying data. i woke this morning and it was not responding. i rebooted and i have a critical alert and the pool is offline

I have an export/disconnect option but i am nervous at this point

This is mostly media and not that critical. Is there a quick way i can dump out my config/setup to share?
 

NugentS

MVP
Joined
Apr 16, 2020
Messages
2,947
Multiple failures - check the cabling, is there any in common?
 

tomkinsg

Cadet
Joined
Aug 11, 2021
Messages
4
i am not sure what info is wanted thar i could provide that would help

I have 4x seagate st4000dm000 drives
the power supply is a 400w output FSP group
the moteherboard is a ASR N3150itx
all discs in the pool are connected directly to the motherboard. i have 3x boot SSD drives attached to an additional sata i/f card.

when i boot now it fails. I get an error ;

Solaris: warning pool "VOL001" has encountered an uncorrectable io error suspended

the boot then fails. I unplugged the new drive and booted and the server started. all drives are listed in disks.

If i re-add either the new disk or the original that was degraded, the disk is seen but this has no effect on teh availability of the pool.

is there anything that can be done?

thanks
 
Joined
Jan 4, 2014
Messages
1,644
The forum rules will give you some clues as to what you should include in your h/w specs e.g. RAM type, processor, etc...

tn06.jpg


The Seagate drives you're using are SMR drives. This is likely the reason you've lost your pool. You need to replace these with CMR drives. If you're not clear on the SMR issue, please review this thread List of known SMR drives,

You have lost your pool. It's time to rebuild it. Replace the disks you are using with CMR disks such as drives from the Seagate NAS range or WD Red Plus or higher spec WD NAS drives.

Other points to consider:
  1. Disk redundancy is not a substitute for an effective backup strategy.
  2. Higher-order ZFS e.g. RAIDZ2 will tolerate more disk failures. The trade-off is disk space.
 
Top