Resilver performance on TrueNAS 13.0-U6

Tasslehoff

Cadet
Joined
Nov 30, 2023
Messages
1
Hi, recently on one of my TrueNAS I got a fault NVMe drive used for logs and cache, so I changed the faulted drive with a new one, cloned partitions structures and made a zfs replace on the faulted device with the right partition on the new one.

Code:
root@drsrv01[~]# zpool status drpool
  pool: drpool
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
        continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
  scan: resilver in progress since Tue Nov 28 17:44:25 2023
        256G scanned at 1.54M/s, 257G issued at 1.54M/s, 81.2T total
        0B resilvered, 0.31% done, no estimated completion time
config:

        NAME                                              STATE     READ WRITE CKSUM
        drpool                                            DEGRADED     0     0     0
          raidz2-0                                        ONLINE       0     0     0
            gptid/0669279c-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/06a65ff5-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/06b6a529-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/06cdfa15-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/06f6faa6-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/071217a4-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/0709e55c-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/07002dfd-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/072ebe3b-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            gptid/073b819f-eded-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
        logs
          mirror-1                                        DEGRADED     0     0     0
            gptid/379f7792-edf9-11eb-80d9-7c10c91fa126    ONLINE       0     0     0
            replacing-1                                   DEGRADED     0     0     0
              435495427710988294                          UNAVAIL      0     0     0  was /dev/gptid/16c5daf8-edf9-11eb-80d9-7c10c91fa126
              gptid/b0142897-8d14-11ee-bf44-7c10c91fa126  ONLINE       0     0     0
        cache
          gptid/44c383dd-edf9-11eb-80d9-7c10c91fa126      UNAVAIL      0     0     0  cannot open
          gptid/4d025c31-edf9-11eb-80d9-7c10c91fa126      ONLINE       0     0     0

errors: No known data errors


As you can see the pool is pretty big and the resilver takes forever, in two days It did only 0.31% :eek:

I was looking for some zfs tuning parameters to speed things out and I found something...
Code:
vfs.zfs.scrub_delay=0
vfs.zfs.top_maxinflight=128
vfs.zfs.resilver_min_time_ms=5000
vfs.zfs.resilver_delay=0


...but seems like vfs.zfs.scrub_delay and vfs.zfs.resilver_delay return "unknown oid" error.

Please note that I'm running the latest Core version (I updated from 13.0-U4 a few days ago).
Is there any other available parameters I can tune to make resilver a little faster?
Do vfs.zfs.scrub_delay and vfs.zfs.resilver_delay parameters been changed with something else?

Thanks for any advice.

Tas
 
Top