My pool had two disks (3T Barracudas, surprise
) start throwing uncorrectable sector errors, so I replaced the one that was worst. I also noticed that replication hadn't been running since I updated the push machine from 9.3 to to 9.10 a few weeks ago. During the resilver, I ended up with some minor data corruption, and I am trying to make sure I can "save" the rest in case the other drive craps out before the resilver completes since its error count keeps going up.
I updated my replication pull target to 9.10 and moved the system dataset to make replication run again, but the one affected snapshot fails to send and replication just runs over and over.
This is the state of the pool. I appear to have the rest of the data replicated now, but even after I swap the disks and save the pool, I will still need to fix this somehow.
I continually get emails with this:
I updated my replication pull target to 9.10 and moved the system dataset to make replication run again, but the one affected snapshot fails to send and replication just runs over and over.
This is the state of the pool. I appear to have the rest of the data replicated now, but even after I swap the disks and save the pool, I will still need to fix this somehow.
Code:
[root@freenas] ~# zpool status -v zfspool pool: zfspool state: ONLINE status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scan: resilver in progress since Thu May 19 06:30:07 2016 2.85T scanned out of 4.01T at 79.4M/s, 4h14m to go 612G resilvered, 71.08% done config: NAME STATE READ WRITE CKSUM zfspool ONLINE 0 0 8 raidz1-0 ONLINE 0 0 16 gptid/ae451901-b3d1-11e4-b68a-001e4fb0f51d ONLINE 0 0 0 ada1 ONLINE 0 0 0 (resilvering) gptid/1d7db6e9-add2-11e2-ab62-525400390d09 ONLINE 8 0 0 gptid/1e1b3145-add2-11e2-ab62-525400390d09 ONLINE 0 0 0 errors: Permanent errors have been detected in the following files: zfspool/home@auto-20160516.1822-7d:/path/to/badfile.gz
I continually get emails with this:
Code:
Hello, The replication failed for the local ZFS zfspool/home while attempting to apply incremental send of snapshot auto-20160515.1822-7d -> auto-20160516.1822-7d to 10.0.0.52
- Is there something like zfs_send_corrupt_data tunable for ZFSoL that will override this behavior and allow me to send the rest of the intact data?
- Can I tell ZFS to ignore the corrupt file in the snapshot and carry on?
- If I can stop the replication in order to unlock and delete the affected snapshot and its children, how would I do that?