Good morning!
I have a bunch of SMR disks lying around, and I figured I'd try and do something with them. Since SMR is becoming more prevalent, we'll run into enterprise solutions employing the technology for storage sooner or later, so NAS software should be figuring out ways to deal with it as well.
Predictably, I keep running into errors like this:
Surprisingly, there's nothing much that the base configuration of TrueNAS Scale allows me to do about this. The timeout in question appears to be hardcoded in a kernel driver somewhere. I suspect it's related to
Now I'm no wizkid, and recompiling the kernel to see if increasing this value might make SMR disks useable is something I'd rather not try - also because I'm not sure what'll happen next time I update the TrueNAS Scale box I'm performing this experiment on.
So I have a couple of questions:
I have a bunch of SMR disks lying around, and I figured I'd try and do something with them. Since SMR is becoming more prevalent, we'll run into enterprise solutions employing the technology for storage sooner or later, so NAS software should be figuring out ways to deal with it as well.
Predictably, I keep running into errors like this:
Code:
Aug 9 17:00:10 nas kernel: sd 8:0:0:0: attempting task abort!scmd(0x00000000c4e14ac5), outstanding for 1420 ms & timeout 1000 ms Aug 9 17:00:10 nas kernel: sd 8:0:0:0: tag#35 CDB: ATA command pass through(16) 85 08 0e 00 d5 00 01 00 e0 00 4f 00 c2 00 b0 00 Aug 9 17:00:10 nas kernel: scsi target8:0:0: handle(0x000a), sas_address(0x5001e677b7d18fe8), phy(8) Aug 9 17:00:10 nas kernel: scsi target8:0:0: enclosure logical id(0x5001e677b7d18fff), slot(8) Aug 9 17:00:10 nas kernel: sd 8:0:0:0: device_block, handle(0x000a) Aug 9 17:00:10 nas kernel: sd 8:0:0:0: task abort: SUCCESS scmd(0x00000000c4e14ac5) Aug 9 17:00:10 nas kernel: sd 8:0:0:0: device_unblock and setting to running, handle(0x000a) Aug 9 17:00:11 nas kernel: sd 8:0:0:0: Power-on or device reset occurred
Surprisingly, there's nothing much that the base configuration of TrueNAS Scale allows me to do about this. The timeout in question appears to be hardcoded in a kernel driver somewhere. I suspect it's related to
Code:
/usr/src/linux-headers-5.10.142+truenas/include/linux/blkdev.h:BLK_MIN_SG_TIMEOUT
Now I'm no wizkid, and recompiling the kernel to see if increasing this value might make SMR disks useable is something I'd rather not try - also because I'm not sure what'll happen next time I update the TrueNAS Scale box I'm performing this experiment on.
So I have a couple of questions:
- Is there way to change this timeout other than recompile the kernel?
- Might future releases have such a feature? Is that workable, should I post a feature request (https://ixsystems.atlassian.net/jira/software/c/projects/NAS/issues)?
- What else could I try to make SMR disks play well with ZFS?