gdarends
Explorer
- Joined
- Jan 20, 2015
- Messages
- 72
Update on HDD tests.
I ran "smartctl -t long" on all drives. It took about 10hrs. All drives came out without issues except for one that had 1 error. (This is one of the newly bought drives btw)
I tried to run the solnet-array-test, but couldn't. It only runs on TrueNAS Core, not Scale.
I ran some fio tests. Some tests were giving very weak results while some were ridiculous (I think because of caching)
I ran this one multiple times right after each other and I think this was hitting the cache.
This test I set primarycache to none on the dataset before running.
I ran "smartctl -t long" on all drives. It took about 10hrs. All drives came out without issues except for one that had 1 error. (This is one of the newly bought drives btw)
SMART overall-health self-assessment test result: PASSED
....
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x000b 100 100 016 Pre-fail Always - 0
2 Throughput_Performance 0x0005 136 136 054 Pre-fail Offline - 80
3 Spin_Up_Time 0x0007 180 180 024 Pre-fail Always - 468 (Average 391)
4 Start_Stop_Count 0x0012 100 100 000 Old_age Always - 15
5 Reallocated_Sector_Ct 0x0033 100 100 005 Pre-fail Always - 0
7 Seek_Error_Rate 0x000b 100 100 067 Pre-fail Always - 0
8 Seek_Time_Performance 0x0005 119 119 020 Pre-fail Offline - 35
9 Power_On_Hours 0x0012 100 100 000 Old_age Always - 64
10 Spin_Retry_Count 0x0013 100 100 060 Pre-fail Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always - 15
192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age Always - 15
193 Load_Cycle_Count 0x0012 100 100 000 Old_age Always - 15
194 Temperature_Celsius 0x0002 125 125 000 Old_age Always - 48 (Min/Max 22/52)
196 Reallocated_Event_Count 0x0032 100 100 000 Old_age Always - 0
197 Current_Pending_Sector 0x0022 100 100 000 Old_age Always - 0
198 Offline_Uncorrectable 0x0008 100 100 000 Old_age Offline - 0
199 UDMA_CRC_Error_Count 0x000a 200 200 000 Old_age Always - 0
SMART Error Log Version: 1
ATA Error Count: 1
CR = Command Register [HEX]
FR = Features Register [HEX]
SC = Sector Count Register [HEX]
SN = Sector Number Register [HEX]
CL = Cylinder Low Register [HEX]
CH = Cylinder High Register [HEX]
DH = Device/Head Register [HEX]
DC = Device Command Register [HEX]
ER = Error register [HEX]
ST = Status register [HEX]
Powered_Up_Time is measured from power on, and printed as
DDd+hh:mm:SS.sss where DD=days, hh=hours, mm=minutes,
SS=sec, and sss=millisec. It "wraps" after 49.710 days.
Error 1 occurred at disk power-on lifetime: 32834 hours (1368 days + 2 hours)
When the command that caused the error occurred, the device was active or idle.
After command completion occurred, registers were:
ER ST SC SN CL CH DH
-- -- -- -- -- -- --
84 51 00 00 00 00 00
Commands leading to the command that caused the error were:
CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
-- -- -- -- -- -- -- -- ---------------- --------------------
ec 00 00 00 00 00 00 00 00:25:40.465 IDENTIFY DEVICE
ec 00 00 00 00 00 00 00 00:25:40.463 IDENTIFY DEVICE
ec 00 00 00 00 00 00 00 00:25:40.462 IDENTIFY DEVICE
ec 00 00 00 00 00 00 00 00:25:40.460 IDENTIFY DEVICE
ef 10 02 00 00 00 00 00 00:25:40.215 SET FEATURES [Enable SATA feature]
I tried to run the solnet-array-test, but couldn't. It only runs on TrueNAS Core, not Scale.
I ran some fio tests. Some tests were giving very weak results while some were ridiculous (I think because of caching)
root@truenas[/mnt/baki]# fio --name=random-write --ioengine=posixaio --rw=randwrite --bs=4k --numjobs=1 --size=4g --iodepth=1 --runtime=60 --time_based --end_fsync=1
Run status group 0 (all jobs):
WRITE: bw=26.9MiB/s (28.2MB/s), 26.9MiB/s-26.9MiB/s (28.2MB/s-28.2MB/s), io=1635MiB (1714MB), run=60693-60693msec
root@truenas[/mnt/baki]# fio --name=random-write --ioengine=posixaio --rw=randwrite --bs=4k --size=4g --numjobs=1 --iodepth=1 --runtime=60 --time_based --end_fsync=1
Run status group 0 (all jobs):
WRITE: bw=25.8MiB/s (27.0MB/s), 25.8MiB/s-25.8MiB/s (27.0MB/s-27.0MB/s), io=1557MiB (1632MB), run=60438-60438msec
root@truenas[/mnt/baki]# fio --name=random-write --ioengine=posixaio --rw=randwrite --bs=64k --size=256m --numjobs=16 --iodepth=16 --runtime=60 --time_based --end_fsync=1
Run status group 0 (all jobs):
WRITE: bw=222MiB/s (233MB/s), 13.6MiB/s-16.6MiB/s (14.3MB/s-17.4MB/s), io=16.8GiB (18.0GB), run=64412-77161msec
root@truenas[/mnt/baki]# fio --name=random-write --ioengine=posixaio --rw=randwrite --bs=1m --size=16g --numjobs=1 --iodepth=1 --runtime=60 --time_based --end_fsync=1
Run status group 0 (all jobs):
WRITE: bw=358MiB/s (375MB/s), 358MiB/s-358MiB/s (375MB/s-375MB/s), io=24.1GiB (25.9GB), run=68886-68886msec
root@truenas[/mnt/baki]# fio --iodepth=64 --ioengine=libaio --direct=1 --rw=randwrite --bs=4k --size=2G --numjobs=64 --runtime=20 --group_reporting --name=test-rand-write
Run status group 0 (all jobs):
WRITE: bw=28.0MiB/s (29.4MB/s), 28.0MiB/s-28.0MiB/s (29.4MB/s-29.4MB/s), io=561MiB (588MB), run=20009-20009msec
root@truenas[/mnt/baki]# fio --iodepth=64 --ioengine=libaio --direct=1 --rw=write --bs=1m --size=2g --numjobs=4 --runtime=20 --group_reporting --name=test-write
Run status group 0 (all jobs):
WRITE: bw=476MiB/s (500MB/s), 476MiB/s-476MiB/s (500MB/s-500MB/s), io=8192MiB (8590MB), run=17196-17196msec
I ran this one multiple times right after each other and I think this was hitting the cache.
root@truenas[/mnt/baki]# fio --iodepth=64 --ioengine=libaio --direct=1 --rw=read --bs=1m --size=2g --numjobs=4 --runtime=10 --group_reporting --name=test-read
Run status group 0 (all jobs):
READ: bw=11.3GiB/s (12.1GB/s), 11.3GiB/s-11.3GiB/s (12.1GB/s-12.1GB/s), io=8192MiB (8590MB), run=711-711msec
Run status group 0 (all jobs):
READ: bw=1451MiB/s (1522MB/s), 1451MiB/s-1451MiB/s (1522MB/s-1522MB/s), io=8192MiB (8590MB), run=5645-5645msec
Run status group 0 (all jobs):
READ: bw=7056MiB/s (7399MB/s), 7056MiB/s-7056MiB/s (7399MB/s-7399MB/s), io=8192MiB (8590MB), run=1161-1161msec
Run status group 0 (all jobs):
READ: bw=7564MiB/s (7932MB/s), 7564MiB/s-7564MiB/s (7932MB/s-7932MB/s), io=8192MiB (8590MB), run=1083-1083msec
This test I set primarycache to none on the dataset before running.
root@truenas[/mnt/baki]# fio --directory=/baki/fio --name=randread --ioengine=libaio --iodepth=32 --rw=randread --bs=1M --size=256M --numjobs=32 --time_based --runtime=300
Run status group 0 (all jobs):
READ: bw=177MiB/s (185MB/s), 4466KiB/s-7787KiB/s (4573kB/s-7974kB/s), io=51.8GiB (55.7GB), run=300008-300234msec