Slow transfer speeds on gigabit ethernet

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
What speed do you need?
I was hoping for minimum 500mbps

Like I said earlier 10gig fibre or copper will be faster but your storage pool is too slow and you will still be disappointed.
I do have 3 cache ssd's, and I am using my NAS for playing games so I don't need high transfer speeds on my smb share only my iSCSI need needs to be fast.
 
Joined
Jan 18, 2017
Messages
525
Units of measure are very important for us to understand what you want to do. You said your internet connection is 500mbps, you can currently write data to your NAS at about 900mbps. Which means you have already reached your objective.
 

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
I was hoping for minimum 500mbps

Good news, you're already getting 1000Mbps (megaBITS per second). There are eight bits in a byte. You need to divide or multiply by eight when converting.

If you want 500MB/s (megaBYTES per second) speeds, you will need to run a 10Gbps (10 gigaBITS per second) network on both your desktop and server, and substantially increase the speed of your storage pool. Cache SSDs will only help to a limited degree after the game files are repeatedly read - you would likely need to make an all-SSD pool as well as increase your CPU speed since the Xeon 5500 series is getting a bit long in the tooth.

I am using my NAS for playing games

It's vanishingly unlikely that even with 10Gbps networking and an all-SSD pool the iSCSI share would be faster than just installing your hard drives or SSDs locally to your gaming machine.
 

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
Units of measure are very important for us to understand what you want to do. You said your internet connection is 500mbps, you can currently write data to your NAS at about 900mbps. Which means you have already reached your objective.
But my switch can do 1 gigaBYTE and my pc can handle 1 gigaBYTE, but the file explorer says 100 megaBYTE.
Could you explain why I only got 100 megaBYTES in my file transfer?
Thanks.
 

Samuel Tai

Never underestimate your own stupidity
Moderator
Joined
Apr 24, 2020
Messages
5,399
No, your switch and PC NIC can only do 1 GigaBIT per second. File Explorer says 100 MByte, which is 800 MegaBIT, well within the bandwidth of your switch and PC NIC.
 

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
that is very confusing, don't understand why we can't just use either byte or bit.
So I should install fiber or buy a 10 gig ethernet card?
 

Samuel Tai

Never underestimate your own stupidity
Moderator
Joined
Apr 24, 2020
Messages
5,399
Blame it on marketing and telecomm tradition. Transfer bandwidth has always been measured and advertised in units of bits/second (base 10). Disk storage historically been measured in units of bytes (8 bits, base 2). So a MB (actually a MiB) is 2^20 bytes, or 8,388,608 bits.
 

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
Does anybody know why my pc freezes if I do a heavy file transfer and this comes up in the console:
1663342271912.png
 

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
Blame it on marketing and telecomm tradition. Transfer bandwidth has always been measured and advertised in units of bits/second (base 10). Disk storage historically been measured in units of bytes (8 bits, base 2). So a MB (actually a MiB) is 2^20 bytes, or 8,388,608 bits.
To add to the mess, disk storage is measured in the 2^X fashion, but sold in 10^X fashion - a "MB" as shown by a computer is actually a "mebibyte" or "MiB" of 8,388,608 bits or 1,048,576 bytes - but the drive is being sold by the megabyte (MB) which is 1,000,000 bytes.

That's why your "1TB" SSD shows up in your computer as "931GB" - it's actually 1,000,000,000,000 (10^12) bytes as opposed to 1,099,511,627,776 (2^40) bytes.

Does anybody know why my pc freezes if I do a heavy file transfer and this comes up in the console:

Couple different causes could be in play here (network hardware is getting overwhelmed by the packets, pool isn't able to keep up with the incoming data, delayed ACK is being hit and timing out, network interface on your PC is being shared by iSCSI and other traffic) but firstly - I see two different IPs calling a ping timeout there (.47 and .76) so time out (heh) for an important question:

Have you connected two different Windows PCs/iSCSI initiators to this single LUN?
 

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
Have you connected two different Windows PCs/iSCSI initiators to this single LUN?
Yes I wanted to test if this happend with two computers, but this happend before I connected two pc to the iscsi share.

Is it maybe the pool that can't handle the incoming data? each hard drive can take 150 mbps each, and I have 4, is it overloading the hard drives even with the 3 ssd I use for cache?
 

ChrisRJ

Wizard
Joined
Oct 23, 2020
Messages
1,919
You need to use the correct speed abbreviation (as layed out by @HoneyBadger above), if you want proper responses. 150 mbps does not exist in practice; in theory it means 150 milli bit per second.
 
Last edited:

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
Based on the fio line, your entire pool is capable of 150MB/s - not each disk.

This seems, as the kids would say, "a little sus" as my 2-way mirror-pool can beat those numbers, and you have a four-disk stripe which should theoretically be able to push a much higher rate of throughput.

Code:
root@freenas:/mnt/MirrorVolume/test # fio --name TEST --eta-newline=5s --filename=fio-tempfile.dat --rw=randwrite --size=50g --io_size=1500g --blocksize=128k --iodepth=1 --direct=1 --numjobs=1 --group_reporting --runtime=60
TEST: (g=0): rw=randwrite, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=psync, iodepth=1
fio-3.0
Starting 1 process
TEST: Laying out IO file (1 file / 51200MiB)
Jobs: 1 (f=1): [w(1)][13.1%][r=0KiB/s,w=148MiB/s][r=0,w=1184 IOPS][eta 00m:53s]
Jobs: 1 (f=1): [w(1)][23.0%][r=0KiB/s,w=141MiB/s][r=0,w=1125 IOPS][eta 00m:47s]
Jobs: 1 (f=1): [w(1)][31.7%][r=0KiB/s,w=155MiB/s][r=0,w=1243 IOPS][eta 00m:41s]
Jobs: 1 (f=1): [w(1)][42.6%][r=0KiB/s,w=133MiB/s][r=0,w=1064 IOPS][eta 00m:35s]
Jobs: 1 (f=1): [w(1)][52.5%][r=0KiB/s,w=127MiB/s][r=0,w=1015 IOPS][eta 00m:29s]
Jobs: 1 (f=1): [w(1)][62.3%][r=0KiB/s,w=146MiB/s][r=0,w=1166 IOPS][eta 00m:23s]
Jobs: 1 (f=1): [w(1)][71.7%][r=0KiB/s,w=139MiB/s][r=0,w=1115 IOPS][eta 00m:17s]
Jobs: 1 (f=1): [w(1)][81.7%][r=0KiB/s,w=126MiB/s][r=0,w=1010 IOPS][eta 00m:11s]
Jobs: 1 (f=1): [w(1)][90.2%][r=0KiB/s,w=141MiB/s][r=0,w=1126 IOPS][eta 00m:06s]
Jobs: 1 (f=1): [w(1)][100.0%][r=0KiB/s,w=141MiB/s][r=0,w=1128 IOPS][eta 00m:00s]
TEST: (groupid=0, jobs=1): err= 0: pid=54757: Fri Sep 16 15:15:41 2022
  write: IOPS=1290, BW=161MiB/s (169MB/s)(9681MiB/60001msec)
    clat (usec): min=27, max=340794, avg=767.57, stdev=3622.40
     lat (usec): min=30, max=340797, avg=771.09, stdev=3622.46
    clat percentiles (usec):
     |  1.00th=[   30],  5.00th=[   31], 10.00th=[   58], 20.00th=[   68],
     | 30.00th=[   79], 40.00th=[   86], 50.00th=[   91], 60.00th=[  101],
     | 70.00th=[  113], 80.00th=[  586], 90.00th=[ 1762], 95.00th=[ 4080],
     | 99.00th=[11076], 99.50th=[12649], 99.90th=[16057], 99.95th=[22152],
     | 99.99th=[99091]
   bw (  KiB/s): min= 1024, max=2324409, per=99.53%, avg=164440.47, stdev=212517.62, samples=119
   iops        : min=    8, max=18159, avg=1284.18, stdev=1660.30, samples=119
  lat (usec)   : 50=9.45%, 100=49.57%, 250=15.24%, 500=3.34%, 750=7.23%
  lat (usec)   : 1000=1.28%
  lat (msec)   : 2=4.52%, 4=4.25%, 10=3.71%, 20=1.34%, 50=0.04%
  lat (msec)   : 100=0.01%, 250=0.01%, 500=0.01%
  cpu          : usr=0.86%, sys=7.33%, ctx=91342, majf=0, minf=0
  IO depths    : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
     submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     issued rwt: total=0,77448,0, short=0,0,0, dropped=0,0,0
     latency   : target=0, window=0, percentile=100.00%, depth=1

Run status group 0 (all jobs):
  WRITE: bw=161MiB/s (169MB/s), 161MiB/s-161MiB/s (169MB/s-169MB/s), io=9681MiB (10.2GB), run=60001-60001msec
 

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
Based on the fio line, your entire pool is capable of 150MB/s - not each disk.

This seems, as the kids would say, "a little sus" as my 2-way mirror-pool can beat those numbers, and you have a four-disk stripe which should theoretically be able to push a much higher rate of throughput.
Here is the part number for my 4x 4tb drives in a stripe: WDC WD40EFRX-68WT0N0
remember that I also have 750gb of ssd cache.

So I would expect too get higher transfer speeds.
I'm also considering upgrading to 10gbps nic card directly connected beetwen my NAS and pc.
 

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
Here is the part number for my 4x 4tb drives in a stripe: WDC WD40EFRX-68WT0N0

EFRX isn't SMR so you aren't dealing with that thankfully.

Try fio --name=testwrites --rw=write --bs=16k --size=16g --numjobs=1 --iodepth=1 --runtime=60 --time_based just to storm your pool with 16K async writes (which is what your iSCSI ZVOL will do under load) - make sure it can keep up with draining to disk.

Code:
root@freenas:/mnt/MirrorVolume/test # fio --name=testwrites --rw=write --bs=16k --size=16g --numjobs=1 --iodepth=1 --runtime=60 --time_based
testwrites: (g=0): rw=write, bs=(R) 16.0KiB-16.0KiB, (W) 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=psync, iodepth=1
fio-3.0
Starting 1 process
Jobs: 1 (f=1): [W(1)][100.0%][r=0KiB/s,w=377MiB/s][r=0,w=24.1k IOPS][eta 00m:00s]
testwrites: (groupid=0, jobs=1): err= 0: pid=57107: Fri Sep 16 15:40:42 2022
  write: IOPS=23.0k, BW=360MiB/s (377MB/s)(21.1GiB/60001msec)
    clat (usec): min=7, max=399120, avg=41.20, stdev=478.90
     lat (usec): min=8, max=399121, avg=41.80, stdev=478.92
    clat percentiles (usec):
     |  1.00th=[    9],  5.00th=[    9], 10.00th=[    9], 20.00th=[    9],
     | 30.00th=[    9], 40.00th=[    9], 50.00th=[   11], 60.00th=[   24],
     | 70.00th=[   74], 80.00th=[   91], 90.00th=[   92], 95.00th=[   96],
     | 99.00th=[  145], 99.50th=[  151], 99.90th=[  255], 99.95th=[  375],
     | 99.99th=[  709]
   bw (  KiB/s): min=70958, max=893732, per=99.62%, avg=367083.10, stdev=112177.85, samples=119
   iops        : min= 4433, max=55858, avg=22942.29, stdev=7011.19, samples=119
  lat (usec)   : 10=49.37%, 20=10.43%, 50=2.19%, 100=33.91%, 250=3.99%
  lat (usec)   : 500=0.08%, 750=0.01%, 1000=0.01%
  lat (msec)   : 2=0.01%, 4=0.01%, 10=0.01%, 20=0.01%, 50=0.01%
  lat (msec)   : 500=0.01%
  cpu          : usr=11.20%, sys=43.61%, ctx=514482, majf=0, minf=0
  IO depths    : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
     submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     issued rwt: total=0,1381888,0, short=0,0,0, dropped=0,0,0
     latency   : target=0, window=0, percentile=100.00%, depth=1

Run status group 0 (all jobs):
  WRITE: bw=360MiB/s (377MB/s), 360MiB/s-360MiB/s (377MB/s-377MB/s), io=21.1GiB (22.6GB), run=60001-60001msec


remember that I also have 750gb of ssd cache.

So I would expect too get higher transfer speeds.
I'm also considering upgrading to 10gbps nic card directly connected beetwen my NAS and pc.

SSD cache is probably irrelevant here because the "cache" devices are for read caching, and even if they were for writes, ZFS separate log devices are for synchronous writes only and not a true "cache" - async writes are already faster than any sync write will ever be.

Upgrading to 10Gbps will do nothing if your pool cannot keep up, so we need to focus on why your disks aren't performing at full throttle. Are you using the ICH10R ports on your server motherboard? The SAS controller is an older LSI1068E which could be hamstringing things. Disable hyper-threading, maybe even remove a single CPU and run from a single socket (3x8GB would be 24GB) to avoid potential NUMA issues?
 

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
Try fio --name=testwrites --rw=write --bs=16k --size=16g --numjobs=1 --iodepth=1 --runtime=60 --time_based just to storm your pool with 16K async writes (which is what your iSCSI ZVOL will do under load) - make sure it can keep up with draining to disk.
Results:
Code:
 
root@truenas[~]# fio --name=testwrites --rw=write --bs=16k --size=16g --numjobs=1 --iodepth=1 --runtime=60 --time_based
testwrites: (g=0): rw=write, bs=(R) 16.0KiB-16.0KiB, (W) 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=psync, iodepth=1
fio-3.28
Starting 1 process
testwrites: Laying out IO file (1 file / 16384MiB)
Jobs: 1 (f=1): [W(1)][100.0%][eta 00m:00s]
testwrites: (groupid=0, jobs=1): err= 0: pid=22883: Fri Sep 16 13:21:45 2022
  write: IOPS=17.7k, BW=276MiB/s (290MB/s)(16.2GiB/60001msec); 0 zone resets
    clat (usec): min=7, max=19510k, avg=55.38, stdev=20521.75
     lat (usec): min=7, max=19510k, avg=55.63, stdev=20521.75
    clat percentiles (usec):
     |  1.00th=[    8],  5.00th=[    8], 10.00th=[    8], 20.00th=[    8],
     | 30.00th=[    8], 40.00th=[    8], 50.00th=[    9], 60.00th=[    9],
     | 70.00th=[   10], 80.00th=[   12], 90.00th=[   32], 95.00th=[   42],
     | 99.00th=[  192], 99.50th=[  221], 99.90th=[  255], 99.95th=[  816],
     | 99.99th=[ 1303]
   bw (  KiB/s): min=24701, max=1323808, per=100.00%, avg=644930.63, stdev=462648.32, samples=51
   iops        : min= 1543, max=82738, avg=40307.75, stdev=28915.53, samples=51
  lat (usec)   : 10=70.80%, 20=14.99%, 50=10.95%, 100=0.97%, 250=2.17%
  lat (usec)   : 500=0.05%, 750=0.01%, 1000=0.01%
  lat (msec)   : 2=0.04%, 4=0.01%, 10=0.01%, 20=0.01%, 50=0.01%
  lat (msec)   : 750=0.01%, 1000=0.01%, >=2000=0.01%
  cpu          : usr=2.99%, sys=22.60%, ctx=20832, majf=0, minf=0
  IO depths    : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
     submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     complete  : 0=0.0%, 4=100.0 


SSD cache is probably irrelevant here because the "cache" devices are for read caching, and even if they were for writes, ZFS separate log devices are for synchronous writes only and not a true "cache" - async writes are already faster than any sync write will ever be.
I was hoping cache would work as a temporary storage device until it has fully transfered the data over to the hdd's.

Upgrading to 10Gbps will do nothing if your pool cannot keep up, so we need to focus on why your disks aren't performing at full throttle.
I have some more hdd's I can put into my NAS.

Are you using the ICH10R ports on your server motherboard?
only for the cache drives, I tried to plug my hdd's into there but it seems like the LSI controller can max handle 2tb drives.

Disable hyper-threading, maybe even remove a single CPU and run from a single socket (3x8GB would be 24GB) to avoid potential NUMA issues?
I only used one prossecor for a while but thought a second would fix the issue(suprise it didn't), and my half of my RAM would be gone.
I can try disabling hyper-threading if you wan't me to?
 

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
Something seems amiss in your fio results, especially this line for your latencies:

Code:
     lat (usec): min=7, max=19510k, avg=55.63, stdev=20521.75


That's telling me that your max latency was 19510K microseconds - that being 19,510ms or around 19.5 seconds.

Reseat your cables to see that they're fastened securely and pull your drives SMART results with smartctl -a /dev/ada0 on through ada3

I was hoping cache would work as a temporary storage device until it has fully transfered the data over to the hdd's.

Unfortunately that isn't how ZFS works, there is no traditional "write cache" that is filled and then tiered down. You could set them as a separate pool and benchmark them with the same lines to see if the performance is better. Good news is that ZFS uses RAM for asynchronous writes which will be much faster than any persistent storage media.

I have some more hdd's I can put into my NAS.

You might be able to get to 500MB/s of writes with eight or ten drives, but I think you have other issues as shown by the latency numbers.

only for the cache drives, I tried to plug my hdd's into there but it seems like the LSI controller can max handle 2tb drives.

You're correct, it's an older SAS1 so it would fail at just over 2TB.

I only used one prossecor for a while but thought a second would fix the issue(suprise it didn't), and my half of my RAM would be gone.
I can try disabling hyper-threading if you wan't me to?

Worth a shot. Disable any power-savings measures in BIOS as well (set to Max Performance/Disabled) and see if that changes latencies.
 

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
You might be able to get to 500MB/s of writes with eight or ten drives, but I think you have other issues as shown by the latency numbers.
I have 14 drives just laying here, I could put them all in my NAS, if the result would be better.

You're correct, it's an older SAS1 so it would fail at just over 2TB.
yep thats sounds correct.

Worth a shot. Disable any power-savings measures in BIOS as well (set to Max Performance/Disabled) and see if that changes latencies.
I don't see a hyper threading, but I see something similar(highlighted in the screenshot) if it's the same as hyper threading?
1663367002488.png


That's telling me that your max latency was 19510K microseconds - that being 19,510ms or around 19.5 seconds.

Reseat your cables to see that they're fastened securely and pull your drives SMART results with smartctl -a /dev/ada0 on through ada3
here are the 4 result's(the list is kinda long):

Code:

root@truenas[~]# smartctl -a /dev/ada0
smartctl 7.2 2021-09-14 r5236 [FreeBSD 13.1-RELEASE-p1 amd64] (local build)
Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF INFORMATION SECTION ===
Model Family:     Western Digital Red
Device Model:     WDC WD40EFRX-68WT0N0
Serial Number:    WD-WCC4E0967267
LU WWN Device Id: 5 0014ee 2b46f7c0f
Firmware Version: 80.00A80
User Capacity:    4,000,787,030,016 bytes [4.00 TB]
Sector Sizes:     512 bytes logical, 4096 bytes physical
Rotation Rate:    5400 rpm
Device is:        In smartctl database [for details use: -P show]
ATA Version is:   ACS-2 (minor revision not indicated)
SATA Version is:  SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is:    Fri Sep 16 15:11:28 2022 PDT
SMART support is: Available - device has SMART capability.
SMART support is: Enabled

=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED

General SMART Values:
Offline data collection status:  (0x00) Offline data collection activity
                                        was never started.
                                        Auto Offline Data Collection: Disabled.
Self-test execution status:      (   0) The previous self-test routine completed
                                        without error or no self-test has ever
                                        been run.
Total time to complete Offline
data collection:                (55260) seconds.
Offline data collection
capabilities:                    (0x7b) SMART execute Offline immediate.
                                        Auto Offline data collection on/off support.
                                        Suspend Offline collection upon new
                                        command.
                                        Offline surface scan supported.
                                        Self-test supported.
                                        Conveyance Self-test supported.
                                        Selective Self-test supported.
SMART capabilities:            (0x0003) Saves SMART data before entering
                                        power-saving mode.
                                        Supports SMART auto save timer.
Error logging capability:        (0x01) Error logging supported.
                                        General Purpose Logging supported.
Short self-test routine
recommended polling time:        (   2) minutes.
Extended self-test routine
recommended polling time:        ( 552) minutes.
Conveyance self-test routine
recommended polling time:        (   5) minutes.
SCT capabilities:              (0x703d) SCT Status supported.
                                        SCT Error Recovery Control supported.
                                        SCT Feature Control supported.
                                        SCT Data Table supported.

SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  WHEN_FAILED RAW_VALUE
  1 Raw_Read_Error_Rate     0x002f   200   200   051    Pre-fail  Always       -       11
  3 Spin_Up_Time            0x0027   181   169   021    Pre-fail  Always       -       7908
  4 Start_Stop_Count        0x0032   100   100   000    Old_age   Always       -       264
  5 Reallocated_Sector_Ct   0x0033   200   200   140    Pre-fail  Always       -       0
  7 Seek_Error_Rate         0x002e   200   200   000    Old_age   Always       -       0
  9 Power_On_Hours          0x0032   006   006   000    Old_age   Always       -       69154
 10 Spin_Retry_Count        0x0032   100   100   000    Old_age   Always       -       0
 11 Calibration_Retry_Count 0x0032   100   100   000    Old_age   Always       -       0
 12 Power_Cycle_Count       0x0032   100   100   000    Old_age   Always       -       260
192 Power-Off_Retract_Count 0x0032   200   200   000    Old_age   Always       -       191
193 Load_Cycle_Count        0x0032   193   193   000    Old_age   Always       -       22830
194 Temperature_Celsius     0x0022   122   104   000    Old_age   Always       -       30
196 Reallocated_Event_Count 0x0032   200   200   000    Old_age   Always       -       0
197 Current_Pending_Sector  0x0032   200   200   000    Old_age   Always       -       0
198 Offline_Uncorrectable   0x0030   100   253   000    Old_age   Offline      -       0
199 UDMA_CRC_Error_Count    0x0032   200   200   000    Old_age   Always       -       0
200 Multi_Zone_Error_Rate   0x0008   100   253   000    Old_age   Offline      -       0

SMART Error Log Version: 1
No Errors Logged

SMART Self-test log structure revision number 1
Num  Test_Description    Status                  Remaining  LifeTime(hours)  LBA_of_first_error
# 1  Short offline       Completed without error       00%      3482         -

SMART Selective self-test log data structure revision number 1
 SPAN  MIN_LBA  MAX_LBA  CURRENT_TEST_STATUS
    1        0        0  Not_testing
    2        0        0  Not_testing


Code:
Selective self-test flags (0x0):
  After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.

root@truenas[~]# smartctl -a /dev/ada1
smartctl 7.2 2021-09-14 r5236 [FreeBSD 13.1-RELEASE-p1 amd64] (local build)
Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF INFORMATION SECTION ===
Model Family:     Western Digital Red
Device Model:     WDC WD40EFRX-68WT0N0
Serial Number:    WD-WCC4E1071370
LU WWN Device Id: 5 0014ee 2b47afc2f
Firmware Version: 80.00A80
User Capacity:    4,000,787,030,016 bytes [4.00 TB]
Sector Sizes:     512 bytes logical, 4096 bytes physical
Rotation Rate:    5400 rpm
Device is:        In smartctl database [for details use: -P show]
ATA Version is:   ACS-2 (minor revision not indicated)
SATA Version is:  SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is:    Fri Sep 16 15:13:24 2022 PDT
SMART support is: Available - device has SMART capability.
SMART support is: Enabled

=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED

General SMART Values:
Offline data collection status:  (0x00) Offline data collection activity
                                        was never started.
                                        Auto Offline Data Collection: Disabled.
Self-test execution status:      (   0) The previous self-test routine completed
                                        without error or no self-test has ever
                                        been run.
Total time to complete Offline
data collection:                (54780) seconds.
Offline data collection
capabilities:                    (0x7b) SMART execute Offline immediate.
                                        Auto Offline data collection on/off support.
                                        Suspend Offline collection upon new
                                        command.
                                        Offline surface scan supported.
                                        Self-test supported.
                                        Conveyance Self-test supported.
                                        Selective Self-test supported.
SMART capabilities:            (0x0003) Saves SMART data before entering
                                        power-saving mode.
                                        Supports SMART auto save timer.
Error logging capability:        (0x01) Error logging supported.
                                        General Purpose Logging supported.
Short self-test routine
recommended polling time:        (   2) minutes.
Extended self-test routine
recommended polling time:        ( 548) minutes.
Conveyance self-test routine
recommended polling time:        (   5) minutes.
SCT capabilities:              (0x703d) SCT Status supported.
                                        SCT Error Recovery Control supported.
                                        SCT Feature Control supported.
                                        SCT Data Table supported.

SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  WHEN_FAILED RAW_VALUE
  1 Raw_Read_Error_Rate     0x002f   200   200   051    Pre-fail  Always       -       0
  3 Spin_Up_Time            0x0027   178   165   021    Pre-fail  Always       -       8083
  4 Start_Stop_Count        0x0032   100   100   000    Old_age   Always       -       254
  5 Reallocated_Sector_Ct   0x0033   200   200   140    Pre-fail  Always       -       0
  7 Seek_Error_Rate         0x002e   200   200   000    Old_age   Always       -       0
  9 Power_On_Hours          0x0032   005   005   000    Old_age   Always       -       69548
 10 Spin_Retry_Count        0x0032   100   100   000    Old_age   Always       -       0
 11 Calibration_Retry_Count 0x0032   100   100   000    Old_age   Always       -       0
 12 Power_Cycle_Count       0x0032   100   100   000    Old_age   Always       -       248
192 Power-Off_Retract_Count 0x0032   200   200   000    Old_age   Always       -       188
193 Load_Cycle_Count        0x0032   193   193   000    Old_age   Always       -       22937
194 Temperature_Celsius     0x0022   123   103   000    Old_age   Always       -       29
196 Reallocated_Event_Count 0x0032   200   200   000    Old_age   Always       -       0
197 Current_Pending_Sector  0x0032   200   200   000    Old_age   Always       -       0
198 Offline_Uncorrectable   0x0030   100   253   000    Old_age   Offline      -       0
199 UDMA_CRC_Error_Count    0x0032   200   200   000    Old_age   Always       -       0
200 Multi_Zone_Error_Rate   0x0008   200   200   000    Old_age   Offline      -       0

SMART Error Log Version: 1
No Errors Logged

SMART Self-test log structure revision number 1

Num  Test_Description    Status                  Remaining  LifeTime(hours)  LBA_of_first_error
# 1  Short offline       Completed without error       00%      3875         -
# 2  Extended offline    Interrupted (host reset)      90%      2827         -
# 3  Extended offline    Aborted by host               90%      2827         -
# 4  Extended offline    Completed without error       00%      2812         -

SMART Selective self-test log data structure revision number 1
 SPAN  MIN_LBA  MAX_LBA  CURRENT_TEST_STATUS
    1        0        0  Not_testing
    2        0        0  Not_testing


Code:
If Selective self-test is pending on power-up, resume after 0 minute delay.

root@truenas[~]# smartctl -a /dev/ada2
smartctl 7.2 2021-09-14 r5236 [FreeBSD 13.1-RELEASE-p1 amd64] (local build)
Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF INFORMATION SECTION ===
Model Family:     Western Digital Red
Device Model:     WDC WD40EFRX-68WT0N0
Serial Number:    WD-WCC4E0972453
LU WWN Device Id: 5 0014ee 209c599d5
Firmware Version: 80.00A80
User Capacity:    4,000,787,030,016 bytes [4.00 TB]
Sector Sizes:     512 bytes logical, 4096 bytes physical
Rotation Rate:    5400 rpm
Device is:        In smartctl database [for details use: -P show]
ATA Version is:   ACS-2 (minor revision not indicated)
SATA Version is:  SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is:    Fri Sep 16 15:14:44 2022 PDT
SMART support is: Available - device has SMART capability.
SMART support is: Enabled

=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED

General SMART Values:
Offline data collection status:  (0x00) Offline data collection activity
                                        was never started.
                                        Auto Offline Data Collection: Disabled.
Self-test execution status:      (   0) The previous self-test routine completed
                                        without error or no self-test has ever
                                        been run.
Total time to complete Offline
data collection:                (50580) seconds.
Offline data collection
capabilities:                    (0x7b) SMART execute Offline immediate.
                                        Auto Offline data collection on/off support.
                                        Suspend Offline collection upon new
                                        command.
                                        Offline surface scan supported.
                                        Self-test supported.
                                        Conveyance Self-test supported.
                                        Selective Self-test supported.
SMART capabilities:            (0x0003) Saves SMART data before entering
                                        power-saving mode.
                                        Supports SMART auto save timer.
Error logging capability:        (0x01) Error logging supported.
                                        General Purpose Logging supported.
Short self-test routine
recommended polling time:        (   2) minutes.
Extended self-test routine
recommended polling time:        ( 506) minutes.
Conveyance self-test routine
recommended polling time:        (   5) minutes.
SCT capabilities:              (0x703d) SCT Status supported.
                                        SCT Error Recovery Control supported.
                                        SCT Feature Control supported.
                                        SCT Data Table supported.

SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  WHEN_FAILED RAW_VALUE
  1 Raw_Read_Error_Rate     0x002f   200   200   051    Pre-fail  Always       -       13
  3 Spin_Up_Time            0x0027   177   166   021    Pre-fail  Always       -       8116
  4 Start_Stop_Count        0x0032   100   100   000    Old_age   Always       -       284
  5 Reallocated_Sector_Ct   0x0033   200   200   140    Pre-fail  Always       -       0
  7 Seek_Error_Rate         0x002e   200   200   000    Old_age   Always       -       0
  9 Power_On_Hours          0x0032   006   006   000    Old_age   Always       -       69081
 10 Spin_Retry_Count        0x0032   100   100   000    Old_age   Always       -       0
 11 Calibration_Retry_Count 0x0032   100   100   000    Old_age   Always       -       0
 12 Power_Cycle_Count       0x0032   100   100   000    Old_age   Always       -       251
192 Power-Off_Retract_Count 0x0032   200   200   000    Old_age   Always       -       189
193 Load_Cycle_Count        0x0032   193   193   000    Old_age   Always       -       22842
194 Temperature_Celsius     0x0022   122   104   000    Old_age   Always       -       30
196 Reallocated_Event_Count 0x0032   200   200   000    Old_age   Always       -       0
197 Current_Pending_Sector  0x0032   200   200   000    Old_age   Always       -       0
198 Offline_Uncorrectable   0x0030   100   253   000    Old_age   Offline      -       0
199 UDMA_CRC_Error_Count    0x0032   200   200   000    Old_age   Always       -       0
200 Multi_Zone_Error_Rate   0x0008   100   253   000    Old_age   Offline      -       0

SMART Error Log Version: 1
No Errors Logged

SMART Self-test log structure revision number 1
Num  Test_Description    Status                  Remaining  LifeTime(hours)  LBA_of_first_error
# 1  Short offline       Completed without error       00%      3408         -

SMART Selective self-test log data structure revision number 1
 SPAN  MIN_LBA  MAX_LBA  CURRENT_TEST_STATUS
    1        0        0  Not_testing
    2        0        0  Not_testing


Code:

root@truenas[~]# smartctl -a /dev/ada3
smartctl 7.2 2021-09-14 r5236 [FreeBSD 13.1-RELEASE-p1 amd64] (local build)
Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF INFORMATION SECTION ===
Model Family:     Western Digital Red
Device Model:     WDC WD40EFRX-68WT0N0
Serial Number:    WD-WCC4E1042632
LU WWN Device Id: 5 0014ee 2b47afca3
Firmware Version: 80.00A80
User Capacity:    4,000,787,030,016 bytes [4.00 TB]
Sector Sizes:     512 bytes logical, 4096 bytes physical
Rotation Rate:    5400 rpm
Device is:        In smartctl database [for details use: -P show]
ATA Version is:   ACS-2 (minor revision not indicated)
SATA Version is:  SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is:    Fri Sep 16 15:15:44 2022 PDT
SMART support is: Available - device has SMART capability.
SMART support is: Enabled

=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED

General SMART Values:
Offline data collection status:  (0x00) Offline data collection activity
                                        was never started.
                                        Auto Offline Data Collection: Disabled.
Self-test execution status:      (   0) The previous self-test routine completed
                                        without error or no self-test has ever
                                        been run.
Total time to complete Offline
data collection:                (52680) seconds.
Offline data collection
capabilities:                    (0x7b) SMART execute Offline immediate.
                                        Auto Offline data collection on/off support.
                                        Suspend Offline collection upon new
                                        command.
                                        Offline surface scan supported.
                                        Self-test supported.
                                        Conveyance Self-test supported.
                                        Selective Self-test supported.
SMART capabilities:            (0x0003) Saves SMART data before entering
                                        power-saving mode.
                                        Supports SMART auto save timer.
Error logging capability:        (0x01) Error logging supported.
                                        General Purpose Logging supported.
Short self-test routine
recommended polling time:        (   2) minutes.
Extended self-test routine
recommended polling time:        ( 527) minutes.
Conveyance self-test routine
recommended polling time:        (   5) minutes.
SCT capabilities:              (0x703d) SCT Status supported.
                                        SCT Error Recovery Control supported.
                                        SCT Feature Control supported.
                                        SCT Data Table supported.

SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  WHEN_FAILED RAW_VALUE
  1 Raw_Read_Error_Rate     0x002f   200   200   051    Pre-fail  Always       -       1
  3 Spin_Up_Time            0x0027   176   163   021    Pre-fail  Always       -       8166
  4 Start_Stop_Count        0x0032   100   100   000    Old_age   Always       -       280
  5 Reallocated_Sector_Ct   0x0033   200   200   140    Pre-fail  Always       -       0
  7 Seek_Error_Rate         0x002e   200   200   000    Old_age   Always       -       0
  9 Power_On_Hours          0x0032   006   006   000    Old_age   Always       -       69131
 10 Spin_Retry_Count        0x0032   100   100   000    Old_age   Always       -       0
 11 Calibration_Retry_Count 0x0032   100   100   000    Old_age   Always       -       0
 12 Power_Cycle_Count       0x0032   100   100   000    Old_age   Always       -       253
192 Power-Off_Retract_Count 0x0032   200   200   000    Old_age   Always       -       190
193 Load_Cycle_Count        0x0032   193   193   000    Old_age   Always       -       22989
194 Temperature_Celsius     0x0022   122   104   000    Old_age   Always       -       30
196 Reallocated_Event_Count 0x0032   200   200   000    Old_age   Always       -       0
197 Current_Pending_Sector  0x0032   200   200   000    Old_age   Always       -       0
198 Offline_Uncorrectable   0x0030   100   253   000    Old_age   Offline      -       0
199 UDMA_CRC_Error_Count    0x0032   200   200   000    Old_age   Always       -       0
200 Multi_Zone_Error_Rate   0x0008   100   253   000    Old_age   Offline      -       0

SMART Error Log Version: 1
No Errors Logged

SMART Self-test log structure revision number 1
No self-tests have been logged.  [To run self-tests, use: smartctl -t]

SMART Selective self-test log data structure revision number 1
 SPAN  MIN_LBA  MAX_LBA  CURRENT_TEST_STATUS
    1        0        0  Not_testing
    2        0        0  Not_testing
 
Last edited:

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
I have 14 drives just laying here, I could put them all in my NAS, if the result would be better.

If the problem is purely in disk throughput then that would solve it - but if the issue lies elsewhere, it won't help.

Question - is the ICH10R in AHCI mode in the BIOS? Ensure that it's not in IDE or RAID. Another suspect might be drive write cache being disabled on the SATA ports somehow (a possibility if the controller is trying to pretend to do RAID)

I don't see a hyper threading, but I see something similar(highlighted in the screenshot) if it's the same as hyper threading?

SMT is a generic term, that's the one to try disabling.

here are the 4 result's(the list is kinda long):
(snipped)

No errors showing, but no tests run either. A SMART long test against each drive might be a good place to start.
 

Ikon

Dabbler
Joined
Aug 26, 2022
Messages
49
Question - is the ICH10R in AHCI mode in the BIOS? Ensure that it's not in IDE or RAID. Another suspect might be drive write cache being disabled on the SATA ports somehow (a possibility if the controller is trying to pretend to do RAID)
Im pretty sure it's in IDE mode, I'm gonna try AHCI mode..
EDIT: I have the option beetwen: Intel AHCI ROM and BIOS Native Module, wich one?


SMT is a generic term, that's the one to try disabling.
will do.
EDIT: it did not solve my problem, it just made my pc freeze earlier and half of my core's are missing in the truenas UI
 
Last edited:
Top