Hi.
This is the second time I've checked into my FreeNAS box (running FreeNAS-9.2.1.2-RELEASE-x64) and found that the CPU utilization was at 100%...
...had been at 100% for several days...
Based on some of the other posts that have this kind of problem, I've collected the following data:
Letting 'zpool iostat 1' run for a few seconds:
And the cause of the 100% CPU utilization (the full output is available here):
The hardware specs are as follows:
CPU
Intel Core i3-4130
Motherboard
SUPERMICRO MBD-X10SLM-F-O
RAM
16GB Crucial DDR3 SDRAM ECC Unbuffered DDR3 1600 (CT2KIT51272BD160B)
HDD
6x HGST/Hitachi Ultrastar 7K3000 2TB (HUA723020ALA641) in a ZRAID2 configuration
I have ZFS Scrubs scheduled for 4am on the 1st and 15th day of the month, every month, and every day except Friday and Saturday, with a 10 day threshold.
I have a recursive periodic snapshot scheduled for every 6 hours and available to run at any time during the day, every day of the week. This snapshot is then replicated to another FreeNAS box (also 9.2.1.2, same HW config, except only 8GB of RAM, and 6x 3TB WD Reds).
I have a long SMART test scheduled for all disks at 4am on the 8th and 22nd of the month, every month, and every day. The short tests are scheduled for 2am, every other day of the month, every day of the week.
So...any ideas what keeps causing this problem? My guess is that it's something related to the scrub, though I'm not sure what or why.
This is the second time I've checked into my FreeNAS box (running FreeNAS-9.2.1.2-RELEASE-x64) and found that the CPU utilization was at 100%...

...had been at 100% for several days...

Based on some of the other posts that have this kind of problem, I've collected the following data:
Code:
badass@freenas_1:~ % zpool status pool: volume_0 state: ONLINE scan: scrub repaired 0 in 2h41m with 0 errors on Sun Mar 23 06:41:20 2014 config: NAME STATE READ WRITE CKSUM volume_0 ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 gptid/6a557eb9-a100-11e3-827a-002590d97629 ONLINE 0 0 0 gptid/6aaa0492-a100-11e3-827a-002590d97629 ONLINE 0 0 0 gptid/6afdd223-a100-11e3-827a-002590d97629 ONLINE 0 0 0 gptid/6b50e62f-a100-11e3-827a-002590d97629 ONLINE 0 0 0 gptid/6ba1e0ae-a100-11e3-827a-002590d97629 ONLINE 0 0 0 gptid/6bf4aa35-a100-11e3-827a-002590d97629 ONLINE 0 0 0 errors: No known data errors
Code:
badass@freenas_1:~ % sudo gstat dT: 1.001s w: 1.000s L(q) ops/s r/s kBps ms/r w/s kBps ms/w %busy Name 0 0 0 0 0.0 0 0 0.0 0.0| ada0 0 0 0 0 0.0 0 0 0.0 0.0| ada0p1 0 0 0 0 0.0 0 0 0.0 0.0| ada0p2 0 0 0 0 0.0 0 0 0.0 0.0| ada1 0 0 0 0 0.0 0 0 0.0 0.0| ada2 0 0 0 0 0.0 0 0 0.0 0.0| ada3 0 0 0 0 0.0 0 0 0.0 0.0| ada4 0 0 0 0 0.0 0 0 0.0 0.0| ada5 0 0 0 0 0.0 0 0 0.0 0.0| ada1p1.eli 0 0 0 0 0.0 0 0 0.0 0.0| gptid/6a557eb9-a100-11e3-827a-002590d97629 0 0 0 0 0.0 0 0 0.0 0.0| ada1p1 0 0 0 0 0.0 0 0 0.0 0.0| ada1p2 0 0 0 0 0.0 0 0 0.0 0.0| ada2p1 0 0 0 0 0.0 0 0 0.0 0.0| ada2p2 0 0 0 0 0.0 0 0 0.0 0.0| ada3p1 0 0 0 0 0.0 0 0 0.0 0.0| ada3p2 0 0 0 0 0.0 0 0 0.0 0.0| ada4p1 0 0 0 0 0.0 0 0 0.0 0.0| ada4p2 0 0 0 0 0.0 0 0 0.0 0.0| ada5p1 0 0 0 0 0.0 0 0 0.0 0.0| ada5p2 0 0 0 0 0.0 0 0 0.0 0.0| ada2p1.eli 0 0 0 0 0.0 0 0 0.0 0.0| gptid/6aaa0492-a100-11e3-827a-002590d97629 0 0 0 0 0.0 0 0 0.0 0.0| ada3p1.eli 0 0 0 0 0.0 0 0 0.0 0.0| gptid/6afdd223-a100-11e3-827a-002590d97629 0 0 0 0 0.0 0 0 0.0 0.0| ada4p1.eli 0 0 0 0 0.0 0 0 0.0 0.0| gptid/6b50e62f-a100-11e3-827a-002590d97629 0 0 0 0 0.0 0 0 0.0 0.0| ada5p1.eli 0 0 0 0 0.0 0 0 0.0 0.0| gptid/6ba1e0ae-a100-11e3-827a-002590d97629 0 0 0 0 0.0 0 0 0.0 0.0| gptid/6bf4aa35-a100-11e3-827a-002590d97629 0 0 0 0 0.0 0 0 0.0 0.0| da0 0 0 0 0 0.0 0 0 0.0 0.0| da0s1 0 0 0 0 0.0 0 0 0.0 0.0| da0s2
Letting 'zpool iostat 1' run for a few seconds:
Code:
badass@freenas_1:~ % zpool iostat 1 capacity operations bandwidth pool alloc free read write read write ---------- ----- ----- ----- ----- ----- ----- volume_0 4.71T 6.16T 2 12 245K 319K volume_0 4.71T 6.16T 0 210 0 1.17M volume_0 4.71T 6.16T 0 0 0 0 volume_0 4.71T 6.16T 0 0 0 0 volume_0 4.71T 6.16T 0 0 0 0 volume_0 4.71T 6.16T 0 0 0 12.0K volume_0 4.71T 6.16T 0 180 0 847K volume_0 4.71T 6.16T 0 0 0 0 volume_0 4.71T 6.16T 0 0 0 0 volume_0 4.71T 6.16T 0 0 0 0 volume_0 4.71T 6.16T 0 0 0 0 volume_0 4.71T 6.16T 0 180 0 847K volume_0 4.71T 6.16T 0 0 0 0 volume_0 4.71T 6.16T 0 0 0 0
And the cause of the 100% CPU utilization (the full output is available here):
Code:
badass@freenas_1:~ % ps auxH USER PID %CPU %MEM VSZ RSS TT STAT STARTED TIME COMMAND root 25427 100.0 0.6 127812 94084 ?? R Wed04AM 2326:35.16 zpool history volume_0 root 27963 100.0 0.6 127812 94084 ?? R Mon04AM 5205:01.55 zpool history volume_0 root 75495 100.0 0.6 127812 94084 ?? R 4:00AM 894:33.66 zpool history volume_0 root 76201 98.8 0.6 127812 94084 ?? R Tue04AM 3765:31.39 zpool history volume_0
The hardware specs are as follows:
CPU
Intel Core i3-4130
Motherboard
SUPERMICRO MBD-X10SLM-F-O
RAM
16GB Crucial DDR3 SDRAM ECC Unbuffered DDR3 1600 (CT2KIT51272BD160B)
HDD
6x HGST/Hitachi Ultrastar 7K3000 2TB (HUA723020ALA641) in a ZRAID2 configuration
I have ZFS Scrubs scheduled for 4am on the 1st and 15th day of the month, every month, and every day except Friday and Saturday, with a 10 day threshold.
I have a recursive periodic snapshot scheduled for every 6 hours and available to run at any time during the day, every day of the week. This snapshot is then replicated to another FreeNAS box (also 9.2.1.2, same HW config, except only 8GB of RAM, and 6x 3TB WD Reds).
I have a long SMART test scheduled for all disks at 4am on the 8th and 22nd of the month, every month, and every day. The short tests are scheduled for 2am, every other day of the month, every day of the week.
So...any ideas what keeps causing this problem? My guess is that it's something related to the scrub, though I'm not sure what or why.