[root@FreeNAS /mnt/WDVolume_A/Data/scripts]# bash arc_stats.sh arc_stats.sh: line 2: $'\r': command not found arc_stats.sh: line 5: $'\r': command not found
And you shouldn't use bash arc_stats.sh but just ./arc_stats.sh ;)
[root@FreeNAS /mnt/WDVolume_A/Data/scripts]# ./arc_stats.sh bash: ./arc_stats.sh: Permission denied
Guessing you pasted on a Windows system? There's extraneous \r characters (carriage return) mixed in with your \n's (newline)...
Try:
# tr -d '\015' < arc_stats.sh > arc_stats_fixed.sh
Now it's working! I'm coming from OSX system! Can you explain to me what have you change with this command?
"tr" - UNIX translate characters
"-d" - don't translate, rather, delete
"'\015'" - octal code for carriage return
"<" - input redirect
">" - output redirect
The owner of the dataset is root and i'm try to execute it with root user... :/
arc_summary.py System Memory: 0.61% 192.87 MiB Active, 4.20% 1.31 GiB Inact 93.95% 29.22 GiB Wired, 0.01% 3.33 MiB Cache 1.23% 393.14 MiB Free, 0.00% 0 Bytes Gap Real Installed: 32.00 GiB Real Available: 99.79% 31.93 GiB Real Managed: 97.41% 31.10 GiB Logical Total: 32.00 GiB Logical Used: 94.71% 30.31 GiB Logical Free: 5.29% 1.69 GiB Kernel Memory: 395.89 MiB Data: 93.21% 369.02 MiB Text: 6.79% 26.88 MiB Kernel Memory Map: 31.10 GiB Size: 88.37% 27.49 GiB Free: 11.63% 3.62 GiB Page: 1 ------------------------------------------------------------------------ ARC Summary: (HEALTHY) Storage pool Version: 5000 Filesystem Version: 5 Memory Throttle Count: 0 ARC Misc: Deleted: 15.14m Mutex Misses: 2.81k Evict Skips: 2.81k ARC Size: 88.10% 26.52 GiB Target Size: (Adaptive) 88.18% 26.55 GiB Min Size (Hard Limit): 12.50% 3.76 GiB Max Size (High Water): 8:1 30.10 GiB ARC Size Breakdown: Recently Used Cache Size: 63.60% 16.88 GiB Frequently Used Cache Size: 36.40% 9.66 GiB ARC Hash Breakdown: Elements Max: 658.06k Elements Current: 99.13% 652.30k Collisions: 3.77m Chain Max: 6 Chains: 45.69k Page: 2 ------------------------------------------------------------------------ ARC Total accesses: 459.71m Cache Hit Ratio: 63.90% 293.77m Cache Miss Ratio: 36.10% 165.94m Actual Hit Ratio: 62.04% 285.22m Data Demand Efficiency: 80.88% 271.41m Data Prefetch Efficiency: 48.70% 16.93m CACHE HITS BY CACHE LIST: Anonymously Used: 2.73% 8.03m Most Recently Used: 9.89% 29.04m Most Frequently Used: 87.20% 256.18m Most Recently Used Ghost: 0.10% 300.29k Most Frequently Used Ghost: 0.08% 222.67k CACHE HITS BY DATA TYPE: Demand Data: 74.73% 219.52m Prefetch Data: 2.81% 8.25m Demand Metadata: 22.34% 65.64m Prefetch Metadata: 0.13% 367.23k CACHE MISSES BY DATA TYPE: Demand Data: 31.27% 51.89m Prefetch Data: 5.23% 8.68m Demand Metadata: 63.36% 105.15m Prefetch Metadata: 0.13% 214.72k Page: 3 ------------------------------------------------------------------------ Page: 4 ------------------------------------------------------------------------ DMU Prefetch Efficiency: 7.99b Hit Ratio: 0.45% 35.97m Miss Ratio: 99.55% 7.96b Page: 5 ------------------------------------------------------------------------ Page: 6 ------------------------------------------------------------------------ ZFS Tunable (sysctl): kern.maxusers 2379 vm.kmem_size 33397690368 vm.kmem_size_scale 1 vm.kmem_size_min 0 vm.kmem_size_max 1319413950874 vfs.zfs.vol.unmap_enabled 1 vfs.zfs.vol.mode 2 vfs.zfs.sync_pass_rewrite 2 vfs.zfs.sync_pass_dont_compress 5 vfs.zfs.sync_pass_deferred_free 2 vfs.zfs.zio.exclude_metadata 0 vfs.zfs.zio.use_uma 1 vfs.zfs.cache_flush_disable 0 vfs.zfs.zil_replay_disable 0 vfs.zfs.version.zpl 5 vfs.zfs.version.spa 5000 vfs.zfs.version.acl 1 vfs.zfs.version.ioctl 6 vfs.zfs.debug 0 vfs.zfs.super_owner 0 vfs.zfs.min_auto_ashift 9 vfs.zfs.max_auto_ashift 13 vfs.zfs.vdev.write_gap_limit 4096 vfs.zfs.vdev.read_gap_limit 32768 vfs.zfs.vdev.aggregation_limit 131072 vfs.zfs.vdev.trim_max_active 64 vfs.zfs.vdev.trim_min_active 1 vfs.zfs.vdev.scrub_max_active 2 vfs.zfs.vdev.scrub_min_active 1 vfs.zfs.vdev.async_write_max_active 10 vfs.zfs.vdev.async_write_min_active 1 vfs.zfs.vdev.async_read_max_active 3 vfs.zfs.vdev.async_read_min_active 1 vfs.zfs.vdev.sync_write_max_active 10 vfs.zfs.vdev.sync_write_min_active 10 vfs.zfs.vdev.sync_read_max_active 10 vfs.zfs.vdev.sync_read_min_active 10 vfs.zfs.vdev.max_active 1000 vfs.zfs.vdev.async_write_active_max_dirty_percent60 vfs.zfs.vdev.async_write_active_min_dirty_percent30 vfs.zfs.vdev.mirror.non_rotating_seek_inc1 vfs.zfs.vdev.mirror.non_rotating_inc 0 vfs.zfs.vdev.mirror.rotating_seek_offset1048576 vfs.zfs.vdev.mirror.rotating_seek_inc 5 vfs.zfs.vdev.mirror.rotating_inc 0 vfs.zfs.vdev.trim_on_init 1 vfs.zfs.vdev.larger_ashift_minimal 0 vfs.zfs.vdev.bio_delete_disable 0 vfs.zfs.vdev.bio_flush_disable 0 vfs.zfs.vdev.cache.bshift 16 vfs.zfs.vdev.cache.size 0 vfs.zfs.vdev.cache.max 16384 vfs.zfs.vdev.metaslabs_per_vdev 200 vfs.zfs.vdev.trim_max_pending 10000 vfs.zfs.txg.timeout 5 vfs.zfs.trim.enabled 1 vfs.zfs.trim.max_interval 1 vfs.zfs.trim.timeout 30 vfs.zfs.trim.txg_delay 32 vfs.zfs.space_map_blksz 4096 vfs.zfs.spa_slop_shift 5 vfs.zfs.spa_asize_inflation 24 vfs.zfs.deadman_enabled 1 vfs.zfs.deadman_checktime_ms 5000 vfs.zfs.deadman_synctime_ms 1000000 vfs.zfs.debug_flags 0 vfs.zfs.recover 0 vfs.zfs.spa_load_verify_data 1 vfs.zfs.spa_load_verify_metadata 1 vfs.zfs.spa_load_verify_maxinflight 10000 vfs.zfs.ccw_retry_interval 300 vfs.zfs.check_hostid 1 vfs.zfs.mg_fragmentation_threshold 85 vfs.zfs.mg_noalloc_threshold 0 vfs.zfs.condense_pct 200 vfs.zfs.metaslab.bias_enabled 1 vfs.zfs.metaslab.lba_weighting_enabled 1 vfs.zfs.metaslab.fragmentation_factor_enabled1 vfs.zfs.metaslab.preload_enabled 1 vfs.zfs.metaslab.preload_limit 3 vfs.zfs.metaslab.unload_delay 8 vfs.zfs.metaslab.load_pct 50 vfs.zfs.metaslab.min_alloc_size 33554432 vfs.zfs.metaslab.df_free_pct 4 vfs.zfs.metaslab.df_alloc_threshold 131072 vfs.zfs.metaslab.debug_unload 0 vfs.zfs.metaslab.debug_load 0 vfs.zfs.metaslab.fragmentation_threshold70 vfs.zfs.metaslab.gang_bang 16777217 vfs.zfs.free_bpobj_enabled 1 vfs.zfs.free_max_blocks 18446744073709551615 vfs.zfs.no_scrub_prefetch 0 vfs.zfs.no_scrub_io 0 vfs.zfs.resilver_min_time_ms 3000 vfs.zfs.free_min_time_ms 1000 vfs.zfs.scan_min_time_ms 1000 vfs.zfs.scan_idle 50 vfs.zfs.scrub_delay 4 vfs.zfs.resilver_delay 2 vfs.zfs.top_maxinflight 32 vfs.zfs.delay_scale 500000 vfs.zfs.delay_min_dirty_percent 60 vfs.zfs.dirty_data_sync 67108864 vfs.zfs.dirty_data_max_percent 10 vfs.zfs.dirty_data_max_max 4294967296 vfs.zfs.dirty_data_max 3428664115 vfs.zfs.max_recordsize 1048576 vfs.zfs.zfetch.array_rd_sz 1048576 vfs.zfs.zfetch.max_distance 8388608 vfs.zfs.zfetch.min_sec_reap 2 vfs.zfs.zfetch.max_streams 8 vfs.zfs.prefetch_disable 0 vfs.zfs.mdcomp_disable 0 vfs.zfs.nopwrite_enabled 1 vfs.zfs.dedup.prefetch 1 vfs.zfs.l2c_only_size 0 vfs.zfs.mfu_ghost_data_lsize 15740569600 vfs.zfs.mfu_ghost_metadata_lsize 360044032 vfs.zfs.mfu_ghost_size 16100613632 vfs.zfs.mfu_data_lsize 11272137728 vfs.zfs.mfu_metadata_lsize 674260992 vfs.zfs.mfu_size 12361233408 vfs.zfs.mru_ghost_data_lsize 10592845824 vfs.zfs.mru_ghost_metadata_lsize 1808963072 vfs.zfs.mru_ghost_size 12401808896 vfs.zfs.mru_data_lsize 15186720768 vfs.zfs.mru_metadata_lsize 46192640 vfs.zfs.mru_size 15576815616 vfs.zfs.anon_data_lsize 0 vfs.zfs.anon_metadata_lsize 0 vfs.zfs.anon_size 3983360 vfs.zfs.l2arc_norw 1 vfs.zfs.l2arc_feed_again 1 vfs.zfs.l2arc_noprefetch 1 vfs.zfs.l2arc_feed_min_ms 200 vfs.zfs.l2arc_feed_secs 1 vfs.zfs.l2arc_headroom 2 vfs.zfs.l2arc_write_boost 8388608 vfs.zfs.l2arc_write_max 8388608 vfs.zfs.arc_meta_limit 8080987136 vfs.zfs.arc_free_target 56573 vfs.zfs.arc_shrink_shift 7 vfs.zfs.arc_average_blocksize 8192 vfs.zfs.arc_min 4040493568 vfs.zfs.arc_max 32323948544 Page: 7 ------------------------------------------------------------------------
Well, you have a pool that's bigger than 100 TB and only 32 GB of RAM, that and the fact the stats are worse with time tell me you don't have enough RAM.
Your stats aren't good but they're not critical either, you can live with that if the perfs are ok for you as is ;)
Is it possible to make arc not to cache Files larger than 3GB ?
Should I turn autotune on?
Would I benefit from L2ARC?
My specs
FreeNAS 9.10
MOBO: SuperMicro X10SL7-F
CPU: Intel(R) Xeon(R) CPU E3-1226 v3 @ 3.30GHz
RAM: 32 GB ECC (KVR1333D3E9SK2/16G)
HBA: 2x IBM M1015 (IT)
STORAGE: 2x 500SSD (Mirror)
STORAGE: 24 x 5TB WD RED (4 x 6xRAID-Z2)
BOOT: 2x16GB Sandisk Cruzer Fit USB 2.0
PSU: Seasonic SS-660XP2
UPS: Cyberpower CP1500AVR
Well my machine has been rock solid past year. WD REDs seem to be not as power hungry.I'd be holding my breath every time I had to reboot that, given the number of disks and the PSU :o
I've got a similarly spec'ed machine (m/board, cpu, ram) with the same PSU but less than half the number of drives you're running.
I agree, my plan is to upgrade this server to supermicro chassis next year.Ah yeah, didn't noticed it but I confirm, you definitely want to do something about that. It's not because you don't have problems now that you'll not have any tomorrow and/or that you don't abuse the PSU (and being a good quality PSU it didn't died on you for now but still not a good idea).
The problem is mainly the spin-up current, it's huge, something like 2.5 A for the WD drives and 3 A for the Seagate drives (more info if your want) so with 24 drives it's around 700 W for the WDs and 850 W for the Seagates and that's just for the drives, you need to add all the others components in your system, which should be something like 100 W just eyeballing it. So, yeah, I'd not use anything under 850-900 W on this server.
In the end you do what you want but I really encourage you to look at this before it's too late.