Kevin Horton
Guru
- Joined
 - Dec 2, 2015
 
- Messages
 - 730
 
I'm pondering adding an SSD for L2ARC dedicated to metadata to hopefully speed up rsync.  I use rsync to back up to two different pairs of striped disks for offsite backup (I also have two automatically replicated RAIDZ2 onsite backups).  The data is mostly static, and rsync takes many hours to determine which files have changed since the last backup.  @Constantin reported a very significant improvement in rsync perf, and I hope my use case is similar enough to get a worthwhile benefit.
Current HW:
Motherboard: Supermicro X10SRH-cF
CPU: Intel E5-1650v4
RAM: 32GB
Storage: main and backkup pools, each with 8 x 4TB WD Red, in RAIDZ2
Jails: Plex, homebridge and nextcloud, none of which are used heavily
Server use-case; static data storage, Apple Time Machine backup destination, run Plex, homebridge and Nextcloud servers
I'm looking for any recommendations on suitable SSDs for L2ARC. I'm also pondering whether I would need to buy more RAM. My current plan is to try the L2ARC with my current 32GB of RAM, and increase that to 64GB (or more) if I see signs that RAM has become a bottleneck.
	
		
			
		
		
	
			
			Current HW:
Motherboard: Supermicro X10SRH-cF
CPU: Intel E5-1650v4
RAM: 32GB
Storage: main and backkup pools, each with 8 x 4TB WD Red, in RAIDZ2
Jails: Plex, homebridge and nextcloud, none of which are used heavily
Server use-case; static data storage, Apple Time Machine backup destination, run Plex, homebridge and Nextcloud servers
Code:
arc_summary.py
System Memory:
        1.00%   317.91  MiB Active,     14.24%  4.41    GiB Inact
        81.79%  25.36   GiB Wired,      0.00%   0       Bytes Cache
        2.97%   942.17  MiB Free,       -0.00%  -221184 Bytes Gap
        Real Installed:                         32.00   GiB
        Real Available:                 99.53%  31.85   GiB
        Real Managed:                   97.34%  31.00   GiB
        Logical Total:                          32.00   GiB
        Logical Used:                   83.33%  26.67   GiB
        Logical Free:                   16.67%  5.33    GiB
Kernel Memory:                                  673.78  MiB
        Data:                           93.72%  631.49  MiB
        Text:                           6.28%   42.29   MiB
Kernel Memory Map:                              31.00   GiB
        Size:                           7.71%   2.39    GiB
        Free:                           92.29%  28.61   GiB
                                                                Page:  1
------------------------------------------------------------------------
ARC Summary: (HEALTHY)
        Storage pool Version:                   5000
        Filesystem Version:                     5
        Memory Throttle Count:                  0
ARC Misc:
        Deleted:                                29.96m
        Mutex Misses:                           30.95k
        Evict Skips:                            30.95k
ARC Size:                               56.24%  16.88   GiB
        Target Size: (Adaptive)         56.10%  16.83   GiB
        Min Size (Hard Limit):          12.50%  3.75    GiB
        Max Size (High Water):          8:1     30.00   GiB
ARC Size Breakdown:
        Recently Used Cache Size:       44.81%  7.56    GiB
        Frequently Used Cache Size:     55.19%  9.31    GiB
ARC Hash Breakdown:
        Elements Max:                           2.25m
        Elements Current:               56.09%  1.26m
        Collisions:                             14.41m
        Chain Max:                              7
        Chains:                                 156.31k
                                                                Page:  2
------------------------------------------------------------------------
ARC Total accesses:                                     2.43b
        Cache Hit Ratio:                94.82%  2.31b
        Cache Miss Ratio:               5.18%   126.04m
        Actual Hit Ratio:               94.58%  2.30b
        Data Demand Efficiency:         88.82%  51.34m
        Data Prefetch Efficiency:       2.56%   15.98m
        CACHE HITS BY CACHE LIST:
          Anonymously Used:             0.15%   3.55m
          Most Recently Used:           8.15%   188.15m
          Most Frequently Used:         91.59%  2.11b
          Most Recently Used Ghost:     0.09%   1.97m
          Most Frequently Used Ghost:   0.02%   461.47k
        CACHE HITS BY DATA TYPE:
          Demand Data:                  1.98%   45.60m
          Prefetch Data:                0.02%   409.43k
          Demand Metadata:              97.59%  2.25b
          Prefetch Metadata:            0.42%   9.73m
        CACHE MISSES BY DATA TYPE:
          Demand Data:                  4.55%   5.74m
          Prefetch Data:                12.35%  15.57m
          Demand Metadata:              79.42%  100.10m
          Prefetch Metadata:            3.67%   4.63m
                                                                Page:  3
------------------------------------------------------------------------
                                                                Page:  4
------------------------------------------------------------------------
DMU Prefetch Efficiency:                        516.21m
        Hit Ratio:                      2.03%   10.50m
        Miss Ratio:                     97.97%  505.71m
                                                                Page:  5
------------------------------------------------------------------------
                                                                Page:  6
------------------------------------------------------------------------
ZFS Tunable (sysctl):
        kern.maxusers                           2374
        vm.kmem_size                            33290891264
        vm.kmem_size_scale                      1
        vm.kmem_size_min                        0
        vm.kmem_size_max                        1319413950874
        vfs.zfs.vol.immediate_write_sz          32768
        vfs.zfs.vol.unmap_sync_enabled          0
        vfs.zfs.vol.unmap_enabled               1
        vfs.zfs.vol.recursive                   0
        vfs.zfs.vol.mode                        2
        vfs.zfs.sync_pass_rewrite               2
        vfs.zfs.sync_pass_dont_compress         5
        vfs.zfs.sync_pass_deferred_free         2
        vfs.zfs.zio.dva_throttle_enabled        1
        vfs.zfs.zio.exclude_metadata            0
        vfs.zfs.zio.use_uma                     1
        vfs.zfs.zil_slog_bulk                   786432
        vfs.zfs.cache_flush_disable             0
        vfs.zfs.zil_replay_disable              0
        vfs.zfs.version.zpl                     5
        vfs.zfs.version.spa                     5000
        vfs.zfs.version.acl                     1
        vfs.zfs.version.ioctl                   7
        vfs.zfs.debug                           0
        vfs.zfs.super_owner                     0
        vfs.zfs.immediate_write_sz              32768
        vfs.zfs.standard_sm_blksz               131072
        vfs.zfs.dtl_sm_blksz                    4096
        vfs.zfs.min_auto_ashift                 12
        vfs.zfs.max_auto_ashift                 13
        vfs.zfs.vdev.queue_depth_pct            1000
        vfs.zfs.vdev.write_gap_limit            4096
        vfs.zfs.vdev.read_gap_limit             32768
        vfs.zfs.vdev.aggregation_limit_non_rotating131072
        vfs.zfs.vdev.aggregation_limit          1048576
        vfs.zfs.vdev.trim_max_active            64
        vfs.zfs.vdev.trim_min_active            1
        vfs.zfs.vdev.scrub_max_active           2
        vfs.zfs.vdev.scrub_min_active           1
        vfs.zfs.vdev.async_write_max_active     10
        vfs.zfs.vdev.async_write_min_active     1
        vfs.zfs.vdev.async_read_max_active      3
        vfs.zfs.vdev.async_read_min_active      1
        vfs.zfs.vdev.sync_write_max_active      10
        vfs.zfs.vdev.sync_write_min_active      10
        vfs.zfs.vdev.sync_read_max_active       10
        vfs.zfs.vdev.sync_read_min_active       10
        vfs.zfs.vdev.max_active                 1000
        vfs.zfs.vdev.async_write_active_max_dirty_percent60
        vfs.zfs.vdev.async_write_active_min_dirty_percent30
        vfs.zfs.vdev.mirror.non_rotating_seek_inc1
        vfs.zfs.vdev.mirror.non_rotating_inc    0
        vfs.zfs.vdev.mirror.rotating_seek_offset1048576
        vfs.zfs.vdev.mirror.rotating_seek_inc   5
        vfs.zfs.vdev.mirror.rotating_inc        0
        vfs.zfs.vdev.trim_on_init               1
        vfs.zfs.vdev.bio_delete_disable         0
        vfs.zfs.vdev.bio_flush_disable          0
        vfs.zfs.vdev.cache.bshift               16
        vfs.zfs.vdev.cache.size                 0
        vfs.zfs.vdev.cache.max                  16384
        vfs.zfs.vdev.default_ms_shift           29
        vfs.zfs.vdev.min_ms_count               16
        vfs.zfs.vdev.max_ms_count               200
        vfs.zfs.vdev.trim_max_pending           10000
        vfs.zfs.txg.timeout                     5
        vfs.zfs.trim.enabled                    1
        vfs.zfs.trim.max_interval               1
        vfs.zfs.trim.timeout                    30
        vfs.zfs.trim.txg_delay                  32
        vfs.zfs.spa_min_slop                    134217728
        vfs.zfs.spa_slop_shift                  5
        vfs.zfs.spa_asize_inflation             24
        vfs.zfs.deadman_enabled                 1
        vfs.zfs.deadman_checktime_ms            5000
        vfs.zfs.deadman_synctime_ms             1000000
        vfs.zfs.debug_flags                     0
        vfs.zfs.debugflags                      0
        vfs.zfs.recover                         0
        vfs.zfs.spa_load_verify_data            1
        vfs.zfs.spa_load_verify_metadata        1
        vfs.zfs.spa_load_verify_maxinflight     10000
        vfs.zfs.max_missing_tvds_scan           0
        vfs.zfs.max_missing_tvds_cachefile      2
        vfs.zfs.max_missing_tvds                0
        vfs.zfs.spa_load_print_vdev_tree        0
        vfs.zfs.ccw_retry_interval              300
        vfs.zfs.check_hostid                    1
        vfs.zfs.mg_fragmentation_threshold      85
        vfs.zfs.mg_noalloc_threshold            0
        vfs.zfs.condense_pct                    200
        vfs.zfs.metaslab_sm_blksz               4096
        vfs.zfs.metaslab.bias_enabled           1
        vfs.zfs.metaslab.lba_weighting_enabled  1
        vfs.zfs.metaslab.fragmentation_factor_enabled1
        vfs.zfs.metaslab.preload_enabled        1
        vfs.zfs.metaslab.preload_limit          3
        vfs.zfs.metaslab.unload_delay           8
        vfs.zfs.metaslab.load_pct               50
        vfs.zfs.metaslab.min_alloc_size         33554432
        vfs.zfs.metaslab.df_free_pct            4
        vfs.zfs.metaslab.df_alloc_threshold     131072
        vfs.zfs.metaslab.debug_unload           0
        vfs.zfs.metaslab.debug_load             0
        vfs.zfs.metaslab.fragmentation_threshold70
        vfs.zfs.metaslab.force_ganging          16777217
        vfs.zfs.free_bpobj_enabled              1
        vfs.zfs.free_max_blocks                 18446744073709551615
        vfs.zfs.zfs_scan_checkpoint_interval    7200
        vfs.zfs.zfs_scan_legacy                 0
        vfs.zfs.no_scrub_prefetch               0
        vfs.zfs.no_scrub_io                     0
        vfs.zfs.resilver_min_time_ms            3000
        vfs.zfs.free_min_time_ms                1000
        vfs.zfs.scan_min_time_ms                1000
        vfs.zfs.scan_idle                       50
        vfs.zfs.scrub_delay                     4
        vfs.zfs.resilver_delay                  2
        vfs.zfs.top_maxinflight                 32
        vfs.zfs.delay_scale                     500000
        vfs.zfs.delay_min_dirty_percent         60
        vfs.zfs.dirty_data_sync                 67108864
        vfs.zfs.dirty_data_max_percent          10
        vfs.zfs.dirty_data_max_max              4294967296
        vfs.zfs.dirty_data_max                  3419890892
        vfs.zfs.max_recordsize                  1048576
        vfs.zfs.default_ibs                     15
        vfs.zfs.default_bs                      9
        vfs.zfs.zfetch.array_rd_sz              1048576
        vfs.zfs.zfetch.max_idistance            67108864
        vfs.zfs.zfetch.max_distance             8388608
        vfs.zfs.zfetch.min_sec_reap             2
        vfs.zfs.zfetch.max_streams              8
        vfs.zfs.prefetch_disable                0
        vfs.zfs.send_holes_without_birth_time   1
        vfs.zfs.mdcomp_disable                  0
        vfs.zfs.per_txg_dirty_frees_percent     30
        vfs.zfs.nopwrite_enabled                1
        vfs.zfs.dedup.prefetch                  1
        vfs.zfs.dbuf_cache_lowater_pct          10
        vfs.zfs.dbuf_cache_hiwater_pct          10
        vfs.zfs.dbuf_cache_shift                5
        vfs.zfs.dbuf_cache_max_bytes            1006785920
        vfs.zfs.arc_min_prescient_prefetch_ms   6
        vfs.zfs.arc_min_prefetch_ms             1
        vfs.zfs.l2c_only_size                   0
        vfs.zfs.mfu_ghost_data_esize            495983616
        vfs.zfs.mfu_ghost_metadata_esize        4735586816
        vfs.zfs.mfu_ghost_size                  5231570432
        vfs.zfs.mfu_data_esize                  10333800448
        vfs.zfs.mfu_metadata_esize              230963712
        vfs.zfs.mfu_size                        10951674368
        vfs.zfs.mru_ghost_data_esize            418293248
        vfs.zfs.mru_ghost_metadata_esize        12361377792
        vfs.zfs.mru_ghost_size                  12779671040
        vfs.zfs.mru_data_esize                  3002670080
        vfs.zfs.mru_metadata_esize              426496
        vfs.zfs.mru_size                        5335509504
        vfs.zfs.anon_data_esize                 0
        vfs.zfs.anon_metadata_esize             0
        vfs.zfs.anon_size                       132692992
        vfs.zfs.l2arc_norw                      1
        vfs.zfs.l2arc_feed_again                1
        vfs.zfs.l2arc_noprefetch                1
        vfs.zfs.l2arc_feed_min_ms               200
        vfs.zfs.l2arc_feed_secs                 1
        vfs.zfs.l2arc_headroom                  2
        vfs.zfs.l2arc_write_boost               8388608
        vfs.zfs.l2arc_write_max                 8388608
        vfs.zfs.arc_meta_limit                  8054287360
        vfs.zfs.arc_free_target                 173196
        vfs.zfs.arc_kmem_cache_reap_retry_ms    1000
        vfs.zfs.compressed_arc_enabled          1
        vfs.zfs.arc_grow_retry                  60
        vfs.zfs.arc_shrink_shift                7
        vfs.zfs.arc_average_blocksize           8192
        vfs.zfs.arc_no_grow_shift               5
        vfs.zfs.arc_min                         4027143680
        vfs.zfs.arc_max                         32217149440
        vfs.zfs.abd_chunk_size                  4096
                                                                Page:  7
------------------------------------------------------------------------
I'm looking for any recommendations on suitable SSDs for L2ARC. I'm also pondering whether I would need to buy more RAM. My current plan is to try the L2ARC with my current 32GB of RAM, and increase that to 64GB (or more) if I see signs that RAM has become a bottleneck.
			
				Last edited: