- Joined
- Apr 16, 2020
- Messages
- 2,947
Anyone,
If you look at my specs below you can see I have two optane drives. Each acts as a Log for a different pool
SSDPool (6*SATA SSD, Mirrored Devs) & BigPool (6*HDD, Mirrored Devs)
The Optanes are 250+GB, but a ZIL I understand needs a mere 16GB or so. Currently the optanes are running with a 120GB partition.
From the command line I have run
gpart create -s GPT nvd0
gpart add -t freebsd-zfs -a 1m -l "name" -s 120G nvd0
And then added this as a log using zpool add "Pool" log nvd0p1
It occurs to me that I could (at least in theory until I try it)
1. Remove both logs
2. Wipe both Optanes nvd0 & nvd1
3. Create 2 * 120 (ish) GB partitions on each Optane using gpart
4. zpool add "Pool1" log mirror nvd0p1 nvd1p1
5. zpool add "Pool2" log mirror nvd0p2 nvd1p2
That way the ZIL / Slogs would be mirrored (in case of failure)
Anyone got any thoughts as to good / bad idea?
--------------------------------------------------------
As a second question - I have got my paws on a pair of Intel DC S3610 SSD's (860GB) in good condition
My large HDD pool has video data on it (not much help) but also a large number of small files in a host of sub folders that can take a while to look through from a client.
If I understand things correctly adding these SSD's as a mirror special vdev to the HDD pool, and then copying off and back on again all the data would cache the metadata on the SSD's and potentially speed some operations up (my mileage may vary).
Almost as importantly, as the Pool is mirrored vdevs (3*2) and the special vdev would be mirrored then the vdev could be removed somehow from the pool if the effect is not as obvious as I would like. I am not sure I would want to go ahead if I can't remove the fusion (special) vdev from the pool
Output from a couple of commands below (assuming I have posted it correctly)
If you look at my specs below you can see I have two optane drives. Each acts as a Log for a different pool
SSDPool (6*SATA SSD, Mirrored Devs) & BigPool (6*HDD, Mirrored Devs)
The Optanes are 250+GB, but a ZIL I understand needs a mere 16GB or so. Currently the optanes are running with a 120GB partition.
From the command line I have run
gpart create -s GPT nvd0
gpart add -t freebsd-zfs -a 1m -l "name" -s 120G nvd0
And then added this as a log using zpool add "Pool" log nvd0p1
It occurs to me that I could (at least in theory until I try it)
1. Remove both logs
2. Wipe both Optanes nvd0 & nvd1
3. Create 2 * 120 (ish) GB partitions on each Optane using gpart
4. zpool add "Pool1" log mirror nvd0p1 nvd1p1
5. zpool add "Pool2" log mirror nvd0p2 nvd1p2
That way the ZIL / Slogs would be mirrored (in case of failure)
Anyone got any thoughts as to good / bad idea?
--------------------------------------------------------
As a second question - I have got my paws on a pair of Intel DC S3610 SSD's (860GB) in good condition
My large HDD pool has video data on it (not much help) but also a large number of small files in a host of sub folders that can take a while to look through from a client.
If I understand things correctly adding these SSD's as a mirror special vdev to the HDD pool, and then copying off and back on again all the data would cache the metadata on the SSD's and potentially speed some operations up (my mileage may vary).
Almost as importantly, as the Pool is mirrored vdevs (3*2) and the special vdev would be mirrored then the vdev could be removed somehow from the pool if the effect is not as obvious as I would like. I am not sure I would want to go ahead if I can't remove the fusion (special) vdev from the pool
Output from a couple of commands below (assuming I have posted it correctly)
Code:
1k: 25996
2k: 96317
4k: 30507
8k: 8270
16k: 12312
32k: 41484
64k: 61987
128k: 29278
256k: 68105
512k: 59129
1M: 23653
2M: 13284
4M: 12278
8M: 8403
16M: 4052
32M: 2500
64M: 887
128M: 973
256M: 10093
512M: 4869
1G: 1807
2G: 978
4G: 75
8G: 18
16G: 5
32G: 4
64G: 1
128G: 1
Traversing all blocks ...
bp count: 115609922
ganged count: 0
bp logical: 16163205622784 avg: 139808
bp physical: 15831147257344 avg: 136935 compression: 1.02
bp allocated: 15842228080640 avg: 137031 compression: 1.02
bp deduped: 0 ref>1: 0 deduplication: 1.00
Normal class: 15842201255936 used: 44.16%
additional, non-pointer bps of type 0: 501995
number of (compressed) bytes: number of bps
17: 35 *
18: 645 *
19: 131 *
20: 200 *
21: 294 *
22: 171 *
23: 209 *
24: 170 *
25: 81 *
26: 537 *
27: 440 *
28: 20033 *********
29: 90890 ****************************************
30: 32 *
31: 366 *
32: 330 *
33: 1095 *
34: 235 *
35: 249 *
36: 95 *
37: 313 *
38: 351 *
39: 564 *
40: 1474 *
41: 1088 *
42: 1094 *
43: 637 *
44: 326 *
45: 174 *
46: 211 *
47: 277 *
48: 364 *
49: 609 *
50: 613 *
51: 2388 **
52: 2034 *
53: 15682 *******
54: 46653 *********************
55: 7557 ****
56: 20175 *********
57: 3764 **
58: 16107 ********
59: 904 *
60: 782 *
61: 734 *
62: 974 *
63: 897 *
64: 856 *
65: 975 *
66: 1424 *
67: 1670 *
68: 20794 **********
69: 3523 **
70: 2182 *
71: 2710 **
72: 2008 *
73: 2741 **
74: 4311 **
75: 3816 **
76: 2524 **
77: 2892 **
78: 4608 ***
79: 34884 ****************
80: 3507 **
81: 5909 ***
82: 19271 *********
83: 2895 **
84: 1957 *
85: 2041 *
86: 5233 ***
87: 3211 **
88: 10550 *****
89: 21068 **********
90: 7270 ****
91: 10659 *****
92: 3203 **
93: 2695 **
94: 2642 **
95: 5120 ***
96: 5205 ***
97: 2563 **
98: 2096 *
99: 1801 *
100: 1823 *
101: 2001 *
102: 1694 *
103: 2214 *
104: 1928 *
105: 1257 *
106: 1522 *
107: 1755 *
108: 1708 *
109: 2313 **
110: 1780 *
111: 12902 ******
112: 15300 *******
Dittoed blocks on same vdev: 70889
Blocks LSIZE PSIZE ASIZE avg comp %Total Type
- - - - - - - unallocated
2 32K 8K 24K 12K 4.00 0.00 object directory
10 5K 5K 120K 12K 1.00 0.00 object array
1 16K 4K 12K 12K 4.00 0.00 packed nvlist
- - - - - - - packed nvlist size
68 2.12M 348K 1.02M 15.4K 6.25 0.00 L1 bpobj
2.58K 330M 23.0M 68.9M 26.7K 14.38 0.00 L0 bpobj
2.65K 332M 23.3M 69.9M 26.4K 14.26 0.00 bpobj
- - - - - - - bpobj header
- - - - - - - SPA space map header
41 656K 164K 492K 12K 4.00 0.00 L2 SPA space map
588 9.19M 2.61M 7.82M 13.6K 3.53 0.00 L1 SPA space map
14.8K 162M 89.2M 268M 18.0K 1.82 0.00 L0 SPA space map
15.4K 172M 91.9M 276M 17.9K 1.87 0.00 SPA space map
12 312K 312K 312K 26K 1.00 0.00 ZIL intent log
105 13.1M 420K 840K 8K 32.00 0.00 L5 DMU dnode
105 13.1M 420K 840K 8K 32.00 0.00 L4 DMU dnode
105 13.1M 420K 840K 8K 32.00 0.00 L3 DMU dnode
106 13.2M 424K 852K 8.04K 32.00 0.00 L2 DMU dnode
304 38M 7.53M 15.1M 50.9K 5.05 0.00 L1 DMU dnode
79.6K 1.24G 343M 688M 8.65K 3.71 0.00 L0 DMU dnode
80.3K 1.33G 352M 706M 8.80K 3.87 0.00 DMU dnode
106 408K 408K 852K 8.04K 1.00 0.00 DMU objset
- - - - - - - DSL directory
49 25.5K 3K 48K 1003 8.50 0.00 DSL directory child map
47 42K 16.5K 180K 3.83K 2.55 0.00 DSL dataset snap map
60 374K 96.5K 408K 6.80K 3.87 0.00 DSL props
- - - - - - - DSL dataset
- - - - - - - ZFS znode
- - - - - - - ZFS V0 ACL
35 1.09M 140K 280K 8K 8.00 0.00 L3 ZFS plain file
21.9K 701M 91.6M 183M 8.35K 7.66 0.00 L2 ZFS plain file
581K 18.1G 4.90G 9.79G 17.3K 3.71 0.07 L1 ZFS plain file
87.4M 14.3T 14.2T 14.2T 166K 1.01 98.30 L0 ZFS plain file
88.0M 14.4T 14.2T 14.2T 165K 1.01 98.37 ZFS plain file
1 32K 4K 8K 8K 8.00 0.00 L2 ZFS directory
89.4K 2.79G 358M 715M 8K 8.00 0.00 L1 ZFS directory
672K 3.10G 848M 2.79G 4.25K 3.74 0.02 L0 ZFS directory
761K 5.89G 1.18G 3.49G 4.69K 5.00 0.02 ZFS directory
45 45K 39K 360K 8K 1.15 0.00 ZFS master node
- - - - - - - ZFS delete queue
9 288K 36K 72K 8K 8.00 0.00 L4 zvol object
30 960K 300K 600K 20K 3.20 0.00 L3 zvol object
953 29.8M 12.9M 25.8M 27.7K 2.31 0.00 L2 zvol object
98.6K 3.08G 1.12G 2.24G 23.2K 2.76 0.02 L1 zvol object
21.3M 341G 234G 234G 11.0K 1.46 1.59 L0 zvol object
21.4M 344G 235G 236G 11.0K 1.46 1.60 zvol object
- - - - - - - zvol prop
- - - - - - - other uint8[]
- - - - - - - other uint64[]
- - - - - - - other ZAP
- - - - - - - persistent error log
1 32K 12K 36K 36K 2.67 0.00 L1 SPA history
128 16M 1.48M 4.44M 35.5K 10.81 0.00 L0 SPA history
129 16.0M 1.49M 4.48M 35.5K 10.74 0.00 SPA history
- - - - - - - SPA history offsets
- - - - - - - Pool properties
- - - - - - - DSL permissions
- - - - - - - ZFS ACL
- - - - - - - ZFS SYSACL
- - - - - - - FUID table
- - - - - - - FUID table size
64 34.5K 3K 12K 192 11.50 0.00 DSL dataset next clones
- - - - - - - scan work queue
263 166K 61.5K 432K 1.64K 2.70 0.00 ZFS user/group/project used
- - - - - - - ZFS user/group/project quota
- - - - - - - snapshot refcount tags
- - - - - - - DDT ZAP algorithm
- - - - - - - DDT statistics
- - - - - - - System attributes
- - - - - - - SA master node
45 67.5K 65.5K 360K 8K 1.03 0.00 SA attr registration
90 1.41M 348K 720K 8K 4.14 0.00 SA attr layouts
- - - - - - - scan translations
- - - - - - - deduplicated block
252 194K 119K 1.39M 5.67K 1.63 0.00 DSL deadlist map
- - - - - - - DSL deadlist map hdr
22 13.5K 3K 12K 558 4.50 0.00 DSL dir clones
- - - - - - - bpobj subobj
- - - - - - - deferred free
- - - - - - - dedup ditto
21 672K 84K 252K 12K 8.00 0.00 L1 other
277 1.33M 210K 756K 2.73K 6.48 0.00 L0 other
298 1.99M 294K 1008K 3.38K 6.91 0.00 other
105 13.1M 420K 840K 8K 32.00 0.00 L5 Total
114 13.4M 456K 912K 8K 30.11 0.00 L4 Total
170 15.2M 860K 1.68M 10.1K 18.05 0.00 L3 Total
23.0K 745M 105M 210M 9.14K 7.09 0.00 L2 Total
770K 24.1G 6.37G 12.7G 17.0K 3.78 0.09 L1 Total
109M 14.7T 14.4T 14.4T 135K 1.02 99.91 L0 Total
110M 14.7T 14.4T 14.4T 134K 1.02 100.00 Total
Block Size Histogram
block psize lsize asize
size Count Size Cum. Count Size Cum. Count Size Cum.
512: 241K 120M 120M 238K 119M 119M 0 0 0
1K: 147K 175M 295M 141K 167M 286M 0 0 0
2K: 210K 572M 867M 205K 560M 846M 0 0 0
4K: 6.29M 25.2G 26.0G 157K 798M 1.61G 6.07M 24.3G 24.3G
8K: 7.87M 80.3G 106G 83.2K 940M 2.52G 8.22M 81.5G 106G
16K: 9.02M 146G 252G 21.5M 345G 348G 9.46M 156G 262G
32K: 697K 32.3G 284G 894K 29.3G 377G 701K 32.4G 294G
64K: 1.45M 133G 417G 86.3K 7.95G 385G 1.45M 133G 427G
128K: 74.5M 9.32T 9.72T 77.1M 9.64T 10.0T 74.5M 9.32T 9.73T
256K: 19.2K 7.04G 9.73T 11 4.05M 10.0T 19.4K 7.10G 9.74T
512K: 9.33M 4.67T 14.4T 9.37M 4.68T 14.7T 9.33M 4.67T 14.4T
1M: 0 0 14.4T 0 0 14.7T 0 0 14.4T
2M: 0 0 14.4T 0 0 14.7T 0 0 14.4T
4M: 0 0 14.4T 0 0 14.7T 0 0 14.4T
8M: 0 0 14.4T 0 0 14.7T 0 0 14.4T
16M: 0 0 14.4T 0 0 14.7T 0 0 14.4T
root@freenas[~]#