Poor write performance

Status
Not open for further replies.

sfcredfox

Patron
Joined
Aug 26, 2014
Messages
340
what were you using as an HBA
I am using an HP H221 I think. It's HPs 6G rebranded LSI, a 9207-8e I believe. Flashed to IT P20.

Now there were some weird things about those cards where HP was selling two different LSI part numbers as the same thing (H221). My card can only be upgraded with the firmware from an LSI 9205-8e, so really that's what mine is. New ones were the 9207. Make sense, nope...HP can be stupid some times. I found a blog taking about it which was the only way I figured it out.

I really want to know what your numbers are with the 300GB 6G drives. I only had 12, so I had to extrapolate what the performance would have been with a full shelf of them, based on the increase of a mirrored pair. I started with 4 pairs, then 5, then 6. I saw about a 4-5 percent increase each time I added a pair, but who knows if that would have been a linear increase all the way to 24 disks.

I used VMware IO analyzer, ran that for a long while to get the cache at peak performance and then ran a 1 hour SQL16K test against the datastore on the 6G pool. Got around 16,500 IOPS.

Here's the so what of all this:

With proper caching of the dataset, I was able to equal or exceed the performance of the old SAS drives with 12 SATA 2TB disks. Now, if the data isn't cached, of course it sucks, it's SATA. People like to point out that using cached performance data isn't good for anything, but for my purpose, I argue that I'm not solely interested in the performance data purely of the disk subsystem, but how the entire FreeNAS system supports my applications in VMware. So in that regard, if the hot portions of my application database end up cached, that's exactly what will happen in real life, and that's what I want to base my performance expectations on. Otherwise, all storage venders would suck unless we went back to the days of three racks full of disks to obtain the performance of TrueNAS/Nimble/etc. who can 'cheat' with caching and get far better performance using so fewer disks.

I also imagine your data will also back up the idea of more vdevs usually leads to higher write throughput, with proper low latency SLOG, like you added :)

Sorry, long post for a simple question
 

sfcredfox

Patron
Joined
Aug 26, 2014
Messages
340
If you get time for all this, you could also use that shelf of SAS disks to compare the mirrored performance results to a Z1/Z2 config. With exact same everything else, that would be very revealing of relative performance between the configs. It would be very interesting. Obviously, I'm a nerd..."interesting data"...
 
Joined
Dec 29, 2014
Messages
1,135
I am using an HP H221 I think. It's HPs 6G rebranded LSI, a 9207-8e I believe. Flashed to IT P20.

If it takes 9205 code, then that means your card PCIe 2.0. The 9507 is PCIe 3.0. I am at least happy to know that the card is able to control the D2700. The first one of these I got (MSA20) needed a special RAID card to even talk to the drive. It was an SA642, I think.

I also imagine your data will also back up the idea of more vdevs usually leads to higher write throughput, with proper low latency SLOG, like you added :)

I'll have to see what this one can do once I get it alive again. The D2700 is attached to my backup FreeNAS, so it won't have the 900P available to it. I am going to play around with using a HW RAID controller mirror as an SLOG versus not. I may try the RZ2 versus mirror thing as well just for chuckles.

Sorry, long post for a simple question

I tell my customers that when I do this sort of thing, it is due to engineer syndrome. It means I am genetically incapable of giving short answers to questions, particularly technical ones. :smile:
 
Joined
Dec 29, 2014
Messages
1,135
I have the results of some testing I have done. First I'll describe the files servers. These are creatively named freenas and freenas2. :smile:

freenas2 is my primary one.

Cisco UCS C240 M3S
FreeNAS 11.1-U4
Dual E5-2660 v2 CPU @ 2.20GHz
128G ECC RAM DDR3-1333
Chelsio T520-CR 10G-SR SFP+

t5nex0: <Chelsio T520-CR> mem 0xfb300000-0xfb37ffff,0xfa000000-0xfaffffff,0xfbb0
4000-0xfbb05fff irq 64 at device 0.4 numa-domain 1 on pci14
cxl0: <port 0> numa-domain 1 on t5nex0
cxl0: 16 txq, 8 rxq (NIC)
cxl1: <port 1> numa-domain 1 on t5nex0
cxl1: 16 txq, 8 rxq (NIC)
t5nex0: PCIe gen3 x8, 2 ports, 18 MSI-X interrupts, 51 eq, 17 iq

Drives controlled by LSI 9271-8i in JBOD mode

AVAGO MegaRAID SAS FreeBSD mrsas driver version: 06.712.04.00-fbsd
mrsas0: <AVAGO Thunderbolt SAS Controller> port 0xf000-0xf0ff mem 0xfbc60000-0xfbc63fff,0xfbc00000-0xfbc3ffff irq 56 at device 0.0 n
uma-domain 1 on pci13
mrsas0: Using MSI-X with 16 number of vectors
mrsas0: FW supports <16> MSIX vector,Online CPU 40 Current MSIX <16>
mrsas0: MSI-x interrupts setup success

17 x Cisco 1TB 7.2K SATA drives in internal drive bays

da0 at mrsas0 bus 1 scbus1 target 27 lun 0
da0: <ATA ST91000640NS CC03> Fixed Direct Access SPC-4 SCSI device
da0: 150.000MB/s transfers
da0: 953869MB (1953525168 512 byte sectors)

ZFS Pool where VMware virtual machines are stored shared via NFS v3
NAME STATE READ WRITE CKSUM
RAIDZ2-I ONLINE 0 0 0
raidz2-0 ONLINE 0 0 0
gptid/bd041ac6-9e63-11e7-a091-e4c722848f30 ONLINE 0 0 0
gptid/bdef2899-9e63-11e7-a091-e4c722848f30 ONLINE 0 0 0
gptid/bed51d90-9e63-11e7-a091-e4c722848f30 ONLINE 0 0 0
gptid/bfb76075-9e63-11e7-a091-e4c722848f30 ONLINE 0 0 0
gptid/c09c704a-9e63-11e7-a091-e4c722848f30 ONLINE 0 0 0
gptid/c1922b7c-9e63-11e7-a091-e4c722848f30 ONLINE 0 0 0
gptid/c276eb75-9e63-11e7-a091-e4c722848f30 ONLINE 0 0 0
gptid/c3724eeb-9e63-11e7-a091-e4c722848f30 ONLINE 0 0 0
raidz2-1 ONLINE 0 0 0
gptid/a1b7ef4b-3c2a-11e8-978a-e4c722848f30 ONLINE 0 0 0
gptid/a2eb419f-3c2a-11e8-978a-e4c722848f30 ONLINE 0 0 0
gptid/a41758d7-3c2a-11e8-978a-e4c722848f30 ONLINE 0 0 0
gptid/a5444dfb-3c2a-11e8-978a-e4c722848f30 ONLINE 0 0 0
gptid/a6dcd16f-3c2a-11e8-978a-e4c722848f30 ONLINE 0 0 0
gptid/a80cd73c-3c2a-11e8-978a-e4c722848f30 ONLINE 0 0 0
gptid/a94711a5-3c2a-11e8-978a-e4c722848f30 ONLINE 0 0 0
gptid/aaa6631d-3c2a-11e8-978a-e4c722848f30 ONLINE 0 0 0
logs
gptid/6f76bc3b-5aee-11e8-8c41-e4c722848f30 ONLINE 0 0 0
spares
gptid/4abff125-23a2-11e8-a466-e4c722848f30 AVAIL

SLOG - Intel Optane 900P 280G

nvme0: <Generic NVMe Device> mem 0xdf010000-0xdf013fff irq 40 at device 0.0 numa-domain 0 on pci5
-----------------------------------------------------------
freenas is the secondary box that I don't use all that much. I may try and set up some automated replication to it at some point, but haven't been inspired to do so yet.

Cisco UCS C240 M3S
FreeNAS 11.1-U4
Dual E5-2620 CPU @ 2.00GHz
64G ECC RAM DDR3-1333
Chelsio T520-CR 10G-SR SFP+

t5nex0: <Chelsio T520-CR> mem 0xfb300000-0xfb37ffff,0xfa000000-0xfaffffff,0xfbb0
4000-0xfbb05fff irq 64 at device 0.4 numa-domain 1 on pci14
cxl0: <port 0> numa-domain 1 on t5nex0
cxl0: 16 txq, 8 rxq (NIC)
cxl1: <port 1> numa-domain 1 on t5nex0
cxl1: 16 txq, 8 rxq (NIC)
t5nex0: PCIe gen3 x8, 2 ports, 18 MSI-X interrupts, 51 eq, 17 iq

HP D2700 external enclosure with 25 SAS/SATA slots controlled by LSI 9207-8e. The D2700 has two controllers, and each one is connected to a port on the 9207-8e

mps0: <Avago Technologies (LSI) SAS2308> port 0x7000-0x70ff mem 0xdf140000-0xdf14ffff,0xdf100000-0xdf13ffff irq 32 at device 0.0 num
a-domain 0 on pci3
mps0: Firmware: 20.00.07.00, Driver: 21.02.00.00-fbsd
mps0: IOCCapabilities: 5a85c<ScsiTaskFull,DiagTrace,SnapBuf,EEDP,TransRetry,EventReplay,MSIXIndex,HostDisc>

ses0 at mps0 bus 0 scbus0 target 33 lun 0
ses0: <HP D2700 SAS AJ941A 0149> Fixed Enclosure Services SPC-3 SCSI device
ses0: 600.000MB/s transfers
ses0: Command Queueing enabled
ses0: SCSI-3 ENC Device
ses1 at mps0 bus 0 scbus0 target 59 lun 0
ses1: GEOM: da0: the secondary GPT header is not in the last LBA.
<HP D2700 SAS AJ941A 0149> Fixed Enclosure Services SPC-3 SCSI device
ses1: 600.000MB/s transfers
ses1: Command Queueing enabled
ses1: SCSI-3 ENC Device

25 x 300G 10k SAS drives in the D2700 look like this.

da5: <HP DG0300FAMWN HPDF> Fixed Direct Access SPC-3 SCSI device
da5: 600.000MB/s transfers
da5: Command Queueing enabled
da5: 286102MB (585937500 512 byte sectors)

ZFS Pool where VMware virtual machines are stored shared via NFS v3
NAME STATE READ WRITE CKSUM
TEST2 ONLINE 0 0 0
raidz2-0 ONLINE 0 0 0
gptid/07379c8a-60ef-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/09d7ac93-60ef-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/0c7f06f4-60ef-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/0f28ba50-60ef-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/11cb5ce8-60ef-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/1478482b-60ef-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
raidz2-1 ONLINE 0 0 0
gptid/a2e860a8-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/a59c35e2-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/aa159c4a-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/acd5a2e2-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/af81bbb2-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/b23b1e54-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
raidz2-2 ONLINE 0 0 0
gptid/f486e8de-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/f73b2f3f-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/f9eabce3-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/fc9f41a5-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/ff7e9906-60f5-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/029453a0-60f6-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
raidz2-3 ONLINE 0 0 0
gptid/084c70aa-60f6-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/0aff24d6-60f6-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/0dbd832b-60f6-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/10670ef6-60f6-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/1327e8dc-60f6-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
gptid/15e5089e-60f6-11e8-bc93-f44e059f1a98 ONLINE 0 0 0
spares
gptid/1be4166c-60f6-11e8-bc93-f44e059f1a98 AVAIL


SLOG drive(s) controlled by LSI 9271-8i

AVAGO MegaRAID SAS FreeBSD mrsas driver version: 06.712.04.00-fbsd
mrsas0: <AVAGO Thunderbolt SAS Controller> port 0xf000-0xf0ff mem 0xfbc60000-0xf
bc63fff,0xfbc00000-0xfbc3ffff irq 56 at device 0.0 numa-domain 1 on pci13
mrsas0: Using MSI-X with 16 number of vectors
mrsas0: FW supports <16> MSIX vector,Online CPU 24 Current MSIX <16>
mrsas0: MSI-x interrupts setup success

SLOG using RAID1 with ST9146803SS drives (146G SAS 10K):
da50 at mrsas0 bus 0 scbus1 target 0 lun 0
da50: <LSI MR9271-8i 3.46> Fixed Direct Access SPC-3 SCSI device
da50: 150.000MB/s transfers
da50: 139236MB (285155328 512 byte sectors)

SLOG using RAID1 with ST9300653SS drives (300G SAS 15K): *** This is RAID0 in tests with VM-UCS2 because I don't have enough 300G 15K SAS drives
da51 at mrsas0 bus 0 scbus1 target 1 lun 0
da51: <LSI MR9271-8i 3.46> Fixed Direct Access SPC-3 SCSI device
da51: 150.000MB/s transfers
da51: 285148MB (583983104 512 byte sectors)
-----------------------------------------------------------
I have 2 ESXi hosts that have enough local storage to hold the 3 VM's that normally have running all the time. Those a FreeBSD box (my mail server), Vcenter (small), and APC PC Network Shutdown (CentOS).

VM-DL360-G7
HP DL360 G7 - ESXi 6.0.0 - Dual X5672 CPU - 72GB RAM - P410i RAID - 4 x WDC WD7500BPKT RAID5
Mellanox Technologies MT26448 [ConnectX EN 10GigE , PCIe 2.0 5GT/s]

VM-UCS2
Cisco UCS C220 M3S - ESXi 6.5.0 - Dual E5-2670 CPU - 128G RAM - LSI 9266-8i RAID
UCS VIC P81E 10G NIC
-----------------------------------------------------------
Network switch is an HP S5820 (really an H3C OEM'd for HP). All connections between hosts and switch are 10G SR SFP using OM-3 fiber
-----------------------------------------------------------

Enough table setting, and on to the tests.

DL360-G7 client, freenas (secondary) server

No SLOG: 6.0G down, 342M up (peak) with ~= 250M avg up
Cisco 146G 10K SAS RAID1 SLOG: 6.0G down, 474M up (almost constant)
Cisco 300G 15K SAS RAID1 SLOG: 6.0G down, 730M up (almost constant)

DL360-G7 client, freenas2 (primary) server

6.6G down (peak) with ~= 5G avg, 4.9G up (peak) with ~= 4G average


VM-UCS2 client, freenas (secondary) server

No SLOG: 7.6G down (peak) 7.2G avg (sustained), 344M up (peak) ~= 212M avg (flucuates)
Cisco 146G 10K SAS RAID1 SLOG: 7.4G down (sustained), 637M (peak) up ~= 432M avg (sustained)
Cisco 300G 15K SAS RAID0 SLOG: 7.5G down (sustained), 847M (peak) up ~= 760M avg (sustained)


VM-UCS2 client, freenas2 (primary) server

8.3G down (peak) 7.8G avg (sustained), 4.2G up (peak) ~= 3G avg (sustained)
-----------------------------------------------------------
First off I have to say how impressed I am at the read performance FreeNAS is able to deliver even when using some older drives/controllers. Next, I was astonished by how much difference the SLOG made for NFS write performance. I know most people prefer iSCSI for backing VMware data stores, but I use NFS because I understand it better as a grumpy old Unix guy. I am mostly a network infrastructure person these days, so there isn't a huge motivation to learn iSCSI if I can get NFS to do what I need. I am going to try this again soon as I bought some used S3500 drives to use as an SLOG on the secondary host. I can certainly run some other tests if anybody would like. I know the storage Vmotion stuff isn't the most objective measure, but that is the thing that matters most to me so it seemed the ideal test.

One thing that puzzles me a little bit is how large the difference in write performance is between the 146G 10K drives and 300G 15K drives. How much is rotational speed, and how much is that the drive is larger? I certainly don't understand the architecture of ZFS well enough to answer that, but I am curious.
 
Last edited:

sfcredfox

Patron
Joined
Aug 26, 2014
Messages
340
I think that confirms at least what we expected: that the 6G drives do a little better than the 3G drives.

The super high read speeds might be a measure of the cache more than the drive pool itself?

What tool did you use and what parameters?

Have you given any though to using MVware's IO Analyzer VM? It's a free download, you import it, login to it, then use your web browser to tell it what parameters to test.

That's where I told it to run a SQL 16K load for 15 minutes, then the same thing for an hour. It showed me how the system would support VMs using a combination of the caching and the writing at the same time. I also tested by deploying four of them (two on each ESX) and running all the tests at the same time to simulate a busy environment.

You might find the numbers slightly more reflective of the performance of your VMs.

I'd also like to compare your numbers on the 300GB 10K drives if they were in mirrors and see if your Z2 pool is doing just as good or better than mirrors?
 
Joined
Dec 29, 2014
Messages
1,135
The super high read speeds might be a measure of the cache more than the drive pool itself?

That is certainly possible. Both boxes have a decent amount of RAM (128/64).

What tool did you use and what parameters?

I was looking at the network reporting in FreeNAS.

Have you given any though to using MVware's IO Analyzer VM? It's a free download, you import it, login to it, then use your web browser to tell it what parameters to test.

I have thought about it. It took way more time to do the inexact tests that I have already run. I will probably wait until my used S3500's arrive for a better SLOG in the secondary FreeNAS. FYI, I blame you for that! :smile:

I'd also like to compare your numbers on the 300GB 10K drives if they were in mirrors and see if your Z2 pool is doing just as good or better than mirrors?

I'll give that a whirl too when I have all the parts and some more time to fiddle with the testing.
 

Chris Moore

Hall of Famer
Joined
May 2, 2015
Messages
10,080
All I can say is WOW! Just adding this card as an SLOG (ZIL in the GUI when you extend the volume) increases the write performance on the pool by 4-5X. I mean, WOW! Thanks for the suggestions. I definitely won't forget about this. It was about $370 on Amazon, but worth every penny.
What kind of drive did you get?
 

Chris Moore

Hall of Famer
Joined
May 2, 2015
Messages
10,080
Joined
Dec 29, 2014
Messages
1,135

sfcredfox

Patron
Joined
Aug 26, 2014
Messages
340
Joined
Dec 29, 2014
Messages
1,135
I think I got 100MB/s writing to those, if you got a pair, that will likely be your throughput for the pool, or combined if you stipe them

One of my S3500's arrived today, and I have done a little testing with it. Here are the results:

S3500 SLOG (single) - JBOD 1.1G upload (1.0G sustained), 3.6G download (3.0G sustained)
S3500 SLOG (single) - RAID0 1.4G upload (1.2G sustained), 3.6G download (3.0G sustained)

Involving the RAID controller on the SLOG did seem to help. In either case, this is quite a jump from using separate HDD's for an SLOG. They weren't super expensive, so I think that was probably worth the cost. This is my secondary unit, so I am trying to keep from going too crazy on it.
 

Chris Moore

Hall of Famer
Joined
May 2, 2015
Messages
10,080
One of my S3500's arrived today, and I have done a little testing with it. Here are the results:

S3500 SLOG (single) - JBOD 1.1G upload (1.0G sustained), 3.6G download (3.0G sustained)
S3500 SLOG (single) - RAID0 1.4G upload (1.2G sustained), 3.6G download (3.0G sustained)

Involving the RAID controller on the SLOG did seem to help. In either case, this is quite a jump from using separate HDD's for an SLOG. They weren't super expensive, so I think that was probably worth the cost. This is my secondary unit, so I am trying to keep from going too crazy on it.
You might want to compare them with the results of this drive:
https://www.ebay.com/itm/HGST-100GB...27395-Enterprise-High-Endurance-/273242307438

I picked the ones I have up for $50 each and I wish I had bought more. This is the performance test I did with one of them:
https://forums.freenas.org/index.ph...-and-finding-the-best-slog.63521/#post-455075

It is a SAS drive though, so it will need an actual SAS port to plug into.
 
Joined
Dec 29, 2014
Messages
1,135
You might want to compare them with the results of this drive:
https://www.ebay.com/itm/HGST-100GB...27395-Enterprise-High-Endurance-/273242307438

I picked the ones I have up for $50 each and I wish I had bought more.

Yes, that is considerably better than what I am getting off these drives. All the slots are SAS/SATA, so that wouldn't be a problem. Now I have to figure out some way to get room to put the D2700 back in the rack. I didn't think I was going to use it again, but the 9207-8e's combined with the SSD SLOG make it more than viable for a secondary server.
 
Joined
Dec 29, 2014
Messages
1,135
Last entry for the moment on this. The 2nd S3500 arrived today, and here are the results from that.

S3500 SLOG (dual) - JBOD 2.2G upload (2.0G sustained), 7.5G download (6.2G sustained)
S3500 SLOG (dual) - RAID0 2.3G upload (2.1G sustained), 7.5G download (6.2G sustained)

It looks like the RAID0 give it a little boost, but I will likely stay with striping the SSD SLOG's via ZFS. I did order a couple of the Hitachi SSD's that @Chris Moore suggested, so I will give those a whirl when they come in.
 

Chris Moore

Hall of Famer
Joined
May 2, 2015
Messages
10,080
Last entry for the moment on this. The 2nd S3500 arrived today, and here are the results from that.

S3500 SLOG (dual) - JBOD 2.2G upload (2.0G sustained), 7.5G download (6.2G sustained)
S3500 SLOG (dual) - RAID0 2.3G upload (2.1G sustained), 7.5G download (6.2G sustained)

It looks like the RAID0 give it a little boost, but I will likely stay with striping the SSD SLOG's via ZFS. I did order a couple of the Hitachi SSD's that @Chris Moore suggested, so I will give those a whirl when they come in.
We have a system at work that uses a pair of those Hitachi drives in each 16 bay drive shelf and there are 12 drive shelves.
 
Joined
Dec 29, 2014
Messages
1,135
The two Hitachi SSD's arrived today, and they definitely performed better than the SATA ones. No surprise there. The write performance was just about a straight line, but two different peaks. The first peak was 3.0G with ~=2.7G sustained. Then it peaked at 4.3G sustaining around 4.0G. I suspect that is once the ARC allocated as much as it could of the 64G RAM this box has. With the two SSD's in a striped SLOG, the performance was not the far off from the Intel 900P. In any case, that makes my secondary a pretty darn viable NAS. So I am happy!
 

Chris Moore

Hall of Famer
Joined
May 2, 2015
Messages
10,080
With the two SSD's in a striped SLOG, the performance was not the far off from the Intel 900P. In any case, that makes my secondary a pretty darn viable NAS. So I am happy!
That makes me happy too. Do you want to add that information to the other thread where we were bench-marking SLOG devices?
 
Status
Not open for further replies.
Top