SSD Only build, where are we at?

Status
Not open for further replies.

diskdiddler

Wizard
Joined
Jul 9, 2014
Messages
2,377
Last time I saw a thread discussing this, some of the developers or cyberjock raised several concerns about potential design issues with an all SSD NAS. (sorry, I don't recall the issues or the location of the post, I did have a search)

While I'm not in the market, yet. I'm curious if this is a viable option in 9, 12, 24, 36 months? When do we think FreeNAS will work well with an all SSD build?
Based on hardware improvements in the industry, rumours, hypothetical claims, etc, it does sound like we may see 4/8/10TB 'storage' SSDs at a reasonable price in less than 18 months.
($500 US Crucial 4TB announced a month or two ago, so we're getting there)

I'd love a near silent, lower power, cooler NAS with no moving parts. (Oddly enough, performance isn't my interest over 100MB/s)
 

cyberjock

Inactive Account
Joined
Mar 25, 2012
Messages
19,525
The design issues with all-SSD really revolve around having experience with the specific hardware (SSDs) that you intend to use, in conjunction with properly utilizing FreeNAS with those SSDs. This is not an issue that is resolved with software updates or more code. This is an issue with the hardware itself, how SSDs work, their lifespan and longevity while storing your data, etc.

To be short, it's totally viable right now. You just have to have the experience and knowledge to properly manage SSD based ZFS systems. Most people aren't going to buy a 1/2 dozen of 5 or 6 different SSDs, test them in a bunch of different environments to figure out what hardware works with ZFS and what doesn't, and then pick from one of the few that are actually acceptable.

So unless you're ready to go that route, you're simply better off buying hardware from a company like iXsystems that has already spent gobs of money on testing and validating to ensure your SSDs can handle ZFS well.

Of course, you can go buy a bunch of whatever your favorite brand of SSD is and use those, but from what I've personally seen, it doesn't seem to work out so well once you actually start using the system. Most people that I've worked with that has failed SSD zpools built and tested the zpool, and all looked so wonderful. Then one day (always on a Friday evening before a 3 day weekend it seems) things aren't going so well, you can't mount the zpool, you can't get to your data, and you've relied on the system for so long that you have no spare server to use and you can't figure out if you've done an update to your backups in over 3 months. Whoops!

It really boils down to either building it right by doing your own R&D (very expensive this way), buying something from a company like iXsystems that did the R&D for you (probably more than you want to pay), taking some serious chances that things could go badly for you with no warning (probably more risk than you want to accept), or simply not doing an all-SSD zpool (Why do this? Are you not a nerd that has a need for all-SSD for the e-penis factor?).

I've got a bunch of older SSDs lying around here. I've been considering making an all-SSD zpool of them in my main box. It would be more of an experiment than anything else since it would be different models (but all the same brand.. Intel) and I'd be surprised if the darn thing lasted a month. The thing is I really don't have high expectations for the longevity, so I really can't convince myself to spend the time on it when I can't expect it to have a particularly useful ROI.

I too would love a near-silent, low power, cool NAS with little or no moving parts. The problem is that without all of the validation and verification of the hardware, you're quickly on an island where something goes wrong that nobody else has ever seen, and the only good solution is to rebuild and restore from backup. *I*, personally, am against scenarios where you are expecting to have to restore from backup in a short period of time. ;)

The only SSD zpool I've worked with that wasn't made by iXsystems or pre-built by someone else is a little 30GB I have that uses SLC memory. Good for at least 100k writes for each memory cell, and has 99% of the life remaining per SMART. Not overly concerned about it up and dying. Of course, you can't buy 4TB SLC-based SSDs at any price. They simply don't exist.

It's once of those things that works and either works well for the long term or quickly turns into a nightmare because you're spending so much time trying to figure out why the heck stuff doesn't work right, play nice, etc.
 

diskdiddler

Wizard
Joined
Jul 9, 2014
Messages
2,377
Well considering what a resliver actually does to a new drive and looking at my reports of writes to my disks, I certainly can imagine an MLC or TLC drive might really take a beating in a FreeNAS setup. There's at one guy on the forums who claims he's cooked 5 SSDs so far, if I recall (but I don't think he's running SSD only, it's his boot or cache drive)

I'm just curious if I ran 6 SSDs with 2 disk redundancy, just how risky it is. My assumption, based on posts on these forums and replies like yours, is that it would currently be more risky than running with 6 regular hard drives.
Perhaps we need a minimum of 10 disks with SSDs with 3 or 4 disk redundancy?? I simply don't know.

Either way, within the next 12 to 18 months (and I suspect sooner) you're going to have well paid, high end nerd, cashed up posters in this very forum, with money to burn who are going to want to try to do a build with just SSDs. Within 3 years, I suspect it will be commonplace for members to be asking about SSD only builds.
So ultimately that's the future that's coming, be it SATA based SSDs or something else, but something will need to change. Maybe they'll start making "WD Red, NAS SSDs" in 3 years?
 

titan_rw

Guru
Joined
Sep 1, 2012
Messages
586
I have a small ssd only pool that I use for deduplicated backups. It's simply two 960 gig ssd's in a stripe.

I was concerned about write endurance, as I send over 300 gigs of backups over the network to it each day. Most of it is deduplicated though, however there are tons of small random writes going on. These are TLC ssd's that have a certain amount of MLC / SLC (don't remember which) flash that acts as a buffer before writes hit the main TLC flash.

Here's host writes vs nand writes for the last 120 some odd days on both drives added together.

uaW0yiL.png


You can see freenas has written about 8tb to the ssd's, whereas that's only resulted in about 1.2 tb of actual nand writes.

Even the host writes are only averaging about 33 gigs a day, which is a little high for a consumer workload, but not terrible. Most consumer workloads are figured at about 20 gigs a day I think.

But I'm actually only doing about 5 gigs a day per drive of actual TLC wearing out writes.

Yes, I realize a striped pool is not good for redundancy, but for this particular data, I don't really care. It was more of a 'was if' thing for me.
 

David E

Contributor
Joined
Nov 1, 2013
Messages
119
@titan_rw How did you get the data for, and build that graph? Very interested.

My next rebuild in the fall will most likely be fully SSD based, and I may also in the interim try out a smaller SSD build to get some experience with it.
 

jgreco

Resident Grinch
Joined
May 29, 2011
Messages
18,680
I'm just curious if I ran 6 SSDs with 2 disk redundancy, just how risky it is. My assumption, based on posts on these forums and replies like yours, is that it would currently be more risky than running with 6 regular hard drives.

Cyberjock is a bit of a pessimist who sees the world through some harshly black and white glasses.

The six hard drives would have better endurance characteristics than the SSD's, but that could be fine for many scenarios. I've been running a small mirror SSD pool of 850 Evo M.2's for several months and it's pretty nice. I may end up expanding it, but for what we do around here I'm not really convinced that it is a win over local hypervisor SSD datastores.

Perhaps we need a minimum of 10 disks with SSDs with 3 or 4 disk redundancy?? I simply don't know.

Yeah, if I end up running this to its logical conclusion, it'll be ten SSD's of 3-way mirrors (3 vdevs and a spare).

Either way, within the next 12 to 18 months (and I suspect sooner) you're going to have well paid, high end nerd, cashed up posters in this very forum, with money to burn who are going to want to try to do a build with just SSDs. Within 3 years, I suspect it will be commonplace for members to be asking about SSD only builds.

People are already doing this for jails. I don't really see the difference.

So ultimately that's the future that's coming, be it SATA based SSDs or something else, but something will need to change. Maybe they'll start making "WD Red, NAS SSDs" in 3 years?

Actually I'm waiting for the hard drive manufacturers to WAKE THE FSCK UP and move back to larger form factors for HDD storage. Now that we've figured out that SSD will win the speed war, the smaller sizes of hard drives do not make as much sense.
 

David E

Contributor
Joined
Nov 1, 2013
Messages
119
The six hard drives would have better endurance characteristics than the SSD's, but that could be fine for many scenarios. I've been running a small mirror SSD pool of 850 Evo M.2's for several months and it's pretty nice. I may end up expanding it, but for what we do around here I'm not really convinced that it is a win over local hypervisor SSD datastores.

How does that perform? Also IMO the real losses vs just having SSDs in your hypervisor is global snapshotting/ease of backups, and fast load balancing of compute resources via DRS. But you pay for that with *much* slower performance than having the disks in the box with the compute. I've never tried VSAN, but I'm interested in how it would compare for this use case.
 

jgreco

Resident Grinch
Joined
May 29, 2011
Messages
18,680
Well, I can tell you that sticking some Intel 535's in RAID1 on an LSI 3108 with 2GB of cache is smashing reads through at 460MB/sec, whereas FreeNAS and the Samsung 850's are around 300MB/sec. Since a lot of our NAS is virtualized, and the workload isn't highly variable, there's not as much of a pressing need for being able to idle hypervisors. The 3108 plus SSD solution is a lot less expensive in capex and opex for a small number of hypervisors; a decent FreeNAS system and 10G networking is $$$$. But options are always nice. Also it's worth noting that the I/O configuration I've got for FreeNAS+ESXi is almost certainly far from optimal, so there's a very real possibility that the FreeNAS solution could go faster.
 

cyberjock

Inactive Account
Joined
Mar 25, 2012
Messages
19,525
Cyberjock is a bit of a pessimist who sees the world through some harshly black and white glasses.

The six hard drives would have better endurance characteristics than the SSD's, but that could be fine for many scenarios. I've been running a small mirror SSD pool of 850 Evo M.2's for several months and it's pretty nice. I may end up expanding it, but for what we do around here I'm not really convinced that it is a win over local hypervisor SSD datastores.

That's a load of crap. I don't try to compare SSDs to platter disks. There's so many variables that unless you're going to start quoting numbers, you can GTFO with your justifications. :P
 

jgreco

Resident Grinch
Joined
May 29, 2011
Messages
18,680
That's a load of crap.

Read it in context. Do you really deny that you tend to see things in absolutist terms?

I don't try to compare SSDs to platter disks. There's so many variables that unless you're going to start quoting numbers, you can GTFO with your justifications. :p

I never said you did.
 

cyberjock

Inactive Account
Joined
Mar 25, 2012
Messages
19,525
Do you really deny that you tend to see things in absolutist terms?

For some things, I do. For many, many things, absolutely not. You only see me discuss absolutist things very vocally because when you cannot provide absolutes, trying to take a side between one or the other is really more about opinion than anything else.

As a scientist for many years, very little is in absolutes. It's more of what tends to work the best for most people. For example, we all know that the recommended temp is <40C for hard drives, but if someone is here running 50C I can't say "well, crap, your drives are gonna die in the next year... good job".

It's easy to argue the absolute stuff. It's virtually impossible to argue the stuff that is open to conjecture, opinion, twisted facts, etc.

I see all the "shades of grey" very well. But quite often it's a very twisty and turny road, and if you look away from the road for even one second... disaster strikes. So I strive to keep people in the areas that are black or white because the grey area is where people most often fail. Especially people who aren't an expert in a topic. And not surprisingly, very few people asking questions here are experts on FreeNAS or ZFS.
 

cyberjock

Inactive Account
Joined
Mar 25, 2012
Messages
19,525
I never said you did.

Really?
diskdiddler said:
I'm just curious if I ran 6 SSDs with 2 disk redundancy, just how risky it is. My assumption, based on posts on these forums and replies like yours, is that it would currently be more risky than running with 6 regular hard drives.

Cyberjock is a bit of a pessimist who sees the world through some harshly black and white glasses.

Sure looks like someone asked a question comparing SSDs to regular disks, and the first word to the answer had my name in it...
 

jgreco

Resident Grinch
Joined
May 29, 2011
Messages
18,680
Really?
diskdiddler said:
I'm just curious if I ran 6 SSDs with 2 disk redundancy, just how risky it is. My assumption, based on posts on these forums and replies like yours, is that it would currently be more risky than running with 6 regular hard drives.

Cyberjock is a bit of a pessimist who sees the world through some harshly black and white glasses.

Sure looks like someone asked a question comparing SSDs to regular disks, and the first word to the answer had my name in it...

Yes. Someone asked a question comparing SSDs to regular disks, and you came out with this jaw-dropping response that included all sorts of FUD:

Most people aren't going to buy a 1/2 dozen of 5 or 6 different SSDs, test them in a bunch of different environments to figure out what hardware works with ZFS and what doesn't, and then pick from one of the few that are actually acceptable.

"what hardware works with ZFS"....?!?!? Any quality SSD is likely to be fine. I can't think of a contemporary SSD that doesn't "work with ZFS". This is FUD of the worst sort.

So unless you're ready to go that route, you're simply better off buying hardware from a company like iXsystems that has already spent gobs of money on testing and validating to ensure your SSDs can handle ZFS well.

"Handle ZFS well." It's a NAS. Even for those of us with 10G ports, "handle ZFS well" is kind of a very strange thing to say, because a NAS is limited by nature to some rather modest limits. You're not talking to a bunch of corporate types who are looking for a hot shit software SAN that can run iSCSI on a pair of 40G uplinks to a cluster of ESXi running heavy transactional data; for that, I'd agree, see iXsystems. You're talking to a forum with what's largely a bunch of home hobbyists who are looking to store mostly long term archival data on predominantly 1G networks. There's no need to suggest that somehow there are SSDs that can't "handle ZFS well." Any quality modern SSD should be able to perform NAS duties. It doesn't have to be the best choice or the highest performing choice, it merely has to work well enough as a NAS drive at a low price point. More FUD.

It really boils down to either building it right by doing your own R&D (very expensive this way), buying something from a company like iXsystems that did the R&D for you (probably more than you want to pay), taking some serious chances that things could go badly for you with no warning (probably more risk than you want to accept), or simply not doing an all-SSD zpool (Why do this? Are you not a nerd that has a need for all-SSD for the e-penis factor?).

That's such a narrow view of the world. There are other benefits of an all-SSD pool. Lower noise. Less sensitive to heat. More responsiveness. There are other use models for FreeNAS than just storing massive amounts of data. Consider something like storing office files. You don't need a lot of space, but high reliability and good responsiveness are nice qualities.

I've got a bunch of older SSDs lying around here. I've been considering making an all-SSD zpool of them in my main box. It would be more of an experiment than anything else since it would be different models (but all the same brand.. Intel) and I'd be surprised if the darn thing lasted a month.

Perhaps you should actually try that little experiment. Why the hell would you be "surprised if the darn thing lasted a month"? I mean, sure, you can probably make it burn through its write endurance in that period if you try, that's not really that hard, but part of being a storage admin is understanding your workloads and how to balance competing interests to arrive at something that's suitable to your needs.

Our new hypervisors out in Ashburn, with an LSI 3108 RAID controller, I've been putting Intel 535's in them. Why? Because I did the math and discovered that our write levels were such that we were in the ballpark of the 40GB/day writes that they're spec'd for. So I deployed five in each server, two RAID1's and a spare. The numbers suggest the busier of the two datastores will last the full five years, perhaps just barely. But I don't even really care if it only were to last two, because at that time, the cost to replace will be very modest, and in the meantime, the cost differential between those $150 SSD's and the next step up (Intel DC S3500 480GB) at $320, is so great and the rate at which SSD prices are dropping is so extreme, that this is actually the smart thing to do.

I too would love a near-silent, low power, cool NAS with little or no moving parts. The problem is that without all of the validation and verification of the hardware, you're quickly on an island where something goes wrong that nobody else has ever seen, and the only good solution is to rebuild and restore from backup. *I*, personally, am against scenarios where you are expecting to have to restore from backup in a short period of time. ;)

That's just more FUD. It's 2016, dude. SSD technology is well understood, we have things like media wearout indicators, and SSD's have been deployed in a huge range of storage roles. Lots of users here on the forums have actually been deploying SSD's for their jails, which is arguably a more challenging environment than mere pool storage, and we're not hearing about all the daily failures.

The only SSD zpool I've worked with that wasn't made by iXsystems or pre-built by someone else is a little 30GB I have that uses SLC memory. Good for at least 100k writes for each memory cell, and has 99% of the life remaining per SMART. Not overly concerned about it up and dying. Of course, you can't buy 4TB SLC-based SSDs at any price. They simply don't exist.

We already understand that SLC isn't actually practical or even necessarily better than MLC, and manufacturers have been shipping enterprise grade MLC since... 2010? Like literally "before FreeNAS"? Those amazing DC S3710 SSD's with the multiple petabyte write endurance? Those use MLC. SLC vs MLC is a settled thing, and it doesn't even matter which is "better".

Basically, to make a long story short, no one gives a crap about SLC anymore. I don't know how it is that you missed out on the last half a decade, but amazing strides have been made. We've got 512GB microSD cards out there. The price on MLC has plummeted while performance and endurance have slowly increased at the low end, and at the top end, it's an unrecognizable market.

It's once of those things that works and either works well for the long term or quickly turns into a nightmare because you're spending so much time trying to figure out why the heck stuff doesn't work right, play nice, etc.

That might have been the case 5 years ago. That's not the case today. All the major issues are known. As with HDD's, there may well be things to learn, such as WDIDLE or other tweaking of the sorts we suggest for HDD's, but you need to remember that this community is perfectly capable of dealing with that sort of thing, and in fact asking some questions about that was actually the opener to this thread.

So, here, here's the non-FUD answer to the question that the poster asked.

The six hard drives would have better endurance characteristics than the SSD's, but that could be fine for many scenarios.

That's a perfect assessment of the realities.
 

jgreco

Resident Grinch
Joined
May 29, 2011
Messages
18,680
Well considering what a resliver actually does to a new drive and looking at my reports of writes to my disks, I certainly can imagine an MLC or TLC drive might really take a beating in a FreeNAS setup. There's at one guy on the forums who claims he's cooked 5 SSDs so far, if I recall (but I don't think he's running SSD only, it's his boot or cache drive)

Yes, you can cook SSD's if you're not careful or you don't care. But you also have to remember that SSD specifications are sometimes pretty conservative. For example, we deployed an OCZ Vertex 3 60GB in ... 2012? as L2ARC on a machine. The specifications for the OCZ Vertex 3 included 20GB/day of write endurance and a 3 year warranty, which is about 22TB, but the reported stats from our unit is actually 699TB (699027GB) written(!). I almost wouldn't believe that if I hadn't checked the stats a few times along the way, and if I wasn't aware of things like The SSD Endurance Experiment, I wouldn't believe it. As it is, I have to attribute a fair chunk of that to the compression features of the SandForce controller, I don't really think that the drive would have sustained 700TB of incompressible data.

The number of actual post-burnin-but-less-than-rated-endurance SSD failures we've seen, I can count on one hand. In that way, they tend to be far more reliable than hard drives, which crap out for random reasons.

Some HDD's will actually report write statistics, such as the Seagate 4TB's here, which report 73871328331 LBA's written over 24300 hours, which works out to 37GB/day (73871328331 lbas / 24300 hrs * 24 hrs in a day * 512 bytes per lba). So you can potentially do some math to see out what your write statistics might be like.
 

cyberjock

Inactive Account
Joined
Mar 25, 2012
Messages
19,525
Feel free to consider it FUD... but I've seen quite a few system (both iXsystems and custom built) and I can tell you that there are reasons to avoid some models, firmwares, etc. I'm just not going to get into the details because it's very complicated.

Good luck to anyone that wants to run SSDs though!
 

jgreco

Resident Grinch
Joined
May 29, 2011
Messages
18,680
Even the host writes are only averaging about 33 gigs a day, which is a little high for a consumer workload, but not terrible. Most consumer workloads are figured at about 20 gigs a day I think.

But I'm actually only doing about 5 gigs a day per drive of actual TLC wearing out writes.

But the endurance is typically figured not on the actual TLC write load but rather on the writes sent to the disk. That can be frustrating.

What SSD are you using? The SanDisk Ultra II? I've been avoiding that because while it's really cheap ($200 for 960GB last Black Friday) there's no endurance rating available, plus it doesn't provide Device Activity Signal, and I like blinky lights.

The Intel 535's are rated for 40GB/day endurance (and have a gimpy but functional DAS). But cap out at 480GB.

The Samsung 850 Pro's are really nice with a 10 year warranty and 300TBW, or 82TB/day. Unfortunately no DAS.

The 850 Evo 500's are pretty good too, with a 150TBW and 5 year warranty, the same 82TB/day.
 

jgreco

Resident Grinch
Joined
May 29, 2011
Messages
18,680
Feel free to consider it FUD... but I've seen quite a few system (both iXsystems and custom built) and I can tell you that there are reasons to avoid some models, firmwares, etc. I'm just not going to get into the details because it's very complicated.

Good luck to anyone that wants to run SSDs though!

You can say the same thing about HDD's. Geeeez.
 

titan_rw

Guru
Joined
Sep 1, 2012
Messages
586
Some HDD's will actually report write statistics, such as the Seagate 4TB's here, which report 73871328331 LBA's written over 24300 hours, which works out to 37GB/day (73871328331 lbas / 24300 hrs * 24 hrs in a day * 512 bytes per lba). So you can potentially do some math to see out what your write statistics might be like.

I have a bunch of Seagate drives that report Read / Written LBAs. Here's some, with power on hours to give some indication of their age:

Code:
POH:11542 WriteGB:20900
POH:10284 WriteGB:23573
POH:21599 WriteGB:49808
POH:18322 WriteGB:49106
POH:14744 WriteGB:51840
POH:11348 WriteGB:20910
POH:13387 WriteGB:27241
POH:30148 WriteGB:29683
POH:21703 WriteGB:32851
POH: 9399 WriteGB: 9695
POH:11942 WriteGB:19034
POH: 3993 WriteGB:16248
POH:11738 WriteGB:21045
POH:15400 WriteGB:14876
POH:13262 WriteGB:34937
POH:17722 WriteGB:50099


Also note that most of the drives had at least an initial run of badblocks run on them, which does 4 full write passes. Being 3tb drives, that would be ~12tb of initial writes.

Even so, this is well within SSD's write endurance. Especially considering you wouldn't be running badblocks on an ssd.
 

titan_rw

Guru
Joined
Sep 1, 2012
Messages
586
But the endurance is typically figured not on the actual TLC write load but rather on the writes sent to the disk. That can be frustrating.

What SSD are you using? The SanDisk Ultra II? I've been avoiding that because while it's really cheap ($200 for 960GB last Black Friday) there's no endurance rating available, plus it doesn't provide Device Activity Signal, and I like blinky lights.

The Intel 535's are rated for 40GB/day endurance (and have a gimpy but functional DAS). But cap out at 480GB.

The Samsung 850 Pro's are really nice with a 10 year warranty and 300TBW, or 82TB/day. Unfortunately no DAS.

The 850 Evo 500's are pretty good too, with a 150TBW and 5 year warranty, the same 82TB/day.

Yes, they actually are Ultra II's. I got them on the same black friday sale. In Canada they were a lot more than $200, but still very well priced.

Mine are mounted internally, so no blinky lights at all. I thought I had found some kind of write endurance specs at some point, but I'm not totally sure.
 

titan_rw

Guru
Joined
Sep 1, 2012
Messages
586
@titan_rw How did you get the data for, and build that graph? Very interested.

My next rebuild in the fall will most likely be fully SSD based, and I may also in the interim try out a smaller SSD build to get some experience with it.

I just used a script that pulls the info from smartctl. The script is run as a cron job every day and appends to a .txt file. Then I manually update the spreadsheet by re-opening the .txt file in excel and copy / pasting the new data. There's probably a better way to do the last part.
 
Status
Not open for further replies.
Top