zpool import hangs forever, help please!!!

Status
Not open for further replies.

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
I am *so* tired of ZFS and its "superiority". It was supposed to the the uber fs and in reality, for me, it just has been a nightmare after a nigthmare. Now it looks like I am in deep do-do and have no idea how to get out.

I am running FN8.0 and within it I created a raidz1 (tank1 - yes, how original...) using 4x 2Tb WD EARS disks.
It was all more or less OK (some minor corruption when I pushed the system heavily - I only have 2Gb of RAM) but nothing serious that scrub couldn't handle. Seeing that more RAM is better, I upped up to 4 Gb.
Then, suddenly, when I started to migrate data into tank1, ZFS gave out a series of nasty messages stating that the pool wasn't available and froze the system.
I rebooted and then the system hanged at the "mounting file system" stage.
Fine, I figure something got corrupted in the USB stick.
Re-installed on the stick from a CDROM, loaded the configuration from backup and... same stalling point.
Fine, I re-installed on the stick from a second time (no restore config file this time), removed the 2 extra Gb and then it loaded (I would suspect this had something to do with the extra RAM which was supposed to help ZFS). Of course, no zpool.
I did an auto import from the GUI... hangs forever.
I did a "zpool imort" from CLI and apparently tank1 is there, all OK and ONLINE.
I then did a "zpool import -fF tank1" and it also hangs forever... WTF!!!

So, I am now suck with a raidz1 pool containing about 4.5 Tb worth of info that is non-accessible and apparently non-recoverable!

So much for the *uber* FS.

Any help would be immensly appreciated indeed!
 
Joined
May 27, 2011
Messages
566
great opening, moving on...

'tank1', hell mine is called 'storage'

a lack of memory will not cause corruption, it sounds like you have major hardware issues. it doesn't sound like a software issue it sounds like you started getting corruption from a hardware failure and instead of acting on that, you added memory. I'm not sure what to tell you. i would run a memtest on your current hardware, and on different hardware, install Freenas 8.0, once installed and booting properly, shut it off and add your disks. you may have accidentally installed FreeNAS onto one of your 2 TB drives and had a second one fail on you, removing 2 disks from your raidz which will kill it.

let me know what happens with the memtest and the new hardware.
 

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
great opening, moving on...

'tank1', hell mine is called 'storage'

a lack of memory will not cause corruption, it sounds like you have major hardware issues. it doesn't sound like a software issue it sounds like you started getting corruption from a hardware failure and instead of acting on that, you added memory. I'm not sure what to tell you. i would run a memtest on your current hardware, and on different hardware, install Freenas 8.0, once installed and booting properly, shut it off and add your disks. you may have accidentally installed FreeNAS onto one of your 2 TB drives and had a second one fail on you, removing 2 disks from your raidz which will kill it.

let me know what happens with the memtest and the new hardware.

Yes, I thought about that. Allow me to clarify.
The MOBO I am using won't handle 4Gb (for some weird reason) although it is supposed to, however, with 2Gb it is OK. The minor corruption (on 2Gb) happened while I was pumping data full speed into tank1, and I mean thousands of small files directly from a hdd attached to the mobo into tank1 using cp through CLI. Under those conditions memory utilization went to 100% and stayed there. I am not surprised that out of more than 100.000 files about 40 got corrupted. No biggie. But, to take care of this issue, I upped to 4Gb... big mistake!
However, the mayor melt-down happened when I was on 4 Gb.
Asap I started to pump data in, it crashed.
Now I am back on 2 Gbs and a couple of memtests I run indicate that RAM is OK. I am confident on this configuration since I have been using it for days on full data load almost without any hiccups as I mentioned above. Heck! I had 5 DVD drives attached all pumping data into tank1 (big files this time) and it didn't even blink (inactive memory stayed at 1.5 Gbs).

Wrt overwriting the raidz1, nope. I am 100% sure. Only on the USB stick. I was *very* careful not to screw-up. Additionally, zpool import does indicate that tank1 is OK and all 4 hdds are ONLINE. I know, zpool import is not exactly a fine tool for diagnosing software errors, but at least it is something.

So, I have to conclude that in this "stable" config (ie 2Gb), it is purely a ZFS issue.
I would have loved to been able to use zpool imort -fFX, but it is not implemented.

Any ideas?
 

freeflow

Dabbler
Joined
May 29, 2011
Messages
38
I'm asuming you haven't yet managed to reimport your zpool

From the cli

zpool status

To confirm that there are no pools available
then try

zpool import -a -f

followed by

zpool status

Hopefully your zpool will be imported and its status is OK. IF not report back what you get from the above status and import commands.

If you do successfully import your zpool then use 'zpool export' to export it again and then use the auto import feature from the GUI
 
Joined
May 27, 2011
Messages
566
So, I have to conclude that in this "stable" config (ie 2Gb), it is purely a ZFS issue.
I would have loved to been able to use zpool imort -fFX, but it is not implemented.

Any ideas?

well i gave my advice, you choose not to take it. it's not a software issue, when i made the migration to 8, i copied 5 TB at about 2.5 gb/s my controller overheated (a few times before i realized), 0 corrupted files. if zfs corrupted files under heavy load, it would not be used in production Anywhere. zfs was not designed by 4th graders. promise it's not zfs that's at fault.
 

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
I'm asuming you haven't yet managed to reimport your zpool

From the cli

zpool status

To confirm that there are no pools available
then try

zpool import -a -f

followed by

zpool status

Hopefully your zpool will be imported and its status is OK. IF not report back what you get from the above status and import commands.

If you do successfully import your zpool then use 'zpool export' to export it again and then use the auto import feature from the GUI

Thank you kindly. Still trying without any results. Output as follows:

1 - zpool status
no pools available

2 - zpool import -a -f
hangs

3 - if I run a simple zpool import to see what's available I get the standard (and correct) output:

pool: tank1
id: bunch of #
state: ONLINE
action: The pool can be imported using its name or numeric identifier
config:
tank1: ONLINE
raidz1: ONLINE

gpt/ada0 ONLINE
gpt/ada1 ONLINE
gpt/ada2 ONLINE
gpt/ada3 ONLINE

Been doing a lot of reading wrt these types of "hangings" and they would seem to be related to data corruption. If this is the case, rolling back transactions to a point at which the pool was synched and dumping the rest (i.e. corrupted ones) should restore the pool. However, it would seem that the only command capable of doing this automagically is zpool import -fFX <poolname> which is not implemented in FreeBSD.

Tx! again.
 

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
well i gave my advice, you choose not to take it. it's not a software issue, when i made the migration to 8, i copied 5 TB at about 2.5 gb/s my controller overheated (a few times before i realized), 0 corrupted files. if zfs corrupted files under heavy load, it would not be used in production Anywhere. zfs was not designed by 4th graders. promise it's not zfs that's at fault.

Tx! for your input, it is certainly appreciated. However, pls note that although your system may have been able to handle it, I don't have an endless pit of $$$ to build a super-duper box to match the uber fs. I am restricted to what I can get on the cheap. If this hardware produces a small corruption of about 40 files in 100.000 then so be it. I will have to live with it. For example, I simply cannot afford to pay for a higher-endish PCI-SATA controller the same price as a consumer hdd.
Additionally, I do understand that in my config (i.e. 2Gb) I am severely pushing ZFS and some "instability" is to be expected (as indicated by current literature).
Wrt ZFS used in "prod" environments, sure, however, pls note that such a prod environments usually have uudles of $$$ to burn and can afford uber-hardware. Stability also comes from this. Additionally, all prod environments have a decent tape backup system that can handle Tb's no sweat. I can't afford this either, hence, raidz1.

In other words, I can tolerate a lot of hicc-ups, lack of performance and nuances.
However, what I can't tolerate is my current state of affairs where the pool is inaccessible and I cant even repair it.

ZFS may not have been designed by 4th graders, but it is certainly a retarded system, if you ask me. Sure, it may have all kinds of bells and whistles, but sorely lacks recovery tools. It is simply pathetic that the only automated way to --perhaps-- restore a pool in my conditions is to run the mythical "zpool import -fFX" command (I am referring to the X) which is un-documented (even in Solaris) and not implemented anywhere else that I know of. Otherwise, if you try to roll it back manually, you are stuck in ZFS geekland. Good luck trying to solve your problem. Just google it.

Or I could try to check the rollback point before a corruption using the -n switch... which is not implemented either (see http://solori.wordpress.com/2010/07/15/zfs-pool-import-fails-after-power-outage/).

The whole point of ZFS was that it was supposed to hide complexity so that script-kiddies could use it. Well, that's not the case. Not by a long shot.

It is simply retarded that an uber-journaling-transactional-CRC-checked-to-death fs can't roll-back corrupted transactions automatically (remember, we are talking raid1 so there is plenty of redundancy where to check corruption from). It is also retarded that it is not possible to check it while not imported. And don't get me started with the fact that a pool cannot be added hdd's to. Or that it cannot recognize 4k hdds or that it cannot self-tune RAM utilization.

It is ludicrous that one has to spend countless days (yes days) trying to solve problems that the ZFS was supposed to protect (or at least minimize) against. Actually, ZFS creates more problems that it solves.

And yes, I have even tried Solaris 11 but the controller (a Sil 3114 chip) is not supported by Solaris; sure, why would the most widespread chip on the market be supported by Solaris? Too much common sense I guess. Btw, there are a few rumors around saying that the Sil 3114 is supported by Solaris. Do me a favor, if come across such a mythical Solaris driver, pls let me know (in any case, if a driver actually appears will have to reflash the controller since Solaris don't linke hardware raid either - nice huh?).

And yes, I have also tried the latest FreeBSD 8.2 without any success. The mysterious X switch is not implemented either.

I'll try with Linux now, but I don't expect any difference since as far as I know, it's not implemented either.

So, am I pissed off? Sure! ZFS is not living up to expectations. Heck! Not even close!

Again, tx! for your feedback.
 

freeflow

Dabbler
Joined
May 29, 2011
Messages
38
I ran an 8x500GB raid z2 with 2GB ram without any problems for several months. I strongly suspect you are having some sort of hardware problem.

Some alternative approaches

1. try an Ubuntu install/live disk (but only install with your target drive installed, disconnect your zfs raid).

In ubuntu open a terminal

sudo bash
>enter password
zpool status
You will be told there is no zfs and be told how to install the zfs-fuse package. Follow the instructions

Then try importing your pool and/or looking at the logs to see if there are any error messages which might give you a clue.

2. zfsguru experimental with zfs v28

google zfsguru and download the relevant iso then follow instructions. If you can't import from the GUI you will at least be able to try the command line as zfs v28 will have the F option for import.

3. Connect your disks to your main PC. Boot to a live version of freenas 7 or 8. Se if you can import your pool.

One or more of the above will help you work out what is going wrongor even retrieve your pool.

Please remember, if you do successfully import your pool. export it again before trying the FreeNAS 8 GUI.

Finally have a look at the -D flag for import. It might just help.

And remember. There are millions of users of zfs who don't have any problems.
 

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
I ran an 8x500GB raid z2 with 2GB ram without any problems for several months. I strongly suspect you are having some sort of hardware problem.

Some alternative approaches

[snip]

WOW! lots of alternatives to follow-up on. Tx!. This will take a while. Will report back asap done.

Oh, a slightly off-topic question if you don't mind. Did you stress the system? I mean *really* stress it? In my case I was pushing thousand of files directly from a Hdd on the mobo into the raidz1 using cp in the CLI. The memory utilization went straight up to 100% and stayed there for the duration. And I mean a *lot* of files. About 1/2 Tb worth of them. Had to cool down the source hdd by taking it out of the box and adding fans. However, in contrast, cping a few large files (0.5 to 1.5 Gb) does not even register a blip in the system. That's why I am saying that I wasn't surprised of the extra-low corruption level. I don't think this level of stress is common on low-end systems. The literature for ZFS clearly states that it requires a lot of RAM otherwise it is unstable. However, having said that, I don't discard the possibility of a hardware issue, there is always a chance. Anyhoo, tx! again and will post asap I get results (one way or the other).
 

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
Perhaps you'll be get more help by posting to the FreeBSD filesystems mailing list. The guys responsible for ZFS in FreeBSD will see your message.

Paul

Tx! good idea. However, I did cross-post to several forums, including FreeBSD. I totally missed the mail list. They are much more active, you are absolutely right! Will keep it in mind for the next catastrophe!
 

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
I ran an 8x500GB raid z2 with 2GB ram without any problems for several months. I strongly suspect you are having some sort of hardware problem.

Some alternative approaches

1. try an Ubuntu install/live disk (but only install with your target drive installed, disconnect your zfs raid).

In ubuntu open a terminal

sudo bash
>enter password
zpool status
You will be told there is no zfs and be told how to install the zfs-fuse package. Follow the instructions

Then try importing your pool and/or looking at the logs to see if there are any error messages which might give you a clue.

WOOHOO!! Success!

OK the Ubuntu thingy did it, well, sort of. Check the terminal output:

globus@globus-desktop:~$ sudo bash
[sudo] password for globus:
root@globus-desktop:~# zpool status
no pools available
root@globus-desktop:~# zpool import
pool: tank1
id: 6042935247182975919
state: ONLINE
status: The pool was last accessed by another system.
action: The pool can be imported using its name or numeric identifier and
the '-f' flag.
see: http://www.sun.com/msg/ZFS-8000-EY
config:

tank1 ONLINE
raidz1-0 ONLINE
disk/by-id/ata-WDC_WD20EARS-00MVWB0_WD-WCAZA4832167-part2 ONLINE
disk/by-id/ata-WDC_WD20EARS-00MVWB0_WD-WCAZA5639660-part2 ONLINE
disk/by-id/ata-WDC_WD20EARS-00MVWB0_WD-WCAZA5621172-part2 ONLINE
disk/by-id/ata-WDC_WD20EARS-00MVWB0_WD-WMAZA4291636-part2 ONLINE


root@globus-desktop:~# zpool import -f tank1
cannot import 'tank1': I/O error
Recovery is possible, but will result in some data loss.
Returning the pool to its state as of 2011-06-13
should correct the problem. Approximately 1 seconds of data
must be discarded, irreversibly. Recovery can be attempted
by executing 'zpool import -F tank1'. A scrub of the pool
is strongly recommended after recovery.
root@globus-desktop:~# zpool import -F tank1
cannot import 'tank1': pool may be in use from other system, it was last accessed by (hostid: 0xc96589ee) on Mon Jun 13 21:11:41 2011
use '-f' to import anyway
root@globus-desktop:~# zpool import -f tank1
cannot import 'tank1': I/O error
Recovery is possible, but will result in some data loss.
Returning the pool to its state as of 2011-06-13
should correct the problem. Approximately 1 seconds of data
must be discarded, irreversibly. Recovery can be attempted
by executing 'zpool import -F tank1'. A scrub of the pool
is strongly recommended after recovery.


root@globus-desktop:~# zpool import -fFX tank1
Pool tank1 returned to its state as of 2011-06-13.
Discarded approximately 1 seconds of transactions.
root@globus-desktop:~# zpool import
root@globus-desktop:~# zpool status
pool: tank1
state: ONLINE
status: The pool is formatted using an older on-disk format. The pool can
still be used, but some features are unavailable.
action: Upgrade the pool using 'zpool upgrade'. Once this is done, the
pool will no longer be accessible on older software versions.
scrub: none requested
config:

NAME STATE READ WRITE CKSUM
tank1 ONLINE 0 0 0
raidz1-0 ONLINE 0 0 0
disk/by-id/ata-WDC_WD20EARS-00MVWB0_WD-WCAZA4832167-part2 ONLINE 0 0 0
disk/by-id/ata-WDC_WD20EARS-00MVWB0_WD-WCAZA5639660-part2 ONLINE 0 0 0
disk/by-id/ata-WDC_WD20EARS-00MVWB0_WD-WCAZA5621172-part2 ONLINE 0 0 0
disk/by-id/ata-WDC_WD20EARS-00MVWB0_WD-WMAZA4291636-part2 ONLINE 0 0 0

errors: No known data errors
root@globus-desktop:~#


======================

In other words, obviously the Ubuntu distro has a much more advanced version of ZFS than FN 8.0 which does recognize that there has been data corruption and correctly diagnoses that it needs to ditch the corrupted data. However, the -f and -F switches do bupkus!
It's the magic "X" switch in conjunction with the -f and -F that did the trick.

As an aside, I think this validates (at least partially) my theory that the RAM upgrade from 2Gb to 4Gb went screwy but beyond this, the hardware is "sufficiently" operational for my needs. Apparently, there are no other errors in the raidz1 (running scrub as I type).

Anyhoo I owe you a debt of gratitude!

BIG TX!
 

pauldonovan

Explorer
Joined
May 31, 2011
Messages
76
Glad you got it sorted out.

BTW, the zpool import -X flag is implemented in the ZFS v28 patch for FreeBSD-CURRENT. FreeBSD 8.2 only has v15, although there is also a patch for 8.2 as well. iX Systems have imported that patch into the FreeNAS 8.1 branch so it (-X) should be available in FreeNAS at some point in the future.

Paul
 
Joined
May 27, 2011
Messages
566
first up, instead of

zpool import -a -f



try



zpool import -f <your pool name here>



the -a may be your issue, it's a long shot but I'm still trying to help.





I do understand that in my config (i.e. 2Gb) I am severely pushing ZFS and some "instability" is to be expected (as indicated by current literature).



it's not pushing zfs in the slightest, and the literature is talking about performance, not stability.





ZFS may not have been designed by 4th graders, but it is certainly a retarded system, if you ask me. Sure, it may have all kinds of bells and whistles, but sorely lacks recovery tools. It is simply pathetic that the only automated way to --perhaps-- restore a pool in my conditions is to run the mythical "zpool import -fFX" command (I am referring to the X) which is un-documented (even in Solaris) and not implemented anywhere else that I know of. Otherwise, if you try to roll it back manually, you are stuck in ZFS geekland. Good luck trying to solve your problem. Just google it.

the -F option isn't in this version of zfs. FreeBSD uses version 15, NexentaStor, from the article uses a much more up to date version of zfs. if you really want to use it, grab a live disk of solaris express 11.



The whole point of ZFS was that it was supposed to hide complexity so that script-kiddies could use it. Well, that's not the case. Not by a long shot.

not in the slightest, zfs was never designed for ease of use. FreeNAS and other distributions put a nice wrapper around it but zfs was not designed to be pretty.



It is simply retarded that an uber-journaling-transactional-CRC-checked-to-death fs can't roll-back corrupted transactions automatically (remember, we are talking raid1 so there is plenty of redundancy where to check corruption from). It is also retarded that it is not possible to check it while not imported.

the later versions of zfs do with the -F option, however, this version does not. raid1 has 1 disk of redundancy, if you're seeing corruption and you don't handle it, you run the risk of having the issues you have.



And don't get me started with the fact that a pool cannot be added hdd's to.

you can add disks to a pool, you just can't add disks to a raidz(2) vdev. you can expand your storage just fine. ZFS was not designed for the end user, it was designed for mass storage in a closest where storage isn't added disk by disk, but in significant chunks.



Or that it cannot recognize 4k hdds

as far as i know 4k drives do not advertise that they are 4k it's not part of the SATA standard, there is no mechanism for a drive to advertise 512 vs 4k.



or that it cannot self-tune RAM utilization.

it does a fantastic job if it's given enough memory. otherwise it struggles.





And yes, I have even tried Solaris 11 but the controller (a Sil 3114 chip) is not supported by Solaris; sure, why would the most widespread chip on the market be supported by Solaris? Too much common sense I guess. Btw, there are a few rumors around saying that the Sil 3114 is supported by Solaris. Do me a favor, if come across such a mythical Solaris driver, pls let me know (in any case, if a driver actually appears will have to reflash the controller since Solaris don't linke hardware raid either - nice huh?).

my Sil 3114 card worked great under OpenSolaris 2009.06, i'll plug it into my Oracle Solaris Express 11 2010.11 system when i have some time to let you know if it works or not.



So, am I pissed off? Sure! ZFS is not living up to expectations. Heck! Not even close!

Sorry this has been a negative experience for you. you started with a lot of incorrect preconceptions. you may want to look into openfiler, it is linux based and does not use zfs.
 

freeflow

Dabbler
Joined
May 29, 2011
Messages
38
I think you now need to revise your opinions. ZFS has delivered fully. Your hardware seems to have been the problem.
 

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
Hi matthewowen01,

Tx! for the input, I'll reply to the comments for other people's sake only, but please note that this issue has now become academic for me since my problem is solved. I also won't re-quote your comments since it will get too confusing. Anyhoo, on with the task:

1 - Regarding zpool import switches: I have tried all of them in any combination you may think of, according to the zpool manual, the "official" Solaris manual, the "official" BSD manual and some other manuals and, let me be absolutely clear on this point, *none* of them worked as expected. I tried them in FreeNAS 8.0 and FreeBSD 8.2. The *only* switch that actually worked was the -X switch that is currently unavailable in the above mentioned OS's.

2 - The -X switch is an un-documented switch. It is not in the manuals. Anywhere in any official manual. I have a deep issue with an FS where the *only* tool that works is an un-documented and un-tested one. And for that matter, it does not even work well. Furthermore, if you check postings all over the net, you will start to see a pattern. And the pattern is always the same. When zpool import fails, people try -X. From what I have seen, it has a 50/50 chance of success. If it does not succeed, then you have to roll back manually. Again, check the posts on how to roll back manually. Or better, don't bother, let's just say that in order to do that, you will need to relinquish remote control of your system to a ZFS geek, it's the only way. So, am I impressed? Most definitively no. It is not acceptable for such a sophisticated FS on its 15th! version not to have good, documented and automated tools for rolling back. It is simply something that one does not do. Not for a FS that is supposed to be "mainframe quality" or "datacenter quality".

3 - Low level corruption. Again, check the posts all over the net. It boils down to this, ZFS is incapable of detecting silent corruption, even though it was supposed to have been build to handle this. Not only can't do it, but is quite difficult to troubleshoot systems with ZFS and such issues. And I am not talking low end hardware, I am talking high end hardware. Now, some people say that ZFS is simply exposing underlying hardware issues that other FS's don't see. Fair enough, but then why it can only detect this kind of corruption on a scrub? I mean, isn't ZFS supposed to be scrubbing all the time on the fly? And how is it possible that this kind of corruption does not happen with other high end FS? I would like to point out that although I am not an IT geek person, I have been around high-end IT systems for over 20 years and I can assure you that there are more ZFS-related problems out there than there are UFS, for example. Then why is the industry using ZFS? Simply because of storage space management capability. Not because of of inherent data safety issues. That's entirely covered by other layers and backups. To my knowledge there are no large IT departments or divisions that use naked ZFS systems for storage. There are always other layers, ZFS is being used *only* as container provider. Again, I am not impressed.

4 - RAM, again, check posts all over the net. Same pattern appears. Low RAM is OK as long as you are not pushing the system. Under such conditions, ZFS bearly uses RAM. However, push the system to a high level and ZFS becomes a RAM hog. Check for data-migration related posts, not daily utilization. You will be surprised of what you will find. In real life (not the theory of what the manual says), you need plenty of RAM if you want your data to be safe. However, the more RAM you have the higher the chance of silent corruption. On a side note, it goes without saying that yes! the more performance you want the more RAM you need. This is not the issue.

5 - Ease of use. Sorry but theoretically speaking, yes it was. I'll refer you to the horse's mouth: the very Sun announcement of ZFS. Check http://web.archive.org/web/20060428092023/http://www.sun.com/2004-0914/feature/ where it stresses "Simple administration
ZFS automates and consolidates complicated storage administration concepts, reducing administrative overhead by 80 percent. " So much for "simple".

6 - Raidz1 and redundancy. There is plenty of redundancy in a standard Zpool without going to a raidz1. I am not talking only about data redundancy (which in a standard zpool isn't) but about indexing and checksuming. If we throw raid1 on top of that, it makes for quite a bit of redundancy to check and double check. ZFS is not doing that (see silent corruption above). ZFS was supposed to be "self-healing" (again, check Sun's announcement). It is not. The -F option does quite little in reality (just Google it) and why did we had to wait until version 15!!!! to have such a switch?. And you still can't scrub it unless imported and you can't import it because it is not scrubbed. Talking about the chiken and the egg! Heck! A lousy NTFS can be cheked and salvaged off-line *even when both tables are corrupted*. ZFS cannot. No, I am definitively not impressed.

7 - Disk additions. Yes, you can expand storage by replacing disks with a higher capacity ones. And yes, you can add top vdevs. However, you can't add disks to any type of raidz since the block pointer rewrite functionality is not "yet" implemented. Yes, I know that people considers ZFS a "mainframe animal" however, I'll point you back to the horse's mouth, where Sun stated: "ZFS meets the needs of a file system for everything from desktops to data centers. " Sure, the statement is true, as long as you don't have any need that is not covered in ZFS. It's like Ford said: People can have a Ford-T model in any color they want, as long as it is black. So, no, again, I am not impressed.

8 - Wrt the 4K hdd issues. Again, sure. Such hdd's do not advertise it. However, the Linux community has come across similar issues many times and have always found ways around it, by hard coding, testing or discovering. Basically it is just Oracle's lazyness that prevents ZFS from discovering and acting upon 4k hdds. And this is not a trivial issue because you have to manually create a zpool with an ashift of 12 in order to get better performance, if not, your performance will get hit by as much as 50%! Now, manual creation of such pools is not trivial, nor well documented, nor understood nor well tested nor anything close to a user-friendly set of instructions. And the same goes for FreeNAS (7 or 8) and FreeBSD and any other OS that uses ZFS to my knowledge. So, no, I am most definitively not impressed.

9 - RAM self-tunning. Yes, once it has set parameters, it tunes itself marvelously. However, and here is the pickle, *without* initial parameters it does bupkus well. Proof? Sure, just start using ZFS undera any OS (except Solaris) and you will find out quite quickly that you get the standard ZFS message saying that it's ussing less than 512 Mb of RAM and you should expect instability. I am talking about parameters such as:

#ZFS kernel tune
vm.kmem_size="512M"
vfs.zfs.arc_min="128M"
vfs.zfs.arc_max="128M"
vfs.zfs.prefetch_disable="0"
vfs.zfs.zil_disable="0"
zfs.zfs.txg.timout="30"
vfs.zfs.vdev.max_pending="35"
vfs.zfs.vdev.min_pending="4"

Can you imagine the user's cry if the evil, evil, MS would have relesed its Windoze Datacentre version *without* a RAM tunning algo? Again, not impressed.

10 - Sil 3114. It works for you in Solaris? Good for you! You are lucky! However, there is NO official support in Express 11 for such cards. Period!. In order to try to get it to work I had to manually load the Si3124 driver which has been reported that "may" work with a Sil3114. Well, it does not. At least not with my card. Sure, the driver loads and sure, the PCI card is visible in Express, however, no hdd is visible nor accessible. Again, not impressed, and I like Solaris.

11 - Openfiler, yes, I looked into it. It's OK, but I discarded it since, precisely, it does not have ZFS. Originally, looking at ZFS, it looked *so* superior that how could I not try it out!? In retrospect, it was a grave mistake. Now it's too late. I am in. I simply don't have the budget to buy a brand new array of hdd's to re-migrate about 5 Tb of data from a raidz1 into a plain raid5. Is ZFS oversold in its capacity? In my view, absolutely yes! It is not user friendly, it is finiky and most definitively un-forgiving. This is not to say that people may not get lucky, particularly if you can afford good quality hardware. However, for the daily enthusiast, the diy type of person, I would definitively discourage its use.

12 - Lastly, I am not saying tha ZFS is not a wonderful idea. All I am saying is that it should be realistically treated for what it is: a beta sofware still heavily under development.

Anyhoo, my 2c.

Again, I would like to reiterate that despite our differences in opinion, I certainly value very much indeed your willingness to help a fellow ZFS'er in need.
Tx! again!.

Again, I would like to reiterate that despite our differences in opinion, I certainly value very much indeed your willingness to help a fellow ZFS'er in need.
Tx! again!.
 

globus999

Contributor
Joined
Jun 9, 2011
Messages
105
I think you now need to revise your opinions. ZFS has delivered fully. Your hardware seems to have been the problem.

Hi, and again, many tx! for all your help. Couldn't have done it without you.
However, wrt my opinions, sorry but I have to answer on the negative, pls see the post above to Matt.

Yes, the main corruption issue was the upgrade from 2Gb to 4Gb, and yes, this was most definitively a hardware issue. Wrt low level corruption, I still contend that it is a memory issue under heavy load. Now that I have finished heavy data migration and am back to day to day operations, I have run a few tests and nope, there is no low level corruption.

In other words:

Low RAM + super heavy data migration (i.e. high number of files) = low level corruption
Low RAM + daily utilization = no low level corruption.

Those are my conclusions.... so far (however I do reserve the right to self-correct me if I find otherwise in the future) :)

Again, many, many tx!
 

Martijn81

Cadet
Joined
Jun 20, 2011
Messages
2
9 - RAM self-tunning. Yes, once it has set parameters, it tunes itself marvelously. However, and here is the pickle, *without* initial parameters it does bupkus well. Proof? Sure, just start using ZFS undera any OS (except Solaris) and you will find out quite quickly that you get the standard ZFS message saying that it's ussing less than 512 Mb of RAM and you should expect instability.

I made the mistake of installing the wrong architecture on a Intel 64 bit CPU. I installed I386 instead of AMD64. I also got that standard ZFS messages when running i386.
Maybe you installed the wrong architecture?

No tuning required when running on AMD64.
 

joeschmuck

Old Man
Moderator
Joined
May 28, 2011
Messages
10,994
You know FreeNAS is in Beta and not formal release plus it's free. You don't have to use it. Many people have informed you that your system is likely the cause of your problems and I think you need to move to UDF if you will be running 2GB RAM.

As for ZFS, there is nothing wrong with it, it was designed for the administrator level computer user, not the typical home user so the interface is purely command line like the old DOS days. I thing the developers are doing a great job and eventually they will have a nice final product but until then it's Beta.

I saw your trouble ticket 524 which slammed the developers for a unacceptable product. After reading several of your postings I see you as a bitchy spoiled kid with nothing nice to say to anyone. My words of advice is if you really feel that way, move on to WHS 2011, it's a nice new product and I'm sure it works great, then you can contact MS when something doesn't work to your satisfaction.
 
Status
Not open for further replies.
Top