RickH
Explorer
- Joined
- Oct 31, 2014
- Messages
- 61
I've been experimenting with FreeNAS on some older hardware for a while and I finally decided to implement FreeNAS in a production environment. I work for a small family owned service company that scans documents for our clients. The new storage server is going to be used for several roles:
Being a small business, money is always an issue, but I've put together the following server to try and meet our needs:
I have everything assembled and have been doing some initial testing over the past few days. I have set up an 11 drive RAID Z3 volume (I have ordered a 12th drive to use as a hot spare) using the Intel 313 SSD as a log drive (I have sync=always enabled for the ISCSI Dataset). The NIC's are set up in a LCAP aggregation and overall performance has been excellent. I have been able to max out the Gigabit connections on all of my test clients and have seen excellent performance even with multiple clients simultaneously uploading and downloading files.
My only concern is a 'SCSI Status Error' that keeps showing up in my logs. (see attached files)
I see this error repeated continuously for several of my drives:
I have run and attached the following hoping someone out there has some suggestions.
You can see that this 'reset' seems to interrupt the SMART short tests I have scheduled, but I don't see any other obvious errors on any of the drives.
NOTE: Please disregard /dev/da11 - this drive isn't part of the pool, it was just a spare I stuck in the chasis to test the midplane and experiment with jails...
The only reference I can find to this error in the forums has been attributed to either a power supply issue or possible cabling issues....
I'm running dual power supplies so I'm fairly confident that isn't the issue. I did have to purchase a SFF-8484 to SFF-8087 cable to connect my SAS9210 to the midplane so there's a small chance that the issue is there (although the fact it's only showing on the WD drives makes me wonder). I have ordered a new cable just to make sure but I'm really hoping someone out there has some experience with this problem.
1) Backup location for our main servers (all being backed up using ShadowProtect) over CIFS
2) Working storage for our scanning software - lots of simultaneous users reading and writing small image files (typically 30-60 kb in size) over CIFS
3) ESXi datastore over ISCSI
2) Working storage for our scanning software - lots of simultaneous users reading and writing small image files (typically 30-60 kb in size) over CIFS
3) ESXi datastore over ISCSI
Being a small business, money is always an issue, but I've put together the following server to try and meet our needs:
Dell FS12-SC (basically a C2100 that was sold in bulk to big data-centers)
24GB ECC ram
2x onboard Intel PRO1000 NIC
Intel PRO1000 4-port PCI-e NIC
LSI SAS9210-8i
Generic 8GB SSD for boot
Intel 313 24GB SSD for Log
Drives (because of cost concerns I had to re-use some drives, but all passed a full surface SMART test prior to being installed)12 hot swap 3.5" trays, dual redundant power supplies, 2 internal 2.5" trays,
and a Newisys SAS/SATA midplane
2x Xeon L5410 @ 2.33 GHzand a Newisys SAS/SATA midplane
24GB ECC ram
2x onboard Intel PRO1000 NIC
Intel PRO1000 4-port PCI-e NIC
LSI SAS9210-8i
Generic 8GB SSD for boot
Intel 313 24GB SSD for Log
11x 2TB drives (Mixture of WD Black, WD Green, and Hitachi)
I have everything assembled and have been doing some initial testing over the past few days. I have set up an 11 drive RAID Z3 volume (I have ordered a 12th drive to use as a hot spare) using the Intel 313 SSD as a log drive (I have sync=always enabled for the ISCSI Dataset). The NIC's are set up in a LCAP aggregation and overall performance has been excellent. I have been able to max out the Gigabit connections on all of my test clients and have seen excellent performance even with multiple clients simultaneously uploading and downloading files.
My only concern is a 'SCSI Status Error' that keeps showing up in my logs. (see attached files)
I see this error repeated continuously for several of my drives:
"SCSI sense: UNIT ATTENTION asc:29,0 (Power on, reset, or bus device reset occurred)"
This error seems to only occur on my WD drives (both the Green's and the Black's) and I have yet to see it on any of the 3 Hitachi's. I don't notice any performance issues and a full scrub reveals no data or checksum errors.I have run and attached the following hoping someone out there has some suggestions.
camcontrol devlist
zpool status
smartctl -a /dev/da## (for each drive)
zpool status
smartctl -a /dev/da## (for each drive)
You can see that this 'reset' seems to interrupt the SMART short tests I have scheduled, but I don't see any other obvious errors on any of the drives.
NOTE: Please disregard /dev/da11 - this drive isn't part of the pool, it was just a spare I stuck in the chasis to test the midplane and experiment with jails...
The only reference I can find to this error in the forums has been attributed to either a power supply issue or possible cabling issues....
I'm running dual power supplies so I'm fairly confident that isn't the issue. I did have to purchase a SFF-8484 to SFF-8087 cable to connect my SAS9210 to the midplane so there's a small chance that the issue is there (although the fact it's only showing on the WD drives makes me wonder). I have ordered a new cable just to make sure but I'm really hoping someone out there has some experience with this problem.
Attachments
Last edited: