storage increase

Status
Not open for further replies.

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
can someone tell me what i am doing wrong if i am at all please...

by using this calculator, http://biduleohm.free.fr/zfsraidsarc/ i came up with this.

Drive size: TB 4
Number of drives: 6
RAID type: (requires at least 5 drives) raidz3
Drive MTBF: Mh
MTTPR: h
Rebuild speed: MiB/s

TiB TB %
Drive size 3.638 4 N/A
Total parity space 10.91 12 50
Total data space 10.91 12 50
Total RAID space 21.83 24 100
Metadata overhead 0.175 0.192 1.6
Blocks overhead 0 0 0
Total overhead 0.175 0.192 1.6
Minimum recommended free space 2.148 2.362 19.68
Usable data space 8.592 9.446 78.72

MTTR: 97.02 h
MTTDL: 1.9 x 10^14 h

now i added these drives by offlining removing and replacing 1 at a time. doing its resilvering process. after all drives were replaced it shows in my pool 18.6 TiB.

however my problem is my dataset is still remaining at 8.4TiB available.(drives i had in before were 2tbx6)

my dataset options are all set to 0 for unlimited. now even though it says 8.4 TiB available will it use the other 10.2 as needed, or will i have to increase my main dataset quota that holds a smb share?

if there is another option i am missing thats much better than please tell me. i plan to increase again soon to 6tb drives. 6x6 and want to utilize all the space i can for this particular dataset. it is not a media server by the way.

thank you in advance.
 

Ericloewe

Server Wrangler
Moderator
Joined
Feb 15, 2014
Messages
20,194
FreeNAS version?

Also, can you provide a screenshot? That seems rather unusual.
 

bigphil

Patron
Joined
Jan 30, 2014
Messages
486
my dataset options are all set to 0 for unlimited. now even though it says 8.4 TiB available will it use the other 10.2 as needed, or will i have to increase my main dataset quota that holds a smb share?
If you have a quota on a parent then all descendants will be limited to that quota amount. So...if the dataset in question is the descendant of a parent dataset that has a quota, this would be expected to see and you'll need to increase the quota limit.
 

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
there are no quotas set at all.

upload_2017-2-15_18-13-23.png


FreeNAS-9.3-STABLE-201605170422
 

Ericloewe

Server Wrangler
Moderator
Joined
Feb 15, 2014
Messages
20,194
Post the output of zpool status please.
 

bigphil

Patron
Joined
Jan 30, 2014
Messages
486
Please post the output from "zfs get all pool"
 

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
Code:
scan: scrub repaired 0 in 3h43m with 0 errors on Sun Feb  5 03:43:29 2017   
config:                                                                       
                                                                              
        NAME                                            STATE     READ WRITE CKS
UM                                                                             
        pool                                            ONLINE       0     0   
0                                                                             
          raidz3-0                                      ONLINE       0     0   
0                                                                             
            gptid/5b728858-38f3-11e6-875f-0030487e5312  ONLINE       0     0   
0                                                                             
            gptid/181156e1-2b97-11e6-875f-0030487e5312  ONLINE       0     0   
0                                                                             
            gptid/bb388ad3-1276-11e6-875f-0030487e5312  ONLINE       0     0   
0                                                                             
            gptid/eb763fba-1727-11e6-875f-0030487e5312  ONLINE       0     0   
0                                                                             
            gptid/2b27b81d-3120-11e6-875f-0030487e5312  ONLINE       0     0   
0                                                                             
            gptid/c07e49f2-10d2-11e6-875f-0030487e5312  ONLINE       0     0   
0                                                                             
                                                                              
errors: No known data errors
 
Last edited:

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
Code:
zfs get all pool
NAME  PROPERTY              VALUE                  SOURCE
pool  type                  filesystem             -
pool  creation              Tue Apr 14 19:33 2015  -
pool  used                  1.49T                  -
pool  available             8.42T                  -
pool  referenced            180K                   -
pool  compressratio         1.05x                  -
pool  mounted               yes                    -
pool  quota                 none                   local
pool  reservation           none                   local
pool  recordsize            128K                   default
pool  mountpoint            /mnt/pool              default
pool  sharenfs              off                    default
pool  checksum              on                     default
pool  compression           lz4                    local
pool  atime                 on                     default
pool  devices               on                     default
pool  exec                  on                     default
pool  setuid                on                     default
pool  readonly              off                    default
pool  jailed                off                    default
pool  snapdir               hidden                 default
pool  aclmode               passthrough            local
pool  aclinherit            passthrough            local
pool  canmount              on                     default
pool  xattr                 off                    temporary
pool  copies                1                      default
pool  version               5                      -
pool  utf8only              off                    -
pool  normalization         none                   -
pool  casesensitivity       sensitive              -
pool  vscan                 off                    default
pool  nbmand                off                    default
pool  sharesmb              off                    default
pool  refquota              none                   local
pool  refreservation        none                   local
pool  primarycache          all                    default
pool  secondarycache        all                    default
pool  usedbysnapshots       0                      -
pool  usedbydataset         180K                   -
pool  usedbychildren        1.49T                  -
pool  usedbyrefreservation  0                      -
pool  logbias               latency                default
pool  dedup                 off                    default
pool  mlslabel                                     -
pool  sync                  standard               default
pool  refcompressratio      1.00x                  -
pool  written               180K                   -
pool  logicalused           1.55T                  -
pool  logicalreferenced     9.50K                  -
pool  volmode               default                default
pool  filesystem_limit      none                   default
pool  snapshot_limit        none                   default
pool  filesystem_count      none                   default
pool  snapshot_count        none                   default
pool  redundant_metadata    all                    default
 
Last edited:

Jailer

Not strong, but bad
Joined
Sep 12, 2014
Messages
4,977
You have a RAIDz3 pool consisting of 6 4TB drives. Half your drives are being used for parity. The available space being shown is correct.
 

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
hmmm ok. i am going to look at my remote backup and see what that has. i started that one with 4tb drives from the get go.
 

Ericloewe

Server Wrangler
Moderator
Joined
Feb 15, 2014
Messages
20,194
[CODE][/CODE] tags please.
 

bigphil

Patron
Joined
Jan 30, 2014
Messages
486
I'm no mathematician, but something seems off in the GUI, but you are running an old version of FreeNAS, so maybe that's it. The first line is listing the Pool available space and I don't know how it could be 18.6TiB when your max would be 10.9TiB with your disk layout. output of "zpool list" please.
 

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
Code:
zpool list
NAME           SIZE  ALLOC   FREE  EXPANDSZ   FRAG    CAP  DEDUP  HEALTH  ALTROO                    T
freenas-boot  7.19G  5.15G  2.04G         -      -    71%  1.00x  ONLINE  -
pool          21.8T  3.17T  18.6T         -     4%    14%  1.00x  ONLINE  /mnt
 

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
i have 3 systems. i tried to update my newest one that is a plex server to 9.10.2u1 and it hung on loading kernel modules. had to go back to 9.10.1-u2. the other two i was going update if all was well but it wasnt so i am not. cant chance these two to be updated and it fails if the main one isnt even in front of me. im doing this all remotely.
 

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
my remote backup server is having gui issues and wont load my pool. cant even check for updates. thats running the same version as well. sits at loading then 5-10 mins later Sorry, an error occurred.
 

tvalenti03

Dabbler
Joined
Feb 15, 2017
Messages
12
Code:
zpool list                                                    
NAME           SIZE  ALLOC   FREE  EXPANDSZ   FRAG    CAP  DEDUP  HEALTH  ALTROO
T                                                                              
freenas-boot  7.19G  3.00G  4.19G         -      -    41%  1.00x  ONLINE  -    
pool          18.1T   936G  17.2T         -     1%     5%  1.00x  ONLINE  /mnt 


this is from my remote backup. its running 5 drives instead of 6 but they are 4tb each.
 

bigphil

Patron
Joined
Jan 30, 2014
Messages
486
Ok, I found out whats going on...."zpool list" is what the GUI first line is showing and this reports actual physical space available to the pool (6x4=24TB or 21.8TiB). So, like @Jailer said, your available space is correct for your disk layout (as the output of "zfs list" shows). It's explained in the zpool(8) man page
 
Status
Not open for further replies.
Top