SOLVED "zpool status -v" and "zpool status -x" give conflicting information under 11.2-RC1

Status
Not open for further replies.

Doogie

Dabbler
Joined
Mar 9, 2014
Messages
24
I had a backplane fail on an 11.2-RC1 test system. In the process of troubleshooting, one of the disks was physically removed from a 10 HDD RAIDZ1 array before being offlined. This HDD was RMAd and no longer available. A replacement drive verified as good was used for replacement and physically connected. The array remains degraded, and the documentation does not address my current situation of a drive that is listed as "UNAVAIL" and not having an "option" button to select "replace" as per documentation (https://doc.freenas.org/11.2/storage.html#replacing-a-failed-disk). I have rebooted and attempted to troubleshoot by reviewing prior posts on this forum as well as reading the documentation, to no avail.

upload_2018-11-9_8-5-55.png



My main question is how to address this issue. However, as a secondary concern, I am seeing conflicting information from the command line. "zpool status -x" states the pool is healthy, but the -v parameter status the pool is degraded (which it is):

---------------------------------------------------------
Code:
root@Offsite1:~ # zpool status -x
all pools are healthy

root@Offsite1:~ # zpool status -v
  pool: Volume
 state: DEGRADED
status: Some supported features are not enabled on the pool. The pool can
   still be used, but some features are unavailable.
action: Enable all features using 'zpool upgrade'. Once this is done,
   the pool may no longer be accessible by software that does not support
   the features. See zpool-features(7) for details.
  scan: resilvered 458G in 6 days 15:23:37 with 0 errors on Tue Nov  6 03:50:44 2018
config:

   NAME											STATE	 READ WRITE CKSUM
   Volume										  DEGRADED	 0	 0	 0
	 raidz1-0									  DEGRADED	 0	 0	 0
	   gptid/a7601136-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0
	   replacing-1								 UNAVAIL	  0	 0	 0
		 10210022842724997111					  UNAVAIL	  0	 0	 0  was /dev/gptid/0f38e63e-daab-11e8-a7a0-60a44c42e357
		 9750869321900414361					   UNAVAIL	  0	 0	 0  was /dev/gptid/66e9fe9b-dc71-11e8-b980-60a44c42e357
	   gptid/a9369f8b-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0
	   gptid/aa46726d-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0
	   gptid/ab281f76-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0
	   gptid/ac375292-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0
	   gptid/ad432bb1-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0
	   gptid/ae18ccc3-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0
	   gptid/aef869b3-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0
	   gptid/afe7a69f-9014-11e7-9b37-60a44c42e357  ONLINE	   0	 0	 0

errors: No known data errors

root@Offsite1:~ # glabel status
									  Name  Status  Components
gptid/afe7a69f-9014-11e7-9b37-60a44c42e357	 N/A  ada0p2
gptid/a7601136-9014-11e7-9b37-60a44c42e357	 N/A  ada1p2
gptid/a9369f8b-9014-11e7-9b37-60a44c42e357	 N/A  da0p2
gptid/aa46726d-9014-11e7-9b37-60a44c42e357	 N/A  da1p2
gptid/ab281f76-9014-11e7-9b37-60a44c42e357	 N/A  da2p2
gptid/ac375292-9014-11e7-9b37-60a44c42e357	 N/A  da3p2
gptid/ad432bb1-9014-11e7-9b37-60a44c42e357	 N/A  da4p2
gptid/ae18ccc3-9014-11e7-9b37-60a44c42e357	 N/A  da5p2
gptid/aef869b3-9014-11e7-9b37-60a44c42e357	 N/A  da6p2
gptid/1aeb4cce-cf82-11e6-bb7c-0021ccdbce2c	 N/A  da8p1
gptid/afd55912-9014-11e7-9b37-60a44c42e357	 N/A  ada0p1

root@Offsite1:~ # camcontrol devlist
<ATA ST8000DM002-1YW1 DN02>		at scbus0 target 25 lun 0 (pass0,da0)
<ATA ST8000DM002-1YW1 DN02>		at scbus0 target 26 lun 0 (pass1,da1)
<ATA ST8000DM002-1YW1 DN02>		at scbus0 target 27 lun 0 (pass2,da2)
<ATA ST8000DM002-1YW1 DN02>		at scbus0 target 28 lun 0 (pass3,da3)
<ATA ST8000DM002-1YW1 DN02>		at scbus0 target 29 lun 0 (pass4,da4)
<ATA ST8000DM002-1YW1 DN02>		at scbus0 target 30 lun 0 (pass5,da5)
<ATA ST8000DM002-1YW1 DN02>		at scbus0 target 31 lun 0 (pass6,da6)
<ATA ST8000DM004-2CX1 0001>		at scbus0 target 37 lun 0 (pass7,da7)
<ST8000DM002-1YW112 DN02>		  at scbus4 target 0 lun 0 (pass8,ada0)
<ST8000DM002-1YW112 DN02>		  at scbus5 target 0 lun 0 (pass9,ada1)
< Patriot Memory PMAP>			 at scbus7 target 0 lun 0 (pass10,da8)


----------------------------------------------------

I kindly request the assistance of the FreeNAS community in helping me get the pool back to healthy status. How can I remove the "UNAVAIL" entries and resilver using my new disk, either from the command line or GUI? Thanks!

Jeff Arnholt
 

Chris Moore

Hall of Famer
Joined
May 2, 2015
Messages
10,080
this is just one of the many reasons why you're not supposed to use RAIDz1. now you have no redundancy. if another disk fails, you lose the entire storage pool. start making a backup.

Sent from my SAMSUNG-SGH-I537 using Tapatalk
 

Chris Moore

Hall of Famer
Joined
May 2, 2015
Messages
10,080
PS. you should submit a bug report

Sent from my SAMSUNG-SGH-I537 using Tapatalk
 

Doogie

Dabbler
Joined
Mar 9, 2014
Messages
24
This is a test system. I have multiple redundant copies of the data. However, it takes me about 11 days to restore 60TB over my network connection, so I'd prefer to repair the pool rather than deleting and restoring it.

Any thoughts as to how to remove and restore a missing disk? What are you supposed to do when the "UNAVAIL" device has no options to remove, replace, etc?
 

Ericloewe

Server Wrangler
Moderator
Joined
Feb 15, 2014
Messages
20,194
Nothing to see here. From the zpool(8) man page:

-x Only display status for pools that are exhibiting errors or
are otherwise unavailable. Warnings about pools not using
the latest on-disk format
, having non-native block size or
disabled features will not be included.
 

Ericloewe

Server Wrangler
Moderator
Joined
Feb 15, 2014
Messages
20,194
I agree that it is odd behavior, but given the aversion to removing features in ZFS, I think it's easiest to just forget that option exists.
 

Doogie

Dabbler
Joined
Mar 9, 2014
Messages
24
I was able to solve my problem of removing UNAVAIL ghost drive(s) through a post on the Ubuntu forums.

(see http://askubuntu.com/questions/305830/replacing-a-dead-disk-in-a-zpool)

zpool offline [pool] [ghost_guid]
(for example: zpool offline Volume 18058307076531860683)

One this is done, drive should change from UNAVAIL to OFFLINE in Storage | Pools | Pool Status and have options buttons restored in GUI. Choose "replace" using new drive under dropdown (for example, "da7"), and resilvering starts. For 60TB pool, takes 48 hrs.

Command to do this instead is: zpool replace [pool] [ghost_guid] [new dev]
(for example: zpool replace Volume 18058307076531860683 /dev/da15)

It may be necessary to first detach UNAVAIL device: zpool detach Volume /dev/gptid/[string]
(for example: zpool detach Volume /dev/gptid/0f38e63e-daab-11e8-a7a0-60a44c42e357)

I don't see a way for me to marked this "solved", so moderators, if possible, please do so. Thank you.
 
Last edited:

Ericloewe

Server Wrangler
Moderator
Joined
Feb 15, 2014
Messages
20,194
Oh, you had a different problem and I completely missed it. D'oh.
 

Ericloewe

Server Wrangler
Moderator
Joined
Feb 15, 2014
Messages
20,194
(For future reference, you should have an edit thread option near the top, you can change the prefix to SOLVED.)
 
Status
Not open for further replies.
Top