I have a problem on one of my systems with zfs. I used to have zpool created 
with 3 luns on SAN. I did not have to put any raid or anything on it since it 
was already using raid on SAN. Anyway server rebooted and I cannot zee my 
pools. 
When I do try to import it it does fail. I am using EMC Clarion as SAN and 
powerpath
# zpool list
no pools available
# zpool import -f
  pool: mypool
  id: 4148251638983938048
state: FAULTED
status: One or more devices are missing from the system.
action: The pool cannot be imported. Attach the missing
  devices and try again.
  see: http://www.sun.com/msg/ZFS-8000-3C
config:
  mypool UNAVAIL insufficient replicas
  emcpower0a UNAVAIL cannot open
  emcpower2a UNAVAIL cannot open
  emcpower3a ONLINE

I think I am able to see all the luns and I should be able to access them on my 
sun box.
# powermt display dev=all
Pseudo name=emcpower0a
CLARiiON ID=APM00070202835 [NRHAPP02]
Logical device ID=6006016045201A001264FB20990FDC11 [LUN 13]
state=alive; policy=CLAROpt; priority=0; queued-IOs=0
Owner: default=SP B, current=SP B
==============================================================================
---------------- Host --------------- - Stor - -- I/O Path - -- Stats ---
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
3074 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c2t5006016041E035A4d0s0 SP A4 active 
alive 0 0
3074 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c2t5006016941E035A4d0s0 SP B5 active 
alive 0 0
3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c3t5006016141E035A4d0s0 SP A5 
active alive 0 0
3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c3t5006016841E035A4d0s0 SP B4 
active alive 0 0


Pseudo name=emcpower1a
CLARiiON ID=APM00070202835 [NRHAPP02]
Logical device ID=6006016045201A004C1388343C10DC11 [LUN 14]
state=alive; policy=CLAROpt; priority=0; queued-IOs=0
Owner: default=SP B, current=SP B
==============================================================================
---------------- Host --------------- - Stor - -- I/O Path - -- Stats ---
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
3074 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c2t5006016041E035A4d1s0 SP A4 active 
alive 0 0
3074 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c2t5006016941E035A4d1s0 SP B5 active 
alive 0 0
3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c3t5006016141E035A4d1s0 SP A5 
active alive 0 0
3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c3t5006016841E035A4d1s0 SP B4 
active alive 0 0


Pseudo name=emcpower3a
CLARiiON ID=APM00070202835 [NRHAPP02]
Logical device ID=6006016045201A00A82C68514E86DC11 [LUN 7]
state=alive; policy=CLAROpt; priority=0; queued-IOs=0
Owner: default=SP B, current=SP B
==============================================================================
---------------- Host --------------- - Stor - -- I/O Path - -- Stats ---
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
3074 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c2t5006016041E035A4d3s0 SP A4 active 
alive 0 0
3074 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c2t5006016941E035A4d3s0 SP B5 active 
alive 0 0
3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c3t5006016141E035A4d3s0 SP A5 
active alive 0 0
3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c3t5006016841E035A4d3s0 SP B4 
active alive 0 0


Pseudo name=emcpower2a
CLARiiON ID=APM00070202835 [NRHAPP02]
Logical device ID=600601604B141B00C2F6DB2AC349DC11 [LUN 24]
state=alive; policy=CLAROpt; priority=0; queued-IOs=0
Owner: default=SP B, current=SP B
==============================================================================
---------------- Host --------------- - Stor - -- I/O Path - -- Stats ---
### HW Path I/O Paths Interf. Mode State Q-IOs Errors
==============================================================================
3074 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c2t5006016041E035A4d2s0 SP A4 active 
alive 0 0
3074 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c2t5006016941E035A4d2s0 SP B5 active 
alive 0 0
3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c3t5006016141E035A4d2s0 SP A5 
active alive 0 0
3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 c3t5006016841E035A4d2s0 SP B4 
active alive 0 0


So format does show them as well
bash-3.00# echo | format
Searching for disks...done
AVAILABLE DISK SELECTIONS:
  0. c1t0d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424>
  /[EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL 
PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
  1. c1t1d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424>
  /[EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL 
PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
  2. c1t2d0 <SEAGATE-ST973401LSUN72G-0556-68.37GB>
  /[EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL 
PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
  3. c1t3d0 <SEAGATE-ST973401LSUN72G-0556-68.37GB>
  /[EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL 
PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
  4. c2t5006016941E035A4d0 <DGC-RAID5-0324 cyl 51198 alt 2 hd 256 sec 16>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],0
  5. c2t5006016041E035A4d0 <DGC-RAID5-0324 cyl 51198 alt 2 hd 256 sec 16>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],0
  6. c2t5006016941E035A4d1 <DGC-RAID5-0324 cyl 32766 alt 2 hd 64 sec 10>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],1
  7. c2t5006016041E035A4d1 <DGC-RAID5-0324 cyl 32766 alt 2 hd 64 sec 10>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],1
  8. c2t5006016041E035A4d2 <DGC-RAID5-0324 cyl 32766 alt 2 hd 256 sec 10>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],2
  9. c2t5006016941E035A4d2 <DGC-RAID5-0324 cyl 32766 alt 2 hd 256 sec 10>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],2
  10. c2t5006016041E035A4d3 <DGC-RAID5-0324 cyl 63998 alt 2 hd 256 sec 16>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],3
  11. c2t5006016941E035A4d3 <DGC-RAID5-0324 cyl 63998 alt 2 hd 256 sec 16>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],3
  12. c3t5006016841E035A4d0 <DGC-RAID5-0324 cyl 51198 alt 2 hd 256 sec 16>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],0
  13. c3t5006016141E035A4d0 <DGC-RAID5-0324 cyl 51198 alt 2 hd 256 sec 16>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],0
  14. c3t5006016141E035A4d1 <DGC-RAID5-0324 cyl 32766 alt 2 hd 64 sec 10>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],1
  15. c3t5006016841E035A4d1 <DGC-RAID5-0324 cyl 32766 alt 2 hd 64 sec 10>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],1
  16. c3t5006016141E035A4d2 <DGC-RAID5-0324 cyl 32766 alt 2 hd 256 sec 10>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],2
  17. c3t5006016841E035A4d2 <DGC-RAID5-0324 cyl 32766 alt 2 hd 256 sec 10>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],2
  18. c3t5006016841E035A4d3 <DGC-RAID5-0324 cyl 63998 alt 2 hd 256 sec 16>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],3
  19. c3t5006016141E035A4d3 <DGC-RAID5-0324 cyl 63998 alt 2 hd 256 sec 16>
  /[EMAIL PROTECTED],700000/[EMAIL PROTECTED],2/SUNW,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0/[EMAIL PROTECTED],3
  20. emcpower0a <DGC-RAID5-0324 cyl 51198 alt 2 hd 256 sec 16>
  /pseudo/[EMAIL PROTECTED]
  21. emcpower1a <DGC-RAID5-0324 cyl 32766 alt 2 hd 64 sec 10>
  /pseudo/[EMAIL PROTECTED]
  22. emcpower2a <DGC-RAID5-0324 cyl 32766 alt 2 hd 256 sec 10>
  /pseudo/[EMAIL PROTECTED]
  23. emcpower3a <DGC-RAID5-0324 cyl 63998 alt 2 hd 256 sec 16>
  /pseudo/[EMAIL PROTECTED]
Specify disk (enter its number): Specify disk (enter its number):



Now the fun part with troubleshooting this.

When I do zdb on emcpower3a which seems to be ok from zpool perspective I get 
the following output:
bash-3.00# zdb -lv /dev/dsk/emcpower3a
--------------------------------------------
LABEL 0
--------------------------------------------
  version=3
  name='mypool'
  state=0
  txg=4367380
  pool_guid=4148251638983938048
  top_guid=9690155374174551757
  guid=9690155374174551757
  vdev_tree
  type='disk'
  id=2
  guid=9690155374174551757
  path='/dev/dsk/emcpower3a'
  whole_disk=0
  metaslab_array=1813
  metaslab_shift=30
  ashift=9
  asize=134208815104
--------------------------------------------
LABEL 1
--------------------------------------------
  version=3
  name='mypool'
  state=0
  txg=4367380
  pool_guid=4148251638983938048
  top_guid=9690155374174551757
  guid=9690155374174551757
  vdev_tree
  type='disk'
  id=2
  guid=9690155374174551757
  path='/dev/dsk/emcpower3a'
  whole_disk=0
  metaslab_array=1813
  metaslab_shift=30
  ashift=9
  asize=134208815104
--------------------------------------------
LABEL 2
--------------------------------------------
  version=3
  name='mypool'
  state=0
  txg=4367380
  pool_guid=4148251638983938048
  top_guid=9690155374174551757
  guid=9690155374174551757
  vdev_tree
  type='disk'
  id=2
  guid=9690155374174551757
  path='/dev/dsk/emcpower3a'
  whole_disk=0
  metaslab_array=1813
  metaslab_shift=30
  ashift=9
  asize=134208815104
--------------------------------------------
LABEL 3
--------------------------------------------
  version=3
  name='mypool'
  state=0
  txg=4367380
  pool_guid=4148251638983938048
  top_guid=9690155374174551757
  guid=9690155374174551757
  vdev_tree
  type='disk'
  id=2
  guid=9690155374174551757
  path='/dev/dsk/emcpower3a'
  whole_disk=0
  metaslab_array=1813
  metaslab_shift=30
  ashift=9
  asize=134208815104



But when I do zdb on emcpower0a which seems to be not that ok and get the 
following output:
bash-3.00# zdb -lv /dev/dsk/emcpower0a
--------------------------------------------
LABEL 0
--------------------------------------------
  version=3
  name='mypool'
  state=0
  txg=4367379
  pool_guid=4148251638983938048
  top_guid=14125143252243381576
  guid=14125143252243381576
  vdev_tree
  type='disk'
  id=0
  guid=14125143252243381576
  path='/dev/dsk/emcpower0a'
  whole_disk=0
  metaslab_array=13
  metaslab_shift=29
  ashift=9
  asize=107365269504
  DTL=727
--------------------------------------------
LABEL 1
--------------------------------------------
  version=3
  name='mypool'
  state=0
  txg=4367379
  pool_guid=4148251638983938048
  top_guid=14125143252243381576
  guid=14125143252243381576
  vdev_tree
  type='disk'
  id=0
  guid=14125143252243381576
  path='/dev/dsk/emcpower0a'
  whole_disk=0
  metaslab_array=13
  metaslab_shift=29
  ashift=9
  asize=107365269504
  DTL=727
--------------------------------------------
LABEL 2
--------------------------------------------
failed to read label 2
--------------------------------------------
LABEL 3
--------------------------------------------
failed to read label 3


that also is the same for emcpower2a in my pool.

Is there a way to be able to fix failed LABELs 2 and 3? I know you need 4 of 
them, but is there a way to reconstruct them in any way? Or is my pool lost 
completely and I need to recreate it? It would be off that reboot of a server 
could cause such disaster. But I was unable to find anywhere where people would 
be able to repair or recreate those LABELS. How would I recover my zpools? Any 
help or suggestion is greatly appreciated.

Regards,

Chris


_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to