Hi all,

I recently created a RAID-Z1 pool out of a set of 7 SCSI disks, using the following command:

# zpool create magicant raidz c5t0d0 c5t1d0 c5t2d0 c5t3d0 c5t4d0 c5t5d0 c5t6d0

It worked fine, but I was slightly confused by the size yield (99 GB vs the 116 GB I had on my other RAID-Z1 pool of same-sized disks).

I thought one of the disks might have been to blame, so I tried swapping it out - it turned out my replacement disk was a dud (zpool wasn't happy about that, and eventually offlined the disk). Oh well, swap the old one back in, no harm done.

Reboot, and ZFS informs me that I'm missing another, unrelated disk (c5t1d0 was the one I tried swapping out unsuccessfully - c5t3d0 is the one it complained about, which I had swapped for another disk before any problems began or any data was in the pool, with no problems - ZFS scrubbed and was happy).

It continually claimed the device was unavailable and so the pool was in degraded mode - attempting to replace the disk with itself yielded an error about the disk being in use by the same pool which claimed the disk was unavailable. Unmount the pool, same error persists, zpool replace continues to give that error, despite repeated zpool offline magicant  c5t3d0 followed by zpool online [etc].

I try exporting and re-importing the pool - the export went fine. The import threw the confusing error which is the point of this email:

# zpool import
  pool: magicant
    id: 3232403590553596936
 state: FAULTED
status: One or more devices are missing from the system.
action: The pool cannot be imported. Attach the missing
        devices and try again.
   see: http://www.sun.com/msg/ZFS-8000-6X
config:

        magicant    UNAVAIL   missing device
          raidz1    ONLINE
            c5t0d0  ONLINE
            c5t1d0  ONLINE
            c5t2d0  ONLINE
            c5t3d0  ONLINE
            c5t4d0  ONLINE
            c5t5d0  ONLINE
            c5t6d0  ONLINE

        Additional devices are known to be part of this pool, though their
        exact configuration cannot be determined.

So, to summarize:
7-scsi-disk raidz1 zpool is created
c5t3d0 is swapped out for another disk of identical size, zfs is happy and functions fine after a scrub
c5t1d0 is swapped out for another disk of identical size (which happened to be a dud), Solaris didn't like that, so I put the original back in and rebooted
On boot, zpool claims c5t3d0 is unavailable, while format and cfgadm both agree that the disk still exists and is dandy. zpool replace pool c5t3d0 c5t3d0 claims it's in use by that pool, zpool offline pool c5t3d0 followed by zpool online pool c5t3d0 doesn't help. zpool export pool worked, but then zpool import pool threw the above error.

Is this a bug, or am I missing something obvious?

snv 44, x86.

- Rich
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to