Eric and Richard - thanks for your responses. I tried both:
echo ::spa -c | mcb zdb -C (not much of a man page for this one!) and was able to match the POOL id from the log (hex 4fcdc2c9d60a5810) with both outputs. As Richard pointed out, I needed to convert the hex value to decimal to get a match with the zdb output. In neither case, however, was I able to get a match with the disk vdev id from the fmdump output. It turns out that a disk in this machine was replaced about a month ago, and sure enough the vdev that was complaining at the time was the 0x179e471c0732582 vdev that is now missing. What's confusing is that the fmd message I posted about is dated Oct 22 whereas the original error and replacement happened back in September. An "fmadm faulty" on the machine currently doesn't return any issues. After physically replacing the bad drive and issuing the "zpool replace" command, I think that we probably issued the "fmadm repair <uuid>" command in line with what Sun has asked us to do in the past. In our experience, if you don't do this then fmd will re-issue duplicate complaints regarding hardware failures after every reboot until you do. In this case, perhaps a "repair" wasn't really the appropriate command since we actually replaced the drive. Would a "fmadm flush" have been better? Perhaps a clean reboot is in order? So, it looks like the root problem here is that fmd is confused rather than there being a real issue with ZFS. Despite this, we're happy to know that we can now match vdevs against physical devices using either the mdb trick or zdb. We've followed Eric's work on ZFS device enumeration for the Fishwork project with great interest - hopefully this will eventually get extended to the fmdump output as suggested. Sean Walmsley -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss