On 7/13/07, [EMAIL PROTECTED] <[EMAIL PROTECTED]> wrote: > > Can you post a "powermt display dev=all", a zpool status and format > command?
Sure. There are no pools to give status on because I can't import them. For the others: # powermt display dev=all Pseudo name=emcpower0a CLARiiON ID=APM00043600837 [########] Logical device ID=600601600C4912003AB4B247BA2BDA11 [LUN 46] state=alive; policy=CLAROpt; priority=0; queued-IOs=0 Owner: default=SP B, current=SP B ============================================================================== ---------------- Host --------------- - Stor - -- I/O Path - -- Stats --- ### HW Path I/O Paths Interf. Mode State Q-IOs Errors ============================================================================== 3073 [EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 c2t500601613060099Cd1s0 SP A1 active alive 0 0 3073 [EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 c2t500601693060099Cd1s0 SP B1 active alive 0 0 3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 c3t500601603060099Cd1s0 SP A0 active alive 0 0 3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 c3t500601683060099Cd1s0 SP B0 active alive 0 0 Pseudo name=emcpower1a CLARiiON ID=APM00043600837 [########] Logical device ID=600601600C4912004C5CFDFFB62BDA11 [LUN 0] state=alive; policy=CLAROpt; priority=0; queued-IOs=0 Owner: default=SP A, current=SP A ============================================================================== ---------------- Host --------------- - Stor - -- I/O Path - -- Stats --- ### HW Path I/O Paths Interf. Mode State Q-IOs Errors ============================================================================== 3073 [EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 c2t500601613060099Cd0s0 SP A1 active alive 0 0 3073 [EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 c2t500601693060099Cd0s0 SP B1 active alive 0 0 3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 c3t500601603060099Cd0s0 SP A0 active alive 0 0 3072 [EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 c3t500601683060099Cd0s0 SP B0 active alive 0 0 AVAILABLE DISK SELECTIONS: 0. c1t0d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 1. c1t1d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 2. c2t500601613060099Cd0 <DGC-RAID 5-0219-500.00GB> /[EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0/[EMAIL PROTECTED],0 3. c2t500601693060099Cd0 <DGC-RAID 5-0219-500.00GB> /[EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0/[EMAIL PROTECTED],0 4. c2t500601613060099Cd1 <DGC-RAID 5-0219-500.00GB> /[EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0/[EMAIL PROTECTED],1 5. c2t500601693060099Cd1 <DGC-RAID 5-0219-500.00GB> /[EMAIL PROTECTED],600000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0/[EMAIL PROTECTED],1 6. c3t500601683060099Cd0 <DGC-RAID 5-0219-500.00GB> /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0/[EMAIL PROTECTED],0 7. c3t500601603060099Cd0 <DGC-RAID 5-0219-500.00GB> /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0/[EMAIL PROTECTED],0 8. c3t500601683060099Cd1 <DGC-RAID 5-0219-500.00GB> /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0/[EMAIL PROTECTED],1 9. c3t500601603060099Cd1 <DGC-RAID 5-0219-500.00GB> /[EMAIL PROTECTED],700000/[EMAIL PROTECTED]/[EMAIL PROTECTED],0/[EMAIL PROTECTED],1 10. emcpower0a <DGC-RAID 5-0219-500.00GB> /pseudo/[EMAIL PROTECTED] 11. emcpower1a <DGC-RAID 5-0219-500.00GB> /pseudo/[EMAIL PROTECTED] > > [EMAIL PROTECTED] wrote on 07/13/2007 09:38:01 AM: > > > How much fun can you have with a simple thing like powerpath? > > > > Here's the story: I have a (remote) system with access to a couple > > of EMC LUNs. Originally, I set it up with mpxio and created a simple > > zpool containing the two LUNs. > > > > It's now been reconfigured to use powerpath instead of mpxio. > > > > My problem is that I can't import the pool. I get: > > > > pool: ###### > > id: ################### > > state: FAULTED > > status: One or more devices are missing from the system. > > action: The pool cannot be imported. Attach the missing > > devices and try again. > > see: http://www.sun.com/msg/ZFS-8000-3C > > config: > > > > disk00 UNAVAIL insufficient replicas > > c3t50060xxxxxxxxxxCd1 ONLINE > > c3t50060xxxxxxxxxxCd0 UNAVAIL cannot open > > > > Now, it's working up to the point at which it's worked out that > > the bits of the pool are in the right places. It just can't open > > all the bits. Why is that? > > > > I notice that it's using the underlying cXtXdX device names > > rather than the virtual emcpower{0,1} names. However, rather > > more worrying is that if I try to create a new pool, then it correctly > > fails if I use the cXtXdX device (warning me that it contains > > part of a pool) but if I go through the emcpower devices > > then I don't get a warning. > > > > (One other snippet - the cXtXdX device nodes look > > slightly odd, in that some of them look like the traditional > > SMI labelled nodes, while some are more in an EFI style > > with a device node for the disk.) > > > > Is there any way to fix this or are we going to have to > > start over? > > > > If we do start over, is powerpath going to behave itself > > or might this sort of issue bite us again in the future? > > > > Thanks for any help or suggestions from any > > powerpath experts. > > > > -- > > -Peter Tribble > > http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/ > > _______________________________________________ > > zfs-discuss mailing list > > zfs-discuss@opensolaris.org > > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > -- -Peter Tribble http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/ _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss