Hi all,

IHACWHAC (I have a colleague who has a customer - hello, if you're listening :-) who's trying to build and test a scenario where he can salvage the data off the (internal ?) disks of a T2000 in case the sysboard and with it the on-board raid controller dies.

If I understood correctly, he replaces the motherboard, does some magic to get the raid config back, but even when raidctl says "I'm fine", zpool complains that it cannot open one of the replicas:

# raidctl
RAID    Volume  RAID            RAID            Disk
Volume  Type    Status          Disk            Status
------------------------------------------------------
c0t0d0  IM      OK              c0t0d0          OK
                                c0t1d0          OK
c0t2d0  IM      OK              c0t2d0          OK
                                c0t3d0          OK

# zpool status -x
  pool: dpool
 state: FAULTED
status: One or more devices could not be opened.  There are insufficient
        replicas for the pool to continue functioning.
action: Attach the missing device and online it using 'zpool online'.
   see: http://www.sun.com/msg/ZFS-8000-D3
 scrub: none requested
config:

        NAME        STATE     READ WRITE CKSUM
        dpool       UNAVAIL      0     0     0  insufficient replicas
          c0t2d0    UNAVAIL      0     0     0  cannot open
#

what the customer does to achieve this is documented in the attachment (sorry about the German comments, but I thought translating them would have been a bit much to ask).

TIA for any comments, etc.
Michael
--
Michael Schuster                  (+49 89) 46008-2974 / x62974
visit the online support center:  http://www.sun.com/osc/

Recursion, n.: see 'Recursion'
-------- Original Message --------
 
ich habe das system booten können habe! Aber alle ZFS-Daten inkl. darauf 
liegender Zonen sind dabei verloren gegangen:
 
{0} ok select  /[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED]/[EMAIL 
PROTECTED] scsi-pci1000,56 LSI,1064E
scsi-pci1000,56 ?
{0} ok show-volumes
Volume 0 Target 83  Type IM (Integrated Mirroring)
  Optimal  Enabled  Inactive
  2 Members                                         143112591 Blocks, 73 GB
  Disk 1
    Primary  Online
    Target 2        FUJITSU MAY2073RCSUN72G 0401
  Disk 0
    Secondary  Online
    Target 3        FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 82  Type IM (Integrated Mirroring)
  Optimal  Enabled  Inactive
  2 Members                                         143112591 Blocks, 73 GB
  Disk 2
    Primary  Online
    Target 0        FUJITSU MAY2073RCSUN72G 0401
  Disk 3
    Secondary  Online
    Target 1        FUJITSU MAY2073RCSUN72G 0401
{0} ok
{0} ok 0 activate-volume
Volume 0 is now activated
{0} ok show-volumes
Volume 0 Target 82  Type IM (Integrated Mirroring)
  Optimal  Enabled  Inactive
  2 Members                                         143112591 Blocks, 73 GB
  Disk 2
    Primary  Online
    Target 0        FUJITSU MAY2073RCSUN72G 0401
  Disk 3
    Secondary  Online
    Target 1        FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 2  Type IM (Integrated Mirroring)
  Degraded  Enabled  Resync In Progress                    <----- 
Anzeigefehler? (Volume 1 statt 0)
  2 Members                                         143112591 Blocks, 73 GB
  Disk 1
    Primary  Online
    Target 4        FUJITSU MAY2073RCSUN72G 0401
  Disk 0
    Secondary  Online  Out Of Sync
    Target 3        FUJITSU MAY2073RCSUN72G 0401
{0} ok 1 activate-volume
RAID volume request failed, ioc-status = 0007              <----- deswegen 
diese Fehlermeldung!
Volume 1 could not be activated
{0} ok 0 activate-volume                                   <----- deswegen noch 
mal Volume 0 !!!
Volume 0 is now activated
{0} ok show-volumes
Volume 0 Target 2  Type IM (Integrated Mirroring)
  Degraded  Enabled  Resync In Progress                    <----- jetzt stimmt 
die Anzeige wieder!
  2 Members                                         143112591 Blocks, 73 GB
  Disk 1
    Primary  Online
    Target 4        FUJITSU MAY2073RCSUN72G 0401
  Disk 0
    Secondary  Online  Out Of Sync
    Target 3        FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 0  Type IM (Integrated Mirroring)
  Degraded  Enabled                                        <----- hier auch. 
Die Spiegelung von Volume 1 beginnt erst wenn die von Volume 0 abgeschlossen 
wird.
  2 Members                                         143112591 Blocks, 73 GB
  Disk 2
    Primary  Online
    Target 5        FUJITSU MAY2073RCSUN72G 0401
  Disk 3
    Secondary  Online  Out Of Sync
    Target 1        FUJITSU MAY2073RCSUN72G 0401
{0} ok
.
.
.
{0} ok show-volumes
Volume 0 Target 2  Type IM (Integrated Mirroring)
  Optimal  Enabled                                         <----- Die 
Spiegelung von Volume 0 ist abgeschlossen.
  2 Members                                         143112591 Blocks, 73 GB
  Disk 1
    Primary  Online
    Target 4        FUJITSU MAY2073RCSUN72G 0401
  Disk 0
    Secondary  Online
    Target 3        FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 0  Type IM (Integrated Mirroring)
  Degraded  Enabled  Resync In Progress                    <----- Die 
Spiegelung von Volume 1 ist nun angelaufen.
  2 Members                                         143112591 Blocks, 73 GB
  Disk 2
    Primary  Online
    Target 5        FUJITSU MAY2073RCSUN72G 0401
  Disk 3
    Secondary  Online  Out Of Sync
    Target 1        FUJITSU MAY2073RCSUN72G 0401
{0} ok boot
Boot device: disk  File and args:
ERROR: /[EMAIL PROTECTED]: Last Trap: Fast Data Access MMU Miss      <----- 
Bootversuche scheitern obwohl die Boorplatten (Volume 0) optimal sind!
[Exception handlers interrupted, please file a bug]
[type 'resume' to attempt a normal recovery]
{0} ok
.
.
.
{0} ok show-volumes
Volume 0 Target 2  Type IM (Integrated Mirroring)
  Optimal  Enabled
  2 Members                                         143112591 Blocks, 73 GB
  Disk 1
    Primary  Online
    Target 4        FUJITSU MAY2073RCSUN72G 0401
  Disk 0
    Secondary  Online
    Target 3        FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 0  Type IM (Integrated Mirroring)
  Optimal  Enabled
  2 Members                                         143112591 Blocks, 73 GB
  Disk 2
    Primary  Online
    Target 5        FUJITSU MAY2073RCSUN72G 0401
  Disk 3
    Secondary  Online
    Target 1        FUJITSU MAY2073RCSUN72G 0401
{0} ok boot
Boot device: disk  File and args:
ERROR: /[EMAIL PROTECTED]: Last Trap: Fast Data Access MMU Miss      <----- 
Bootversuche scheitern auch alle Volumes optimal sind!
[Exception handlers interrupted, please file a bug]
[type 'resume' to attempt a normal recovery]
{0} ok
{0} ok reset-all                                           <----- mit 
reset-all...
 
SC Alert: Host System has Reset
 
SC Alert: Host system has shut down.
cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu 
cpu cpu cpu cpu Device: pci
/[EMAIL PROTECTED]: Device 0 pci
.
.
.
Sun Fire T200, No Keyboard
Copyright 2006 Sun Microsystems, Inc.  All rights reserved.
OpenBoot 4.20.4, 8184 MB memory available, Serial #64571856.
Ethernet address 0:3:ba:d9:49:d0, Host ID: 83d949d0.
 
 
 
{0} ok boot
Boot device: disk  File and args:
Loading ufs-file-system package 1.4 04 Aug 1995 13:02:54.
FCode UFS Reader 1.12 00/07/17 15:48:16.
Loading: /platform/SUNW,Sun-Fire-T200/ufsboot
Loading: /platform/sun4v/ufsboot
SunOS Release 5.10 Version Generic_118833-17 64-bit                    <----- 
Server bootet nun...
Copyright 1983-2005 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
Hostname: fmuw31
checking ufs filesystems
/dev/rdsk/c0t0d0s5: is logging.
 
fmuw31 console login: Jul 17 18:28:48 fmuw31 sendmail[315]: My unqualified host 
name (fmuw31) unknown; sleeping for retry
Jul 17 18:28:48 fmuw31 sendmail[316]: My unqualified host name (fmuw31) 
unknown; sleeping for retry
 
SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major         <----- 
Aber es gibt ein ZFS-Problem!
EVENT-TIME: Mon Jul 17 18:28:50 CEST 2006
PLATFORM: SUNW,Sun-Fire-T200, CSN: -, HOSTNAME: fmuw31
SOURCE: zfs-diagnosis, REV: 1.0
EVENT-ID: 37c8cdbf-addb-4d34-c180-d1b2379e27b2
DESC: A ZFS pool failed to open.  Refer to http://sun.com/msg/ZFS-8000-CS for 
more information.
AUTO-RESPONSE: No automated response will occur.
IMPACT: The pool data is unavailable
REC-ACTION: Run 'zpool status -x' and either attach the missing device or
            restore from backup.
 
SC Alert: Host detected fault, MSGID: ZFS-8000-CS
 
fmuw31 console login: root
Password:
Jul 17 18:29:48 fmuw31 login: ROOT LOGIN /dev/console
Last login: Fri Jul 14 17:02:59 from 53.1.224.107
Sun Microsystems Inc.   SunOS 5.10      Generic January 2005
#
#
# zpool list
NAME                    SIZE    USED   AVAIL    CAP  HEALTH     ALTROOT
dpool                      -       -       -      -  FAULTED    -      <----- 
Datenverlust!!!
# zfs list
no datasets available                                                  <----- 
Datenverlust!!!
#
 
 
-----> Bemerkung zur ZFS Konfiguration:
Auf dem zweiten Spiegelpaar das über den onboard RAID-Controller auf dem 
urspünglichen Server erstellt wurde,
wurde ein ZFS-Pool der die ganze Platte / Volume erfaste (68GB) erzeugt und 
ZFS-Filesystemeangelegt:
 
# zpool create dpool c0t2d0
# zpool status
  pool: dpool
 state: ONLINE
 scrub: none requested
config:
 
        NAME        STATE     READ WRITE CKSUM
        dpool       ONLINE       0     0     0
          c0t2d0    ONLINE       0     0     0
 
errors: No known data errors
# zpool list
NAME                    SIZE    USED   AVAIL    CAP  HEALTH     ALTROOT
dpool                    68G     79K   68.0G     0%  ONLINE     -
# zfs create dpool/zones
# zfs list
NAME                   USED  AVAIL  REFER  MOUNTPOINT
dpool                  106K  66.9G  26.5K  /dpool
dpool/zones           24.5K  66.9G  24.5K  /dpool/zones
# zfs set mountpoint=/zones dpool/zones
# zfs create dpool/zones/fmuw31-vs01
# zfs create dpool/zones/fmuw31-vs02
# zfs create dpool/zones/fmuw31-vs03
# zfs list
NAME                   USED  AVAIL  REFER  MOUNTPOINT
dpool                  204K  66.9G  24.5K  /dpool
dpool/zones            102K  66.9G  28.5K  /zones
dpool/zones/fmuw31-vs01  24.5K  66.9G  24.5K  /zones/fmuw31-vs01
dpool/zones/fmuw31-vs02  24.5K  66.9G  24.5K  /zones/fmuw31-vs02
dpool/zones/fmuw31-vs03  24.5K  66.9G  24.5K  /zones/fmuw31-vs03
#
...
es wurden Virtuelle-server / Zonen auf diesen ZFSs erstellt... die 
Plattenbelegung sah wie folgt aus:
 
# df -k
Filesystem            kbytes    used   avail capacity  Mounted on
/dev/dsk/c0t0d0s0    20654577 3680834 16767198    19%    /
/devices                   0       0       0     0%    /devices
ctfs                       0       0       0     0%    /system/contract
proc                       0       0       0     0%    /proc
mnttab                     0       0       0     0%    /etc/mnttab
swap                 22868632    1032 22867600     1%    /etc/svc/volatile
objfs                      0       0       0     0%    /system/object
/platform/SUNW,Sun-Fire-T200/lib/libc_psr/libc_psr_hwcap1.so.1
                     20654577 3680834 16767198    19%    
/platform/sun4v/lib/libc_psr.so.1
/platform/SUNW,Sun-Fire-T200/lib/sparcv9/libc_psr/libc_psr_hwcap1.so.1
                     20654577 3680834 16767198    19%    
/platform/sun4v/lib/sparcv9/libc_psr.so.1
fd                         0       0       0     0%    /dev/fd
/dev/dsk/c0t0d0s4    8262325   70108 8109594     1%    /var
swap                 22868040     440 22867600     1%    /tmp
swap                 22867656      56 22867600     1%    /var/run
/dev/dsk/c0t0d0s5    24677176   24489 24405916     1%    /vol1
dpool                70189056      24 70026796     1%    /dpool
dpool/zones          70189056      28 70026796     1%    /zones
dpool/zones/fmuw31-vs01
                     20971520   80148 20891371     1%    /zones/fmuw31-vs01
dpool/zones/fmuw31-vs02
                     70189056   81881 70026796     1%    /zones/fmuw31-vs02
dpool/zones/fmuw31-vs03
                     70189056      24 70026796     1%    /zones/fmuw31-vs03
#
 
... und jetzt ist das alles weg!
 
 
 
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to