On Wed, Sep 07, 2022 at 09:13:12PM +0200, Danny Koenig wrote:
> 
> 
> On Wed, 2022-09-07 at 19:24 +0200, Danny Koenig wrote:
> > On Wed, 2022-09-07 at 18:18 +0100, Zé Loff wrote:
> > > 
> > > On Wed, Sep 07, 2022 at 06:14:59PM +0100, Zé Loff wrote:
> > > > On Wed, Sep 07, 2022 at 06:42:54PM +0200, Danny König wrote:
> > > > > Hi,
> > > > >  
> > > > > first time I need your support regarding a broken softraid0.
> > > > > After a
> > > > > normal reboot the softraid0 (RAID 1, two disks sd0 and sd1,
> > > > > softraid0
> > > > > as sd5) was broken what means no mounts, no nothing.
> > > > > 
> > > > >  
> > > > > root@core-fs:~ => disklabel -p g /dev/sd0c 
> > > > > # /dev/sd0c: 
> > > > > type: SCSI 
> > > > > disk: SCSI disk 
> > > > > label: ST16000NM003G-2K 
> > > > > duid: b1c256d6dc122fa7 
> > > > > flags: 
> > > > > bytes/sector: 512 
> > > > > sectors/track: 63 
> > > > > tracks/cylinder: 255 
> > > > > sectors/cylinder: 16065 
> > > > > cylinders: 1945332 
> > > > > total sectors: 31251759104 # total bytes: 14902.0G 
> > > > > boundstart: 0 
> > > > > boundend: 31251759104 
> > > > > drivedata: 0
> > > > > 16 partitions: 
> > > > > # size offset fstype [fsize bsize cpg] 
> > > > > a: 14902.0G 0 RAID 
> > > > > c: 14902.0G 0 unused
> > > > > 
> > > > > 
> > > > > 
> > > > > root@core-fs:~ => disklabel -p g /dev/sd1c 
> > > > > # /dev/sd1c: 
> > > > > type: SCSI 
> > > > > disk: SCSI disk 
> > > > > label: ST16000NM003G-2K 
> > > > > duid: acedcb2d1684fb18 
> > > > > flags: 
> > > > > bytes/sector: 512 
> > > > > sectors/track: 63 
> > > > > tracks/cylinder: 255 
> > > > > sectors/cylinder: 16065 
> > > > > cylinders: 1945332 
> > > > > total sectors: 31251759104 # total bytes: 14902.0G 
> > > > > boundstart: 0 
> > > > > boundend: 31251759104 
> > > > > drivedata: 0
> > > > > 16 partitions: 
> > > > > # size offset fstype [fsize bsize cpg] 
> > > > > a: 14902.0G 0 RAID 
> > > > > c: 14902.0G 0 unused 
> > > > > 
> > > > > 
> > > > > 
> > > > > root@core-fs:~ => disklabel -p g /dev/sd5c 
> > > > > # /dev/sd5c:
> > > > > type: SCSI
> > > > > disk: SCSI disk
> > > > > label: SR RAID 1
> > > > > duid: 0000000000000000
> > > > > flags:
> > > > > bytes/sector: 512
> > > > > sectors/track: 255
> > > > > tracks/cylinder: 511
> > > > > sectors/cylinder: 130305
> > > > > cylinders: 239835
> > > > > total sectors: 31251758576 # total bytes: 14902.0G
> > > > > boundstart: 0
> > > > > boundend: 31251758576
> > > > > drivedata: 0 
> > > > > 
> > > > > 16 partitions:
> > > > > #                size           offset  fstype [fsize bsize  
> > > > > cpg]
> > > > >   c:         14902.0G                0  unused
> > > > > 
> > > > > 
> > > > > 
> > > > >  
> > > > > root@core-fs:~ => bioctl softraid0 
> > > > > Volume Status Size Device 
> > > > > softraid0 0 Online 16000900390912 sd5 RAID1 
> > > > > 0 Online 16000900390912 0:0.0 noencl <sd0a> 
> > > > > 1 Online 16000900390912 0:1.0 noencl <sd1a>
> > > > > 
> > > > > 
> > > > >  
> > > > > Thats what we got, but no mounts are possible.
> > > > >  
> > > > > Any ideas how to become the data back available?
> > > > >  
> > > > > Thank you very much.
> > > > >  
> > > > > Regards,
> > > > > Danny
> > > > > 
> > > > 
> > > > sd5's disklabel looks pretty empty (duid included), but you
> > > > already
> > > > know
> > > > that...
> > > > Do you have a backup disklabel, e.g., in /var/backups?
> > 
> > 
> > Yes, that's the output:
> > 
> > # /dev/rsd5c:
> > type: SCSI
> > disk: SCSI disk
> > label: SR RAID 1C
> > duid: 1c318a61d9aa3255
> > flags:
> > bytes/sector: 512
> > sectors/track: 63
> > tracks/cylinder: 255
> > sectors/cylinder: 16065
> > cylinders: 1945331
> > total sectors: 31251758576
> > boundstart: 0
> > boundend: 31251758576
> > drivedata: 0 
> > 
> > 16 partitions:
> > #                size           offset  fstype [fsize bsize   cpg]
> >   c:      31251758576                0  4.2BSD   2048 16384 12960
> > 
> 
> Restored from /var/backups/disklabel.sd5.current by using "disklabel -e
> sd5" as recommended. Sadly this didn't changed anything. Still can't
> mount.
> 
> Current state:
> 
> root@core-fs:~ => disklabel sd5c          
> # /dev/rsd5c:
> type: SCSI
> disk: SCSI disk
> label: SR RAID 1C
> duid: 2765b6ceb44e92fb
> flags:
> bytes/sector: 512
> sectors/track: 63
> tracks/cylinder: 255
> sectors/cylinder: 16065
> cylinders: 1945331
> total sectors: 31251758576
> boundstart: 0
> boundend: 31251758576
> drivedata: 0 
> 
> 16 partitions:
> #                size           offset  fstype [fsize bsize   cpg]
>   c:      31251758576                0  4.2BSD   2048 16384 12960
> 
> 
> Any suggestions?
> 
> Just wondering about the size of 32 TB. Can this be correct? It should
> be RAID 1, two disks with each a size of 16 TB.
> 

It is correct. Those are sectors, not bytes. See the "total sectors"-line
in your disklabel output. If you run

# disklabel -p t sd5

it should report 14.55 TB (which equals 16 TiB).

(The option "-p t" means "print disklabel using unit t(erabyte)".)

(sectors * bytes/sector) / 1024^3 => (31251758576 * 512) / 1024^3 = 14902.00 GB
(sectors * bytes/sector) / 1000^3 => (31251758576 * 512) / 1000^3 = 16000.90 GiB

(sectors * bytes/sector) / 1024^4 => (31251758576 * 512) / 1024^4 = 14.55 TB
(sectors * bytes/sector) / 1000^4 => (31251758576 * 512) / 1000^4 = 16.00 TiB

(I'm no expert. Just chipping in. Good luck and don't panic. And DON'T
blindly type in commands or do procedures that people have suggested!)

Erling

Reply via email to