That is correct. It's a raid 6 disk shelf with one volume connected
via fibre.
Lars-Gunnar Persson
Den 2. mars. 2009 kl. 16.57 skrev Blake <blake.ir...@gmail.com>:
It looks like you only have one physical device in this pool. Is
that correct?
On Mon, Mar 2, 2009 at 9:01 AM, Lars-Gunnar Persson
<lars-gunnar.pers...@nersc.no> wrote:
Hey to everyone on this mailing list (since this is my first post)!
We've a Sun Fire X4100 M2 server running Solaris 10 u6 and after
some system
work this weekend we have a problem with only one ZFS volume.
We have a pool called /Data with many file systems and two volumes.
The
status of my zpool is:
-bash-3.00$ zpool status
pool: Data
state: ONLINE
scrub: scrub in progress, 5.99% done, 13h38m to go
config:
NAME STATE READ WRITE CKSUM
Data ONLINE 0 0 0
c4t5000402001FC442Cd0 ONLINE 0 0 0
errors: No known data errors
Yesterday I started the scrub process because I read that was a
smart thing
to do after a zpool export and zpool import procedure. I did this
because I
wanted to move the zpool to another OS installation but changed my
mind and
did a zpool import on the same OS as I did an export.
After checking as much information as I could find on the web, I
was advised
to to run the zpool scrub after an import.
Well, the problem now is that one volume in this zpool is not
working. I've
shared it via iscsi to a Linux host (all of this was working on
Friday). The
Linux host reports that it can't find a partition table. Here is
the log
from the Linux host:
Mar 2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte
hdwr sectors
(268435 MB)
Mar 2 11:09:36 eva kernel: SCSI device sdb: drive cache: write
through
Mar 2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte
hdwr sectors
(268435 MB)
Mar 2 11:09:37 eva kernel: SCSI device sdb: drive cache: write
through
Mar 2 11:09:37 eva kernel: sdb: unknown partition table
Mar 2 11:09:37 eva kernel: Attached scsi disk sdb at scsi28,
channel 0, id
0, lun 0
So I checked the status on my Solaris server and I found this
information a
bit strange;:
-bash-3.00$ zfs list Data/subversion1
NAME USED AVAIL REFER MOUNTPOINT
Data/subversion1 22.5K 519G 22.5K -
How can it bed 519GB available on a volume that is 250GB in size?
Here are
more details:
-bash-3.00$ zfs get all Data/subversion1
NAME PROPERTY VALUE SOURCE
Data/subversion1 type volume -
Data/subversion1 creation Wed Apr 2 9:06 2008 -
Data/subversion1 used 22.5K -
Data/subversion1 available 519G -
Data/subversion1 referenced 22.5K -
Data/subversion1 compressratio 1.00x -
Data/subversion1 reservation 250G local
Data/subversion1 volsize 250G -
Data/subversion1 volblocksize 8K -
Data/subversion1 checksum on default
Data/subversion1 compression off default
Data/subversion1 readonly off default
Data/subversion1 shareiscsi off local
Will this be fixed after the scrub process is finished tomorrow or
is this
volume lost forever?
Hoping for some quick answers as the data is quite important for us.
Regards,
Lars-Gunnar Persson
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss