Lars-Gunnar,

On Tue, Mar 03, 2009 at 11:18:27AM +0100, Lars-Gunnar Persson wrote:
> -bash-3.00$ zfs list -o  
> name,type,used,avail,ratio,compression,reserv,volsize Data/subversion1
> NAME                TYPE   USED  AVAIL  RATIO  COMPRESS  RESERV  VOLSIZE
> Data/subversion1  volume  22.5K   511G  1.00x       off    250G     250G

This shows that the volume still exists.
Correct me if I am wrong here :
- Did you mean that the contents of the volume subversion1 are corrupted ?

What does that volume have on it ? Does it contain a filesystem which can
can be mounted on Solaris ? If so, we could try mounting it locally on the
Solaris box. This is to rule out any iSCSI issues.

Also, do you have any snapshots of the volume ? If so, you could rollback
to the latest snapshot. But, that would mean we lose some amount of data.

Also, you mentioned that the volume was in use for a year. But, I see in the
above output that it has only about 22.5K used. Is that correct ? I would
have expected it to be higher.

You should also check what 'zpool history -i ' says.

Thanks and regards,
Sanjeev
>
> I've also learned the the AVAIL column reports what's available in the  
> zpool and NOT what's available in the ZFS volume.
>
> -bash-3.00$ sudo zpool status -v
> Password:
>   pool: Data
>  state: ONLINE
>  scrub: scrub in progress, 5.86% done, 12h46m to go
> config:
>
>         NAME                     STATE     READ WRITE CKSUM
>         Data                     ONLINE       0     0     0
>           c4t5000402001FC442Cd0  ONLINE       0     0     0
>
> errors: No known data errors
>
> Interesting thing here is that the scrub process should be finished  
> today but the progress is much slower than reported here. And will the  
> scrub process help anything in my case?
>
>
> -bash-3.00$ sudo fmdump
> TIME                 UUID                                 SUNW-MSG-ID
> Nov 15 2007 10:16:38 8aa789d2-7f3a-45d5-9f5c-c101d73b795e ZFS-8000-CS
> Oct 14 09:31:40.8179 8c7d9847-94b7-ec09-8da7-c352de405b78 FMD-8000-2K
>
> bash-3.00$ sudo fmdump -ev
> TIME                 CLASS                                 ENA
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e6850ff400401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e6850ff400401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e6897db600401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e688d11500401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68926e600401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e6897db600401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68a3d3900401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68bc67400001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68d8bb600401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68da5b500001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68da5b500001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68f0c9800401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68da5b500001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e6897db600401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68e981900001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68f0c9800401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e690a11000401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68f0c9800401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e690385500001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e690385500001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e690a11000401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e692a4ca00001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68bc67400001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e690a11000401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e6850ff400401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e6850ff400401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68bc67400001
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e690385500001
> Nov 15 2007 09:33:52 ereport.fs.zfs.data                    
> 0x915e6850ff400401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68a3d3900401
> Nov 15 2007 09:33:52 ereport.fs.zfs.io                      
> 0x915e68a3d3900401
> Nov 15 2007 10:16:12 ereport.fs.zfs.vdev.open_failed        
> 0x0533bb1b56400401
> Nov 15 2007 10:16:12 ereport.fs.zfs.zpool                   
> 0x0533bb1b56400401
> Oct 14 09:31:31.6092 ereport.fm.fmd.log_append              
> 0x02eb96a8b6502801
> Oct 14 09:31:31.8643 ereport.fm.fmd.mod_init                
> 0x02ec89eadd100401
>
>
> On 3. mars. 2009, at 08.10, Lars-Gunnar Persson wrote:
>
>> I've turned off iSCSI sharing at the moment.
>>
>> My first question is: how can zfs report available is larger than  
>> reservation on a zfs volume? I also know that used mshould be larger  
>> than 22.5 K. Isn't this strange?
>>
>> Lars-Gunnar Persson
>>
>> Den 3. mars. 2009 kl. 00.38 skrev Richard Elling 
>> <richard.ell...@gmail.com>:
>>
>>> Lars-Gunnar Persson wrote:
>>>> Hey to everyone on this mailing list (since this is my first post)!
>>>
>>> Welcome!
>>>
>>>>
>>>> We've a Sun Fire X4100 M2 server running Solaris 10 u6 and after  
>>>> some system work this weekend we have a problem with only one ZFS  
>>>> volume.
>>>>
>>>> We have a pool called /Data with many file systems and two  
>>>> volumes. The status of my zpool is:
>>>>
>>>> -bash-3.00$ zpool status
>>>> pool: Data
>>>> state: ONLINE
>>>> scrub: scrub in progress, 5.99% done, 13h38m to go
>>>> config:
>>>>
>>>> NAME STATE READ WRITE CKSUM
>>>> Data ONLINE 0 0 0
>>>> c4t5000402001FC442Cd0 ONLINE 0 0 0
>>>>
>>>> errors: No known data errors
>>>>
>>>>
>>>> Yesterday I started the scrub process because I read that was a  
>>>> smart thing to do after a zpool export and zpool import procedure.  
>>>> I did this because I wanted to move the zpool to another OS  
>>>> installation but changed my mind and did a zpool import on the  
>>>> same OS as I did an export.
>>>>
>>>> After checking as much information as I could find on the web, I  
>>>> was advised to to run the zpool scrub after an import.
>>>>
>>>> Well, the problem now is that one volume in this zpool is not  
>>>> working. I've shared it via iscsi to a Linux host (all of this was  
>>>> working on Friday). The Linux host reports that it can't find a  
>>>> partition table. Here is the log from the Linux host:
>>>>
>>>> Mar 2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte  
>>>> hdwr sectors (268435 MB)
>>>> Mar 2 11:09:36 eva kernel: SCSI device sdb: drive cache: write  
>>>> through
>>>> Mar 2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte  
>>>> hdwr sectors (268435 MB)
>>>> Mar 2 11:09:37 eva kernel: SCSI device sdb: drive cache: write  
>>>> through
>>>> Mar 2 11:09:37 eva kernel: sdb: unknown partition table
>>>> Mar 2 11:09:37 eva kernel: Attached scsi disk sdb at scsi28,  
>>>> channel 0, id 0, lun 0
>>>>
>>>>
>>>> So I checked the status on my Solaris server and I found this  
>>>> information a bit strange;:
>>>>
>>>> -bash-3.00$ zfs list Data/subversion1
>>>> NAME USED AVAIL REFER MOUNTPOINT
>>>> Data/subversion1 22.5K 519G 22.5K -
>>>>
>>>> How can it bed 519GB available on a volume that is 250GB in size?  
>>>> Here are more details:
>>>>
>>>> -bash-3.00$ zfs get all Data/subversion1
>>>> NAME PROPERTY VALUE SOURCE
>>>> Data/subversion1 type volume -
>>>> Data/subversion1 creation Wed Apr 2 9:06 2008 -
>>>> Data/subversion1 used 22.5K -
>>>> Data/subversion1 available 519G -
>>>> Data/subversion1 referenced 22.5K -
>>>> Data/subversion1 compressratio 1.00x -
>>>> Data/subversion1 reservation 250G local
>>>> Data/subversion1 volsize 250G -
>>>> Data/subversion1 volblocksize 8K -
>>>> Data/subversion1 checksum on default
>>>> Data/subversion1 compression off default
>>>> Data/subversion1 readonly off default
>>>> Data/subversion1 shareiscsi off local
>>>
>>> It does not appear that Data/subversion1 is being shared via iscsi?
>>> -- richard
>>>
>>>
>> _______________________________________________
>> zfs-discuss mailing list
>> zfs-discuss@opensolaris.org
>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>
>
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

-- 
----------------
Sanjeev Bagewadi
Solaris RPE 
Bangalore, India
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to