Re: [zfs-discuss] zfs corruption w/ sil3114 sata controllers

2007-10-31 Thread Nigel Smith
Ok, this is a strange problem!
You seem to have tried & eliminated all the possible issues
that the community has suggested!

I was hoping you would see some errors logged in
'/var/adm/messages' that would give a clue.

Your original 'zpool status' said 140 errors.
Over what time period are these occurring?
I'm wondering if the errors are occurring at a
constant steady rate or if there are bursts of error?
Maybe you could monitor zpool status while generating
activity with "dd" or similar.
You could use "zpool iostat " to monitor
bandwidth and see if it is reasonably steady or erratic.

>From your "prtconf -D" we see the 3114 card is using
the "ata" driver, as expected.
I believe the driver can talk to the disk drive
in either PIO or DMA mode, so you could try 
changing that in the "ata.conf" file. See here for details:
http://docs.sun.com/app/docs/doc/819-2254/ata-7d?a=view

I've just had a quick look at the source code for
the ata driver, and there does seem to be specific support
for the Silicon Image chips in the drivers:
http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/controller/ata/sil3xxx.c
and
http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/controller/ata/sil3xxx.h
The file "sil3xxx.h" does mention:
  "Errata Sil-AN-0109-B2 (Sil3114 Rev 0.3)
  To prevent erroneous ERR set for queued DMA transfers
  greater then 8k, FIS reception for FIS0cfg needs to be set
  to Accept FIS without Interlock"
..which I read as meaning there have being some 'issues'
with this chip. And it sounds similar to the issue mention on
the link that Tomasz supplied:
http://home-tj.org/wiki/index.php/Sil_m15w

If you decide to try a different SATA controller card, possible options are:

1. The si3124 driver, which supports SiI-3132 (PCI-E)
   and SiI-3124 (PCI-X) devices.
   
2. The AHCI driver, which supports the Intel ICH6 and latter devices, often
   found on motherboard.
   
4. The NV_SATA driver which supports Nvidia ck804/mcp55 devices.

Regards
Nigel Smith
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs corruption w/ sil3114 sata controllers

2007-10-31 Thread Edward Saipetch
Nigel,

Thanks for the response!  Basically my last method of testing was to 
sftp a few 50-100MB files to /tank over a couple of minutes and force a 
scrub after.  The very first time this happened, I was using it as a NAS 
device dumping data to it for over a week.  I went to a customer's site 
to show him how cool zfs was and upon running zpool status, I saw the 
data corruption status and telling me to restore from a backup.  Running 
zpool status without a scrub shows no errors.

I tried mirrored devices, no raid whatsoever and raidz, all with the 
same results.  All the motherboards I've been using only have PCI since 
I was hoping I could create a low cost solution as a POC.  I'll test 
changing the transfer mode a bit later.  Other people have had better 
luck, what other debugging can be done?  I'm willing to even let someone 
have remote access to the box if they want.

Nigel Smith wrote:
> Ok, this is a strange problem!
> You seem to have tried & eliminated all the possible issues
> that the community has suggested!
>
> I was hoping you would see some errors logged in
> '/var/adm/messages' that would give a clue.
>
> Your original 'zpool status' said 140 errors.
> Over what time period are these occurring?
> I'm wondering if the errors are occurring at a
> constant steady rate or if there are bursts of error?
> Maybe you could monitor zpool status while generating
> activity with "dd" or similar.
> You could use "zpool iostat " to monitor
> bandwidth and see if it is reasonably steady or erratic.
>
> >From your "prtconf -D" we see the 3114 card is using
> the "ata" driver, as expected.
> I believe the driver can talk to the disk drive
> in either PIO or DMA mode, so you could try 
> changing that in the "ata.conf" file. See here for details:
> http://docs.sun.com/app/docs/doc/819-2254/ata-7d?a=view
>
> I've just had a quick look at the source code for
> the ata driver, and there does seem to be specific support
> for the Silicon Image chips in the drivers:
> http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/controller/ata/sil3xxx.c
> and
> http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/controller/ata/sil3xxx.h
> The file "sil3xxx.h" does mention:
>   "Errata Sil-AN-0109-B2 (Sil3114 Rev 0.3)
>   To prevent erroneous ERR set for queued DMA transfers
>   greater then 8k, FIS reception for FIS0cfg needs to be set
>   to Accept FIS without Interlock"
> ..which I read as meaning there have being some 'issues'
> with this chip. And it sounds similar to the issue mention on
> the link that Tomasz supplied:
> http://home-tj.org/wiki/index.php/Sil_m15w
>
> If you decide to try a different SATA controller card, possible options are:
>
> 1. The si3124 driver, which supports SiI-3132 (PCI-E)
>and SiI-3124 (PCI-X) devices.
>
> 2. The AHCI driver, which supports the Intel ICH6 and latter devices, often
>found on motherboard.
>
> 4. The NV_SATA driver which supports Nvidia ck804/mcp55 devices.
>
> Regards
> Nigel Smith
>  
>  
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>   

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs corruption w/ sil3114 sata controllers

2007-10-31 Thread Mario Goebbels
I haven't seen the beginning of this discussion, but seeing SiI sets the
fire alarm off here.

The Silicon Image chipsets are renowned to be crap and causing data
corruption. At least the variants that usually go onto mainboards. Based
on this, I suggest that you should get a different card.

-mg



signature.asc
Description: OpenPGP digital signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs corruption w/ sil3114 sata controllers

2007-10-31 Thread Edward Saipetch
Mario,

I don't have any issues getting a new card.  The root of the discussion 
started because people did indeed post that they had good luck with 
them.  In fact, when I went out there and google'd to find which cards 
would worked well, it seemed to be at the top of the list.  I'm 
interested to know if it's something I can help resolve so other people 
don't have this problem or make sure people don't run into the same 
issue I do.

Mario Goebbels wrote:
> I haven't seen the beginning of this discussion, but seeing SiI sets the
> fire alarm off here.
>
> The Silicon Image chipsets are renowned to be crap and causing data
> corruption. At least the variants that usually go onto mainboards. Based
> on this, I suggest that you should get a different card.
>
> -mg
>   
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] DESC: A ZFS pool failed to open.

2007-10-31 Thread Vahid Moghaddasi
Hi,
All the pools seem healthy and zfs file systems are all fine according to 
"zpool status -x" but during the boot we get the following error:

fmadm faulty also returns this:
 --
degraded zfs://pool=pe09_01
 8f5e62aa-c0af-4536-cef3-8e9d9169ea92
 --
degraded zfs://pool=re09_01
 eca6f995-12ba-ce3b-9a5b-d33f1b6580ac
 --

Thanks,

SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY:
Major
EVENT-TIME: Fri Oct 26 18:25:33 EDT
2007
PLATFORM: SUNW,Sun-Fire-T200, CSN: -, HOSTNAME:
myserver
SOURCE: zfs-diagnosis, REV:
1.0
EVENT-ID: 8168b743-76b2-ccc0-9816-
e7387c93267b
DESC: A ZFS pool failed to open.  Refer to http://sun.com/msg/ZFS-8000-CS
for more information.
AUTO-RESPONSE: No automated response will
occur.
IMPACT: The pool data is
unavailable
REC-ACTION: Run 'zpool status -x' and either attach the missing device
or
restore from
backup.

SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY:
Major
EVENT-TIME: Fri Oct 26 18:25:34 EDT
2007
PLATFORM: SUNW,Sun-Fire-T200, CSN: -, HOSTNAME:
myserver
SOURCE: zfs-diagnosis, REV:
1.0
EVENT-ID: 5c6e92ad-560b-ee04-
b324-8f32f869
DESC: A ZFS pool failed to open.  Refer to http://sun.com/msg/ZFS-8000-CS
for more information.
AUTO-RESPONSE: No automated response will
occur.
IMPACT: The pool data is
unavailable
REC-ACTION: Run 'zpool status -x' and either attach the missing device
or
restore from backup.

Is my pool bad or this is just a bug? How make it go away?
Thanks,
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] hierarchical share of zfs filesystem

2007-10-31 Thread Jan Hendrik Mangold
I have a quick question about ZFS filesystems and snapshots and NFS  
shares.


Is it possible to have one mount point for a set of ZFS filesystems?  
Lets say I have /export/homes as a ZFS filesystem and NFS shared,  I  
create a new ZFS filesystem /export/homes/user1 I need to explicitly  
mount it. Only the mount point is visible under /export/homes. This  
has been standard NFS with devices mounted, ie. submounts are not  
exported, but that becomes very unhandy with dynamic ZFS and  
snapshots under NFS?


Again, the question is if I can export all ZFS filesystems under a  
specific mount point.

--
Jan Hendrik Mangold
 Sun Microsystems
Senior Lead Staff Engineer
650-585-5484 (x81371)
http://learning.central/wiki/display/~13922
"idle hands are the developers workshop"


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] hierarchical share of zfs filesystem

2007-10-31 Thread Darren J Moffat
Jan Hendrik Mangold wrote:
> I have a quick question about ZFS filesystems and snapshots and NFS shares.
> 
> Is it possible to have one mount point for a set of ZFS filesystems? 
> Lets say I have /export/homes as a ZFS filesystem and NFS shared,  I 
> create a new ZFS filesystem /export/homes/user1 I need to explicitly 
> mount it. Only the mount point is visible under /export/homes. This has 
> been standard NFS with devices mounted, ie. submounts are not exported, 
> but that becomes very unhandy with dynamic ZFS and snapshots under NFS?

Solved by PSARC 2007/416 NFSv4 Mirror-mounts which integrated on Monday 
last week.

http://opensolaris.org/os/community/on/flag-days/pages/2007102201/

All going well this should be in snv_77.

-- 
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] hierarchical share of zfs filesystem

2007-10-31 Thread Jan Hendrik Mangold


On 31 Oct 2007, at 11:14, Darren J Moffat wrote:

Solved by PSARC 2007/416 NFSv4 Mirror-mounts which integrated on  
Monday last week.


that is great news. Especially since it confirms that I didn't do  
anything wrong, but that this feature just wasn't available yet ... :)

--
Jan Hendrik Mangold
 Sun Microsystems
Senior Lead Staff Engineer
650-585-5484 (x81371)
http://learning.central/wiki/display/~13922
"idle hands are the developers workshop"


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS Quota Oddness

2007-10-31 Thread Ben Rockwood
I've run across an odd issue with ZFS Quota's.  This is an snv_43 system with 
several zones/zfs datasets, but only one effected.  The dataset shows 10GB 
used, 12GB refered but when counting the files only has 6.7GB of data:

zones/ABC10.8G  26.2G  12.0G  /zones/ABC
zones/[EMAIL PROTECTED]14.7M  -  12.0G  -

[xxx:/zones/ABC/.zfs/snapshot/now] root# gdu --max-depth=1 -h .
43k ./dev
6.7G./root
1.5k./lu
6.7G.

I don't understand what might the cause this disparity.  This is an older box, 
snv_43.  Any bugs that might apply, fixed or in progress?

Thanks.

benr.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS: Cannot share a zfs dataset added to a labeled zone

2007-10-31 Thread Danny Hayes
- I set the mount point as follows.

zfs set mountpoint=/zone/restricted/root/data zone/data

- I then added the dataset to the restricted zone using zonecfg. The full path 
to the dataset is now /zone/restricted/root/zone/restricted/root/data. I am not 
sure if that is what you intended, but it is a result of adding it as a dataset 
to the zone after setting the mountpoint.

- I updated the /zone/restricted/etc/dfs/dfstab with the following line.

/usr/bin/share -F nfs -o rw /zone/restricted/root/zone/data

- During reboot I receive the following error.

cannot mount 'zone/data': mountpoint or dataset is busy
svc:/system/filesystem/local:default: WARNING: /usr/sbin/zfs mount -a failed: 
exit status 1
Oct 31 14:43:08 svc.startd[19960]: svc:/system/filesystem/local:default: Method 
"/lib/svc/method/fs-local" failed with exit status 95.
Oct 31 14:43:08 svc.startd[19960]: system/filesystem/local:default failed 
fatally: transitioned to maintenance (see 'svcs -xv' for details)

- This is exactly the same problem that prompted the original message. Service 
fail during boot which prevent opening a console. This only occurs when you try 
to share the dataset. If you remove the line from 
/zone/restricted/etc/dfs/dfstab and reboot the zone everything works fine. Any 
ideas what I am doing wrong?
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS Quota Oddness

2007-10-31 Thread Bill Sommerfeld
On Wed, 2007-10-31 at 14:15 -0700, Ben Rockwood wrote:
> I've run across an odd issue with ZFS Quota's.  This is an snv_43 system with 
> several zones/zfs datasets, but only one effected.  The dataset shows 10GB 
> used, 12GB refered but when counting the files only has 6.7GB of data:
> 
> zones/ABC10.8G  26.2G  12.0G  /zones/ABC
> zones/[EMAIL PROTECTED]14.7M  -  12.0G  -
> 
> [xxx:/zones/ABC/.zfs/snapshot/now] root# gdu --max-depth=1 -h .
> 43k ./dev
> 6.7G./root
> 1.5k./lu
> 6.7G.
> 
> I don't understand what might the cause this disparity.  

is the pool built out of three-disk raidz1 groups and predates snv_42?

if so, it's probably some facet of:

6288488 du reports misleading size on RAID-Z



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS Training

2007-10-31 Thread Victor Engle
This class looks pretty good...

http://www.sun.com/training/catalog/courses/SA-229-S10.xml



On 10/31/07, Lisa Richards <[EMAIL PROTECTED]> wrote:
>
>
>
>
> Is there a class on ZFS installation and administration ?
>
>
>
> Lisa Richards
>
> Zykis Corporation
>
> [EMAIL PROTECTED]
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS Training

2007-10-31 Thread Lisa Richards
Is there a class on ZFS installation and administration ?

Lisa Richards
Zykis Corporation
[EMAIL PROTECTED]
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] the number of mount points

2007-10-31 Thread Hidehiko Jono
Hi,

IHAC who wants to use ZFS for user's home directory.
He is worry about the number of mount points.
Does ZFS have any limitation about the number of mount points in a server?

Regards,
Jono
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss