Re: [zfs-discuss] zfs corruption w/ sil3114 sata controllers
Ok, this is a strange problem! You seem to have tried & eliminated all the possible issues that the community has suggested! I was hoping you would see some errors logged in '/var/adm/messages' that would give a clue. Your original 'zpool status' said 140 errors. Over what time period are these occurring? I'm wondering if the errors are occurring at a constant steady rate or if there are bursts of error? Maybe you could monitor zpool status while generating activity with "dd" or similar. You could use "zpool iostat " to monitor bandwidth and see if it is reasonably steady or erratic. >From your "prtconf -D" we see the 3114 card is using the "ata" driver, as expected. I believe the driver can talk to the disk drive in either PIO or DMA mode, so you could try changing that in the "ata.conf" file. See here for details: http://docs.sun.com/app/docs/doc/819-2254/ata-7d?a=view I've just had a quick look at the source code for the ata driver, and there does seem to be specific support for the Silicon Image chips in the drivers: http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/controller/ata/sil3xxx.c and http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/controller/ata/sil3xxx.h The file "sil3xxx.h" does mention: "Errata Sil-AN-0109-B2 (Sil3114 Rev 0.3) To prevent erroneous ERR set for queued DMA transfers greater then 8k, FIS reception for FIS0cfg needs to be set to Accept FIS without Interlock" ..which I read as meaning there have being some 'issues' with this chip. And it sounds similar to the issue mention on the link that Tomasz supplied: http://home-tj.org/wiki/index.php/Sil_m15w If you decide to try a different SATA controller card, possible options are: 1. The si3124 driver, which supports SiI-3132 (PCI-E) and SiI-3124 (PCI-X) devices. 2. The AHCI driver, which supports the Intel ICH6 and latter devices, often found on motherboard. 4. The NV_SATA driver which supports Nvidia ck804/mcp55 devices. Regards Nigel Smith This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs corruption w/ sil3114 sata controllers
Nigel, Thanks for the response! Basically my last method of testing was to sftp a few 50-100MB files to /tank over a couple of minutes and force a scrub after. The very first time this happened, I was using it as a NAS device dumping data to it for over a week. I went to a customer's site to show him how cool zfs was and upon running zpool status, I saw the data corruption status and telling me to restore from a backup. Running zpool status without a scrub shows no errors. I tried mirrored devices, no raid whatsoever and raidz, all with the same results. All the motherboards I've been using only have PCI since I was hoping I could create a low cost solution as a POC. I'll test changing the transfer mode a bit later. Other people have had better luck, what other debugging can be done? I'm willing to even let someone have remote access to the box if they want. Nigel Smith wrote: > Ok, this is a strange problem! > You seem to have tried & eliminated all the possible issues > that the community has suggested! > > I was hoping you would see some errors logged in > '/var/adm/messages' that would give a clue. > > Your original 'zpool status' said 140 errors. > Over what time period are these occurring? > I'm wondering if the errors are occurring at a > constant steady rate or if there are bursts of error? > Maybe you could monitor zpool status while generating > activity with "dd" or similar. > You could use "zpool iostat " to monitor > bandwidth and see if it is reasonably steady or erratic. > > >From your "prtconf -D" we see the 3114 card is using > the "ata" driver, as expected. > I believe the driver can talk to the disk drive > in either PIO or DMA mode, so you could try > changing that in the "ata.conf" file. See here for details: > http://docs.sun.com/app/docs/doc/819-2254/ata-7d?a=view > > I've just had a quick look at the source code for > the ata driver, and there does seem to be specific support > for the Silicon Image chips in the drivers: > http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/controller/ata/sil3xxx.c > and > http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/controller/ata/sil3xxx.h > The file "sil3xxx.h" does mention: > "Errata Sil-AN-0109-B2 (Sil3114 Rev 0.3) > To prevent erroneous ERR set for queued DMA transfers > greater then 8k, FIS reception for FIS0cfg needs to be set > to Accept FIS without Interlock" > ..which I read as meaning there have being some 'issues' > with this chip. And it sounds similar to the issue mention on > the link that Tomasz supplied: > http://home-tj.org/wiki/index.php/Sil_m15w > > If you decide to try a different SATA controller card, possible options are: > > 1. The si3124 driver, which supports SiI-3132 (PCI-E) >and SiI-3124 (PCI-X) devices. > > 2. The AHCI driver, which supports the Intel ICH6 and latter devices, often >found on motherboard. > > 4. The NV_SATA driver which supports Nvidia ck804/mcp55 devices. > > Regards > Nigel Smith > > > This message posted from opensolaris.org > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs corruption w/ sil3114 sata controllers
I haven't seen the beginning of this discussion, but seeing SiI sets the fire alarm off here. The Silicon Image chipsets are renowned to be crap and causing data corruption. At least the variants that usually go onto mainboards. Based on this, I suggest that you should get a different card. -mg signature.asc Description: OpenPGP digital signature ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs corruption w/ sil3114 sata controllers
Mario, I don't have any issues getting a new card. The root of the discussion started because people did indeed post that they had good luck with them. In fact, when I went out there and google'd to find which cards would worked well, it seemed to be at the top of the list. I'm interested to know if it's something I can help resolve so other people don't have this problem or make sure people don't run into the same issue I do. Mario Goebbels wrote: > I haven't seen the beginning of this discussion, but seeing SiI sets the > fire alarm off here. > > The Silicon Image chipsets are renowned to be crap and causing data > corruption. At least the variants that usually go onto mainboards. Based > on this, I suggest that you should get a different card. > > -mg > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] DESC: A ZFS pool failed to open.
Hi, All the pools seem healthy and zfs file systems are all fine according to "zpool status -x" but during the boot we get the following error: fmadm faulty also returns this: -- degraded zfs://pool=pe09_01 8f5e62aa-c0af-4536-cef3-8e9d9169ea92 -- degraded zfs://pool=re09_01 eca6f995-12ba-ce3b-9a5b-d33f1b6580ac -- Thanks, SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major EVENT-TIME: Fri Oct 26 18:25:33 EDT 2007 PLATFORM: SUNW,Sun-Fire-T200, CSN: -, HOSTNAME: myserver SOURCE: zfs-diagnosis, REV: 1.0 EVENT-ID: 8168b743-76b2-ccc0-9816- e7387c93267b DESC: A ZFS pool failed to open. Refer to http://sun.com/msg/ZFS-8000-CS for more information. AUTO-RESPONSE: No automated response will occur. IMPACT: The pool data is unavailable REC-ACTION: Run 'zpool status -x' and either attach the missing device or restore from backup. SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major EVENT-TIME: Fri Oct 26 18:25:34 EDT 2007 PLATFORM: SUNW,Sun-Fire-T200, CSN: -, HOSTNAME: myserver SOURCE: zfs-diagnosis, REV: 1.0 EVENT-ID: 5c6e92ad-560b-ee04- b324-8f32f869 DESC: A ZFS pool failed to open. Refer to http://sun.com/msg/ZFS-8000-CS for more information. AUTO-RESPONSE: No automated response will occur. IMPACT: The pool data is unavailable REC-ACTION: Run 'zpool status -x' and either attach the missing device or restore from backup. Is my pool bad or this is just a bug? How make it go away? Thanks, This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] hierarchical share of zfs filesystem
I have a quick question about ZFS filesystems and snapshots and NFS shares. Is it possible to have one mount point for a set of ZFS filesystems? Lets say I have /export/homes as a ZFS filesystem and NFS shared, I create a new ZFS filesystem /export/homes/user1 I need to explicitly mount it. Only the mount point is visible under /export/homes. This has been standard NFS with devices mounted, ie. submounts are not exported, but that becomes very unhandy with dynamic ZFS and snapshots under NFS? Again, the question is if I can export all ZFS filesystems under a specific mount point. -- Jan Hendrik Mangold Sun Microsystems Senior Lead Staff Engineer 650-585-5484 (x81371) http://learning.central/wiki/display/~13922 "idle hands are the developers workshop" ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] hierarchical share of zfs filesystem
Jan Hendrik Mangold wrote: > I have a quick question about ZFS filesystems and snapshots and NFS shares. > > Is it possible to have one mount point for a set of ZFS filesystems? > Lets say I have /export/homes as a ZFS filesystem and NFS shared, I > create a new ZFS filesystem /export/homes/user1 I need to explicitly > mount it. Only the mount point is visible under /export/homes. This has > been standard NFS with devices mounted, ie. submounts are not exported, > but that becomes very unhandy with dynamic ZFS and snapshots under NFS? Solved by PSARC 2007/416 NFSv4 Mirror-mounts which integrated on Monday last week. http://opensolaris.org/os/community/on/flag-days/pages/2007102201/ All going well this should be in snv_77. -- Darren J Moffat ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] hierarchical share of zfs filesystem
On 31 Oct 2007, at 11:14, Darren J Moffat wrote: Solved by PSARC 2007/416 NFSv4 Mirror-mounts which integrated on Monday last week. that is great news. Especially since it confirms that I didn't do anything wrong, but that this feature just wasn't available yet ... :) -- Jan Hendrik Mangold Sun Microsystems Senior Lead Staff Engineer 650-585-5484 (x81371) http://learning.central/wiki/display/~13922 "idle hands are the developers workshop" ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS Quota Oddness
I've run across an odd issue with ZFS Quota's. This is an snv_43 system with several zones/zfs datasets, but only one effected. The dataset shows 10GB used, 12GB refered but when counting the files only has 6.7GB of data: zones/ABC10.8G 26.2G 12.0G /zones/ABC zones/[EMAIL PROTECTED]14.7M - 12.0G - [xxx:/zones/ABC/.zfs/snapshot/now] root# gdu --max-depth=1 -h . 43k ./dev 6.7G./root 1.5k./lu 6.7G. I don't understand what might the cause this disparity. This is an older box, snv_43. Any bugs that might apply, fixed or in progress? Thanks. benr. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] NFS: Cannot share a zfs dataset added to a labeled zone
- I set the mount point as follows. zfs set mountpoint=/zone/restricted/root/data zone/data - I then added the dataset to the restricted zone using zonecfg. The full path to the dataset is now /zone/restricted/root/zone/restricted/root/data. I am not sure if that is what you intended, but it is a result of adding it as a dataset to the zone after setting the mountpoint. - I updated the /zone/restricted/etc/dfs/dfstab with the following line. /usr/bin/share -F nfs -o rw /zone/restricted/root/zone/data - During reboot I receive the following error. cannot mount 'zone/data': mountpoint or dataset is busy svc:/system/filesystem/local:default: WARNING: /usr/sbin/zfs mount -a failed: exit status 1 Oct 31 14:43:08 svc.startd[19960]: svc:/system/filesystem/local:default: Method "/lib/svc/method/fs-local" failed with exit status 95. Oct 31 14:43:08 svc.startd[19960]: system/filesystem/local:default failed fatally: transitioned to maintenance (see 'svcs -xv' for details) - This is exactly the same problem that prompted the original message. Service fail during boot which prevent opening a console. This only occurs when you try to share the dataset. If you remove the line from /zone/restricted/etc/dfs/dfstab and reboot the zone everything works fine. Any ideas what I am doing wrong? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS Quota Oddness
On Wed, 2007-10-31 at 14:15 -0700, Ben Rockwood wrote: > I've run across an odd issue with ZFS Quota's. This is an snv_43 system with > several zones/zfs datasets, but only one effected. The dataset shows 10GB > used, 12GB refered but when counting the files only has 6.7GB of data: > > zones/ABC10.8G 26.2G 12.0G /zones/ABC > zones/[EMAIL PROTECTED]14.7M - 12.0G - > > [xxx:/zones/ABC/.zfs/snapshot/now] root# gdu --max-depth=1 -h . > 43k ./dev > 6.7G./root > 1.5k./lu > 6.7G. > > I don't understand what might the cause this disparity. is the pool built out of three-disk raidz1 groups and predates snv_42? if so, it's probably some facet of: 6288488 du reports misleading size on RAID-Z ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS Training
This class looks pretty good... http://www.sun.com/training/catalog/courses/SA-229-S10.xml On 10/31/07, Lisa Richards <[EMAIL PROTECTED]> wrote: > > > > > Is there a class on ZFS installation and administration ? > > > > Lisa Richards > > Zykis Corporation > > [EMAIL PROTECTED] > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS Training
Is there a class on ZFS installation and administration ? Lisa Richards Zykis Corporation [EMAIL PROTECTED] ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] the number of mount points
Hi, IHAC who wants to use ZFS for user's home directory. He is worry about the number of mount points. Does ZFS have any limitation about the number of mount points in a server? Regards, Jono ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss