Re: [zfs-discuss] pool wide corruption, "Bad exchange descriptor"
Well, I see no takers or even a hint... I've been playing with zdb to try to examine the pool, but I get: # zdb -b pool4_green zdb: can't open pool4_green: Bad exchange descriptor # zdb -d pool4_green zdb: can't open pool4_green: Bad exchange descriptor So I'm not sure how to debug using zdb. Is there something better or something else I should be looking at? The disks are all there, all online. How can I at least rollback to the last consistent bit of data on there? Or is all hope lost and I lost over 600GB of data? The worst part is that there are no errors in the logs and it just "disappeared" without a trace. The only logs are from subsequent reboots where it says a ZFS pool failed to open. It does not give me a warm & fuzzy about using ZFS as I've relied on it heavily in the past 5 years. Any advice would be well appreciated... On 7/2/2010 3:01 AM, Brian Kolaci wrote: I've recently acquired some storage and have been trying to copy data from a remote data center to hold backup data. The copies had been going for weeks, with about 600GB transferred so far, and then I noticed the throughput on the router stopped. I see a pool disappeared. # zpool status -x pool: pool4_green state: FAULTED status: The pool metadata is corrupted and the pool cannot be opened. action: Destroy and re-create the pool from a backup source. see: http://www.sun.com/msg/ZFS-8000-72 scrub: none requested config: NAME STATE READ WRITE CKSUM pool4_green FAULTED 0 0 1 corrupted data raidz2 ONLINE 0 0 6 c10t0d0 ONLINE 0 0 0 c10t1d0 ONLINE 0 0 0 c10t2d0 ONLINE 0 0 0 c10t3d0 ONLINE 0 0 0 c10t4d0 ONLINE 0 0 0 c10t5d0 ONLINE 0 0 0 c10t6d0 ONLINE 0 0 0 I powered down the system and all the disk systems and powered up fresh. I tried to clear the error: #zpool clear pool4_green internal error: Bad exchange descriptor Abort (core dumped) So then I took a look with zdb: #zdb -vvv pool4_green version=15 name='pool4_green' state=0 txg=83 pool_guid=3115817837859301858 hostid=237914636 hostname='galaxy' vdev_tree type='root' id=0 guid=3115817837859301858 children[0] type='raidz' id=0 guid=10261633106033684483 nparity=2 metaslab_array=24 metaslab_shift=36 ashift=9 asize=6997481881600 is_log=0 children[0] type='disk' id=0 guid=11313548069045029894 path='/dev/dsk/c10t0d0s0' devid='id1,s...@n60026b9040e26100139d8514065a1d67/a' phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@0,0:a' whole_disk=1 children[1] type='disk' id=1 guid=5547727760941401848 path='/dev/dsk/c10t1d0s0' devid='id1,s...@n60026b9040e26100139d851d06ec511d/a' phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@1,0:a' whole_disk=1 children[2] type='disk' id=2 guid=8407102896612298450 path='/dev/dsk/c10t2d0s0' devid='id1,s...@n60026b9040e26100139d85260770c1bf/a' phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@2,0:a' whole_disk=1 children[3] type='disk' id=3 guid=17509238716791782209 path='/dev/dsk/c10t3d0s0' devid='id1,s...@n60026b9040e26100139d852f07fea314/a' phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@3,0:a' whole_disk=1 children[4] type='disk' id=4 guid=18419120996062075464 path='/dev/dsk/c10t4d0s0' devid='id1,s...@n60026b9040e26100139d8537086c271f/a' phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@4,0:a' whole_disk=1 children[5] type='disk' id=5 guid=8308368067368943006 path='/dev/dsk/c10t5d0s0' devid='id1,s...@n60026b9040e26100139d85440934f640/a' phys_path='/p...@0,0/pci8086,3...@4/p
Re: [zfs-discuss] pool wide corruption, "Bad exchange descriptor"
On Jul 6, 2010, at 6:30 PM, Brian Kolaci wrote: > Well, I see no takers or even a hint... > > I've been playing with zdb to try to examine the pool, but I get: > > # zdb -b pool4_green > zdb: can't open pool4_green: Bad exchange descriptor > > # zdb -d pool4_green > zdb: can't open pool4_green: Bad exchange descriptor > > So I'm not sure how to debug using zdb. Is there something better or > something else I should be looking at? > The disks are all there, all online. How can I at least rollback to the last > consistent bit of data on there? > Or is all hope lost and I lost over 600GB of data? > > The worst part is that there are no errors in the logs and it just > "disappeared" without a trace. > The only logs are from subsequent reboots where it says a ZFS pool failed to > open. > > It does not give me a warm & fuzzy about using ZFS as I've relied on it > heavily in the past 5 years. > > Any advice would be well appreciated... You can download build 134 LiveCD boot off it and try 'zpool import -nfF pool4_green' for a start. regards victor > On 7/2/2010 3:01 AM, Brian Kolaci wrote: >> I've recently acquired some storage and have been trying to copy data from a >> remote data center to hold backup data. The copies had been going for >> weeks, with about 600GB transferred so far, and then I noticed the >> throughput on the router stopped. I see a pool disappeared. >> >> # zpool status -x >> >> pool: pool4_green >> state: FAULTED >> status: The pool metadata is corrupted and the pool cannot be opened. >> action: Destroy and re-create the pool from a backup source. >> see: http://www.sun.com/msg/ZFS-8000-72 >> scrub: none requested >> config: >> >>NAME STATE READ WRITE CKSUM >>pool4_green FAULTED 0 0 1 corrupted data >> raidz2 ONLINE 0 0 6 >>c10t0d0 ONLINE 0 0 0 >>c10t1d0 ONLINE 0 0 0 >>c10t2d0 ONLINE 0 0 0 >>c10t3d0 ONLINE 0 0 0 >>c10t4d0 ONLINE 0 0 0 >>c10t5d0 ONLINE 0 0 0 >>c10t6d0 ONLINE 0 0 0 >> >> I powered down the system and all the disk systems and powered up fresh. >> I tried to clear the error: >> >> #zpool clear pool4_green >> internal error: Bad exchange descriptor >> Abort (core dumped) >> >> So then I took a look with zdb: >> >> #zdb -vvv pool4_green >>version=15 >>name='pool4_green' >>state=0 >>txg=83 >>pool_guid=3115817837859301858 >>hostid=237914636 >>hostname='galaxy' >>vdev_tree >>type='root' >>id=0 >>guid=3115817837859301858 >>children[0] >>type='raidz' >>id=0 >>guid=10261633106033684483 >>nparity=2 >>metaslab_array=24 >>metaslab_shift=36 >>ashift=9 >>asize=6997481881600 >>is_log=0 >>children[0] >>type='disk' >>id=0 >>guid=11313548069045029894 >>path='/dev/dsk/c10t0d0s0' >>devid='id1,s...@n60026b9040e26100139d8514065a1d67/a' >> >> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@0,0:a' >>whole_disk=1 >>children[1] >>type='disk' >>id=1 >>guid=5547727760941401848 >>path='/dev/dsk/c10t1d0s0' >>devid='id1,s...@n60026b9040e26100139d851d06ec511d/a' >> >> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@1,0:a' >>whole_disk=1 >>children[2] >>type='disk' >>id=2 >>guid=8407102896612298450 >>path='/dev/dsk/c10t2d0s0' >>devid='id1,s...@n60026b9040e26100139d85260770c1bf/a' >> >> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@2,0:a' >>whole_disk=1 >>children[3] >>type='disk' >>id=3 >>guid=17509238716791782209 >>path='/dev/dsk/c10t3d0s0' >>devid='id1,s...@n60026b9040e26100139d852f07fea314/a' >> >> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@3,0:a' >>whole_disk=1 >>children[4] >>type='disk' >>id=4 >>guid=18419120996062075464 >>path='/dev/dsk/c10t4d0s0' >>devid='id1,s...@n60026b9040e26100139d8537086c271f/a' >>
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
On Jul 4, 2010, at 4:58 AM, Andrew Jones wrote: > Victor, > > The zpool import succeeded on the next attempt following the crash that I > reported to you by private e-mail! >From the threadlist it looked like system was pretty low on memory with stacks >of userland stuff swapped out, hence system was not responsive, but it was >able to complete inconsistent dataset processing in the end. > > For completeness, this is the final status of the pool: > > > pool: tank > state: ONLINE > scan: resilvered 1.50K in 165h28m with 0 errors on Sat Jul 3 08:02:30 2010 > config: > >NAMESTATE READ WRITE CKSUM >tankONLINE 0 0 0 > raidz2-0 ONLINE 0 0 0 >c0t0d0 ONLINE 0 0 0 >c0t1d0 ONLINE 0 0 0 >c0t2d0 ONLINE 0 0 0 >c0t3d0 ONLINE 0 0 0 >c0t4d0 ONLINE 0 0 0 >c0t5d0 ONLINE 0 0 0 >c0t6d0 ONLINE 0 0 0 >c0t7d0 ONLINE 0 0 0 >cache > c2t0d0ONLINE 0 0 0 > > errors: No known data errors > Good. Run 'zpool scrub' to make sure there are no other errors. regards victor > Thank you very much for your help. We did not need to add additional RAM to > solve this, in the end. Instead, we needed to persist with the import through > several panics to finally work our way through the large inconsistent > dataset; it is unclear whether the resilvering caused additional processing > delay. Unfortunately, the delay made much of the data quite stale, now that > it's been recovered. > > It does seem that zfs would benefit tremendously from a better (quicker and > more intuitive?) set of recovery tools, that are available to a wider range > of users. It's really a shame, because the features and functionality in zfs > are otherwise absolutely second to none. > > /Andrew[i][/i][i][/i][i][/i][i][/i][i][/i] > -- > This message posted from opensolaris.org > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Kernel Panic on zpool clean
On Jul 3, 2010, at 1:20 PM, George wrote: >> Because of that I'm thinking that I should try >> to change the hostid when booted from the CD to be >> the same as the previously installed system to see if >> that helps - unless that's likely to confuse it at >> all...? > > I've now tried changing the hostid using the code from > http://forums.sun.com/thread.jspa?threadID=5075254 NB: you need to leave this > running in a separate terminal. > > This changes the start of "zpool import" to > > pool: storage2 >id: 14701046672203578408 > state: FAULTED > status: The pool metadata is corrupted. > action: The pool cannot be imported due to damaged devices or data. >The pool may be active on another system, but can be imported using >the '-f' flag. > see: http://www.sun.com/msg/ZFS-8000-72 > > > but otherwise nothing is changed with respect to trying to import or clear > the pool. The pool is 8TB and the machine has 4GB but as far as I can see via > top the commands aren't failing due to a lack of memory. > > I'm a bit stumped now. The only thing else I can think to try is inserting > c9t4d4 (the new drive) and removing c6t4d0 (which should be fine). The > problem with this though is that it relies on c7t4d0 (which is faulty) and so > it assumes that the errors can be cleared, the replace stopped and the drives > swapped back before further errors happen. I think it is quite likely to be possible to get readonly access to your data, but this requires modified ZFS binaries. What is your pool version? What build do you have installed on your system disk or available as LiveCD? regards victor ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS recovery tools
On Jul 4, 2010, at 1:33 AM, R. Eulenberg wrote: > R. Eulenberg web.de> writes: > >> op> I was setting up a new systen (osol 2009.06 >>> and updating to op> the lastest version of osol/dev - snv_134 - >>> with op> deduplication) and then I tried to import my >>> backup zpool, but op> it does not work. op> # zpool import -f tank1 op> cannot import 'tank1': one or more devices >>> is currently unavailable op> Destroy and re-create the pool from a backup >>> source op> Any other option (-F, -X, -V, -D) and any >>> combination of them op> doesn't helps too. >> R., please let us know if the 'zdb -e -bcsvL >>> ' incantation Sigbjorn suggested ends up working for you or not. >> >> Hi, >> >> it seems that you are talking about my case of trouble. >> I answered in this thread: >> >> http://www.opensolaris.org/jive/thread.jspa?messageID=485342&tstart=0#485342 >> >> I hope of any ideas helping me. >> >> Regards >> Ron > > > Hi, > > several days ago I added this: > > set /zfs/:zfs_recover=/1/ > set aok=/1/ > > to the /etc/system file and ran: > > zdb -e -bcsvL tank1 > settings in the /etc/systm file do not affect zdb in any way. In recent builds (135+) there's an option -A for zdb that allows to simulate setting one or both of these parameters, but anyway it is not useful in your case. > without an output and without a prompt (prozess hangs up) How did you arrive at this conclusion? Did you check "pstack `pgrep zdb`/1" output a few times? > and result of running: > > zdb -eC tank1 > > was the same. > I hope you could help me because the problem isn't solved. > > Regards > Ron > > > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Kernel Panic on zpool clean
> I think it is quite likely to be possible to get readonly access to > your data, but this requires modified ZFS binaries. What is your pool > version? What build do you have installed on your system disk or > available as LiveCD? Sorry, but does this mean if ZFS can't write to the drives, access to the pool won't be possible? If so, that's rather scary... Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Kernel Panic on zpool clean
On Jun 28, 2010, at 11:27 PM, George wrote: > Again this core dumps when I try to do "zpool clear storage2" > > Does anyone have any suggestions what would be the best course of action now? Do you have any crahsdumps saved? First one is most interesting one... ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Hashing files rapidly on ZFS
Daniel Carosone wrote: > Something similar would be useful, and much more readily achievable, > from ZFS from such an application, and many others. Rather than a way > to compare reliably between two files for identity, I'ld liek a way to > compare identity of a single file between two points in time. If my > application can tell quickly that the file content is unaltered since > last time I saw the file, I can avoid rehashing the content and use a > stored value. If I can achieve this result for a whole directory > tree, even better. This would be great for any kind of archiving software. Aren't zfs checksums already ready to solve this? If a file changes, it's dnodes' checksum changes, the checksum of the directory it is in and so forth all the way up to the uberblock. There may be ways a checksum changes without a real change in the files content, but the other way round should hold. If the checksum didn't change, the file didn't change. So the only missing link is a way to determine zfs's checksum for a file/directory/dataset. Am I missing something here? Of course atime update should be turned off, otherwise the checksum will get changed by the archiving agent. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Remove non-redundant disk
Hello list, This has probably been discussed, however I would like to bring it up again, so that the powers that be, know someone else is looking for this feature. I would like to be able to shrink a pool and remove a non-redundant disk. Is this something that is in the works? It would be fantastic if I had this capability. Thanks! - Cassandra Unix Administrator "From a little spark may burst a mighty flame." -Dante Alighieri ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Remove non-redundant disk
- Original Message - > Hello list, > > This has probably been discussed, however I would like to bring it up > again, so that the powers that be, know someone else is looking for > this feature. > > I would like to be able to shrink a pool and remove a non-redundant > disk. > > Is this something that is in the works? > > It would be fantastic if I had this capability. You're a little unclear on what you want, but it seems to me you want to change a raidz2 to a raidz1 or something like that. That is, AFAIK, in the works, with the block rewrite functionality. As with most other parts of progress in OpenSolaris, nothing is clear about when or if this will get integrated into the system. For now, you can't change a raidz(n) VDEV and you can't detach a VDEV from a pool. The only way is to build a new pool and move the data with things like zfs send/receive. You can also remove a drive from a raidz(n), and reducing its redundancy, but you can't change the value of n. Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Remove non-redundant disk
The pool is not redundant, so I would suppose, yes, is is Raid-1 on the software level. I have a few drives, which are on a specific array, which I would like to remove from this pool. I have discovered the "replace" command, and I am going to try and replace, 1 for 1, the drives I would like to remove. However, it would be nice if there were a way to simply "remove" the disks, if space allowed. - Cassandra (609) 243-2413 Unix Administrator "From a little spark may burst a mighty flame." -Dante Alighieri On Tue, Jul 6, 2010 at 11:55 AM, Roy Sigurd Karlsbakk wrote: > - Original Message - > > Hello list, > > > > This has probably been discussed, however I would like to bring it up > > again, so that the powers that be, know someone else is looking for > > this feature. > > > > I would like to be able to shrink a pool and remove a non-redundant > > disk. > > > > Is this something that is in the works? > > > > It would be fantastic if I had this capability. > > You're a little unclear on what you want, but it seems to me you want to > change a raidz2 to a raidz1 or something like that. That is, AFAIK, in the > works, with the block rewrite functionality. As with most other parts of > progress in OpenSolaris, nothing is clear about when or if this will get > integrated into the system. > > For now, you can't change a raidz(n) VDEV and you can't detach a VDEV from > a pool. The only way is to build a new pool and move the data with things > like zfs send/receive. You can also remove a drive from a raidz(n), and > reducing its redundancy, but you can't change the value of n. > > Vennlige hilsener / Best regards > > roy > -- > Roy Sigurd Karlsbakk > (+47) 97542685 > r...@karlsbakk.net > http://blogg.karlsbakk.net/ > -- > I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det > er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av > idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate > og relevante synonymer på norsk. > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Consequences of resilvering failure
I'm just about to start using ZFS in a RAIDZ configuration for a home file server (mostly holding backups), and I wasn't clear on what happens if data corruption is detected while resilvering. For example: let's say I'm using RAIDZ1 and a drive fails. I pull it and put in a new one. While resilvering, ZFS detects corrupt data on one of the remaining disks. Will the resilvering continue, with some files marked as containing errors, or will it simply fail? (I found this process[1] to repair damaged data, but I wasn't sure what would happen if it was detected in the middle of resilvering.) I will of course have a backup of the pool, but I may opt for additional backup if the entire pool could be lost due to data corruption (as opposed to just a few files potentially being lost). Thanks, Michael [1] http://dlc.sun.com/osol/docs/content/ZFSADMIN/gbbwl.html ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Legacy MountPoint for /rpool/ROOT
I have two different servers with ZFS root but both of them has different mountpoint for rpool/ROOT one is /rpool/ROOT and other is legacy. Whats the difference between the two and which is the one we should keep. And why there is 3 different zfs datasets rpool, rpool/ROOT and rpool/ROOT/zfsBE ? # zfs list -r rpool NAME USED AVAIL REFER MOUNTPOINT rpool 24.1G 110G98K/rpool rpool/ROOT6.08G 110G21K/rpool/ROOT rpool/ROOT/zfsBE 6.08G 110G 6.08G / ** zfs list -r rpool NAMEUSED AVAIL REFER MOUNTPOINT rpool 34.0G 99.9G94K /rpool rpool/ROOT 27.9G 99.9G18K legacy rpool/ROOT/s10s_u6 27.9G 99.9G 27.9G / -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Legacy MountPoint for /rpool/ROOT
I have two different servers with ZFS root but both of them has different mountpoint for rpool/ROOT one is /rpool/ROOT and other is legacy. Whats the difference between the two and which is the one we should keep. And why there is 3 different zfs datasets rpool, rpool/ROOT and rpool/ROOT/zfsBE ? Can anyone explain this to me . Thanx in advance # zfs list -r rpool NAME USED AVAIL REFER MOUNTPOINT rpool 24.1G 110G98K/rpool rpool/ROOT6.08G 110G21K/rpool/ROOT rpool/ROOT/zfsBE 6.08G 110G 6.08G / ** zfs list -r rpool NAMEUSED AVAIL REFER MOUNTPOINT rpool 34.0G 99.9G94K /rpool rpool/ROOT 27.9G 99.9G18K legacy rpool/ROOT/s10s_u6 27.9G 99.9G 27.9G / -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Help with Faulted Zpool Call for Help(Cross post)
Hello list, I posted this a few days ago on opensolaris-discuss@ list I am posting here, because there my be too much noise on other lists I have been without this zfs set for a week now. My main concern at this point,is it even possible to recover this zpool. How does the metadata work? what tool could is use to rebuild the corrupted parts or even find out what parts are corrupted. most but not all of these disks were Hitachi Retail 1TB didks. I have a Fileserver that runs FreeBSD 8.1 (zfs v14) after a poweroutage, I am unable to import my zpool named Network my pool is made up of 6 1TB disks configured in raidz. there is ~1.9TB of actual data on this pool. I have loaded Open Solaris svn_134 on a seprate boot disk, in hopes of recovering my zpool. on Open Solaris 134, I am not able to import my zpool almost everything I try gives me cannot import 'Network': I/O error I have done quite a bit of searching, and I found that import -fFX Network should work however after ~ 20 hours this hard locks Open Solaris (however it does return a ping) here is a list of commands that I have run on Open Solaris http://www.puffybsd.com/zfsv14.txt if anyone could help me use zdb or mdb to recover my pool I would very much appreciate it. I believe the metadata is corrupt on my zpool -- Sam Fourman Jr. Fourman Networks http://www.fourmannetworks.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Remove non-redundant disk
- Original Message - > The pool is not redundant, so I would suppose, yes, is is Raid-1 on > the software level. > > I have a few drives, which are on a specific array, which I would like > to remove from this pool. > > I have discovered the "replace" command, and I am going to try and > replace, 1 for 1, the drives I would like to remove. > > However, it would be nice if there were a way to simply "remove" the > disks, if space allowed. zfs attach new drive, zfs detach old drive, that will replace the drive without much hazzle Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Legacy MountPoint for /rpool/ROOT
On 07/ 6/10 10:56 AM, Ketan wrote: I have two different servers with ZFS root but both of them has different mountpoint for rpool/ROOT one is /rpool/ROOT and other is legacy. It should be legacy. Whats the difference between the two and which is the one we should keep. And why there is 3 different zfs datasets rpool, rpool/ROOT and rpool/ROOT/zfsBE ? rpool is the top-level dataset of the pool (every pool has one). rpool/ROOT is just a grouping mechanism for the roots of all BEs on the system. rpool/ROOT/zfsBE is the root file system of the BE called "zfsBE". lori # zfs list -r rpool NAME USED AVAIL REFER MOUNTPOINT rpool 24.1G 110G98K/rpool rpool/ROOT6.08G 110G21K/rpool/ROOT rpool/ROOT/zfsBE 6.08G 110G 6.08G / ** zfs list -r rpool NAMEUSED AVAIL REFER MOUNTPOINT rpool 34.0G 99.9G94K /rpool rpool/ROOT 27.9G 99.9G18K legacy rpool/ROOT/s10s_u6 27.9G 99.9G 27.9G / ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Consequences of resilvering failure
- Original Message - > I'm just about to start using ZFS in a RAIDZ configuration for a home > file server (mostly holding backups), and I wasn't clear on what > happens if data corruption is detected while resilvering. For example: > let's say I'm using RAIDZ1 and a drive fails. I pull it and put in a > new one. While resilvering, ZFS detects corrupt data on one of the > remaining disks. Will the resilvering continue, with some files marked > as containing errors, or will it simply fail? The pool will remain available, but you will have data corruption. The simple way to avoid this, is to use a raidz2, where the chances are far lower for data corruption. Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Remove non-redundant disk
I tried zfs replace, however the new drive is slightly smaller, and even with a -f, it refuses to replace the drive. I guess i will have to export the pool and destroy this one to get my drives back. Still would like the ability to shrink a pool. - Cassandra (609) 243-2413 Unix Administrator "From a little spark may burst a mighty flame." -Dante Alighieri On Tue, Jul 6, 2010 at 1:02 PM, Roy Sigurd Karlsbakk wrote: > - Original Message - > > The pool is not redundant, so I would suppose, yes, is is Raid-1 on > > the software level. > > > > I have a few drives, which are on a specific array, which I would like > > to remove from this pool. > > > > I have discovered the "replace" command, and I am going to try and > > replace, 1 for 1, the drives I would like to remove. > > > > However, it would be nice if there were a way to simply "remove" the > > disks, if space allowed. > > zfs attach new drive, zfs detach old drive, that will replace the drive > without much hazzle > > Vennlige hilsener / Best regards > > roy > -- > Roy Sigurd Karlsbakk > (+47) 97542685 > r...@karlsbakk.net > http://blogg.karlsbakk.net/ > -- > I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det > er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av > idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate > og relevante synonymer på norsk. > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
> > Good. Run 'zpool scrub' to make sure there are no > other errors. > > regards > victor > Yes, scrubbed successfully with no errors. Thanks again for all of your generous assistance. /AJ -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS fsck?
Hi all With several messages in here about troublesome zpools, would there be a good reason to be able to fsck a pool? As in, check the whole thing instead of having to boot into live CDs and whatnot? Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS fsck?
- Original Message - > From: "Roy Sigurd Karlsbakk" > To: "OpenSolaris ZFS discuss" > Sent: Tuesday, 6 July, 2010 6:35:51 PM > Subject: [zfs-discuss] ZFS fsck? > Hi all > > With several messages in here about troublesome zpools, would there be > a good reason to be able to fsck a pool? As in, check the whole thing > instead of having to boot into live CDs and whatnot? > > Vennlige hilsener / Best regards > > roy Scrub? :) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS fsck?
On Tue, 6 Jul 2010, Roy Sigurd Karlsbakk wrote: Hi all With several messages in here about troublesome zpools, would there be a good reason to be able to fsck a pool? As in, check the whole thing instead of having to boot into live CDs and whatnot? You can do this with "zpool scrub". It visits every allocated block and verifies that everything is correct. It's not the same as fsck in that scrub can detect and repair problems with the pool still online and all datasets mounted, whereas fsck cannot handle mounted filesystems. If you really want to use it on an exported pool, you can use zdb, although it might take some time. Here's an example on a small empty pool: # zpool create -f mypool raidz c4t1d0s0 c4t2d0s0 c4t3d0s0 c4t4d0s0 c4t5d0s0 # zpool list mypool NAME SIZE ALLOC FREECAP DEDUP HEALTH ALTROOT mypool 484M 280K 484M 0% 1.00x ONLINE - # zpool export mypool # zdb -ebcc mypool Traversing all blocks to verify checksums and verify nothing leaked ... No leaks (block sum matches space maps exactly) bp count: 48 bp logical:378368 avg: 7882 bp physical:39424 avg:821 compression: 9.60 bp allocated: 185344 avg: 3861 compression: 2.04 bp deduped: 0ref>1: 0 deduplication: 1.00 SPA allocated: 185344 used: 0.04% # ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS fsck?
> You can do this with "zpool scrub". It visits every allocated block > and > verifies that everything is correct. It's not the same as fsck in that > scrub can detect and repair problems with the pool still online and > all > datasets mounted, whereas fsck cannot handle mounted filesystems. > > If you really want to use it on an exported pool, you can use zdb, > although it might take some time. Here's an example on a small empty > pool: > > # zpool create -f mypool raidz c4t1d0s0 c4t2d0s0 c4t3d0s0 c4t4d0s0 > c4t5d0s0 > # zpool list mypool > NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT > mypool 484M 280K 484M 0% 1.00x ONLINE - > # zpool export mypool > # zdb -ebcc mypool ... what I'm saying is that there are several posts in here where the only solution is to boot onto a live cd and then do an import, due to metadata corruption. This should be doable from the installed system Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS fsck?
On Tue, 6 Jul 2010, Roy Sigurd Karlsbakk wrote: what I'm saying is that there are several posts in here where the only solution is to boot onto a live cd and then do an import, due to metadata corruption. This should be doable from the installed system Ah, I understand now. A couple of things worth noting: - if the root filesystem in a boot pool cannot be mounted, it's problematic to access the tools necessary to repair it. So going to a livecd (or a network boot for that matter) is the best way forward. - if the tools available to failsafe are insufficient to repair a pool, then booting off a livecd/network is the only way forward. It is also worth pointing out here that the 134a build has the pool recovery code built-in. The "-F" option to zpool import only became available after build 128 or 129. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] NexentaStor 3.0.3 vs OpenSolaris - Patches more up to date?
Release Notes information: If there are new features, each release is added to http://www.nexenta.com/corp/documentation/release-notes-support. If just bug fixes, then the Changelog listing is updated: http://www.nexenta.com/corp/documentation/nexentastor-changelog Regards, Spandana ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Help with Faulted Zpool Call for Help(Cross post)
Hi Sam, In general, FreeBSD uses different device naming conventions and power failures seem to clobber disk labeling. The "I/O error" message also points to problems accessing these disks. I'm not sure if this helps, but I see that the 6 disks from the zdb -e output are indicated as c7t0d0p0 --> c7t5d0p0, and the paths resemble /dev/dsk/c7t1d0p0, which makes sense. When you look at the individual labels using zdb -l /dev/dsk/c7t1d0p0, the physical path for this disk looks identical, but the path=/dev/da4 and so on. I'm not familiar with the /dev/da* device naming. When I use zdb -l on my pool's disks, I see the same phys_path and the same path= for each disk. I hoping someone else who has patched their disks back together after a power failure can comment. Thanks, Cindy On 07/06/10 11:02, Sam Fourman Jr. wrote: Hello list, I posted this a few days ago on opensolaris-discuss@ list I am posting here, because there my be too much noise on other lists I have been without this zfs set for a week now. My main concern at this point,is it even possible to recover this zpool. How does the metadata work? what tool could is use to rebuild the corrupted parts or even find out what parts are corrupted. most but not all of these disks were Hitachi Retail 1TB didks. I have a Fileserver that runs FreeBSD 8.1 (zfs v14) after a poweroutage, I am unable to import my zpool named Network my pool is made up of 6 1TB disks configured in raidz. there is ~1.9TB of actual data on this pool. I have loaded Open Solaris svn_134 on a seprate boot disk, in hopes of recovering my zpool. on Open Solaris 134, I am not able to import my zpool almost everything I try gives me cannot import 'Network': I/O error I have done quite a bit of searching, and I found that import -fFX Network should work however after ~ 20 hours this hard locks Open Solaris (however it does return a ping) here is a list of commands that I have run on Open Solaris http://www.puffybsd.com/zfsv14.txt if anyone could help me use zdb or mdb to recover my pool I would very much appreciate it. I believe the metadata is corrupt on my zpool ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] NexentaStor 3.0.3 vs OpenSolaris - Patches more up to date?
On Tue, Jul 6, 2010 at 4:06 PM, Spandana Goli wrote: > Release Notes information: > If there are new features, each release is added to > http://www.nexenta.com/corp/documentation/release-notes-support. > > If just bug fixes, then the Changelog listing is updated: > http://www.nexenta.com/corp/documentation/nexentastor-changelog Is there a bug tracker were one can objectively list all the bugs (with details) that went into a release ? "Many bug fixes" is a bit too general. -- Giovanni Tirloni gtirl...@sysdroid.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zpool throughput: snv 134 vs 138 vs 143
Hi all, I've noticed something strange in the throughput in my zpool between different snv builds, and I'm not sure if it's an inherent difference in the build or a kernel parameter that is different in the builds. I've setup two similiar machines and this happens with both of them. Each system has 16 2TB Samsung HD203WI drives (total) directly connected to two LSI 3081E-R 1068e cards with IT firmware in one raidz3 vdev. In both computers, after a fresh installation of snv 134, the throughput is a maximum of about 300 MB/s during scrub or something like "dd if=/dev/zero bs=1024k of=bigfile". If I bfu to snv 138, I then get throughput of about 700 MB/s with both scrub or a single thread dd. I assumed at first this was some sort of bug or regression in 134 that made it slow. However, I've now tested also from the fresh 134 installation, compiling the OS/Net build 143 from the mercurial repository and booting into it, after which the dd throughput is still only about 300 MB/s just like snv 134. The scrub throughput in 143 is even slower, rarely surpassing 150 MB/s. I wonder if the scrubbing being extra slow here is related to the additional statistics displayed during the scrub that didn't used to be shown. Is there some kind of debug option that might be enabled in the 134 build and persist if I compile snv 143 which would be off if I installed a 138 through bfu? If not, it makes me think that the bfu to 138 is changing the configuration somewhere to make it faster rather than fixing a bug or being a debug flag on or off. Does anyone have any idea what might be happening? One thing I haven't tried is bfu'ing to 138, and from this faster working snv 138 installing the snv 143 build, which may possibly create a 143 that performs faster if it's simply a configuration parameter. I'm not sure offhand if installing source-compiled ON builds from a bfu'd rpool is supported, although I suppose it's simple enough to try. Thanks, Chad Cantwell ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] pool wide corruption, "Bad exchange descriptor"
On Jul 6, 2010, at 7:30 AM, Brian Kolaci wrote: > Well, I see no takers or even a hint... > > I've been playing with zdb to try to examine the pool, but I get: > > # zdb -b pool4_green > zdb: can't open pool4_green: Bad exchange descriptor For the archives, EBADE "Bad exchange descriptor" was repurposed as ECKSUM in zio.h. This agrees with the zpool status below. I recommend Victor's advice. -- richard > > # zdb -d pool4_green > zdb: can't open pool4_green: Bad exchange descriptor > > So I'm not sure how to debug using zdb. Is there something better or > something else I should be looking at? > The disks are all there, all online. How can I at least rollback to the last > consistent bit of data on there? > Or is all hope lost and I lost over 600GB of data? > > The worst part is that there are no errors in the logs and it just > "disappeared" without a trace. > The only logs are from subsequent reboots where it says a ZFS pool failed to > open. > > It does not give me a warm & fuzzy about using ZFS as I've relied on it > heavily in the past 5 years. > > Any advice would be well appreciated... > > On 7/2/2010 3:01 AM, Brian Kolaci wrote: >> I've recently acquired some storage and have been trying to copy data from a >> remote data center to hold backup data. The copies had been going for >> weeks, with about 600GB transferred so far, and then I noticed the >> throughput on the router stopped. I see a pool disappeared. >> >> # zpool status -x >> >> pool: pool4_green >> state: FAULTED >> status: The pool metadata is corrupted and the pool cannot be opened. >> action: Destroy and re-create the pool from a backup source. >> see: http://www.sun.com/msg/ZFS-8000-72 >> scrub: none requested >> config: >> >>NAME STATE READ WRITE CKSUM >>pool4_green FAULTED 0 0 1 corrupted data >> raidz2 ONLINE 0 0 6 >>c10t0d0 ONLINE 0 0 0 >>c10t1d0 ONLINE 0 0 0 >>c10t2d0 ONLINE 0 0 0 >>c10t3d0 ONLINE 0 0 0 >>c10t4d0 ONLINE 0 0 0 >>c10t5d0 ONLINE 0 0 0 >>c10t6d0 ONLINE 0 0 0 >> >> I powered down the system and all the disk systems and powered up fresh. >> I tried to clear the error: >> >> #zpool clear pool4_green >> internal error: Bad exchange descriptor >> Abort (core dumped) >> >> So then I took a look with zdb: >> >> #zdb -vvv pool4_green >>version=15 >>name='pool4_green' >>state=0 >>txg=83 >>pool_guid=3115817837859301858 >>hostid=237914636 >>hostname='galaxy' >>vdev_tree >>type='root' >>id=0 >>guid=3115817837859301858 >>children[0] >>type='raidz' >>id=0 >>guid=10261633106033684483 >>nparity=2 >>metaslab_array=24 >>metaslab_shift=36 >>ashift=9 >>asize=6997481881600 >>is_log=0 >>children[0] >>type='disk' >>id=0 >>guid=11313548069045029894 >>path='/dev/dsk/c10t0d0s0' >>devid='id1,s...@n60026b9040e26100139d8514065a1d67/a' >> >> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@0,0:a' >>whole_disk=1 >>children[1] >>type='disk' >>id=1 >>guid=5547727760941401848 >>path='/dev/dsk/c10t1d0s0' >>devid='id1,s...@n60026b9040e26100139d851d06ec511d/a' >> >> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@1,0:a' >>whole_disk=1 >>children[2] >>type='disk' >>id=2 >>guid=8407102896612298450 >>path='/dev/dsk/c10t2d0s0' >>devid='id1,s...@n60026b9040e26100139d85260770c1bf/a' >> >> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@2,0:a' >>whole_disk=1 >>children[3] >>type='disk' >>id=3 >>guid=17509238716791782209 >>path='/dev/dsk/c10t3d0s0' >>devid='id1,s...@n60026b9040e26100139d852f07fea314/a' >> >> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@3,0:a' >>whole_disk=1 >>children[4] >>type='disk' >>id=4 >>guid=18419120996062075464 >>path='/dev/dsk/c10t4d0s0' >>devid='id1,s...
Re: [zfs-discuss] Help with Faulted Zpool Call for Help(Cross post)
On Jul 6, 2010, at 10:02 AM, Sam Fourman Jr. wrote: > Hello list, > > I posted this a few days ago on opensolaris-discuss@ list > I am posting here, because there my be too much noise on other lists > > I have been without this zfs set for a week now. > My main concern at this point,is it even possible to recover this zpool. > > How does the metadata work? what tool could is use to rebuild the > corrupted parts > or even find out what parts are corrupted. > > > most but not all of these disks were Hitachi Retail 1TB didks. > > > I have a Fileserver that runs FreeBSD 8.1 (zfs v14) > after a poweroutage, I am unable to import my zpool named Network > my pool is made up of 6 1TB disks configured in raidz. > there is ~1.9TB of actual data on this pool. > > I have loaded Open Solaris svn_134 on a seprate boot disk, > in hopes of recovering my zpool. > > on Open Solaris 134, I am not able to import my zpool > almost everything I try gives me cannot import 'Network': I/O error > > I have done quite a bit of searching, and I found that import -fFX > Network should work > however after ~ 20 hours this hard locks Open Solaris (however it does > return a ping) > > here is a list of commands that I have run on Open Solaris > > http://www.puffybsd.com/zfsv14.txt You ran "zdb -l /dev/dsk/c7t5d0s2" which is not the same as "zdb -l /dev/dsk/c7t5d0p0" because of the default partitioning. In Solaris c*t*d*p* are fdisk partitions and c*t*d*s* are SMI or EFI slices. This why label 2&3 could not be found and can be part of the problem to start. Everything in /dev/dsk and /dev/rdsk is a symlink or directory, so you can fake them out with a temporary directory and clever use of the "zpool import -d" command. Examples are in the archives. -- richard > > if anyone could help me use zdb or mdb to recover my pool > I would very much appreciate it. > > I believe the metadata is corrupt on my zpool > > > -- > > Sam Fourman Jr. > Fourman Networks > http://www.fourmannetworks.com > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Richard Elling rich...@nexenta.com +1-760-896-4422 ZFS and NexentaStor training, Rotterdam, July 13-15, 2010 http://nexenta-rotterdam.eventbrite.com/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Lost ZIL Device
Hello All, I've recently run into an issue I can't seem to resolve. I have been running a zpool populated with two RAID-Z1 VDEVs and a file on the (separate) OS drive for the ZIL: raidz1-0 ONLINE c12t0d0 ONLINE c12t1d0 ONLINE c12t2d0 ONLINE c12t3d0 ONLINE raidz1-2 ONLINE c12t4d0 ONLINE c12t5d0 ONLINE c13t0d0 ONLINE c13t1d0 ONLINE logs /ZIL-Log.img This was running on Nexenta Community Edition v3. Everything was going smoothly until today when the OS hard drive crashed and I was not able to boot from it any longer. I had migrated this setup from an OpenSolaris install some months back and I still had the old drive intact. I put it in the system, booted it up and tried to import the zpool. Unfortunately, I have not been successful. Previously when migrating from OSOL to Nexenta I was able to get the new system to recognize and import the ZIL device file. Since it has been lost in the drive crash I have not been able to duplicate that success. Here is the output from a 'zpool import' command: pool: tank id: 9013303135438223804 state: UNAVAIL status: The pool was last accessed by another system. action: The pool cannot be imported due to damaged devices or data. see: http://www.sun.com/msg/ZFS-8000-EY config: tank UNAVAIL missing device raidz1-0 ONLINE c12t0d0 ONLINE c12t1d0 ONLINE c12t5d0 ONLINE c12t3d0 ONLINE raidz1-2 ONLINE c12t4d0 ONLINE c12t2d0 ONLINE c13t0d0 ONLINE c13t1d0 ONLINE I created a new file for the ZIL (using mkfile) and tried to specify it for inclusion with -d but it doesn't get recognized. Probably because it was never part of the original zpool. I also symlinked the new ZIL file into /dev/dsk but that didn't make any difference either. Any suggestions? Andrew Kener ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Remove non-redundant disk
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- > boun...@opensolaris.org] On Behalf Of Cassandra Pugh > > I would like to be able to shrink a pool and remove a non-redundant > disk. > > Is this something that is in the works? I think the request is to remove vdev's from a pool. Not currently possible. Is this in the works? ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Lost ZIL Device
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- > boun...@opensolaris.org] On Behalf Of Andrew Kener > > the OS hard drive crashed [and log device] Here's what I know: In zpool >= 19, if you import this, it will prompt you to confirm the loss of the log device, and then it will import. Here's what I have heard: The ability to import with a failed log device as described above, was created right around zpool 14 or 15, not quite sure which. Here's what I don't know: If the failed zpool was some version which was too low ... and you try to import on an OS which is capable of a much higher version of zpool ... Can the newer OS handle it just because the newer OS is able to handle a newer version of zpool? Or maybe the version of the failed pool is the one that matters, regardless of what the new OS is capable of doing now? ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Expected throughput
Under FreeBSD I've seen zpool scrub sustain nearly 500 MB/s in pools with large files (a pool with eight MIRROR vdevs on two Silicon Image 3124 controllers). You need to carefully look for bottlenecks in the hardware. You don't indicate how the disks are attached. I would measure the total bandwidth of the disks outside of ZFS (ie, dd if=disk of=/dev/null bs=128k, one per disk on all disks simultaneously). -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss