Re: [zfs-discuss] pool wide corruption, "Bad exchange descriptor"

2010-07-06 Thread Brian Kolaci

 Well, I see no takers or even a hint...

I've been playing with zdb to try to examine the pool, but I get:

# zdb -b pool4_green
zdb: can't open pool4_green: Bad exchange descriptor

# zdb -d pool4_green
zdb: can't open pool4_green: Bad exchange descriptor

So I'm not sure how to debug using zdb.  Is there something better or something 
else I should be looking at?
The disks are all there, all online.  How can I at least rollback to the last 
consistent bit of data on there?
Or is all hope lost and I lost over 600GB of data?

The worst part is that there are no errors in the logs and it just 
"disappeared" without a trace.
The only logs are from subsequent reboots where it says a ZFS pool failed to 
open.

It does not give me a warm & fuzzy about using ZFS as I've relied on it heavily 
in the past 5 years.

Any advice would be well appreciated...

On 7/2/2010 3:01 AM, Brian Kolaci wrote:

I've recently acquired some storage and have been trying to copy data from a 
remote data center to hold backup data.  The copies had been going for weeks, 
with about 600GB transferred so far, and then I noticed the throughput on the 
router stopped.  I see a pool disappeared.

# zpool status -x

  pool: pool4_green
 state: FAULTED
status: The pool metadata is corrupted and the pool cannot be opened.
action: Destroy and re-create the pool from a backup source.
   see: http://www.sun.com/msg/ZFS-8000-72
 scrub: none requested
config:

NAME STATE READ WRITE CKSUM
pool4_green  FAULTED  0 0 1  corrupted data
  raidz2 ONLINE   0 0 6
c10t0d0  ONLINE   0 0 0
c10t1d0  ONLINE   0 0 0
c10t2d0  ONLINE   0 0 0
c10t3d0  ONLINE   0 0 0
c10t4d0  ONLINE   0 0 0
c10t5d0  ONLINE   0 0 0
c10t6d0  ONLINE   0 0 0

I powered down the system and all the disk systems and powered up fresh.
I tried to clear the error:

#zpool clear pool4_green
internal error: Bad exchange descriptor
Abort (core dumped)

So then I took a look with zdb:

#zdb -vvv pool4_green
version=15
name='pool4_green'
state=0
txg=83
pool_guid=3115817837859301858
hostid=237914636
hostname='galaxy'
vdev_tree
type='root'
id=0
guid=3115817837859301858
children[0]
type='raidz'
id=0
guid=10261633106033684483
nparity=2
metaslab_array=24
metaslab_shift=36
ashift=9
asize=6997481881600
is_log=0
children[0]
type='disk'
id=0
guid=11313548069045029894
path='/dev/dsk/c10t0d0s0'
devid='id1,s...@n60026b9040e26100139d8514065a1d67/a'

phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@0,0:a'
whole_disk=1
children[1]
type='disk'
id=1
guid=5547727760941401848
path='/dev/dsk/c10t1d0s0'
devid='id1,s...@n60026b9040e26100139d851d06ec511d/a'

phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@1,0:a'
whole_disk=1
children[2]
type='disk'
id=2
guid=8407102896612298450
path='/dev/dsk/c10t2d0s0'
devid='id1,s...@n60026b9040e26100139d85260770c1bf/a'

phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@2,0:a'
whole_disk=1
children[3]
type='disk'
id=3
guid=17509238716791782209
path='/dev/dsk/c10t3d0s0'
devid='id1,s...@n60026b9040e26100139d852f07fea314/a'

phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@3,0:a'
whole_disk=1
children[4]
type='disk'
id=4
guid=18419120996062075464
path='/dev/dsk/c10t4d0s0'
devid='id1,s...@n60026b9040e26100139d8537086c271f/a'

phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@4,0:a'
whole_disk=1
children[5]
type='disk'
id=5
guid=8308368067368943006
path='/dev/dsk/c10t5d0s0'
devid='id1,s...@n60026b9040e26100139d85440934f640/a'

phys_path='/p...@0,0/pci8086,3...@4/p

Re: [zfs-discuss] pool wide corruption, "Bad exchange descriptor"

2010-07-06 Thread Victor Latushkin

On Jul 6, 2010, at 6:30 PM, Brian Kolaci wrote:

> Well, I see no takers or even a hint...
> 
> I've been playing with zdb to try to examine the pool, but I get:
> 
> # zdb -b pool4_green
> zdb: can't open pool4_green: Bad exchange descriptor
> 
> # zdb -d pool4_green
> zdb: can't open pool4_green: Bad exchange descriptor
> 
> So I'm not sure how to debug using zdb.  Is there something better or 
> something else I should be looking at?
> The disks are all there, all online.  How can I at least rollback to the last 
> consistent bit of data on there?
> Or is all hope lost and I lost over 600GB of data?
> 
> The worst part is that there are no errors in the logs and it just 
> "disappeared" without a trace.
> The only logs are from subsequent reboots where it says a ZFS pool failed to 
> open.
> 
> It does not give me a warm & fuzzy about using ZFS as I've relied on it 
> heavily in the past 5 years.
> 
> Any advice would be well appreciated...

You can download build 134 LiveCD boot off it and try 'zpool import -nfF 
pool4_green' for a start.

regards
victor


> On 7/2/2010 3:01 AM, Brian Kolaci wrote:
>> I've recently acquired some storage and have been trying to copy data from a 
>> remote data center to hold backup data.  The copies had been going for 
>> weeks, with about 600GB transferred so far, and then I noticed the 
>> throughput on the router stopped.  I see a pool disappeared.
>> 
>> # zpool status -x
>> 
>>  pool: pool4_green
>> state: FAULTED
>> status: The pool metadata is corrupted and the pool cannot be opened.
>> action: Destroy and re-create the pool from a backup source.
>>   see: http://www.sun.com/msg/ZFS-8000-72
>> scrub: none requested
>> config:
>> 
>>NAME STATE READ WRITE CKSUM
>>pool4_green  FAULTED  0 0 1  corrupted data
>>  raidz2 ONLINE   0 0 6
>>c10t0d0  ONLINE   0 0 0
>>c10t1d0  ONLINE   0 0 0
>>c10t2d0  ONLINE   0 0 0
>>c10t3d0  ONLINE   0 0 0
>>c10t4d0  ONLINE   0 0 0
>>c10t5d0  ONLINE   0 0 0
>>c10t6d0  ONLINE   0 0 0
>> 
>> I powered down the system and all the disk systems and powered up fresh.
>> I tried to clear the error:
>> 
>> #zpool clear pool4_green
>> internal error: Bad exchange descriptor
>> Abort (core dumped)
>> 
>> So then I took a look with zdb:
>> 
>> #zdb -vvv pool4_green
>>version=15
>>name='pool4_green'
>>state=0
>>txg=83
>>pool_guid=3115817837859301858
>>hostid=237914636
>>hostname='galaxy'
>>vdev_tree
>>type='root'
>>id=0
>>guid=3115817837859301858
>>children[0]
>>type='raidz'
>>id=0
>>guid=10261633106033684483
>>nparity=2
>>metaslab_array=24
>>metaslab_shift=36
>>ashift=9
>>asize=6997481881600
>>is_log=0
>>children[0]
>>type='disk'
>>id=0
>>guid=11313548069045029894
>>path='/dev/dsk/c10t0d0s0'
>>devid='id1,s...@n60026b9040e26100139d8514065a1d67/a'
>>
>> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@0,0:a'
>>whole_disk=1
>>children[1]
>>type='disk'
>>id=1
>>guid=5547727760941401848
>>path='/dev/dsk/c10t1d0s0'
>>devid='id1,s...@n60026b9040e26100139d851d06ec511d/a'
>>
>> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@1,0:a'
>>whole_disk=1
>>children[2]
>>type='disk'
>>id=2
>>guid=8407102896612298450
>>path='/dev/dsk/c10t2d0s0'
>>devid='id1,s...@n60026b9040e26100139d85260770c1bf/a'
>>
>> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@2,0:a'
>>whole_disk=1
>>children[3]
>>type='disk'
>>id=3
>>guid=17509238716791782209
>>path='/dev/dsk/c10t3d0s0'
>>devid='id1,s...@n60026b9040e26100139d852f07fea314/a'
>>
>> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@3,0:a'
>>whole_disk=1
>>children[4]
>>type='disk'
>>id=4
>>guid=18419120996062075464
>>path='/dev/dsk/c10t4d0s0'
>>devid='id1,s...@n60026b9040e26100139d8537086c271f/a'
>> 

Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)

2010-07-06 Thread Victor Latushkin

On Jul 4, 2010, at 4:58 AM, Andrew Jones wrote:

> Victor,
> 
> The zpool import succeeded on the next attempt following the crash that I 
> reported to you by private e-mail! 

>From the threadlist it looked like system was pretty low on memory with stacks 
>of userland stuff swapped out, hence system was not responsive, but it was 
>able to complete inconsistent dataset processing in the end.

> 
> For completeness, this is the final status of the pool:
> 
> 
>  pool: tank
> state: ONLINE
> scan: resilvered 1.50K in 165h28m with 0 errors on Sat Jul  3 08:02:30 2010
> config:
> 
>NAMESTATE READ WRITE CKSUM
>tankONLINE   0 0 0
>  raidz2-0  ONLINE   0 0 0
>c0t0d0  ONLINE   0 0 0
>c0t1d0  ONLINE   0 0 0
>c0t2d0  ONLINE   0 0 0
>c0t3d0  ONLINE   0 0 0
>c0t4d0  ONLINE   0 0 0
>c0t5d0  ONLINE   0 0 0
>c0t6d0  ONLINE   0 0 0
>c0t7d0  ONLINE   0 0 0
>cache
>  c2t0d0ONLINE   0 0 0
> 
> errors: No known data errors
> 

Good. Run 'zpool scrub' to make sure there are no other errors.

regards
victor

> Thank you very much for your help. We did not need to add additional RAM to 
> solve this, in the end. Instead, we needed to persist with the import through 
> several panics to finally work our way through the large inconsistent 
> dataset; it is unclear whether the resilvering caused additional processing 
> delay. Unfortunately, the delay made much of the data quite stale, now that 
> it's been recovered.
> 
> It does seem that zfs would benefit tremendously from a better (quicker and 
> more intuitive?) set of recovery tools, that are available to a wider range 
> of users. It's really a shame, because the features and functionality in zfs 
> are otherwise absolutely second to none.
> 
> /Andrew[i][/i][i][/i][i][/i][i][/i][i][/i]
> -- 
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel Panic on zpool clean

2010-07-06 Thread Victor Latushkin

On Jul 3, 2010, at 1:20 PM, George wrote:

>> Because of that I'm thinking that I should try
>> to change the hostid when booted from the CD to be
>> the same as the previously installed system to see if
>> that helps - unless that's likely to confuse it at
>> all...?
> 
> I've now tried changing the hostid using the code from 
> http://forums.sun.com/thread.jspa?threadID=5075254 NB: you need to leave this 
> running in a separate terminal.
> 
> This changes the start of "zpool import" to
> 
>  pool: storage2
>id: 14701046672203578408
> state: FAULTED
> status: The pool metadata is corrupted.
> action: The pool cannot be imported due to damaged devices or data.
>The pool may be active on another system, but can be imported using
>the '-f' flag.
>   see: http://www.sun.com/msg/ZFS-8000-72
> 
> 
> but otherwise nothing is changed with respect to trying to import or clear 
> the pool. The pool is 8TB and the machine has 4GB but as far as I can see via 
> top the commands aren't failing due to a lack of memory.
> 
> I'm a bit stumped now. The only thing else I can think to try is inserting 
> c9t4d4 (the new drive) and removing c6t4d0 (which should be fine). The 
> problem with this though is that it relies on c7t4d0 (which is faulty) and so 
> it assumes that the errors can be cleared, the replace stopped and the drives 
> swapped back before further errors happen.

I think it is quite likely to be possible to get readonly access to your data, 
but this requires modified ZFS binaries. What is your pool version? What build 
do you have installed on your system disk or available as LiveCD?

regards
victor

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS recovery tools

2010-07-06 Thread Victor Latushkin

On Jul 4, 2010, at 1:33 AM, R. Eulenberg wrote:

> R. Eulenberg  web.de> writes:
> 
>> 
   op> I was setting up a new systen (osol 2009.06
>>> and updating to
   op> the lastest version of osol/dev - snv_134 -
>>> with
   op> deduplication) and then I tried to import my
>>> backup zpool, but
   op> it does not work.  
 
   op> # zpool import -f tank1 
   op> cannot import 'tank1': one or more devices
>>> is currently unavailable 
   op> Destroy and re-create the pool from a backup
>>> source
 
   op> Any other option (-F, -X, -V, -D) and any
>>> combination of them
   op> doesn't helps too.
>> 
 R., please let us know if the 'zdb -e -bcsvL
>>> ' incantation
 Sigbjorn suggested ends up working for you or not.
>> 
>> Hi,
>> 
>> it seems that you are talking about my case of trouble. 
>> I answered in this thread:
>> 
>> http://www.opensolaris.org/jive/thread.jspa?messageID=485342&tstart=0#485342
>> 
>> I hope of any ideas helping me.
>> 
>> Regards
>> Ron
> 
> 
> Hi,
> 
> several days ago I added this:
> 
> set /zfs/:zfs_recover=/1/
> set aok=/1/
> 
> to the /etc/system file and ran:
> 
> zdb -e -bcsvL tank1
> 

settings in the /etc/systm file do not affect zdb in any way. In recent builds 
(135+) there's an option -A for zdb that allows to simulate setting one or both 
of these parameters, but anyway it is not useful in your case.

> without an output and without a prompt (prozess hangs up)

How did you arrive at this conclusion? Did you check "pstack `pgrep zdb`/1" 
output a few times?

> and result of running:
> 
> zdb -eC tank1
> 
> was the same.
> I hope you could help me because the problem isn't solved.
> 
> Regards
> Ron 
> 
> 
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel Panic on zpool clean

2010-07-06 Thread Roy Sigurd Karlsbakk
> I think it is quite likely to be possible to get readonly access to
> your data, but this requires modified ZFS binaries. What is your pool
> version? What build do you have installed on your system disk or
> available as LiveCD?

Sorry, but does this mean if ZFS can't write to the drives, access to the pool 
won't be possible? If so, that's rather scary...

Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel Panic on zpool clean

2010-07-06 Thread Victor Latushkin

On Jun 28, 2010, at 11:27 PM, George wrote:

> Again this core dumps when I try to do "zpool clear storage2"
> 
> Does anyone have any suggestions what would be the best course of action now?

Do you have any crahsdumps saved? First one is most interesting one...
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Hashing files rapidly on ZFS

2010-07-06 Thread Arne Jansen
Daniel Carosone wrote:
> Something similar would be useful, and much more readily achievable,
> from ZFS from such an application, and many others.  Rather than a way
> to compare reliably between two files for identity, I'ld liek a way to
> compare identity of a single file between two points in time.  If my
> application can tell quickly that the file content is unaltered since
> last time I saw the file, I can avoid rehashing the content and use a
> stored value. If I can achieve this result for a whole directory
> tree, even better.

This would be great for any kind of archiving software. Aren't zfs checksums
already ready to solve this? If a file changes, it's dnodes' checksum changes,
the checksum of the directory it is in and so forth all the way up to the
uberblock.
There may be ways a checksum changes without a real change in the files content,
but the other way round should hold. If the checksum didn't change, the file
didn't change.
So the only missing link is a way to determine zfs's checksum for a
file/directory/dataset. Am I missing something here? Of course atime update
should be turned off, otherwise the checksum will get changed by the archiving
agent.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Remove non-redundant disk

2010-07-06 Thread Cassandra Pugh
Hello list,

This has probably been discussed, however I would like to bring it up again,
so that the powers that be, know someone else is looking for this feature.

I would like to be able to shrink a pool and remove a non-redundant disk.

Is this something that is in the works?

It would be fantastic if I had this capability.

Thanks!


-
Cassandra
Unix Administrator


"From a little spark may burst a mighty flame."
-Dante Alighieri
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Remove non-redundant disk

2010-07-06 Thread Roy Sigurd Karlsbakk
- Original Message -
> Hello list,
> 
> This has probably been discussed, however I would like to bring it up
> again, so that the powers that be, know someone else is looking for
> this feature.
> 
> I would like to be able to shrink a pool and remove a non-redundant
> disk.
> 
> Is this something that is in the works?
> 
> It would be fantastic if I had this capability.

You're a little unclear on what you want, but it seems to me you want to change 
a raidz2 to a raidz1 or something like that. That is, AFAIK, in the works, with 
the block rewrite functionality. As with most other parts of progress in 
OpenSolaris, nothing is clear about when or if this will get integrated into 
the system.

For now, you can't change a raidz(n) VDEV and you can't detach a VDEV from a 
pool. The only way is to build a new pool and move the data with things like 
zfs send/receive. You can also remove a drive from a raidz(n), and reducing its 
redundancy, but you can't change the value of n.

Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Remove non-redundant disk

2010-07-06 Thread Cassandra Pugh
The pool is not redundant, so I would suppose, yes, is is Raid-1 on the
software level.

I have a few drives, which are on a specific array, which I would like to
remove from this pool.

I have discovered the "replace" command, and I am going to try and replace,
1 for 1, the drives I would like to remove.

However, it would be nice  if there were a way to simply "remove" the disks,
if space allowed.


-
Cassandra
(609) 243-2413
Unix Administrator


"From a little spark may burst a mighty flame."
-Dante Alighieri


On Tue, Jul 6, 2010 at 11:55 AM, Roy Sigurd Karlsbakk wrote:

> - Original Message -
> > Hello list,
> >
> > This has probably been discussed, however I would like to bring it up
> > again, so that the powers that be, know someone else is looking for
> > this feature.
> >
> > I would like to be able to shrink a pool and remove a non-redundant
> > disk.
> >
> > Is this something that is in the works?
> >
> > It would be fantastic if I had this capability.
>
> You're a little unclear on what you want, but it seems to me you want to
> change a raidz2 to a raidz1 or something like that. That is, AFAIK, in the
> works, with the block rewrite functionality. As with most other parts of
> progress in OpenSolaris, nothing is clear about when or if this will get
> integrated into the system.
>
> For now, you can't change a raidz(n) VDEV and you can't detach a VDEV from
> a pool. The only way is to build a new pool and move the data with things
> like zfs send/receive. You can also remove a drive from a raidz(n), and
> reducing its redundancy, but you can't change the value of n.
>
> Vennlige hilsener / Best regards
>
> roy
> --
> Roy Sigurd Karlsbakk
> (+47) 97542685
> r...@karlsbakk.net
> http://blogg.karlsbakk.net/
> --
> I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det
> er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av
> idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate
> og relevante synonymer på norsk.
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Consequences of resilvering failure

2010-07-06 Thread Michael Johnson
I'm just about to start using ZFS in a RAIDZ configuration for a home file 
server (mostly holding backups), and I wasn't clear on what happens if data 
corruption is detected while resilvering.  For example: let's say I'm using 
RAIDZ1 and a drive fails.  I pull it and put in a new one.  While resilvering, 
ZFS detects corrupt data on one of the remaining disks.  Will the resilvering 
continue, with some files marked as containing errors, or will it simply fail?

(I found this process[1] to repair damaged data, but I wasn't sure what would 
happen if it was detected in the middle of resilvering.)

I will of course have a backup of the pool, but I may opt for additional backup 
if the entire pool could be lost due to data corruption (as opposed to just a 
few files potentially being lost).

Thanks,
Michael

[1] http://dlc.sun.com/osol/docs/content/ZFSADMIN/gbbwl.html


  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Legacy MountPoint for /rpool/ROOT

2010-07-06 Thread Ketan
I have two different servers with ZFS root but both of them has different 
mountpoint for rpool/ROOT  one is /rpool/ROOT and other is legacy. Whats 
the difference between the two and which is the one we should keep. 

And why there is 3 different zfs datasets  rpool, rpool/ROOT and 
rpool/ROOT/zfsBE ? 


# zfs list -r rpool
NAME USED  AVAIL  REFER  MOUNTPOINT
rpool  24.1G   110G98K/rpool
rpool/ROOT6.08G   110G21K/rpool/ROOT
rpool/ROOT/zfsBE   6.08G   110G  6.08G  /

**

zfs list -r rpool
NAMEUSED  AVAIL  REFER  MOUNTPOINT
rpool 34.0G  99.9G94K  /rpool
rpool/ROOT   27.9G  99.9G18K  legacy
rpool/ROOT/s10s_u6  27.9G  99.9G  27.9G  /
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Legacy MountPoint for /rpool/ROOT

2010-07-06 Thread Ketan
I have two different servers with ZFS root but both of them has different 
mountpoint for rpool/ROOT  one is /rpool/ROOT and other is legacy. Whats 
the difference between the two and which is the one we should keep. 

And why there is 3 different zfs datasets  rpool, rpool/ROOT and 
rpool/ROOT/zfsBE ? 

Can anyone explain this to me . Thanx in advance

# zfs list -r rpool
NAME USED  AVAIL  REFER  MOUNTPOINT
rpool  24.1G   110G98K/rpool
rpool/ROOT6.08G   110G21K/rpool/ROOT
rpool/ROOT/zfsBE   6.08G   110G  6.08G  /

**

zfs list -r rpool
NAMEUSED  AVAIL  REFER  MOUNTPOINT
rpool 34.0G  99.9G94K  /rpool
rpool/ROOT   27.9G  99.9G18K  legacy
rpool/ROOT/s10s_u6  27.9G  99.9G  27.9G  /
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Help with Faulted Zpool Call for Help(Cross post)

2010-07-06 Thread Sam Fourman Jr.
Hello list,

I posted this a few days ago on opensolaris-discuss@ list
I am posting here, because there my be too much noise on other lists

I have been without this zfs set for a week now.
My main concern at this point,is it even possible to recover this zpool.

How does the metadata work? what tool could is use to rebuild the
corrupted parts
or even find out what parts are corrupted.


most but not all of these disks were Hitachi Retail 1TB didks.


I have a Fileserver that runs FreeBSD 8.1 (zfs v14)
after a poweroutage, I am unable to import my zpool named Network
my pool is made up of 6 1TB disks configured in raidz.
there is ~1.9TB of actual data on this pool.

I have loaded Open Solaris svn_134 on a seprate boot disk,
in hopes of recovering my zpool.

on Open Solaris 134, I am not able to import my zpool
almost everything I try gives me cannot import 'Network': I/O error

I have done quite a bit of searching, and I found that import -fFX
Network should work
however after ~ 20 hours this hard locks Open Solaris (however it does
return a ping)

here is a list of commands that I have run on Open Solaris

http://www.puffybsd.com/zfsv14.txt

if anyone could help me use zdb or mdb to recover my pool
I would very much appreciate it.

I believe the metadata is corrupt on my zpool


-- 

Sam Fourman Jr.
Fourman Networks
http://www.fourmannetworks.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Remove non-redundant disk

2010-07-06 Thread Roy Sigurd Karlsbakk
- Original Message -
> The pool is not redundant, so I would suppose, yes, is is Raid-1 on
> the software level.
> 
> I have a few drives, which are on a specific array, which I would like
> to remove from this pool.
> 
> I have discovered the "replace" command, and I am going to try and
> replace, 1 for 1, the drives I would like to remove.
> 
> However, it would be nice if there were a way to simply "remove" the
> disks, if space allowed.

zfs attach new drive, zfs detach old drive, that will replace the drive without 
much hazzle
 
Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Legacy MountPoint for /rpool/ROOT

2010-07-06 Thread Lori Alt

On 07/ 6/10 10:56 AM, Ketan wrote:

I have two different servers with ZFS root but both of them has different 
mountpoint for rpool/ROOT  one is /rpool/ROOT and other is legacy.


It should be legacy.


  Whats
the difference between the two and which is the one we should keep.

And why there is 3 different zfs datasets  rpool, rpool/ROOT and 
rpool/ROOT/zfsBE ?
   


rpool is the top-level dataset of the pool (every pool has one).

rpool/ROOT is just a grouping mechanism for the roots of all BEs on the 
system.


rpool/ROOT/zfsBE is the root file system of the BE called "zfsBE".

lori



# zfs list -r rpool
NAME USED  AVAIL  REFER  MOUNTPOINT
rpool  24.1G   110G98K/rpool
rpool/ROOT6.08G   110G21K/rpool/ROOT
rpool/ROOT/zfsBE   6.08G   110G  6.08G  /

**

zfs list -r rpool
NAMEUSED  AVAIL  REFER  MOUNTPOINT
rpool 34.0G  99.9G94K  /rpool
rpool/ROOT   27.9G  99.9G18K  legacy
rpool/ROOT/s10s_u6  27.9G  99.9G  27.9G  /
   


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Consequences of resilvering failure

2010-07-06 Thread Roy Sigurd Karlsbakk
- Original Message -
> I'm just about to start using ZFS in a RAIDZ configuration for a home
> file server (mostly holding backups), and I wasn't clear on what
> happens if data corruption is detected while resilvering. For example:
> let's say I'm using RAIDZ1 and a drive fails. I pull it and put in a
> new one. While resilvering, ZFS detects corrupt data on one of the
> remaining disks. Will the resilvering continue, with some files marked
> as containing errors, or will it simply fail?

The pool will remain available, but you will have data corruption. The simple 
way to avoid this, is to use a raidz2, where the chances are far lower for data 
corruption.

Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Remove non-redundant disk

2010-07-06 Thread Cassandra Pugh
I tried zfs replace, however the new drive is slightly smaller, and even
with a -f, it refuses to replace the drive.

I guess i will have to export the pool and destroy this one to get my drives
back.

Still would like the ability to shrink a pool.
-
Cassandra
(609) 243-2413
Unix Administrator


"From a little spark may burst a mighty flame."
-Dante Alighieri


On Tue, Jul 6, 2010 at 1:02 PM, Roy Sigurd Karlsbakk wrote:

> - Original Message -
> > The pool is not redundant, so I would suppose, yes, is is Raid-1 on
> > the software level.
> >
> > I have a few drives, which are on a specific array, which I would like
> > to remove from this pool.
> >
> > I have discovered the "replace" command, and I am going to try and
> > replace, 1 for 1, the drives I would like to remove.
> >
> > However, it would be nice if there were a way to simply "remove" the
> > disks, if space allowed.
>
> zfs attach new drive, zfs detach old drive, that will replace the drive
> without much hazzle
>
> Vennlige hilsener / Best regards
>
> roy
> --
> Roy Sigurd Karlsbakk
> (+47) 97542685
> r...@karlsbakk.net
> http://blogg.karlsbakk.net/
> --
> I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det
> er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av
> idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate
> og relevante synonymer på norsk.
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)

2010-07-06 Thread Andrew Jones
> 
> Good. Run 'zpool scrub' to make sure there are no
> other errors.
> 
> regards
> victor
> 

Yes, scrubbed successfully with no errors. Thanks again for all of your 
generous assistance.

/AJ
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS fsck?

2010-07-06 Thread Roy Sigurd Karlsbakk
Hi all

With several messages in here about troublesome zpools, would there be a good 
reason to be able to fsck a pool? As in, check the whole thing instead of 
having to boot into live CDs and whatnot?
 
Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS fsck?

2010-07-06 Thread iMx
- Original Message -
> From: "Roy Sigurd Karlsbakk" 
> To: "OpenSolaris ZFS discuss" 
> Sent: Tuesday, 6 July, 2010 6:35:51 PM
> Subject: [zfs-discuss] ZFS fsck?
> Hi all
> 
> With several messages in here about troublesome zpools, would there be
> a good reason to be able to fsck a pool? As in, check the whole thing
> instead of having to boot into live CDs and whatnot?
> 
> Vennlige hilsener / Best regards
> 
> roy

Scrub? :)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS fsck?

2010-07-06 Thread Mark J Musante

On Tue, 6 Jul 2010, Roy Sigurd Karlsbakk wrote:


Hi all

With several messages in here about troublesome zpools, would there be a 
good reason to be able to fsck a pool? As in, check the whole thing 
instead of having to boot into live CDs and whatnot?


You can do this with "zpool scrub".  It visits every allocated block and 
verifies that everything is correct.  It's not the same as fsck in that 
scrub can detect and repair problems with the pool still online and all 
datasets mounted, whereas fsck cannot handle mounted filesystems.


If you really want to use it on an exported pool, you can use zdb, 
although it might take some time.  Here's an example on a small empty 
pool:


# zpool create -f mypool raidz c4t1d0s0 c4t2d0s0 c4t3d0s0 c4t4d0s0 c4t5d0s0
# zpool list mypool
NAME SIZE  ALLOC   FREECAP  DEDUP  HEALTH  ALTROOT
mypool   484M   280K   484M 0%  1.00x  ONLINE  -
# zpool export mypool
# zdb -ebcc mypool

Traversing all blocks to verify checksums and verify nothing leaked ...

No leaks (block sum matches space maps exactly)

bp count:  48
bp logical:378368  avg:   7882
bp physical:39424  avg:821 compression:   9.60
bp allocated:  185344  avg:   3861 compression:   2.04
bp deduped: 0ref>1:  0   deduplication:   1.00
SPA allocated: 185344 used:  0.04%

#
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS fsck?

2010-07-06 Thread Roy Sigurd Karlsbakk
> You can do this with "zpool scrub". It visits every allocated block
> and
> verifies that everything is correct. It's not the same as fsck in that
> scrub can detect and repair problems with the pool still online and
> all
> datasets mounted, whereas fsck cannot handle mounted filesystems.
> 
> If you really want to use it on an exported pool, you can use zdb,
> although it might take some time. Here's an example on a small empty
> pool:
> 
> # zpool create -f mypool raidz c4t1d0s0 c4t2d0s0 c4t3d0s0 c4t4d0s0
> c4t5d0s0
> # zpool list mypool
> NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT
> mypool 484M 280K 484M 0% 1.00x ONLINE -
> # zpool export mypool
> # zdb -ebcc mypool
...

what I'm saying is that there are several posts in here where the only solution 
is to boot onto a live cd and then do an import, due to metadata corruption. 
This should be doable from the installed system

Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS fsck?

2010-07-06 Thread Mark J Musante

On Tue, 6 Jul 2010, Roy Sigurd Karlsbakk wrote:

what I'm saying is that there are several posts in here where the only 
solution is to boot onto a live cd and then do an import, due to 
metadata corruption. This should be doable from the installed system


Ah, I understand now.

A couple of things worth noting:

- if the root filesystem in a boot pool cannot be mounted, it's 
problematic to access the tools necessary to repair it.  So going to a 
livecd (or a network boot for that matter) is the best way forward.


- if the tools available to failsafe are insufficient to repair a pool, 
then booting off a livecd/network is the only way forward.


It is also worth pointing out here that the 134a build has the pool 
recovery code built-in.  The "-F" option to zpool import only became 
available after build 128 or 129.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NexentaStor 3.0.3 vs OpenSolaris - Patches more up to date?

2010-07-06 Thread Spandana Goli
Release Notes information:

If there are new features, each release is added to
http://www.nexenta.com/corp/documentation/release-notes-support.

If just bug fixes, then the Changelog listing is updated:
http://www.nexenta.com/corp/documentation/nexentastor-changelog

Regards,
Spandana
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Help with Faulted Zpool Call for Help(Cross post)

2010-07-06 Thread Cindy Swearingen

Hi Sam,

In general, FreeBSD uses different device naming conventions and power
failures seem to clobber disk labeling. The "I/O error" message also
points to problems accessing these disks.

I'm not sure if this helps, but I see that the 6 disks from the zdb -e
output are indicated as c7t0d0p0 --> c7t5d0p0, and the paths resemble
/dev/dsk/c7t1d0p0, which makes sense.

When you look at the individual labels using zdb -l /dev/dsk/c7t1d0p0,
the physical path for this disk looks identical, but the path=/dev/da4
and so on. I'm not familiar with the /dev/da* device naming.

When I use zdb -l on my pool's disks, I see the same phys_path and
the same path= for each disk.

I hoping someone else who has patched their disks back together after
a power failure can comment.

Thanks,

Cindy

On 07/06/10 11:02, Sam Fourman Jr. wrote:

Hello list,

I posted this a few days ago on opensolaris-discuss@ list
I am posting here, because there my be too much noise on other lists

I have been without this zfs set for a week now.
My main concern at this point,is it even possible to recover this zpool.

How does the metadata work? what tool could is use to rebuild the
corrupted parts
or even find out what parts are corrupted.


most but not all of these disks were Hitachi Retail 1TB didks.


I have a Fileserver that runs FreeBSD 8.1 (zfs v14)
after a poweroutage, I am unable to import my zpool named Network
my pool is made up of 6 1TB disks configured in raidz.
there is ~1.9TB of actual data on this pool.

I have loaded Open Solaris svn_134 on a seprate boot disk,
in hopes of recovering my zpool.

on Open Solaris 134, I am not able to import my zpool
almost everything I try gives me cannot import 'Network': I/O error

I have done quite a bit of searching, and I found that import -fFX
Network should work
however after ~ 20 hours this hard locks Open Solaris (however it does
return a ping)

here is a list of commands that I have run on Open Solaris

http://www.puffybsd.com/zfsv14.txt

if anyone could help me use zdb or mdb to recover my pool
I would very much appreciate it.

I believe the metadata is corrupt on my zpool



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NexentaStor 3.0.3 vs OpenSolaris - Patches more up to date?

2010-07-06 Thread Giovanni Tirloni
On Tue, Jul 6, 2010 at 4:06 PM, Spandana Goli  wrote:
> Release Notes information:
> If there are new features, each release is added to
> http://www.nexenta.com/corp/documentation/release-notes-support.
>
> If just bug fixes, then the Changelog listing is updated:
> http://www.nexenta.com/corp/documentation/nexentastor-changelog

Is there a bug tracker were one can objectively list all the bugs
(with details) that went into a release ?

"Many bug fixes" is a bit too general.

-- 
Giovanni Tirloni
gtirl...@sysdroid.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool throughput: snv 134 vs 138 vs 143

2010-07-06 Thread Chad Cantwell
Hi all,

I've noticed something strange in the throughput in my zpool between
different snv builds, and I'm not sure if it's an inherent difference
in the build or a kernel parameter that is different in the builds.
I've setup two similiar machines and this happens with both of them.
Each system has 16 2TB Samsung HD203WI drives (total) directly connected
to two LSI 3081E-R 1068e cards with IT firmware in one raidz3 vdev.

In both computers, after a fresh installation of snv 134, the throughput
is a maximum of about 300 MB/s during scrub or something like
"dd if=/dev/zero bs=1024k of=bigfile".

If I bfu to snv 138, I then get throughput of about 700 MB/s with both
scrub or a single thread dd.

I assumed at first this was some sort of bug or regression in 134 that
made it slow.  However, I've now tested also from the fresh 134
installation, compiling the OS/Net build 143 from the mercurial
repository and booting into it, after which the dd throughput is still
only about 300 MB/s just like snv 134.  The scrub throughput in 143
is even slower, rarely surpassing 150 MB/s.  I wonder if the scrubbing
being extra slow here is related to the additional statistics displayed
during the scrub that didn't used to be shown.

Is there some kind of debug option that might be enabled in the 134 build
and persist if I compile snv 143 which would be off if I installed a 138
through bfu?  If not, it makes me think that the bfu to 138 is changing
the configuration somewhere to make it faster rather than fixing a bug or
being a debug flag on or off.  Does anyone have any idea what might be
happening?  One thing I haven't tried is bfu'ing to 138, and from this
faster working snv 138 installing the snv 143 build, which may possibly
create a 143 that performs faster if it's simply a configuration parameter.
I'm not sure offhand if installing source-compiled ON builds from a bfu'd
rpool is supported, although I suppose it's simple enough to try.

Thanks,
Chad Cantwell
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] pool wide corruption, "Bad exchange descriptor"

2010-07-06 Thread Richard Elling
On Jul 6, 2010, at 7:30 AM, Brian Kolaci wrote:

> Well, I see no takers or even a hint...
> 
> I've been playing with zdb to try to examine the pool, but I get:
> 
> # zdb -b pool4_green
> zdb: can't open pool4_green: Bad exchange descriptor

For the archives, EBADE "Bad exchange descriptor" was repurposed
as ECKSUM in zio.h. This agrees with the zpool status below. I 
recommend Victor's advice.
 -- richard

> 
> # zdb -d pool4_green
> zdb: can't open pool4_green: Bad exchange descriptor
> 
> So I'm not sure how to debug using zdb.  Is there something better or 
> something else I should be looking at?
> The disks are all there, all online.  How can I at least rollback to the last 
> consistent bit of data on there?
> Or is all hope lost and I lost over 600GB of data?
> 
> The worst part is that there are no errors in the logs and it just 
> "disappeared" without a trace.
> The only logs are from subsequent reboots where it says a ZFS pool failed to 
> open.
> 
> It does not give me a warm & fuzzy about using ZFS as I've relied on it 
> heavily in the past 5 years.
> 
> Any advice would be well appreciated...
> 
> On 7/2/2010 3:01 AM, Brian Kolaci wrote:
>> I've recently acquired some storage and have been trying to copy data from a 
>> remote data center to hold backup data.  The copies had been going for 
>> weeks, with about 600GB transferred so far, and then I noticed the 
>> throughput on the router stopped.  I see a pool disappeared.
>> 
>> # zpool status -x
>> 
>>  pool: pool4_green
>> state: FAULTED
>> status: The pool metadata is corrupted and the pool cannot be opened.
>> action: Destroy and re-create the pool from a backup source.
>>   see: http://www.sun.com/msg/ZFS-8000-72
>> scrub: none requested
>> config:
>> 
>>NAME STATE READ WRITE CKSUM
>>pool4_green  FAULTED  0 0 1  corrupted data
>>  raidz2 ONLINE   0 0 6
>>c10t0d0  ONLINE   0 0 0
>>c10t1d0  ONLINE   0 0 0
>>c10t2d0  ONLINE   0 0 0
>>c10t3d0  ONLINE   0 0 0
>>c10t4d0  ONLINE   0 0 0
>>c10t5d0  ONLINE   0 0 0
>>c10t6d0  ONLINE   0 0 0
>> 
>> I powered down the system and all the disk systems and powered up fresh.
>> I tried to clear the error:
>> 
>> #zpool clear pool4_green
>> internal error: Bad exchange descriptor
>> Abort (core dumped)
>> 
>> So then I took a look with zdb:
>> 
>> #zdb -vvv pool4_green
>>version=15
>>name='pool4_green'
>>state=0
>>txg=83
>>pool_guid=3115817837859301858
>>hostid=237914636
>>hostname='galaxy'
>>vdev_tree
>>type='root'
>>id=0
>>guid=3115817837859301858
>>children[0]
>>type='raidz'
>>id=0
>>guid=10261633106033684483
>>nparity=2
>>metaslab_array=24
>>metaslab_shift=36
>>ashift=9
>>asize=6997481881600
>>is_log=0
>>children[0]
>>type='disk'
>>id=0
>>guid=11313548069045029894
>>path='/dev/dsk/c10t0d0s0'
>>devid='id1,s...@n60026b9040e26100139d8514065a1d67/a'
>>
>> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@0,0:a'
>>whole_disk=1
>>children[1]
>>type='disk'
>>id=1
>>guid=5547727760941401848
>>path='/dev/dsk/c10t1d0s0'
>>devid='id1,s...@n60026b9040e26100139d851d06ec511d/a'
>>
>> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@1,0:a'
>>whole_disk=1
>>children[2]
>>type='disk'
>>id=2
>>guid=8407102896612298450
>>path='/dev/dsk/c10t2d0s0'
>>devid='id1,s...@n60026b9040e26100139d85260770c1bf/a'
>>
>> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@2,0:a'
>>whole_disk=1
>>children[3]
>>type='disk'
>>id=3
>>guid=17509238716791782209
>>path='/dev/dsk/c10t3d0s0'
>>devid='id1,s...@n60026b9040e26100139d852f07fea314/a'
>>
>> phys_path='/p...@0,0/pci8086,3...@4/pci1028,1...@0/s...@3,0:a'
>>whole_disk=1
>>children[4]
>>type='disk'
>>id=4
>>guid=18419120996062075464
>>path='/dev/dsk/c10t4d0s0'
>>devid='id1,s...

Re: [zfs-discuss] Help with Faulted Zpool Call for Help(Cross post)

2010-07-06 Thread Richard Elling

On Jul 6, 2010, at 10:02 AM, Sam Fourman Jr. wrote:

> Hello list,
> 
> I posted this a few days ago on opensolaris-discuss@ list
> I am posting here, because there my be too much noise on other lists
> 
> I have been without this zfs set for a week now.
> My main concern at this point,is it even possible to recover this zpool.
> 
> How does the metadata work? what tool could is use to rebuild the
> corrupted parts
> or even find out what parts are corrupted.
> 
> 
> most but not all of these disks were Hitachi Retail 1TB didks.
> 
> 
> I have a Fileserver that runs FreeBSD 8.1 (zfs v14)
> after a poweroutage, I am unable to import my zpool named Network
> my pool is made up of 6 1TB disks configured in raidz.
> there is ~1.9TB of actual data on this pool.
> 
> I have loaded Open Solaris svn_134 on a seprate boot disk,
> in hopes of recovering my zpool.
> 
> on Open Solaris 134, I am not able to import my zpool
> almost everything I try gives me cannot import 'Network': I/O error
> 
> I have done quite a bit of searching, and I found that import -fFX
> Network should work
> however after ~ 20 hours this hard locks Open Solaris (however it does
> return a ping)
> 
> here is a list of commands that I have run on Open Solaris
> 
> http://www.puffybsd.com/zfsv14.txt

You ran "zdb -l /dev/dsk/c7t5d0s2" which is not the same as
"zdb -l /dev/dsk/c7t5d0p0" because of the default partitioning.
In Solaris c*t*d*p* are fdisk partitions and c*t*d*s* are SMI or
EFI slices. This why label 2&3 could not be found and can be
part of the problem to start.

Everything in /dev/dsk and /dev/rdsk is a symlink or directory, 
so you can fake them out with a temporary directory and 
clever use of the "zpool import -d" command.  Examples are
in the archives.
 -- richard 

> 
> if anyone could help me use zdb or mdb to recover my pool
> I would very much appreciate it.
> 
> I believe the metadata is corrupt on my zpool
> 
> 
> -- 
> 
> Sam Fourman Jr.
> Fourman Networks
> http://www.fourmannetworks.com
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

-- 
Richard Elling
rich...@nexenta.com   +1-760-896-4422
ZFS and NexentaStor training, Rotterdam, July 13-15, 2010
http://nexenta-rotterdam.eventbrite.com/




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Lost ZIL Device

2010-07-06 Thread Andrew Kener
Hello All,

 

I've recently run into an issue I can't seem to resolve.  I have been
running a zpool populated with two RAID-Z1 VDEVs and a file on the
(separate) OS drive for the ZIL:

 

raidz1-0   ONLINE

c12t0d0  ONLINE

c12t1d0  ONLINE

c12t2d0  ONLINE

c12t3d0  ONLINE

raidz1-2   ONLINE

c12t4d0  ONLINE

c12t5d0  ONLINE

c13t0d0  ONLINE

c13t1d0  ONLINE

logs

/ZIL-Log.img

 

This was running on Nexenta Community Edition v3.  Everything was going
smoothly until today when the OS hard drive crashed and I was not able to
boot from it any longer.  I had migrated this setup from an OpenSolaris
install some months back and I still had the old drive intact.  I put it in
the system, booted it up and tried to import the zpool.  Unfortunately, I
have not been successful.  Previously when migrating from OSOL to Nexenta I
was able to get the new system to recognize and import the ZIL device file.
Since it has been lost in the drive crash I have not been able to duplicate
that success.

 

Here is the output from a 'zpool import' command:

 

pool: tank

id: 9013303135438223804

state: UNAVAIL

status: The pool was last accessed by another system.

action: The pool cannot be imported due to damaged devices or data.

   see: http://www.sun.com/msg/ZFS-8000-EY

config:

 

tank UNAVAIL  missing device

  raidz1-0   ONLINE

c12t0d0  ONLINE

c12t1d0  ONLINE

c12t5d0  ONLINE

c12t3d0  ONLINE

  raidz1-2   ONLINE

c12t4d0  ONLINE

c12t2d0  ONLINE

c13t0d0  ONLINE

c13t1d0  ONLINE

 

I created a new file for the ZIL (using mkfile) and tried to specify it for
inclusion with -d but it doesn't get recognized.  Probably because it was
never part of the original zpool.  I also symlinked the new ZIL file into
/dev/dsk but that didn't make any difference either.  

 

Any suggestions?

 

Andrew Kener

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Remove non-redundant disk

2010-07-06 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Cassandra Pugh
> 
> I would like to be able to shrink a pool and remove a non-redundant
> disk.
> 
> Is this something that is in the works?

I think the request is to remove vdev's from a pool.  Not currently
possible.  Is this in the works?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Lost ZIL Device

2010-07-06 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Andrew Kener
> 
> the OS hard drive crashed [and log device]

Here's what I know:  In zpool >= 19, if you import this, it will prompt you
to confirm the loss of the log device, and then it will import.

Here's what I have heard:  The ability to import with a failed log device as
described above, was created right around zpool 14 or 15, not quite sure
which.

Here's what I don't know:  If the failed zpool was some version which was
too low ... and you try to import on an OS which is capable of a much higher
version of zpool ... Can the newer OS handle it just because the newer OS is
able to handle a newer version of zpool?  Or maybe the version of the failed
pool is the one that matters, regardless of what the new OS is capable of
doing now?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Expected throughput

2010-07-06 Thread James Van Artsdalen
Under FreeBSD I've seen zpool scrub sustain nearly 500 MB/s in pools with large 
files (a pool with eight MIRROR vdevs on two Silicon Image 3124 controllers).

You need to carefully look for bottlenecks in the hardware.  You don't indicate 
how the disks are attached.  I would measure the total bandwidth of the disks 
outside of ZFS (ie, dd if=disk of=/dev/null bs=128k, one per disk on all disks 
simultaneously).
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss