Re: ZFS log device removal?

2010-10-20 Thread Nikolay Denev
On 19 Oct, 2010, at 07:24 , Adam Stylinski wrote:

> I hope I'm not asking anything too obvious, but I just saw this commit via
> the svn-src-all list:
> http://lists.freebsd.org/pipermail/svn-src-all/2010-October/030466.html
> 
> And I was wondering if this perhaps means that log device removal is now
> possible on 8-STABLE.  Or was this just a small code revision toward that
> feature?
> 
> Also I apologize in advance if my mailer does something weird and sends this
> message more than once, please be sure to tell me.
> 

Hello,

This only adds the ability to offline a log device, but not to remove it.
Also pools with missing log devices can't be imported. These features are added 
in ZFS v19.

Regards,
Nikolay___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


'zfs send -i': destination has been modified

2010-10-20 Thread Dan Langille
I am trying to do a 'zfs send -i' and failing.

This is my simple proof of concept test:

Create the data
# zfs create storage/a
# touch /storage/a/1
# touch /storage/a/2
# touch /storage/a/3

Snapshot
# zfs snapshot storage/a...@2010.10.19

send
# zfs send storage/a...@2010.10.19 | zfs receive -v storage/compressed/a
receiving full stream of storage/a...@2010.10.19 into
storage/compressed/a...@2010.10.19
received 252KB stream in 2 seconds (126KB/sec)
#



Create one more file and snapshot that

# touch /storage/a/4
# zfs snapshot storage/a...@2010.10.20


send it

# zfs send -i storage/a...@2010.10.19  storage/a...@2010.10.20 | zfs receive -v
storage/compressed/a
receiving incremental stream of storage/a...@2010.10.20 into
storage/compressed/a...@2010.10.20
received 250KB stream in 3 seconds (83.4KB/sec)

What do we have?

# find /storage/compressed/a
/storage/compressed/a
/storage/compressed/a/1
/storage/compressed/a/2
/storage/compressed/a/3
/storage/compressed/a/4


Of note:

* FreeBSD 8.1-STABLE
* ZFS filesystem version 4.
* ZFS pool version 15.
* zfs send is on compression off
* zfs receive has compression on

What I actually want to do and what fails:


# zfs snapshot storage/bac...@2010.10.19

# zfs send storage/bac...@2010.10.19 | zfs receive -v
storage/compressed/bacula
receiving full stream of storage/bac...@2010.10.19 into
storage/compressed/bac...@2010.10.19
received 4.38TB stream in 42490 seconds (108MB/sec)

# zfs snapshot storage/bac...@2010.10.20

# zfs send -i storage/bac...@2010.10.19 storage/bac...@2010.10.20 | zfs
receive -v storage/compressed/bacula
receiving incremental stream of storage/bac...@2010.10.20 into
storage/compressed/bac...@2010.10.20
cannot receive incremental stream: destination storage/compressed/bacula
has been modified
since most recent snapshot
warning: cannot send 'storage/bac...@2010.10.20': Broken pipe

I have no idea why this fails.  Clues please?

To my knowledge, the destination has not been written to.


-- 
Dan Langille -- http://langille.org/

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: 'zfs send -i': destination has been modified

2010-10-20 Thread Ruben de Groot
On Wed, Oct 20, 2010 at 08:32:33AM -0400, Dan Langille typed:
> I am trying to do a 'zfs send -i' and failing.
> 
> This is my simple proof of concept test:
> 
> Create the data
> # zfs create storage/a
> # touch /storage/a/1
> # touch /storage/a/2
> # touch /storage/a/3
> 
> Snapshot
> # zfs snapshot storage/a...@2010.10.19
> 
> send
> # zfs send storage/a...@2010.10.19 | zfs receive -v storage/compressed/a
> receiving full stream of storage/a...@2010.10.19 into
> storage/compressed/a...@2010.10.19
> received 252KB stream in 2 seconds (126KB/sec)
> #
> 
> 
> 
> Create one more file and snapshot that
> 
> # touch /storage/a/4
> # zfs snapshot storage/a...@2010.10.20
> 
> 
> send it
> 
> # zfs send -i storage/a...@2010.10.19  storage/a...@2010.10.20 | zfs receive 
> -v
> storage/compressed/a
> receiving incremental stream of storage/a...@2010.10.20 into
> storage/compressed/a...@2010.10.20
> received 250KB stream in 3 seconds (83.4KB/sec)
> 
> What do we have?
> 
> # find /storage/compressed/a
> /storage/compressed/a
> /storage/compressed/a/1
> /storage/compressed/a/2
> /storage/compressed/a/3
> /storage/compressed/a/4
> 
> 
> Of note:
> 
> * FreeBSD 8.1-STABLE
> * ZFS filesystem version 4.
> * ZFS pool version 15.
> * zfs send is on compression off
> * zfs receive has compression on
> 
> What I actually want to do and what fails:
> 
> 
> # zfs snapshot storage/bac...@2010.10.19
> 
> # zfs send storage/bac...@2010.10.19 | zfs receive -v
> storage/compressed/bacula
> receiving full stream of storage/bac...@2010.10.19 into
> storage/compressed/bac...@2010.10.19
> received 4.38TB stream in 42490 seconds (108MB/sec)
> 
> # zfs snapshot storage/bac...@2010.10.20
> 
> # zfs send -i storage/bac...@2010.10.19 storage/bac...@2010.10.20 | zfs
> receive -v storage/compressed/bacula
> receiving incremental stream of storage/bac...@2010.10.20 into
> storage/compressed/bac...@2010.10.20
> cannot receive incremental stream: destination storage/compressed/bacula
> has been modified
> since most recent snapshot
> warning: cannot send 'storage/bac...@2010.10.20': Broken pipe
> 
> I have no idea why this fails.  Clues please?
> 
> To my knowledge, the destination has not been written to.

Has any read operation been done on the destination (ie: updated atime) ?

Ruben

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: 'zfs send -i': destination has been modified

2010-10-20 Thread Dan Langille

On Wed, October 20, 2010 8:44 am, Ruben de Groot wrote:
> On Wed, Oct 20, 2010 at 08:32:33AM -0400, Dan Langille typed:
>> I am trying to do a 'zfs send -i' and failing.
>>
>> This is my simple proof of concept test:
>>
>> Create the data
>> # zfs create storage/a
>> # touch /storage/a/1
>> # touch /storage/a/2
>> # touch /storage/a/3
>>
>> Snapshot
>> # zfs snapshot storage/a...@2010.10.19
>>
>> send
>> # zfs send storage/a...@2010.10.19 | zfs receive -v storage/compressed/a
>> receiving full stream of storage/a...@2010.10.19 into
>> storage/compressed/a...@2010.10.19
>> received 252KB stream in 2 seconds (126KB/sec)
>> #
>>
>>
>>
>> Create one more file and snapshot that
>>
>> # touch /storage/a/4
>> # zfs snapshot storage/a...@2010.10.20
>>
>>
>> send it
>>
>> # zfs send -i storage/a...@2010.10.19  storage/a...@2010.10.20 | zfs receive
>> -v
>> storage/compressed/a
>> receiving incremental stream of storage/a...@2010.10.20 into
>> storage/compressed/a...@2010.10.20
>> received 250KB stream in 3 seconds (83.4KB/sec)
>>
>> What do we have?
>>
>> # find /storage/compressed/a
>> /storage/compressed/a
>> /storage/compressed/a/1
>> /storage/compressed/a/2
>> /storage/compressed/a/3
>> /storage/compressed/a/4
>>
>>
>> Of note:
>>
>> * FreeBSD 8.1-STABLE
>> * ZFS filesystem version 4.
>> * ZFS pool version 15.
>> * zfs send is on compression off
>> * zfs receive has compression on
>>
>> What I actually want to do and what fails:
>>
>>
>> # zfs snapshot storage/bac...@2010.10.19
>>
>> # zfs send storage/bac...@2010.10.19 | zfs receive -v
>> storage/compressed/bacula
>> receiving full stream of storage/bac...@2010.10.19 into
>> storage/compressed/bac...@2010.10.19
>> received 4.38TB stream in 42490 seconds (108MB/sec)
>>
>> # zfs snapshot storage/bac...@2010.10.20
>>
>> # zfs send -i storage/bac...@2010.10.19 storage/bac...@2010.10.20 | zfs
>> receive -v storage/compressed/bacula
>> receiving incremental stream of storage/bac...@2010.10.20 into
>> storage/compressed/bac...@2010.10.20
>> cannot receive incremental stream: destination storage/compressed/bacula
>> has been modified
>> since most recent snapshot
>> warning: cannot send 'storage/bac...@2010.10.20': Broken pipe
>>
>> I have no idea why this fails.  Clues please?
>>
>> To my knowledge, the destination has not been written to.
>
> Has any read operation been done on the destination (ie: updated atime) ?

Not that I know of.  But I do think that is the issue.  Thank you.  Adding
a -F option to the receive helps:

# zfs send -i storage/bac...@2010.10.19 storage/bac...@2010.10.20 | zfs
receive -vF storage/compressed/bacula
receiving incremental stream of storage/bac...@2010.10.20 into
storage/compressed/bac...@2010.10.20
received 20.0GB stream in 303 seconds (67.5MB/sec)

Just after I sent my email, I found this post:

 http://lists.freebsd.org/pipermail/freebsd-current/2007-July/075774.html

Problem solved.  :)


-- 
Dan Langille -- http://langille.org/

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: 'zfs send -i': destination has been modified

2010-10-20 Thread Derek Buttineau
On 2010-10-20, at 8:54 AM, Dan Langille wrote:

> Not that I know of.  But I do think that is the issue.  Thank you.  Adding
> a -F option to the receive helps:
> 
> # zfs send -i storage/bac...@2010.10.19 storage/bac...@2010.10.20 | zfs
> receive -vF storage/compressed/bacula
> receiving incremental stream of storage/bac...@2010.10.20 into
> storage/compressed/bac...@2010.10.20
> received 20.0GB stream in 303 seconds (67.5MB/sec)
> 
> Just after I sent my email, I found this post:
> 
> http://lists.freebsd.org/pipermail/freebsd-current/2007-July/075774.html
> 
> Problem solved.  :)

Good workaround, forcing the receive, but I'd still like to figure out what's 
causing the filesystem to be modified on the receiving end.  But this will be 
useful if I can't determine what is causing it to differ.

Thanks,

Derek

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: 'zfs send -i': destination has been modified

2010-10-20 Thread Derek Buttineau
Seeing similar here Dan, my destination filesystems are becoming modified 
sometime after the previous snapshot has been sent so the incremental fails to 
be received.  However, the server I'm sending to is not in use so I can't 
explain why it's changing.  

When I run a zdiff on the receiving system's filesystem, it can't find any 
changes in the file, but the byte difference between the snapshot and current 
can range from anywhere from 32KB to 101MB so far in my testing. 

If I rollback the filesystem to the last received snapshot it works, but I have 
no explanation for it changing.

I disabled everything that was running on the receiving system yesterday 
afternoon and there's been no change since, so now I need to step through and 
see if I can figure out what process was causing the byte difference on the 
filesystems.

Derek

On 2010-10-20, at 8:32 AM, Dan Langille wrote:

> I am trying to do a 'zfs send -i' and failing.
> 
> This is my simple proof of concept test:
> 
> Create the data
> # zfs create storage/a
> # touch /storage/a/1
> # touch /storage/a/2
> # touch /storage/a/3
> 
> Snapshot
> # zfs snapshot storage/a...@2010.10.19
> 
> send
> # zfs send storage/a...@2010.10.19 | zfs receive -v storage/compressed/a
> receiving full stream of storage/a...@2010.10.19 into
> storage/compressed/a...@2010.10.19
> received 252KB stream in 2 seconds (126KB/sec)
> #
> 
> 
> 
> Create one more file and snapshot that
> 
> # touch /storage/a/4
> # zfs snapshot storage/a...@2010.10.20
> 
> 
> send it
> 
> # zfs send -i storage/a...@2010.10.19  storage/a...@2010.10.20 | zfs receive 
> -v
> storage/compressed/a
> receiving incremental stream of storage/a...@2010.10.20 into
> storage/compressed/a...@2010.10.20
> received 250KB stream in 3 seconds (83.4KB/sec)
> 
> What do we have?
> 
> # find /storage/compressed/a
> /storage/compressed/a
> /storage/compressed/a/1
> /storage/compressed/a/2
> /storage/compressed/a/3
> /storage/compressed/a/4
> 
> 
> Of note:
> 
> * FreeBSD 8.1-STABLE
> * ZFS filesystem version 4.
> * ZFS pool version 15.
> * zfs send is on compression off
> * zfs receive has compression on
> 
> What I actually want to do and what fails:
> 
> 
> # zfs snapshot storage/bac...@2010.10.19
> 
> # zfs send storage/bac...@2010.10.19 | zfs receive -v
> storage/compressed/bacula
> receiving full stream of storage/bac...@2010.10.19 into
> storage/compressed/bac...@2010.10.19
> received 4.38TB stream in 42490 seconds (108MB/sec)
> 
> # zfs snapshot storage/bac...@2010.10.20
> 
> # zfs send -i storage/bac...@2010.10.19 storage/bac...@2010.10.20 | zfs
> receive -v storage/compressed/bacula
> receiving incremental stream of storage/bac...@2010.10.20 into
> storage/compressed/bac...@2010.10.20
> cannot receive incremental stream: destination storage/compressed/bacula
> has been modified
> since most recent snapshot
> warning: cannot send 'storage/bac...@2010.10.20': Broken pipe
> 
> I have no idea why this fails.  Clues please?
> 
> To my knowledge, the destination has not been written to.
> 
> 
> -- 
> Dan Langille -- http://langille.org/
> 
> ___
> freebsd-stable@freebsd.org mailing list
> http://lists.freebsd.org/mailman/listinfo/freebsd-stable
> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


latest -stable: still waiting after ...

2010-10-20 Thread Daniel Braniss
hi,
with the latest -stable, the boot process gets stuck with
...
ugen2.2:  at usbus2
uhub6:  on 
usbus2
uhub6: 3 ports with 3 removable, self powered
ugen3.2:  at usbus3
ukbd0:  on usbus3
kbd2 at ukbd0
ums0:  on usbus3
ums0: 3 buttons and [Z] coordinates ID=0 <- stuck here
run_interrupt_driven_hooks: still waiting after 60 seconds for xpt_config  
SMP: AP CPU #1 Launched!
SMP: AP CPU #7 Launched!
SMP: AP CPU #3 Launched!


this does not happen with and older -stable (August) kernel
Cheers,
danny


___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: Has anyone usd hast in production yet - opinions ?

2010-10-20 Thread Pete French
> Being the author of many problem reports I can say that most of them were not
> critical and for marginal cases (like some issues with hooks or a race that
> showed up when changing HAST role in loop -- you would never do this in
> production). And fixes were committed in several days after a report. I don't
> know any open issue.

Ah, thanks for this. I have just updated everything to the latest -STABLE,
and am thinking of shifting to hast sometime next week.

cheers,

-pete.
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Spurious reboot in 8.1-RELEASE when reading from ZFS pool with > 9 disks

2010-10-20 Thread Sean Thomas Caron

Hi folks,

I've been playing with ZFS in 8.1-RELEASE (amd64) on a Sun Fire X4500  
with 16 GB RAM and in general it seems to work well when used as  
recommended.


BUT...

In spite of the suggestion of Sun and FreeBSD developers to the  
contrary, I have been trying to create some zraid pools of size  
greater than 9 disks and it seems to give some trouble.


If I try to create a pool containing more than 9 [500 GB] disks,  
doesn't matter if it is zraid1 or zraid2, the system seems to reboot  
given any amount of sustained reads from the pool (haven't tested  
mirrored pools). Now, I am not sure of:


- Whether the reboot in the zraid1 case is caused by exactly the same  
issue as the reboot in the zraid2 case


- Whether this is an issue of total number of member disks, or total  
amount of disk space in the pool. All I have to work with at the  
moment is 500 GB drives.


I am not doing any sysctl tuning; just running with the defaults or  
what the system automatically sizes. I tried playing around with  
tuning some sysctl parameters including setting arc_max to be very  
small and it didn't seem to help any; pools of greater than 9 disks in  
size are always unstable.


Writes seem to work OK; I can, say, pull stuff from over the network  
and save it to the pool, or I can do something like,


dd if=/dev/random of=/mybigpool/bigfile bs=1024 size=10M

and it will write data all day pretty happily. But if I try to read  
back from the pool, for example,


dd if=/mybigpool/bigfile of=/dev/null bs=1024

or even to just do something like,

cp /mybigpool/bigfile /mybigpool/bigfile_2

the system reboots pretty much immediately. I never see anything on  
the console at all; it just reboots.


Even if I build a new kernel with debugging options:

options KDB
options DDB

the system still just reboots; I never see anything on the console and  
I never get to the debugger.


So, as I say, very easy to reproduce the problem, just create a zraid  
pool of any type with more than 9 member disks, dump some data to it,  
then try to read it back, and the machine will reboot.


If I create a pool with 9 or fewer disks, the system seems perfectly  
stable. I was never able to reproduce the reboot behavior as long as  
the pools contained 9 or fewer drives, beating on it fairly hard with  
iozone and multiple current dd operations writing large files to and  
from memory.


Just wondering if anyone's seen this problem before and as to whether  
or not it is a known bug and may have been fixed in STABLE or CURRENT?  
Should I report this as a bug? Should I just create pools of 9 or  
fewer drives? Not sure if my customer is going to want to use STABLE  
or CURRENT in production but I wanted to run this by the list just to  
see.


Best,

-Sean
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


RE: Spurious reboot in 8.1-RELEASE when reading from ZFS pool with > 9 disks

2010-10-20 Thread Sean Thomas Caron

Hi Lawrence,

Interesting; have you tried this for raidz2 as well?

I just created a raidz2 pool with 5 disks and then added another 5  
disk raidz2 to it, so, total of 10 disks in the pool (though this is  
ultimately a losing strategy unless the number of disks is >> 9  
because two drives are lost for parity in each sub-raid in the pool).


It (seemed) just slightly more stable than creating a single raidz2  
pool with > 9 disks but it still crashes.


I guess this does allow me to say its more an issue of number of  
devices in the pool versus capacity of the pool because with the  
parity drives taken out, the pool with two 5-disk raidz2s has less  
total capacity than a pool with a single 9-disk raidz2.


Just out of idle curiousity, I also tried it with raidz1 on my system.  
Again, I created a 5-disk pool, raidz1 this time, then added another  
5-disk raidz1 to the pool for, again, total of 10 disks.


Again, a bit of a losing strategy versus creating one great big raidz  
unless the number of disks is >> 9 because of losing a disk in each  
sub-raidz1 in the pool for parity but less so of course than raidz2.


This seemed to crash too, same behavior.

Are you using 8.1-RELEASE or STABLE or ...?

Best,

-Sean



I have a 16 disk pool, if you create it with

zpool create poolname raidz disk1 disk2 disk3 etc

then

zpool add poolname raidz disk8 disk9 disk10 etc

You get the full size pool and no issues.

  pool: tank
 state: ONLINE
 scan: scrub repaired 0 in 0h0m with 0 errors on Wed Oct 20 14:54:08 2010
config:

NAMESTATE READ WRITE CKSUM
tankONLINE   0 0 0
  raidz1-0  ONLINE   0 0 0
da0 ONLINE   0 0 0
da1 ONLINE   0 0 0
da2 ONLINE   0 0 0
da3 ONLINE   0 0 0
da4 ONLINE   0 0 0
da5 ONLINE   0 0 0
da6 ONLINE   0 0 0
da7 ONLINE   0 0 0
  raidz1-1  ONLINE   0 0 0
da8 ONLINE   0 0 0
da9 ONLINE   0 0 0
da10ONLINE   0 0 0
da11ONLINE   0 0 0
da12ONLINE   0 0 0
da13ONLINE   0 0 0
da14ONLINE   0 0 0
da15ONLINE   0 0 0

errors: No known data errors



___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: Spurious reboot in 8.1-RELEASE when reading from ZFS pool with > 9 disks

2010-10-20 Thread Boris Kochergin
 Ahoy. I just thought I'd add a data point to the mix. I have an 
11-disk v13 pool comprised of 400-GB disks on an 8.1 amd64 system and 
the machine behaves just fine with it:


# zpool status
  pool: archive
 state: ONLINE
status: The pool is formatted using an older on-disk format.  The pool can
still be used, but some features are unavailable.
action: Upgrade the pool using 'zpool upgrade'.  Once this is done, the
pool will no longer be accessible on older software versions.
 scrub: resilver completed after 0h0m with 0 errors on Fri Oct  8 
17:56:52 2010

config:

NAMESTATE READ WRITE CKSUM
archive ONLINE   0 0 0
  raidz1ONLINE   0 0 0
ad4 ONLINE   0 0 0  133K resilvered
ad6 ONLINE   0 0 0  84K resilvered
ad8 ONLINE   0 0 0  85.5K resilvered
ad10ONLINE   0 0 0  84.5K resilvered
ad12ONLINE   0 0 0  88K resilvered
ad14ONLINE   0 0 0  83.5K resilvered
ad16ONLINE   0 0 0  83K resilvered
ad18ONLINE   0 0 0  84.5K resilvered
ad20ONLINE   0 0 0  85.5K resilvered
ad22ONLINE   0 0 0  84K resilvered
ad24ONLINE   0 0 0  86.5K resilvered

errors: No known data errors

-Boris

On 10/20/10 13:26, Sean Thomas Caron wrote:

Hi Lawrence,

Interesting; have you tried this for raidz2 as well?

I just created a raidz2 pool with 5 disks and then added another 5 
disk raidz2 to it, so, total of 10 disks in the pool (though this is 
ultimately a losing strategy unless the number of disks is >> 9 
because two drives are lost for parity in each sub-raid in the pool).


It (seemed) just slightly more stable than creating a single raidz2 
pool with > 9 disks but it still crashes.


I guess this does allow me to say its more an issue of number of 
devices in the pool versus capacity of the pool because with the 
parity drives taken out, the pool with two 5-disk raidz2s has less 
total capacity than a pool with a single 9-disk raidz2.


Just out of idle curiousity, I also tried it with raidz1 on my system. 
Again, I created a 5-disk pool, raidz1 this time, then added another 
5-disk raidz1 to the pool for, again, total of 10 disks.


Again, a bit of a losing strategy versus creating one great big raidz 
unless the number of disks is >> 9 because of losing a disk in each 
sub-raidz1 in the pool for parity but less so of course than raidz2.


This seemed to crash too, same behavior.

Are you using 8.1-RELEASE or STABLE or ...?

Best,

-Sean



I have a 16 disk pool, if you create it with

zpool create poolname raidz disk1 disk2 disk3 etc

then

zpool add poolname raidz disk8 disk9 disk10 etc

You get the full size pool and no issues.

  pool: tank
 state: ONLINE
 scan: scrub repaired 0 in 0h0m with 0 errors on Wed Oct 20 14:54:08 
2010

config:

NAMESTATE READ WRITE CKSUM
tankONLINE   0 0 0
  raidz1-0  ONLINE   0 0 0
da0 ONLINE   0 0 0
da1 ONLINE   0 0 0
da2 ONLINE   0 0 0
da3 ONLINE   0 0 0
da4 ONLINE   0 0 0
da5 ONLINE   0 0 0
da6 ONLINE   0 0 0
da7 ONLINE   0 0 0
  raidz1-1  ONLINE   0 0 0
da8 ONLINE   0 0 0
da9 ONLINE   0 0 0
da10ONLINE   0 0 0
da11ONLINE   0 0 0
da12ONLINE   0 0 0
da13ONLINE   0 0 0
da14ONLINE   0 0 0
da15ONLINE   0 0 0

errors: No known data errors



___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: Spurious reboot in 8.1-RELEASE when reading from ZFS pool with > 9 disks

2010-10-20 Thread Jeremy Chadwick
On Wed, Oct 20, 2010 at 11:27:38AM -0400, Sean Thomas Caron wrote:
> I've been playing with ZFS in 8.1-RELEASE (amd64) on a Sun Fire
> X4500 with 16 GB RAM and in general it seems to work well when used
> as recommended.
> 
> BUT...
> 
> In spite of the suggestion of Sun and FreeBSD developers to the
> contrary, I have been trying to create some zraid pools of size
> greater than 9 disks and it seems to give some trouble.
> 
> If I try to create a pool containing more than 9 [500 GB] disks,
> doesn't matter if it is zraid1 or zraid2, the system seems to reboot
> given any amount of sustained reads from the pool (haven't tested
> mirrored pools). Now, I am not sure of:
> 
> - Whether the reboot in the zraid1 case is caused by exactly the
> same issue as the reboot in the zraid2 case
> 
> - Whether this is an issue of total number of member disks, or total
> amount of disk space in the pool. All I have to work with at the
> moment is 500 GB drives.
> 
> I am not doing any sysctl tuning; just running with the defaults or
> what the system automatically sizes. I tried playing around with
> tuning some sysctl parameters including setting arc_max to be very
> small and it didn't seem to help any; pools of greater than 9 disks
> in size are always unstable.
> 
> Writes seem to work OK; I can, say, pull stuff from over the network
> and save it to the pool, or I can do something like,
> 
> dd if=/dev/random of=/mybigpool/bigfile bs=1024 size=10M
> 
> and it will write data all day pretty happily. But if I try to read
> back from the pool, for example,
> 
> dd if=/mybigpool/bigfile of=/dev/null bs=1024
> 
> or even to just do something like,
> 
> cp /mybigpool/bigfile /mybigpool/bigfile_2
> 
> the system reboots pretty much immediately. I never see anything on
> the console at all; it just reboots.
> 
> Even if I build a new kernel with debugging options:
> 
> options KDB
> options DDB
> 
> the system still just reboots; I never see anything on the console
> and I never get to the debugger.
> 
> So, as I say, very easy to reproduce the problem, just create a
> zraid pool of any type with more than 9 member disks, dump some data
> to it, then try to read it back, and the machine will reboot.
> 
> If I create a pool with 9 or fewer disks, the system seems perfectly
> stable. I was never able to reproduce the reboot behavior as long as
> the pools contained 9 or fewer drives, beating on it fairly hard
> with iozone and multiple current dd operations writing large files
> to and from memory.
> 
> Just wondering if anyone's seen this problem before and as to
> whether or not it is a known bug and may have been fixed in STABLE
> or CURRENT? Should I report this as a bug? Should I just create
> pools of 9 or fewer drives? Not sure if my customer is going to want
> to use STABLE or CURRENT in production but I wanted to run this by
> the list just to see.

There are users here using FreeBSD ZFS with *lots* of disks (I think
someone was using 32 disks at one point) reliably.  Some of them post
here regularly (with other issues that don't consist of sporadic
reboots).

The kernel options may not be sufficient.  I'm used to using these:

# Debugging options
options BREAK_TO_DEBUGGER   # Sending a serial BREAK drops to DDB
options KDB # Enable kernel debugger support
options KDB_TRACE   # Print stack trace automatically on 
panic
options DDB # Support DDB
options GDB # Support remote GDB

And in /etc/rc.conf, setting:

ddb_enable="yes"

Next: arc_max isn't "technically" a sysctl, meaning it can't be changed
in real-time, so I'm not sure how you managed to do that.  Validation:

sysctl: oid 'vfs.zfs.arc_max' is a read only tunable
sysctl: Tunable values are set in /boot/loader.conf

Your system may be reporting something relating to kmem exhaustion but
is then auto-rebooting so fast that you can't see the message on VGA
console.  Do you have serial console?

Please try setting the following tunables in /boot/loader.conf and
reboot the machine, then see if the same problem persists.

vm.kmem_size="16384M"
vfs.zfs.arc_max="14336M"
vfs.zfs.prefetch_disable="1"
vfs.zfs.zio.use_uma="0"
vfs.zfs.txg.timeout="5"

I would also advocate you try 8.1-STABLE as there have been many changes
in ZFS since then (and I'm not just referring to the v15 import),
including how the ARC gets sized/adjusted.  CURRENT is highly
bleeding-edge, so I would start or stick with STABLE.

Finally, there's always the possibility that the PSU has some sort of
load problem with that many disks all being accessed at the same time.
I imagine the power draw of that system is quite high.  I can't imagine
Sun shipping a box with a insufficient PSU, but then again power draw
changes depending on the RPM of the disks used and many other things.

-- 
| Jeremy Chadwick   j...@parodius.com |
| Parodius Networking

Re: kpanic on install >32GB of RAM [SEC=UNCLASSIFIED]

2010-10-20 Thread Andriy Gapon
on 20/10/2010 21:05 Sean Bruno said the following:
> HP testers report that 7 nor 8 install without the kernel
> 'insta-panicing' across all CPUS at the same time.
> 
> I'm going to ask that they try the latest 9 snap shot and then kill the
> project at this time.

So you are not willing to try the patch?

-- 
Andriy Gapon
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: kpanic on install >32GB of RAM [SEC=UNCLASSIFIED]

2010-10-20 Thread Sean Bruno
On Tue, 2010-10-19 at 06:01 -0700, John Baldwin wrote:
> On Tuesday, October 19, 2010 1:16:07 am Andriy Gapon wrote:
> > on 19/10/2010 06:11 Wilkinson, Alex said the following:
> > > Will it work something like http://freshmeat.net/projects/mcelog/ ?
> > 
> > jhb has a port of this, yes.
> 
> The relevant bits are at http://www.FreeBSD.org/~jhb/mcelog/
> 
> I should probably make this into a port, but I really need to e-mail the 
> mcelog folks about integrating the patches into their tree.
> 


Q:  Has anyone installed any version of FreeBSD on a DL980 G7 from HP
yet?  Has anyone installed any version of FreeBSD on a DL580 G7 from HP?

HP testers report that 7 nor 8 install without the kernel
'insta-panicing' across all CPUS at the same time.

I'm going to ask that they try the latest 9 snap shot and then kill the
project at this time.

sean

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: kpanic on install >32GB of RAM [SEC=UNCLASSIFIED]

2010-10-20 Thread Andriy Gapon
on 20/10/2010 21:16 Sean Bruno said the following:
> Huh ... did I miss it?  I'm willing to try a lot.

Perhaps then :-)
This was my post:
http://thread.gmane.org/gmane.os.freebsd.stable/72450/focus=72515

It had a link to a parallel discussion on current:
http://thread.gmane.org/gmane.os.freebsd.current/128099/focus=128576

Which had a link to:
http://people.freebsd.org/~avg/uma-many-cpus.diff

Which is the patch :)
-- 
Andriy Gapon
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: kpanic on install >32GB of RAM [SEC=UNCLASSIFIED]

2010-10-20 Thread Sean Bruno
On Wed, 2010-10-20 at 11:11 -0700, Andriy Gapon wrote:
> on 20/10/2010 21:05 Sean Bruno said the following:
> > HP testers report that 7 nor 8 install without the kernel
> > 'insta-panicing' across all CPUS at the same time.
> > 
> > I'm going to ask that they try the latest 9 snap shot and then kill the
> > project at this time.
> 
> So you are not willing to try the patch?
> 


Huh ... did I miss it?  I'm willing to try a lot.

sean

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: kpanic on install >32GB of RAM [SEC=UNCLASSIFIED]

2010-10-20 Thread Sean Bruno
On Wed, 2010-10-20 at 11:20 -0700, Andriy Gapon wrote:
> on 20/10/2010 21:16 Sean Bruno said the following:
> > Huh ... did I miss it?  I'm willing to try a lot.
> 
> Perhaps then :-)
> This was my post:
> http://thread.gmane.org/gmane.os.freebsd.stable/72450/focus=72515
> 
> It had a link to a parallel discussion on current:
> http://thread.gmane.org/gmane.os.freebsd.current/128099/focus=128576
> 
> Which had a link to:
> http://people.freebsd.org/~avg/uma-many-cpus.diff
> 
> Which is the patch :)

So here's the problem.

Machine in in Houston, can't netboot/nfsboot it.  The basic installers
all will get me to the Beastie loader, but fail after loading the kernel
and attempting to boot.

I guess, I could replace the kernel on the CD and have them reburn it?

Sean

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: kpanic on install >32GB of RAM [SEC=UNCLASSIFIED]

2010-10-20 Thread Boris Samorodov
On Wed, 20 Oct 2010 11:28:21 -0700 Sean Bruno wrote:

> So here's the problem.

> Machine in in Houston, can't netboot/nfsboot it.  The basic installers
> all will get me to the Beastie loader, but fail after loading the kernel
> and attempting to boot.

> I guess, I could replace the kernel on the CD and have them reburn it?

You may create a booting USB stick and send them the image.

-- 
WBR, Boris Samorodov (bsam)
Research Engineer, http://www.ipt.ru Telephone & Internet SP
FreeBSD Committer, http://www.FreeBSD.org The Power To Serve
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: Spurious reboot in 8.1-RELEASE when reading from ZFS pool with > 9 disks

2010-10-20 Thread Sean Thomas Caron

Hi Jeremy,

Thanks for the very helpful response!

I added all debugging options that you specified to my kernel and  
rebuilt; then set the kernel parameters as you mention (I was being a  
bit lazy earlier when I called them sysctls; I always tuned them in  
loader.conf; just that you can view their values with sysctl).


Rebooted the system with the new kernel and set up a 11-disk zraid2  
pool again then started beating on it. At first it seemed to be a bit  
more resilient with this set of kernel parameters but eventually it  
too failed out.


Again I just got a straight up reboot, no debugger, no output to the  
console flashed by as far as I can tell.


I don't have a serial console hooked up right now but it's probably  
possible to do so through the ILOM or equivalent; I will have to look  
into that further.


This is pretty wierd.

I am thinking there might be some memory starting to go in this  
system; never seen failing memory in an ECC box cause reboots this  
consistently and only under such specific conditions but I suppose it  
isn't completely out of the question. I'll talk to my customer and see  
what they can do about the hardware; maybe they have some spares.


I will also try 8.1-STABLE when I have a chance and see if that works better.

But it's definitely helpful to know that folks have > 9 disk raidz  
pools up and running on FreeBSD 8.x with no trouble - that it "should  
work". And the list of tunables is very useful; nice to have something  
to work with that I can have a bit more confidence in outside of my  
own guessing :)


I will report back to the list when I have more information.

Thanks!

-Sean


Quoting Jeremy Chadwick :


There are users here using FreeBSD ZFS with *lots* of disks (I think
someone was using 32 disks at one point) reliably.  Some of them post
here regularly (with other issues that don't consist of sporadic
reboots).

The kernel options may not be sufficient.  I'm used to using these:

# Debugging options
options BREAK_TO_DEBUGGER   # Sending a serial BREAK drops to DDB
options KDB # Enable kernel debugger support
options KDB_TRACE   # Print stack trace  
automatically on panic

options DDB # Support DDB
options GDB # Support remote GDB

And in /etc/rc.conf, setting:

ddb_enable="yes"

Next: arc_max isn't "technically" a sysctl, meaning it can't be changed
in real-time, so I'm not sure how you managed to do that.  Validation:

sysctl: oid 'vfs.zfs.arc_max' is a read only tunable
sysctl: Tunable values are set in /boot/loader.conf

Your system may be reporting something relating to kmem exhaustion but
is then auto-rebooting so fast that you can't see the message on VGA
console.  Do you have serial console?

Please try setting the following tunables in /boot/loader.conf and
reboot the machine, then see if the same problem persists.

vm.kmem_size="16384M"
vfs.zfs.arc_max="14336M"
vfs.zfs.prefetch_disable="1"
vfs.zfs.zio.use_uma="0"
vfs.zfs.txg.timeout="5"

I would also advocate you try 8.1-STABLE as there have been many changes
in ZFS since then (and I'm not just referring to the v15 import),
including how the ARC gets sized/adjusted.  CURRENT is highly
bleeding-edge, so I would start or stick with STABLE.

Finally, there's always the possibility that the PSU has some sort of
load problem with that many disks all being accessed at the same time.
I imagine the power draw of that system is quite high.  I can't imagine
Sun shipping a box with a insufficient PSU, but then again power draw
changes depending on the RPM of the disks used and many other things.

--
| Jeremy Chadwick   j...@parodius.com |
| Parodius Networking   http://www.parodius.com/ |
| UNIX Systems Administrator  Mountain View, CA, USA |
| Making life hard for others since 1977.  PGP: 4BD6C0CB |

___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"






___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"