Victor,
I've reproduced the crash and have vmdump.0 and dump device files. How do I
query the stack on crash for your analysis? What other analysis should I
provide?
Thanks
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-
> -Original Message-
> From: Erik Trimble [mailto:erik.trim...@oracle.com]
> Sent: 星期四, 七月 01, 2010 11:45
> To: Fred Liu
> Cc: Bob Friesenhahn; 'OpenSolaris ZFS discuss'
> Subject: Re: [zfs-discuss] OCZ Vertex 2 Pro performance numbers
>
> On 6/30/2010 7:17 PM, Fred Liu wrote:
> > See. T
On 6/30/2010 7:17 PM, Fred Liu wrote:
> See. Thanks.
> Does it have the hardware functionality to detect the power outage and do
> force cache flush when the cache is enabled?
> Any more detailed info about the capacity (farad) of this supercap and how
> long one discharge will be?
>
> Thanks.
>
On 6/30/2010 2:01 PM, valrh...@gmail.com wrote:
Another question on SSDs in terms of performance vs. capacity.
Between $150 and $200, there are at least three SSDs that would fit the rough
specifications for the L2ARC on my system:
1. Crucial C300, 64 GB: $150: medium performance, medium capac
Hi Eff,
There are a significant number of variables to work through with dedup and
compression enabled. So the first suggestion I have is to disable those
features for now so your not working with too many elements.
With those features set aside an NTFS cluster operation does not = a 64k raw
See. Thanks.
Does it have the hardware functionality to detect the power outage and do force
cache flush when the cache is enabled?
Any more detailed info about the capacity (farad) of this supercap and how long
one discharge will be?
Thanks.
Fred
-Original Message-
From: Bob Friesenh
On Wed, 30 Jun 2010, Fred Liu wrote:
Any duration limit on the supercap? How long can it sustain the data?
A supercap on a SSD drive only needs to sustain the data until it has
been saved (perhaps 10 milliseconds). It is different than a RAID
array battery.
Bob
--
Bob Friesenhahn
bfrie...
Any duration limit on the supercap? How long can it sustain the data?
Thanks.
Fred
-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of David Magda
Sent: 星期六, 六月 26, 2010 21:48
To: Arne Jansen
Cc: 'OpenSolaris ZFS discuss
Ragnar Sundblad wrote:
I was referring to the case where zfs has written data to the drive but
still hasen't issued a cache flush, and before the cache flush the drive
is reset. If zfs finally issues a cache flush and then isn't informed
that the drive has been reset, data is lost.
I hope this
On 07/ 1/10 01:36 AM, Tony MacDoodle wrote:
Hello,
Has anyone encountered the following error message, running Solaris 10
u8 in an LDom.
bash-3.00# devfsadm
devfsadm: write failed for /dev/.devfsadm_dev.lock: Bad exchange
descriptor
Not specifically. But it is clear from what follows th
Aha:
http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6794136
I think I'll try booting from a b134 Live CD and see that will let me fix
things.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.
Just in case any stray searches finds it way here, this is what happened to my
pool: http://phrenetic.to/zfs
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinf
On 30 jun 2010, at 22.46, Garrett D'Amore wrote:
> On Wed, 2010-06-30 at 22:28 +0200, Ragnar Sundblad wrote:
>
>> To be safe, the protocol needs to be able to discover that the devices
>> (host or disk) has been disconnected and reconnected or has been reset
>> and that either parts assumptions
On Wed, 2010-06-30 at 16:41 -0500, Nicolas Williams wrote:
> On Wed, Jun 30, 2010 at 01:35:31PM -0700, valrh...@gmail.com wrote:
> > Finally, for my purposes, it doesn't seem like a ZIL is necessary? I'm
> > the only user of the fileserver, so there probably won't be more than
> > two or three comp
On Wed, Jun 30, 2010 at 01:35:31PM -0700, valrh...@gmail.com wrote:
> Finally, for my purposes, it doesn't seem like a ZIL is necessary? I'm
> the only user of the fileserver, so there probably won't be more than
> two or three computers, maximum, accessing stuff (and writing stuff)
> remotely.
It
I have a Opensolaris snv_134 machine with 2 x 1.5TB drives. One is a Samsung
Silencer the other is a dreaded Western Digital Green.
I'm testing the mirror for failure by simply yanking out the SATA cable while
the machine is running. The system never skips a beat, which is great. But the
reconn
Another question on SSDs in terms of performance vs. capacity.
Between $150 and $200, there are at least three SSDs that would fit the rough
specifications for the L2ARC on my system:
1. Crucial C300, 64 GB: $150: medium performance, medium capacity.
2. OCZ Vertex 2, 50 GB: $180: higher performa
On Wed, 2010-06-30 at 22:28 +0200, Ragnar Sundblad wrote:
> To be safe, the protocol needs to be able to discover that the devices
> (host or disk) has been disconnected and reconnected or has been reset
> and that either parts assumptions about the state of the other has to
> be invalidated.
>
>
Thanks to everyone for such helpful and detailed answers. Contrary to some of
the trolls in other threads, I've had a fantastic experience here, and am
grateful to the community.
Based on the feedback, I'll upgrade my machine to 8 GB of RAM. I only have two
slots on the motherboard, and either
On 12 apr 2010, at 22.32, Carson Gaspar wrote:
> Carson Gaspar wrote:
>> Miles Nordin wrote:
"re" == Richard Elling writes:
>>> How do you handle the case when a hotplug SATA drive is powered off
>>> unexpectedly with data in its write cache? Do you replay the writes, or do
>>> they g
> This may not work for you, but it worked for me, and I was pleasantly
> surprised. Replace a drive with itself.
>
> zpool replace tank c0t2d0 c0t2d0
I tried that - it didn't work - I replaced the drive with a new one, that
worked, and then I made a new zpool on the old drive with zfs-fuse in L
- Original Message -
> I think zfs on ubuntu currently is a rather bad idea. See test below
> with ubuntu Lucid 10.04 (amd64)
Quick update on this - it seems this is due to a bug in the Linux kernel where
it can't deal with partition changes on a drive with mounted filesystems. I'm
not 1
On Wed, Jun 30, 2010 at 12:54 PM, Edward Ned Harvey
wrote:
>> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
>> boun...@opensolaris.org] On Behalf Of Asif Iqbal
>>
>> would be nice if i could pipe the zfs send stream to a split and then
>> send of those splitted stream over the
>>
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Roy Sigurd Karlsbakk
>
> There was some messup with switching of drives and an unexpected
> reboot, so I suddenly have a drive in my pool that is partly
> resilvered. zfs status shows the pool
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Asif Iqbal
>
> would be nice if i could pipe the zfs send stream to a split and then
> send of those splitted stream over the
> network to a remote system. it would help sending it over to remo
On Wed, Jun 30, 2010 at 09:47:15AM -0700, Edward Ned Harvey wrote:
> > From: Arne Jansen [mailto:sensi...@gmx.net]
> >
> > Edward Ned Harvey wrote:
> > > Due to recent experiences, and discussion on this list, my colleague
> > and
> > > I performed some tests:
> > >
> > > Using solaris 10, fully u
> From: Arne Jansen [mailto:sensi...@gmx.net]
>
> Edward Ned Harvey wrote:
> > Due to recent experiences, and discussion on this list, my colleague
> and
> > I performed some tests:
> >
> > Using solaris 10, fully upgraded. (zpool 15 is latest, which does
> not
> > have log device removal that wa
I searched and searched but was not able to find your added text in
this long quoted message. Please re-submit using the english language
in simple ASCII text intended for humans.
Thanks,
Bob
On Wed, 30 Jun 2010, Eric Andersen wrote:
On Jun 28, 2010, at 10:03 AM, zfs-discuss-requ...@opens
On Jun 28, 2010, at 10:03 AM, zfs-discuss-requ...@opensolaris.org wrote:
> Send zfs-discuss mailing list submissions to
> zfs-discuss@opensolaris.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> or, via ema
On Tue, Jun 29, 2010 at 11:25 AM, Patrick Donnelly wrote:
> I googled around but couldn't find anything on whether someone has
> good or bad experiences with the Caviar *Blue* drives? I saw in the
> archives Caviar Blacks are *not* recommended for ZFS arrays (excluding
> apparently RE3 and RE4?).
Hey Mitch,
The zpool split feature is available in the OpenSolaris release if
you upgrade to build 131.
You can read about the requirements here:
http://hub.opensolaris.org/bin/view/Community+Group+zfs/docs
See the ZFS Admin Guide, page 89-90
Thanks,
Cindy
On 06/29/10 13:37, Mitchell Petty
Hello,
Has anyone encountered the following error message, running Solaris 10 u8 in
an LDom.
bash-3.00# devfsadm
devfsadm: write failed for /dev/.devfsadm_dev.lock: Bad exchange descriptor
bash-3.00# zpool status -v rpool
pool: rpool
state: DEGRADED
status: One or more devices has experienced a
Well, I was doing a ZFS send / receive to backup a large (60 GB) of
data, which never completed. A zpool clear at that point just hung and I
had to reboot the system, after which it appeared to come up clean. As
soon as I tried the backup again I noticed the pool reported the error
you see belo
Interesting, this time it worked! Does specifying the device to clear
cause the command to behave differently? I had assumed w/out the device
specification, the clear would just apply to all devices in the pool
(which are just the one).
Thanks,
Brian
Cindy Swearingen wrote:
Hi Brian,
Becaus
Hi,
Is "zpool split" available ? If not when will it be ? If it is what
are the prerequisites ?
Thanks In Advance ,
Mitch
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/lis
Hi list,
I googled around but couldn't find anything on whether someone has
good or bad experiences with the Caviar *Blue* drives? I saw in the
archives Caviar Blacks are *not* recommended for ZFS arrays (excluding
apparently RE3 and RE4?). Specifically I'm looking to buy Western
Digital Caviar Bl
Hi Cindy,
The scrub didn't help and yes, this is an external USB device.
Thanks,
Brian
Cindy Swearingen wrote:
Hi Brian,
You might try running a scrub on this pool.
Is this an external USB device?
Thanks,
Cindy
On 06/29/10 09:16, Brian Leonard wrote:
Hi,
I have a zpool which is currentl
> Please try
>
> zdb -U /dev/null -ebcsv storage2
r...@crypt:~# zdb -U /dev/null -ebcsv storage2
zdb: can't open storage2: No such device or address
If I try
r...@crypt:~# zdb -C storage2
Then it prints what appears to be a valid configuration but then the same error
message about being unab
On Jun 30, 2010, at 10:48 AM, George wrote:
>> I suggest you to try running 'zdb -bcsv storage2' and
>> show the result.
>
> r...@crypt:/tmp# zdb -bcsv storage2
> zdb: can't open storage2: No such device or address
>
> then I tried
>
> r...@crypt:/tmp# zdb -ebcsv storage2
> zdb: can't open sto
39 matches
Mail list logo