On January 28, 2007 7:57:31 PM -0800 "Anantha N. Srirama"
<[EMAIL PROTECTED]> wrote:
Agreed, I guess I didn't articulate my point/thought very well. The best
config is to present JBoDs and let ZFS provide the data protection. This
has been a very stimulating conversation thread; it is shedding ne
Agreed, I guess I didn't articulate my point/thought very well. The best config
is to present JBoDs and let ZFS provide the data protection. This has been a
very stimulating conversation thread; it is shedding new light into how to best
use ZFS.
This message posted from opensolaris.org
_
Hello Francois,
Friday, January 26, 2007, 4:09:43 PM, you wrote:
FD> On Fri, 2007-01-26 at 06:16 -0800, Jeffery Malloch wrote:
>> Hi Folks,
>>
>> I am currently in the midst of setting up a completely new file server using
>> a pretty well loaded Sun T2000 (8x1GHz, 16GB RAM) connected to an Eng
Hello Anantha,
Friday, January 26, 2007, 5:06:46 PM, you wrote:
ANS> All my feedback is based on Solaris 10 Update 2 (aka 06/06) and
ANS> I've no comments on NFS. I strongly recommend that you use ZFS
ANS> data redundancy (z1, z2, or mirror) and simply delegate the
ANS> Engenio to stripe the data
Hello Jeff,
Saturday, January 27, 2007, 8:27:09 AM, you wrote:
JB> You're all correct. File data is never byte-swapped. Most metadata
JB> needs to be byte-swapped, but it's generally only 1-2% of your space.
JB> So the overhead shouldn't be significant, even if you never rewrite.
I remember s
Hello Richard,
Friday, January 26, 2007, 11:36:07 PM, you wrote:
RE> We've been talking a lot recently about failure rates and types of
RE> failures. As you may know, I do look at field data and generally don't
RE> ask the group for more data. But this time, for various reasons (I
RE> might hav
Anton B. Rang wrote:
How badly can you mess up a JBOD?
Two words: vibration, cooling.
Three more: power, signal quality.
I've seen even individual drive cases with bad enough signal quality to cause
bit errors.
Yep, if I crank up the amp to over 1kW, then on some frequencies, I see lots
of
On January 28, 2007 4:59:48 PM +0100 Pawel Jakub Dawidek <[EMAIL PROTECTED]>
wrote:
On Fri, Jan 26, 2007 at 06:08:50PM -0800, Darren Dunham wrote:
> 3. I created file system with huge amount of data, where most of the
> data is read-only. I change my server from intel to sparc64 machine.
> Adapt
[EMAIL PROTECTED] wrote on 01/27/2007 06:48:17 AM:
> is it planned to add some other compression algorithm to zfs ?
>
> lzjb is quite good and especially performing very well, but i`d like
> to have better compression (bzip2?) - no matter how worse
> performance drops with this.
>
> regards
On Sun, Jan 28, 2007 at 06:19:25AM -0800, Anantha N. Srirama wrote:
>
> You're right that storage level snapshots are filesystem agnostic. I'm
> not sure why you believe you won't be able to restore individual files
> by using a NetApp snapshot? In the case of ZFS you'd take a periodic
> snapshot
On Sat, Jan 27, 2007 at 04:15:30PM -0800, Anantha N. Srirama wrote:
>
> I'm not sure what benefit you forsee by running a COW filesystem
> (ZFS) on a COW array (NetApp).
The application requires a filesystem with POSIX semantics. My first
choice would be NFS from the Netapp, but this won't work
Too bad...I was in the situation where every "zpool ..." command was stuck (as
well as df command) and my hope was - it's a known/fixed bug. I could not save
the core files, not sure I can reproduce the bug.
Thank you for quick reply,
[i]-- leon[/i]
This message posted from opensolaris.org
_
Hi Leon,
This was fixed in March 2006, and is in S10_U2.
Neil.
Leon Koll wrote On 01/28/07 08:58,:
Hello,
what is the status of the bug 6381203 fix in S10 u3 ?
("deadlock due to i/o while assigning (tc_lock held)")
Was it integrated? Is there a patch?
Thanks,
[i]-- leon[/i]
This message
On Fri, Jan 26, 2007 at 06:08:50PM -0800, Darren Dunham wrote:
> > What do you guys think about implementing 'zfs/zpool rewrite' command?
> > It'll read every block older than the date when the command was executed
> > and write it again (using standard ZFS COW mechanism, simlar to how
> > resilver
Hello,
what is the status of the bug 6381203 fix in S10 u3 ?
("deadlock due to i/o while assigning (tc_lock held)")
Was it integrated? Is there a patch?
Thanks,
[i]-- leon[/i]
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs
You're right that storage level snapshots are filesystem agnostic. I'm not sure
why you believe you won't be able to restore individual files by using a NetApp
snapshot? In the case of ZFS you'd take a periodic snapshot and use it to
restore files, in the case of NetApp you can do the same (of c
>
>On 28-Jan-07, at 7:59 AM, [EMAIL PROTECTED] wrote:
>
>>
>>>
>>> On 27-Jan-07, at 10:15 PM, Anantha N. Srirama wrote:
>>>
... ZFS will not stop alpha particle induced memory corruption
after data has been received by server and verified to be correct.
Sadly I've been hit with that
On 28-Jan-07, at 7:59 AM, [EMAIL PROTECTED] wrote:
On 27-Jan-07, at 10:15 PM, Anantha N. Srirama wrote:
... ZFS will not stop alpha particle induced memory corruption
after data has been received by server and verified to be correct.
Sadly I've been hit with that as well.
My brother poi
>> Take note though, that giving zfs the entire disk gives a possible
>> performance win, as zfs will only enable the write cache for the disk
>> if it is given the entire disk.
>
>really?
>why this?
In the old days, Sun never enabled the write cache on devices because
of reliability issues. (S
> Take note though, that giving zfs the entire disk gives a possible
> performance win, as zfs will only enable the write cache for the disk
> if it is given the entire disk.
really?
why this?
is this tuneable somehow/somewhere? can i enabyle writecache if only using a
dedicated partition ?
[EMAIL PROTECTED] wrote:
> Alpha particles which "hit" CPUs must have their origin inside said CPU.
>
> (Alpha particles do not penentrate skin, paper, let alone system cases
> or CPU packagaging)
Gamma rays cannot be shielded in a senseful way.
Jörg
--
EMail:[EMAIL PROTECTED] (home) Jörg Sch
[EMAIL PROTECTED] wrote:
On 27-Jan-07, at 10:15 PM, Anantha N. Srirama wrote:
... ZFS will not stop alpha particle induced memory corruption
after data has been received by server and verified to be correct.
Sadly I've been hit with that as well.
My brother points out that you ca
>
>On 27-Jan-07, at 10:15 PM, Anantha N. Srirama wrote:
>
>> ... ZFS will not stop alpha particle induced memory corruption
>> after data has been received by server and verified to be correct.
>> Sadly I've been hit with that as well.
>
>
>My brother points out that you can use a rad hardened
23 matches
Mail list logo