Re: [zfs-discuss] AHCI or IDE?

2010-12-16 Thread Deano
AHCI or IDE is at the device level not zfs level, as such it's a property of
the OS layer. I suspect all OSs that zfs run on support AHCI directly
(OpenSolaris do). 

Always use it instead of IDE emulation (which is what is happening what you
select IDE in the firmware, it is purely for old OS support with no direct
SATA support)

HTH,
Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Alexander Lesle
Sent: 16 December 2010 19:43
To: zfs-discuss@opensolaris.org
Subject: [zfs-discuss] AHCI or IDE?

Hello All,

I want to build a home file and media server now. After experiment with a
Asus Board and running in unsolve problems I have bought this
Supermicro Board X8SIA-F with Intel i3-560 and 8 GB Ram
http://www.supermicro.com/products/motherboard/Xeon3000/3400/X8SIA.cfm?IPMI=
Y
also the LSI HBA SAS 9211-8i
http://lsi.com/storage_home/products_home/host_bus_adapters/sas_hbas/interna
l/sas9211-8i/index.html

rpool = 2vdev mirror
tank = 2 x 2vdev mirror. For the future I want to have the option to
expand up to 12 x 2vdev mirror.

After reading the board manual I found at page 4-9 where I can set
SATA#1 from IDE to AHCI.

Can zfs handle AHCI for rpool?
Can zfs handle AHCI for tank?

Thx for helping.
-- 
Regarrds
Alexander

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2010-12-20 Thread Deano
Hi,
Which brings up an interesting question... 

IF it were fixed in for example illumos or freebsd is there a plan for how
to handle possible incompatible zfs implementations?

Currently the basic version numbering only works as it implies only one
stream of development, now with multiple possible stream does ZFS need to
move to a feature bit system or are we going to have to have forks or
multiple incompatible versions?

Thanks,
Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Phil Harman
Sent: 20 December 2010 10:43
To: Lanky Doodle
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] A few questions

> Why does resilvering take so long in raidz anyway?

Because it's broken. There were some changes a while back that made it more
broken.

There has been a lot of discussion, anecdotes and some data on this list. 

The resilver doesn't do a single pass of the drives, but uses a "smarter"
temporal algorithm based on metadata.

However, the current implentation has difficulty finishing the job if
there's a steady flow of updates to the pool.

As far as I'm aware, the only way to get bounded resilver times is to stop
the workload until resilvering is completed.

The problem exists for mirrors too, but is not as marked because mirror
reconstruction is inherently simpler.

I believe Oracle is aware of the problem, but most of the core ZFS team has
left. And of course, a fix for Oracle Solaris no longer means a fix for the
rest of us.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2010-12-21 Thread Deano
On Dec 20, 2010, at 7:31 AM, Phil Harman  wrote:

> If you only have a few slow drives, you don't have performance.

> Like trying to win the Indianapolis 500 with a tricycle...

 

Well you can put a jet engine on a tricycle and perhaps win it… Or you can 
change the race course to only allow a tricycle space to move. In the context 
of storage we have 2 factors hardware and software, having faster and more 
reliable spindles is no reason to suggest that better software can’t be used to 
beat it. The simple example is ZIL SSD, where using some software and  even a 
cheap commodity SSD will outperform sync writes than any amount of expensive 
spindle drives. Before ZIL software is was easy to argue that the only way of 
speeding up writes was more faster spindles.

 

The question therefore is, is there room in the software implementation to 
achieve performance and reliability numbers similar to expensive drives whilst 
using relative cheap drives?

 

ZFS is good but IMHO easy to see how it can be improved to better meet this 
situation, I can’t currently say when this line of thinking and code will move 
from research to production level use (tho I have a pretty good idea ;) ) but I 
wouldn’t bet on the status quo lasting much longer. In some ways the removal of 
OpenSolaris may actually be a good thing, as its catalyized a number of 
developers from the view that zfs is Oracle led, to thinking “what can we do 
with zfs code as a base”?

 

Ffor example how about sticking a cheap 80GiB commodity SSD in the storage 
case. When a resilver or defrag is required, use it as a scratch space to give 
you a block of fast IOPs storage space to accelerate the slow parts. When its 
done secure erase and power it down, ready for the next time a resilver needs 
to happen. The hardware is available, just needs someone to write the software…

 

 

Bye,

Deano

 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2010-12-21 Thread Deano
Doh sorry about that, the threading got very confused on my mail reader!

 

Bye,

Deano

 

From: Phil Harman [mailto:phil.har...@gmail.com] 
Sent: 21 December 2010 13:12
To: Deano
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] A few questions

 

On 21/12/2010 13:05, Deano wrote: 

On Dec 20, 2010, at 7:31 AM, Phil Harman  wrote:

> If you only have a few slow drives, you don't have performance.

> Like trying to win the Indianapolis 500 with a tricycle...


Actually, I didn't say that, Richard did :)




Well you can put a jet engine on a tricycle and perhaps win it… Or you can 
change the race course to only allow a tricycle space to move. In the context 
of storage we have 2 factors hardware and software, having faster and more 
reliable spindles is no reason to suggest that better software can’t be used to 
beat it. The simple example is ZIL SSD, where using some software and  even a 
cheap commodity SSD will outperform sync writes than any amount of expensive 
spindle drives. Before ZIL software is was easy to argue that the only way of 
speeding up writes was more faster spindles.

 

The question therefore is, is there room in the software implementation to 
achieve performance and reliability numbers similar to expensive drives whilst 
using relative cheap drives?

 

ZFS is good but IMHO easy to see how it can be improved to better meet this 
situation, I can’t currently say when this line of thinking and code will move 
from research to production level use (tho I have a pretty good idea ;) ) but I 
wouldn’t bet on the status quo lasting much longer. In some ways the removal of 
OpenSolaris may actually be a good thing, as its catalyized a number of 
developers from the view that zfs is Oracle led, to thinking “what can we do 
with zfs code as a base”?

 

Ffor example how about sticking a cheap 80GiB commodity SSD in the storage 
case. When a resilver or defrag is required, use it as a scratch space to give 
you a block of fast IOPs storage space to accelerate the slow parts. When its 
done secure erase and power it down, ready for the next time a resilver needs 
to happen. The hardware is available, just needs someone to write the software…

 

 

Bye,

Deano

 

 
 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] stupid ZFS question - floating point operations

2010-12-22 Thread Deano
There are no floating points operations in zfs, however even if there would
that wouldn't be a bad thing, as modern CPU are float monsters indeed its
likely some things would be faster if converted to use the float ALU (note
however those operations would have to account for the different properties
of the ALUs).

The most complex ALU parts of zfs are the Galois field math in the RAID
codes and the checksum operations. All of which are integer ops (tho it
might be interesting to see if the Galois field could be accelerated using
the float ALU).

Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Jerry Kemp
Sent: 22 December 2010 19:44
To: zfs-discuss@opensolaris.org
Subject: [zfs-discuss] stupid ZFS question - floating point operations

I have a coworker, who's primary expertise is in another flavor of Unix.

This coworker lists floating point operations as one of ZFS detriments.

I's not really sure what he means specifically, or where he got this
reference from.

In an effort to refute what I believe is an error or misunderstanding on
his part, I have spent time on Yahoo, Google, the ZFS section of
OpenSolaris.org, etc.  I really haven't turned up much of anything that
would prove or disprove his comments.  The one thing I haven't done is
to go through the ZFS source code, but its been years since I have done
any serious programming.

If someone from Oracle, or anyone on this mailing list could point me
towards any documentation, or give me a definitive word, I would sure
appreciate it.  If there were floating point operations going on within
ZFS, at this point I am uncertain as to what they would be.

TIA for any comments,

Jerry
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] stupid ZFS question - floating point operations

2010-12-22 Thread Deano


-Original Message-
From: Peter Jeremy [mailto:peter.jer...@alcatel-lucent.com] 
Sent: 22 December 2010 21:17
To: Deano
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] stupid ZFS question - floating point operations

On 2010-Dec-23 04:48:19 +0800, Deano  wrote:
> modern CPU are float monsters indeed its likely some things would be 
>faster if converted to use the float ALU
Peter wrote
> _Some_ modern CPUs are good at FP, a lot aren't.  The SPARC T-1 was
particularly
> poor as it only had a single FPU.  Likewise, performance in the x86 world
is highly
> variable, depending on the vendor and core you pick.  AFAIK, iA64 and PPC
are
> consistently good - but neither are commonly found in conjunction with
ZFS.  You
> may also need to allow for software assist:  Very few CPUs implement all
of the
> IEEE FP standard in hardware and most (including SPARC) require software
to
> implement parts of the standard.  If your algorithm happens to make
significant use
> of things other than normalised numbers and zero, your performance may be
severely
> affected by the resultant traps and software assistance.
I can't speak for old architecture like SPARC but all modern ALU designs
support most of the subset of useful IEEE in hardware and at high speed.
In particular x86 has extremely good float ALU performance, compared to some
architectures it is relatively low but certainly not something to avoid.

A CPU design that isn't at *least* 5 GFLOPS per core is archaic. This is
only accelerating due to the consumer market that many CPUs end up in. From
graphics to video decoding to audio synthesis, floating point math
dominants.

Not to say they aren't able to perform very many integer ALU ops as well,
just that the old mantra that FPU is to be avoided hasn't been true for
years.

> Any use of floating point within the kernel also means changes to when FPU
context
> is saved - and, unless this can be implemented lazily, it will adversely
impact the
> cost of all context switches and potentially system calls.
Of course the cost of the extra register movement involved in context
switches is a concern, but this cost can be evaluated against the gains. I
like to see someone actually profile the costs in SunOs as many kernel
architectures I know accept FPU (and other specialist registers) restoration
when needed as a worthwhile cost.

Bye,
Deano

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Looking for 3.5" SSD for ZIL

2010-12-23 Thread Deano
If anybody does know of any source to the secure erase/reformatters, I'll
happily volunteer to do the port and then maintain it.


I'm currently in talks with several SSD and driver chip hardware peeps with
regard getting datasheets for some SSD products etc. for the purpose of
better support under the OI/Solaris driver model but these things can take a
while to obtain, so if anybody knows of existing open source versions I'll
jump on it.

 

Thanks,

Deano

 

From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Garrett D'Amore
Sent: 23 December 2010 15:22
To: Erik Trimble; Christopher George
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Looking for 3.5" SSD for ZIL

 

 

We should get the reformatter(s) ported to illumos/solaris, if source is
available.  Something to consider.

  - Garrett

-Original Message-
From: zfs-discuss-boun...@opensolaris.org on behalf of Erik Trimble
Sent: Wed 12/22/2010 10:36 PM
To: Christopher George
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Looking for 3.5" SSD for ZIL

On 12/22/2010 7:05 AM, Christopher George wrote:
>> I'm not sure if TRIM will work with ZFS.
> Neither ZFS nor the ZIL code in particular support TRIM.
>
>> I was concerned that with trim support the SSD life and
>> write throughput will get affected.
> Your concerns about sustainable write performance (IOPS)
> for a Flash based SSD are valid, the resulting degradation
> will vary depending on the controller used.
>
> Best regards,
>
> Christopher George
> Founder/CTO
> www.ddrdrive.com

Christopher is correct, in that SSDs will suffer from (non-trivial)
performance degredation after they've exhausted their free list, and
haven't been told to reclaim emptied space.  True battery-backed DRAM is
the only permanent solution currently available which never runs into
this problem.  Even TRIM-supported SSDs eventually need reconditioning.

However, this *can* be overcome by frequently re-formatting the SSD (not
the Solaris format, a low-level format using a vendor-supplied
utility).  It's generally a simple thing, but requires pulling the SSD
from the server, connecting it to either a Linux or Windows box, running
the reformatter, then replacing the SSD.  Which, is a PITA.

But, still a bit cheaper than buying a DDRdrive. 


--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Looking for 3.5" SSD for ZIL

2010-12-23 Thread Deano
In an ideal world, if we could obtain details on how to reset/format blocks of 
a SSD, we could do it automatically running behind the ZIL. As a log its going 
in one direction, a background task could clean up behind it, making the 
performance lowing over time a non-issue for the ZIL. A first start may be 
calling unmap/trim on those blocks (which I was surprised to find in the source 
is already coded up in the SATA driver, just not used yet) but really a reset 
would be better.

But as you say a tool to say if its need doing would be a good start. They 
certainly exist in closed source form...

Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org 
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Ray Van Dolson
Sent: 23 December 2010 15:46
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Looking for 3.5" SSD for ZIL

On Thu, Dec 23, 2010 at 07:35:29AM -0800, Deano wrote:
> If anybody does know of any source to the secure erase/reformatters,
> I’ll happily volunteer to do the port and then maintain it.
> 
> I’m currently in talks with several SSD and driver chip hardware
> peeps with regard getting datasheets for some SSD products etc. for
> the purpose of better support under the OI/Solaris driver model but
> these things can take a while to obtain, so if anybody knows of
> existing open source versions I’ll jump on it.
> 
> Thanks,
> Deano

A tool to help the end user know *when* they should run the reformatter
tool would be helpful too.

I know we can just wait until performance "degrades", but it would be
nice to see what % of blocks are in use, etc.

Ray
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Looking for 3.5" SSD for ZIL

2010-12-23 Thread Deano
Secure Erase is currently a entire drive function, its writes all the cell
resetting it. It also updates the firmware GC maps so it knows the drive is
clean. Trim just gives more info the firmware that a block is unused (as
normally a delete is just updating an index table and the firmware has no
way of knowing which cells are no longer needed by the OS).

Currently firmware is meant to help conventional file system usage. However
ZIL isn't normal usage and as such *IF* and it's a big if, we can
effectively bypass the firmware trying to be clever or at least help it be
clever then we can avoid the downgrade over time. In particular if we could
secure erase a few cells as once as required, the lifetime would be much
longer, I'd even argue that taking the wear leveling off the drives hand
would be useful in the ZIL case. The other thing is the slow down only
occurs once the SSD fills and has to start getting clever where to put
things and which cells to change, for a ZIL that is again something we could
avoid in software fairly easy.

Its also worth putting this in perspective, a complete secure erase every
night to restore performance to your ZIL would still let the SSD last for
*years*. And given how cheap some SSD are, it is probably still cheaper to
effectively burn the ZIL out and just replace it once a year. Maybe not a
classic level of RAID but the very essence of the idea, lots of cheap can be
better than expensive if you know what you are doing.

Bye,
Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Christopher George
Sent: 23 December 2010 16:46
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Looking for 3.5" SSD for ZIL

> However, this *can* be overcome by frequently re-formatting the SSD (not 
> the Solaris format, a low-level format using a vendor-supplied utility).

For those looking to "Secure Erase" a OCZ SandForce based SSD to reclaim 
performance, the following OCZ Forum thread might be of interest:

http://www.ocztechnologyforum.com/forum/showthread.php?75773-Secure-Erase-TR
IM-and-anything-else-Sandforce

OCZ uses the term "DuraClass" as a catch-all for algorithms controlling wear

leveling, drive longevity...  There is a direct correlation between Secure
Erase 
frequency and expected SSD lifetime.

Thread #1 detailing a recommended frequency of Secure Erase use:

"3) Secure erase a drive every 6 months to free up previously read only 
blocks, secure erase every 2 days to get round Duraclass and you will kill
the 
drive very quickly"

Thread #5 explaining DuraClass and relationship to TRIM:

"Duraclass is limiting the speed of the drive NOT TRIM. TRIM is used along 
with wear levelling."

Thread #6 provides more details of DuraClass and TRIM:

"Now Duraclass monitors all writes and control's encryption and compression,

this is what effects the speed of the blocks being written to..NOT the fact
they 
have been TRIM'd or not TRIM'd."

"You guys have become fixated at TRIM not speeding up the drive and forget 
that Duraclass controls all writes incurred by the drive once a GC map has 
been written."

Above excerpts written by a OCZ employed thread moderator (Tony).

Best regards,

Christopher George
Founder/CTO
www.ddrdrive.com
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2011-01-05 Thread Deano
Edward Ned Harvey wrote
> I don't know if anyone has real numbers, dollars contributed or number of
> developer hours etc, but I think it's fair to say that oracle is probably
> contributing more to the closed source ZFS right now, than the rest of the
> world is contributing to the open source ZFS right now.  Also, we know
that
> the closed source ZFS right now is more advanced than the open source ZFS
> (zpool 31 vs 28).  Oracle closed source ZFS is ahead, and probably
> developing faster too, than the open source ZFS right now.

> If anyone has any good way to draw more contributors into the open source
> tree, that would also be useful and appreciated.  Gosh, it would be nice
to
> get major players like Dell, HP, IBM, Apple contributing into that
project.

This is something that Illumos/Open source ZFS needs to decide what it
wants, effectively we can't innovate ZFS without breaking capability...
because our Illumos ZPool version 29 (if we innovate) will not be Oracle
Zpool version 29.

If we want open-source ZFS to we need to make that choice and let everyone
know, apart from submitting bug fixes to zpool v28, are I'm not sure if
other changed would be welcome?

So honestly do we want to innovate ZFS (I do) or do we just want to follow
Oracle?

Bye,
Deano

de...@cloudpixies.com


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2011-01-05 Thread Deano

Edward Ned Harvey wrote
> From: Deano [mailto:de...@rattie.demon.co.uk]
> Sent: Wednesday, January 05, 2011 9:16 AM
> 
> So honestly do we want to innovate ZFS (I do) or do we just want to follow
> Oracle?

> Well, you can't follow Oracle.  Unless you wait till they release
something,
> reverse engineer it, and attempt to reimplement it.  I am quite sure
you'll
> be sued if you do that.
>
> If you want forward development in the open source tree, you basically
have
> only one option:  Some major contributor must have a financial interest,
and
> commit to a real concerted development effort, with their own roadmap,
which
> is intentionally designed NOT to overlap with the Oracle roadmap.
> Otherwise, the code will stagnate.

Why does it need a big backer? Erm ZFS isn't that large or amazingly complex
code. It is *good* code but take 100s of developers and a fortune to
develop? Erm nope (which I'd bet it never had at Sun either).

Why not overlap Oracle? what has it got to do with Oracle if we have split
into ZFS (Oracle) and "OpenZFS" in future. "OpenZFS" will get whatever
features developers feel that want or they need to develop for it.

This is the fundamental choice of Open source ZFS, illumos and OpenIndiania
(and other distributions) have to decide, what is there purpose? Is it a
free compatible (though trailing) version of Oracle Solaris OR a platform
that shared an ancestor with Oracle Solaris via Sun OpenSolaris but now is
its own evolutionary species, with no more connection than I have with a
15th cousin removed on my great, great, great, grandfathers side.

This isn't even a theoretical what if situation for me, I have a major
modification to ZFS (still being developed), it has no basis on Oracle or
anybody elses needs just mine. It is what I felt I needed and ZFS was the
right base for it. Now will that go into "OpenZFS"? Honestly I don't know
yet, because not sure it would be wanted (it will be incompatible with
Oracle ZFS) and personally, commercially I'm not sure if it's the right move
to open source the feature.

I bet I'm not the only small developer out there in a similar situation, the
landscape is very unclear about what actually the community wants to do
going forward, and whether we will have or even want "OpenZFS" and Oracle
ZFS or Oracle ZFS and 90% compatibles (always trailing) or Oracle ZFS + DevA
ZFS + DevB ZFS + DevC ZFS.

Bye,
Deano

de...@cloudpixies.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zil and root on the same SSD disk

2011-01-12 Thread Deano
>From my notes from mirroring a new install (I install first then mirror).
You won't need pfexec if your super-user.
Inside format, fdisk twice first time delete anything there, second time it
will ask you if you want to install a default Solaris2 setup.
Obviously change the disk id to match your system.

pfexec format
fdisk (delete if required) and install standard Solaris2
pfexec prtvtoc /dev/rdsk/c1d0s2 | pfexec fmthard -s - /dev/rdsk/c2d0s2
pfexec zpool attach -f rpool c1d0s0 c2d0s0

I've never put a ZIL on the same disk, so can't help there sorry.
HTH,
Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Per Jorgensen
Sent: 12 January 2011 20:51
To: zfs-discuss@opensolaris.org
Subject: [zfs-discuss] zil and root on the same SSD disk

Hi

got a brand new server with 14 x 2TB disk, and 2X160GB SSD disk , my plan
was to install opensolaris on one of the SSD disk and then zfs mirror the
root disk onto the second SSD disk , but since the server will handle some
sync NFS write i also want to add a zil log on the same SSD disks, also
mirrored. I don't have in depth knowledge on solaris or opensolaris , and
don't know much about how partitions works on solaris. I have tried the last
2 days solving this , and reading howtos on the net , but have not been able
to find any  step by step howto explaining how I should do it. :-(

I have tried creating on big solaris2 partition , and then used 140G to the
system , and then make a new slice the my zil , witch looked to work , at
least i could add the 20G slice as log device to my pool , but then the
problem came when i tried to to clone the partition map from the system SSD
to the second SSD disk with this command

prtvtoc /dev/rdsk/c7t0d0s2 | fmthard -s - /dev/rdsk/c7t1d0s2

it said something about overlapping partitions/slices ( unfortunately i did
not save the output of the error )

but what is the right way to do this , should i create one big solaris2
partitions and then slice it up or should I create 2 separate partitions ,
and what is the right way to mirror both the system disk and zil log to the
second SSD disk. And is i possible to do the partitioning from the installer
? ( I am installing opensolaris from the Live CD)

please , hoping someone could provide me with a nice step by step howto on
this , or maybe guide be through. 

br
Per Jorgensen
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] mix 4k wdd drives whit 512 WD drives ?

2011-01-26 Thread Deano
If 512B drive is used with 4KiB ashift, there may be a performance lost (as
more data will be pushed and read from the drive).

It not so much the drives are lying, just they don't implement the ATAPI
standard (for SATA) to tell the OS they are really 4KiB drives. A firmware
update would fix this. The modern OpenSolaris children (S11E, OpenIndiana,
etc.) all read the physical size correctly when its provided via the drive.
Illumos will be providing support at the driver level via a driver look up
table quite soon for drives that don't tell the OS the physical block size
and I would expect other OS's will also hide the problem at the driver level
going forward.

For now zpool pool fixes are a reasonable work-around. The previous linked
zpool forces 4KiB which might not be ideal if you are also using non 4KiB
drives. A link to a zpool with a command line block-size is on the
OpenIndiana wiki at this link
http://wiki.openindiana.org/oi/SATA+-+Advanced+Format+and+4K+Sector+drives

HTH,
Deano
de...@cloudpixies.com

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Tomas Ögren
Sent: 26 January 2011 17:01
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] mix 4k wdd drives whit 512 WD drives ?

On 26 January, 2011 - Benji sent me these 0,8K bytes:

> Those WD20EARS emulate 512 bytes sectors, so yes you can freely mix
> and match them with other "regular" 512 bytes drives. Some have
> reported slower read/write speeds but nothing catastrophic.

For some workloads, 3x slower than it should be.

> Or you can create a new 4K aligned pool (composed of only 4K drives!)
> to really take advantage of them. For that, you will need a modified
> zpool command to sets the ashift value of the pool to 12.

A 4k aligned pool will work perfectly on a 512b aligned disk, it's just
the other way that's bad. I guess ZFS could start defaulting to 4k, but
ideally it should do the right thing depending on content (although
that's hard for disks that are lying).

/Tomas
-- 
Tomas Ögren, st...@acc.umu.se, http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Lower latency ZIL Option?: SSD behind Controller BB Write Cache

2011-01-28 Thread Deano
Hi Edward,
Do you have a source for the 8KiB block size data? whilst we can't avoid the
SSD controller in theory we can change the smallest size we present to the
SSD to 8KiB fairly easily... I wonder if that would help the controller do a
better job (especially with TRIM)

I might have to do some test, so far the assumption (even inside sun's sd
driver) is that SSD are really 4KiB even when the claim 512B, perhaps we
should have an 8KiB option...

Thanks,
Deano
de...@cloudpixies.com

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Edward Ned Harvey
Sent: 28 January 2011 13:25
To: 'Eff Norwood'; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Lower latency ZIL Option?: SSD behind Controller
BB Write Cache

> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Eff Norwood
> 
> We tried all combinations of OCZ SSDs including their PCI based SSDs and
> they do NOT work as a ZIL. After a very short time performance degrades
> horribly and for the OCZ drives they eventually fail completely. 

This was something interesting I found recently.  Apparently for flash
manufacturers, flash hard drives are like the pimple on the butt of the
elephant. A vast majority of the flash production in the world goes into
devices like smartphones, cameras, tablets, etc.  Only a slim minority goes
into hard drives.  As a result, they optimize for these other devices, and
one of the important side effects is that standard flash chips use an 8K
page size.  But hard drives use either 4K or 512B.  

The SSD controller secretly remaps blocks internally, and aggregates small
writes into a single 8K write, so there's really no way for the OS to know
if it's writing to a 4K block which happens to be shared with another 4K
block in the 8K page.  So it's unavoidable, and whenever it happens, the
drive can't simply write.  It must read modify write, which is obviously
much slower.

Also if you look up the specs of a SSD, both for IOPS and/or sustainable
throughput...  They lie.  Well, technically they're not lying because
technically it is *possible* to reach whatever they say.  Optimize your
usage patterns and only use blank drives which are new from box, or have
been fully TRIM'd.  Pt...  But in my experience, reality is about 50% of
whatever they say.

Presently, the only way to deal with all this is via the TRIM command, which
cannot eliminate the read/modify/write, but can reduce their occurrence.
Make sure your OS supports TRIM.  I'm not sure at what point ZFS added TRIM,
or to what extent...  Can't really measure the effectiveness myself.

Long story short, in the real world, you can expect the DDRDrive to crush
and shame the performance of any SSD you can find.  It's mostly a question
of PCIe slot versus SAS/SATA slot, and other characteristics you might care
about, like external power, etc.



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS and TRIM

2011-01-29 Thread Deano
Whilst the driver supports TRIM, ZFS doesn't yet. So in practice it's not
supported.

Bye,
Deano
de...@cloudpixies.com


-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Brandon High
Sent: 29 January 2011 18:40
To: Edward Ned Harvey
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] ZFS and TRIM

On Sat, Jan 29, 2011 at 8:31 AM, Edward Ned Harvey
 wrote:
> What is the status of ZFS support for TRIM?

I believe it's been supported for a while now.
http://www.c0t0d0s0.org/archives/6792-SATA-TRIM-support-in-Opensolaris.html

-B

-- 
Brandon High : bh...@freaks.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS and TRIM

2011-02-04 Thread Deano
Even without TRIM and lots of use, SSD are still likely to help as ZIL and
L2ARC cache units better than spindle disks, however don't expect the same
performance you got after a fresh wipe/install.

It makes sense to go with the brands with the best garbage collector you can
and also if you can leave more space unused, that helps.

Bye,
Deano


-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Orvar Korvar
Sent: 04 February 2011 13:20
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] ZFS and TRIM

So, the bottom line is that Solaris 11 Express can not use TRIM and SSD? Is
that the conclusion? So, it might not be a good idea to use a SSD?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS/Drobo (Newbie) Question

2011-02-07 Thread Deano
In zfs terminology each of the groups you have is a VDEV and a zpool can be
made of a number of VDEVs. This zpool can then be mounted as a single
filesystem, or you can split it into as many filesystems as you wish.

So the answer is yes to all the configurations you asked about and a lot
more :)

Bye,
Deano
de...@cloudpixies.com

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Gaikokujin
Kyofusho
Sent: 05 February 2011 17:55
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] ZFS/Drobo (Newbie) Question

Thank you kebabber. I will try out indiana and virtual box to play around
with it a bit.

Just to make sure I understand your example, if I say had a 4x2tb drives,
2x750gb, 2x1.5tb drives etc then i could make 3 groups (perhaps 1 raidz1 + 1
mirrored + 1 mirrored), in terms of accessing them would they just be
mounted like 3 partitions or could it all be accessed like one big
partition?

Anywho, I have indiana DL'ing now (very slow connection so thought I would
post while i wait).

Cheers,

-Gaiko
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SIL3114 and sparc solaris 10

2011-02-23 Thread Deano
That card works on OI-148 x86 if its SATA mode (tested it myself, RAID does
not), which should also work on the SPARC version of the OS (common driver).

Would mean upgrading to OpenIndiana and using the in development SPARC SKU
but if you're adventurous and nothing else works...

Bye,
Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Mauricio Tavares
Sent: 23 February 2011 13:18
To: zfs-discuss@opensolaris.org
Subject: [zfs-discuss] SIL3114 and sparc solaris 10

Perhaps a bit off-topic (I asked on the rescue list -- 
http://web.archiveorange.com/archive/v/OaDWVGdLhxWVWIEabz4F -- and was 
told to try here), but I am kinda shooting in the dark: I have been 
finding online scattered and vague info stating that this card can be 
made to work with a sparc solaris 10 box 
(http://old.nabble.com/eSATA-or-firewire-in-Solaris-Sparc-system-td27150246.
html 
is the only link I can offer right now). Can anyone confirm or deny that?
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] best migration path from Solaris 10

2011-03-19 Thread Deano
Nexenta are a great company (I'm no way affiliated with them btw), if for no
other reason being willing to invest in Illumos and by that OpenIndiana and
NCP (for which they charge nothing). If you need a large enterprise
commercially backed storage server system, NextentaStor is the answer.

If you want a CLI OS, NCP or OpenIndiana text only will fulfill that spot
(there is also illumos-extra which is even more stripped down to bare
minimum), if a GUI or desktop is required OpenIndiana has that covered.

Whilst there isn't yet official commercial support for OpenIndiana, that is
something that has been brought up and the feeling is that it is something
that would like to be offered at some point in time, it is just the
logistics and organization of that support has to be setup. If it became a
deal breaker I'd suggest talking to a senior person of OI, as it possible
something could be arranged.

I expect we are no more than 6 months away from Illumos being the defacto
open source foundation for all the major distributions, which will then
start freeing up a lot of developer resources to continue and improve
things, beyond this initial get everything together period.

ZFS open source future is being hammered out, with hopefully a cross
platform working body to promote and move it forward appearing. I know
everyone involved is both eager to make sure it's truly open and portable
(with support for Illumos/OI, FreeBSD, Linux and OS-X in working or
in-progress) and that it has a future regardless of any code drops from
Oracle (they are of course welcome if they come, but clearly at this point
it can't be relied upon).

Personally I think OpenIndiana with Illumos foundations, is a great
enterprise quality open source OS with a great future and a bunch of really
committed guys and gals behind it.

My 2P,
Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Garrett D'Amore
Sent: 18 March 2011 22:15
To: Paul B. Henson
Cc: openindiana-disc...@openindiana.org; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] best migration path from Solaris 10

Thanks for thinking about us, Paul.

A few quick thoughts:

a) Nexenta Core Platform is a bare-bones OS.  No GUI, in other words (no
X11.)  It might well suit you.

b) NCP 3 will not have an upgrade path to NCP 4.  Its simply too much
change in the underlying packaging.

c) NCP 4 is still 5-6 months away.  We're still developing it.

d) NCP 4 will make much more use of the illumos userland, and only use
Debian when illumos doesn't have an equivalent.

e) NCP comes entirely unsupported.  NexentaStor is a commercial product
with real support behind it, though.

f) *Today*, NexentaStor 3 has newer code in it than NCP.  That will be
changing, as we will be keeping the two much more closely in sync
starting with 3.1.

g) If you want to self support, OpenIndiana or NCP are both good
options.  NCP has debian packaging, and lacks a bunch of the GUI
goodies.  NCP 3 is not as new as OI, but is probably a bit more proven.

Hopefully the additional information is helpful to you.

- Garrett


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [OpenIndiana-discuss] best migration path from Solaris 10

2011-03-19 Thread Deano
That’s not quite right, Illumos is open source continuation of ONNV, which
is the core foundation, however it doesn't include other consolidation that
made up OpenSolaris.

OpenIndiana does, it takes all those consolidations and produces a working
OS you can install. Of course the biggest and most important of those
consolidation is Illumos. The reason OI is only now slowly moving to
Illumos, it has many other consolidations that also required work before a
safe move could be made, and also keeping OpenIndiana in line with the
closed source variants.

There are other forks for both areas (ONNV replacements and installable OS
distributions), if for some reason Illumos and OpenIndiana aren't suitable.

HTH,
Deano



-Original Message-
From: Pasi Kärkkäinen [mailto:pa...@iki.fi] 
Sent: 19 March 2011 14:58
To: Michael DeMan
Cc: openindiana-disc...@openindiana.org; zfs-discuss@opensolaris.org
Subject: Re: [OpenIndiana-discuss] [zfs-discuss] best migration path from
Solaris 10

On Fri, Mar 18, 2011 at 06:26:37PM -0700, Michael DeMan wrote:
> ZFSv28 is in HEAD now and will be out in 8.3.
> 
> ZFS + HAST in 9.x means being able to cluster off different hardware.
> 
> In regards to OpenSolaris and Indiana - can somebody clarify the
relationship there?  It was clear with OpenSolaris that the latest/greatest
ZFS would always be available since it was a guinea-pig product for cost
conscious folks and served as an excellent area for Sun to get marketplace
feedback and bug fixes done before rolling updates into full Solaris.
> 
> To me it seems that Open Indiana is basically a green branch off of a dead
tree - if I am wrong, please enlighten me.
> 

Illumos project was started as a fork of OpenSolaris when Oracle was still
publishing OpenSolaris sources.

Then Oracle closed OpenSolaris development, and decided to call upcoming
(closed) versions "Solaris 11 Express",
with no source included.

Illumos project continued the development based on the latest published
OpenSolaris sources, 
and a bit later OpenIndiana *distribution* was announced to deliver a binary
distro based on OpenSolaris/Illumos.

So in short Illumos is the development project, which hosts the new sources,
and OpenIndiana is a binary distro based on it.


-- Pasi

> On Mar 18, 2011, at 6:16 PM, Roy Sigurd Karlsbakk wrote:
> 
> >> I think we all feel the same pain with Oracle's purchase of Sun.
> >> 
> >> FreeBSD that has commercial support for ZFS maybe?
> > 
> > Fbsd currently has a very old zpool version, not suitable for running
with SLOGs, since if you lose it, you may lose the pool, which isn't very
amusing...
> > 
> > Vennlige hilsener / Best regards
> > 
> > roy
> > --
> > Roy Sigurd Karlsbakk
> > (+47) 97542685
> > r...@karlsbakk.net
> > http://blogg.karlsbakk.net/
> > --
> > I all pedagogikk er det essensielt at pensum presenteres intelligibelt.
Det er et elementært imperativ for alle pedagoger å unngå eksessiv
anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller
eksisterer adekvate og relevante synonymer på norsk.
> 
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
OpenIndiana-discuss mailing list
openindiana-disc...@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] best migration path from Solaris 10

2011-03-23 Thread Deano
OpenIndiana and others (i.e. Benunix) are distributions that actively
support full desktop workstations based on the Illumos base.

It is true, that the storage server application is a popular one and so has
supporters both commercially and others. ZFS is amazing and quite rightly it
stands out, it works even better when used with zones, crossbow, dtrace,
etc. and so its obvious to see what it's a focus and often seems the only
priority. 

However is isn't the only interest, by a long shot.

The SFE package repositories has many packages available to install for when
the binary packaging aren't up to date. OpenIndiana is hard at work trying
to build bigger binary repositories with more apps and newer versions.
A simple "pkg install APPLICATION" is the aim for the majority of main
applications.

Is it not moving fast enough, or missing the packages you need?
Well that's the beauty of Open Source, we welcome and have systems to help
newcomers add and update the packages and applications they want, so we all
benefit. 

Ultimately I'd (and I'm sure many would) like to have a level of binary
repositories similar to Debian, with stable and faster changing place repos
and support for many different applications, however that requires a lot of
work and manpower.

Bye,
Deano

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Fajar A. Nugraha
Sent: 23 March 2011 01:09
To: Jeff Bacon
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] best migration path from Solaris 10

On Wed, Mar 23, 2011 at 7:33 AM, Jeff Bacon 
wrote:
>> I've also started conversations with Pogo about offering an
> OpenIndiana
>> based workstation, which might be another option if you prefer more of

> Sometimes I'm left wondering if anyone uses the non-Oracle versions for
> anything but file storage... ?

Seeing that userland programs for *Solaris and derivatives (GUI,
daemons, tools, etc) is usually late compared to bleeding-edge Linux
distros (e.g. Ubuntu), with no particular dedicated team working on
improvement there, I'm guessing the answer will be "highly unlikely".

-- 
Fajar
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [illumos-Developer] ZFS working group and feature flags proposal

2011-05-25 Thread Deano

Hi Matt,

That's looks really good, I've been meaning to implement a ZFS compressor
(using a two pass, LZ4 + Arithmetic Entropy), so nice to see a route with
which this can be done.

One question, is the extendibility of RAID and other similar systems, my
quick perusal makes me thinks this is handled by simple asserting a new
feature using the extension mechanism, but perhaps I've missed something? Do
you see it being able to handle this situation?
Its of course a slightly tricky one, as not only does it change data but
potentially data layout as well...

Great work ZFS working group :) Nice to see ZFS's future coming together!

Bye,
Deano

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss