> Not to be picky, but the X2100 and X2200 series are
> NOT
> designed/targeted for disk serving (they don't even
> have redundant power
> supplies).  They're compute-boxes.  The X4100/X4200
> are what you are
> looking for to get a flexible box more oriented
> towards disk i/o and
> expansion.

I don't see those as being any better suited to external discs other than:

#1 - They have the capacity for redundant PSUs, which is irrelevant to my needs.
#2 - They only have PCI Express slots, and I can't find any good external SATA 
interface cards on PCI Express

I can't wrap my head around the idea that I should buy a lot more than I need, 
which still doesn't serve my purposes. The 4 disks in an x4100 still aren't 
enough, and the machine is a fair amount more costly. I just need mirrored boot 
drives, and an external disk array.

> That said (if you're set on an X2200 M2), you are
> probably better off
> getting a PCI-E SCSI controller, and then attaching
> it to an external
> SCSI->SATA JBOD.  There are plenty of external JBODs
> out there which use
> Ultra320/Ultra160 as a host interface and SATA as a
> drive interface.
> Sun will sell you a supported SCSI controller with
> the X2200 M2 (the
> "Sun StorageTek PCI-E Dual Channel Ultra320 SCSI
> HBA").
>
> SCSI is far better for a host attachment mechanism
> than eSATA if you
> plan on doing more than a couple of drives, which it
> sounds like you
> are. While the SCSI HBA is going to cost quite a bit
> more than an eSATA
> HBA, the external JBODs run about the same, and the
> total difference is
> going to be $300 or so across the whole setup (which
> will cost you $5000
> or more fully populated). So the cost to use SCSI vs
> eSATA as the host-
> attach is a rounding error.

I understand your comments in some ways, in others I do not. It sounds like 
we're moving backwards in time. Exactly why is SCSI "better" than SAS/SATA for 
external devices? From my experience (with other OSs/hardware platforms) the 
opposite is true. A nice SAS/SATA controller with external ports (especially 
those that allow multiple SAS/SATA drives via one cable - whichever tech you 
use) works wonderfully for me, and I get a nice thin/clean cable which makes 
cable management much more "enjoyable" in higher density situations.
 
I also don't agree with the logic "just spend a mere $300 extra to use older 
technology!"

$300 may not be much to large business, but things like this nickle and dime 
small business owners. There's a lot of things I'd prefer to spend $300 on than 
an expensive SCSI HBA which offers no advantages over a SAS counterpart, in 
fact offers disadvantages instead. 

Your input is of course highly valued, and it's quite possible I'm missing an 
important piece to the puzzle somewhere here, but I am not convinced this is 
the ideal solution - simply a "stick with the old stuff, it's easier" solution, 
which I am very much against.

Thanks,
David
 
 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to