On Wed, 1 Oct 2014 09:28:12 +0200 Kasper Dieter wrote:

> On Tue, Sep 30, 2014 at 04:38:41PM +0200, Mark Nelson wrote:
> > On 09/29/2014 03:58 AM, Dan Van Der Ster wrote:
> > > Hi Emmanuel,
> > > This is interesting, because we?ve had sales guys telling us that
> > > those Samsung drives are definitely the best for a Ceph journal O_o !
> > 
> > Our sales guys or Samsung sales guys?  :)  If it was ours, let me know.
> > 
> > > The conventional wisdom has been to use the Intel DC S3700 because
> > > of its massive durability.
> > 
> > The S3700 is definitely one of the better drives on the market for
> > Ceph journals.  Some of the higher end PCIE SSDs have pretty high
> > durability (and performance) as well, but cost more (though you can
> > save SAS bay space, so it's a trade-off).
> Intel P3700 could be an alternative with 10 Drive-Writes/Day for 5 years
> (see attachment)
> 
They're certainly nice and competitively priced (TBW/$ wise at least). 
However as I said in another thread, once your SSDs start to outlive your
planned server deployment time (in our case 5 years) that's probably good
enough.

It's all about finding the balance between cost, speed (BW and IOPS),
durability and space.

For example I'm currently building a cluster based on 2U, 12 hotswap bays
servers (because I already had 2 floating around) and am using 4 100GB DC
S3700 (at US$200 each) and 8 HDDS in them. 
Putting in a 400GB DC P3700 (US$1200( instead and 4 more HDDs would have
pushed me over the budget and left me with a less than 30% "used" SSD 5
years later, at a time when we clearly can expect these things to be
massively faster and cheaper.

Now if you're actually having a cluster that would wear out a P3700 in 5
years (or you're planning to run your machines until they burst into
flames), then that's another story. ^.^

Christian

> -Dieter
> 
> > 
> > >
> > > Anyway, I?m curious what do the SMART counters say on your SSDs??
> > > are they really failing due to worn out P/E cycles or is it
> > > something else?
> > >
> > > Cheers, Dan
> > >
> > >
> > >> On 29 Sep 2014, at 10:31, Emmanuel Lacour <elac...@easter-eggs.com>
> > >> wrote:
> > >>
> > >>
> > >> Dear ceph users,
> > >>
> > >>
> > >> we are managing ceph clusters since 1 year now. Our setup is
> > >> typically made of Supermicro servers with OSD sata drives and
> > >> journal on SSD.
> > >>
> > >> Those SSD are all failing one after the other after one year :(
> > >>
> > >> We used Samsung 850 pro (120Go) with two setup (small nodes with 2
> > >> ssd, 2 HD in 1U):
> > >>
> > >> 1) raid 1 :( (bad idea, each SSD support all the OSDs journals
> > >> writes :() 2) raid 1 for OS (nearly no writes) and dedicated
> > >> partition for journals (one per OSD)
> > >>
> > >>
> > >> I'm convinced that the second setup is better and we migrate old
> > >> setup to this one.
> > >>
> > >> Thought, statistics gives 60GB (option 2) to 100 GB (option 1)
> > >> writes per day on SSD on a not really over loaded cluster. Samsung
> > >> claims to give 5 years warranty if under 40GB/day. Those numbers
> > >> seems very low to me.
> > >>
> > >> What are your experiences on this? What write volumes do you
> > >> encounter, on wich SSD models, which setup and what MTBF?
> > >>
> > >>
> > >> --
> > >> Easter-eggs                              Spécialiste GNU/Linux
> > >> 44-46 rue de l'Ouest  -  75014 Paris  -  France -  Métro Gaité
> > >> Phone: +33 (0) 1 43 35 00 37    -   Fax: +33 (0) 1 43 35 00 76
> > >> mailto:elac...@easter-eggs.com  -   http://www.easter-eggs.com
> > >> _______________________________________________
> > >> ceph-users mailing list
> > >> ceph-users@lists.ceph.com
> > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > >
> > > _______________________________________________
> > > ceph-users mailing list
> > > ceph-users@lists.ceph.com
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > >
> > 
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


-- 
Christian Balzer        Network/Systems Engineer                
ch...@gol.com           Global OnLine Japan/Fusion Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to