I had a question about one of the points on the slide. Slide 24, last bullet point, says:

 * If you use XFS, don't put your OSD journal as a file on the disk
     o Use a separate partition, the first partition!
     o We still need to reinstall our whole cluster to re-partition the
       OSDs


Do you have any details on that?

I am doing that, mostly because my rados bench test indicated that it was faster than using a journal partition. Now I'm having time-out issues with OSD during recovery, when there really shouldn't be any, so I'm curious.


Thanks for any info.


*Craig Lewis*
Senior Systems Engineer
Office +1.714.602.1309
Email cle...@centraldesktop.com <mailto:cle...@centraldesktop.com>

*Central Desktop. Work together in ways you never thought possible.*
Connect with us Website <http://www.centraldesktop.com/> | Twitter <http://www.twitter.com/centraldesktop> | Facebook <http://www.facebook.com/CentralDesktop> | LinkedIn <http://www.linkedin.com/groups?gid=147417> | Blog <http://cdblog.centraldesktop.com/>

On 4/1/14 07:18 , Dan Van Der Ster wrote:
Hi,

On 1 Apr 2014 at 15:59:07, Andrey Korolyov (and...@xdel.ru <mailto:and...@xdel.ru>) wrote:
On 04/01/2014 03:51 PM, Robert Sander wrote:
> On 01.04.2014 13:38, Karol Kozubal wrote:
>
>> I am curious to know what is the largest known ceph production deployment?
>
> I would assume it is the CERN installation.
>
> Have a look at the slides from Frankfurt Ceph Day:
>
> http://www.slideshare.net/Inktank_Ceph/scaling-ceph-at-cern
>
> Regards
>

Just curious, how CERN guys built the network topology to prevent
possible cluster splits, because split in the middle will cause huge
downtime even for a relatively short split time enough to mark half of
those 1k OSDs as down by remaining MON majority.

The mons are distributed around the data centre, across N switches.
The OSDs are across a few switches --- actually, we could use CRUSH rules to replicate across switches but didn't do so because of an (unconfirmed) fear that the uplinks would become a bottleneck. So a switch or routing outage scenario is clearly a point of failure where some PGs could become stale, but we've been lucky enough not to suffer from that yet.

BTW, this 3PB cluster was built to test the scalability of Ceph's implementation, not because we have 3PB of data to store in Ceph today (most of the results of those tests are discussed in that presentation.). And we are currently partitioning this cluster down into a smaller production instance for Cinder and other instances for ongoing tests.

BTW#2, I don't think the CERN cluster is the largest. Isn't DreamHost's bigger?

Cheers, Dan

-- Dan van der Ster || Data & Storage Services || CERN IT Department --


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to