Hi John,

Thanks for the responses.

For (a), I remember reading somewhere that one can only run a max of 1
monitor/node, I assume that that implies the single monitor process will be
responsible for ALL ceph clusters on that node, correct?

So (b) isn't really a Ceph issue, that's nice to know. Any recommendations
on the minimum kernel/glibc version and min RAM size requirements where
Ceph can be run on a single client in native mode? Reason I ask this is in
a few deployment scenarios (especially non-standard like telco platforms),
hardware gets added gradually, so its more important to be able to scale
the cluster out gracefully. I actually see Ceph as an alternative to SAN,
using JBODs from machines to create a larg(ish) storage cluster. Plus,
usually, the clients would probably be running on the same hardware as the
OSD/MON, because space on the chassis is at a premium.

(d) I was thinking about single node failure scenarios, with 3 nodes,
wouldn't a failure of 1 node cause PAXOS to not work?



Thanks,
Hari





On Fri, Jul 26, 2013 at 10:00 AM, John Wilkins <john.wilk...@inktank.com>wrote:

> (a) Yes. See
> http://ceph.com/docs/master/rados/configuration/ceph-conf/#running-multiple-clusters
> and
> http://ceph.com/docs/master/rados/deployment/ceph-deploy-new/#naming-a-cluster
> (b) Yes. See
> http://wiki.ceph.com/03FAQs/01General_FAQ#How_Can_I_Give_Ceph_a_Try.3F
>  Mounting kernel modules on the same node as Ceph Daemons can cause
> older kernels to deadlock.
> (c) Someone else can probably answer that better than me.
> (d) At least three. Paxos requires a simple majority, so 2 out of 3 is
> sufficient. See
> http://ceph.com/docs/master/rados/configuration/mon-config-ref/#background
> particularly the monitor quorum section.
>
> On Wed, Jul 24, 2013 at 4:03 PM, Hariharan Thantry <than...@gmail.com>
> wrote:
> > Hi folks,
> >
> > Some very basic questions.
> >
> > (a) Can I be running more than 1 ceph cluster on the same node (assume
> that
> > I have no more than 1 monitor/node, but storage is contributed by one
> node
> > into more than 1 cluster)
> > (b) Are there any issues with running Ceph clients on the same node as
> the
> > other Ceph storage cluster entities (OSD/MON?)
> > (c) Is the best way to access Ceph storage cluster in native mode by
> > multiple clients through hosting a shared-disk filesystem on top of the
> RBD
> > (like OCFS2?). What if these clients were running inside VMs? Could one
> then
> > create independent partitions on top of rbd and give a partition to each
> of
> > the VMs?
> > (d) Isn't the realistic minimum for # of monitors in a cluster at least 4
> > (to guard against one failure?)
> >
> >
> > Thanks,
> > Hari
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
>
>
>
> --
> John Wilkins
> Senior Technical Writer
> Intank
> john.wilk...@inktank.com
> (415) 425-9599
> http://inktank.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to