Thank you so much. This is helpful not only for me, but for all beginners.

Raj


On Fri, Aug 23, 2013 at 5:31 PM, Kasper Dieter <dieter.kas...@ts.fujitsu.com
> wrote:

> Once the cluster is created on Ceph server nodes with MONs and OSDs on it
> you have to copy the config + auth info to the clients:
>
> #--- on server node, e.g.:
> scp /etc/ceph/ceph.conf         client-1:/etc/ceph
> scp /etc/ceph/keyring.bin       client-1:/etc/ceph
> scp /etc/ceph/ceph.conf         client-2:/etc/ceph
> scp /etc/ceph/keyring.bin       client-2:/etc/ceph
>
> #--- on client node(s):
> modprobe -v rbd
> modprobe -v ceph        # only, if you want to run CephFS
> rados lspools
> rbd create -c /etc/ceph/ceph.conf  --size 1024000 --pool rbd    rbd-64k
>   --order 16 --keyring /etc/ceph/keyring.bin
> rbd create -c /etc/ceph/ceph.conf  --size 1024000 --pool rbd    rbd-128k
>  --order 17 --keyring /etc/ceph/keyring.bin
> rbd create -c /etc/ceph/ceph.conf  --size 1024000 --pool rbd    rbd-256k
>  --order 18 --keyring /etc/ceph/keyring.bin
> rbd create -c /etc/ceph/ceph.conf  --size 1024000 --pool rbd    rbd-4m
>  --order 22 --keyring /etc/ceph/keyring.bin
> rbd map rbd-64k
> rbd map rbd-128k
> rbd map rbd-256k
> rbd map rbd-4m
> rbd showmapped
>
> id pool   image       snap device
> 5  rbd    rbd-64k     -    /dev/rbd5
> 6  rbd    rbd-128k    -    /dev/rbd6
> 7  rbd    rbd-256k    -    /dev/rbd7
> 8  rbd    rbd-4m      -    /dev/rbd8
>
>
> Now, your application can direct access the Rados Block Devices /dev/rbdX
>
> Regards,
> -Dieter
>
>
>
> On Fri, Aug 23, 2013 at 01:31:05PM +0200, raj kumar wrote:
> >    Thank you Sir. I appreciate your help on this.
> >    I upgraded the kernel to 3.4.53-8.
> >    For second point, I want to give a client(which is not kvm) a block
> >    storage. So without iscsi how the client will access the ceph cluster
> and
> >    allocated block device.  and can you please let me know the flow to
> >    provision the block storage. creating rbd image and map in one of the
> mon
> >    host is right?  the ceph doc is not very clear on this.
> >    Regards
> >    Raj
> >
> >    On Fri, Aug 23, 2013 at 4:03 PM, Kasper Dieter
> >    <[1]dieter.kas...@ts.fujitsu.com> wrote:
> >
> >      On Thu, Aug 22, 2013 at 03:32:35PM +0200, raj kumar wrote:
> >      >    ceph cluster is running fine in centos6.4.
> >      >    Now I would like to export the block device to client using
> rbd.
> >      >    my question is,
> >      >    1. I used to modprobe rbd in one of the monitor host. But I got
> >      error,
> >      >       FATAL: Module rbd not found
> >      >       I could not find rbd module. How can i do this?
> >
> >      # cat /etc/centos-release
> >      CentOS release 6.4 (Final)
> >
> >      # updatedb
> >      # locate rbd.ko
> >      /lib/modules/3.8.13/kernel/drivers/block/rbd.ko
> >
> >      # locate virtio_blk.ko
> >
>  /lib/modules/2.6.32-358.14.1.el6.x86_64/kernel/drivers/block/virtio_blk.ko
> >
>  /lib/modules/2.6.32-358.el6.x86_64/kernel/drivers/block/virtio_blk.ko
> >      /lib/modules/3.8.13/kernel/drivers/block/virtio_blk.ko
> >
> >      Well, the standard CentOS-6.4 kernel does not include 'rbd.ko'.
> >      For some reasons the 'Enterprise distros' (RHEL, SLES) disabled the
> Ceph
> >      Kernel
> >      components by default, although the CephFS (= ceph.ko) is in the
> >      upstream Kernel
> >      until 2.6.34, and the Block-Device (= rbd.ko) until 2.6.37.
> >
> >      We build our own Kernel 3.8.13 (a good mixture of recent & muture)
> and
> >      put it into CentOS-6.4.
> >      >    2. Once the rbd is created. Do we need to create iscsi target
> in
> >      one of a
> >      >    monitor host and present the lun to client. If so what if the
> >      monitor host
> >      >    goes down. so what is the best practice to provide a lun to
> >      clients.
> >      >    thanks
> >      This depends on your Client.
> >      Using
> >        "RADOS - Block-Layer - RBD-Driver - iSCSI-TGT // iSCSI-INI -
> Client"
> >      is a waste of stack overhead.
> >      If the client is kvm-qemu you can use
> >        "RADOS // librbd - kvm-qemu"
> >      or
> >        "RADOS // Block-Layer - RBD-Driver - Client"
> >
> >      The "//" symbolized the border between Server-nodes and
> client-nodes.
> >
> >      -Dieter
> >
> >      >    Raj
> >
> >      > _______________________________________________
> >      > ceph-users mailing list
> >      > [2]ceph-users@lists.ceph.com
> >      > [3]http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> > References
> >
> >    Visible links
> >    1. mailto:dieter.kas...@ts.fujitsu.com
> >    2. mailto:ceph-users@lists.ceph.com
> >    3. http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to