Once the cluster is created on Ceph server nodes with MONs and OSDs on it you have to copy the config + auth info to the clients:
#--- on server node, e.g.: scp /etc/ceph/ceph.conf client-1:/etc/ceph scp /etc/ceph/keyring.bin client-1:/etc/ceph scp /etc/ceph/ceph.conf client-2:/etc/ceph scp /etc/ceph/keyring.bin client-2:/etc/ceph #--- on client node(s): modprobe -v rbd modprobe -v ceph # only, if you want to run CephFS rados lspools rbd create -c /etc/ceph/ceph.conf --size 1024000 --pool rbd rbd-64k --order 16 --keyring /etc/ceph/keyring.bin rbd create -c /etc/ceph/ceph.conf --size 1024000 --pool rbd rbd-128k --order 17 --keyring /etc/ceph/keyring.bin rbd create -c /etc/ceph/ceph.conf --size 1024000 --pool rbd rbd-256k --order 18 --keyring /etc/ceph/keyring.bin rbd create -c /etc/ceph/ceph.conf --size 1024000 --pool rbd rbd-4m --order 22 --keyring /etc/ceph/keyring.bin rbd map rbd-64k rbd map rbd-128k rbd map rbd-256k rbd map rbd-4m rbd showmapped id pool image snap device 5 rbd rbd-64k - /dev/rbd5 6 rbd rbd-128k - /dev/rbd6 7 rbd rbd-256k - /dev/rbd7 8 rbd rbd-4m - /dev/rbd8 Now, your application can direct access the Rados Block Devices /dev/rbdX Regards, -Dieter On Fri, Aug 23, 2013 at 01:31:05PM +0200, raj kumar wrote: > Thank you Sir. I appreciate your help on this. > > I upgraded the kernel to 3.4.53-8. > > For second point, I want to give a client(which is not kvm) a block > > storage. So without iscsi how the client will access the ceph cluster and > > allocated block device. and can you please let me know the flow to > > provision the block storage. creating rbd image and map in one of the mon > > host is right? the ceph doc is not very clear on this. > > Regards > > Raj > > > > On Fri, Aug 23, 2013 at 4:03 PM, Kasper Dieter > > <[1]dieter.kas...@ts.fujitsu.com> wrote: > > > > On Thu, Aug 22, 2013 at 03:32:35PM +0200, raj kumar wrote: > > > ceph cluster is running fine in centos6.4. > > > Now I would like to export the block device to client using rbd. > > > my question is, > > > 1. I used to modprobe rbd in one of the monitor host. But I got > > error, > > > FATAL: Module rbd not found > > > I could not find rbd module. How can i do this? > > > > # cat /etc/centos-release > > CentOS release 6.4 (Final) > > > > # updatedb > > # locate rbd.ko > > /lib/modules/3.8.13/kernel/drivers/block/rbd.ko > > > > # locate virtio_blk.ko > > > /lib/modules/2.6.32-358.14.1.el6.x86_64/kernel/drivers/block/virtio_blk.ko > /lib/modules/2.6.32-358.el6.x86_64/kernel/drivers/block/virtio_blk.ko > > /lib/modules/3.8.13/kernel/drivers/block/virtio_blk.ko > > > > Well, the standard CentOS-6.4 kernel does not include 'rbd.ko'. > > For some reasons the 'Enterprise distros' (RHEL, SLES) disabled the Ceph > > Kernel > > components by default, although the CephFS (= ceph.ko) is in the > > upstream Kernel > > until 2.6.34, and the Block-Device (= rbd.ko) until 2.6.37. > > > > We build our own Kernel 3.8.13 (a good mixture of recent & muture) and > > put it into CentOS-6.4. > > > 2. Once the rbd is created. Do we need to create iscsi target in > > one of a > > > monitor host and present the lun to client. If so what if the > > monitor host > > > goes down. so what is the best practice to provide a lun to > > clients. > > > thanks > > This depends on your Client. > > Using > > "RADOS - Block-Layer - RBD-Driver - iSCSI-TGT // iSCSI-INI - Client" > > is a waste of stack overhead. > > If the client is kvm-qemu you can use > > "RADOS // librbd - kvm-qemu" > > or > > "RADOS // Block-Layer - RBD-Driver - Client" > > > > The "//" symbolized the border between Server-nodes and client-nodes. > > > > -Dieter > > > > > Raj > > > > > _______________________________________________ > > > ceph-users mailing list > > > [2]ceph-users@lists.ceph.com > > > [3]http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > References > > Visible links > 1. mailto:dieter.kas...@ts.fujitsu.com > 2. mailto:ceph-users@lists.ceph.com > 3. http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com