Once the cluster is created on Ceph server nodes with MONs and OSDs on it
you have to copy the config + auth info to the clients:

#--- on server node, e.g.:
scp /etc/ceph/ceph.conf         client-1:/etc/ceph
scp /etc/ceph/keyring.bin       client-1:/etc/ceph
scp /etc/ceph/ceph.conf         client-2:/etc/ceph
scp /etc/ceph/keyring.bin       client-2:/etc/ceph

#--- on client node(s):
modprobe -v rbd
modprobe -v ceph        # only, if you want to run CephFS
rados lspools
rbd create -c /etc/ceph/ceph.conf  --size 1024000 --pool rbd    rbd-64k     
--order 16 --keyring /etc/ceph/keyring.bin
rbd create -c /etc/ceph/ceph.conf  --size 1024000 --pool rbd    rbd-128k    
--order 17 --keyring /etc/ceph/keyring.bin
rbd create -c /etc/ceph/ceph.conf  --size 1024000 --pool rbd    rbd-256k    
--order 18 --keyring /etc/ceph/keyring.bin
rbd create -c /etc/ceph/ceph.conf  --size 1024000 --pool rbd    rbd-4m      
--order 22 --keyring /etc/ceph/keyring.bin
rbd map rbd-64k
rbd map rbd-128k
rbd map rbd-256k
rbd map rbd-4m
rbd showmapped

id pool   image       snap device    
5  rbd    rbd-64k     -    /dev/rbd5 
6  rbd    rbd-128k    -    /dev/rbd6 
7  rbd    rbd-256k    -    /dev/rbd7 
8  rbd    rbd-4m      -    /dev/rbd8 


Now, your application can direct access the Rados Block Devices /dev/rbdX

Regards,
-Dieter



On Fri, Aug 23, 2013 at 01:31:05PM +0200, raj kumar wrote:
>    Thank you Sir. I appreciate your help on this.                             
>   
>    I upgraded the kernel to 3.4.53-8.                                         
>   
>    For second point, I want to give a client(which is not kvm) a block        
>   
>    storage. So without iscsi how the client will access the ceph cluster and  
>   
>    allocated block device.  and can you please let me know the flow to        
>   
>    provision the block storage. creating rbd image and map in one of the mon  
>   
>    host is right?  the ceph doc is not very clear on this.                    
>   
>    Regards                                                                    
>   
>    Raj                                                                        
>   
>                                                                               
>   
>    On Fri, Aug 23, 2013 at 4:03 PM, Kasper Dieter                             
>   
>    <[1]dieter.kas...@ts.fujitsu.com> wrote:                                   
>   
>                                                                               
>   
>      On Thu, Aug 22, 2013 at 03:32:35PM +0200, raj kumar wrote:               
>   
>      >    ceph cluster is running fine in centos6.4.                          
>   
>      >    Now I would like to export the block device to client using rbd.    
>   
>      >    my question is,                                                     
>   
>      >    1. I used to modprobe rbd in one of the monitor host. But I got     
>   
>      error,                                                                   
>   
>      >       FATAL: Module rbd not found                                      
>   
>      >       I could not find rbd module. How can i do this?                  
>   
>                                                                               
>   
>      # cat /etc/centos-release                                                
>   
>      CentOS release 6.4 (Final)                                               
>   
>                                                                               
>   
>      # updatedb                                                               
>   
>      # locate rbd.ko                                                          
>   
>      /lib/modules/3.8.13/kernel/drivers/block/rbd.ko                          
>   
>                                                                               
>   
>      # locate virtio_blk.ko                                                   
>   
>      
> /lib/modules/2.6.32-358.14.1.el6.x86_64/kernel/drivers/block/virtio_blk.ko 
>      /lib/modules/2.6.32-358.el6.x86_64/kernel/drivers/block/virtio_blk.ko    
>   
>      /lib/modules/3.8.13/kernel/drivers/block/virtio_blk.ko                   
>   
>                                                                               
>   
>      Well, the standard CentOS-6.4 kernel does not include 'rbd.ko'.          
>   
>      For some reasons the 'Enterprise distros' (RHEL, SLES) disabled the Ceph 
>   
>      Kernel                                                                   
>   
>      components by default, although the CephFS (= ceph.ko) is in the         
>   
>      upstream Kernel                                                          
>   
>      until 2.6.34, and the Block-Device (= rbd.ko) until 2.6.37.              
>   
>                                                                               
>   
>      We build our own Kernel 3.8.13 (a good mixture of recent & muture) and   
>   
>      put it into CentOS-6.4.                                                  
>   
>      >    2. Once the rbd is created. Do we need to create iscsi target in    
>   
>      one of a                                                                 
>   
>      >    monitor host and present the lun to client. If so what if the       
>   
>      monitor host                                                             
>   
>      >    goes down. so what is the best practice to provide a lun to         
>   
>      clients.                                                                 
>   
>      >    thanks                                                              
>   
>      This depends on your Client.                                             
>   
>      Using                                                                    
>   
>        "RADOS - Block-Layer - RBD-Driver - iSCSI-TGT // iSCSI-INI - Client"   
>   
>      is a waste of stack overhead.                                            
>   
>      If the client is kvm-qemu you can use                                    
>   
>        "RADOS // librbd - kvm-qemu"                                           
>   
>      or                                                                       
>   
>        "RADOS // Block-Layer - RBD-Driver - Client"                           
>   
>                                                                               
>   
>      The "//" symbolized the border between Server-nodes and client-nodes.    
>   
>                                                                               
>   
>      -Dieter                                                                  
>   
>                                                                               
>   
>      >    Raj                                                                 
>   
>                                                                               
>   
>      > _______________________________________________                        
>   
>      > ceph-users mailing list                                                
>   
>      > [2]ceph-users@lists.ceph.com                                           
>   
>      > [3]http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com              
>   
> 
> References
> 
>    Visible links
>    1. mailto:dieter.kas...@ts.fujitsu.com
>    2. mailto:ceph-users@lists.ceph.com
>    3. http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to