Hi, i have done a short look into RBD + iSCSI, and i found TGT + librbd.
https://github.com/fujita/tgt http://stgt.sourceforge.net/ i didn't take a deeper look into it but i like to test it in the next monthor so, it looks easy to me https:// github.com/fujita/tgt/blob/master/doc/README.rbd cheers Ansgar 2013/3/16 Bond, Darryl <db...@nrggos.com.au> > I have a small 3 node ceph cluster with 6 OSDs on each node > I would like to re-export some rbd images via LIO. > Is it recommended to run RBD/LIO on one of the cluster nodes? > > Preliminary tests show that it works fine. I have seen reports (that I > can't find) that it is not recommended to run the RBD kernel module on an > OSD node. > > Has anyone used multiple hosts to do iSCSI multipathing to a singe RBD > image for vmware? > My thoughts are to export the same RBD image via LIO from 2 hosts. It is > easy to configure LIO to use the same iSCSI target address on both hosts. > > I could then configure vmware storage with the two ceph nodes as a > primary/secondary failover. > > Regards > Darryl > > > The contents of this electronic message and any attachments are intended > only for the addressee and may contain legally privileged, personal, > sensitive or confidential information. If you are not the intended > addressee, and have received this email, any transmission, distribution, > downloading, printing or photocopying of the contents of this message or > attachments is strictly prohibited. Any legal privilege or confidentiality > attached to this message and attachments is not waived, lost or destroyed > by reason of delivery to any person other than intended addressee. If you > have received this message and are not the intended addressee you should > notify the sender by return email and destroy all copies of the message and > any attachments. Unless expressly attributed, the views expressed in this > email do not necessarily represent the views of the company. > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com