1. How about enabling trim/discard support in virtio-SCSI and using fstrim? That might work for you.
4. Well you can mount them rw in multiple vm's with predictably bad results, so I don't see any reason why you could not specify ro as a mount option and do ok. Sent from my iPad > On Oct 21, 2013, at 12:09 AM, Jon <three1...@gmail.com> wrote: > > Hello, > > Are there any current Perl modules for Ceph? I found a thread [1] from 2011 > with a version of Ceph::RADOS, but it only has functions to deal with pools, > and the ->list_pools function causes a seg. fault. > > I'm interested in controlling Ceph via script / application and I was > wondering [hoping] if anyone else had a current module before I go > reinventing the wheel. (My wheel would likely leverage calls to system() and > use the rbd/rados/ceph functions directly initially... I'm not proficient > with C/XS) > > I've been primarily using OpenNebula, though I've evaluated OpenStack, > CloudStack, and even Eucalyptus and they all seem to meet ($x-1)/$x criteria > (one project seems to do one thing better than another, but they all are > missing one feature that another project has--this is a generalization, but > this isn't the OpenNebula mailing list). What I'm looking to do at the > moment is simplify my lab deployments. My current workflow only takes 10 > minutes or so to deploy a new vm: > > 1) dump xml of existing vm (usually the "base" vm that the template was > created from, I actually have a "template" that I just copy and modify now) > 2) clone rbd to new vm (usually using vmname) > 3) edit vm template to relflect new values > -- change name of vm to the new vmname > -- remove specific identifiers (MAC, etc. unnecessary when copying > "template") > -- update disk to reflect new rbd > 4) login to console and "pre-provision" vm > -- update system > -- assign hostname > -- generate ssh-keys (I remove the sshd host keys when "sysprepping" for > cloning, ubuntu I know for sure doesn't regenerate the keys on boot, I > _THINK_ RHEL might) > > I actually already did this work on automating deployments[2], but that was > back when I was primarily using qcow2 images. It leverages guestfish to do > all of the vm :management" (setting IP, hostname, generating ssh host keys, > etc). But now I want to leverage my Ceph cluster for images. > > Couple of tangentially related questions that I don't think warrant a whole > thread: > > 1) Is it possible to zero and compress rbds? (I like to use virt-sysprep and > virt-sparcify to prepare my images, then, when I was using qcow images, I > would compress them before cloning) > 2) has anyone used virt-sysprep|virt-sparcify against rbd images? I suppose > if I'm creating a template image, I could create the qcow image then convert > it to an rbd, but qcow-img creates format 1 images. > 3) anyone know of a way to create format 2 images with qemu-img? When I > specify -O rbd qemu-img seg faults, and rbd2 is an invalid format. > 4) Is it possible to mount an RBD to multiple vms as readonly? I'm thinking > like readonly iso images converted to rbds? (is it even possible to convert > an iso to an image?) > > > Thanks for your help. > > Best Regards, > Jon A > > [1] http://www.spinics.net/lists/ceph-devel/msg04147.html > [2] https://github.com/three18ti/PrepVM-App > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com