Dear,
 
In your last message i has understood, the exclusive-lock is work in kernel 4.9 
or higher and this could help-me, with don't permission write in two machines, 
but this feature only avaible in kernel 4.12, is right ?
 
I will reading more about the pacemaker, in my environment testing, i Would use 
the heartbeat, but the pacemaker it seams to be one alternative better.

 
 
Thanks a Lot
Marcelo >>>By default, since 4.9, > > >> > > when the exclusive-lock feature is enabled, only a single client > > >> > > can write to > > >> > the > > >> > > block device at a time -
Em 14/07/2017, Nick Fisk <n...@fisk.me.uk> escreveu:
> > > > -----Original Message----- > > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Jason Dillaman > > Sent: 14 July 2017 16:40 > > To: li...@marcelofrota.info > > Cc: ceph-users <ceph-users@lists.ceph.com> > > Subject: Re: [ceph-users] Ceph mount rbd > > > > On Fri, Jul 14, 2017 at 9:44 AM,  <li...@marcelofrota.info> wrote: > > > Gonzalo, > > > > > > > > > > > > You are right, i told so much about my enviroment actual and maybe i > > > didn't know explain my problem the better form, with ceph in the > > > moment, mutiple hosts clients can mount and write datas in my system > > > and this is one problem, because i could have filesystem corruption. > > > > > > > > > > > > Example, today, if runing the comand in two machines in the same time, > > > it will work. > > > > > > > > > > > > mount /dev/rbd0 /mnt/veeamrepo > > > > > > cd /mnt/veeamrepo ; touch testfile.txt > > > > > > > > > > > > I need ensure, only one machine will can execute this. > > > > > > > A user could do the same thing with any number of remote block devices (i.e. I could map an iSCSI target multiple times). As I said > > before, you can use the "exclusive" option available since kernel 4.12, roll your own solution using the advisory locks available from > > the rbd CLI, or just use CephFS if you want to be able to access a file system on multiple hosts. > > Pacemaker, will also prevent a RBD to be mounted multiple times, if you want to manage the fencing outside of Ceph. > > > > > > > > > Thanks a lot, > > > > > > Marcelo > > > > > > > > > Em 14/07/2017, Gonzalo Aguilar Delgado <gagui...@aguilardelgado.com> > > > escreveu: > > > > > > > > >> Hi, > > >> > > >> Why you would like to maintain copies by yourself. You replicate on > > >> ceph and then on different files inside ceph? Let ceph take care of counting. > > >> Create a pool with 3 or more copies and let ceph take care of what's > > >> stored and where. > > >> > > >> Best regards, > > >> > > >> > > >> El 13/07/17 a las 17:06, li...@marcelofrota.info escribió: > > >> > > > >> > I will explain More about my system actual, in the moment i have 2 > > >> > machines using drbd in mode master/slave and i running the > > >> > aplication in machine master, but existing 2 questions importants > > >> > in my enviroment with drbd actualy : > > >> > > > >> > 1 - If machine one is master and mounting partitions, the slave > > >> > don't can mount the system, Unless it happens one problem in > > >> > machine master, this is one mode, to prevent write in filesystem > > >> > incorrect > > >> > > > >> > 2 - When i write data in machine master in drbd, the drbd write > > >> > datas in slave machine Automatically, with this, if one problem > > >> > happens in node master, the machine slave have coppy the data. > > >> > > > >> > In the moment, in my enviroment testing with ceph, using the > > >> > version > > >> > 4.10 of kernel and i mount the system in two machines in the same > > >> > time, in production enviroment, i could serious problem with this > > >> > comportament. > > >> > > > >> > How can i use the ceph and Ensure that I could get these 2 > > >> > behaviors kept in a new environment with Ceph? > > >> > > > >> > Thanks a lot, > > >> > > > >> > Marcelo > > >> > > > >> > > > >> > Em 28/06/2017, Jason Dillaman <jdill...@redhat.com> escreveu: > > >> > > ... additionally, the forthcoming 4.12 kernel release will > > >> > > support non-cooperative exclusive locking. By default, since 4.9, > > >> > > when the exclusive-lock feature is enabled, only a single client > > >> > > can write to > > >> > the > > >> > > block device at a time -- but they will cooperatively pass the > > >> > > lock > > >> > back > > >> > > and forth upon write request. With the new "rbd map" option, you > > >> > > can > > >> > map a > > >> > > image on exactly one host and prevent other hosts from mapping > > >> > > the > > >> > image. > > >> > > If that host should die, the exclusive-lock will automatically > > >> > > become available to other hosts for mapping. > > >> > > > > >> > > Of course, I always have to ask the use-case behind mapping the > > >> > > same > > >> > image > > >> > > on multiple hosts. Perhaps CephFS would be a better fit if you > > >> > > are > > >> > trying > > >> > > to serve out a filesystem? > > >> > > > > >> > > On Wed, Jun 28, 2017 at 6:25 PM, Maged Mokhtar > > >> > <mmokh...@petasan.org> wrote: > > >> > > > > >> > > > On 2017-06-28 22:55, li...@marcelofrota.info wrote: > > >> > > > > > >> > > > Hi People, > > >> > > > > > >> > > > I am testing the new enviroment, with ceph + rbd with ubuntu > > >> > 16.04, and i > > >> > > > have one question. > > >> > > > > > >> > > > I have my cluster ceph and mount the using the comands to ceph > > >> > > > in > > >> > my linux > > >> > > > enviroment : > > >> > > > > > >> > > > rbd create veeamrepo --size 20480 rbd --image veeamrepo info > > >> > > > modprobe rbd rbd map veeamrepo rbd feature disable veeamrepo > > >> > > > exclusive-lock object-map fast-diff deep-flatten mkdir > > >> > > > /mnt/veeamrepo mount /dev/rbd0 /mnt/veeamrepo > > >> > > > > > >> > > > The comands work fine, but i have one problem, in the moment, i > > >> > can mount > > >> > > > the /mnt/veeamrepo in the same time in 2 machines, and this is > > >> > > > a > > >> > bad option > > >> > > > for me in the moment, because this could generate one > > >> > > > filesystem > > >> > corrupt. > > >> > > > > > >> > > > I need only one machine to be allowed to mount and write at a time. > > >> > > > > > >> > > > Example if machine1 mount the /mnt/veeamrepo and machine2 try > > >> > mount, one > > >> > > > error would be displayed, show message the machine can not > > >> > > > mount, > > >> > because > > >> > > > the system already mounted in machine1. > > >> > > > > > >> > > > Someone, could help-me with this or give some tips, for > > >> > > > solution my problem. ? > > >> > > > > > >> > > > Thanks a lot > > >> > > > > > >> > > > _______________________________________________ > > >> > > > ceph-users mailing list > > >> > > > ceph-users@lists.ceph.com > > >> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >> > > > > > >> > > > > > >> > > > > > >> > > > You can use Pacemaker to map the rbd and mount the filesystem > > >> > > > on 1 > > >> > server > > >> > > > and in case of failure switch to another server. > > >> > > > > > >> > > > > > >> > > > _______________________________________________ > > >> > > > ceph-users mailing list > > >> > > > ceph-users@lists.ceph.com > > >> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >> > > > > > >> > > > > > >> > > > > >> > > > > >> > > -- > > >> > > Jason > > >> > > > >> > > > >> > > > >> > _______________________________________________ > > >> > ceph-users mailing list > > >> > ceph-users@lists.ceph.com > > >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >> > > >> ---- > > >> > > >> _______________________________________________ > > >> ceph-users mailing list > > >> ceph-users@lists.ceph.com > > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > > _______________________________________________ > > > ceph-users mailing list > > > ceph-users@lists.ceph.com > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > > > > -- > > Jason > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to