I don't know enough on ocfs to help. Sounds like you have unconccurent writes though
Sent from TypeMail On Oct 15, 2015, 1:53 AM, at 1:53 AM, gjprabu <gjpr...@zohocorp.com> wrote: >Hi Tyler, > > > > Can please send me the next setup action to be taken on this issue. > > > >Regards > >Prabu > > > > > >---- On Wed, 14 Oct 2015 13:43:29 +0530 gjprabu ><gjpr...@zohocorp.com> wrote ---- > > > > >Hi Tyler, > > > >Thanks for your reply. We have disabled rbd_cache but still issue is >persist. Please find our configuration file. > > > ># cat /etc/ceph/ceph.conf > >[global] > >fsid = 944fa0af-b7be-45a9-93ff-b9907cfaee3f > >mon_initial_members = integ-hm5, integ-hm6, integ-hm7 > >mon_host = 192.168.112.192,192.168.112.193,192.168.112.194 > >auth_cluster_required = cephx > >auth_service_required = cephx > >auth_client_required = cephx > >filestore_xattr_use_omap = true > >osd_pool_default_size = 2 > > > >[mon] > >mon_clock_drift_allowed = .500 > > > >[client] > >rbd_cache = false > > > >-------------------------------------------------------------------------------------- > > > > cluster 944fa0af-b7be-45a9-93ff-b9907cfaee3f > > health HEALTH_OK > >monmap e2: 3 mons at >{integ-hm5=192.168.112.192:6789/0,integ-hm6=192.168.112.193:6789/0,integ-hm7=192.168.112.194:6789/0} > > election epoch 480, quorum 0,1,2 integ-hm5,integ-hm6,integ-hm7 > > osdmap e49780: 2 osds: 2 up, 2 in > > pgmap v2256565: 190 pgs, 2 pools, 1364 GB data, 410 kobjects > > 2559 GB used, 21106 GB / 24921 GB avail > > 190 active+clean > > client io 373 kB/s rd, 13910 B/s wr, 103 op/s > > > > > >Regards > >Prabu > > > >---- On Tue, 13 Oct 2015 19:59:38 +0530 Tyler Bishop ><tyler.bis...@beyondhosting.net> wrote ---- > > > > >You need to disable RBD caching. > > > > > > > > Tyler Bishop >Chief Technical Officer > 513-299-7108 x10 > >tyler.bis...@beyondhosting.net > > > >If you are not the intended recipient of this transmission you are >notified that disclosing, copying, distributing or taking any action in >reliance on the contents of this information is strictly prohibited. > > > > > > > > > > > >From: "gjprabu" <gjpr...@zohocorp.com> > >To: "Frédéric Nass" <frederic.n...@univ-lorraine.fr> > >Cc: "<ceph-users@lists.ceph.com>" ><ceph-users@lists.ceph.com>, "Siva Sokkumuthu" ><sivaku...@zohocorp.com>, "Kamal Kannan Subramani(kamalakannan)" ><ka...@manageengine.com> > >Sent: Tuesday, October 13, 2015 9:11:30 AM > >Subject: Re: [ceph-users] ceph same rbd on multiple client > > > > >Hi , > > > > >We have CEPH RBD with OCFS2 mounted servers. we are facing i/o errors >simultaneously while move the folder using one nodes in the same disk >other nodes data replicating with below said error (Copying is not >having any problem). Workaround if we remount the partition this issue >get resolved but after sometime problem again reoccurred. please help >on this issue. > > > >Note : We have total 5 Nodes, here two nodes working fine other nodes >are showing like below input/output error on moved data's. > > > >ls -althr > >ls: cannot access LITE_3_0_M4_1_TEST: Input/output error > >ls: cannot access LITE_3_0_M4_1_OLD: Input/output error > >total 0 > >d????????? ? ? ? ? ? LITE_3_0_M4_1_TEST > >d????????? ? ? ? ? ? LITE_3_0_M4_1_OLD > > > >Regards > >Prabu > > > > > > >---- On Fri, 22 May 2015 17:33:04 +0530 Frédéric Nass ><frederic.n...@univ-lorraine.fr> wrote ---- > > > > >Hi, > > > >Waiting for CephFS, you can use clustered filesystem like OCFS2 or GFS2 >on top of RBD mappings so that each host can access the same device and >clustered filesystem. > > > >Regards, > > > >Frédéric. > > > >Le 21/05/2015 16:10, gjprabu a écrit : > > > > > >-- Frédéric Nass Sous direction des Infrastructures, Direction du >Numérique, Université de Lorraine. Tél : 03.83.68.53.83 >_______________________________________________ > >ceph-users mailing list > >ceph-users@lists.ceph.com > >http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >Hi All, > > > >We are using rbd and map the same rbd image to the rbd device on two >different client but i can't see the data until i umount and mount -a >partition. Kindly share the solution for this issue. > > > >Example > >create rbd image named foo > >map foo to /dev/rbd0 on server A, mount /dev/rbd0 to /mnt > >map foo to /dev/rbd0 on server B, mount /dev/rbd0 to /mnt > > > >Regards > >Prabu > > > > > > > > >_______________________________________________ ceph-users mailing list >ceph-users@lists.ceph.com >http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > >_______________________________________________ > >ceph-users mailing list > >ceph-users@lists.ceph.com > >http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com