Hi Frederic,

 

           Can you give me some solution, we are spending more time to solve 
this issue.

        

Regards

Prabu









 ---- On Thu, 15 Oct 2015 17:14:13 +0530 Tyler Bishop 
<tyler.bis...@beyondhosting.net> wrote ----




I don't know enough on ocfs to help.  Sounds like you have unconccurent writes 
though

Sent from TypeMail


On Oct 15, 2015, at 1:53 AM, gjprabu <gjpr...@zohocorp.com> wrote:

Hi Tyler,



   Can please send me the next setup action to be taken on this issue.



Regards

Prabu






---- On Wed, 14 Oct 2015 13:43:29 +0530 gjprabu <gjpr...@zohocorp.com> 
wrote ----




Hi Tyler,



         Thanks for your reply. We have disabled rbd_cache but still issue is 
persist. Please find our configuration file.



# cat /etc/ceph/ceph.conf

[global]

fsid = 944fa0af-b7be-45a9-93ff-b9907cfaee3f

mon_initial_members = integ-hm5, integ-hm6, integ-hm7

mon_host = 192.168.112.192,192.168.112.193,192.168.112.194

auth_cluster_required = cephx

auth_service_required = cephx

auth_client_required = cephx

filestore_xattr_use_omap = true

osd_pool_default_size = 2



[mon]

mon_clock_drift_allowed = .500



[client]

rbd_cache = false



--------------------------------------------------------------------------------------



 cluster 944fa0af-b7be-45a9-93ff-b9907cfaee3f

     health HEALTH_OK

     monmap e2: 3 mons at 
{integ-hm5=192.168.112.192:6789/0,integ-hm6=192.168.112.193:6789/0,integ-hm7=192.168.112.194:6789/0}

            election epoch 480, quorum 0,1,2 integ-hm5,integ-hm6,integ-hm7

     osdmap e49780: 2 osds: 2 up, 2 in

      pgmap v2256565: 190 pgs, 2 pools, 1364 GB data, 410 kobjects

            2559 GB used, 21106 GB / 24921 GB avail

                 190 active+clean

  client io 373 kB/s rd, 13910 B/s wr, 103 op/s





Regards

Prabu



 ---- On Tue, 13 Oct 2015 19:59:38 +0530 Tyler Bishop 
<tyler.bis...@beyondhosting.net> wrote ----




You need to disable RBD caching.







 Tyler Bishop
Chief Technical Officer
 513-299-7108 x10
 
tyler.bis...@beyondhosting.net

 
 
If you are not the intended recipient of this transmission you are notified 
that disclosing, copying, distributing or taking any action in reliance on the 
contents of this information is strictly prohibited.

 









From: "gjprabu" <gjpr...@zohocorp.com>

To: "Frédéric Nass" <frederic.n...@univ-lorraine.fr>

Cc: "<ceph-users@lists.ceph.com>" <ceph-users@lists.ceph.com>, 
"Siva Sokkumuthu" <sivaku...@zohocorp.com>, "Kamal Kannan 
Subramani(kamalakannan)" <ka...@manageengine.com>

Sent: Tuesday, October 13, 2015 9:11:30 AM

Subject: Re: [ceph-users] ceph same rbd on multiple client




Hi ,




 We have CEPH  RBD with OCFS2 mounted servers. we are facing i/o errors 
simultaneously while move the folder using one nodes in the same disk other 
nodes data replicating with below said error (Copying is not having any 
problem). Workaround if we remount the partition this issue get resolved but 
after sometime problem again reoccurred. please help on this issue.



Note : We have total 5 Nodes, here two nodes working fine other nodes are 
showing like below input/output error on moved data's.



ls -althr 

ls: cannot access LITE_3_0_M4_1_TEST: Input/output error 

ls: cannot access LITE_3_0_M4_1_OLD: Input/output error 

total 0 

d????????? ? ? ? ? ? LITE_3_0_M4_1_TEST 

d????????? ? ? ? ? ? LITE_3_0_M4_1_OLD 



Regards

Prabu






---- On Fri, 22 May 2015 17:33:04 +0530 Frédéric Nass 
<frederic.n...@univ-lorraine.fr> wrote ----




Hi,



Waiting for CephFS, you can use clustered filesystem like OCFS2 or GFS2 on top 
of RBD mappings so that each host can access the same device and clustered 
filesystem.



Regards,



Frédéric.



Le 21/05/2015 16:10, gjprabu a écrit :





-- Frédéric Nass Sous direction des Infrastructures, Direction du Numérique, 
Université de Lorraine. Tél : 03.83.68.53.83
_______________________________________________ 

ceph-users mailing list 

ceph-users@lists.ceph.com 

http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 


Hi All,



        We are using rbd and map the same rbd image to the rbd device on two 
different client but i can't see the data until i umount and mount -a 
partition. Kindly share the solution for this issue.



Example

create rbd image named foo

map foo to /dev/rbd0 on server A,   mount /dev/rbd0 to /mnt

map foo to /dev/rbd0 on server B,   mount /dev/rbd0 to /mnt



Regards

Prabu








_______________________________________________ ceph-users mailing list 
ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com






_______________________________________________

ceph-users mailing list

ceph-users@lists.ceph.com

http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



















_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to