Hi Taylor,


   Details are below.



ceph -s

    cluster 944fa0af-b7be-45a9-93ff-b9907cfaee3f

     health HEALTH_OK

     monmap e2: 3 mons at 
{integ-hm5=192.168.112.192:6789/0,integ-hm6=192.168.112.193:6789/0,integ-hm7=192.168.112.194:6789/0}

            election epoch 526, quorum 0,1,2 integ-hm5,integ-hm6,integ-hm7

     osdmap e50127: 3 osds: 3 up, 3 in

      pgmap v2923439: 190 pgs, 2 pools, 3401 GB data, 920 kobjects

            6711 GB used, 31424 GB / 40160 GB avail

                 190 active+clean

 client io 35153 kB/s rd, 1912 kB/s wr, 672 op/s



Client is automatically unmounted in our cause.



Is it possible to change the PG_num in the production setup.



Journal stored on SATA 7.2k RPM  6GPS and 1gb network interface.



We are not configured Public and cluster as a separate network and it will be 
transferable via same LAN. Do we need to do this setup for better performance.



Also what is beter i/o operation setting for the crush map.





Still we are getting errors in ceph osd logs ,what need to done for this error.



2015-11-03 13:04:18.809488 7f387019c700  0 bad crc in data 3742210963 != exp 
924878202

2015-11-03 13:04:18.812911 7f387019c700  0 -- 192.168.112.231:6800/49908 
>> 192.168.112.192:0/1457324982 pipe(0x170d2000 sd=44 :6800 s=0 pgs=0 
cs=0 l=0 c=0x1b18bf40).accept peer addr is really 192.168.112.192:0/1457324982 
(socket is 192.168.112.192:47128/0)





Regards

Prabu










---- On Tue, 03 Nov 2015 12:50:40 +0530 Chris Taylor <ctay...@eyonic.com> 
wrote ----




On 2015-11-02 10:19 pm, gjprabu wrote: 



> Hi Taylor, 

> 

> I have checked DNS name and all host resolve to the correct IP. MTU 

> size is 1500 in switch level configuration done. There is no firewall/ 

> selinux is running currently. 

> 

> Also we would like to know below query's which already in the thread. 

> 

> Regards 

> Prabu 

> 

> ---- On Tue, 03 Nov 2015 11:20:07 +0530 CHRIS TAYLOR 

> <ctay...@eyonic.com> wrote ---- 

> 

> I would double check the network configuration on the new node. 

> Including hosts files and DNS names. Do all the host names resolve to 

> the correct IP addresses from all hosts? 

> 

> "... 192.168.112.231:6800/49908 >> 192.168.113.42:0/599324131 ..." 

> 

> Looks like the communication between subnets is a problem. Is 

> xxx.xxx.113.xxx a typo? If that's correct, check MTU sizes. Are they 

> configured correctly on the switch and all NICs? 

> 

> Is there any iptables/firewall rules that could be blocking traffic 

> between hosts? 

> 

> Hope that helps, 

> 

> Chris 

> 

> On 2015-11-02 9:18 pm, gjprabu wrote: 

> 

> Hi, 

> 

> Anybody please help me on this issue. 

> 

> Regards 

> Prabu 

> 

> ---- On Mon, 02 Nov 2015 17:54:27 +0530 GJPRABU 
<gjpr...@zohocorp.com> 

> wrote ---- 

> 

> Hi Team, 

> 

> We have ceph setup with 2 OSD and replica 2 and it is mounted with 

> ocfs2 clients and its working. When we added new osd all the clients 

> rbd mapped device disconnected and got hanged by running rbd ls or rbd 

> map command. We waited for long hours to scale the new osd size but 

> peering not completed event data sync finished, but client side issue 

> was persist and thought to try old osd service stop/start, after some 

> time rbd mapped automatically using existing map script. 

> 

> After service stop/start in old osd again 3rd OSD rebuild and back 

> filling started and after some time clients rbd mapped device 

> disconnected and got hanged by running rbd ls or rbd map command. We 

> thought to wait till to finished data sync in 3'rd OSD and its 

> completed, even though client side rbd not mapped. After we restarted 

> all mon and osd service and client side issue got fixed and mounted 

> rbd. We suspected some issue in our setup. also attached logs for your 

> reference. 

> 



What does 'ceph -s' look like? is the cluster HEALTH_OK? 



> 

> Something we are missing in our setup i don't know, highly appreciated 

> if anybody help us to solve this issue. 

> 

> Before new osd.2 addition : 

> 

> osd.0 - size : 13T and used 2.7 T 

> osd.1 - size : 13T and used 2.7 T 

> 

> After new osd addition : 

> osd.0 size : 13T and used 1.8T 

> osd.1 size : 13T and used 2.1T 

> osd.2 size : 15T and used 2.5T 

> 

> rbd ls 

> repo / integrepository (pg_num: 126) 

> rbd / integdownloads (pg_num: 64) 

> 

> Also we would like to know few clarifications . 

> 

> If any new osd will be added whether all client will be unmounted 

> automatically . 

> 



Clients do not need to unmount images when OSDs are added. 



> While add new osd can we access ( read / write ) from client machines ? 

> 



Clients still have read/write access to RBD images in the cluster while 

adding OSDs and during recovery. 



> How much data will be added in new osd - without change any repilca / 

> pg_num ? 

> 



The data will re-balance between OSDs automatically. I found having more 

PGs help distribute the load more evenly. 



> How long to take finish this process ? 



Depends greatly on the hardware and configuration. Whether Journals on 

SSD or spinning disks, network connectivity, max_backfills, etc. 



> 

> If we missed any common configuration - please share the same . 



I don't see any configuration for public and cluster networks. If you 

are sharing the same network for clients and object replication/recovery 

the cluster re-balancing data between OSDs could cause problems with the 

client traffic. 



Take a look at: 

http://docs.ceph.com/docs/master/rados/configuration/network-config-ref/ 



> 

> ceph.conf 

> [global] 

> fsid = 944fa0af-b7be-45a9-93ff-b9907cfaee3f 

> mon_initial_members = integ-hm5, integ-hm6, integ-hm7 

> mon_host = 192.168.112.192,192.168.112.193,192.168.112.194 

> auth_cluster_required = cephx 

> auth_service_required = cephx 

> auth_client_required = cephx 

> filestore_xattr_use_omap = true 

> osd_pool_default_size = 2 

> 

> [mon] 

> mon_clock_drift_allowed = .500 

> 

> [client] 

> rbd_cache = false 

> 

> Current Logs from new osd also attached old logs. 

> 

> 2015-11-02 12:47:48.481641 7f386f691700 0 bad crc in data 3889133030 != 

> exp 2857248268 

> 2015-11-02 12:47:48.482230 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0xc510580).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:42530/0) 

> 2015-11-02 12:47:48.483951 7f386f691700 0 bad crc in data 3192803598 != 

> exp 1083014631 

> 2015-11-02 12:47:48.484512 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0xc516f60).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:42531/0) 

> 2015-11-02 12:47:48.486284 7f386f691700 0 bad crc in data 133120597 != 

> exp 393328400 

> 2015-11-02 12:47:48.486777 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0xc514620).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:42532/0) 

> 2015-11-02 12:47:48.488624 7f386f691700 0 bad crc in data 3299720069 != 

> exp 211350069 

> 2015-11-02 12:47:48.489100 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0xc513860).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:42533/0) 

> 2015-11-02 12:47:48.490911 7f386f691700 0 bad crc in data 2381447347 != 

> exp 1177846878 

> 2015-11-02 12:47:48.491390 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0xc513700).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:42534/0) 

> 2015-11-02 12:47:48.493167 7f386f691700 0 bad crc in data 2093712440 != 

> exp 2175112954 

> 2015-11-02 12:47:48.493682 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0xc514200).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:42535/0) 

> 2015-11-02 12:47:48.495150 7f386f691700 0 bad crc in data 3047197039 != 

> exp 38098198 

> 2015-11-02 12:47:48.495679 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0xc510b00).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:42536/0) 

> 2015-11-02 12:47:48.497259 7f386f691700 0 bad crc in data 1400444622 != 

> exp 2648291990 

> 2015-11-02 12:47:48.497756 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0x17f7b700).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:42537/0) 

> 2015-11-02 13:02:00.439025 7f386f691700 0 bad crc in data 4159064831 != 

> exp 903679865 

> 2015-11-02 13:02:00.441337 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0x17f7e5c0).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:43128/0) 

> 2015-11-02 13:02:00.442756 7f386f691700 0 bad crc in data 1134831440 != 

> exp 892008036 

> 2015-11-02 13:02:00.443369 7f386f691700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 
cs=0 l=0 c=0x17f7ee00).accept peer addr is really 192.168.113.42:0/599324131 
(socket is 192.168.113.42:43129/0) 

> 2015-11-02 13:08:43.272527 7f387049f700 0 -- 192.168.112.231:6800/49908 

> >> 192.168.112.115:0/4256128918 pipe(0x170ea000 sd=33 :6800 s=0 
pgs=0 cs=0 l=0 c=0x17f7e1a0).accept peer addr is really 
192.168.112.115:0/4256128918 (socket is 192.168.112.115:51660/0) 

> 

> Regards 

> Prabu 

> 

> Regards 

> G.J 

> 

> _______________________________________________ 

> ceph-users mailing list 

> ceph-users@lists.ceph.com 

> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [1] 

> 

> _______________________________________________ 

> ceph-users mailing list 

> ceph-users@lists.ceph.com 

> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [1] 







Links: 

------ 

[1] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 






_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to