My 0.02: 1. Why you need to simultaneously set the map for your purpose ? It’s obvious very important for ceph to have an atomic CLI , but this is just because the map may be changed by cluster itself ( loss node or what), but not for your case. Since the map can be auto-distributed by ceph, I really think it’s a good idea to just change your own code , to have the map changing stuff only happen in one node.
2. We are also evaluating the pros and cons for “local pool” , well, the only pros is you can save the network BW for read. You may want to say latency, I agree with you before but after we have a complete latency breakdown for ceph, showing that the network latency can be neglect , even using full-ssd setup. The question remain is “how much BW can be save?”, well, unless you have some prior statement about the workload that is reading majority, or you still have to use a 10GbE link for Ceph to have a balanced throughput . But the drawback is really obvious, live-migration complexity, management complexity, and etc. Xiaoxi From: ceph-users-boun...@lists.ceph.com [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Rongze Zhu Sent: Friday, July 26, 2013 2:29 PM To: Gregory Farnum Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] add crush rule in one command On Fri, Jul 26, 2013 at 2:27 PM, Rongze Zhu <ron...@unitedstack.com<mailto:ron...@unitedstack.com>> wrote: On Fri, Jul 26, 2013 at 1:22 PM, Gregory Farnum <g...@inktank.com<mailto:g...@inktank.com>> wrote: On Thu, Jul 25, 2013 at 7:41 PM, Rongze Zhu <ron...@unitedstack.com<mailto:ron...@unitedstack.com>> wrote: > Hi folks, > > Recently, I use puppet to deploy Ceph and integrate Ceph with OpenStack. We > put computeand storage together in the same cluster. So nova-compute and > OSDs will be in each server. We will create a local pool for each server, > and the pool only use the disks of each server. Local pools will be used by > Nova for root disk and ephemeral disk. Hmm, this is constraining Ceph quite a lot; I hope you've thought about what this means in terms of data availability and even utilization of your storage. :) We also will create global pool for Cinder, the IOPS of global pool will be betther than local pool. The benefit of local pool is reducing the network traffic between servers and Improving the management of storage. We use one same Ceph Gluster for Nova,Cinder,Glance, and create different pools(and diffenrent rules) for them. Maybe it need more testing :) s/Gluster/Cluster/g > In order to use the local pools, I need add some rules for the local pools > to ensure the local pools using only local disks. There is only way to add > rule in ceph: > > ceph osd getcrushmap -o crush-map > crushtool -c crush-map.txt -o new-crush-map > ceph osd setcrushmap -i new-crush-map > > If multiple servers simultaneously set crush map(puppet agent will do that), > there is the possibility of consistency problems. So if there is an command > for adding rule, which will be very convenient. Such as: > > ceph osd crush add rule -i new-rule-file > > Could I add the command into Ceph? We love contributions to Ceph, and this is an obvious hole in our atomic CLI-based CRUSH manipulation which a fix would be welcome for. Please be aware that there was a significant overhaul to the way these commands are processed internally between Cuttlefish and Dumpling-to-be that you'll need to deal with if you want to cross that boundary. I also recommend looking carefully at how we do the individual pool changes and how we handle whole-map injection to make sure the interface you use and the places you do data extraction makes sense. :) Thank you for your quick reply, it is very useful for me :) -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com -- Rongze Zhu - 朱荣泽 Email: zrz...@gmail.com<mailto:zrz...@gmail.com> Blog: http://way4ever.com Weibo: http://weibo.com/metaxen Github: https://github.com/zhurongze -- Rongze Zhu - 朱荣泽 Email: zrz...@gmail.com<mailto:zrz...@gmail.com> Blog: http://way4ever.com Weibo: http://weibo.com/metaxen Github: https://github.com/zhurongze
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com