I ran a cluster with 2 generations of the same vendor hardware. 24 osd supermicro and 32 osd supermicro (with faster/more RAM and CPU cores). The cluster itself ran decently well, but the load differences was drastic between the 2 types of nodes. It required me to run the cluster with 2 separate config files for each type of node and was an utter PITA when troubleshooting bottlenecks.
Ultimately I moved around hardware and created a legacy cluster on the old hardware and created a new cluster using the newer configuration. In general it was very hard to diagnose certain bottlenecks due to everything just looking so different. The primary one I encountered was snap trimming due to deleted thousands of snapshots/day. If you aren't pushing any limits of Ceph, you will probably be fine. But if you have a really large cluster, use a lot of snapshots, or are pushing your cluster harder than the average user... Then I'd avoid mixing server configurations in a cluster. On Fri, Jun 9, 2017, 1:36 AM Deepak Naidu <dna...@nvidia.com> wrote: > Wanted to check if anyone has a ceph cluster which has mixed vendor > servers both with same disk size i.e. 8TB but different count i.e. Example > 10 OSD servers from Dell with 60 Disk per server and other 10 OSD servers > from HP with 26 Disk per server. > > If so does that change any performance dynamics ? or is it not advisable . > > -- > Deepak > > ----------------------------------------------------------------------------------- > This email message is for the sole use of the intended recipient(s) and > may contain > confidential information. Any unauthorized review, use, disclosure or > distribution > is prohibited. If you are not the intended recipient, please contact the > sender by > reply email and destroy all copies of the original message. > > ----------------------------------------------------------------------------------- > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com