You can reweight the OSD's either automatically based on utilization (ceph osd reweight-by-utilization) or by hand.
See: https://ceph.com/planet/ceph-osd-reweight/ http://docs.ceph.com/docs/master/rados/operations/control/#osd-subsystem It's probably not ideal to have OSD's of such different sizes on a node. ___ John Petrini NOC Systems Administrator // *CoreDial, LLC* // coredial.com // [image: Twitter] <https://twitter.com/coredial> [image: LinkedIn] <http://www.linkedin.com/company/99631> [image: Google Plus] <https://plus.google.com/104062177220750809525/posts> [image: Blog] <http://success.coredial.com/blog> Hillcrest I, 751 Arbor Way, Suite 150, Blue Bell PA, 19422 *P: *215.297.4400 x232 // *F: *215.297.4401 // *E: * jpetr...@coredial.com [image: Exceptional people. Proven Processes. Innovative Technology. Discover CoreDial - watch our video] <http://cta-redirect.hubspot.com/cta/redirect/210539/4c492538-6e4b-445e-9480-bef676787085> The information transmitted is intended only for the person or entity to which it is addressed and may contain confidential and/or privileged material. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited. If you received this in error, please contact the sender and delete the material from any computer. On Fri, Dec 2, 2016 at 12:36 AM, Волков Павел (Мобилон) <vol...@mobilon.ru> wrote: > Good day. > > I have set up the repository ceph and created several pools on the hdd > 4TB. My problem lies in uneven filling hdd. > > > > root@ceph-node1:~# df -H > > Filesystem Size Used Avail Use% Mounted on > > /dev/sda1 236G 2.7G 221G 2% / > > none 4.1k 0 4.1k 0% /sys/fs/cgroup > > udev 30G 4.1k 30G 1% /dev > > tmpfs 6.0G 1.1M 6.0G 1% /run > > none 5.3M 0 5.3M 0% /run/lock > > none 30G 8.2k 30G 1% /run/shm > > none 105M 0 105M 0% /run/user > > */dev/sdf1 4.0T 1.7T 2.4T 42% /var/lib/ceph/osd/ceph-4* > > /dev/sdg1 395G 329G 66G 84% /var/lib/ceph/osd/ceph-5 > > /dev/sdi1 195G 152G 44G 78% /var/lib/ceph/osd/ceph-7 > > */dev/sdd1 4.0T 1.7T 2.4T 41% /var/lib/ceph/osd/ceph-2* > > /dev/sdh1 395G 330G 65G 84% /var/lib/ceph/osd/ceph-6 > > */dev/sdb1 4.0T 1.9T 2.2T 46% /var/lib/ceph/osd/ceph-0* > > */dev/sde1 4.0T 2.1T 2.0T 51% /var/lib/ceph/osd/ceph-3* > > */dev/sdc1 4.0T 1.8T 2.3T 45% /var/lib/ceph/osd/ceph-1* > > > > > > On the test machine, this leads to an overflow error CDM and further > incorrect operation. > > How to make that all hdd filled equally? > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com