Hi Chuck, Thanks for the reply. The reason for such weight distribution seems to do with the ring rebalance command. I've scripted the disk addition (and rebalance) process to the ring using a wrapper command. When I trigger the rebalance after each disk addition, only the first rebalance seems to take effect.
Is there any other way to adjust the weights other than rebalance? Or is there a way to force a rebalance, even if the frequency of the rebalance (as a part of disk addition) is under an hour (the min_part_hours value in ring creation). On May 1, 2014 9:00 PM, "Chuck Thier" <cth...@gmail.com> wrote: > Hi Shyam, > > If I am reading your ring output correctly, it looks like only the devices > in node .202 have a weight set, and thus why all of your objects are going > to that one node. You can update the weight of the other devices, and > rebalance, and things should get distributed correctly. > > -- > Chuck > > > On Thu, May 1, 2014 at 5:28 AM, Shyam Prasad N <nspmangal...@gmail.com>wrote: > >> Hi, >> >> I created a swift cluster and configured the rings like this... >> >> swift-ring-builder object.builder create 10 3 1 >> >> ubuntu-202:/etc/swift$ swift-ring-builder object.builder >> object.builder, build version 12 >> 1024 partitions, 3.000000 replicas, 1 regions, 4 zones, 12 devices, >> 300.00 balance >> The minimum number of hours before a partition can be reassigned is 1 >> Devices: id region zone ip address port replication ip >> replication port name weight partitions balance meta >> 0 1 1 10.3.0.202 6010 >> 10.3.0.202 6010 xvdb 1.00 1024 300.00 >> 1 1 1 10.3.0.202 6020 >> 10.3.0.202 6020 xvdc 1.00 1024 300.00 >> 2 1 1 10.3.0.202 6030 >> 10.3.0.202 6030 xvde 1.00 1024 300.00 >> 3 1 2 10.3.0.212 6010 >> 10.3.0.212 6010 xvdb 1.00 0 -100.00 >> 4 1 2 10.3.0.212 6020 >> 10.3.0.212 6020 xvdc 1.00 0 -100.00 >> 5 1 2 10.3.0.212 6030 >> 10.3.0.212 6030 xvde 1.00 0 -100.00 >> 6 1 3 10.3.0.222 6010 >> 10.3.0.222 6010 xvdb 1.00 0 -100.00 >> 7 1 3 10.3.0.222 6020 >> 10.3.0.222 6020 xvdc 1.00 0 -100.00 >> 8 1 3 10.3.0.222 6030 >> 10.3.0.222 6030 xvde 1.00 0 -100.00 >> 9 1 4 10.3.0.232 6010 >> 10.3.0.232 6010 xvdb 1.00 0 -100.00 >> 10 1 4 10.3.0.232 6020 >> 10.3.0.232 6020 xvdc 1.00 0 -100.00 >> 11 1 4 10.3.0.232 6030 >> 10.3.0.232 6030 xvde 1.00 0 -100.00 >> >> Container and account rings have a similar configuration. >> Once the rings were created and all the disks were added to the rings >> like above, I ran rebalance on each ring. (I ran rebalance after adding >> each of the node above.) >> Then I immediately scp the rings to all other nodes in the cluster. >> >> I now observe that the objects are all going to 10.3.0.202. I don't see >> the objects being replicated to the other nodes. So much so that 202 is >> approaching 100% disk usage, while other nodes are almost completely empty. >> What am I doing wrong? Am I not supposed to run rebalance operation after >> addition of each disk/node? >> >> Thanks in advance for the help. >> >> -- >> -Shyam >> >> _______________________________________________ >> OpenStack-dev mailing list >> OpenStack-dev@lists.openstack.org >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev >> >> > > _______________________________________________ > OpenStack-dev mailing list > OpenStack-dev@lists.openstack.org > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev > >
_______________________________________________ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev