So you're running two workers on the same machine (10.0.0.4), is that correct? Normally you'd run one worker per machine unless there was a particular reason otherwise. What version of Apache Kafka are you using? I'm not clear from your question if the distribution of tasks is presenting a problem to you (if so please describe why), or if you're just interested in the theory behind the rebalancing protocol?
-- Robin Moffatt | Senior Developer Advocate | ro...@confluent.io | @rmoff On Wed, 20 May 2020 at 06:46, Deepak Raghav <deepakragha...@gmail.com> wrote: > Hi > > Please, can anybody help me with this? > > Regards and Thanks > Deepak Raghav > > > > On Tue, May 19, 2020 at 1:37 PM Deepak Raghav <deepakragha...@gmail.com> > wrote: > > > Hi Team > > > > We have two worker node in a cluster and 2 connector with having 10 tasks > > each. > > > > Now, suppose if we have two kafka connect process W1(Port 8080) and > > W2(Port 8078) started already in distribute mode and then register the > > connectors, task of one connector i.e 10 tasks are divided equally > between > > two worker i.e first task of A connector to W1 worker node and sec task > of > > A connector to W2 worker node, similarly for first task of B connector, > > will go to W1 node and sec task of B connector go to W2 node. > > > > e.g > > *#First Connector : * > > { > > "name": "REGION_CODE_UPPER-Cdb_Dchchargeableevent", > > "connector": { > > "state": "RUNNING", > > "worker_id": "10.0.0.4:*8080*" > > }, > > "tasks": [ > > { > > "id": 0, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:*8078*" > > }, > > { > > "id": 1, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 2, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 3, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 4, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 5, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 6, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 7, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 8, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 9, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > } > > ], > > "type": "sink" > > } > > > > > > *#Sec connector* > > > > { > > "name": "REGION_CODE_UPPER-Cdb_Neatransaction", > > "connector": { > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > "tasks": [ > > { > > "id": 0, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 1, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 2, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 3, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 4, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 5, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 6, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 7, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 8, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 9, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > } > > ], > > "type": "sink" > > } > > > > But I have seen a strange behavior, when I just shutdown W2 worker node > > and start it again task are divided but in diff way i.e all the tasks of > A > > connector will get into W1 node and tasks of B Connector into W2 node. > > > > Can you please have a look for this. > > > > *#First Connector* > > > > { > > "name": "REGION_CODE_UPPER-Cdb_Dchchargeableevent", > > "connector": { > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > "tasks": [ > > { > > "id": 0, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 1, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 2, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 3, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 4, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 5, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 6, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 7, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 8, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > }, > > { > > "id": 9, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8080" > > } > > ], > > "type": "sink" > > } > > > > *#Second Connector *: > > > > { > > "name": "REGION_CODE_UPPER-Cdb_Neatransaction", > > "connector": { > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > "tasks": [ > > { > > "id": 0, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 1, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 2, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 3, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 4, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 5, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 6, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 7, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 8, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > }, > > { > > "id": 9, > > "state": "RUNNING", > > "worker_id": "10.0.0.4:8078" > > } > > ], > > "type": "sink" > > } > > > > > > Regards and Thanks > > Deepak Raghav > > > > >