Update: As of now, setting the ports manually in each supervisor and
restarting the supervisors and nimbus made it work.
It doesn't seem to work if you simply expect the default 4 nodes to get
used.

On Tue, Jul 19, 2016 at 1:00 PM, Navin Ipe <[email protected]>
wrote:

> setNumWorkers is already set to 1 for all topologies.
> The problem I'm facing is the same as this person:
> https://groups.google.com/forum/#!topic/storm-user/wLOq1nImRWQ
> I have 5 AWS supervisor nodes running, and theoretically, there should be
> 5*4=20 worker slots. But when I submit 5 topologies on the 6th AWS server
> that contains Nimbus, only 4 topologies get one worker each. The 5th
> topology gets 0 workers.
>
> On Tue, Jul 19, 2016 at 12:56 PM, Sai Dilip Reddy Kiralam <
> [email protected]> wrote:
>
>>
>> what I understood from side is you want to use all the supervisor
>> slots.if i'm right
>> problem will be on coding side.you will specify how many workers you need
>> to run the topology.once check Set workers statement.
>>
>> conf.setNumWorkers(2);
>>
>>
>>
>>
>>
>> *Best regards,*
>>
>> *K.Sai Dilip Reddy.*
>>
>> On Tue, Jul 19, 2016 at 12:43 PM, Navin Ipe <
>> [email protected]> wrote:
>>
>>> There is another person who has faced the same problem earlier:
>>> https://groups.google.com/forum/#!topic/storm-user/wLOq1nImRWQ
>>>
>>> *This is what I've found out till now:*
>>>
>>>    - Theoretically, having 4 supervisors and 4 default slots on each,
>>>    should give you 4*4=16 workers (but in reality, that does not work).
>>>    - Every worker is a JVM (the slots.ports)
>>>    - Number of workers = max number of slots a topology can occupy
>>>    - Single worker JVM executes code of a single topology
>>>
>>>
>>> Is there anyone who has experienced this problem or knows how to solve
>>> it?
>>>
>>>
>>> On Tue, Jul 19, 2016 at 10:16 AM, Navin Ipe <
>>> [email protected]> wrote:
>>>
>>>> Any help please?
>>>>
>>>> On Mon, Jul 18, 2016 at 3:49 PM, Navin Ipe <
>>>> [email protected]> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> On my local system I was able to run 5 topologies only when I
>>>>> increased the number of slots in storm.yaml, to 5.
>>>>>
>>>>> But when submitting my jar to storm on an Amazon node, even though I
>>>>> had 5 supervisor nodes running, only 4 of my topologies were assigned a
>>>>> worker. The 5th topology got 0 workers.
>>>>>
>>>>> Why is this happening? Shouldn't each supervisor have 4 worker slots
>>>>> available by default? So 5 supervisors would have 5*4=20 slots?
>>>>>
>>>>> I've read that I'm expected to configure the slots in the storm.yaml
>>>>> of every supervisor and restart the supervisor. So if I configure each
>>>>> supervisor to have 5 slots, will I be able to run 5*5=25 topologies?
>>>>>
>>>>> How exactly does this slots concept work and what was it meant for?
>>>>>
>>>>> --
>>>>> Regards,
>>>>> Navin
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Regards,
>>>> Navin
>>>>
>>>
>>>
>>>
>>> --
>>> Regards,
>>> Navin
>>>
>>
>>
>
>
> --
> Regards,
> Navin
>



-- 
Regards,
Navin

Reply via email to