Hi Tim,

    Thanks for your response.
    The results are the same.
        4 CPU (*8 cores in total)
        kafka partitions = 4 per topic
        parallesim for job = 3
        task.slot / TM = 4

    Basically this flink application consumes (kafka source) from 2 topics
and produces (kafka sink) onto 1 topic. on 1 consumer topic, the event load
is 100K/sec, while the other source has 1 event / an hour ...
    I'm wondering if parallelism is enabled on multiple sources
irrespective of the partition size.

    What I did is to enable 1 partition for the 2nd topic (1 event/hour)
and 4 partitions for 100K events topic. And deployed  a 3 parallelism job
and the results are the same...

Best Regards
CVP

On Wed, Jan 11, 2017 at 1:11 PM, Till Rohrmann <trohrm...@apache.org> wrote:

> Hi CVP,
>
> changing the parallelism from 1 to 2 with every TM having only one slot
> will inevitably introduce another network shuffle operation between the
> sources and the keyed co flat map. This might be the source of your slow
> down, because before everything was running on one machine without any
> network communication (apart from reading from Kafka).
>
> Do you also observe a further degradation when increasing the parallelism
> from 2 to 4, for example (given that you've increased the number of topic
> partitions to at least the maximum parallelism in your topology)?
>
> Cheers,
> Till
>
> On Tue, Jan 10, 2017 at 11:37 AM, Chakravarthy varaga <
> chakravarth...@gmail.com> wrote:
>
>> Hi Guys,
>>
>>     I understand that you are extremely busy but any pointers here is
>> highly appreciated. I can proceed forward towards concluding the activity !
>>
>> Best Regards
>> CVP
>>
>> On Mon, Jan 9, 2017 at 11:43 AM, Chakravarthy varaga <
>> chakravarth...@gmail.com> wrote:
>>
>>> Anything that I could check or collect for you for investigation ?
>>>
>>> On Sat, Jan 7, 2017 at 1:35 PM, Chakravarthy varaga <
>>> chakravarth...@gmail.com> wrote:
>>>
>>>> Hi Stephen
>>>>
>>>> . Kafka version is: 0.9.0.1 the connector is flinkconsumer09
>>>> . The flatmap n coflatmap are connected by keyBy
>>>> . No data is broadcasted and the data is not exploded based on the
>>>> parallelism
>>>>
>>>> Cvp
>>>>
>>>> On 6 Jan 2017 20:16, "Stephan Ewen" <se...@apache.org> wrote:
>>>>
>>>>> Hi!
>>>>>
>>>>> You are right, parallelism 2 should be faster than parallelism 1 ;-)
>>>>> As ChenQin pointed out, having only 2 Kafka Partitions may prevent further
>>>>> scaleout.
>>>>>
>>>>> Few things to check:
>>>>>   - How are you connecting the FlatMap and CoFlatMap? Default, keyBy,
>>>>> broadcast?
>>>>>   - Broadcast for example would multiply the data based on
>>>>> parallelism, can lead to slowdown when saturating the network.
>>>>>   - Are you using the standard Kafka Source (which Kafka version)?
>>>>>   - Is there any part in the program that multiplies data/effort with
>>>>> higher parallelism (does the FlatMap explode data based on parallelism)?
>>>>>
>>>>> Stephan
>>>>>
>>>>>
>>>>> On Fri, Jan 6, 2017 at 7:27 PM, Chen Qin <qinnc...@gmail.com> wrote:
>>>>>
>>>>>> Just noticed there are only two partitions per topic. Regardless of
>>>>>> how large parallelism set. Only two of those will get partition assigned 
>>>>>> at
>>>>>> most.
>>>>>>
>>>>>> Sent from my iPhone
>>>>>>
>>>>>> On Jan 6, 2017, at 02:40, Chakravarthy varaga <
>>>>>> chakravarth...@gmail.com> wrote:
>>>>>>
>>>>>> Hi All,
>>>>>>
>>>>>>     Any updates on this?
>>>>>>
>>>>>> Best Regards
>>>>>> CVP
>>>>>>
>>>>>> On Thu, Jan 5, 2017 at 1:21 PM, Chakravarthy varaga <
>>>>>> chakravarth...@gmail.com> wrote:
>>>>>>
>>>>>>>
>>>>>>> Hi All,
>>>>>>>
>>>>>>> I have a job as attached.
>>>>>>>
>>>>>>> I have a 16 Core blade running RHEL 7. The taskmanager default
>>>>>>> number of slots is set to 1. The source is a kafka stream and each of 
>>>>>>> the 2
>>>>>>> sources(topic) have 2 partitions each.
>>>>>>>
>>>>>>>
>>>>>>> *What I notice is that when I deploy a job to run with
>>>>>>> #parallelism=2 the total processing time doubles the time it took when 
>>>>>>> the
>>>>>>> same job was deployed with #parallelism=1. It linearly increases with 
>>>>>>> the
>>>>>>> parallelism.*
>>>>>>> Since the numberof slots is set to 1 per TM, I would assume that the
>>>>>>> job would be processed in parallel in 2 different TMs and that each
>>>>>>> consumer in each TM is connected to 1 partition of the topic. This
>>>>>>> therefore should have kept the overall processing time the same or less 
>>>>>>> !!!
>>>>>>>
>>>>>>> The co-flatmap connects the 2 streams & uses ValueState
>>>>>>> (checkpointed in FS). I think this is distributed among the TMs. My
>>>>>>> understanding is that the search of values state could be costly between
>>>>>>> TMs.  Do you sense something wrong here?
>>>>>>>
>>>>>>> Best Regards
>>>>>>> CVP
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>
>>
>

Reply via email to