We don't have any compression on Kafka

On 14 January 2015 at 22:54, Guozhang Wang <wangg...@gmail.com> wrote:

> Did you have compression enabled on Kafka?
>
> On Tue, Jan 13, 2015 at 10:33 AM, dinesh kumar <dinesh...@gmail.com>
> wrote:
>
> > We are using 0.8.1.1 version of Kafka and *not 0.8.2 *as mentioned above.
> >
> > Thanks,
> > Dinesh
> >
> > On 13 January 2015 at 23:35, dinesh kumar <dinesh...@gmail.com> wrote:
> >
> > > Hi Guozhang,
> > > Sorry for the misinformation. We have file sizes around 50 - 100 MB. So
> > we
> > > set *fetch.message.max.bytes* conservatively around 188743680.  Can you
> > > please explain me the reason for this behavior?
> > >
> > > Thanks,
> > > Dinesh
> > >
> > > On 13 January 2015 at 21:42, Guozhang Wang <wangg...@gmail.com> wrote:
> > >
> > >> Dinesh,
> > >>
> > >> Your fetch.message.max.bytes is 188743680 < 155MB, but you said some
> > >> messages can be as large as 180MB. Could you try to set it to be
> larger
> > >> than, say 200MB and see if it helps?
> > >>
> > >> Guozhang
> > >>
> > >> On Tue, Jan 13, 2015 at 4:18 AM, dinesh kumar <dinesh...@gmail.com>
> > >> wrote:
> > >>
> > >> > Hi,
> > >> > I am been facing some JAVA high level consumer related issues lately
> > and
> > >> > would like to understand more on this.
> > >> >
> > >> > We have 9 bare-metals (48 core, 250 GB, Terabytes of Hard disks)
> > running
> > >> > *Kafka
> > >> > 0.8.2* and 5 independent VM (8 core, 60 GB) running zookeeper.
> > >> >
> > >> > I have a topic that has key as metadata and value as a file. The
> file
> > >> can
> > >> > be as large as *180 MB.* We have a topic with 90 partitions.
> Sometimes
> > >> > there will be only one consumer consuming from the topic. When the
> > >> consumer
> > >> > group for my topic has a *lag in the range of 200's* and when I
> start
> > a
> > >> > consumer (no other consumer running before) there is *no data*
> coming
> > >> > through to the consumer.
> > >> >
> > >> > Please find below my consumer parameters.
> > >> >
> > >> > "zookeeper.connect"                => <zookeepers>,
> > >> > "group.id"                         => "default",
> > >> > "consumer.timeout.ms"              => "-1",
> > >> > "auto.offset.reset"                => "smallest",
> > >> > "auto.commit.enable"               => "false",
> > >> > "consumer.timeout.ms"          => "-1",
> > >> > "zookeeper.session.timeout.ms" => "100000",
> > >> > "zookeeper.connection.timeout.ms"  => "6000",
> > >> > "zookeeper.sync.time.ms"           => "2000",
> > >> > "rebalance.backoff.ms"             =>  "20000",
> > >> > "rebalance.max.retries"            => "50"
> > >> > "fetch.message.max.bytes"      => "188743680",
> > >> > "fetch.size"                   => "18874368"
> > >> >
> > >> >
> > >> >
> > >> > This problem occurs only when the *auto.offset.reset *property is
> > >> > *smallest.
> > >> > *I am able to get data if the offset is largest. I tried using the
> > >> *console
> > >> > consumer* for the same topic and consumer group with
> > *--from-beginning*
> > >> > option, I can see the data getting printed. I looked into the
> > >> > ConsoleConsumer code and I saw that there was no
> > >> > *fetch.message.max.bytes *property
> > >> > in the consumer option.
> > >> >
> > >> > So I removed the *fetch.message.max.bytes *from my code and the
> > consumer
> > >> > started working but was throwing exception when the message is
> large.
> > >> >
> > >> > So *fetch.message.max.bytes *seemed to be the problem but I cannot
> do
> > >> > without it as my messages a big files. Can someone explain to me
> what
> > is
> > >> > the issue here? I also adjusted the *fetch.size *parameter according
> > to
> > >> my
> > >> > max message size but it did not help.
> > >> >
> > >> >
> > >> > To summerize, I would like to understand what is happening in the
> > >> consumer
> > >> > end when handling large lags with big *fetch.message.max.bytes. *
> > >> >
> > >> >
> > >> > Thanks,
> > >> > Dinesh
> > >> >
> > >>
> > >>
> > >>
> > >> --
> > >> -- Guozhang
> > >>
> > >
> > >
> >
>
>
>
> --
> -- Guozhang
>

Reply via email to