Might this message be more appropriate for the developer mailing list? Is
it okay to double post in this case?
Thanks again, Dominic.
On Jun 12, 2016 00:59, "Dominic Chambers"
wrote:
I'd like to add a notification API to Kafka so that I can receive the
following notifications for each broker in
Hi Joe,
Thanks for your feedback.
Compatibility is a complex subject and the exact details need to be defined
for each project. Technically, one could claim that making changes so that
a 0.10.1.0 client doesn't work with a 0.10.0.0 broker would be an
incompatible change and should not be allowed,
The consumer instances close I.e leave the group only if they are idle for a
long time..we have expiration threads which monitor this and remove any
consumer instances if they keep sitting . Also , consumers are closed when the
application is shut down. The poll() does receive around 481 record
Rohit,
The 30s number sounds very suspicious because it is exactly the value of
the session timeout. But if you are driving the consumer correctly, you
shouldn't normally hit this timeout. Dana was asking about consumers
leaving gracefully because that is one case where you can inadvertently
trigg
Can anybody help out on this?
From: Rohit Sardesai
Sent: 19 June 2016 11:47:01
To: users@kafka.apache.org
Subject: Re: consumer.poll() takes approx. 30 seconds - 0.9 new consumer api
In my tests , I am using around 24 consumer groups. I never call
consumer.clos
You can adjust request.timeout.ms, which is shared between both new
producer and new consumer. I don't think its quite what you want, but
probably the closest that exists across both clients. There's not much more
than that -- when you say "when the connection to the entire broker cluster
is lost"
Hi Ismael,
Agree on timing is more important. If we give enough heads
up to the users who are on Java 7 thats great but still
shipping this in 0.10.x line is won't be good as it still
perceived as maint release even the release might contain
Great, glad you sorted it out. If the namespace is being omitted
incorrectly from the request the connector is making, please file a bug
report -- I can't think of a reason we'd omit that, but it's certainly
possible it is a bug on our side.
-Ewen
On Wed, Jun 15, 2016 at 7:08 AM, Tauzell, Dave w
The most common use case for Kafka is within a data center, but you can
absolutely produce data across the WAN. You may need to adjust some
settings (e.g. timeouts, max in flight requests per connection if you want
high throughput) to account for operating over the WAN, but you can
definitely do it
Hi Chris,
We should also ensure that auto.create.topics.enable is set to true.
Thank you,
Anirudh
Hi Chris,
If the topic not exist, it will create a new topic with the name which you
give.
Thanks,
Nicole
On Sat, Jun 18, 2016 at 1:55 AM, Chris Barlock wrote:
> If you have a consumer listening
Hi Chris,
If the topic not exist, it will create a new topic with the name which you
give.
Thanks,
Nicole
On Sat, Jun 18, 2016 at 1:55 AM, Chris Barlock wrote:
> If you have a consumer listening on a topic and that topic is deleted is
> the consumer made aware -- perhaps by some exception -- o
Hi Matthias,
I solved the problem with specifying the serders and reading source as
KStream instead of KTable. So, instead of
KTable source = builder.table("topic1");
I added:
KStream source =
builder.stream(longSerde,stringSerde,"topic1");
Thanks
-Adrienne
On Sun,
Hi Yi,
Your observation about accessing the state stores that are already there vs.
keeping state outside of Kafka Streams is a good one. We are currently working
on having the state stores accessible like you mention and should be able to
share some design docs shortly.
Thanks
Eno
> On 19 Ju
Hello,
I am thinking of using the Kafka Steams feature to "unify" our real-time
and scheduled workflow. An example is that in our workflow with stages A-->
B --> C, the A --> B segment can be achieved in real-time, but B-->C
segment is usually a done with a scheduled job, running maybe once per ho
Hello all
I use kafka input and kafka output plugin in logstash. I have high cpu
usage, what can I do to get it better?
Thanks a lot
Hello all
I use kafka input and kafka output plugin in logstash. I have high cpu
usage, what can I do to get it better?
logstash version 2.3.2
logstash-input-kafka 2.0.8
logstash-output-kafka 2.0.5
Thanks a lot
Looks like the producers lose the connection to the brokers.
Do the brokers have enough resources to handle all the producers? Does the
network support that throughput?
On Sun, 19 Jun 2016, 17:27 Avi Asulin, wrote:
> Hi
> We are using kafka 0.8.2 with scala 2.10 version
> We currently have 3 br
Hi
We are using kafka 0.8.2 with scala 2.10 version
We currently have 3 brokers and we are working with ~ 170 producers
We frequently get the Error
ERROR Closing socket for /170.144.181.50 because of error
(kafka.network.Processor)
java.io.IOException: Connection reset by peer
at sun.nio.c
Can you show the full stack trace?
How do you ingest the date into the topic? I also think, you should read
the topic as KStream (instead of KTable).
What de-/serializer do you specify in props. (see
http://docs.confluent.io/3.0.0/streams/developer-guide.html#streams-developer-guide-serdes)
-Ma
Hi,
I want to implement wordcount example with reduce function in KTable.
However, I get the error:
Exception in thread "StreamThread-1"
org.apache.kafka.common.errors.SerializationException: Size of data
received by LongDeserializer is not 8
Here is my code:
KTable source = builder.t
@jun Rao
about this question,can you give me some suggestion?
> 在 2016年6月18日,上午11:26,Kafka 写道:
>
> hello,I have done a series of tests on kafka 0.9.0,and one of the results
> confused me.
>
> test enviroment:
> kafka cluster: 3 brokers,8core cpu / 8g mem /1g netcard
> client:4core cpu
21 matches
Mail list logo