thanks for your reply, Kirsch. yes , i have analyzed using MAT, the output
for memory leak is as follows:
[image: 内嵌图片 1]
2016-08-26 14:34 GMT+08:00 Harald Kirsch :
> There is hardly any way anyone can guess what happens there from just the
> numbers.
>
> What you should do is start Kafka with -X
There is hardly any way anyone can guess what happens there from just
the numbers.
What you should do is start Kafka with -XX:+HeapDumpOnOutOfMemoryError,
possibly even reduce the Xmx to 500MB and let it bomb out. Then you take
a look at the generated heap dump with the Eclipse Memory Analyzer
is there anybody can help me?
2016-08-23 8:49 GMT+08:00 黄川 :
> Hi, I am using kafka_2.11-0.9.0.1, there are multiple of warnings occur
> in kafkaServer.out:
> java.lang.OutOfMemoryError: Java heap space.
>
>
>- the jstat output like this:
>
> # jstat -gc 28591
> S0CS1CS0US1U
hi, guys
I have a three nodes kafka cluster.
Some times, one of the Kafka brokers may disappear from zookeeper
"/brokers/ids", but the process is still live and prints logs seems normal.
Have anyone of you also saw this problem? Is this a bug?
I suspect when session expired, the kafka broker do no
Hello Mate,
Thanks for your detailed response and it surely helps.
WhiteList is the required config for MM from 0.9.0 onwards. And you are
correct that --new-consumer requires --bootstrap-servers rather than
--zookeeper .
However, did you notice that MM picks the topics which are present at the
t
Hi all,
I am using kafka 0.10.0.1, and I set up my listeners like:
listeners=PLAINTEXT://myhostName:9092
then I have one consumer going using the new api. However, I did not see
anything return for the api.
The log from kafka is:
[2016-08-26 14:39:28,548] INFO [GroupCoordinator 0]: Preparing to
Bryan,
https://issues.apache.org/jira/browse/KAFKA-3410 reported a similar issue
but only happened when the leader broker's log was manually deleted. In
your case, was there any data loss in the broker due to things like power
outage?
Thanks,
Jun
On Tue, Aug 23, 2016 at 9:00 AM, Bryan Baugher
Hi Jan,
It seems your main concern is for the changed behavior of time based log
rolling and time based retention. That is actually why we have two
timestamp types. If user set the log.message.timestamp.type to
LogAppendTime, the broker will behave exactly the same as they were, except
the rolling
Hello Bryan,
I think you were encountering
https://issues.apache.org/jira/browse/KAFKA-3410. Maybe you can take a look
on this ticket and see if it matches your scenario.
Guozhang
On Tue, Aug 23, 2016 at 9:00 AM, Bryan Baugher wrote:
> Hi everyone,
>
> Yesterday we had lots of network failures
Hello,
I'm trying to understand best practices related to joining streams using
the Kafka Streams API.
I can configure the topology such that two sources feed into a single
processor:
topologyBuilder
.addSource("A", stringDeserializer, itemDeserializer, "a-topic")
.addSource("B", stringD
Just want to add something:
I you use Kafka Streams DSL, the library is Kafka centric.
However, you could use low-level Processor API to get data into your
topology from other systems. The problem will be missing fault-tolerance
that you would need to code by yourself. When reading from Kafka,
fa
Jan,
Thanks a lot for the feedback. Now I understood your concern better. The
following are my comments.
The first odd thing that you pointed out could be a real concern.
Basically, if a producer publishes messages with really old timestamp, our
default log.roll.hours (7 days) will indeed cause t
Hello:
We are running a data pipeline application stack using Kafka 0.8.2.2 in
production. We have been seeing intermittent CLOSE_WAIT on our kafka
brokers frequently and they fill up the file handles pretty quickly. By the
time the open file count reaches around 40K, the node becomes unresponsive
Does anybody do the following in production? If so, what are your experiences?
1. Use .Net applications for producers or consumers
2. Consume messages across the WAN (across datacenters) - I'm wondering
if MirrorMaker is always a requirement for cross-WAN
-Dave
This e-mail and any
kafka-python includes some benchmarking scripts in
https://github.com/dpkp/kafka-python/tree/master/benchmarks
The concurrency and execution model of the JVM are both significantly
different than python. I would definitely recommend some background reading
on CPython GIL if you are interested on p
Good question. All of them would run in a single thread. That is the model.
Multiple threads would make sense to run separate DAGs.
Eno
> On 25 Aug 2016, at 18:32, Abhishek Agarwal wrote:
>
> Hi Eno,
>
> Thanks for your reply. If my application DAG has three stream processors,
> first of wh
I think what Dana is suggesting is that since Python isn't doing a good job
utilising all the available CPU power, you could run multiple python
processes to process the load. Divide the mongodb collection between, say,
4 parts and process each part with one python process. On kafka side.
Or use a
Hi Eno,
Thanks for your reply. If my application DAG has three stream processors,
first of which is source, would all of them run in single thread? There may
be scenarios wherein I want to have different number of threads for
different processors since some may be CPU bound and some may be IO boun
Hi Abhishek,
- Correct on connecting to external stores. You can use Kafka Connect to get
things in or out. (Note that in the 0.10.1 release KIP-67 allows you to
directly query Kafka Stream's stores so, for some kind of data you don't need
to move it to an external store. This is pushed in trun
I would write a python client that writes dummy data to kafka to measure how
fast you can write to Kafka without MongoDB in the mix. I've been doing load
testing recently can with 3 brokers I can write 100MB/s (using Java clients).
-Dave
-Original Message-
From: Dominik Safaric [mailto
Dear Dana,
> I would recommend
> other tools for bulk transfers.
What tools/languages would you rather recommend then using Python?
I could for sure accomplish the same by using the native Java Kafka Producer
API, but should this really affect the performance under the assumption that
the Ka
python is generally restricted to a single CPU, and kafka-python will max
out a single CPU well before it maxes a network card. I would recommend
other tools for bulk transfers. Otherwise you may find that partitioning
your data set and running separate python processes for each will increase
the o
A few days ago, I’ve started migrating for the purpose of a benchmark onto
stream processing engines an entire collection from MongoDB to a Kafka log.
In summary, the MongoDB collection contains approximately 560 million documents
of mean size 2529 bytes that I am in time of writing this still
Hi,
I was reading up on kafka streams for a project and came across this blog
https://softwaremill.com/kafka-streams-how-does-it-fit-stream-landscape/
I wanted to validate some assertions made in blog, with kafka community
- Kafka streams is kafka-in, kafka-out application. Does the user need
kafk
Hi Umesh,
I am new to kafka as well, and configuring the MirrorMaker. I got mine
working in the following way.
I run the mirror maker instance on the mirror cluster, as in where you want
to mirror the topics to, although I'm not sure it matters.
I use the following options when starting my servi
enable debug logs and check for any exceptions/error messages
log4j file : kafka/config/tools-log4j.properties
On Thu, Aug 25, 2016 at 6:41 PM, Erik Parienty
wrote:
> Thanks a lot but now I get nothing :-)
>
> The command I am running:
> ./kafka-mirror-maker.sh --new.consumer --consumer.confi
Thanks a lot but now I get nothing :-)
The command I am running:
./kafka-mirror-maker.sh --new.consumer --consumer.config
consumer_ssl.properties --producer.config producer_ssl.properties --whitelist
=".*"
consumer_ssl.properties:
bootstrap.servers=XX:9093
group.id=securing-kafka-group
Security is supported in new Consumer API. use "--new.consumer" option to
use enable new consumer inside MirrorMaker.
On Thu, Aug 25, 2016 at 6:08 PM, Erik Parienty
wrote:
> As I understand mirror-maker support consumer ssl
> I tried to set it but I get WARN Property security.protocol is not va
As I understand mirror-maker support consumer ssl
I tried to set it but I get WARN Property security.protocol is not valid
(kafka.utils.VerifiableProperties)
And its connecting without ssl
Hey Folks,
I was trying to understand the behavior of MirrorMaker but looks like I am
missing something here. Please see the steps which I performed :
1) I configured MM on source Kafka cluster
2) Created a topic and pushed some data in it using console producer.
3) My understanding is that MM wou
30 matches
Mail list logo