Questions about Kafka 0.9 API changes

2014-09-22 Thread Valentin
onsumer to switch topics/partitions to the currently needed set or would this be a very expensive operation (i.e. because it would fetch metadata from Kafka to identify the leader for each partition)? Greetings Valentin

Re: Questions about Kafka 0.9 API changes

2014-09-23 Thread Valentin
eConsumer. Greetings Valentin > 2) I am not very familiar with HTTP wrapper on the clients, could someone > who have done so comment here? > > 3) The new KafkaConsumer has not been fully implemented, but you can take a > look at its JavaDoc for examples of using the new client. >

Re: Questions about Kafka 0.9 API changes

2014-09-23 Thread Valentin
ood approach. All in all, the planned design of the new consumer API just doesn't seem to fit my use case well. Which is why I am a bit anxious about the SimpleConsumer API being deprecated. Or am I missing something here? Thanks! Greetings Valentin > On Mon, Sep 22, 2014 at 8:10 AM,

Re: Questions about Kafka 0.9 API changes

2014-09-23 Thread Valentin
ionale is for deprecating the SimpleConsumer API, if there are use cases which just work much better using it. Greetings Valentin On 23/09/14 18:16, Guozhang Wang wrote: > Hello, > > For your use case, with the new consumer you can still create a new > consumer instance for each topi

Re: Questions about Kafka 0.9 API changes

2014-09-25 Thread Valentin
well as my current SimpleConsumer approach :| Or am I missing something here? Greetings Valentin On Wed, 24 Sep 2014 17:44:15 -0700, Jun Rao wrote: > Valentin, > > As Guozhang mentioned, to use the new consumer in the SimpleConsumer way, > you would subscribe to a set of topic partiti

Re: Questions about Kafka 0.9 API changes

2014-09-28 Thread Valentin
Hi Jun, ok, I created: https://issues.apache.org/jira/browse/KAFKA-1655 Greetings Valentin On Sat, 27 Sep 2014 08:31:01 -0700, Jun Rao wrote: > Valentin, > > That's a good point. We don't have this use case in mind when designing the > new consumer api. A straightforwar

Message routing, Kafka-to-REST and HTTP API tools/frameworks for Kafka?

2015-03-24 Thread Valentin
level and TLS for transport encryption. Offset tracking is done by the connected systems, not by the broker/zookeeper/REST API. We expect a high message volume in the future here, so performance would be a key concern. Greetings Valentin

How can the only replica for a partition not be in the Isr list? (replication factor=1)

2015-04-08 Thread Valentin
--r-- 1 kafka kafka0 Apr 6 10:56 0266.log To fix the issue we had to change unclean.leader.election.enable from false to true and restart all 3 brokers. Is that really the intended approach in such a scenario? Greetings Valentin

Broken auto leader rebalance after using reassign partitions tool

2015-06-15 Thread Valentin
Valentin

Re: Broken auto leader rebalance after using reassign partitions tool

2015-06-16 Thread Valentin
uot;partition": 1, "replicas": [2,4,6] }, { "topic": "T5", "partition": 2, "replicas": [1,3,5] }, { "topic": "T5", "partition": 3, "replicas": [2,4,6] }, { "topic": "T5", "parti

Re: Kafka avro producer

2019-03-27 Thread Valentin
Hi Karishma, you can definitely use Avro without the Confluent schema registry. Just write you own serializer/ deserializer. However , you need to share the avro schema version between your producer and consumer somehow.. and also think about changes on your avro schema. Greets Valentin Von

Problem with jdbc sink connector

2019-04-16 Thread Valentin
alse", > "connection.url": "jdbc:oracle:thin:@X:16767/XXX", > "key.converter.schema.registry.url": "http://localhost:8081";, > "pk.mode": "record_value", > "pk.fields": "ID" > > > when I use "insert.mode": "insert" everything works fine, but when I switch > to „upsert“ I am only able to insert binary files „BLOBs“ up to 32767 bytes. > Otherwise I am getting: > SQLException: ORA-01461: can bind a LONG value only for insert into a LONG > column > > Do you have an idea how to solve this issue? Many thx in advance Valentin

Re: kafka issue

2019-12-09 Thread Valentin
called context() However, keep in mind broker and IP addresses could change. I would not recommend to build a business logic on top of ip addresses. Within a producer, you can manage the destination partition / broker. Regards Valentin -- > Am 10.12.2019 um 04:48 schrieb 朱超 : >

Using Kafka on DC/OS + Marathon

2017-10-02 Thread Valentin Forst
dedicated DC/OS instance for our Kafka-Cluster? Or Kafka-Cluster on its own? - Is there something else we should consider important if using Kafka on DC/OS + Marathon? Thanks in advance for your time. Valentin

Re: Using Kafka on DC/OS + Marathon

2017-10-03 Thread Valentin Forst
have a different DC/OS Cluster 3. Kafka -Cluster on its own Cheers, Valentin > Am 02.10.2017 um 16:35 schrieb David Garcia : > > I’m not sure how your requirements of Kafka are related to your requirements > for marathon. Kafka is a streaming-log system and marathon is a schedul

Re: Using Kafka on DC/OS + Marathon

2017-10-03 Thread Valentin Forst
Hi Avinash, Thanks for this hint. It would have been great, if someone could share experience using this framework on the production environment. Thanks in advance Valentin > Am 02.10.2017 um 19:39 schrieb Avinash Shahdadpuri : > > There is a a native kafka framework which runs on

Re: Using Kafka on DC/OS + Marathon

2017-10-03 Thread Valentin Forst
Hi Sean, Thanks a lot for this info ! Are you running DC/OS in prod? Regards Valentin > Am 03.10.2017 um 15:29 schrieb Sean Glover : > > Hi Valentin, > > Kafka is available on DC/OS in the Catalog (aka Universe) as part of the > `kafka` package. Mesosphere has put a

Kafka Streams with embedded RocksDB

2017-10-06 Thread Valentin Forst
ently I am getting the following error: UnsatisfiedLinkError … Can’t find dependency library. Trying to find rocksdbjni,…dll Thanks in advance Valentin

Re: Kafka Streams with embedded RocksDB

2017-10-07 Thread Valentin Forst
Valentin > Am 06.10.2017 um 18:30 schrieb Ted Yu : > > I assume you have read > https://github.com/facebook/rocksdb/wiki/Building-on-Windows > > Please also see https://github.com/facebook/rocksdb/issues/2531 > > BTW your question should be directed to rocksdb forum.

Re: kafka issue

2019-12-10 Thread Valentin Forst
> > Am 10.12.2019 um 06:36 schrieb Valentin : > > Hi Chao, > > I suppose, you would like to know: > within a consumer group which message is coming from which partition, since > partitions corresponds to broker and broker = ip, right? > > Well, if you really wa

Kafka Connect ignores the bootstrap.servers parameter

2020-04-27 Thread Valentin Kulichenko
Greetings to the Kafka Community! I'm a newbie in Kafka and only recently went beyond a local installation described in the Quickstart. I have faced a weird issue that I can't explain. I want to deploy on two machines: - Machine #1 runs ZooKeeper and a single Kafka broker. I use default configura

Re: Kafka Connect ignores the bootstrap.servers parameter

2020-04-28 Thread Valentin Kulichenko
about the environment but would appreciate any pointers in the meantime. -Val On Tue, Apr 28, 2020 at 12:07 AM Suresh Chidambaram wrote: > Hi Val, > > Could you share the server.properties and zookeeper.properties? > > Thanks > C Suresh > > > On Tuesday, Apr

Re: Kafka Connect ignores the bootstrap.servers parameter

2020-04-28 Thread Valentin Kulichenko
listeners=PLAINTEXT://:9092listeners=PLAINTEXT://98,1,96.147:9092> > > Link to article: > Kafka Listeners - Explained > > | > | > | > | | | > > | > > | > | > | | > Kafka Listeners - Explained > > How to connect clients to Kafka hosted

Kafka won't replicate from a specific broker

2016-12-22 Thread Valentin Golev
Hello, I have a three broker Kafka setup (the ids are 1, 2 (kafka 0.10.1.0) and 1001 (kafka 0.10.0.0)). After a failure of two of them a lot of the partitions have the third one (1001) as their leader. It's like this: Topic: userevents0.open Partition: 5Leader: 1 Replicas: 1,2,1001

Re: Kafka won't replicate from a specific broker

2016-12-22 Thread Valentin Golev
.10.1.1 not considered stable yet? I'm not sure about using it... Maybe downgrade would work? Re: restarting the faulty broker. As I understand, to avoid losing data, I'd have to shut down other parts of the cluster first, right? -Valentin On Thu, Dec 22, 2016 at 9:01 PM, Jan Omar

kafka 0.9.0 java version

2016-01-28 Thread Muresanu A.V. (Andrei Valentin)
Hi all, what is the oracle jdk version that is "supported" by kafka 0.9.0 ? 6/7/8... ATTENTION: The information in this e-mail is confidential and only meant for the intended recipient. If you are not the intended recipient , don'

RE: kafka 0.9.0 java version

2016-01-28 Thread Muresanu A.V. (Andrei Valentin)
://kafka.apache.org/documentation.html#java 2016-01-28 16:58 GMT+01:00 Muresanu A.V. (Andrei Valentin) < andrei.mures...@ing.ro>: > Hi all, > > what is the oracle jdk version that is "supported" b