Re: Very low bytesout for internal repartition topic, is it related to RocksDB?

2019-10-14 Thread Sophie Blee-Goldman
Glad that helped! Honestly I can't say I recognize either that exception but I'm fairly confident it's not directly related to rocksdb or Streams. It sounds like a connection somewhere got screwed up, which would be more of a configuration issue I probably can't help with. Of course it's always p

Re: Very low bytesout for internal repartition topic, is it related to RocksDB?

2019-10-14 Thread Xiyuan Hu
Hi Sophie, Thanks for the information! After setting cache.max.bytes.buffering to zero, commenting out my customized rocksDB config and using default rocksDB config, I do see the repartition topic bytesout goes up. But I noticed that some nodes have IO exception as: An I/O error has occurred while

Re: Very low bytesout for internal repartition topic, is it related to RocksDB?

2019-10-14 Thread Sophie Blee-Goldman
Ah ok, 2.3.0 has a known performance issue in the caching layer which tends to get worse the larger the cache size. That might explain what you're seeing with regards to the traffic correlation. It's fixed in 2.3.1 which should be released very soon, but until then you might want to try turning of

Re: Very low bytesout for internal repartition topic, is it related to RocksDB?

2019-10-14 Thread Xiyuan Hu
Hi Sophie, Thanks for the help! I'm using version 2.3.0. The repartition topic with huge lag is the created during the first reduce method, named "XX-KSTREAM-STATE-STORE-0030-repartition". All other internal topics have almost zero lags. For my case, how could I find out if rocksDB causes the lag

Re: Long delay between incoming and outgoing messages using kafka streams

2019-10-14 Thread Sophie Blee-Goldman
Honestly I can't say whether 256 partitions is enough to trigger the performance issues in 2.3.0 but I'd definitely recommend upgrading as soon as you can, just in case. On a related note, how many instances, with how many threads, are you running? 256 partitions with several subtopologies will res

Re: Very low bytesout for internal repartition topic, is it related to RocksDB?

2019-10-14 Thread Sophie Blee-Goldman
Out of curiosity, which version are you using? There's nothing that really jumps out at me as problematic in your RocksDBConfigSetter, but note that I think you may need to increase the number of threads in the "LOW priority thread pool" in addition to setting the maxBackgroundCompactions -- this

Re: Running a stand alone Mirror Maker 2.0 cluster and Issues

2019-10-14 Thread Ryanne Dolan
> timed out while waiting for producer to flush outstanding Yeah, that's what I'd expect to see if Connect was unable to send records to the downstream remote topics, e.g. if min.in-sync.replicas were misconfigured. Given some data seems to arrive, it's possible that everything is configured corr

Re: Running a stand alone Mirror Maker 2.0 cluster and Issues

2019-10-14 Thread Ryanne Dolan
You can disable ACL sync to avoid that error, but it doesn't step on anything else. That can't be it. Consider enabling TRACE or DEBUG logging within Connect to see what it's doing. Otherwise Connect tends to swallow exceptions and retry silently. Ryanne On Mon, Oct 14, 2019, 2:16 PM Vishal Santo

Re: Running a stand alone Mirror Maker 2.0 cluster and Issues

2019-10-14 Thread Vishal Santoshi
Aah no.. this is more to it. Note sure if related to the above. https://github.com/axbaretto/kafka/blob/master/connect/runtime/src/main/java/org/apache/kafka/connect/runtime/SourceTaskOffsetCommitter.java#L114 Is timing out based on https://github.com/axbaretto/kafka/blob/master/connect/runtime

Re: Stale data in KStream->KTable join

2019-10-14 Thread Sophie Blee-Goldman
I'm not sure you really need to/ should be sending data back to the table with .to(table) -- that was kind of the original issue, since it introduces a cycle* into your topology (which should be a DAG). *Of course it's still technically a cycle if you implement the store updates manually with a tr

Re: Running a stand alone Mirror Maker 2.0 cluster and Issues

2019-10-14 Thread Vishal Santoshi
I think this might be it.. Could you confirm. It seems to be on the path to commit the offsets.. but not sure... [2019-10-14 15:29:14,531] ERROR Scheduler for MirrorSourceConnector caught exception in scheduled task: syncing topic ACLs (org.apache.kafka.connect.mirror.Scheduler:102) java.util.con

Re: [VOTE] 2.3.1 RC1

2019-10-14 Thread Mickael Maison
Hi David, Probably not a blocker, but I noticed the authorization section in the security docs was not updated in 2.3.0 to include authorization details for ElectPreferredLeaders and IncrementalAlterConfigs. I just opened a PR to address that in 2.3: https://github.com/apache/kafka/pull/7508 (and

Re: Running a stand alone Mirror Maker 2.0 cluster and Issues

2019-10-14 Thread Ryanne Dolan
> I do not have a single record in the offsets topic That's definitely not normal. You are correct that without records in that topic, MM2 will restart from EARLIEST. The offsets should be stored periodically and whenever the connectors gracefully shutdown or restart. Is it possible the topics do

Re: [VOTE] 2.3.1 RC1

2019-10-14 Thread Gwen Shapira
David, Why do we have two site-doc packages, one for each Scala version? It is just HTML, right? IIRC, in previous releases we only packaged the docs once? Gwen On Fri, Oct 4, 2019 at 6:52 PM David Arthur wrote: > > Hello all, we identified a few bugs and a dependency update we wanted to > get

Kafka with Monolithic

2019-10-14 Thread Uzama Zaid Mohammed Jaward
Hi all, Can we integrate Kafka with Monolithic Architecture? Is this bad practice?

Re: Running a stand alone Mirror Maker 2.0 cluster and Issues

2019-10-14 Thread Vishal Santoshi
2nd/restore issue ( I think I need to solve the offsets topic issue before I go with the scale up and down issue ) As you had indicated, I went ahead and created the offsets topic. The status of the cluster ( destination ) is thus opic# Partitions# BrokersBrokers Spread %Brokers Skew %Brokers L

Resetting idleStartTime when all partitions are empty

2019-10-14 Thread Javier Holguera
Hi, We have a KStream and a KTable that we are left-joining. The KTable has a "backlog" of records that we want to consume before any of the entries in the KStream is processed. To guarantee that, we have played with the timestamp extraction, setting the time for those records in the "distant" pas

Log some types of API requests on broker

2019-10-14 Thread Ivan Yurchenko
Hi, I'm looking for a good way to log some broker API requests which might be considered important. Topic deletion request is quite good an example of such important requests. This piece [1] is the only thing I found that remotely is what I need. However, it logs all the requests and does it on t