"REPLACE_THREAD - Replaces the thread receiving the exception and
processing continues with the same number of configured threads. (Note:
this can result in duplicate records depending on the application’s
processing mode determined by the PROCESSING_GUARANTEE_CONFIG value)"
If an application gets
Hi Yoda,
For your question:
> If an application gets an uncaught exception, then the failed thread will
be replaced with another thread and it will continue processing messages,
skipping that failed message?
--> Yes, if everything goes well after `replace thread`, you can ignore
this failed messag
Hi Yoda,
What do you mean exactly with "skipping that failed message"?
Do you mean a record consumed from a topic that caused an exception that
killed the stream thread?
If the record killed the stream thread due to an exception, for example,
a deserialization exception, it will probably als
Hi Luke, thank you for your answer.
I will try it, I think I will set an alert if there are too many messages.
To ignore the message should I simply return "replace_thread" in the
handler?
On Tue, Aug 10, 2021 at 12:16 PM Luke Chen wrote:
> Hi Yoda,
> For your question:
> > If an application get
Hi Bruno, thank you for your answer.
I mean that the message that caused the exception was consumed and replaced
thread will continue from the next message. How then does it handle
uncaught exceptions, if it will fail again?
On Tue, Aug 10, 2021 at 12:33 PM Bruno Cadonna wrote:
> Hi Yoda,
>
> W
Hello .
How about starting with 'quick start' ?
https://kafka.apache.org/quickstart
Thanks,
Yu Watanabe
On Tue, Aug 10, 2021 at 10:44 AM Gilbert Flores <
gilbert.flo...@primergrp.com> wrote:
> Hi,
>
> Good day to you. We are on the stage of exploring integration systems and
> we have read and
Hey All,
We are using the low level processor API to create kafka stream
applications. Each app has 1 or more in-memory state stores with caching
disabled and changelog enabled. Some of the apps also have global stores.
We noticed from the node metrics (kubernetes) that the stream applications
are
Hello, I am trying to create a Stream that will accept the data from my topic
and be able to use the message keys in the stream as unique identifiers for the
sensors the data originated from. The data in the messages does not have
anything that would be able to identify which sensor it came from
The keys are already part of the stream. When you run builder.stream or
builder.table it returns a Stream or a Table. From there every
operation has a lambda that accepts both key and value. You can use map for
example to accept the key and do something with that. Let me know if you
have any other
Ah, nevermind, just saw the title that this is KSQLDB.
On Tue, Aug 10, 2021 at 12:09 PM Daniel Hinojosa <
dhinoj...@evolutionnext.com> wrote:
> The keys are already part of the stream. When you run builder.stream or
> builder.table it returns a Stream or a Table. From there every
> operation has
I had to verify myself, I wrote an example that uses a key that is not in
the payload and I did where I specify the KEY and now it is a field. Notice
the state down below
CREATE STREAM my_avro_orders (total BIGINT, shipping VARCHAR, state VARCHAR
KEY, discount DOUBLE, gender VARCHAR)
WITH (kafka_
Glad you figured it out.
If you think the docs should be improved, we take PRs :) -- In the end,
it's a broader community effort to have good docs...
-Matthias
On 6/11/21 12:10 PM, Richard Rossel wrote:
> Thanks Matthias for your help,
> I was able to find the issue, it was actually my Serdes
I have 2 Kafka setup, A and B. A is a cluster with 3 instances running on
the same machine, while B is a stand alone in another machine. I tried to
use Mirror Maker to replicate from A to B. The config file is as follows:
-
clusters = A, BA.bootstrap.servers = host1:9091, host1:9092,
host1:90
13 matches
Mail list logo