Dear Michael Noll,

I have a question; Is it possible converting JSON format to YAML format via
using Kafka Streams?

Best Regards

KIM

2017-03-10 11:36 GMT+09:00 BYEONG-GI KIM <bg...@bluedigm.com>:

> Thank you very much for the information!
>
>
> 2017-03-09 19:40 GMT+09:00 Michael Noll <mich...@confluent.io>:
>
>> There's actually a demo application that demonstrates the simplest use
>> case
>> for Kafka's Streams API:  to read data from an input topic and then write
>> that data as-is to an output topic.
>>
>> https://github.com/confluentinc/examples/blob/3.2.x/kafka-
>> streams/src/test/java/io/confluent/examples/streams/Pas
>> sThroughIntegrationTest.java
>>
>> The code above is for Confluent 3.2 and Apache Kafka 0.10.2.
>>
>> The demo shows how to (1) write a message from a producer to the input
>> topic, (2) use a Kafka Streams app to process that data and write the
>> results back to Kafka, and (3) validating the results with a consumer that
>> reads from the output topic.
>>
>> The GitHub project above includes many more such examples, see
>> https://github.com/confluentinc/examples/tree/3.2.x/kafka-streams.
>> Again,
>> this is for Confluent 3.2 and Kafka 0.10.2.  There is a version
>> compatibility matrix that explains which branches you need to use for
>> older
>> versions of Confluent/Kafka as well as for the very latest development
>> version (aka Kafka's trunk):
>> https://github.com/confluentinc/examples/tree/3.2.x/kafka-
>> streams#version-compatibility
>>
>> Hope this helps!
>> Michael
>>
>>
>>
>>
>> On Thu, Mar 9, 2017 at 9:59 AM, BYEONG-GI KIM <bg...@bluedigm.com> wrote:
>>
>> > Hello.
>> >
>> > I'm a new who started learning the one of the new Kafka functionality,
>> aka
>> > Kafka Stream.
>> >
>> > As far as I know, the simplest usage of the Kafka Stream is to do
>> something
>> > like parsing, which forward incoming data from a topic to another topic,
>> > with a few changing.
>> >
>> > So... Here is what I'd want to do:
>> >
>> > 1. Produce a simple message, like 1, 2, 3, 4, 5, ... from a producer
>> > 2. Let Kafka Stream application consume the message and change the
>> message
>> > like [1], [2], [3], ...
>> > 3. Consume the changed message at a consumer
>> >
>> > I've read the documentation,
>> > https://kafka.apache.org/0102/javadoc/index.html?org/apache/
>> kafka/connect,
>> > but it's unclear for me how to implement it.
>> >
>> > Especially, I could not understand the the
>> > line, builder.stream("my-input-topic").mapValues(value -&gt;
>> > value.length().toString()).to("my-output-topic"). Could someone
>> explain it
>> > and how to implement what I've mentioned?
>> >
>> > Thanks in advance.
>> >
>> > Best regards
>> >
>> > KIM
>> >
>>
>
>

Reply via email to