Yes, of course. You can also re-use any existing JSON and/or YAML library for helping you with that.
Also, in general, an application that uses the Kafka Streams API/library is a normal, standard Java application -- you can of course also use any other Java/Scala/... library for the application's processing needs. -Michael On Tue, Mar 14, 2017 at 9:00 AM, BYEONG-GI KIM <bg...@bluedigm.com> wrote: > Dear Michael Noll, > > I have a question; Is it possible converting JSON format to YAML format via > using Kafka Streams? > > Best Regards > > KIM > > 2017-03-10 11:36 GMT+09:00 BYEONG-GI KIM <bg...@bluedigm.com>: > > > Thank you very much for the information! > > > > > > 2017-03-09 19:40 GMT+09:00 Michael Noll <mich...@confluent.io>: > > > >> There's actually a demo application that demonstrates the simplest use > >> case > >> for Kafka's Streams API: to read data from an input topic and then > write > >> that data as-is to an output topic. > >> > >> https://github.com/confluentinc/examples/blob/3.2.x/kafka- > >> streams/src/test/java/io/confluent/examples/streams/Pas > >> sThroughIntegrationTest.java > >> > >> The code above is for Confluent 3.2 and Apache Kafka 0.10.2. > >> > >> The demo shows how to (1) write a message from a producer to the input > >> topic, (2) use a Kafka Streams app to process that data and write the > >> results back to Kafka, and (3) validating the results with a consumer > that > >> reads from the output topic. > >> > >> The GitHub project above includes many more such examples, see > >> https://github.com/confluentinc/examples/tree/3.2.x/kafka-streams. > >> Again, > >> this is for Confluent 3.2 and Kafka 0.10.2. There is a version > >> compatibility matrix that explains which branches you need to use for > >> older > >> versions of Confluent/Kafka as well as for the very latest development > >> version (aka Kafka's trunk): > >> https://github.com/confluentinc/examples/tree/3.2.x/kafka- > >> streams#version-compatibility > >> > >> Hope this helps! > >> Michael > >> > >> > >> > >> > >> On Thu, Mar 9, 2017 at 9:59 AM, BYEONG-GI KIM <bg...@bluedigm.com> > wrote: > >> > >> > Hello. > >> > > >> > I'm a new who started learning the one of the new Kafka functionality, > >> aka > >> > Kafka Stream. > >> > > >> > As far as I know, the simplest usage of the Kafka Stream is to do > >> something > >> > like parsing, which forward incoming data from a topic to another > topic, > >> > with a few changing. > >> > > >> > So... Here is what I'd want to do: > >> > > >> > 1. Produce a simple message, like 1, 2, 3, 4, 5, ... from a producer > >> > 2. Let Kafka Stream application consume the message and change the > >> message > >> > like [1], [2], [3], ... > >> > 3. Consume the changed message at a consumer > >> > > >> > I've read the documentation, > >> > https://kafka.apache.org/0102/javadoc/index.html?org/apache/ > >> kafka/connect, > >> > but it's unclear for me how to implement it. > >> > > >> > Especially, I could not understand the the > >> > line, builder.stream("my-input-topic").mapValues(value -> > >> > value.length().toString()).to("my-output-topic"). Could someone > >> explain it > >> > and how to implement what I've mentioned? > >> > > >> > Thanks in advance. > >> > > >> > Best regards > >> > > >> > KIM > >> > > >> > > > > > -- *Michael G. Noll* Product Manager | Confluent +1 650 453 5860 | @miguno <https://twitter.com/miguno> Follow us: Twitter <https://twitter.com/ConfluentInc> | Blog <http://www.confluent.io/blog>