Structured Streaming has a foreach sink, where you can essentially do what you want with your data. Its easy to create a Kafka producer, and write the data out to kafka. http://spark.apache.org/docs/latest/structured-streaming-programming-guide.html#using-foreach
On Fri, Jan 13, 2017 at 8:28 AM, Koert Kuipers <ko...@tresata.com> wrote: > how do you do this with structured streaming? i see no mention of writing > to kafka > > On Fri, Jan 13, 2017 at 10:30 AM, Peyman Mohajerian <mohaj...@gmail.com> > wrote: > >> Yes, it is called Structured Streaming: https://docs.databr >> icks.com/_static/notebooks/structured-streaming-kafka.html >> http://spark.apache.org/docs/latest/structured-streaming-pro >> gramming-guide.html >> >> On Fri, Jan 13, 2017 at 3:32 AM, Senthil Kumar <senthilec...@gmail.com> >> wrote: >> >>> Hi Team , >>> >>> Sorry if this question already asked in this forum.. >>> >>> Can we ingest data to Apache Kafka Topic from Spark SQL DataFrame ?? >>> >>> Here is my Code which Reads Parquet File : >>> >>> *val sqlContext = new org.apache.spark.sql.SQLContext(sc);* >>> >>> *val df = sqlContext.read.parquet("..../temp/*.parquet")* >>> >>> *df.registerTempTable("beacons")* >>> >>> >>> I want to directly ingest df DataFrame to Kafka ! Is there any way to >>> achieve this ?? >>> >>> >>> Cheers, >>> >>> Senthil >>> >> >> >