Yes. As my understanding, it would allow me to write SQLs to query a spark context. But, the query needs to be specified within a job & deployed.
What I want is to be able to run multiple dynamic queries specified at runtime from a dashboard. -- Nikhil Bafna On Sat, Feb 21, 2015 at 8:37 PM, Ted Yu <yuzhih...@gmail.com> wrote: > Have you looked at > http://spark.apache.org/docs/1.2.0/api/scala/index.html#org.apache.spark.sql.SchemaRDD > ? > > Cheers > > On Sat, Feb 21, 2015 at 4:24 AM, Nikhil Bafna <nikhil.ba...@flipkart.com> > wrote: > >> >> Hi. >> >> My use case is building a realtime monitoring system over >> multi-dimensional data. >> >> The way I'm planning to go about it is to use Spark Streaming to store >> aggregated count over all dimensions in 10 sec interval. >> >> Then, from a dashboard, I would be able to specify a query over some >> dimensions, which will need re-aggregation from the already computed job. >> >> My query is, how can I run dynamic queries over data in schema RDDs? >> >> -- >> Nikhil Bafna >> > >