Yes. As my understanding, it would allow me to write SQLs to query a spark
context. But, the query needs to be specified within a job & deployed.

What I want is to be able to run multiple dynamic queries specified at
runtime from a dashboard.



--
Nikhil Bafna

On Sat, Feb 21, 2015 at 8:37 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> Have you looked at
> http://spark.apache.org/docs/1.2.0/api/scala/index.html#org.apache.spark.sql.SchemaRDD
> ?
>
> Cheers
>
> On Sat, Feb 21, 2015 at 4:24 AM, Nikhil Bafna <nikhil.ba...@flipkart.com>
> wrote:
>
>>
>> Hi.
>>
>> My use case is building a realtime monitoring system over
>> multi-dimensional data.
>>
>> The way I'm planning to go about it is to use Spark Streaming to store
>> aggregated count over all dimensions in 10 sec interval.
>>
>> Then, from a dashboard, I would be able to specify a query over some
>> dimensions, which will need re-aggregation from the already computed job.
>>
>> My query is, how can I run dynamic queries over data in schema RDDs?
>>
>> --
>> Nikhil Bafna
>>
>
>

Reply via email to