I was not sure if registerAsTable() and then query against that table have additional performance impact and if DSL eliminates that.
On Thu, Jul 31, 2014 at 2:33 PM, Zongheng Yang <zonghen...@gmail.com> wrote: > Looking at what this patch [1] has to do to achieve it, I am not sure > if you can do the same thing in 1.0.0 using DSL only. Just curious, > why don't you use the hql() / sql() methods and pass a query string > in? > > [1] https://github.com/apache/spark/pull/1211/files > > On Thu, Jul 31, 2014 at 2:20 PM, Buntu Dev <buntu...@gmail.com> wrote: > > Thanks Zongheng for the pointer. Is there a way to achieve the same in > 1.0.0 > > ? > > > > > > On Thu, Jul 31, 2014 at 1:43 PM, Zongheng Yang <zonghen...@gmail.com> > wrote: > >> > >> countDistinct is recently added and is in 1.0.2. If you are using that > >> or the master branch, you could try something like: > >> > >> r.select('keyword, countDistinct('userId)).groupBy('keyword) > >> > >> On Thu, Jul 31, 2014 at 12:27 PM, buntu <buntu...@gmail.com> wrote: > >> > I'm looking to write a select statement to get a distinct count on > >> > userId > >> > grouped by keyword column on a parquet file SchemaRDD equivalent of: > >> > SELECT keyword, count(distinct(userId)) from table group by keyword > >> > > >> > How to write it using the chained select().groupBy() operations? > >> > > >> > Thanks! > >> > > >> > > >> > > >> > -- > >> > View this message in context: > >> > > http://apache-spark-user-list.1001560.n3.nabble.com/SchemaRDD-select-expression-tp11069.html > >> > Sent from the Apache Spark User List mailing list archive at > Nabble.com. > > > > >