There will probably be some subquery support in 2.0. That particular query would be more efficient to express as an argmax however. Here is an example in Spark 1.6 <https://databricks-prod-cloudfront.cloud.databricks.com/public/4027ec902e239c93eaaa8714f173bcfc/1023043053387187/3170497669323442/2840265927289860/c072bba513.html> .
On Thu, Feb 25, 2016 at 2:58 PM, Mohammad Tariq <donta...@gmail.com> wrote: > AFAIK, this isn't supported yet. A ticket > <https://issues.apache.org/jira/browse/SPARK-4226> is in progress though. > > > > [image: http://] > > Tariq, Mohammad > about.me/mti > [image: http://] > <http://about.me/mti> > > > On Fri, Feb 26, 2016 at 4:16 AM, Mich Talebzadeh < > mich.talebza...@cloudtechnologypartners.co.uk> wrote: > >> >> >> Hi, >> >> >> >> I guess the following confirms that Spark does bot support sub-queries >> >> >> >> val d = HiveContext.table("test.dummy") >> >> d.registerTempTable("tmp") >> >> HiveContext.sql("select * from tmp where id IN (select max(id) from tmp)") >> >> It crashes >> >> The SQL works OK in Hive itself on the underlying table! >> >> select * from dummy where id IN (select max(id) from dummy); >> >> >> >> Thanks >> >> >> -- >> >> Dr Mich Talebzadeh >> >> LinkedIn >> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> http://talebzadehmich.wordpress.com >> >> NOTE: The information in this email is proprietary and confidential. This >> message is for the designated recipient only, if you are not the intended >> recipient, you should destroy it immediately. Any information in this >> message shall not be understood as given or endorsed by Cloud Technology >> Partners Ltd, its subsidiaries or their employees, unless expressly so >> stated. It is the responsibility of the recipient to ensure that this email >> is virus free, therefore neither Cloud Technology partners Ltd, its >> subsidiaries nor their employees accept any responsibility. >> >> >> >