mudit, this link can guide you - https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started
Thanks, Karthik On Thu, Jul 28, 2016 at 8:54 AM, Mudit Kumar <mudit.ku...@askme.in> wrote: > Yes Mich,exactly. > > Thanks, > Mudit > > From: Mich Talebzadeh <mich.talebza...@gmail.com> > Reply-To: <user@hive.apache.org> > Date: Thursday, July 28, 2016 at 1:08 AM > To: user <user@hive.apache.org> > Subject: Re: Hive on spark > > You mean you want to run Hive using Spark as the execution engine which > uses Yarn by default? > > > Something like below > > hive> select max(id) from oraclehadoop.dummy_parquet; > Starting Spark Job = 8218859d-1d7c-419c-adc7-4de175c3ca6d > Query Hive on Spark job[1] stages: > 2 > 3 > Status: Running (Hive on Spark job[1]) > Job Progress Format > CurrentTime StageId_StageAttemptId: > SucceededTasksCount(+RunningTasksCount-FailedTasksCount)/TotalTasksCount > [StageCost] > 2016-07-27 20:38:17,269 Stage-2_0: 0(+8)/24 Stage-3_0: 0/1 > 2016-07-27 20:38:20,298 Stage-2_0: 8(+4)/24 Stage-3_0: 0/1 > 2016-07-27 20:38:22,309 Stage-2_0: 11(+1)/24 Stage-3_0: 0/1 > 2016-07-27 20:38:23,330 Stage-2_0: 12(+8)/24 Stage-3_0: 0/1 > 2016-07-27 20:38:26,360 Stage-2_0: 17(+7)/24 Stage-3_0: 0/1 > 2016-07-27 20:38:27,386 Stage-2_0: 20(+4)/24 Stage-3_0: 0/1 > 2016-07-27 20:38:28,391 Stage-2_0: 21(+3)/24 Stage-3_0: 0/1 > 2016-07-27 20:38:29,395 Stage-2_0: 24/24 Finished Stage-3_0: 1/1 > Finished > Status: Finished successfully in 13.14 seconds > OK > 100000000 > Time taken: 13.426 seconds, Fetched: 1 row(s) > > > HTH > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > > > http://talebzadehmich.wordpress.com > > > *Disclaimer:* Use it at your own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. > The author will in no case be liable for any monetary damages arising from > such loss, damage or destruction. > > > > On 27 July 2016 at 20:31, Mudit Kumar <mudit.ku...@askme.in> wrote: > >> Hi All, >> >> I need to configure hive cluster based on spark engine (yarn). >> I already have a running hadoop cluster. >> >> Can someone point me to relevant documentation? >> >> TIA. >> >> Thanks, >> Mudit >> > >