Can you try running with -Dspark.query.files instead of -Dqfile from itests directory
Thanks Hari > On Jul 2, 2015, at 10:32 AM, "Sergey Shelukhin" <ser...@hortonworks.com> > wrote: > > Hi. I am trying to run TestSparkCliDriver. > > 1) Spark tests do not appear to support specifying a query like other > tests; when I run mvn test -Phadoop-2 -Dtest=TestSparkCliDriver tests run, > but with > mvn test -Phadoop-2 -Dtest=TestSparkCliDriver -Dqfile=foo.q,bar.q,.. test > just instantly succeeds w/o running any queries. Is there some other way > to specify those? > > 2) When I run all the test, they fail with the below exception > I’ve done a full regular build (mvn clean install … in root and then > itests). Are more steps necessary? > The itests/qtest-spark/../../itests/qtest-spark/target/spark directory > exists and has bunch of stuff, but bin/ subdirectory that it tries to run > from is indeed empty. > > 2015-07-02 10:11:58,678 ERROR [main]: spark.SparkTask > (SessionState.java:printError(987)) - Failed to execute spark task, with > exception 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to > create spark client.)' > org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create spark > client. > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSes > sionImpl.java:57) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSes > sion(SparkSessionManagerImpl.java:114) > at > org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUt > ilities.java:127) > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:101) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:89) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1672) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1431) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1212) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1063) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053) > at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:213) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:311) > at org.apache.hadoop.hive.ql.QTestUtil.createSources(QTestUtil.java:840) > at > org.apache.hadoop.hive.cli.TestSparkCliDriver.<clinit>(TestSparkCliDriver.j > ava:59) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:6 > 2) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp > l.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.junit.internal.runners.SuiteMethod.testFromSuiteMethod(SuiteMethod.java > :35) > at org.junit.internal.runners.SuiteMethod.<init>(SuiteMethod.java:24) > at > org.junit.internal.builders.SuiteMethodBuilder.runnerForClass(SuiteMethodBu > ilder.java:11) > at > org.junit.runners.model.RunnerBuilder.safeRunnerForClass(RunnerBuilder.java > :59) > at > org.junit.internal.builders.AllDefaultPossibilitiesBuilder.runnerForClass(A > llDefaultPossibilitiesBuilder.java:26) > at > org.junit.runners.model.RunnerBuilder.safeRunnerForClass(RunnerBuilder.java > :59) > at org.junit.internal.requests.ClassRequest.getRunner(ClassRequest.java:26) > at > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java > :262) > at > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provid > er.java:153) > at > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java: > 124) > at > org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoad > er(ForkedBooter.java:200) > at > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBoot > er.java:153) > at > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103) > Caused by: java.io.IOException: Cannot run program > “[snip]/itests/qtest-spark/../../itests/qtest-spark/target/spark/bin/spark- > submit": error=2, No such file or directory > at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048) > at > org.apache.hive.spark.client.SparkClientImpl.startDriver(SparkClientImpl.ja > va:415) > at > org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:94 > ) > at > org.apache.hive.spark.client.SparkClientFactory.createClient(SparkClientFac > tory.java:80) > at > org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiv > eSparkClient.java:91) > at > org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSpark > Client(HiveSparkClientFactory.java:65) > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSes > sionImpl.java:55) > ... 33 more > Caused by: java.io.IOException: error=2, No such file or directory > at java.lang.UNIXProcess.forkAndExec(Native Method) > at java.lang.UNIXProcess.<init>(UNIXProcess.java:248) > at java.lang.ProcessImpl.start(ProcessImpl.java:134) > at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029) > ... 39 more > > >