[ https://issues.apache.org/jira/browse/HIVE-15302?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15711207#comment-15711207 ]
Rui Li commented on HIVE-15302: ------------------------------- [~kellyzly], you're right about the ideas. But the needed spark jars may not be the same as those listed in wiki now. Those listed are needed when linking spark to hive side, while spark.yarn.archive and spark.yarn.jars are intended for the containers on YARN side. But I guess the needed jars should be quite similar to those for local mode in our current wiki. bq. because user has already set spark.yarn.jars so they can directory download a spark tarball from webside I'm not sure what you mean here. We still need user to have spark installed in their cluster, either downloaded or built by themselves. But we can relax the limitation that the spark must be built w/o hive, in some cases. > Relax the requirement that HoS needs Spark built w/o Hive > --------------------------------------------------------- > > Key: HIVE-15302 > URL: https://issues.apache.org/jira/browse/HIVE-15302 > Project: Hive > Issue Type: Improvement > Reporter: Rui Li > Assignee: Rui Li > > This requirement becomes more and more unacceptable as SparkSQL becomes > widely adopted. Let's use this JIRA to find out how we can relax the > limitation. -- This message was sent by Atlassian JIRA (v6.3.4#6332)