[ https://issues.apache.org/jira/browse/HIVE-16484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16316894#comment-16316894 ]
Marcelo Vanzin commented on HIVE-16484: --------------------------------------- Yes, that's what the launcher does internally, but also, it's an internal API, sort of. You can try it and it will probably work, but I'd stick with the public API unless the extra fds are a real issue (one extra fd per session doesn't sound too crazy?). > Investigate SparkLauncher for HoS as alternative to bin/spark-submit > -------------------------------------------------------------------- > > Key: HIVE-16484 > URL: https://issues.apache.org/jira/browse/HIVE-16484 > Project: Hive > Issue Type: Bug > Components: Spark > Reporter: Sahil Takiar > Assignee: Sahil Takiar > Attachments: HIVE-16484.1.patch, HIVE-16484.10.patch, > HIVE-16484.2.patch, HIVE-16484.3.patch, HIVE-16484.4.patch, > HIVE-16484.5.patch, HIVE-16484.6.patch, HIVE-16484.7.patch, > HIVE-16484.8.patch, HIVE-16484.9.patch > > > The {{SparkClientImpl#startDriver}} currently looks for the {{SPARK_HOME}} > directory and invokes the {{bin/spark-submit}} script, which spawns a > separate process to run the Spark application. > {{SparkLauncher}} was added in SPARK-4924 and is a programatic way to launch > Spark applications. > I see a few advantages: > * No need to spawn a separate process to launch a HoS --> lower startup time > * Simplifies the code in {{SparkClientImpl}} --> easier to debug > * {{SparkLauncher#startApplication}} returns a {{SparkAppHandle}} which > contains some useful utilities for querying the state of the Spark job > ** It also allows the launcher to specify a list of job listeners -- This message was sent by Atlassian JIRA (v6.4.14#64029)