Hi all, Trying out Spark 1.4 on MapR Hadoop 2.5.1 running in yarn-client mode. Seems the application master doesn't work anymore, I get a 500 connect refused, even when I hit the IP/port of the spark UI directly. The logs don't show much.
I build spark with Java 6, hive & scala 2.10 and 2.11. I've tried with and without -Phadoop-provided *Build command;* ./make-distribution.sh --name mapr4.0.2_yarn_j6_2.10 --tgz -Pyarn -Pmapr4 -Phadoop-2.4 -Pmapr4 -Phive -Phadoop-provided -Dhadoop.version=2.5.1-mapr-1501 -Dyarn.version=2.5.1-mapr-1501 -DskipTests -e -X *Logs from spark shell;* 15/06/03 00:10:56 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:4040 15/06/03 00:10:56 INFO util.Utils: Successfully started service 'SparkUI' on port 4040. 15/06/03 00:10:56 INFO ui.SparkUI: Started SparkUI at http://172.31.10.14:4040 15/06/03 00:10:57 INFO yarn.Client: Requesting a new application from cluster with 71 NodeManagers 15/06/03 00:10:57 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (112640 MB per container) 15/06/03 00:10:57 INFO yarn.Client: Will allocate AM container, with 896 MB memory including 384 MB overhead 15/06/03 00:10:57 INFO yarn.Client: Setting up container launch context for our AM 15/06/03 00:10:57 INFO yarn.Client: Preparing resources for our AM container 15/06/03 00:10:57 INFO yarn.Client: Uploading resource file:///apps/spark/spark-1.4.0-SNAPSHOT-bin-mapr4.0.2_yarn_j6_2.11/lib/spark-assembly-1.4.0-SNAPSHOT-hadoop2.5.1-mapr-1501.jar -> maprfs:/user/nw/.sparkStaging/application_1432690361766_0593/spark-assembly-1.4.0-SNAPSHOT-hadoop2.5.1-mapr-1501.jar 15/06/03 00:10:58 INFO yarn.Client: Uploading resource file:/tmp/spark-5e42f904-ff83-4c93-bd35-4c3e20226a8a/__hadoop_conf__983379693214711.zip -> maprfs:/user/nw/.sparkStaging/application_1432690361766_0593/__hadoop_conf__983379693214711.zip 15/06/03 00:10:58 INFO yarn.Client: Setting up the launch environment for our AM container 15/06/03 00:10:58 INFO spark.SecurityManager: Changing view acls to: nw 15/06/03 00:10:58 INFO spark.SecurityManager: Changing modify acls to: nw 15/06/03 00:10:58 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(nw); users with modify permissions: Set(nw) 15/06/03 00:10:58 INFO yarn.Client: Submitting application 593 to ResourceManager 15/06/03 00:10:58 INFO security.ExternalTokenManagerFactory: Initialized external token manager class - com.mapr.hadoop.yarn.security.MapRTicketManager 15/06/03 00:10:58 INFO impl.YarnClientImpl: Submitted application application_1432690361766_0593 15/06/03 00:10:59 INFO yarn.Client: Application report for application_1432690361766_0593 (state: ACCEPTED) 15/06/03 00:10:59 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: default start time: 1433290258143 final status: UNDEFINED tracking URL: http://qtausc-pphd0101.hadoop.local:8088/proxy/application_1432690361766_0593/ user: nw 15/06/03 00:11:00 INFO yarn.Client: Application report for application_1432690361766_0593 (state: ACCEPTED) 15/06/03 00:11:01 INFO yarn.Client: Application report for application_1432690361766_0593 (state: ACCEPTED) 15/06/03 00:11:02 INFO yarn.Client: Application report for application_1432690361766_0593 (state: ACCEPTED) 15/06/03 00:11:03 INFO yarn.Client: Application report for application_1432690361766_0593 (state: ACCEPTED) 15/06/03 00:11:03 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: ApplicationMaster registered as AkkaRpcEndpointRef(Actor[akka.tcp:// sparkYarnAM@192.168.81.167:36542/user/YarnAM#1631897818]) 15/06/03 00:11:03 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> qtausc-pphd0167.hadoop.local, PROXY_URI_BASES -> http://qtausc-pphd0167.hadoop.local:8088/proxy/application_1432690361766_0593), /proxy/application_1432690361766_0593 15/06/03 00:11:03 INFO ui.JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter 15/06/03 00:11:04 INFO yarn.Client: Application report for application_1432690361766_0593 (state: RUNNING) 15/06/03 00:11:04 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: 192.168.81.167 ApplicationMaster RPC port: 0 queue: default start time: 1433290258143 final status: UNDEFINED tracking URL: http://qtausc-pphd0101.hadoop.local:8088/proxy/application_1432690361766_0593/ user: nw 15/06/03 00:11:04 INFO cluster.YarnClientSchedulerBackend: Application application_1432690361766_0593 has started running. 15/06/03 00:11:04 INFO util.Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 45668. 15/06/03 00:11:04 INFO netty.NettyBlockTransferService: Server created on 45668 15/06/03 00:11:04 INFO storage.BlockManagerMaster: Trying to register BlockManager 15/06/03 00:11:04 INFO storage.BlockManagerMasterEndpoint: Registering block manager 172.31.10.14:45668 with 265.4 MB RAM, BlockManagerId(driver, 172.31.10.14, 45668) 15/06/03 00:11:04 INFO storage.BlockManagerMaster: Registered BlockManager *Logs from AM logs page in YARN;* 15/06/03 10:11:01 INFO yarn.ApplicationMaster: Registered signal handlers for [TERM, HUP, INT] 15/06/03 10:11:02 INFO yarn.ApplicationMaster: ApplicationAttemptId: appattempt_1432690361766_0593_000001 15/06/03 10:11:02 INFO spark.SecurityManager: Changing view acls to: nw 15/06/03 10:11:02 INFO spark.SecurityManager: Changing modify acls to: nw 15/06/03 10:11:02 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(nw); users with modify permissions: Set(nw) 15/06/03 10:11:03 INFO slf4j.Slf4jLogger: Slf4jLogger started 15/06/03 10:11:03 INFO Remoting: Starting remoting 15/06/03 10:11:03 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkYarnAM@192.168.81.167:36542] 15/06/03 10:11:03 INFO util.Utils: Successfully started service 'sparkYarnAM' on port 36542. 15/06/03 10:11:03 INFO yarn.ApplicationMaster: Waiting for Spark driver to be reachable. 15/06/03 10:11:03 INFO yarn.ApplicationMaster: Driver now available: 172.31.10.14:59954 15/06/03 10:11:03 INFO yarn.ApplicationMaster$AMEndpoint: Add WebUI Filter. AddWebUIFilter(org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter,Map(PROXY_HOSTS -> qtausc-pphd0167.hadoop.local, PROXY_URI_BASES -> http://qtausc-pphd0167.hadoop.local:8088/proxy/application_1432690361766_0593 ),/proxy/application_1432690361766_0593) 15/06/03 10:11:03 INFO yarn.YarnRMClient: Registering the ApplicationMaster 15/06/03 10:11:04 INFO yarn.YarnAllocator: Will request 2 executor containers, each with 1 cores and 1408 MB memory including 384 MB overhead 15/06/03 10:11:04 INFO yarn.YarnAllocator: Container request (host: Any, capability: <memory:1408, vCores:1, disks:0.0>) 15/06/03 10:11:04 INFO yarn.YarnAllocator: Container request (host: Any, capability: <memory:1408, vCores:1, disks:0.0>) 15/06/03 10:11:04 INFO yarn.ApplicationMaster: Started progress reporter thread with (heartbeat : 3000, initial allocation : 200) intervals 15/06/03 10:11:04 INFO impl.AMRMClientImpl: Received new token for : qtausc-pphd0146.hadoop.local:55935 15/06/03 10:11:04 INFO impl.AMRMClientImpl: Received new token for : qtausc-pphd0155.hadoop.local:45589 15/06/03 10:11:04 INFO yarn.YarnAllocator: Launching container container_1432690361766_0593_01_000002 for on host qtausc-pphd0146.hadoop.local 15/06/03 10:11:04 INFO yarn.YarnAllocator: Launching ExecutorRunnable. driverUrl: akka.tcp:// sparkDriver@172.31.10.14:59954/user/CoarseGrainedScheduler, executorHostname: qtausc-pphd0146.hadoop.local 15/06/03 10:11:04 INFO yarn.YarnAllocator: Launching container container_1432690361766_0593_01_000003 for on host qtausc-pphd0155.hadoop.local 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Starting Executor Container 15/06/03 10:11:04 INFO yarn.YarnAllocator: Launching ExecutorRunnable. driverUrl: akka.tcp:// sparkDriver@172.31.10.14:59954/user/CoarseGrainedScheduler, executorHostname: qtausc-pphd0155.hadoop.local 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Starting Executor Container 15/06/03 10:11:04 INFO yarn.YarnAllocator: Received 2 containers from YARN, launching executors on 2 of them. 15/06/03 10:11:04 INFO impl.ContainerManagementProtocolProxy: yarn.client.max-nodemanagers-proxies : 500 15/06/03 10:11:04 INFO impl.ContainerManagementProtocolProxy: yarn.client.max-nodemanagers-proxies : 500 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Setting up ContainerLaunchContext 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Setting up ContainerLaunchContext 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Preparing Local resources 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Preparing Local resources 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Prepared Local resources Map(__spark__.jar -> resource { scheme: "maprfs" port: -1 file: "/user/nw/.sparkStaging/application_1432690361766_0593/spark-assembly-1.4.0-SNAPSHOT-hadoop2.5.1-mapr-1501.jar" } size: 124419029 timestamp: 1433290257972 type: FILE visibility: PRIVATE) 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Prepared Local resources Map(__spark__.jar -> resource { scheme: "maprfs" port: -1 file: "/user/nw/.sparkStaging/application_1432690361766_0593/spark-assembly-1.4.0-SNAPSHOT-hadoop2.5.1-mapr-1501.jar" } size: 124419029 timestamp: 1433290257972 type: FILE visibility: PRIVATE) 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Setting up executor with environment: Map(CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>/opt/mapr/lib/*:/opt/mapr/hadoop/hadoop-2.5.1/share/hadoop/yarn/*:/opt/mapr/hadoop/hadoop-2.5.1/share/hadoop/common/lib/*:/opt/mapr/hive/hive-current/lib/*, SPARK_LOG_URL_STDERR -> http://qtausc-pphd0155.hadoop.local:8042/node/containerlogs/container_1432690361766_0593_01_000003/nw/stderr?start=0, SPARK_DIST_CLASSPATH -> /opt/mapr/lib/*:/opt/mapr/hadoop/hadoop-2.5.1/share/hadoop/yarn/*:/opt/mapr/hadoop/hadoop-2.5.1/share/hadoop/common/lib/*:/opt/mapr/hive/hive-current/lib/*, SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1432690361766_0593, SPARK_YARN_CACHE_FILES_FILE_SIZES -> 124419029, SPARK_USER -> nw, SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE, SPARK_YARN_MODE -> true, SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1433290257972, SPARK_LOG_URL_STDOUT -> http://qtausc-pphd0155.hadoop.local:8042/node/containerlogs/container_1432690361766_0593_01_000003/nw/stdout?start=0, SPARK_YARN_CACHE_FILES -> maprfs:/user/nw/.sparkStaging/application_1432690361766_0593/spark-assembly-1.4.0-SNAPSHOT-hadoop2.5.1-mapr-1501.jar#__spark__.jar) 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Setting up executor with environment: Map(CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>/opt/mapr/lib/*:/opt/mapr/hadoop/hadoop-2.5.1/share/hadoop/yarn/*:/opt/mapr/hadoop/hadoop-2.5.1/share/hadoop/common/lib/*:/opt/mapr/hive/hive-current/lib/*, SPARK_LOG_URL_STDERR -> http://qtausc-pphd0146.hadoop.local:8042/node/containerlogs/container_1432690361766_0593_01_000002/nw/stderr?start=0, SPARK_DIST_CLASSPATH -> /opt/mapr/lib/*:/opt/mapr/hadoop/hadoop-2.5.1/share/hadoop/yarn/*:/opt/mapr/hadoop/hadoop-2.5.1/share/hadoop/common/lib/*:/opt/mapr/hive/hive-current/lib/*, SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1432690361766_0593, SPARK_YARN_CACHE_FILES_FILE_SIZES -> 124419029, SPARK_USER -> nw, SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE, SPARK_YARN_MODE -> true, SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1433290257972, SPARK_LOG_URL_STDOUT -> http://qtausc-pphd0146.hadoop.local:8042/node/containerlogs/container_1432690361766_0593_01_000002/nw/stdout?start=0, SPARK_YARN_CACHE_FILES -> maprfs:/user/nw/.sparkStaging/application_1432690361766_0593/spark-assembly-1.4.0-SNAPSHOT-hadoop2.5.1-mapr-1501.jar#__spark__.jar) 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Setting up executor with commands: List({{JAVA_HOME}}/bin/java, -server, -XX:OnOutOfMemoryError='kill %p', -Xms1024m, -Xmx1024m, -Djava.io.tmpdir={{PWD}}/tmp, '-Dspark.driver.port=59954', -Dspark.yarn.app.container.log.dir=<LOG_DIR>, org.apache.spark.executor.CoarseGrainedExecutorBackend, --driver-url, akka.tcp://sparkDriver@172.31.10.14:59954/user/CoarseGrainedScheduler, --executor-id, 2, --hostname, qtausc-pphd0155.hadoop.local, --cores, 1, --app-id, application_1432690361766_0593, --user-class-path, file:$PWD/__app__.jar, 1>, <LOG_DIR>/stdout, 2>, <LOG_DIR>/stderr) 15/06/03 10:11:04 INFO yarn.ExecutorRunnable: Setting up executor with commands: List({{JAVA_HOME}}/bin/java, -server, -XX:OnOutOfMemoryError='kill %p', -Xms1024m, -Xmx1024m, -Djava.io.tmpdir={{PWD}}/tmp, '-Dspark.driver.port=59954', -Dspark.yarn.app.container.log.dir=<LOG_DIR>, org.apache.spark.executor.CoarseGrainedExecutorBackend, --driver-url, akka.tcp://sparkDriver@172.31.10.14:59954/user/CoarseGrainedScheduler, --executor-id, 1, --hostname, qtausc-pphd0146.hadoop.local, --cores, 1, --app-id, application_1432690361766_0593, --user-class-path, file:$PWD/__app__.jar, 1>, <LOG_DIR>/stdout, 2>, <LOG_DIR>/stderr) 15/06/03 10:11:04 INFO impl.ContainerManagementProtocolProxy: Opening proxy : qtausc-pphd0155.hadoop.local:45589 15/06/03 10:11:04 INFO impl.ContainerManagementProtocolProxy: Opening proxy : qtausc-pphd0146.hadoop.local:55935 Any ideas what the problem is? Cheers, ~NW