Hi all, I tried to start a Yarn session on an Amazon EMR cluster with Hadoop 2.6.0 following the instructions provided in this link and using Flink 0.9.1 for Hadoop 2.6.0 https://ci.apache.org/projects/flink/flink-docs-release-0.9/setup/yarn_setup.html Running the following command line: ./bin/yarn-session.sh -n 2 -jm 1024 -tm 2048 generated the following error message ------------ 12:53:47,633 INFO org.apache.hadoop.yarn.client.RMProxy - Connecting to ResourceManager at /0.0.0.0:8032 12:53:47,805 WARN org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 12:53:48,226 WARN org.apache.flink.yarn.FlinkYarnClient - Neither the HADOOP_CONF_DIR nor the YARN_CONF_DIR environment variable is set.The Flink YARN Client needs one of these to be set to properly load the Hadoop configuration for accessing YARN. 12:53:48,227 INFO org.apache.flink.yarn.FlinkYarnClient - Using values: 12:53:48,228 INFO org.apache.flink.yarn.FlinkYarnClient - TaskManager count = 2 12:53:48,229 INFO org.apache.flink.yarn.FlinkYarnClient - JobManager memory = 1024 12:53:48,229 INFO org.apache.flink.yarn.FlinkYarnClient - TaskManager memory = 2048 12:53:48,580 WARN org.apache.flink.yarn.FlinkYarnClient - The file system scheme is 'file'. This indicates that the specified Hadoop configuration path is wrong and the sytem is using the default Hadoop configuration values.The Flink YARN client needs to store its files in a distributed file system 12:53:48,593 INFO org.apache.flink.yarn.Utils - Copying from file:/home/hadoop/flink-0.9.1/lib/flink-dist-0.9.1.jar to file:/home/hadoop/.flink/application_1444046049303_0008/flink-dist-0.9.1.jar 12:53:49,245 INFO org.apache.flink.yarn.Utils - Copying from /home/hadoop/flink-0.9.1/conf/flink-conf.yaml to file:/home/hadoop/.flink/application_1444046049303_0008/flink-conf.yaml 12:53:49,251 INFO org.apache.flink.yarn.Utils - Copying from file:/home/hadoop/flink-0.9.1/lib/flink-python-0.9.1.jar to file:/home/hadoop/.flink/application_1444046049303_0008/flink-python-0.9.1.jar 12:53:49,278 INFO org.apache.flink.yarn.Utils - Copying from file:/home/hadoop/flink-0.9.1/conf/logback.xml to file:/home/hadoop/.flink/application_1444046049303_0008/logback.xml 12:53:49,285 INFO org.apache.flink.yarn.Utils - Copying from file:/home/hadoop/flink-0.9.1/conf/log4j.properties to file:/home/hadoop/.flink/application_1444046049303_0008/log4j.properties 12:53:49,304 INFO org.apache.flink.yarn.FlinkYarnClient - Submitting application master application_1444046049303_0008 12:53:49,347 INFO org.apache.hadoop.yarn.client.api.impl.YarnClientImpl - Submitted application application_1444046049303_0008 12:53:49,347 INFO org.apache.flink.yarn.FlinkYarnClient - Waiting for the cluster to be allocated 12:53:49,349 INFO org.apache.flink.yarn.FlinkYarnClient - Deploying cluster, current state ACCEPTED 12:53:50,351 INFO org.apache.flink.yarn.FlinkYarnClient - Deploying cluster, current state ACCEPTED Error while deploying YARN cluster: The YARN application unexpectedly switched to state FAILED during deployment. Diagnostics from YARN: Application application_1444046049303_0008 failed 1 times due to AM Container for appattempt_1444046049303_0008_000001 exited with exitCode: -1000 For more detailed output, check application tracking page:http://ip-172-31-10-16.us<http://ip-172-31-10-16.us>-west-2.compute.internal:20888/proxy/application_1444046049303_0008/Then, click on links to logs of each attempt. Diagnostics: File file:/home/hadoop/.flink/application_1444046049303_0008/flink-conf.yaml does not exist java.io.FileNotFoundException: File file:/home/hadoop/.flink/application_1444046049303_0008/flink-conf.yaml does not exist at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:539) at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:752) at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:529) at org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:419) at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:251) at org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:61) at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359) at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:357) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:356) at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745)
Failing this attempt. Failing the application. If log aggregation is enabled on your cluster, use this command to further invesitage the issue: yarn logs -applicationId application_1444046049303_0008 org.apache.flink.yarn.FlinkYarnClient$YarnDeploymentException: The YARN application unexpectedly switched to state FAILED during deployment. Diagnostics from YARN: Application application_1444046049303_0008 failed 1 times due to AM Container for appattempt_1444046049303_0008_000001 exited with exitCode: -1000 For more detailed output, check application tracking page:http://ip-172-31-10-16.us<http://ip-172-31-10-16.us>-west-2.compute.internal:20888/proxy/application_1444046049303_0008/Then, click on links to logs of each attempt. Diagnostics: File file:/home/hadoop/.flink/application_1444046049303_0008/flink-conf.yaml does not exist java.io.FileNotFoundException: File file:/home/hadoop/.flink/application_1444046049303_0008/flink-conf.yaml does not exist at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:539) at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:752) at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:529) at org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:419) at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:251) at org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:61) at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359) at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:357) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:356) at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Failing this attempt. Failing the application. If log aggregation is enabled on your cluster, use this command to further invesitage the issue: yarn logs -applicationId application_1444046049303_0008 at org.apache.flink.yarn.FlinkYarnClient.deployInternal(FlinkYarnClient.java:627) at org.apache.flink.yarn.FlinkYarnClient.deploy(FlinkYarnClient.java:335) at org.apache.flink.client.FlinkYarnSessionCli.run(FlinkYarnSessionCli.java:403) at org.apache.flink.client.FlinkYarnSessionCli.main(FlinkYarnSessionCli.java:344) ------------ Could anyone help me to solve this? Thanks & Best Regards, Hanen