Hi Andy,

I remember that I was testing a job with almost exactly the same setup as part 
of the Flink 1.3.2 release testing. The command I used to start my job is 
roughly this:

HADOOP_CONF_DIR=/etc/hadoop/conf bin/flink run -c my.main.Class -m yarn-cluster 
-yn 5 -ys 2 ...

i.e. I export the proper hadoop config dir and I run a per-job YARN cluster. I 
think I also exported the result of "hadoop classpath" as HADOOP_CLASSPATH.

Best,
Aljoscha

> On 10. Oct 2017, at 16:43, Andy M. <ajm2...@gmail.com> wrote:
> 
> Hello,
> 
> Bowen:  Unless I am missing something, it says there needs to be no setup
> on EMR, Each topic says: "You don’t have to configure this manually if you
> are running Flink on EMR."  S3 access from CLI works fine on my clusters.
> 
> Chen: Thank you for this, I will look into this if I am unable to get this
> running on YARN successfully.
> 
> Stephan:  Removing the said library causes the flink
> (flink-1.3.2/bin/flink) bash script to fail.  The underlying Java needs
> this to work.  I tried explicitly setting the classpath for the java call
> as well to point to the hadoop library jars.  This is the original java
> command that I was trying to run:
> 
> java
> -Dlog.file=/home/hadoop/flink-1.3.2/log/flink-hadoop-client-ip-172-31-19-27.log
> -Dlog4j.configuration=file:/home/hadoop/flink-1.3.2/conf/log4j-cli.properties
> -Dlogback.configurationFile=file:/home/hadoop/flink-1.3.2/conf/logback.xml
> -classpath
> /home/hadoop/flink-1.3.2/lib/flink-python_2.11-1.3.2.jar:/home/hadoop/flink-1.3.2/lib/flink-shaded-hadoop2-uber-1.3.2.jar:/home/hadoop/flink-1.3.2/lib/log4j-1.2.17.jar:/home/hadoop/flink-1.3.2/lib/slf4j-log4j12-1.7.7.jar:/home/hadoop/flink-1.3.2/lib/flink-dist_2.11-1.3.2.jar::/etc/hadoop/conf:
> org.apache.flink.client.CliFrontend run -m yarn-cluster -yn 1
> /home/hadoop/flink-consumer.jar
> 
> 
> This is what I changed it too(removing the shadded-hadoop2-uber jar and
> adding in the hadoop folder):
> 
> java
> -Dlog.file=/home/hadoop/flink-1.3.2/log/flink-hadoop-client-ip-172-31-19-27.log
> -Dlog4j.configuration=file:/home/hadoop/flink-1.3.2/conf/log4j-cli.properties
> -Dlogback.configurationFile=file:/home/hadoop/flink-1.3.2/conf/logback.xml
> -classpath
> /home/hadoop/flink-1.3.2/lib/flink-python_2.11-1.3.2.jar:/home/hadoop/flink-1.3.2/lib/log4j-1.2.17.jar:/home/hadoop/flink-1.3.2/lib/slf4j-log4j12-1.7.7.jar:/home/hadoop/flink-1.3.2/lib/flink-dist_2.11-1.3.2.jar:/usr/lib/hadoop/lib/activation-1.1.jar:/usr/lib/hadoop/lib/commons-io-2.4.jar:/usr/lib/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/lib/hadoop/lib/log4j-1.2.17.jar:/usr/lib/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/usr/lib/hadoop/lib/commons-lang-2.6.jar:/usr/lib/hadoop/lib/jackson-xc-1.9.13.jar:/usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/lib/hadoop/lib/commons-logging-1.1.3.jar:/usr/lib/hadoop/lib/java-xmlbuilder-0.4.jar:/usr/lib/hadoop/lib/netty-3.6.2.Final.jar:/usr/lib/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/lib/hadoop/lib/commons-math3-3.1.1.jar:/usr/lib/hadoop/lib/jaxb-api-2.2.2.jar:/usr/lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/api-util-1.0.0-M20.jar:/usr/lib/hadoop/lib/commons-net-3.1.jar:/usr/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/lib/hadoop/lib/protobuf-java-2.5.0.jar:/usr/lib/hadoop/lib/asm-3.2.jar:/usr/lib/hadoop/lib/curator-client-2.7.1.jar:/usr/lib/hadoop/lib/jersey-core-1.9.jar:/usr/lib/hadoop/lib/servlet-api-2.5.jar:/usr/lib/hadoop/lib/avro-1.7.4.jar:/usr/lib/hadoop/lib/curator-framework-2.7.1.jar:/usr/lib/hadoop/lib/jersey-json-1.9.jar:/usr/lib/hadoop/lib/slf4j-api-1.7.10.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/lib/hadoop/lib/curator-recipes-2.7.1.jar:/usr/lib/hadoop/lib/jersey-server-1.9.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop/lib/gson-2.2.4.jar:/usr/lib/hadoop/lib/jets3t-0.9.0.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/guava-11.0.2.jar:/usr/lib/hadoop/lib/jettison-1.1.jar:/usr/lib/hadoop/lib/stax-api-1.0-2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:/usr/lib/hadoop/lib/hamcrest-core-1.3.jar:/usr/lib/hadoop/lib/jetty-6.1.26-emr.jar:/usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/commons-collections-3.2.2.jar:/usr/lib/hadoop/lib/htrace-core-3.1.0-incubating.jar:/usr/lib/hadoop/lib/jetty-util-6.1.26-emr.jar:/usr/lib/hadoop/lib/xz-1.0.jar:/usr/lib/hadoop/lib/commons-compress-1.4.1.jar:/usr/lib/hadoop/lib/httpclient-4.5.3.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:/usr/lib/hadoop/lib/zookeeper-3.4.10.jar:/usr/lib/hadoop/lib/commons-configuration-1.6.jar:/usr/lib/hadoop/lib/httpcore-4.4.4.jar:/usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/lib/commons-digester-1.8.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.9.13.jar:/usr/lib/hadoop/lib/jsr305-3.0.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop/lib/jackson-jaxrs-1.9.13.jar:/usr/lib/hadoop/lib/junit-4.11.jar:/etc/hadoop/conf
> org.apache.flink.client.CliFrontend run -m yarn-cluster -yn 1
> /home/hadoop/flink-consumer.jar
> 
> The later throws the following error:
> 
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in
> [jar:file:/home/hadoop/flink-1.3.2/lib/slf4j-log4j12-1.7.7.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> Exception in thread "main" java.lang.NoClassDefFoundError:
> org/apache/hadoop/util/VersionInfo
>        at
> org.apache.flink.runtime.util.EnvironmentInformation.logEnvironmentInfo(EnvironmentInformation.java:283)
>        at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1124)
> Caused by: java.lang.ClassNotFoundException:
> org.apache.hadoop.util.VersionInfo
>        at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>        at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)
>        at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>        ... 2 more
> 
> Simply removing the .jar from the folder causes the same error.
> 
> Thank you
> 
> 
> 
> 
> On Mon, Oct 9, 2017 at 5:46 AM, Stephan Ewen <se...@apache.org> wrote:
> 
>> Hi!
>> 
>> It looks like multiple Hadoop versions are in the classpath. Flink's hadoop
>> jar and the EMR Hadoop jars.
>> I would simply drop Flink's own Hadoop dependency and only use the EMR
>> Hadoop jars.
>> 
>> Delete the 'flink-shaded-h‌​adoop2-uber' jar from Flink, and make sure the
>> setup is such that the Hadoop lib environment variable is set. Then it
>> should not have conflicts any more.
>> 
>> 
>> 
>> On Sun, Oct 8, 2017 at 12:08 AM, Chen Qin <qinnc...@gmail.com> wrote:
>> 
>>> Attached my side project verified working to deploy jobmanager and
>>> taskmanager as stateless service(non yarn/mesos), configuration here
>>> 
>>> https://github.com/chenqin/flink-jar/tree/master/config/hadoop
>>> 
>>> more detail here
>>> https://github.com/chenqin/flink-jar/blob/master/src/
>>> main/java/FlinkBootstrap.java#L49
>>> 
>>> On Fri, Oct 6, 2017 at 10:26 PM, Bowen Li <bowen...@offerupnow.com>
>> wrote:
>>> 
>>>> Hi Andy,
>>>> 
>>>> I believe it's because you didn't set your s3 impl correctly. Try to
>> set
>>>> your core-site.xml by following https://ci.apache.org/
>>>> projects/flink/flink-docs-release-1.4/ops/deployment/
>>>> aws.html#s3afilesystem-
>>>> recommended
>>>> 
>>>> Bowen
>>>> 
>>>> On Fri, Oct 6, 2017 at 7:59 AM, Andy M. <ajm2...@gmail.com> wrote:
>>>> 
>>>>> Hi Till,
>>>>> 
>>>>> Seems like everything is in line there.  hadoop-common.jar ->
>>>>> hadoop-common-2.7.3-amzn-3.jar
>>>>> 
>>>>> And when i decompiled that jar I see  public void
>>>> addResource(Configuration
>>>>> conf) in org/apache/hadoop/conf/Configuration.java
>>>>> 
>>>>> I agree that an incorrect version of the jar is probably being run,
>> is
>>>>> there a way to limit the classpath for the TaskManager when starting
>>> the
>>>>> job?
>>>>> 
>>>>> Thank you
>>>>> 
>>>>> On Fri, Oct 6, 2017 at 6:49 AM, Till Rohrmann <trohrm...@apache.org>
>>>>> wrote:
>>>>> 
>>>>>> Hi Andy,
>>>>>> 
>>>>>> could you check which Hadoop version this jar
>>>>>> /usr/lib/hadoop/hadoop-common.jar is? Maybe also checking whether
>>> the
>>>>>> contained hadoop Configuration class has the method
>>>>>> Configuration.addResource(Lorg/apache/hadoop/conf/
>> Configuration;)V.
>>>>> Maybe
>>>>>> this jar is the culprit because it comes from a different Hadoop
>>>> version.
>>>>>> 
>>>>>> Cheers,
>>>>>> Till
>>>>>> ​
>>>>>> 
>>>>>> On Thu, Oct 5, 2017 at 4:22 PM, Andy M. <ajm2...@gmail.com> wrote:
>>>>>> 
>>>>>>> Hi Till,
>>>>>>> 
>>>>>>> I believe this is what you are looking for, classpath is much
>>> bigger
>>>>> for
>>>>>>> the task manager.  I can also post the whole log file if needed:
>>>>>>> 
>>>>>>> 2017-10-05 14:17:53,038 INFO  org.apache.flink.yarn.
>>>>>> YarnTaskManagerRunner
>>>>>>>                 -  Classpath:
>>>>>>> flink-consumer.jar:lib/flink-dist_2.11-1.3.2.jar:lib/flink-
>>>>>>> python_2.11-1.3.2.jar:lib/flink-shaded-hadoop2-uber-1.3.
>>>>>>> 2.jar:lib/log4j-1.2.17.jar:lib/slf4j-log4j12-1.7.7.jar:
>>>>>>> log4j.properties:logback.xml:flink.jar:flink-conf.yaml::/
>>>>>>> etc/hadoop/conf:/usr/lib/hadoop/hadoop-common-2.7.3-
>>>>>>> amzn-3-tests.jar:/usr/lib/hadoop/hadoop-annotations-2.7.
>>>>>>> 3-amzn-3.jar:/usr/lib/hadoop/hadoop-distcp.jar:/usr/lib/
>>>>>>> hadoop/hadoop-auth-2.7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-
>>>>>>> nfs-2.7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-streaming-2.7.3-
>>>>>>> amzn-3.jar:/usr/lib/hadoop/hadoop-ant-2.7.3-amzn-3.jar:/
>>>>>>> usr/lib/hadoop/hadoop-distcp-2.7.3-amzn-3.jar:/usr/lib/
>>>>>>> hadoop/hadoop-datajoin.jar:/usr/lib/hadoop/hadoop-
>>>>>>> streaming.jar:/usr/lib/hadoop/hadoop-common.jar:/usr/lib/
>>>>>>> hadoop/hadoop-ant.jar:/usr/lib/hadoop/hadoop-sls.jar:/
>>>>>>> usr/lib/hadoop/hadoop-azure-2.7.3-amzn-3.jar:/usr/lib/
>>>>>>> hadoop/hadoop-nfs.jar:/usr/lib/hadoop/hadoop-extras-2.7.
>>>>>>> 3-amzn-3.jar:/usr/lib/hadoop/hadoop-gridmix.jar:/usr/lib/
>>>>>>> hadoop/hadoop-common-2.7.3-amzn-3.jar:/usr/lib/hadoop/
>>>>>>> hadoop-annotations.jar:/usr/lib/hadoop/hadoop-openstack-2.
>>>>>>> 7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-archives-2.7.3-
>>>>>>> amzn-3.jar:/usr/lib/hadoop/hadoop-azure.jar:/usr/lib/
>>>>>>> hadoop/hadoop-extras.jar:/usr/lib/hadoop/hadoop-openstack.
>>>>>>> jar:/usr/lib/hadoop/hadoop-rumen.jar:/usr/lib/hadoop/
>>>>>>> hadoop-aws-2.7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-
>>>>>>> datajoin-2.7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-
>>>>>>> archives.jar:/usr/lib/hadoop/hadoop-aws.jar:/usr/lib/
>>>>>>> hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-rumen-2.7.3-
>>>>>>> amzn-3.jar:/usr/lib/hadoop/hadoop-sls-2.7.3-amzn-3.jar:/
>>>>>>> usr/lib/hadoop/hadoop-gridmix-2.7.3-amzn-3.jar:/usr/lib/
>>>>>>> hadoop/lib/jettison-1.1.jar:/usr/lib/hadoop/lib/jaxb-api-2.
>>>>>>> 2.2.jar:/usr/lib/hadoop/lib/htrace-core-3.1.0-incubating.
>>>>>>> jar:/usr/lib/hadoop/lib/protobuf-java-2.5.0.jar:/usr/
>>>>>>> lib/hadoop/lib/httpclient-4.5.3.jar:/usr/lib/hadoop/lib/
>>>>>>> httpcore-4.4.4.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.
>>>>>>> 1.jar:/usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.
>>>>>>> jar:/usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/
>>>>>>> lib/activation-1.1.jar:/usr/lib/hadoop/lib/jersey-server-
>>>>>>> 1.9.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:/
>>>>>>> usr/lib/hadoop/lib/guava-11.0.2.jar:/usr/lib/hadoop/lib/
>>>>>>> gson-2.2.4.jar:/usr/lib/hadoop/lib/commons-digester-1.
>>>>>>> 8.jar:/usr/lib/hadoop/lib/jackson-xc-1.9.13.jar:/usr/
>>>>>>> lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/
>>>>>>> apacheds-i18n-2.0.0-M15.jar:/usr/lib/hadoop/lib/commons-
>>>>>>> httpclient-3.1.jar:/usr/lib/hadoop/lib/curator-client-2.7.
>>>>>>> 1.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/
>>>>>>> lib/hadoop/lib/commons-codec-1.4.jar:/usr/lib/hadoop/lib/
>>>>>>> commons-net-3.1.jar:/usr/lib/hadoop/lib/jets3t-0.9.0.jar:/
>>>>>>> usr/lib/hadoop/lib/commons-io-2.4.jar:/usr/lib/hadoop/lib/
>>>>>>> xmlenc-0.52.jar:/usr/lib/hadoop/lib/jersey-json-1.9.
>>>>>>> jar:/usr/lib/hadoop/lib/avro-1.7.4.jar:/usr/lib/hadoop/lib/
>>>>>>> commons-math3-3.1.1.jar:/usr/lib/hadoop/lib/log4j-1.2.17.
>>>>>>> jar:/usr/lib/hadoop/lib/commons-compress-1.4.1.jar:/
>>>>>>> usr/lib/hadoop/lib/jsch-0.1.42.jar:/usr/lib/hadoop/lib/
>>>>>>> jaxb-impl-2.2.3-1.jar:/usr/lib/hadoop/lib/netty-3.6.2.
>>>>>>> Final.jar:/usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/
>>>>>>> lib/hadoop/lib/asm-3.2.jar:/usr/lib/hadoop/lib/api-asn1-
>>>>>>> api-1.0.0-M20.jar:/usr/lib/hadoop/lib/jackson-mapper-asl-
>>>>>>> 1.9.13.jar:/usr/lib/hadoop/lib/jetty-util-6.1.26-emr.jar:
>>>>>>> /usr/lib/hadoop/lib/xz-1.0.jar:/usr/lib/hadoop/lib/
>>>>>>> jackson-core-asl-1.9.13.jar:/usr/lib/hadoop/lib/commons-
>>>>>>> cli-1.2.jar:/usr/lib/hadoop/lib/servlet-api-2.5.jar:/usr/
>>>>>>> lib/hadoop/lib/jetty-6.1.26-emr.jar:/usr/lib/hadoop/lib/
>>>>>>> apacheds-kerberos-codec-2.0.0-M15.jar:/usr/lib/hadoop/lib/
>>>>>>> commons-collections-3.2.2.jar:/usr/lib/hadoop/lib/junit-4.
>>>>>>> 11.jar:/usr/lib/hadoop/lib/commons-configuration-1.6.jar:
>>>>>>> /usr/lib/hadoop/lib/api-util-1.0.0-M20.jar:/usr/lib/hadoop/
>>>>>>> lib/stax-api-1.0-2.jar:/usr/lib/hadoop/lib/slf4j-api-1.7.
>>>>>>> 10.jar:/usr/lib/hadoop/lib/curator-recipes-2.7.1.jar:/
>>>>>>> usr/lib/hadoop/lib/hamcrest-core-1.3.jar:/usr/lib/hadoop/
>>>>>>> lib/jsr305-3.0.0.jar:/usr/lib/hadoop/lib/zookeeper-3.4.10.
>>>>>>> jar:/usr/lib/hadoop/lib/commons-lang-2.6.jar:/usr/lib/
>>>>>>> hadoop/lib/jersey-core-1.9.jar:/usr/lib/hadoop/lib/
>>>>>>> curator-framework-2.7.1.jar:/usr/lib/hadoop/lib/jackson-
>>>>>>> jaxrs-1.9.13.jar:/usr/lib/hadoop/lib/java-xmlbuilder-0.
>>>>>>> 4.jar:/usr/lib/hadoop/lib/commons-logging-1.1.3.jar:/
>>>>>>> usr/lib/hadoop-hdfs/hadoop-hdfs-nfs-2.7.3-amzn-3.jar:/
>>>>>>> usr/lib/hadoop-hdfs/hadoop-hdfs-2.7.3-amzn-3-tests.jar:/
>>>>>>> usr/lib/hadoop-hdfs/hadoop-hdfs.jar:/usr/lib/hadoop-hdfs/
>>>>>>> hadoop-hdfs-nfs.jar:/usr/lib/hadoop-hdfs/hadoop-hdfs-2.7.3-
>>>>>>> amzn-3.jar:/usr/lib/hadoop-hdfs/lib/htrace-core-3.1.0-
>>>>>>> incubating.jar:/usr/lib/hadoop-hdfs/lib/protobuf-java-
>>>>>>> 2.5.0.jar:/usr/lib/hadoop-hdfs/lib/leveldbjni-all-1.8.
>>>>>>> jar:/usr/lib/hadoop-hdfs/lib/jersey-server-1.9.jar:/usr/
>>>>>>> lib/hadoop-hdfs/lib/guava-11.0.2.jar:/usr/lib/hadoop-hdfs/
>>>>>>> lib/commons-codec-1.4.jar:/usr/lib/hadoop-hdfs/lib/
>>>>>>> commons-io-2.4.jar:/usr/lib/hadoop-hdfs/lib/xmlenc-0.52.
>>>>>>> jar:/usr/lib/hadoop-hdfs/lib/netty-all-4.0.23.Final.jar:/
>>>>>>> usr/lib/hadoop-hdfs/lib/xercesImpl-2.9.1.jar:/usr/lib/
>>>>>>> hadoop-hdfs/lib/log4j-1.2.17.jar:/usr/lib/hadoop-hdfs/lib/
>>>>>>> netty-3.6.2.Final.jar:/usr/lib/hadoop-hdfs/lib/xml-apis-
>>>>>>> 1.3.04.jar:/usr/lib/hadoop-hdfs/lib/asm-3.2.jar:/usr/lib/
>>>>>>> hadoop-hdfs/lib/jackson-mapper-asl-1.9.13.jar:/usr/
>>>>>>> lib/hadoop-hdfs/lib/commons-daemon-1.0.13.jar:/usr/lib/
>>>>>>> hadoop-hdfs/lib/jetty-util-6.1.26-emr.jar:/usr/lib/hadoop-
>>>>>>> hdfs/lib/jackson-core-asl-1.9.13.jar:/usr/lib/hadoop-hdfs/
>>>>>>> lib/commons-cli-1.2.jar:/usr/lib/hadoop-hdfs/lib/servlet-
>>>>>>> api-2.5.jar:/usr/lib/hadoop-hdfs/lib/jetty-6.1.26-emr.jar:
>>>>>>> /usr/lib/hadoop-hdfs/lib/jsr305-3.0.0.jar:/usr/lib/
>>>>>>> hadoop-hdfs/lib/commons-lang-2.6.jar:/usr/lib/hadoop-hdfs/
>>>>>>> lib/jersey-core-1.9.jar:/usr/lib/hadoop-hdfs/lib/commons-
>>>>>>> logging-1.1.3.jar:/usr/lib/hadoop-mapreduce/jettison-1.1.
>>>>>>> jar:/usr/lib/hadoop-mapreduce/jaxb-api-2.2.2.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/hadoop-distcp.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/htrace-core-3.1.0-incubating.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/hadoop-auth-2.7.3-amzn-3.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/protobuf-java-2.5.0.jar:/usr/lib/
>>>>>> hadoop-mapreduce/hadoop-
>>>>>>> mapreduce-client-core.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> httpclient-4.5.3.jar:/usr/lib/hadoop-mapreduce/hadoop-
>>>>>>> mapreduce-client-shuffle.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-mapreduce-client-common-2.7.3-amzn-3.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/httpcore-4.4.4.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/snappy-java-1.0.4.1.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-mapreduce-client-core-2.7.3-amzn-3.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/jsp-api-2.1.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> jmespath-java-1.11.160.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-mapreduce-client-hs-plugins.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/activation-1.1.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-mapreduce-client-jobclient-2.7.3-amzn-3-tests.
>>>>>>> jar:/usr/lib/hadoop-mapreduce/jersey-server-1.9.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/hadoop-streaming-2.7.3-amzn-3.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/commons-beanutils-1.7.0.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/hadoop-ant-2.7.3-amzn-3.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/guava-11.0.2.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/gson-2.2.4.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-distcp-2.7.3-amzn-3.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-mapreduce-client-app-2.7.3-amzn-3.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/hadoop-mapreduce-examples-2.7.3-amzn-
>>>>>>> 3.jar:/usr/lib/hadoop-mapreduce/commons-digester-1.
>>>>>> 8.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/hadoop-mapreduce-client-hs-plugins-2.7.3-amzn-
>>>>>>> 3.jar:/usr/lib/hadoop-mapreduce/jackson-xc-1.9.13.
>>>>>>> jar:/usr/lib/hadoop-mapreduce/paranamer-2.3.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/apacheds-i18n-2.0.0-M15.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/commons-httpclient-3.1.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/curator-client-2.7.1.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/hadoop-datajoin.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/hadoop-streaming.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> commons-codec-1.4.jar:/usr/lib/hadoop-mapreduce/commons-
>>>>>>> net-3.1.jar:/usr/lib/hadoop-mapreduce/jets3t-0.9.0.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/commons-io-2.4.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/hadoop-ant.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> xmlenc-0.52.jar:/usr/lib/hadoop-mapreduce/jersey-json-
>>>>>>> 1.9.jar:/usr/lib/hadoop-mapreduce/hadoop-sls.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/avro-1.7.4.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/aws-java-sdk-core-1.11.160.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/aws-java-sdk-s3-1.11.160.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/jackson-dataformat-cbor-2.6.6.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/hadoop-azure-2.7.3-amzn-3.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/commons-math3-3.1.1.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/ion-java-1.0.2.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/azure-storage-2.0.0.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-mapreduce-client-shuffle-2.7.3-amzn-3.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/log4j-1.2.17.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/commons-compress-1.4.1.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/hadoop-extras-2.7.3-amzn-3.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/hadoop-gridmix.jar:/usr/lib/hadoop-mapreduce/jsch-
>>>>>>> 0.1.42.jar:/usr/lib/hadoop-mapreduce/joda-time-2.8.1.jar:
>>>>>>> /usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-hs-2.
>>>>>>> 7.3-amzn-3.jar:/usr/lib/hadoop-mapreduce/jaxb-impl-2.
>>>>>>> 2.3-1.jar:/usr/lib/hadoop-mapreduce/hadoop-mapreduce-
>>>>>>> client-hs.jar:/usr/lib/hadoop-mapreduce/netty-3.6.2.Final.
>>>>>>> jar:/usr/lib/hadoop-mapreduce/mockito-all-1.8.5.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/hadoop-openstack-2.7.3-amzn-3.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-
>>>>>>> common.jar:/usr/lib/hadoop-mapreduce/hadoop-archives-2.7.
>>>>>>> 3-amzn-3.jar:/usr/lib/hadoop-mapreduce/hadoop-azure.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-
>>>>>>> jobclient.jar:/usr/lib/hadoop-mapreduce/hadoop-extras.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/asm-3.2.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/api-asn1-api-1.0.0-M20.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/hadoop-openstack.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-mapreduce-client-jobclient-2.7.3-amzn-3.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/hadoop-rumen.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/aws-java-sdk-kms-1.11.160.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/jackson-mapper-asl-1.9.13.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/hadoop-aws-2.7.3-amzn-3.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/jetty-util-6.1.26-emr.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/hadoop-datajoin-2.7.3-amzn-3.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/xz-1.0.jar:/usr/lib/hadoop-mapreduce/hadoop-
>>>>>>> archives.jar:/usr/lib/hadoop-mapreduce/jackson-core-asl-1.
>>>>>>> 9.13.jar:/usr/lib/hadoop-mapreduce/commons-lang3-3.3.2.
>>>>>>> jar:/usr/lib/hadoop-mapreduce/jackson-core-2.6.6.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/commons-cli-1.2.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/servlet-api-2.5.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> jetty-6.1.26-emr.jar:/usr/lib/hadoop-mapreduce/apacheds-
>>>>>>> kerberos-codec-2.0.0-M15.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-aws.jar:/usr/lib/hadoop-mapreduce/hadoop-auth.
>>>>>>> jar:/usr/lib/hadoop-mapreduce/hadoop-rumen-2.7.3-amzn-3.jar:
>>>>>>> /usr/lib/hadoop-mapreduce/commons-collections-3.2.2.jar:
>>>>>>> /usr/lib/hadoop-mapreduce/hadoop-sls-2.7.3-amzn-3.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/metrics-core-3.0.1.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/jackson-annotations-2.6.6.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/junit-4.11.jar:/usr/lib/hadoop-
>>>> mapreduce/commons-
>>>>>>> configuration-1.6.jar:/usr/lib/hadoop-mapreduce/api-util-
>>>>>>> 1.0.0-M20.jar:/usr/lib/hadoop-mapreduce/stax-api-1.0-2.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/curator-recipes-2.7.1.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/hamcrest-core-1.3.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/jackson-databind-2.6.6.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/jsr305-3.0.0.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> hadoop-mapreduce-client-app.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> zookeeper-3.4.10.jar:/usr/lib/hadoop-mapreduce/commons-lang-
>>>>>>> 2.6.jar:/usr/lib/hadoop-mapreduce/jersey-core-1.9.jar:
>>>>>>> /usr/lib/hadoop-mapreduce/curator-framework-2.7.1.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/jackson-jaxrs-1.9.13.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/hadoop-gridmix-2.7.3-amzn-3.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar:/usr/
>>>>>>> lib/hadoop-mapreduce/java-xmlbuilder-0.4.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/commons-logging-1.1.3.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/lib/protobuf-java-2.5.0.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/lib/leveldbjni-all-1.8.jar:/usr/
>>>>>> lib/hadoop-mapreduce/lib/
>>>>>>> snappy-java-1.0.4.1.jar:/usr/lib/hadoop-mapreduce/lib/
>>>>>>> jersey-server-1.9.jar:/usr/lib/hadoop-mapreduce/lib/
>>>>>>> paranamer-2.3.jar:/usr/lib/hadoop-mapreduce/lib/jersey-
>>>>>>> guice-1.9.jar:/usr/lib/hadoop-mapreduce/lib/commons-io-2.4.
>>>>>>> jar:/usr/lib/hadoop-mapreduce/lib/avro-1.7.4.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/lib/log4j-1.2.17.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/lib/commons-compress-1.4.1.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/lib/javax.inject-1.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/lib/netty-3.6.2.Final.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/lib/aopalliance-1.0.jar:/usr/lib/hadoop-mapreduce/
>>>>>>> lib/asm-3.2.jar:/usr/lib/hadoop-mapreduce/lib/jackson-
>>>>>>> mapper-asl-1.9.13.jar:/usr/lib/hadoop-mapreduce/lib/
>>>>>>> guice-servlet-3.0.jar:/usr/lib/hadoop-mapreduce/lib/xz-1.
>>>>>>> 0.jar:/usr/lib/hadoop-mapreduce/lib/guice-3.0.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/lib/jackson-core-asl-1.9.13.jar:/
>>>>>>> usr/lib/hadoop-mapreduce/lib/junit-4.11.jar:/usr/lib/
>>>>>>> hadoop-mapreduce/lib/hamcrest-core-1.3.jar:/usr/lib/hadoop-
>>>>>>> mapreduce/lib/jersey-core-1.9.jar:/usr/lib/hadoop-yarn/
>>>>>>> hadoop-yarn-server-common.jar:/usr/lib/hadoop-yarn/hadoop-
>>>> yarn-server-
>>>>>>> sharedcachemanager.jar:/usr/lib/hadoop-yarn/hadoop-yarn-
>>>>>>> server-web-proxy-2.7.3-amzn-3.jar:/usr/lib/hadoop-yarn/
>>>>>>> hadoop-yarn-common-2.7.3-amzn-3.jar:/usr/lib/hadoop-yarn/
>>>>>>> hadoop-yarn-server-tests-2.7.3-amzn-3.jar:/usr/lib/hadoop-
>>>>>>> yarn/hadoop-yarn-applications-distributedshell.jar:/usr/lib/
>>>>>>> hadoop-yarn/hadoop-yarn-applications-distributedshell-
>>>>>>> 2.7.3-amzn-3.jar:/usr/lib/hadoop-yarn/hadoop-yarn-
>>>>>>> server-resourcemanager.jar:/usr/lib/hadoop-yarn/hadoop-
>>> yarn-server-
>>>>>>> sharedcachemanager-2.7.3-amzn-3.jar:/usr/lib/hadoop-yarn/
>>>>>>> hadoop-yarn-server-applicationhistoryservice.jar:
>>>>>>> /usr/lib/hadoop-yarn/hadoop-yarn-server-tests.jar:/usr/
>>>>>>> lib/hadoop-yarn/hadoop-yarn-server-nodemanager-2.7.3-amzn-
>>>>>>> 3.jar:/usr/lib/hadoop-yarn/hadoop-yarn-client-2.7.3-amzn-
>>>>>>> 3.jar:/usr/lib/hadoop-yarn/hadoop-yarn-server-
>>>>>>> applicationhistoryservice-2.7.3-amzn-3.jar:/usr/lib/hadoop-
>>>>>>> yarn/hadoop-yarn-server-common-2.7.3-amzn-3.jar:/usr/
>>>>>>> lib/hadoop-yarn/hadoop-yarn-registry.jar:/usr/lib/hadoop-
>>>>>>> yarn/hadoop-yarn-api-2.7.3-amzn-3.jar:/usr/lib/hadoop-
>>>>>>> yarn/hadoop-yarn-server-resourcemanager-2.7.3-amzn-3.
>>>>>>> jar:/usr/lib/hadoop-yarn/hadoop-yarn-applications-
>>>>>>> unmanaged-am-launcher.jar:/usr/lib/hadoop-yarn/hadoop-
>>>>>>> yarn-registry-2.7.3-amzn-3.jar:/usr/lib/hadoop-yarn/
>>>>>>> hadoop-yarn-client.jar:/usr/lib/hadoop-yarn/hadoop-yarn-
>>>>>>> common.jar:/usr/lib/hadoop-yarn/hadoop-yarn-server-web-
>>>>>>> proxy.jar:/usr/lib/hadoop-yarn/hadoop-yarn-server-
>>>>>>> nodemanager.jar:/usr/lib/hadoop-yarn/hadoop-yarn-api.
>>>>>>> jar:/usr/lib/hadoop-yarn/hadoop-yarn-applications-
>>>>>>> unmanaged-am-launcher-2.7.3-amzn-3.jar:/usr/lib/hadoop-
>>>>>>> yarn/lib/jettison-1.1.jar:/usr/lib/hadoop-yarn/lib/jaxb-
>>>>>>> api-2.2.2.jar:/usr/lib/hadoop-yarn/lib/protobuf-java-2.5.0.
>>>>>>> jar:/usr/lib/hadoop-yarn/lib/leveldbjni-all-1.8.jar:/usr/
>>>>>>> lib/hadoop-yarn/lib/activation-1.1.jar:/usr/lib/
>>>>>>> hadoop-yarn/lib/jersey-server-1.9.jar:/usr/lib/hadoop-yarn/
>>>>>>> lib/guava-11.0.2.jar:/usr/lib/hadoop-yarn/lib/jackson-xc-1.
>>>>>>> 9.13.jar:/usr/lib/hadoop-yarn/lib/commons-codec-1.4.jar:/
>>>>>>> usr/lib/hadoop-yarn/lib/jersey-guice-1.9.jar:/usr/lib/
>>>>>>> hadoop-yarn/lib/commons-io-2.4.jar:/usr/lib/hadoop-yarn/
>>>>>>> lib/jersey-json-1.9.jar:/usr/lib/hadoop-yarn/lib/jersey-
>>>>>>> client-1.9.jar:/usr/lib/hadoop-yarn/lib/log4j-1.2.17.
>>>>>>> jar:/usr/lib/hadoop-yarn/lib/commons-compress-1.4.1.jar:/
>>>>>>> usr/lib/hadoop-yarn/lib/javax.inject-1.jar:/usr/lib/hadoop-
>>>>>>> yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/lib/hadoop-yarn/lib/
>>>>>>> netty-3.6.2.Final.jar:/usr/lib/hadoop-yarn/lib/
>>>>>>> aopalliance-1.0.jar:/usr/lib/hadoop-yarn/lib/asm-3.2.jar:/
>>>>>>> usr/lib/hadoop-yarn/lib/zookeeper-3.4.10-tests.jar:/
>>>>>>> usr/lib/hadoop-yarn/lib/jackson-mapper-asl-1.9.13.jar:
>>>>>>> /usr/lib/hadoop-yarn/lib/guice-servlet-3.0.jar:/usr/
>>>>>>> lib/hadoop-yarn/lib/jetty-util-6.1.26-emr.jar:/usr/lib/
>>>>>>> hadoop-yarn/lib/xz-1.0.jar:/usr/lib/hadoop-yarn/lib/guice-
>>>>>>> 3.0.jar:/usr/lib/hadoop-yarn/lib/jackson-core-asl-1.9.13.
>>>>>>> jar:/usr/lib/hadoop-yarn/lib/commons-cli-1.2.jar:/usr/lib/
>>>>>>> hadoop-yarn/lib/servlet-api-2.5.jar:/usr/lib/hadoop-yarn/
>>>>>>> lib/jetty-6.1.26-emr.jar:/usr/lib/hadoop-yarn/lib/commons-
>>>>>>> collections-3.2.2.jar:/usr/lib/hadoop-yarn/lib/stax-api-
>>>>>>> 1.0-2.jar:/usr/lib/hadoop-yarn/lib/jsr305-3.0.0.jar:/
>>>>>>> usr/lib/hadoop-yarn/lib/zookeeper-3.4.10.jar:/usr/lib/
>>>>>>> hadoop-yarn/lib/commons-lang-2.6.jar:/usr/lib/hadoop-yarn/
>>>>>>> lib/jersey-core-1.9.jar:/usr/lib/hadoop-yarn/lib/jackson-
>>>>>>> jaxrs-1.9.13.jar:/usr/lib/hadoop-yarn/lib/commons-
>>>>>>> logging-1.1.3.jar:/usr/lib/hadoop-lzo/lib/hadoop-lzo.jar:
>>>>>>> /usr/lib/hadoop-lzo/lib/hadoop-lzo-0.4.19.jar:/usr/
>>>>>>> share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/
>>>>>>> jmespath-java-1.11.129.jar:/usr/share/aws/emr/emrfs/lib/
>>>>>>> bcpkix-jdk15on-1.51.jar:/usr/share/aws/emr/emrfs/lib/jcl-
>>>>>>> over-slf4j-1.7.21.jar:/usr/share/aws/emr/emrfs/lib/ion-
>>>>>>> java-1.0.2.jar:/usr/share/aws/emr/emrfs/lib/slf4j-api-1.7.
>>>>>>> 21.jar:/usr/share/aws/emr/emrfs/lib/javax.inject-1.jar:/
>>>>>>> usr/share/aws/emr/emrfs/lib/aopalliance-1.0.jar:/usr/
>>>>>>> share/aws/emr/emrfs/lib/bcprov-jdk15on-1.51.jar:/usr/
>>>>>>> share/aws/emr/emrfs/lib/emrfs-hadoop-assembly-2.18.0.jar:/
>>>>>>> usr/share/aws/emr/emrfs/auxlib/*:/usr/share/aws/emr/
>>>>>>> lib/*:/usr/share/aws/emr/ddb/lib/emr-ddb-hadoop.jar:/usr/
>>>>>>> share/aws/emr/goodies/lib/emr-hadoop-goodies.jar:/usr/share/
>>>>>>> aws/emr/kinesis/lib/emr-kinesis-hadoop.jar:/usr/share/
>>>>>>> aws/emr/cloudwatch-sink/lib/cloudwatch-sink-1.0.0.jar:/
>>>>>>> usr/share/aws/emr/cloudwatch-sink/lib/cloudwatch-sink.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/aws-java-sdk-inspector-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-api-
>>>>>>> gateway-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-
>>>>>>> java-sdk-kinesis-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>> aws-java-sdk-
>>>>>>> elasticloadbalancingv2-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-rekognition-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-test-utils-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> elasticbeanstalk-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-cloudhsm-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-iam-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-cognitoidp-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-
>> mechanicalturkrequester-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> marketplacecommerceanalytics-1.11.160.jar:/usr/share/aws/
>>>>>>> aws-java-sdk/aws-java-sdk-snowball-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-polly-1.11.160.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/jmespath-java-1.11.160.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/aws-java-sdk-codebuild-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-emr-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> simpleworkflow-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-machinelearning-1.11.160.jar:/usr/share/aws/
>>>>>>> aws-java-sdk/aws-java-sdk-pinpoint-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-appstream-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-sqs-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-opensdk-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> greengrass-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-
>>>>>>> java-sdk-ecs-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-xray-1.11.160.jar:/usr/share/aws/aws-java-
>>>>>>> sdk/aws-java-sdk-cloudtrail-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-cloudwatchmetrics-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> elasticloadbalancing-1.11.160.jar:/usr/share/aws/aws-java-
>>>>>>> sdk/aws-java-sdk-codepipeline-1.11.160.jar:/usr/share/aws/
>>>>>>> aws-java-sdk/aws-java-sdk-cloudwatch-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-shield-1.11.160.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/aws-java-sdk-config-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>> stepfunctions-1.11.160.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/aws-java-sdk-gamelift-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-rds-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-core-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> datapipeline-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-s3-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-dax-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-opsworks-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-servicecatalog-1.11.160.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/aws-java-sdk-cloudsearch-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-dms-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> directory-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-
>>>>>>> java-sdk-opsworkscm-1.11.160.jar:/usr/share/aws/aws-java-
>>>>>> sdk/aws-java-sdk-
>>>>>>> cloudformation-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-cloudfront-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-budgets-1.11.160.jar:/usr/share/aws/
>>>>>>> aws-java-sdk/aws-java-sdk-clouddirectory-1.11.160.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/aws-java-sdk-importexport-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-lex-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> marketplaceentitlement-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-dynamodb-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-autoscaling-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>> elastictranscoder-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>> organizations-1.11.160.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/aws-java-sdk-workspaces-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-ssm-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> servermigration-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-events-1.11.160.jar:/usr/share/aws/aws-java-
>>>>>> sdk/aws-java-sdk-
>>>>>>> applicationautoscaling-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-health-1.11.160.jar:/usr/share/aws/
>>>>>>> aws-java-sdk/aws-java-sdk-kms-1.11.160.jar:/usr/share/aws/
>>>>>>> aws-java-sdk/aws-java-sdk-logs-1.11.160.jar:/usr/share/
>>>>>>> aws/aws-java-sdk/aws-java-sdk-codestar-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-route53-1.11.160.jar:
>>>>>>> /usr/share/aws/aws-java-sdk/aws-java-sdk-redshift-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> marketplacemeteringservice-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-sns-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-batch-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-waf-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-simpledb-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-codedeploy-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-ec2-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-models-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> devicefarm-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-
>>>>>>> java-sdk-cognitoidentity-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-lexmodelbuilding-1.11.160.jar:
>>>>>>> /usr/share/aws/aws-java-sdk/aws-java-sdk-directconnect-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> elasticache-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-
>>>>>>> java-sdk-costandusagereport-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-discovery-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-
>> resourcegroupstaggingapi-1.11.
>>>>>>> 160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-ses-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-lambda-
>>>>>>> 1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> workdocs-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-
>>>>>>> java-sdk-code-generator-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-cognitosync-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-efs-1.11.160.jar:/
>>>>>>> usr/share/aws/aws-java-sdk/aws-java-sdk-sts-1.11.160.jar:
>>>>>>> /usr/share/aws/aws-java-sdk/aws-java-sdk-athena-1.11.160.
>>>>>>> jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-codecommit-1.
>>>>>>> 11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> storagegateway-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-lightsail-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-acm-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-1.11.160.jar:/usr/share/aws/aws-
>>>>>>> java-sdk/aws-java-sdk-glacier-1.11.160.jar:/usr/share/aws/
>>>>>>> aws-java-sdk/aws-java-sdk-ecr-1.11.160.jar:/usr/share/aws/
>>>>>>> aws-java-sdk/aws-java-sdk-support-1.11.160.jar:/usr/
>>>>>>> share/aws/aws-java-sdk/aws-java-sdk-codegen-maven-plugin-
>>>>>>> 1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-
>>>>>>> elasticsearch-1.11.160.jar:/usr/share/aws/aws-java-sdk/
>>>>>>> aws-java-sdk-iot-1.11.160.jar
>>>>>>> 
>>>>>>> Thank you
>>>>>>> 
>>>>>>> On Thu, Oct 5, 2017 at 5:25 AM, Till Rohrmann <
>>> trohrm...@apache.org>
>>>>>>> wrote:
>>>>>>> 
>>>>>>>> Hi Andy,
>>>>>>>> 
>>>>>>>> the CliFrontend is not executed via Yarn, thus, it is not
>>> affected
>>>> by
>>>>>>>> dependencies which are added due to the underlying Yarn
>> cluster.
>>>>>>> Therefore,
>>>>>>>> it would be helpful to look at the TaskManager logs. Either you
>>>> have
>>>>>>>> enabled log aggregation on your Yarn cluster, then you can
>> obtain
>>>> the
>>>>>>> logs
>>>>>>>> via `yarn logs -applicationId <application ID>` or you have to
>>>>> retrieve
>>>>>>>> them from the machines where they were running (either by going
>>>>>> directly
>>>>>>>> there or via the Yarn web interface).
>>>>>>>> 
>>>>>>>> Cheers,
>>>>>>>> Till
>>>>>>>> 
>>>>>>>> On Wed, Oct 4, 2017 at 4:27 PM, Andy M. <ajm2...@gmail.com>
>>> wrote:
>>>>>>>> 
>>>>>>>>> Hi Till,
>>>>>>>>> 
>>>>>>>>> That is actually the classpath used by the flink bash
>>> script(that
>>>>>>>> launches
>>>>>>>>> the jar using the java command).  I changed the execute to an
>>>> echo,
>>>>>> and
>>>>>>>>> grabbed that for the CLI arguments.
>>>>>>>>> 
>>>>>>>>> I believe this is the class path from the log file(although
>> it
>>>>> might
>>>>>>> not
>>>>>>>> be
>>>>>>>>> the taskmanager log, is that any different from what would be
>>> in
>>>> my
>>>>>>>>> flink-1.3.2/log folder?):
>>>>>>>>> 
>>>>>>>>> 2017-10-02 20:03:26,450 INFO  org.apache.flink.client.
>>>> CliFrontend
>>>>>>>>>                 -  Classpath:
>>>>>>>>> /home/hadoop/flink-1.3.2/lib/flink-python_2.11-1.3.2.jar:/
>>>>>>>>> home/hadoop/flink-1.3.2/lib/flink-shaded-hadoop2-uber-1.3.
>>>>>>>>> 2.jar:/home/hadoop/flink-1.3.2/lib/log4j-1.2.17.jar:/home/
>>>>>>>>> hadoop/flink-1.3.2/lib/slf4j-log4j12-1.7.7.jar:/home/
>>>>>>>>> hadoop/flink-1.3.2/lib/flink-dist_2.11-1.3.2.jar::/etc/
>>>>> hadoop/conf:
>>>>>>>>> 
>>>>>>>>> If that doesn't seem right, and you can point me in the right
>>>>>> direction
>>>>>>>> as
>>>>>>>>> to where the TaskManager logs would be, I would be happy to
>>> grab
>>>>> the
>>>>>>>>> information your looking for.
>>>>>>>>> 
>>>>>>>>> Thank you
>>>>>>>>> 
>>>>>>>>> On Wed, Oct 4, 2017 at 3:27 AM, Till Rohrmann <
>>>>> trohrm...@apache.org>
>>>>>>>>> wrote:
>>>>>>>>> 
>>>>>>>>>> Hi Andy,
>>>>>>>>>> 
>>>>>>>>>> this looks to me indeed like a dependency problem. I assume
>>>> that
>>>>>> EMR
>>>>>>> or
>>>>>>>>>> something else is pulling in an incompatible version of
>>> Hadoop.
>>>>>>>>>> 
>>>>>>>>>> The classpath you've posted, is this the one logged in the
>>> log
>>>>>> files
>>>>>>>>>> (TaskManager log) or did you compile it yourself? In the
>>> latter
>>>>>> case,
>>>>>>>> it
>>>>>>>>>> would also be helpful to get access to the TaskManager
>> logs.
>>>>>>>>>> 
>>>>>>>>>> Cheers,
>>>>>>>>>> Till
>>>>>>>>>> 
>>>>>>>>>> On Mon, Oct 2, 2017 at 10:20 PM, Andy M. <
>> ajm2...@gmail.com>
>>>>>> wrote:
>>>>>>>>>> 
>>>>>>>>>>> Hi Fabian,
>>>>>>>>>>> 
>>>>>>>>>>> 1) I have looked at the linked docs, and from what I can
>>> tell
>>>>> no
>>>>>>>> setup
>>>>>>>>>>> should really need to be done to get Flink working(Other
>>> than
>>>>>>>>> downloading
>>>>>>>>>>> the correct binaries, which I believe I did)
>>>>>>>>>>> 2) I have downloaded the Flink 1.3.2
>>>> binaries(flink-1.3.2-bin-
>>>>>>>>>>> hadoop27-scala_2.11.tgz
>>>>>>>>>>> <http://apache.claz.org/flink/
>> flink-1.3.2/flink-1.3.2-bin-
>>>>>>>>>>> hadoop27-scala_2.11.tgz>)
>>>>>>>>>>> This is for hadoop 2.7.X, which matches EMR 5.8.0.
>>>>>>>>>>> 
>>>>>>>>>>> I appreciate any help or guidance you can provide me in
>>>> fixing
>>>>> my
>>>>>>>>>> problems,
>>>>>>>>>>> please let me know if there is anything else I can
>> provide
>>>> you.
>>>>>>>>>>> 
>>>>>>>>>>> Thank you
>>>>>>>>>>> 
>>>>>>>>>>> On Mon, Oct 2, 2017 at 4:12 PM, Fabian Hueske <
>>>>> fhue...@gmail.com
>>>>>>> 
>>>>>>>>> wrote:
>>>>>>>>>>> 
>>>>>>>>>>>> Hi Andy,
>>>>>>>>>>>> 
>>>>>>>>>>>> I'm not an AWS expert, so I'll just check on some
>> common
>>>>>> issues.
>>>>>>>>>>>> 
>>>>>>>>>>>> I guess you already had a look at the Flink docs for
>>>> AWS/EMR
>>>>>> but
>>>>>>>> I'll
>>>>>>>>>>> post
>>>>>>>>>>>> the link just be to sure [1].
>>>>>>>>>>>> 
>>>>>>>>>>>> Since you are using Flink 1.3.2 (EMR 5.8.0 comes with
>>> Flink
>>>>>>> 1.3.1)
>>>>>>>>> did
>>>>>>>>>>> you
>>>>>>>>>>>> built Flink yourself or did you download the binaries?
>>>>>>>>>>>> Does the Hadoop version of the Flink build match the
>>> Hadoop
>>>>>>> version
>>>>>>>>> of
>>>>>>>>>>> EMR
>>>>>>>>>>>> 5.8.0, i.e., Hadoop 2.7.x?
>>>>>>>>>>>> 
>>>>>>>>>>>> Best, Fabian
>>>>>>>>>>>> 
>>>>>>>>>>>> [1]
>>>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-
>>>>>>>>>>> release-1.3/setup/aws.html
>>>>>>>>>>>> 
>>>>>>>>>>>> 2017-10-02 21:51 GMT+02:00 Andy M. <ajm2...@gmail.com
>>> :
>>>>>>>>>>>> 
>>>>>>>>>>>>> Hi Fabian,
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Sorry, I just realized I forgot to include that part.
>>>> The
>>>>>>> error
>>>>>>>>>>> returned
>>>>>>>>>>>>> is:
>>>>>>>>>>>>> 
>>>>>>>>>>>>> java.lang.NoSuchMethodError:
>>>>>>>>>>>>> org.apache.hadoop.conf.Configuration.addResource(
>>>>>>>>>>>> Lorg/apache/hadoop/conf/
>>>>>>>>>>>>> Configuration;)V
>>>>>>>>>>>>>    at com.amazon.ws.emr.hadoop.fs.
>>>>> EmrFileSystem.initialize(
>>>>>>>>>>>>> EmrFileSystem.java:93)
>>>>>>>>>>>>>    at org.apache.flink.runtime.fs.
>>>> hdfs.HadoopFileSystem.
>>>>>>>>>>>>> initialize(HadoopFileSystem.java:328)
>>>>>>>>>>>>>    at org.apache.flink.core.fs.FileSystem.
>>>>>>>> getUnguardedFileSystem(
>>>>>>>>>>>>> FileSystem.java:350)
>>>>>>>>>>>>>    at org.apache.flink.core.fs.
>>>> FileSystem.get(FileSystem.
>>>>>>>>> java:389)
>>>>>>>>>>>>>    at org.apache.flink.core.fs.Path.
>>>>>>>> getFileSystem(Path.java:293)
>>>>>>>>>>>>>    at org.apache.flink.runtime.state.filesystem.
>>>>>>>>>>>>> FsCheckpointStreamFactory.<init>(
>>>>> FsCheckpointStreamFactory.
>>>>>>>>> java:99)
>>>>>>>>>>>>>    at org.apache.flink.runtime.state.filesystem.
>>>>>>> FsStateBackend.
>>>>>>>>>>>>> createStreamFactory(FsStateBackend.java:282)
>>>>>>>>>>>>>    at org.apache.flink.contrib.streaming.state.
>>>>>>>>> RocksDBStateBackend.
>>>>>>>>>>>>> createStreamFactory(RocksDBStateBackend.java:273
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I believe it has something to do with the classpath,
>>> but
>>>> I
>>>>> am
>>>>>>>>> unsure
>>>>>>>>>>> why
>>>>>>>>>>>> or
>>>>>>>>>>>>> how to fix it.  The classpath being used during the
>>>>> execution
>>>>>>> is:
>>>>>>>>>>>>> /home/hadoop/flink-1.3.2/lib/
>>>> flink-python_2.11-1.3.2.jar:/
>>>>>>>>>>>>> ho‌​me/hadoop/flink-1.3.‌​2/
>>> lib/flink-shaded-h‌​adoop2-
>>>>>>>>>>>>> uber-1.3.2.ja‌​r:/home/hadoop/
>>>>> flink‌​-1.3.2/lib/log4j-1.2‌​.
>>>>>>>>>>>>> 17.jar:/home/hadoop‌​/flink-1.
>>>>> 3.2/lib/slf‌​4j-log4j12-1.7.7.
>>>>>>>>>>>>> jar‌​:/home/hadoop/flink-‌​1.
>>>>> 3.2/lib/flink-dist‌​_2.11-1.3.
>>>>>>>>>>>>> 2.jar::/et‌​c/hadoop/conf:
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I decompiled flink-shaded-h‌​adoop2-uber-1.3.2.ja‌​r
>>> and
>>>>> it
>>>>>>>> seems
>>>>>>>>>> the
>>>>>>>>>>>>> addResource function does seem to be there.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Thank you
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On Mon, Oct 2, 2017 at 3:43 PM, Fabian Hueske <
>>>>>>> fhue...@gmail.com
>>>>>>>>> 
>>>>>>>>>>> wrote:
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Hi Andy,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> can you describe in more detail what exactly isn't
>>>>> working?
>>>>>>>>>>>>>> Do you see error messages in the log files or on
>> the
>>>>>> console?
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Thanks, Fabian
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 2017-10-02 15:52 GMT+02:00 Andy M. <
>>> ajm2...@gmail.com
>>>>> :
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hello,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I am about to deploy my first Flink projects  to
>>>>>>> production,
>>>>>>>>> but
>>>>>>>>>> I
>>>>>>>>>>> am
>>>>>>>>>>>>>>> running into a very big hurdle.  I am unable to
>>>> launch
>>>>> my
>>>>>>>>> project
>>>>>>>>>>> so
>>>>>>>>>>>> it
>>>>>>>>>>>>>> can
>>>>>>>>>>>>>>> write to an S3 bucket.  My project is running on
>> an
>>>> EMR
>>>>>>>>> cluster,
>>>>>>>>>>>> where
>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> have installed Flink 1.3.2.  I am using Yarn to
>>>> launch
>>>>>> the
>>>>>>>>>>>> application,
>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>> it seems to run fine unless I am trying to enable
>>>> check
>>>>>>>>>>>> pointing(with a
>>>>>>>>>>>>>> S3
>>>>>>>>>>>>>>> target).  I am looking to use RocksDB as my
>>>>>> check-pointing
>>>>>>>>>> backend.
>>>>>>>>>>>> I
>>>>>>>>>>>>>> have
>>>>>>>>>>>>>>> asked a few places, and I am still unable to
>> find a
>>>>>>> solution
>>>>>>>> to
>>>>>>>>>>> this
>>>>>>>>>>>>>>> problem.  Here are my steps for creating a
>> cluster,
>>>> and
>>>>>>>>> launching
>>>>>>>>>>> my
>>>>>>>>>>>>>>> application, perhaps I am missing a step.  I'd be
>>>> happy
>>>>>> to
>>>>>>>>>> provide
>>>>>>>>>>>> any
>>>>>>>>>>>>>>> additional information if needed.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> AWS Portal:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>    1) EMR -> Create Cluster
>>>>>>>>>>>>>>>    2) Advanced Options
>>>>>>>>>>>>>>>    3) Release = emr-5.8.0
>>>>>>>>>>>>>>>    4) Only select Hadoop 2.7.3
>>>>>>>>>>>>>>>    5) Next -> Next -> Next -> Create Cluster ( I
>>> do
>>>>> fill
>>>>>>> out
>>>>>>>>>>>>>>> names/keys/etc)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Once the cluster is up I ssh into the Master and
>> do
>>>> the
>>>>>>>>>> following:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>    1  wget
>>>>>>>>>>>>>>> http://apache.claz.org/flink/
>>>>>> flink-1.3.2/flink-1.3.2-bin-
>>>>>>>>>>>>>>> hadoop27-scala_2.11.tgz
>>>>>>>>>>>>>>>    2  tar -xzf flink-1.3.2-bin-hadoop27-
>>>>> scala_2.11.tgz
>>>>>>>>>>>>>>>    3  cd flink-1.3.2
>>>>>>>>>>>>>>>    4  ./bin/yarn-session.sh -n 2 -tm 5120 -s 4
>> -d
>>>>>>>>>>>>>>>    5  Change conf/flink-conf.yaml
>>>>>>>>>>>>>>>    6  ./bin/flink run -m yarn-cluster -yn 1
>>>>>>>>> ~/flink-consumer.jar
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> My conf/flink-conf.yaml I add the following
>> fields:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>    state.backend: rocksdb
>>>>>>>>>>>>>>>    state.backend.fs.checkpointdir:
>>>>> s3:/bucket/location
>>>>>>>>>>>>>>>    state.checkpoints.dir: s3:/bucket/location
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> My program's checkpointing setup:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> env.enableCheckpointing(getCheckpointRate,
>>>>>>>>>>> CheckpointingMode.EXACTLY_
>>>>>>>>>>>>>> ONCE)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> env.getCheckpointConfig.
>>>> enableExternalizedCheckpoints(
>>>>>>>>>>>>>>> ExternalizedCheckpointCleanup.
>>>> RETAIN_ON_CANCELLATION)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> env.getCheckpointConfig.
>>>> setMinPauseBetweenCheckpoints(
>>>>>>>>>>>>>>> getCheckpointMinPause)
>>>>>>>>>>>>>>>    env.getCheckpointConfig.
>> setCheckpointTimeout(
>>>>>>>>>>>> getCheckpointTimeout)
>>>>>>>>>>>>>>>    env.getCheckpointConfig.
>>>>>> setMaxConcurrentCheckpoints(1)
>>>>>>>>>>>>>>>    env.setStateBackend(new
>>>> RocksDBStateBackend("s3://
>>>>>>>>>>>>> bucket/location",
>>>>>>>>>>>>>>> true))
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> 

Reply via email to