Hi Gordon, Here's what I use: - Flink: I didn't build Flink myself. I download http://apache.mirrors.lucidnetworks.net/flink/flink-1.3.0/flink-1.3.0-bin-hadoop27-scala_2.11.tgz from https://flink.apache.org/downloads.html (Hadoop® 2.7, Scala 2.11) - flink-kinesis-connector: I build flink-connector-kinesis_2.11-1.3.0.jar myself, from source code downloaded at *#Source* section in https://flink.apache.org/downloads.html. - Mvn -v: Apache Maven 3.2.5
In short, I didn't build Flink. Most likely that dependencies in either flink-dist or flink-kinesis-connector is not shaded properly? Thanks! Bowen On Mon, Jun 19, 2017 at 9:28 PM, Tzu-Li (Gordon) Tai <tzuli...@apache.org> wrote: > Hi, > > We’ve seen this issue before [1]. The usual reason is that the > httpcomponent dependencies weren’t properly shaded in the flink-dist jar. > Having them properly shaded should solve the issue. > > cc Bowen: > Are you building Flink yourself? If yes, what Maven version are you using? > If you’re using 3.3.x+, after the first build under flink/, make sure to go > to flink-dist/ and build a second time for the dependencies to be properly > shaded. > Alternatively, Maven 3.0.x+ is the recommended version, as 3.3.x has > dependency shading issues. > > If you’re not building Flink yourself, the cause could be that the Flink > 1.3.0 flink-dist jar wasn’t shaded properly, may need to double check. > > Best, > Gordon > > [1] https://issues.apache.org/jira/browse/FLINK-5013 > > On 20 June 2017 at 12:14:27 PM, Ted Yu (yuzhih...@gmail.com) wrote: > > I logged FLINK-6951, referencing this thread. > > We can continue discussion there. > > Thanks > > On Mon, Jun 19, 2017 at 9:06 PM, Bowen Li <bowen...@offerupnow.com> wrote: > > > Thanks, Ted! woo, this is unexpected. https://ci.apache. > > org/projects/flink/flink-docs-release-1.3/setup/aws.html is out of date. > > > > I bet anyone using Kinesis with Flink will run into this issue. I can try > > to build Flink myself and resolve this problem. But talking about a > > feasible permanent solution for all flink-connector-kinesis users. Shall > we > > downgrade aws-java-sdk-kinesis version in flink-connector-kinesis, or > shall > > we upgrade httpcomponents version in Flink? > > > > Bowen > > > > > > On Mon, Jun 19, 2017 at 7:02 PM, Ted Yu <yuzhih...@gmail.com> wrote: > > > > > Here is the dependency in the flink-connector-kinesis module: > > > > > > [INFO] +- com.amazonaws:aws-java-sdk-kinesis:jar:1.10.71:compile > > > [INFO] | \- com.amazonaws:aws-java-sdk-core:jar:1.10.71:compile > > > [INFO] | +- org.apache.httpcomponents:httpclient:jar:4.3.6:compile > > > [INFO] | +- org.apache.httpcomponents:httpcore:jar:4.3.3:compile > > > > > > Checking dependency tree of flink, the highest version is 4.2.x > > > > > > You can try building flink with dependency on 4.3.y of httpclient / > > > httpcore > > > > > > FYI > > > > > > > > > > > > On Mon, Jun 19, 2017 at 4:52 PM, Bowen Li <bowen...@offerupnow.com> > > wrote: > > > > > > > Hi guys, > > > > I'm trying to enable Flink's checkpoint on our Flink app. I got the > > > > following Apache http jar compatibility error, and cannot figure out > > how > > > to > > > > resolve it. > > > > > > > > Here's the stacktrace: > > > > > > > > ``` > > > > > > > > 2017-06-19 15:07:39,828 INFO org.apache.flink.runtime. > > taskmanager.Task > > > > - Source: Custom Source -> (Timestamps/Watermarks, > > Filter > > > > -> Map, Filter -> Map, Filter -> Map) (1/1) > > > > (37ab9429deda28e31fa0ed0ed1568654) switched from RUNNING to FAILED. > > > > java.lang.NoSuchFieldError: INSTANCE > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.http.conn. > > > > SdkConnectionKeepAliveStrategy.getKeepAliveDuration( > > > > SdkConnectionKeepAliveStrategy.java:48) > > > > at > > > > org.apache.http.impl.client.DefaultRequestDirector.execute( > > > > DefaultRequestDirector.java:535) > > > > at > > > > org.apache.http.impl.client.AbstractHttpClient.execute( > > > > AbstractHttpClient.java:906) > > > > at > > > > org.apache.http.impl.client.AbstractHttpClient.execute( > > > > AbstractHttpClient.java:805) > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.http.AmazonHttpClient. > > > > executeOneRequest(AmazonHttpClient.java:837) > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.http.AmazonHttpClient. > > > > executeHelper(AmazonHttpClient.java:607) > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.http. > > > > AmazonHttpClient.doExecute(AmazonHttpClient.java:376) > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.http.AmazonHttpClient. > > > > executeWithTimer(AmazonHttpClient.java:338) > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.http. > > > > AmazonHttpClient.execute(AmazonHttpClient.java:287) > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.services. > > > > kinesis.AmazonKinesisClient.doInvoke(AmazonKinesisClient.java:1940) > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.services. > > > > kinesis.AmazonKinesisClient.invoke(AmazonKinesisClient.java:1910) > > > > at > > > > org.apache.flink.kinesis.shaded.com.amazonaws.services. > > > > kinesis.AmazonKinesisClient.describeStream( > > AmazonKinesisClient.java:656) > > > > at > > > > org.apache.flink.streaming.connectors.kinesis.proxy. > > > > KinesisProxy.describeStream(KinesisProxy.java:361) > > > > at > > > > org.apache.flink.streaming.connectors.kinesis.proxy.KinesisProxy. > > > > getShardsOfStream(KinesisProxy.java:323) > > > > at > > > > org.apache.flink.streaming.connectors.kinesis.proxy. > > > > KinesisProxy.getShardList(KinesisProxy.java:231) > > > > at > > > > org.apache.flink.streaming.connectors.kinesis.internals. > > > > KinesisDataFetcher.discoverNewShardsToSubscribe( > > > > KinesisDataFetcher.java:430) > > > > at > > > > org.apache.flink.streaming.connectors.kinesis. > > FlinkKinesisConsumer.run( > > > > FlinkKinesisConsumer.java:202) > > > > at > > > > org.apache.flink.streaming.api.operators.StreamSource. > > > > run(StreamSource.java:87) > > > > at > > > > org.apache.flink.streaming.api.operators.StreamSource. > > > > run(StreamSource.java:55) > > > > at > > > > org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run( > > > > SourceStreamTask.java:95) > > > > at > > > > org.apache.flink.streaming.runtime.tasks.StreamTask. > > > > invoke(StreamTask.java:262) > > > > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:702) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > > > ``` > > > > > > > > Here's my Flink environment setup: > > > > > > > > - I'm using flink-connector-kinesis_2.11-1.3.0.jar that I built > > from > > > > 1.3.0 source code. > > > > - I followed > > > > https://ci.apache.org/projects/flink/flink-docs- > > > > release-1.3/setup/aws.html#flink-for-hadoop-27 > > > > and added all necessary dependency jars > > > > - My application doesn't use Apache http/core. > > > > > > > > > > > > Has anyone experienced the similar incompatibility issue? > > > > > > > > Thanks! > > > > Bowen > > > > > > > > > >