Updating to latest version worked. Thanks! My git repo was less than 1 day old :-(
On Wed, Jan 6, 2016 at 4:54 AM Stephan Ewen <se...@apache.org> wrote: > "java.lang.NoSuchMethodError" in Java virtually always means that the > code is compiled against a different version than executed. > > The version in "~/git/flink/" must be slightly outdated. Can you pull the > latest update of the 1.0-SNAPSHOT master and rebuild the code? > > Stephan > > On Tue, Jan 5, 2016 at 9:48 PM, Robert Metzger <rmetz...@apache.org> > wrote: > >> Hi Alex, >> >> How recent is your Flink 1.0-SNAPSHOT build? Maybe the code on the >> (local) cluster (in /git/flink/flink-dist/target/ >> flink-1.0-SNAPSHOT-bin/flink-1.0-SNAPSHOT/) is not up to date? >> >> I just tried it locally, and the job seems to execute: >> >> ./bin/flink run >> /home/robert/Downloads/flink-poc/target/flink-poc-1.0-SNAPSHOT.jar >> org.apache.flink.streaming.api.scala.DataStream@436bc3601/05/2016 >> 21:44:09 Job execution switched to status RUNNING. >> 01/05/2016 21:44:09 Source: Custom Source -> Map(1/1) switched to >> SCHEDULED >> 01/05/2016 21:44:09 Source: Custom Source -> Map(1/1) switched to >> DEPLOYING >> 01/05/2016 21:44:09 Source: Custom Source -> Map(1/1) switched to >> RUNNING >> >> By the way, in order to print the stream, you have to call counts.print() >> instead of print(counts). >> >> >> >> >> >> >> On Tue, Jan 5, 2016 at 9:35 PM, Alex Rovner <alex.rov...@magnetic.com> >> wrote: >> >>> I believe I have set the version uniformly, unless I am overlooking >>> something in the pom. Attaching my project. >>> >>> I have tried building with both "mvn clean package" and "mvn clean >>> package -Pbuild-jar" and I get the same exception. >>> >>> I am running my app with the following command: >>> >>> ~/git/flink/flink-dist/target/flink-1.0-SNAPSHOT-bin/flink-1.0-SNAPSHOT/bin/flink >>> run -c com.magnetic.KafkaWordCount >>> ~/git/flink-poc/target/flink-poc-1.0-SNAPSHOT.jar >>> >>> On Tue, Jan 5, 2016 at 12:47 PM Robert Metzger <rmetz...@apache.org> >>> wrote: >>> >>>> I think the problem is that you only set the version of the Kafka >>>> connector to 1.0-SNAPSHOT, not for the rest of the Flink dependencies. >>>> >>>> On Tue, Jan 5, 2016 at 6:18 PM, Alex Rovner <alex.rov...@magnetic.com> >>>> wrote: >>>> >>>>> Thanks Till for the info. I tried switching to 1.0-SNAPSHOT and now >>>>> facing another error: >>>>> >>>>> Caused by: java.lang.NoSuchMethodError: >>>>> org.apache.flink.streaming.api.operators.StreamingRuntimeContext.isCheckpointingEnabled()Z >>>>> at >>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.run(FlinkKafkaConsumer.java:413) >>>>> at >>>>> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:58) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:55) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:218) >>>>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584) >>>>> at java.lang.Thread.run(Thread.java:745) >>>>> >>>>> >>>>> On Tue, Jan 5, 2016 at 3:54 AM Till Rohrmann <trohrm...@apache.org> >>>>> wrote: >>>>> >>>>>> Hi Alex, >>>>>> >>>>>> this is a bug in the `0.10` release. Is it possible for you to switch >>>>>> to version `1.0-SNAPSHOT`. With this version, the error should no longer >>>>>> occur. >>>>>> >>>>>> Cheers, >>>>>> Till >>>>>> >>>>>> On Tue, Jan 5, 2016 at 1:31 AM, Alex Rovner <alex.rov...@magnetic.com >>>>>> > wrote: >>>>>> >>>>>>> Hello Flinkers! >>>>>>> >>>>>>> The below program produces the following error when running locally. >>>>>>> I am building the program using maven, using 0.10.0 and running in >>>>>>> streaming only local mode "start-local-streaming.sh". I have verified >>>>>>> that >>>>>>> kafka and the topic is working properly by using kafka-console-*.sh >>>>>>> scripts. What am I doing wrong? Any help would be appreciated it. >>>>>>> >>>>>>> Caused by: java.lang.NumberFormatException: For input string: "" >>>>>>> >>>>>>> at >>>>>>> java.lang.NumberFormatException.forInputString(NumberFormatException.java:65) >>>>>>> >>>>>>> at java.lang.Long.parseLong(Long.java:601) >>>>>>> >>>>>>> at java.lang.Long.valueOf(Long.java:803) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.streaming.connectors.kafka.internals.ZookeeperOffsetHandler.getOffsetFromZooKeeper(ZookeeperOffsetHandler.java:125) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.streaming.connectors.kafka.internals.ZookeeperOffsetHandler.seekFetcherToInitialOffsets(ZookeeperOffsetHandler.java:88) >>>>>>> >>>>>>> >>>>>>> def main(args: Array[String]) { >>>>>>> val env = StreamExecutionEnvironment.getExecutionEnvironment >>>>>>> >>>>>>> val properties = new Properties(); >>>>>>> properties.setProperty("bootstrap.servers", "localhost:9092"); >>>>>>> properties.setProperty("zookeeper.connect", "localhost:2181"); >>>>>>> properties.setProperty("group.id", "test"); >>>>>>> >>>>>>> val stream = env >>>>>>> .addSource(new FlinkKafkaConsumer082[String]("topic", new >>>>>>> SimpleStringSchema(), properties)) >>>>>>> >>>>>>> val counts = stream.map(f=>f.split(",")) >>>>>>> >>>>>>> print(counts) >>>>>>> >>>>>>> env.execute() >>>>>>> } >>>>>>> >>>>>>> -- >>>>>>> *Alex Rovner* >>>>>>> *Director, Data Engineering * >>>>>>> *o:* 646.759.0052 >>>>>>> >>>>>>> * <http://www.magnetic.com/>* >>>>>>> >>>>>>> >>>>>> -- >>>>> *Alex Rovner* >>>>> *Director, Data Engineering * >>>>> *o:* 646.759.0052 >>>>> >>>>> * <http://www.magnetic.com/>* >>>>> >>>>> >>>> -- >>> *Alex Rovner* >>> *Director, Data Engineering * >>> *o:* 646.759.0052 >>> >>> * <http://www.magnetic.com/>* >>> >>> >> > -- *Alex Rovner* *Director, Data Engineering * *o:* 646.759.0052 * <http://www.magnetic.com/>*