Updating to latest version worked. Thanks! My git repo was less than 1 day
old :-(

On Wed, Jan 6, 2016 at 4:54 AM Stephan Ewen <se...@apache.org> wrote:

> "java.lang.NoSuchMethodError" in Java virtually always means that the
> code is compiled against a different version than executed.
>
> The version in "~/git/flink/" must be slightly outdated. Can you pull the
> latest update of the 1.0-SNAPSHOT master and rebuild the code?
>
> Stephan
>
> On Tue, Jan 5, 2016 at 9:48 PM, Robert Metzger <rmetz...@apache.org>
> wrote:
>
>> Hi Alex,
>>
>> How recent is your Flink 1.0-SNAPSHOT build? Maybe the code on the
>> (local) cluster (in /git/flink/flink-dist/target/
>> flink-1.0-SNAPSHOT-bin/flink-1.0-SNAPSHOT/) is not up to date?
>>
>> I just tried it locally, and the job seems to execute:
>>
>> ./bin/flink run
>> /home/robert/Downloads/flink-poc/target/flink-poc-1.0-SNAPSHOT.jar
>> org.apache.flink.streaming.api.scala.DataStream@436bc3601/05/2016
>> 21:44:09 Job execution switched to status RUNNING.
>> 01/05/2016 21:44:09 Source: Custom Source -> Map(1/1) switched to
>> SCHEDULED
>> 01/05/2016 21:44:09 Source: Custom Source -> Map(1/1) switched to
>> DEPLOYING
>> 01/05/2016 21:44:09 Source: Custom Source -> Map(1/1) switched to
>> RUNNING
>>
>> By the way, in order to print the stream, you have to call counts.print()
>> instead of print(counts).
>>
>>
>>
>>
>>
>>
>> On Tue, Jan 5, 2016 at 9:35 PM, Alex Rovner <alex.rov...@magnetic.com>
>> wrote:
>>
>>> I believe I have set the version uniformly, unless I am overlooking
>>> something in the pom. Attaching my project.
>>>
>>> I have tried building with both "mvn clean package" and "mvn clean
>>> package -Pbuild-jar" and I get the same exception.
>>>
>>> I am running my app with the following command:
>>>
>>> ~/git/flink/flink-dist/target/flink-1.0-SNAPSHOT-bin/flink-1.0-SNAPSHOT/bin/flink
>>> run -c com.magnetic.KafkaWordCount
>>> ~/git/flink-poc/target/flink-poc-1.0-SNAPSHOT.jar
>>>
>>> On Tue, Jan 5, 2016 at 12:47 PM Robert Metzger <rmetz...@apache.org>
>>> wrote:
>>>
>>>> I think the problem is that you only set the version of the Kafka
>>>> connector to 1.0-SNAPSHOT, not for the rest of the Flink dependencies.
>>>>
>>>> On Tue, Jan 5, 2016 at 6:18 PM, Alex Rovner <alex.rov...@magnetic.com>
>>>> wrote:
>>>>
>>>>> Thanks Till for the info. I tried switching to 1.0-SNAPSHOT and now
>>>>> facing another error:
>>>>>
>>>>> Caused by: java.lang.NoSuchMethodError:
>>>>> org.apache.flink.streaming.api.operators.StreamingRuntimeContext.isCheckpointingEnabled()Z
>>>>> at
>>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.run(FlinkKafkaConsumer.java:413)
>>>>> at
>>>>> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:58)
>>>>> at
>>>>> org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:55)
>>>>> at
>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:218)
>>>>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)
>>>>> at java.lang.Thread.run(Thread.java:745)
>>>>>
>>>>>
>>>>> On Tue, Jan 5, 2016 at 3:54 AM Till Rohrmann <trohrm...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Hi Alex,
>>>>>>
>>>>>> this is a bug in the `0.10` release. Is it possible for you to switch
>>>>>> to version `1.0-SNAPSHOT`. With this version, the error should no longer
>>>>>> occur.
>>>>>>
>>>>>> Cheers,
>>>>>> Till
>>>>>>
>>>>>> On Tue, Jan 5, 2016 at 1:31 AM, Alex Rovner <alex.rov...@magnetic.com
>>>>>> > wrote:
>>>>>>
>>>>>>> Hello Flinkers!
>>>>>>>
>>>>>>> The below program produces the following error when running locally.
>>>>>>> I am building the program using maven, using 0.10.0 and running in
>>>>>>> streaming only local mode "start-local-streaming.sh".  I have verified 
>>>>>>> that
>>>>>>> kafka and the topic is working properly by using kafka-console-*.sh
>>>>>>> scripts. What am I doing wrong? Any help would be appreciated it.
>>>>>>>
>>>>>>> Caused by: java.lang.NumberFormatException: For input string: ""
>>>>>>>
>>>>>>> at
>>>>>>> java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
>>>>>>>
>>>>>>> at java.lang.Long.parseLong(Long.java:601)
>>>>>>>
>>>>>>> at java.lang.Long.valueOf(Long.java:803)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.streaming.connectors.kafka.internals.ZookeeperOffsetHandler.getOffsetFromZooKeeper(ZookeeperOffsetHandler.java:125)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.streaming.connectors.kafka.internals.ZookeeperOffsetHandler.seekFetcherToInitialOffsets(ZookeeperOffsetHandler.java:88)
>>>>>>>
>>>>>>>
>>>>>>> def main(args: Array[String]) {
>>>>>>>   val env = StreamExecutionEnvironment.getExecutionEnvironment
>>>>>>>
>>>>>>>   val properties = new Properties();
>>>>>>>   properties.setProperty("bootstrap.servers", "localhost:9092");
>>>>>>>   properties.setProperty("zookeeper.connect", "localhost:2181");
>>>>>>>   properties.setProperty("group.id", "test");
>>>>>>>
>>>>>>>   val stream = env
>>>>>>>     .addSource(new FlinkKafkaConsumer082[String]("topic", new 
>>>>>>> SimpleStringSchema(), properties))
>>>>>>>
>>>>>>>   val counts = stream.map(f=>f.split(","))
>>>>>>>
>>>>>>>   print(counts)
>>>>>>>
>>>>>>>   env.execute()
>>>>>>> }
>>>>>>>
>>>>>>> --
>>>>>>> *Alex Rovner*
>>>>>>> *Director, Data Engineering *
>>>>>>> *o:* 646.759.0052
>>>>>>>
>>>>>>> * <http://www.magnetic.com/>*
>>>>>>>
>>>>>>>
>>>>>> --
>>>>> *Alex Rovner*
>>>>> *Director, Data Engineering *
>>>>> *o:* 646.759.0052
>>>>>
>>>>> * <http://www.magnetic.com/>*
>>>>>
>>>>>
>>>> --
>>> *Alex Rovner*
>>> *Director, Data Engineering *
>>> *o:* 646.759.0052
>>>
>>> * <http://www.magnetic.com/>*
>>>
>>>
>>
> --
*Alex Rovner*
*Director, Data Engineering *
*o:* 646.759.0052

* <http://www.magnetic.com/>*

Reply via email to