Thanks a lot Marton and Max, it worked perfectly. Regards from Italy :)
On Tue, Feb 24, 2015 at 11:31 AM, Max Michels <m...@apache.org> wrote: > Hi Giacomo, > > Congratulations on setting up a Flink cluster with HDFS :) To run the > WordCount example provided with Flink, you should first upload your > input file to HDFS. If you have not done so, please run > > > hdfs dfs -put -p file:///home/user/yourinputfile hdfs:///wc_input > > Then, you can use the Flink command-line tool to submit the WordCount job. > > > ./bin/flink run -v examples/flink-java-examples-*-WordCount.jar > hdfs:///wc_input hdfs:///wc_output > > > This should work if you configured HDFS correctly. If you haven't set > the default hdfs name (fs.default.name), you might have to use the > full HDFS URL. For example, if your namenode's address is > namenode.example.com at port 7777, then use > hdfs://namenode.example.com:7777/wc_input. > > > Kind regards, > Max > > On Tue, Feb 24, 2015 at 11:13 AM, Giacomo Licari > <giacomo.lic...@gmail.com> wrote: > > Hi guys, > > I'm Giacomo from Italy, I'm newbie with Flink. > > > > I setted up a cluster with Hadoop 1.2 and Flink. > > > > I would like to ask to you how to run the WordCount example taking the > input > > file from hdfs (example myuser/testWordCount/hamlet. > > txt) and put the output also inside hdfs (example > > myuser/testWordCount/output.txt). > > > > I successfully run the example on my local filesystem, I would like to > test > > it with HDSF. > > > > Thanks a lot guys, > > Giacomo >