Hi ,
All examples that I found executes mapreduce job on a single file but in my
situation I have more than one.
Suppose I have such folder on HDFS which contains some files:
/my_hadoop_hdfs/my_folder:
/my_hadoop_hdfs/my_folder/file1.txt
/my_hadoop_hdfs/my_folder/file2.txt
/my_hadoop_hdfs/my_folder/file3.txt
how can I execute hadoop mapreduce on file1.txt , file2.txt and file3.txt?
Is it possible to provide to hadoop job folder as parameter and all files
will be produced by mapreduce job?
Thanks In Advance
Oleg.