Oleg, if you specifies the directory mapreduce will take all files. Regards, Zacarias
On Tue, Mar 23, 2010 at 3:08 AM, Oleg Ruchovets <oruchov...@gmail.com> wrote: > Hi , > All examples that I found executes mapreduce job on a single file but in my > situation I have more than one. > > Suppose I have such folder on HDFS which contains some files: > > /my_hadoop_hdfs/my_folder: > /my_hadoop_hdfs/my_folder/file1.txt > /my_hadoop_hdfs/my_folder/file2.txt > /my_hadoop_hdfs/my_folder/file3.txt > > > how can I execute hadoop mapreduce on file1.txt , file2.txt and file3.txt? > > Is it possible to provide to hadoop job folder as parameter and all files > will be produced by mapreduce job? > > Thanks In Advance > Oleg. >