Oleg, if you specifies the directory mapreduce will take all files.

Regards,
Zacarias



On Tue, Mar 23, 2010 at 3:08 AM, Oleg Ruchovets <oruchov...@gmail.com> wrote:
> Hi ,
> All examples that I found executes mapreduce job on a single file but in my
> situation I have more than one.
>
> Suppose I have such folder on HDFS which contains some files:
>
>    /my_hadoop_hdfs/my_folder:
>                /my_hadoop_hdfs/my_folder/file1.txt
>                /my_hadoop_hdfs/my_folder/file2.txt
>                /my_hadoop_hdfs/my_folder/file3.txt
>
>
> how can I execute  hadoop mapreduce on file1.txt , file2.txt and file3.txt?
>
> Is it possible to provide to hadoop job folder as parameter and all files
> will be produced by mapreduce job?
>
> Thanks In Advance
> Oleg.
>

Reply via email to