Nothing that uses MapReduce as an underlying execution engine creates a single
file when running multiple reducers because MapReduce doesn't. The real
question is if you want to keep the file on Hadoop, why worry about whether
it's a single file? Most applications on Hadoop will take a directory as an
input and read all the files contained in it.
Alan.
On May 24, 2013, at 12:11 PM, Mix Nin wrote:
> STORE command produces multiple output files. I want a single output file
> and I tried using command as below
>
> STORE (foreach (group NoNullData all) generate flatten($1)) into 'xxxx';
>
> This command produces one single file but at the same time forces to use
> single reducer which kills performance.
>
> How do I overcome the scenario?
>
> Normally STORE command produces multiple output files, apart from that I
> see another file
> "_SUCCESS" in output directory. I ma generating metadata file ( using
> PigStorage('\t', '-schema') ) in output directory
>
> I thought of using getmerge as follows
>
> *hadoop* fs -*getmerge* <dir_of_input_files> <local file>
>
> But this requires
> 1)eliminating files other than data files in HDFS directory
> 2)It creates a single file in local directory but not in HDFS directory
> 3)I need to again move file from local directory to HDFS directory which
> may take additional time , depending on size of single file
> 4)I need to agin place the files which I eliminated in Step 1
>
>
> Is there an efficient way for my problem?
>
> Thanks