Change the query to something like-

INSERT OVERWRITE DIRECTORY '/outputable.txt'
select concat(col1, ',', col2, ',', col3)  from myoutputtable;

That way columns will be separated by ,.

Thanks,
Vinod

On Tue, Aug 7, 2012 at 10:16 AM, zuohua zhang <zuo...@gmail.com> wrote:

> I used the following that it won't help?
>
> ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
>
> On Mon, Aug 6, 2012 at 9:43 PM, Vinod Singh <vi...@vinodsingh.com> wrote:
>
>> Columns of a Hive table are separated by ^A character. Instead of doing a
>> "SELECT * ", you may like to use concat function to have a separator of
>> your choice.
>>
>> Thanks,
>> Vinod
>>
>>
>> On Tue, Aug 7, 2012 at 9:39 AM, zuohua zhang <zuo...@gmail.com> wrote:
>>
>>> I have used the following to output a hive table to a file:
>>> DROP TABLE IF EXISTS myoutputable;
>>> CREATE TABLE myoutputtable
>>> ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
>>> STORED AS TEXTFILE
>>> AS
>>> select
>>> *
>>> from originaltable;
>>> INSERT OVERWRITE DIRECTORY '/outputable.txt'
>>> select * from myoutputtable;
>>>
>>> then i used
>>> hadoop dfs -getmerge /outputtable.txt /mnt/
>>>
>>> but the /mnt/outputtable.txt file shows strange characters ^A in the
>>> file. What did I do wrong?
>>>
>>
>>
>

Reply via email to