Thanks Marcin, I worked ....I uncompressed file and then loaded file in
hive table.

Now its been quick, few mins.




Regards
Sanjiv Singh
Mob :  +091 9990-447-339

On Wed, Jun 22, 2016 at 7:44 AM, Jörn Franke <jornfra...@gmail.com> wrote:

>
>
> Marcin is correct : either split up the gzip files in smaller files of at
> least on HDFS block or use bzip2 with block compression.
> What is the original format of the table?
>
> On 22 Jun 2016, at 01:50, Marcin Tustin <mtus...@handybook.com> wrote:
>
> This is because a GZ file is not splittable at all. Basically, try
> creating this from an uncompressed file, or even better split up the file
> and put the files in a directory in hdfs/s3/whatever.
>
> On Tue, Jun 21, 2016 at 7:45 PM, @Sanjiv Singh <sanjiv.is...@gmail.com>
> wrote:
>
>> Hi ,
>>
>> I have big compressed data file *my_table.dat.gz* ( approx size 100 GB)
>>
>> # load staging table *STAGE_**my_table* from file *my_table.dat.gz*
>>
>> HIVE>> LOAD DATA  INPATH '/var/lib/txt/*my_table.dat.gz*' OVERWRITE INTO
>> TABLE STAGE_my_table ;
>>
>> *# insert into ORC table "my_table"*
>>
>> HIVE>> INSERT INTO TABLE my_table SELECT * FROM TXT_my_table;
>> ....
>> INFO  : Map 1: 0(+1)/1  Reducer 2: 0/1
>> ....
>>
>>
>> Insertion into orc table in going on since 5-6 hours , Seems everything
>> is going sequential with one mapper reading complete file?
>>
>> Please suggest ? help me in improving ORC table load.
>>
>>
>>
>>
>> Regards
>> Sanjiv Singh
>> Mob :  +091 9990-447-339
>>
>
>
> Want to work at Handy? Check out our culture deck and open roles
> <http://www.handy.com/careers>
> Latest news <http://www.handy.com/press> at Handy
> Handy just raised $50m
> <http://venturebeat.com/2015/11/02/on-demand-home-service-handy-raises-50m-in-round-led-by-fidelity/>
>  led
> by Fidelity
>
>

Reply via email to