Any pointers would appreciate.

On Mon, Dec 1, 2014 at 11:27 AM, unmesha sreeveni <unmeshab...@gmail.com>
wrote:

>
> On Mon, Dec 1, 2014 at 11:15 AM, yogendra reddy <yogendra...@gmail.com>
> wrote:
>
>> hive --orcfiledump
>
>
> ​Hi yogendra​
>
> ​shows ​
> Exception in thread "main" java.io.IOException: Malformed ORC file
> /employeeData/empLargenew.txt. Invalid postscript.
> ​But ​my file is not ORC format it is .csv format
>
> *1,Anne,Admin,50000,A*
> *2,Gokul,Admin,50000,B*
>
> So as a workaround I loaded data into a table
>
> * create external table stagingMB (EmployeeID Int,FirstName
> String,Designation String,Salary Int,Department String) row format
> delimited fields terminated by "," location '/employeeData';*
>
> and from the above table I loaded the data into ORC table
>
>  *create table HiveMB (EmployeeID Int,FirstName String,Designation
> String,Salary Int,Department String) clustered by (Department) into 3
> buckets stored as orc TBLPROPERTIES ('transactional'='true') ;  *
>
> * from stagingMB insert into table HiveMB  select
> employeeid,firstname,designation,salary,department;  *
>
>
> --
> *Thanks & Regards *
>
>
> *Unmesha Sreeveni U.B*
> *Hadoop, Bigdata Developer*
> *Centre for Cyber Security | Amrita Vishwa Vidyapeetham*
> http://www.unmeshasreeveni.blogspot.in/
>
>
>


-- 
*Thanks & Regards *


*Unmesha Sreeveni U.B*
*Hadoop, Bigdata Developer*
*Centre for Cyber Security | Amrita Vishwa Vidyapeetham*
http://www.unmeshasreeveni.blogspot.in/

Reply via email to