Any pointers would appreciate. On Mon, Dec 1, 2014 at 11:27 AM, unmesha sreeveni <unmeshab...@gmail.com> wrote:
> > On Mon, Dec 1, 2014 at 11:15 AM, yogendra reddy <yogendra...@gmail.com> > wrote: > >> hive --orcfiledump > > > Hi yogendra > > shows > Exception in thread "main" java.io.IOException: Malformed ORC file > /employeeData/empLargenew.txt. Invalid postscript. > But my file is not ORC format it is .csv format > > *1,Anne,Admin,50000,A* > *2,Gokul,Admin,50000,B* > > So as a workaround I loaded data into a table > > * create external table stagingMB (EmployeeID Int,FirstName > String,Designation String,Salary Int,Department String) row format > delimited fields terminated by "," location '/employeeData';* > > and from the above table I loaded the data into ORC table > > *create table HiveMB (EmployeeID Int,FirstName String,Designation > String,Salary Int,Department String) clustered by (Department) into 3 > buckets stored as orc TBLPROPERTIES ('transactional'='true') ; * > > * from stagingMB insert into table HiveMB select > employeeid,firstname,designation,salary,department; * > > > -- > *Thanks & Regards * > > > *Unmesha Sreeveni U.B* > *Hadoop, Bigdata Developer* > *Centre for Cyber Security | Amrita Vishwa Vidyapeetham* > http://www.unmeshasreeveni.blogspot.in/ > > > -- *Thanks & Regards * *Unmesha Sreeveni U.B* *Hadoop, Bigdata Developer* *Centre for Cyber Security | Amrita Vishwa Vidyapeetham* http://www.unmeshasreeveni.blogspot.in/