Hi shaik

You have given the table location as '/user/hive/warehouse' . When you load 
data into that table and if you have used overwrite, first the directory is 
deleted and then created and new file is loaded.

So here your hive warehouse dir is same as your table location. So when you 
used Load data with overwrite all the table's dirs also
Got deleted.

You can recover the data only if trash is enabled in hdfs.

 
Regards
Bejoy KS

Sent from handheld, please excuse typos.

-----Original Message-----
From: shaik ahamed <shaik5...@gmail.com>
Date: Thu, 26 Jul 2012 19:39:01 
To: <user@hive.apache.org>
Reply-To: user@hive.apache.org
Subject: unable to see the file

Hi Users,

Before creating the table i enable the below cmds


*set hive.exec.compress.output=true;
set io.seqfile.compression.type=BLOCK;*

             As i created an external table with the below syntax


              *CREATE EXTERNAL TABLE test_data(vender string,supplier
string,order_date string,quantity int) row format delimited fields
terminated by ' ' stored as textfile *
*LOCATION '/user/hive/warehouse';*

After that loaded data into test_data .

I am not able to see the table in the path(HDFS path) which i given before
creating it ..  '/user/hive/warehouse' and the other table which are
already exists that also im not able to see.

please reply me with the above concern and tell me how i can see the table .

how can see get back the old exiting hdfs files

Regards
shaik

Reply via email to