Hi Keith,

That's correct.  Any files that are in a directory that's referenced by Hive 
are assumed to be in the same table.

You'll need to make a directory for each table and then load each CSV file into 
the appropriate directory.

Matt Tucker


-----Original Message-----
From: Keith Wiley [mailto:kwi...@keithwiley.com] 
Sent: Monday, March 12, 2012 3:47 PM
To: user@hive.apache.org
Subject: External table point of clarification

It appears as if the syntax for creating an external table takes a column 
description and a directory where the table (a .csv file) should be found.  
Notably, it doesn't take the path to the file, rather it takes the path to the 
directory containing the file.

...which means that if I have multiple .csv files, each a unique table with a 
unique column description, it is not sufficient to put them in one hdfs 
directory, since the columns must be described when creating the hive table.  
Rather, each .csv file conforming to a unique column description must be placed 
in its own separate hdfs directory, right?


Just clarifying...is this correct?  Can I not put a set of .csv files 
comprising multiple tables of a heterogenous database in a single directory and 
then have hive build a set of tables around those files?

Thank you.

________________________________________________________________________________
Keith Wiley     kwi...@keithwiley.com     keithwiley.com    music.keithwiley.com

"It's a fine line between meticulous and obsessive-compulsive and a slippery 
rope between obsessive-compulsive and debilitatingly slow."
                                           --  Keith Wiley 
________________________________________________________________________________

Reply via email to