Hi Keith, That's correct. Any files that are in a directory that's referenced by Hive are assumed to be in the same table.
You'll need to make a directory for each table and then load each CSV file into the appropriate directory. Matt Tucker -----Original Message----- From: Keith Wiley [mailto:kwi...@keithwiley.com] Sent: Monday, March 12, 2012 3:47 PM To: user@hive.apache.org Subject: External table point of clarification It appears as if the syntax for creating an external table takes a column description and a directory where the table (a .csv file) should be found. Notably, it doesn't take the path to the file, rather it takes the path to the directory containing the file. ...which means that if I have multiple .csv files, each a unique table with a unique column description, it is not sufficient to put them in one hdfs directory, since the columns must be described when creating the hive table. Rather, each .csv file conforming to a unique column description must be placed in its own separate hdfs directory, right? Just clarifying...is this correct? Can I not put a set of .csv files comprising multiple tables of a heterogenous database in a single directory and then have hive build a set of tables around those files? Thank you. ________________________________________________________________________________ Keith Wiley kwi...@keithwiley.com keithwiley.com music.keithwiley.com "It's a fine line between meticulous and obsessive-compulsive and a slippery rope between obsessive-compulsive and debilitatingly slow." -- Keith Wiley ________________________________________________________________________________