Manish,
Just specify the directory under location and not the actual file. In
your case (notice, there is no .csv in the location). If your field
definitions match the data already present in HDFS, this should work just
fine. Make sure all the files under directory input have the same structure.
There is nothing wrong with your SQL statement. This works on CLI fine and
I tried the following. Your issue seems to be related to the underlying
metadata store.
CREATE TABLE page_view(viewTime INT, userid BIGINT,
page_url STRING, referrer_url STRING,
ip STRING COMMENT 'IP Address of the
helps
Nanda Vijaydev
On Mon, May 28, 2012 at 3:28 PM, Igor Tatarinov wrote:
> Try replacing the comma with JOIN
>
> igor
> decide.com
>
> On Mon, May 28, 2012 at 6:48 AM, shan s wrote:
>
>> I need help with a simple subquery. Given below data, I need counts and
>&g
Can you paste a sample line of your data on HDFS and which column you are
trying to query?
Thanks
Nanda Vijaydev
On Mon, May 14, 2012 at 11:40 AM, Sukhendu Chakraborty <
sukhendu.chakrabo...@gmail.com> wrote:
> Are NaNs and/or Infinity supported in HIVE? If yes, I wanted to know
>
Hadoop in general does well with fewer large data files instead of more
smaller data files. RDBMS type of indexing and run time optimization is not
exactly available in Hadoop/Hive yet. So one suggestion is to combine some
of this data, if you can, into fewer tables as you are doing sqoop. Even if
In hive, the raw data is in HDFS and there is a metadata layer that defines
the structure of the raw data. Table is usually a reference to metadata,
probably in a mySQL server and it contains a reference to the location of
the data in HDFS, type of delimiter or serde to use and so on.
1. With hive