gotta start by looking at the logs and run the local client to eliminate
HS2. perhaps running hive as such:
$ hive -hiveconf hive.root.logger=DEBUG,console
do you see any smoking gun?
On Wed, Sep 28, 2016 at 7:34 AM, Jose Rozanec wrote:
> Hi,
>
> We have a Hive cluster (Hive 2.1.0+Tez 0.8.4)
Raymond - you were the closest.
Parquet field names contained '::' ex. bag1::user_name
Hope it will help anyone in the future
Thanks for all your help
Tor
On Sun, Aug 17, 2014 at 7:50 PM, Raymond Lau
wrote:
> Do your field names in your parquet files contain upper case letters by
> any chan
Do your field names in your parquet files contain upper case letters by any
chance ex. userName? Hive will not read the data of external tables if
they are not completely lower case field names, it doesn't convert them
properly in the case of external tables.
On Aug 17, 2014 8:00 AM, "hadoop hive"
Take a small set of data like 2-5 line and insert it...
After that you can try insert first 10 column and then next 10 till you
fund your problematic column
On Aug 17, 2014 8:37 PM, "Tor Ivry" wrote:
> Is there any way to debug this?
>
> We are talking about many fields here.
> How can I see whi
Is there any way to debug this?
We are talking about many fields here.
How can I see which field has the mismatch?
On Sun, Aug 17, 2014 at 4:30 PM, hadoop hive wrote:
> Hi,
>
> You check the data type you have provided while creating external table,
> it should match with data in files.
>
> T
Hi,
You check the data type you have provided while creating external table, it
should match with data in files.
Thanks
Vikas Srivastava
On Aug 17, 2014 7:07 PM, "Tor Ivry" wrote:
> Hi
>
>
>
> I have a hive (0.11) table with the following create syntax:
>
>
>
> CREATE EXTERNAL TABLE events(
>
>
Any ideas regarding this ?
For now, i have resolved this issue by putting the amazon credentials into
the Cloudera Manager Hive service safety valve and deploying the new client
configs to the hive gateway nodes.
But this restricts me to using only one amazon account for the Hive
operations.
- H
e excuse typos
> --
> *From: * Cyril Bogus
> *Date: *Mon, 25 Feb 2013 10:34:29 -0500
> *To: *
> *ReplyTo: * user@hive.apache.org
> *Subject: *Re: Hive queries
>
> I do not get any errors.
> It is only when I run hive and try to query the tables
ote device, Please excuse typos
-Original Message-
From: Cyril Bogus
Date: Mon, 25 Feb 2013 10:34:29
To:
Reply-To: user@hive.apache.org
Subject: Re: Hive queries
I do not get any errors.
It is only when I run hive and try to query the tables I imported. Let's
say I want to only get num
I do not get any errors.
It is only when I run hive and try to query the tables I imported. Let's
say I want to only get numeric tuples for a given table. I cannot find the
table (show tables; is empty) unless I go in the hive home folder and run
hive again. I would expect the state of hive to be t
any errors you see ?
On Mon, Feb 25, 2013 at 8:48 PM, Cyril Bogus wrote:
> Hi everyone,
>
> My setup is Hadoop 1.0.4, Hive 0.9.0, Sqoop 1.4.2-hadoop 1.0.0
> Mahout 0.7
>
> I have imported tables from a remote database directly into Hive using
> Sqoop.
>
> Somehow when I try to run Sqoop from Ha
Hi Manish,
If you have data on your local file system,
You can also do something like following from your local file system, without
doing put or copyFromLocal.
$ hive -e "load data local inpath 'path on local file system' into table ;"
Thanks,
Anurag Tangri
Sent from my iPhone
On Feb 16, 201
Dude sorry for the off topic, but having a rocketmail account is
awesome. I wish I still had mine.
On Sat, Feb 16, 2013 at 9:16 PM, manishbh...@rocketmail.com
wrote:
>
> When you want to move data from external system to hive, this means moving
> data to HDFS first and then point the Hive table t
When you want to move data from external system to hive, this means moving data
to HDFS first and then point the Hive table to the file in HDFS where you have
exported the data.
So, you have couple of commands like -copyFromLocal and fget which move the
file to hdfs. If you intent to move in re
[-user@hive, +user@sqoop]
Hi Cyrille,
this seems to me more a Sqoop issue than Hive issue, so I've moved this email
to user@sqoop mailing list. I'm keeping user@hive in Bcc so that the mailing
list will get the memo. Please join the user@sqoop mailing list [1] to receive
additional feedback.
>
Hi Jarcec,
I did try Sqoop. I am running sqoop 1.4.2 --hadoop1.0.0 along with hadoop
1.0.4 But I keep running on the following exception.
Exception in thread "main" java.lang.IncompatibleClassChangeError: Found
class org.apache.hadoop.mapreduce.JobContext, but interface was expected
So I wrote a
Hi Cyrille,
I'm not exactly sure what exactly you mean, so I'm more or less blindly
shooting, but maybe Apache Sqoop [1] might help you?
Jarcec
Links:
1: http://sqoop.apache.org/
On Fri, Feb 15, 2013 at 01:44:45PM -0500, Cyrille Djoko wrote:
> I am looking for a relatively efficient way of tran
ormance of "group by" queries. Are you
referring to the use of map side aggregation? Any resources you can point me to
where I can study this further?
Thanks
Anand
From: Bejoy Ks [mailto:bejoy...@yahoo.com]
Sent: Sunday, April 01, 2012 5:35 PM
To: user@hive.apache.org
Subject: Re: H
Anand
You can optimize pretty much all hive queries. Based on your queries you
need to do the optimizations. For example Group By has some specific way to be
optimized. Some times Distribute By comes in handy for optimizing some queries.
Skew joins are good to balace the reducer loads. etc
Anand,
best place to understand the join queries on hive is from the presentation
by Namit Jain from Facebook.
Here is the pdf
https://cwiki.apache.org/Hive/presentations.data/Hive%20Summit%202011-join.pdf
you can search the video on youtube. Its very well described
On Sun, Apr 1, 2012 at 11:59
Sent: Saturday, October 1, 2011 4:21 AM
Subject: Re: Hive queries not retrieving any values from LZO compressed tables
Hello Bejoy -
Even though you defined your target table by specifying the input format, you
still have to ensure the output of your population query is compressed. Make
sure the
Hello Bejoy -
Even though you defined your target table by specifying the input format,
you still have to ensure the output of your population query is compressed.
Make sure the following are set before executing your "insert overwrite
table " query.
set mapred.output.compress=true;
set mapred.o
Sorry i should've given more details.
The query was limited by a partition range; I just omitted the WHERE
clause in the mail.
The table is not that big. For each day, there is one gzipped file.
The largest file is about 250MB (close to 2GB uncompressed).
I did intend to count and that was just to
Hey Vijay,
You can go to the mapred ui, normally it runs on port 50030 of the namenode
and see how many map jobs got created for your submitted query.
You said that the events table has daily partitions but the example query
that you have does not prune the partitions by specifying a WHERE clause
24 matches
Mail list logo