Re: Tez : Anyway to avoid creating subdirectories by "Insert with union all² ?

2015-08-19 Thread Gopal Vijayaraghavan
> Is there anyway to avoid creating sub-directories? Or this is by design >and can not be changed? This is because of the way file-formats generate hadoop name files without collisions. For instance, any change to that would break Parquet-MR for Tez. That's why we generate a compatible, but colli

Tez : Anyway to avoid creating subdirectories by "Insert with union all” ?

2015-08-19 Thread Jim Green
Hi Team, Below insert with union-all will create sub-directories: set hive.execution.engine=tez; create table h1_passwords_target like h1_passwords; insert overwrite table h1_passwords_target select * from (select * from h1_passwords limit 1 union all select * from h1_passwords limit 2 ) sub

Newly added columns couldn't be correctly mapped to the fields of Parquet files

2015-08-19 Thread Frank Dai
I have several Parquet files on Amazon S3, which have three fields, username STRING, request STRING and timestamp STRING. If I create a table in Hive and add a partition, everything is fine. CREATE EXTERNAL TABLE IF NOT EXISTS access_logs(username STRING, request STRING, `timestamp` STRING) PARTI

Re: hiveserver2 hangs

2015-08-19 Thread Sanjeev Verma
Can somebody gives me some pointer to looked upon? On Wed, Aug 19, 2015 at 9:26 AM, Sanjeev Verma wrote: > Hi > We are experiencing a strange problem with the hiveserver2, in one of the > job it gets the GC limit exceed from mapred task and hangs even having > enough heap available.we are not ab

Re: Hive query over JDBC not honoring fetch size

2015-08-19 Thread Emil Berglind
Also, I tried setting the "hive.fetch.task.conversion" property in the JDBC URL, like so: jdbc:hive2:// 192.168.132.128:1/default?hive.fetch.task.conversion=none, but it is still creating mapreduce tasks for the query, so it effectively seems to be ignoring that property. On Wed, Aug 19, 2015

hiveserver2 hangs

2015-08-19 Thread Sanjeev Verma
Hi We are experiencing a strange problem with the hiveserver2, in one of the job it gets the GC limit exceed from mapred task and hangs even having enough heap available.we are not able to identify what causing this issue. Could anybody help me identify the issue and let me know what pointers I nee

Re: HiveServer2 & Kerberos

2015-08-19 Thread Jary Du
Correct me if I am wrong, my understanding is that after using kerberos authentication, you probably don’t need the password. Hope it helps Thanks, Jary > On Aug 19, 2015, at 9:09 AM, Loïc Chanel wrote: > > By the way, thanks a lot for your help, because your solution works, but I'm > still

Re: Hive query over JDBC not honoring fetch size

2015-08-19 Thread Emil Berglind
When I run the "SELECT * FROM " query it is running it as a mapreduce job. I can see it in the Yarn Manager and also in the Tez UI. This is also when the fetch size is not honored and it tries to basically return all results at once. Is there a way to make this work? On Wed, Aug 19, 2015 at 10:53

Re: HiveServer2 & Kerberos

2015-08-19 Thread Loïc Chanel
By the way, thanks a lot for your help, because your solution works, but I'm still interested in knowing what is the password I did not enter. Thanks again, Loïc Loïc CHANEL Engineering student at TELECOM Nancy Trainee at Worldline - Villeurbanne 2015-08-19 18:07 GMT+02:00 Loïc Chanel : > All

Re: HiveServer2 & Kerberos

2015-08-19 Thread Loïc Chanel
All right, but then, what is the password hive asks for ? Hive's one ? How do I know its value ? Loïc CHANEL Engineering student at TELECOM Nancy Trainee at Worldline - Villeurbanne 2015-08-19 17:51 GMT+02:00 Jary Du : > For Beeline connection string, it should be "!connect > jdbc:hive2://:/;pri

Re: HiveServer2 & Kerberos

2015-08-19 Thread Jary Du
For Beeline connection string, it should be "!connect jdbc:hive2://:/;principal=”. Please make sure it is the hive’s principal, not the user’s. And when you kinit, it should be kinit user’s keytab, not the hive’s keytab. > On Aug 19, 2015, at 8:46 AM, Loïc Chanel wrote: > > Yeah, I forgo

Re: HiveServer2 & Kerberos

2015-08-19 Thread Loïc Chanel
Yeah, I forgot to mention it, but each time I did a kinit user/hive before launching beeline, as I read somewhere that Beeline does not handle Kerberos connection. So, as I can make klist before launching beeline and having a good result, the problem does not come from this. Thanks a lot for your

Re: HiveServer2 & Kerberos

2015-08-19 Thread Jary Du
"The Beeline client must have a valid Kerberos ticket in the ticket cache before attempting to connect." (http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1.3/bk_dataintegration/content/ch_using-hive-clients-examples.html) So you need kinit first to have the valid Kerberos ticket int the tick

Re: HiveServer2 & Kerberos

2015-08-19 Thread Loïc Chanel
Hi again, As I searched another way to make some requests with Kerberos enabled for security on HiveServer, I found that this request should do the same : !connect jdbc:hive2:// 192.168.6.210:1/default;principal=user/h...@westeros.wl org.apache.hive.jdbc.HiveDriver But now I've got another err

Re: Hive query over JDBC not honoring fetch size

2015-08-19 Thread Prem Yadav
actually it should be something like getHandleIdentifier()=hfhkjhfjhkjfh-dsdsad-sdsd--dsada: fetchResults() On Wed, Aug 19, 2015 at 3:49 PM, Prem Yadav wrote: > Hi Emil, > for either of the queries, there will be no mapreduce job. the query > engine understands that in both case, it need not do

Re: Hive query over JDBC not honoring fetch size

2015-08-19 Thread Prem Yadav
Hi Emil, for either of the queries, there will be no mapreduce job. the query engine understands that in both case, it need not do any computation and just needs to fetch all the data from the files. The fetch size should be honored in both cases. Hope you are using hiveserver2. You can try connec

Hive query over JDBC not honoring fetch size

2015-08-19 Thread Emil Berglind
I have a small Java app that I wrote that uses JDBC to run a hive query. The Hive table that I'm running it against has 30+ million rows, and I want to pull them all back to verify the data. If I run a simple "SELECT * FROM " and set a fetch size of 30,000 then the fetch size is not honored and it

Re: Question about PredicateTransitivePropagate

2015-08-19 Thread 孙若曦
I saw the same result. But I debugged a little bit and figured out that it was the PPD optimizer did the transitivity propagation, not PredicateTransitivePropagate. -- Original -- From: "Gopal Vijayaraghavan"; Date: Wed, Aug 19, 2015 02:52 PM To: "user";

HiveServer2 & Kerberos

2015-08-19 Thread Loïc Chanel
Hi all, I have a little issue with HiveServer2 since I have enabled Kerberos. I'm unable to connect to the service via Beeline. When doing !connect jdbc:hive2://192.168.6.210:1 hive hive org.apache.hive.jdbc.HiveDriver I keep receiving the same error : Error: Could not open client transport wi