> Is there anyway to avoid creating sub-directories? Or this is by design
>and can not be changed?
This is because of the way file-formats generate hadoop name files without
collisions.
For instance, any change to that would break Parquet-MR for Tez. That's
why we generate a compatible, but colli
Hi Team,
Below insert with union-all will create sub-directories:
set hive.execution.engine=tez;
create table h1_passwords_target like h1_passwords;
insert overwrite table h1_passwords_target
select * from
(select * from h1_passwords limit 1
union all
select * from h1_passwords limit 2 ) sub
I have several Parquet files on Amazon S3, which have three fields, username
STRING, request STRING and timestamp STRING.
If I create a table in Hive and add a partition, everything is fine.
CREATE EXTERNAL TABLE IF NOT EXISTS access_logs(username STRING,
request STRING, `timestamp` STRING)
PARTI
Can somebody gives me some pointer to looked upon?
On Wed, Aug 19, 2015 at 9:26 AM, Sanjeev Verma
wrote:
> Hi
> We are experiencing a strange problem with the hiveserver2, in one of the
> job it gets the GC limit exceed from mapred task and hangs even having
> enough heap available.we are not ab
Also, I tried setting the "hive.fetch.task.conversion" property in the JDBC
URL, like so: jdbc:hive2://
192.168.132.128:1/default?hive.fetch.task.conversion=none, but it is
still creating mapreduce tasks for the query, so it effectively seems to be
ignoring that property.
On Wed, Aug 19, 2015
Hi
We are experiencing a strange problem with the hiveserver2, in one of the
job it gets the GC limit exceed from mapred task and hangs even having
enough heap available.we are not able to identify what causing this issue.
Could anybody help me identify the issue and let me know what pointers I
nee
Correct me if I am wrong, my understanding is that after using kerberos
authentication, you probably don’t need the password.
Hope it helps
Thanks,
Jary
> On Aug 19, 2015, at 9:09 AM, Loïc Chanel wrote:
>
> By the way, thanks a lot for your help, because your solution works, but I'm
> still
When I run the "SELECT * FROM " query it is running it as a
mapreduce job. I can see it in the Yarn Manager and also in the Tez UI.
This is also when the fetch size is not honored and it tries to basically
return all results at once. Is there a way to make this work?
On Wed, Aug 19, 2015 at 10:53
By the way, thanks a lot for your help, because your solution works, but
I'm still interested in knowing what is the password I did not enter.
Thanks again,
Loïc
Loïc CHANEL
Engineering student at TELECOM Nancy
Trainee at Worldline - Villeurbanne
2015-08-19 18:07 GMT+02:00 Loïc Chanel :
> All
All right, but then, what is the password hive asks for ? Hive's one ? How
do I know its value ?
Loïc CHANEL
Engineering student at TELECOM Nancy
Trainee at Worldline - Villeurbanne
2015-08-19 17:51 GMT+02:00 Jary Du :
> For Beeline connection string, it should be "!connect
> jdbc:hive2://:/;pri
For Beeline connection string, it should be "!connect
jdbc:hive2://:/;principal=”.
Please make sure it is the hive’s principal, not the user’s. And when you
kinit, it should be kinit user’s keytab, not the hive’s keytab.
> On Aug 19, 2015, at 8:46 AM, Loïc Chanel wrote:
>
> Yeah, I forgo
Yeah, I forgot to mention it, but each time I did a kinit user/hive before
launching beeline, as I read somewhere that Beeline does not handle
Kerberos connection.
So, as I can make klist before launching beeline and having a good result,
the problem does not come from this. Thanks a lot for your
"The Beeline client must have a valid Kerberos ticket in the ticket cache
before attempting to connect."
(http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1.3/bk_dataintegration/content/ch_using-hive-clients-examples.html)
So you need kinit first to have the valid Kerberos ticket int the tick
Hi again,
As I searched another way to make some requests with Kerberos enabled for
security on HiveServer, I found that this request should do the same :
!connect jdbc:hive2://
192.168.6.210:1/default;principal=user/h...@westeros.wl
org.apache.hive.jdbc.HiveDriver
But now I've got another err
actually it should be something like
getHandleIdentifier()=hfhkjhfjhkjfh-dsdsad-sdsd--dsada:
fetchResults()
On Wed, Aug 19, 2015 at 3:49 PM, Prem Yadav wrote:
> Hi Emil,
> for either of the queries, there will be no mapreduce job. the query
> engine understands that in both case, it need not do
Hi Emil,
for either of the queries, there will be no mapreduce job. the query engine
understands that in both case, it need not do any computation and just
needs to fetch all the data from the files.
The fetch size should be honored in both cases. Hope you are using
hiveserver2.
You can try connec
I have a small Java app that I wrote that uses JDBC to run a hive query.
The Hive table that I'm running it against has 30+ million rows, and I want
to pull them all back to verify the data. If I run a simple "SELECT * FROM
" and set a fetch size of 30,000 then the fetch size is not honored
and it
I saw the same result. But I debugged a little bit and figured out that it was
the PPD optimizer did the transitivity propagation, not
PredicateTransitivePropagate.
-- Original --
From: "Gopal Vijayaraghavan";
Date: Wed, Aug 19, 2015 02:52 PM
To: "user";
Hi all,
I have a little issue with HiveServer2 since I have enabled Kerberos. I'm
unable to connect to the service via Beeline. When doing
!connect jdbc:hive2://192.168.6.210:1 hive hive
org.apache.hive.jdbc.HiveDriver
I keep receiving the same error :
Error: Could not open client transport wi
19 matches
Mail list logo