HI,
I see the SUM(double_column) displays the result in scientific notation in
the hive cli. Is there any way to customize the number display in hive CLI?
--
Thanks,
Pandeeswaran
Hi,
I am seeing the double values are displayed as scientifi not
Thanks Owen and Hongzhan.
I understand the behavior now.
On Tue, Aug 13, 2013 at 6:28 AM, hongzhan li wrote:
> if you select all the columns ,the orc will not faster than textfile.but
> if you select some column (not all of the colimns),orc will run faster.
> —
>
>
> On Mon, Aug 12, 2013 at 6:
CONFIDENTIALITY NOTICE
==
This email message and any attachments are for the exclusive use of the
intended recipient(s) and may contain confidential and privileged information.
Any unauthorized review, use, disclosure or distribution is prohibited. If you
are not the inten
if you select all the columns ,the orc will not faster than textfile.but if you
select some column (not all of the colimns),orc will run faster.
—
On Mon, Aug 12, 2013 at 6:40 PM, pandees waran wrote:
> Hi,
> Currently, we use TEXTFILE format in hive 0.8 ,while creating the
> external tables in
I am sorry if I am missing something very basic with hive
authorization on 0.9 or before.
I went through the existing security in hive, before hiveserver2 (i.e.
hove 0.9 or before) and I see all the permission related stuff like
Users, Groups, and Roles
Now, as far as I understand, hive query is
Hi
Hive version 0.9.0 (hive-common-0.9.0-cdh4.1.2.jar)
hive.security.authorization.enabled
true
enable or disable the hive client authorization
Linux User = hiveuser1 (no hive permissions)
CASE 1
hive -e "select * from outpdir_ptitle_explanation_parsed limit 10"
Authorization failed:
Ok, given the large number of doubles in the schema and bzip2 compression,
I can see why the text would be smaller.
ORC doesn't do compression on floats or doubles, although there is a jira
to do so. (https://issues.apache.org/jira/browse/HIVE-3889)
Bzip is a very aggressive compressor. We should
Hi Owen,
Thanks for your response.
My structure is like:
a)Textfile:
CREATE EXTERNAL TABLE test_textfile (
COL1 BIGINT,
COL2 STRING,
COL3 BIGINT,
COL4 STRING,
COL5 STRING,
COL6 BIGINT,
COL7 BIGINT,
COL8 BIGINT,
COL9 BIGINT,
COl10 BIGINT,
COl11 BIGINT,
Pandees,
I've never seen a table that was larger with ORC than with text. Can you
share your text's file schema with us? Is the table very small? How many
rows and GB are the tables? The overhead for ORC is typically small, but as
Ed says it is possible for rare cases for the overhead to dominate
You can create CustomAuthenticator implements
PasswordAuthenticationProvider.
However, there is a bug to use CustomAuthenticator,
you should apply patch from https://issues.apache.org/jira/browse/HIVE-4778
And then register on hive-site.xml:
hive.server2
Thanks Edward. I shall try compression besides orc and let you know. And
also, it looks like the cpu usage is lesser while querying orc rather
than text file.
But the total time taken by the query time is slightly more in orc than
text file. Could you please explain the difference between cumul
Colmnar formats do not always beat row wise storage. Many times gzip plus
block storage will compress something better then columnar storage
especially when you have repeated data in different columns.
Based on what you are saying it could be possible that you missed a setting
and the ocr are not
Hi,
how to provide a simple propertyfile or database user/password
authentication for HiveServer2?
Since it is a small research project, we don't have Kerberos or LDAP.
I already found this [0] presentation about it, but it's not in English :(.
On the Cloudera reference manual [1] I found smth. a
Hi,
Currently, we use TEXTFILE format in hive 0.8 ,while creating the
external tables in intermediate processing .
I have read about ORC in 0.11. I have created the same table in 0.11
with ORC format.
Without any compression, the ORC file(totally 3 files) occupied the
space twice more than the TEX
14 matches
Mail list logo