Re: bloom filter used in 0.14?

2016-01-28 Thread Gopal Vijayaraghavan
> So I am questioning whether it is enabled on the version I am on, which >is 0.14. Does anyone know? https://issues.apache.org/jira/browse/HIVE-9188 - fix-version (1.2.0) The version you are using does not have bloom filter support. It should be ignoring the parameter and not generating any bl

Re: bloom filter used in 0.14?

2016-01-28 Thread Prasanth Jayachandran
Hi To get optimal performance from bloom filter, make sure the records of col1 are sorted. Sorted on the column of interest will efficiently prune stripes and row groups. If the records that you are searching for is spread across row groups (10K rows by default) or stripes (64MB by default) the

bloom filter used in 0.14?

2016-01-28 Thread Frank Luo
All, I have a huge table that I periodically want to do select on some particular value. For example, supposing I have a table for the entire world population. Then I know the id of “1234” is criminal, hence I want to pull out his information from the table. Without any optimization, I have to

Re: Stored By

2016-01-28 Thread Gabriel Balan
Hi Why not write your own storage handler extending AccumuloStorageHandler and overriding getInputFormatClass() to return your HiveAccumuloTableInputFormat subclass. hth Gabriel Balan On 1/21/2016 10:46 AM, peter.mar...@baesystems.com wrote: Hi, So I am using the AccumuloStorageHandler to

Re:Two questions about working with Hive using jdbc

2016-01-28 Thread Todd
Can someone help answer the questions? Thanks -- 发自我的网易邮箱平板适配版 在 2016-01-28 22:11:29,Todd 写道: Hi, I am using Hive 0.14, and I am using JDBC to connect the Hive thrift server to do queries things, I encounter two issues- 1. When the query is issued,how can i get the job id(mapreduce th

RE: Backing up hive database

2016-01-28 Thread Mich Talebzadeh
Hi Techie, Have you decided on your HA approach by any chance? Dr Mich Talebzadeh LinkedIn https://www.linkedin.com/profile/view?id=AAEWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw Sybase ASE 15 Gol

Re: "Create external table" nulling data from source table

2016-01-28 Thread Gopal Vijayaraghavan
> And again: the same row is correct if I export a small set of data, and >incorrect if I export a large set - so I think that file/data size has >something to do with this. My Phoenix vs LLAP benchmark hit size related issues in ETL. In my case, the tipping point was >1 hdfs block per CSV file.

CVE-2015-7521: Apache Hive authorization bug disclosure

2016-01-28 Thread Sushanth Sowmyan
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 CVE-2015-7521: Apache Hive authorization bug disclosure Severity: Important Vendor: The Apache Software Foundation Versions Affected: Apache Hive 1.0.0 - 1.0.1 Apache Hive 1.1.0 - 1.1.1 Apache Hive 1.2.0 - 1.2.1 Description: Some partition-level o

RE: "Create external table" nulling data from source table

2016-01-28 Thread Riesland, Zack
Thanks Nick, I don't think there should be any newline chars - the columns that are turning null are mostly Decimal, and none of my strings are longer than one word (no spaces). And again: the same row is correct if I export a small set of data, and incorrect if I export a large set - so I thi

Re: "Create external table" nulling data from source table

2016-01-28 Thread Nicholas Hakobian
Do you have any fields with embedded newline characters? If so, certain hive output formats will parse the newline character as the end of row, and when importing, chances are the missing fields (now part of the next row) will be padded with nulls. This happens in Hive as well if you are using a Te

"Create external table" nulling data from source table

2016-01-28 Thread Riesland, Zack
First time posting to this list. Please forgive me if I break etiquette. I'm looking for some help with getting data from hive to hbase. I'm using HDP 2.2.8. I have a compressed (zlib), orc-based hive table with 12 columns and billions of rows. In order to get the data into hbase, I have to cr

Two questions about working with Hive using jdbc

2016-01-28 Thread Todd
Hi, I am using Hive 0.14, and I am using JDBC to connect the Hive thrift server to do queries things, I encounter two issues- 1. When the query is issued,how can i get the job id(mapreduce that run the query),so that I can get a chance to be able to kill the job. 2. I want to execute a sql file