Hello All,
*Short version*: Looking to leverage hbase fuzzy row scans
on fields in a composite rowkey ( on hbase 0.94.6 ).
*Long version*: Our hbase store design is heavily dependent on a bucketed
composite rowkey that uses all the lovely features in hbase to spread the
read/write
Hi, Narayanan:
The current problem is that for a generic solution, there is no way that we
know that element in the Json is an array. Keep in mind that in any element of
Json, it could be any valid structure. So it could be array, another structure,
or map etc.
You know your data, so you can sa
Thanks Peyman.
Actually the problem with Hive-Json-Serde is that we need to provide
the entire schema upfront while creating the table.
My requirement is that we just project/aggregate on the fields using
get_json_object after creating the external table without schema. This
way the external tabl
perhaps: https://github.com/rcongiu/Hive-JSON-Serde
On Mon, Apr 7, 2014 at 6:52 PM, Narayanan K wrote:
> Hi all
>
> I am using get_json_object to read a json text file. I have created
> the external table as below :
>
> CREATE EXTERNAL TABLE EXT_TABLE ( json string)
> PARTITIONED BY (dt string)
Hi all
I am using get_json_object to read a json text file. I have created
the external table as below :
CREATE EXTERNAL TABLE EXT_TABLE ( json string)
PARTITIONED BY (dt string)
LOCATION '/users/abc/';
The json data has some fields that are not simple fields but fields
which are nested fields
Hi,
I have a use-case where I want a UDF to return dynamic number of columns
based on parameter.
Consider a udf function called return_cols which takes two params, fields
and a comma-separated list of output fields eg. select
return_cols(table-field, "foo,bar") from table or select
return_cols(ta
Thanks Eugene.
I had gone over the link previously. I am not looking for a way to do this
via Pig CLI. My data is going to be read by a *MapReduce job* (by Pig i
meant custom Pig source code) and hence need input readers and output
writers. I need to be able to write using HCatOutputFormat.
I cur
Dear Friends,
I have created the indexes but I am not able to use them. Does hive have an
index optimizer?
For a simple query in Oracle/Postgres, when we do EXPLAIN it gives the
Query Plan with indexes but with Hive it does not show me the indexes. How
can we make sure that the indexes are being
If you are writing from Pig using HCatStorer you don't need to create
HCatSchema.
https://cwiki.apache.org/confluence/display/Hive/HCatalog+LoadStore#HCatalogLoadStore-Usage.1has
examples on how to do it.
So if you create a Hive table that use ORC you should be able to write your
Pig cursor to tha