Hi,
I have directories in HDFS that get populated by nightly ETL job on which we
have Hive external tables. The problem we are having is that when the table
schema changes such that there are new columns in between existing columns.
Is there a straightforward way of having the older data files to
Is there any support for putting an encrypted password to my metastore db in
hive-site.xml?
Maybe an integration with jasypt?
--
Frank Maritato
You can also try using a custom reducer script, as follows:
FROM (
SELECT groupCol, metric, otherFieldYouCareAbout
FROM MyTable
DISTRIBUTE BY groupCol
SORT BY groupCol ASC, metric DESC
) t1
REDUCE *
USING 'myGroupingReduceScript.py'
AS groupCol, metric, otherFieldYouC
I'm beginning to suspect this myself. We have a import job which has
many smaller files. We've been merging them into a single log file and
partitioning by day however I've seen this and other errors (usually
memory related errors) posted by hive and the load fails.
Our latest error has been
It seems that the more dynamic partitions are imported the fewer I am able to
import respectively the smaller the files have to be.
Any clues?
Original-Nachricht
> Datum: Wed, 13 Jul 2011 09:45:27 +0200
> Von: "labtrax"
> An: user@hive.apache.org
> Betreff: Re: dynamic partiti
Hi,
I allways get
java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException:
Hive Runtime Error while processing row (tag=0)
{"key":{},"value":{"_col0":"1129","_col1":"Campaign","_col2":"34811433","_col3":"group","_col4":"1271859453","_col5":"Soundso","_col6":"93709590","_col