Hi,
I apologize for the wide distribution and if this is not the right mailing
list for this.
We write Avro files to Parquet and load them to HDFS so they can be
accessed via an EXTERNAL Hive table. These records have two timestamp
fields which are expressed in the Avro schema as type = long and
Did you guys already look at Dr Elephant?
https://engineering.linkedin.com/blog/2016/04/dr-elephant-open-source-self-serve-performance-tuning-hadoop-spark
Not sure if there is anything you might find useful, but I would be interested
in hearing about good and bad about Dr Elephant w/ Hive.
Sent
Hi,
I am interested in working on a project that takes a large number of Hive
queries (as well as their meta data like amount of resources used etc) and
find out common sub queries and expensive query groups etc.
Are there any existing work in this domain? Happy to collaborate as well
if there a
Thank you Gopal for pointing the root cause. After running command alter table
xxx compact ‘major’ to request a force compaction, total length is right !
Is there any way to do compact immediately after insert values.
Bob He
Thanks
On 25 Jul 2018, at 1:45 PM, Gopal Vijayaraghavan wrote:
> Sea
Happy to help! :)
Proust (Feng Guizhou) [FDS Payment and Marketing] ezt
írta (időpont: 2018. júl. 24., Ke 12:17):
> Just FYI, I'm able to make a custom UDF to apply the thread-safe code
> changes.
>
> Thanks a lot for your help
>
>
> Guizhou
> --
> *From:* Proust (Fe
Hi Gopal,
Thanks for your reply! One more question, does the effect of using pure-java
version is the same as that of using SnappyCodec? Or, in other words, is there
any difference between these two methods, about the compression result and
effect?
Looking forward to your reply and help.
B