you can add the buckets to a paritions no problems with that.
But to have a bucketed map join what you need is, both the tables need to
bucketed and they need to be in the multiplication factor of each other
like if you have X number of buckets on table A then table B will need NX
number of partit
I tried this interesting idea but also felt a little confusing.
I guess you'll need to change the table schema so that it has both buckets
and partitions.
And to take advantage of the buckets inside the partitions, for example
using the bucket map join, you'll need to specify one particular parti
Currently, hive storage handler reads rows one by one.
https://issues.apache.org/jira/browse/HIVE-3603 is for setting cache
size, which is not yet fixed.
2013/5/1 kulkarni.swar...@gmail.com :
> That depends on how dynamic your data is. If it is pretty static, you can
> also consider using somethi
Thanks for the solution & the tip. :)
From: Sanjay Subramanian
To: "user@hive.apache.org" ; sumit ghosh
Sent: Tuesday, 30 April 2013 7:44 PM
Subject: Re: Variable resolution Fails
+1 agreed
Also as a general script programming practice I check if the var
Naidu:
Please don't hijack existing thread. Your questions are not directly related to
Hive.
Cheers
On May 1, 2013, at 12:53 AM, Naidu MS wrote:
> Hi i have two questions regarding hdfs and jps utility
>
> I am new to Hadoop and started leraning hadoop from the past week
>
> 1.when ever i s
*Could not synchronize with target -- * this is for when jps could not
connect to application waiting for direct attach
this is fixed in sun jdk 1.6 update 7. you may want to check which version
of jdk you are using
it is not required to format the namenode again and again. When you format
the nam
Hi i have two questions regarding hdfs and jps utility
I am new to Hadoop and started leraning hadoop from the past week
1.when ever i start start-all.sh and jps in console it showing the
processes started
*naidu@naidu:~/work/hadoop-1.0.4/bin$ jps*
*22283 NameNode*
*23516 TaskTracker*
*26711 Jps