By instance I mean a set of mapreduce jobs (3 in this case)..when
executing in Cli only one instance runs and output is displayed on the
screen, but this is not the case when using with PowerPivot (multiple
instance one after the other and contains the same no. of HDFS read
write...etc)...a
There are multiple instances of 3 mpareduce jobs (executing one after the
other) on running the single query using powerpivot.
I can find out next instance when this throws up in the screen after like
2 instance of the 3mapreduce jobs.
Hive history
file=/tmp/hadoop/hive_job_log_hadoop_201206121120_
Yes understood. I do not have a problem in defining the parameters in the
code. But the problem is, I am using PowerPivot as the visualization engine.
Now, when I give the query as a set like:
add jar /usr/local/hadoop/src/retweetlink1.jar;
create temporary function link as
Like is there anyway to make the .hiverc file be executed even in
hiveserver instance.
simple way like this
hive --service hiveserver -i .hiverc
doesnot work Nithin
Any other way Nitin, I just want to add a single jar file and do not know
much about custom hive build. And this requirement may vary at some other
point of time. Its not a good way of building hive each time I need a new
jar to be added.
I have a jar file : 'twittergen.jar', now how can I add it to hive lib.
Kindly help. I need the function to be used across sections when running a
server instance. Now stuck up with this.
Ya UDF do not live across section. But what if I just want the temporary
function to be created each time of a new session. This is what is done
with help of .hiverc. But again this is working only with cli mode not in
server mode.
BTW I am interested to know how to build the function into hive, k
Nithin,
Any idea on invoking .hiverc when running : /usr/hive/bin/hive --service
hiveserver
This works when I am using Hive cli.
i.e. When I give: select link(tweet) from tweetsdata; in the cli and
defined the function 'link' in .hiverc
But when i run /usr/hive --service hiveserver
And use the function in PowerPivot, then it says that 'link' is not defined.
K..so i have created a file 'sample.hiverc' in the home directory..how do I
run this particular file
Hi Nitin
Can u kindly help me (briefly) on how to add to hiverc...no such location
exsist in my machine
Hi
I am using Hive with Microsoft PowerPivot as the visualization tool.
When I am running a query involving UDF like this from PowerPivot:
add jar /usr/local/hadoop/src/retweetlink1.jar;
create temporary function link as 'retweetlink';
Followed by a select statement, the query executes fine for t
OK I am getting a little confused now.
Consider that I am working on a scenario where there is no limit with
memory available.
In such scenario, is there any advantage of storing data in HDFS in
compressed format. Any advantage, like, if node 1 has data available and it
is executing a particular t
Any idea about lzo or bzip2...any of these splittable??
Kindly check out Apache Mahout and whether it satisfies your needs.
Hi Bejoy
I would like to make this clear.
There is no gain on processing throughput/time on compressing the data
stored in HDFS (not talking about intermediate compression)...wright??
And do I need to add the lzo libraries in Hadoop_Home/lib/native for all
the nodes (including the slave nodes)??
k...understood...so you load the compressed data into memory (thereby
decreasing the size of file needed to be loaded) and then apply
decompression algorithm to get the uncompressed data. is this what happens?
Thanks for the response.
1)How do I use the Gz compression and does it come with Hadoop. Or else how
do I build a compression method for using in Hive. I would like to run
evaluation across compression methods.
What is the default compression used in Hadoop.
2)Kindly bear with me if this question
Thanks all
All help is greatly appreciated. Pl. feel free to post whatever comes to
your mind.
Learned a lot from this conversation.
Pl. post any findings on this topic : Hive as a warehouse - limitations
Thanks
t;
> Regards
> Bejoy KS
>
> ------
> *From:* Sreenath Menon
> *To:* user@hive.apache.org
> *Sent:* Monday, June 4, 2012 2:42 PM
> *Subject:* Front end visualization tool with Hive (when using as a
> warehouse)
>
> Hi all
>
> I am new to hive and am working on analysis of
Hi all
I am new to hive and am working on analysis of twitter data with Hive and
Hadoop in a 27node cluster.
At present am using Microsoft powerpivot as the visualization tool for
visual representation of analysis done using Hive and have got some really
good results and I am stunned by the scalab
21 matches
Mail list logo