at 12:04, Ankit Bhatnagar wrote:
> r u looking for
>
> hive -e "*set hive.cli.print.header=true*; < query> " > output
>
>
> On Tuesday, January 12, 2016 10:14 PM, Sreenath
> wrote:
>
>
> Hi All,
>
> Is there a way we can write the hive column
Hi All,
Is there a way we can write the hive column headers also along with the
output when we are overwriting a query's output to an HDFS or local
directory ?
--
Sreenath S Kamath
Bangalore
Ph No:+91-9590989106
finishes in 20 min. So was wondering if
someone can help me in understanding what hive is doing in this 2 hrs ?
Would really appreciate some help here . Thanks in advance
Best,
Sreenath
>
>
>
>
>
> Anyone has idea why dropping a table could take up to 60 seconds?
>
>
>
> ---
>
> [image: cid:image001.jpg@01CE92B5.CB034C90]
>
> *Dima Machlin, **Big Data Architect*
>
> 15 Abba Eban Blvd. PO Box 4125, Herzliya 46140 IL
>
> P: +972-9-9518147 |M: +972-54-5671337|F: +972-9-9584736
>
> *Pursway.com* <http://www.pursway.com/>
>
>
>
--
Sreenath S Kamath
Bangalore
Ph No:+91-9590989106
Location
>> ELSE
>> (
>> SELECT TOP (1)
>> T2.location
>> FROM#temp1 AS T2
>> WHERE T2.record < T1.record
>> AND T2.fk = T1.fk
>> AND T2.location != -1
>> ORDER BY T2.Record DESC
>> )
>> ENDFROM#temp1 AS T1
>>
>> Thank you for your help in advance!
>>
>
>
>
> --
> Nitin Pawar
>
--
Sreenath S Kamath
Bangalore
Ph No:+91-9590989106
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:88)
... 9 more
Any idea why this would happen ?
--
Sreenath S Kamath
Bangalore
05-02');
>
> it returns below exception:
>
> Error: Error while processing statement: FAILED: SemanticException
> [Error 10014]: Line 1:38 Wrong arguments ''20014-03-02'': The arguments
> for IN should be the same type! Types are: {date IN (string, string)
gt;
> On Thu, May 22, 2014 at 12:20 PM, Sreenath wrote:
>
>> Ok what is the result you are expecting once i run this command ?
>>
>>
>> On Thu, May 22, 2014 at 12:17 PM, Nagarjuna Vissarapu <
>> nagarjuna.v...@gmail.com> wrote:
>>
>>> I think
Thu, May 22, 2014 at 11:58 AM, Shengjun Xin wrote:
>
>> Are datanodes dead?
>>
>>
>> On Thu, May 22, 2014 at 2:23 PM, Sreenath wrote:
>>
>>> Hi All,
>>>
>>> We are running a hadoop cluster and many of our hive queries are failing
Hi,
No the data nodes are not dead and the HDFS is almost 70% free.
Is it related to some network Issues?
On Thu, May 22, 2014 at 11:58 AM, Shengjun Xin wrote:
> Are datanodes dead?
>
>
> On Thu, May 22, 2014 at 2:23 PM, Sreenath wrote:
>
>> Hi All,
>>
>> We
:3290)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2200(DFSClient.java:2783)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2987)
Any IDead why this might be happening ?
--
Sreenath S Kamath
Bangalore
Ph No:+91-9590989106
parameters not
getting set properly. CAn you guys help us out ?
--
Sreenath S Kamath
Hi all,
Is there a way in hive by which i can redirect the output of a query to
some folder that already has some data.
I know 'INSERT OVERWRITE DIRECTORY' but this overwrites the existing
content
thanks
Is adding too many files(python udfs) an overhead for hive query execution
?
--
Sreenath S Kamath
You can actually create a partitioned and keep on adding data hourly into
various folder s
On 21-Apr-2014 6:26 PM, "Kishore kumar" wrote:
>
> Hi Experts,
>
> We are running four node cluster which is installed cdh4.5 with cm4.8, We
> have large size files in zip format in s3, we want to analyze t
Hi all,
I have a partitioned table in hive where each partition will have 630 gzip
compressed files each of average size 100kb. If I query over these files
using hive it will generate exactly 630 mappers i.e one mapper for one file.
Now as an experiment i tried reading those files with pig and pig
sure that will do what is required thanks nitin
On Fri, Feb 14, 2014 at 4:38 PM, Nitin Pawar wrote:
> hive -e " select * from table limit 3" > localfile
>
> will above do for what you want to do?
>
>
> On Fri, Feb 14, 2014 at 4:36 PM, Sreenath wrote:
>
>&
meaning i want the output in a local directory without it scanning through
the whole table
On Fri, Feb 14, 2014 at 4:27 PM, Nitin Pawar wrote:
> what do you mean workaround for this ? its not a problem
>
>
> On Fri, Feb 14, 2014 at 4:26 PM, Sreenath wrote:
>
>> hi all,
>
file.
>> Thats why the mapreduce job to create the file.
>>
>> Where as for other way when its displayed on console, its just hdfs cat
>> operation
>>
>>
>> On Fri, Feb 14, 2014 at 4:13 PM, Sreenath wrote:
>>
>>> i have two queries
>>>
and scans through all the rows launching map-reduce jobs
why the differenceinexecution of both queries?
--
Sreenath S Kamath
Bangalore
Ph No:+91-9590989106
should not disseminate, distribute or copy
>> this e-mail. Please notify the sender immediately and destroy all copies of
>> this message and any attachments.
>>
>> WARNING: Computer viruses can be transmitted via email. The recipient
>> should check this email and any attachments for the presence of viruses.
>> The company accepts no liability for any damage caused by any virus
>> transmitted by this email.
>>
>> www.wipro.com
>>
>
>
--
Sreenath S Kamath
Bangalore
Ph No:+91-9590989106
Hi,
Has anyone written a python udf that can take the input as the group by of
a subquery and process this set to output something. If yes can you please
share a query that uses this udf ?
--
Sreenath S Kamath
Bangalore
Ph No:+91-9590989106
an
over head to the cluster.
--
Sreenath S Kamath
Bangalore
, David Morel wrote:
> On 25 Nov 2013, at 11:50, Sreenath wrote:
>
> hi all,
>>
>> We are using hive for Ad-hoc querying and have a hive table which is
>> partitioned on two fields (date,id).Now for each date there are around
>> 1400
>> ids so on a single day ar
'/Analyze%2F2013%2F10%2F03%2F465' - Unexpected response code 404, expected
200
Anyone has a quick fix for this ?
--
Sreenath S Kamath
Bangalore
Ph No:+91-9590989106
By instance I mean a set of mapreduce jobs (3 in this case)..when
executing in Cli only one instance runs and output is displayed on the
screen, but this is not the case when using with PowerPivot (multiple
instance one after the other and contains the same no. of HDFS read
write...etc)...a
There are multiple instances of 3 mpareduce jobs (executing one after the
other) on running the single query using powerpivot.
I can find out next instance when this throws up in the screen after like
2 instance of the 3mapreduce jobs.
Hive history
file=/tmp/hadoop/hive_job_log_hadoop_201206121120_
Yes understood. I do not have a problem in defining the parameters in the
code. But the problem is, I am using PowerPivot as the visualization engine.
Now, when I give the query as a set like:
add jar /usr/local/hadoop/src/retweetlink1.jar;
create temporary function link as
Like is there anyway to make the .hiverc file be executed even in
hiveserver instance.
simple way like this
hive --service hiveserver -i .hiverc
doesnot work Nithin
Any other way Nitin, I just want to add a single jar file and do not know
much about custom hive build. And this requirement may vary at some other
point of time. Its not a good way of building hive each time I need a new
jar to be added.
I have a jar file : 'twittergen.jar', now how can I add it to hive lib.
Kindly help. I need the function to be used across sections when running a
server instance. Now stuck up with this.
Ya UDF do not live across section. But what if I just want the temporary
function to be created each time of a new session. This is what is done
with help of .hiverc. But again this is working only with cli mode not in
server mode.
BTW I am interested to know how to build the function into hive, k
Nithin,
Any idea on invoking .hiverc when running : /usr/hive/bin/hive --service
hiveserver
This works when I am using Hive cli.
i.e. When I give: select link(tweet) from tweetsdata; in the cli and
defined the function 'link' in .hiverc
But when i run /usr/hive --service hiveserver
And use the function in PowerPivot, then it says that 'link' is not defined.
K..so i have created a file 'sample.hiverc' in the home directory..how do I
run this particular file
Hi Nitin
Can u kindly help me (briefly) on how to add to hiverc...no such location
exsist in my machine
s fine for the 1st and 2nd
round of map-reduce jobs, but when the session changes then throws an error
like retweetlink class is not found.
How do I make an UDF support across multiple sessions.
Thanks
Sreenath Mullassery
OK I am getting a little confused now.
Consider that I am working on a scenario where there is no limit with
memory available.
In such scenario, is there any advantage of storing data in HDFS in
compressed format. Any advantage, like, if node 1 has data available and it
is executing a particular t
Any idea about lzo or bzip2...any of these splittable??
Kindly check out Apache Mahout and whether it satisfies your needs.
Hi Bejoy
I would like to make this clear.
There is no gain on processing throughput/time on compressing the data
stored in HDFS (not talking about intermediate compression)...wright??
And do I need to add the lzo libraries in Hadoop_Home/lib/native for all
the nodes (including the slave nodes)??
k...understood...so you load the compressed data into memory (thereby
decreasing the size of file needed to be loaded) and then apply
decompression algorithm to get the uncompressed data. is this what happens?
Thanks for the response.
1)How do I use the Gz compression and does it come with Hadoop. Or else how
do I build a compression method for using in Hive. I would like to run
evaluation across compression methods.
What is the default compression used in Hadoop.
2)Kindly bear with me if this question
Thanks all
All help is greatly appreciated. Pl. feel free to post whatever comes to
your mind.
Learned a lot from this conversation.
Pl. post any findings on this topic : Hive as a warehouse - limitations
Thanks
, 2012 at 3:49 PM, Bejoy Ks wrote:
> Hi Sreenath
>
> If you are looking at a UI for queries then Cloudera's hue is the
> best choice. Also you do have odbc connectors that integrates BI tools like
> microstrategy, tableau etc with hive.
&g
Hi all
I am new to hive and am working on analysis of twitter data with Hive and
Hadoop in a 27node cluster.
At present am using Microsoft powerpivot as the visualization tool for
visual representation of analysis done using Hive and have got some really
good results and I am stunned by the scalab
Hi Bejoy
Thanks for the response. Can you provide help on writing a custom serde for
the mentioned scenario(or for a general case) and how to include it in hive
directory.
From: Bejoy Ks
To: "user@hive.apache.org" ; m sreenath
Sent: Wednes
Hi
I have a large data-set of the form :
A 12323:213
B adasd
C sedweqwd
A 12323:213
B adasd
C sedweqwd
The general form is :
\t
\t
\t\n
In the given format, I need the part from the three rows in a table
containing three columns (time,name,content). Can regex be used for thi
48 matches
Mail list logo