Re: Writing hive column headers in 'Insert overwrite query'

2016-01-12 Thread Sreenath
at 12:04, Ankit Bhatnagar wrote: > r u looking for > > hive -e "*set hive.cli.print.header=true*; < query> " > output > > > On Tuesday, January 12, 2016 10:14 PM, Sreenath > wrote: > > > Hi All, > > Is there a way we can write the hive column

Writing hive column headers in 'Insert overwrite query'

2016-01-12 Thread Sreenath
Hi All, Is there a way we can write the hive column headers also along with the output when we are overwriting a query's output to an HDFS or local directory ? -- Sreenath S Kamath Bangalore Ph No:+91-9590989106

Hive Start Up Time Manifolds Greater than Execution Time

2015-09-18 Thread Sreenath
finishes in 20 min. So was wondering if someone can help me in understanding what hive is doing in this 2 hrs ? Would really appreciate some help here . Thanks in advance Best, Sreenath

Re: dropping tables can take long time

2014-10-22 Thread Sreenath
> > > > > > Anyone has idea why dropping a table could take up to 60 seconds? > > > > --- > > [image: cid:image001.jpg@01CE92B5.CB034C90] > > *Dima Machlin, **Big Data Architect* > > 15 Abba Eban Blvd. PO Box 4125, Herzliya 46140 IL > > P: +972-9-9518147 |M: +972-54-5671337|F: +972-9-9584736 > > *Pursway.com* <http://www.pursway.com/> > > > -- Sreenath S Kamath Bangalore Ph No:+91-9590989106

Re: Correlated Subqueries Workaround in Hive!

2014-09-15 Thread Sreenath
Location >> ELSE >> ( >> SELECT TOP (1) >> T2.location >> FROM#temp1 AS T2 >> WHERE T2.record < T1.record >> AND T2.fk = T1.fk >> AND T2.location != -1 >> ORDER BY T2.Record DESC >> ) >> ENDFROM#temp1 AS T1 >> >> Thank you for your help in advance! >> > > > > -- > Nitin Pawar > -- Sreenath S Kamath Bangalore Ph No:+91-9590989106

Error Configuring Object In Hive Joins

2014-09-03 Thread Sreenath
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:88) ... 9 more Any idea why this would happen ? -- Sreenath S Kamath Bangalore

Re: hive query with in statement

2014-08-12 Thread Sreenath
05-02'); > > it returns below exception: > > Error: Error while processing statement: FAILED: SemanticException > [Error 10014]: Line 1:38 Wrong arguments ''20014-03-02'': The arguments > for IN should be the same type! Types are: {date IN (string, string)

Re: Reducers Failing

2014-05-21 Thread Sreenath
gt; > On Thu, May 22, 2014 at 12:20 PM, Sreenath wrote: > >> Ok what is the result you are expecting once i run this command ? >> >> >> On Thu, May 22, 2014 at 12:17 PM, Nagarjuna Vissarapu < >> nagarjuna.v...@gmail.com> wrote: >> >>> I think

Re: Reducers Failing

2014-05-21 Thread Sreenath
Thu, May 22, 2014 at 11:58 AM, Shengjun Xin wrote: > >> Are datanodes dead? >> >> >> On Thu, May 22, 2014 at 2:23 PM, Sreenath wrote: >> >>> Hi All, >>> >>> We are running a hadoop cluster and many of our hive queries are failing

Re: Reducers Failing

2014-05-21 Thread Sreenath
Hi, No the data nodes are not dead and the HDFS is almost 70% free. Is it related to some network Issues? On Thu, May 22, 2014 at 11:58 AM, Shengjun Xin wrote: > Are datanodes dead? > > > On Thu, May 22, 2014 at 2:23 PM, Sreenath wrote: > >> Hi All, >> >> We

Reducers Failing

2014-05-21 Thread Sreenath
:3290) at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2200(DFSClient.java:2783) at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2987) Any IDead why this might be happening ? -- Sreenath S Kamath Bangalore Ph No:+91-9590989106

Error's While Using hive.hadoop.supports.splittable.combineinputformat

2014-05-19 Thread Sreenath
parameters not getting set properly. CAn you guys help us out ? -- Sreenath S Kamath

Insert into directory

2014-05-15 Thread Sreenath
Hi all, Is there a way in hive by which i can redirect the output of a query to some folder that already has some data. I know 'INSERT OVERWRITE DIRECTORY' but this overwrites the existing content thanks

Python UDFS

2014-04-28 Thread Sreenath
Is adding too many files(python udfs) an overhead for hive query execution ? -- Sreenath S Kamath

Re: Analyzing data resides on s3 from local hadoop cluster

2014-04-24 Thread Sreenath
You can actually create a partitioned and keep on adding data hourly into various folder s On 21-Apr-2014 6:26 PM, "Kishore kumar" wrote: > > Hi Experts, > > We are running four node cluster which is installed cdh4.5 with cm4.8, We > have large size files in zip format in s3, we want to analyze t

Hive vs Pig against number of files spawned

2014-03-31 Thread Sreenath
Hi all, I have a partitioned table in hive where each partition will have 630 gzip compressed files each of average size 100kb. If I query over these files using hive it will generate exactly 630 mappers i.e one mapper for one file. Now as an experiment i tried reading those files with pig and pig

Re: INSERT OVERWRITE LOCAL DIRECTORY

2014-02-14 Thread Sreenath
sure that will do what is required thanks nitin On Fri, Feb 14, 2014 at 4:38 PM, Nitin Pawar wrote: > hive -e " select * from table limit 3" > localfile > > will above do for what you want to do? > > > On Fri, Feb 14, 2014 at 4:36 PM, Sreenath wrote: > >&

Re: INSERT OVERWRITE LOCAL DIRECTORY

2014-02-14 Thread Sreenath
meaning i want the output in a local directory without it scanning through the whole table On Fri, Feb 14, 2014 at 4:27 PM, Nitin Pawar wrote: > what do you mean workaround for this ? its not a problem > > > On Fri, Feb 14, 2014 at 4:26 PM, Sreenath wrote: > >> hi all, >

Re: INSERT OVERWRITE LOCAL DIRECTORY

2014-02-14 Thread Sreenath
file. >> Thats why the mapreduce job to create the file. >> >> Where as for other way when its displayed on console, its just hdfs cat >> operation >> >> >> On Fri, Feb 14, 2014 at 4:13 PM, Sreenath wrote: >> >>> i have two queries >>>

INSERT OVERWRITE LOCAL DIRECTORY

2014-02-14 Thread Sreenath
and scans through all the rows launching map-reduce jobs why the differenceinexecution of both queries? -- Sreenath S Kamath Bangalore Ph No:+91-9590989106

Re: Date format in Hive

2013-12-24 Thread Sreenath
should not disseminate, distribute or copy >> this e-mail. Please notify the sender immediately and destroy all copies of >> this message and any attachments. >> >> WARNING: Computer viruses can be transmitted via email. The recipient >> should check this email and any attachments for the presence of viruses. >> The company accepts no liability for any damage caused by any virus >> transmitted by this email. >> >> www.wipro.com >> > > -- Sreenath S Kamath Bangalore Ph No:+91-9590989106

Python Udfs

2013-12-24 Thread Sreenath
Hi, Has anyone written a python udf that can take the input as the group by of a subquery and process this set to output something. If yes can you please share a query that uses this udf ? -- Sreenath S Kamath Bangalore Ph No:+91-9590989106

Python UDF's In Hive

2013-12-12 Thread Sreenath
an over head to the cluster. -- Sreenath S Kamath Bangalore

Re: Hive query taking a lot of time just to launch map-reduce jobs

2013-11-25 Thread Sreenath
, David Morel wrote: > On 25 Nov 2013, at 11:50, Sreenath wrote: > > hi all, >> >> We are using hive for Ad-hoc querying and have a hive table which is >> partitioned on two fields (date,id).Now for each date there are around >> 1400 >> ids so on a single day ar

Hive query taking a lot of time just to launch map-reduce jobs

2013-11-25 Thread Sreenath
'/Analyze%2F2013%2F10%2F03%2F465' - Unexpected response code 404, expected 200 Anyone has a quick fix for this ? -- Sreenath S Kamath Bangalore Ph No:+91-9590989106

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-12 Thread Sreenath Menon
By instance I mean a set of mapreduce jobs (3 in this case)..when executing in Cli only one instance runs and output is displayed on the screen, but this is not the case when using with PowerPivot (multiple instance one after the other and contains the same no. of HDFS read write...etc)...a

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-12 Thread Sreenath Menon
There are multiple instances of 3 mpareduce jobs (executing one after the other) on running the single query using powerpivot. I can find out next instance when this throws up in the screen after like 2 instance of the 3mapreduce jobs. Hive history file=/tmp/hadoop/hive_job_log_hadoop_201206121120_

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-12 Thread Sreenath Menon
Yes understood. I do not have a problem in defining the parameters in the code. But the problem is, I am using PowerPivot as the visualization engine. Now, when I give the query as a set like: add jar /usr/local/hadoop/src/retweetlink1.jar; create temporary function link as

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
Like is there anyway to make the .hiverc file be executed even in hiveserver instance. simple way like this hive --service hiveserver -i .hiverc doesnot work Nithin

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
Any other way Nitin, I just want to add a single jar file and do not know much about custom hive build. And this requirement may vary at some other point of time. Its not a good way of building hive each time I need a new jar to be added.

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
I have a jar file : 'twittergen.jar', now how can I add it to hive lib. Kindly help. I need the function to be used across sections when running a server instance. Now stuck up with this.

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
Ya UDF do not live across section. But what if I just want the temporary function to be created each time of a new session. This is what is done with help of .hiverc. But again this is working only with cli mode not in server mode. BTW I am interested to know how to build the function into hive, k

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
Nithin, Any idea on invoking .hiverc when running : /usr/hive/bin/hive --service hiveserver

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
This works when I am using Hive cli. i.e. When I give: select link(tweet) from tweetsdata; in the cli and defined the function 'link' in .hiverc But when i run /usr/hive --service hiveserver And use the function in PowerPivot, then it says that 'link' is not defined.

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
K..so i have created a file 'sample.hiverc' in the home directory..how do I run this particular file

Re: Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
Hi Nitin Can u kindly help me (briefly) on how to add to hiverc...no such location exsist in my machine

Hive UDF support across sections...when running in hiveserver...help

2012-06-11 Thread Sreenath Menon
s fine for the 1st and 2nd round of map-reduce jobs, but when the session changes then throws an error like retweetlink class is not found. How do I make an UDF support across multiple sessions. Thanks Sreenath Mullassery

Re: Compressed data storage in HDFS - Error

2012-06-08 Thread Sreenath Menon
OK I am getting a little confused now. Consider that I am working on a scenario where there is no limit with memory available. In such scenario, is there any advantage of storing data in HDFS in compressed format. Any advantage, like, if node 1 has data available and it is executing a particular t

Re: Compressed data storage in HDFS - Error

2012-06-08 Thread Sreenath Menon
Any idea about lzo or bzip2...any of these splittable??

Re: How to apply data mining on Hive?

2012-06-07 Thread Sreenath Menon
Kindly check out Apache Mahout and whether it satisfies your needs.

Re: Compressed data storage in HDFS - Error

2012-06-06 Thread Sreenath Menon
Hi Bejoy I would like to make this clear. There is no gain on processing throughput/time on compressing the data stored in HDFS (not talking about intermediate compression)...wright?? And do I need to add the lzo libraries in Hadoop_Home/lib/native for all the nodes (including the slave nodes)??

Re: Compressed data storage in HDFS - Error

2012-06-06 Thread Sreenath Menon
k...understood...so you load the compressed data into memory (thereby decreasing the size of file needed to be loaded) and then apply decompression algorithm to get the uncompressed data. is this what happens?

Re: Compressed data storage in HDFS - Error

2012-06-06 Thread Sreenath Menon
Thanks for the response. 1)How do I use the Gz compression and does it come with Hadoop. Or else how do I build a compression method for using in Hive. I would like to run evaluation across compression methods. What is the default compression used in Hadoop. 2)Kindly bear with me if this question

Re: Front end visualization tool with Hive (when using as a warehouse)

2012-06-06 Thread Sreenath Menon
Thanks all All help is greatly appreciated. Pl. feel free to post whatever comes to your mind. Learned a lot from this conversation. Pl. post any findings on this topic : Hive as a warehouse - limitations Thanks

Re: Front end visualization tool with Hive (when using as a warehouse)

2012-06-04 Thread Sreenath Menon
, 2012 at 3:49 PM, Bejoy Ks wrote: > Hi Sreenath > > If you are looking at a UI for queries then Cloudera's hue is the > best choice. Also you do have odbc connectors that integrates BI tools like > microstrategy, tableau etc with hive. &g

Front end visualization tool with Hive (when using as a warehouse)

2012-06-04 Thread Sreenath Menon
Hi all I am new to hive and am working on analysis of twitter data with Hive and Hadoop in a 27node cluster. At present am using Microsoft powerpivot as the visualization tool for visual representation of analysis done using Hive and have got some really good results and I am stunned by the scalab

Re: Challenge : How to load data of multiple rows into a single row in hive table (can regex be used in this scenario)

2012-05-02 Thread m sreenath
Hi Bejoy     Thanks for the response. Can you provide help on writing a custom serde for the mentioned scenario(or for a general case) and how to include it in hive directory. From: Bejoy Ks To: "user@hive.apache.org" ; m sreenath Sent: Wednes

Challenge : How to load data of multiple rows into a single row in hive table (can regex be used in this scenario)

2012-05-01 Thread m sreenath
Hi I have a large data-set of the form : A    12323:213 B    adasd C    sedweqwd A    12323:213 B    adasd C    sedweqwd The general form is : \t \t \t\n In the given format, I need the part from the three rows in a table containing three columns (time,name,content). Can regex be used for thi