Re: Hive table creation error

2012-04-05 Thread Ashwanth Kumar
Seems most likely that the CSV file block is lost or unreadable. Quick questions? - What is the dfs.replication in your Hadoop cluster? - Can you share the Query used for creating the table? - Can you load the same data from local path rather than from the HDFS? On Fri, Apr 6, 2012 at 1:44 AM,

Re: Reflect() with Apache Commons Lang

2012-04-05 Thread Edward Capriolo
Are you adding commons-lang to the classpath? one thing to look out for is hive includes a section of commons-lang in hive-exec.jar which can make it problematic to get another commons-lang on the classpath. Edward On Thu, Apr 5, 2012 at 5:09 PM, Tucker, Matt wrote: > I’m trying to convert sanit

Reflect() with Apache Commons Lang

2012-04-05 Thread Tucker, Matt
I'm trying to convert sanitized HTML data into standard HTML for a query, and found that the Commons Lang project has existing methods to do this. I downloaded the jar, with the thought of using the reflect() UDF to call the unescapeHtml function, but it doesn't work (java.lang.ClassNotFoundExc

Re: Initializing hive sessions with Jars/temp functions

2012-04-05 Thread Sam William
Thanks very much . This is great !! On Apr 5, 2012, at 12:39 PM, Nitin Pawar wrote: > you can have a common a hiverc which everyone has access to and then alias > your hive command to use that hiverc > > ex: store hiverc in /etc/ with read permission for all along with the udfs > then alias h

Hive table creation error

2012-04-05 Thread Manish Bhoge
When i am creating table from comma delimited csv file then i am getting below error: Failed to open file '/user/manish/minfo.csv': Could not read block Block(genStamp=1094, blockId=6807603852292275080L, numBytes=44429, token='AA', startOffset=0, path=u'/user/manish/minfo.csv', nodes=[Da

Re: Initializing hive sessions with Jars/temp functions

2012-04-05 Thread Nitin Pawar
you can have a common a hiverc which everyone has access to and then alias your hive command to use that hiverc ex: store hiverc in /etc/ with read permission for all along with the udfs then alias hive command with hive -i /etc/hiverc On Fri, Apr 6, 2012 at 1:05 AM, Sam William wrote: > Hi, >

Re: Initializing hive sessions with Jars/temp functions

2012-04-05 Thread Edward Capriolo
Use hiverc files https://issues.apache.org/jira/browse/HIVE-1414 Or patch the function into hive. On Thu, Apr 5, 2012 at 3:35 PM, Sam William wrote: > Hi, >   I have this external jar with UDFs  .  I do not want to  everyone in the > company using these functions to  run   add jar blah.jar; crea

Initializing hive sessions with Jars/temp functions

2012-04-05 Thread Sam William
Hi, I have this external jar with UDFs . I do not want to everyone in the company using these functions to run add jar blah.jar; create temporary function blah as 'com.foo.bar.blahblah'; . I want every session to be preinitialized with these 2 statements ?what are my options ?

Re: Getting CNF during map reduce execution

2012-04-05 Thread shashwat shriparv
Your hive jars in hadoop and restart the hadoop then try. On Fri, Apr 6, 2012 at 12:49 AM, Carl Steinbach wrote: > HI Buddhika, > > Which version of Hive are you running? > > Thanks. > > Carl > > On Thu, Apr 5, 2012 at 10:12 AM, buddhika chamith > wrote: > >> Hi All, >> >> I am trying to use hi

Re: Getting CNF during map reduce execution

2012-04-05 Thread Carl Steinbach
HI Buddhika, Which version of Hive are you running? Thanks. Carl On Thu, Apr 5, 2012 at 10:12 AM, buddhika chamith wrote: > Hi All, > > I am trying to use hive inside an OSGi enviornment. So far I faced several > OSGi issues in hive and some issues related to JDO implementation used > (datanuc

Re: SerDe loading external scheme

2012-04-05 Thread Roberto Congiu
I think the additions you make to the Configuration object in the client are done after the job.xml is written to disk and sent to the rest of the cluster, but you could add the external resource URL in hive-site.xml, add a SET external.resource.url=... in the hive query or on your hiverc, or pass

Re: Thrift Server not starting with Hive 0.8.1 and hadoop-1.0.1

2012-04-05 Thread Shin Chan
Hello Shashwat Thanks for your reply Infact yes , its started :) I thought i would say some message in console Hi , Thrift has been started :) Anyways my bad didnt not check it . Thanks Edward and to you also Shashwat :) - Original Message - From: shashwat shriparv Sent: 04/05/12

Re: Thrift Server not starting with Hive 0.8.1 and hadoop-1.0.1

2012-04-05 Thread buddhika chamith
Hi Shin, Not sure what do you mean by hangs here. I think this is about what the server would output on proper startup. (it is at least for me. :)) Did you try invoking hive sever using hive jdbc client [1] or equivalent? If not can you try and see whether it works. Regards Buddhika [1] https:/

Re: Thrift Server not starting with Hive 0.8.1 and hadoop-1.0.1

2012-04-05 Thread shashwat shriparv
It does not hangs rather it is started, just open another shell and type netstat -nl | grep 1 if something shows up means the server is running, and also in new terminal windows type bin/hive and the start createing your creating tables. On Thu, Apr 5, 2012 at 10:18 PM, Shin Chan wrote: > He

Re: Thrift Server not starting with Hive 0.8.1 and hadoop-1.0.1

2012-04-05 Thread Shin Chan
Hello Edward Thank You for your reply I have dowloaded hadoop 0.20 version. Setup hadoop for pesudo mode exported HADOOP_HOME , HIVE_HOME variables Now when in start thrift server it still hangs $hive --service hiveserver -verbose Starting Hive Thrift Server Starting hive server on por

SerDe loading external scheme

2012-04-05 Thread Koert Kuipers
I am working on a hive SerDe where both SerDe and RecordReader need to have access to an external resource with information. This external resource could be on hdfs, in hbase, or on a http server. This situation is very similar to what haivvreo does. The way i go about it right now is that i store

Re: Thrift Server not starting with Hive 0.8.1 and hadoop-1.0.1

2012-04-05 Thread Edward Capriolo
In a nutshell the hive community is trying to catch up with all the versions of hadoop. 0.20.205, 1.0, 0.23.1, I have lost track myself. Every time I consider a major like 0.20.205 I find it does not work with hive. Then I watch issues open up to get hive to work with 0.20.205, before those issues

Thrift Server not starting with Hive 0.8.1 and hadoop-1.0.1

2012-04-05 Thread Shin Chan
Hello All I am trying to start Thrift server on HIVE My configuration for Hadoop is Pseudo mode When i say hive --service hiveserver It just prints Starting Hive Thrift Server WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please use org.apache.hadoop.log.metrics.E

Re: Why BucketJoinMap consume too much memory

2012-04-05 Thread Nitin Pawar
Can you tell me the size of table b? If you are doing bucketing and still size b table is huge then it will reach this problem On Thu, Apr 5, 2012 at 4:22 PM, binhnt22 wrote: > Thank Nitin, > > ** ** > > I tried but no luck. Here’s hive log, please spend a little time to view > it. > >

RE: Why BucketJoinMap consume too much memory

2012-04-05 Thread binhnt22
Thank Nitin, I tried but no luck. Here's hive log, please spend a little time to view it. hive> set hive.optimize.bucketmapjoin = true; hive> set hive.enforce.bucketing=true; hive> set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat; hive> select /*+ MAPJOIN(b) */ *

Re: Why BucketJoinMap consume too much memory

2012-04-05 Thread Nitin Pawar
can you try adding these settings set hive.enforce.bucketing=true; hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat; I have tried bucketing with 1000 buckets and with more than 1TB data tables .. they do go through fine On Thu, Apr 5, 2012 at 3:37 PM, binhnt22 wrote: > H

RE: Why BucketJoinMap consume too much memory

2012-04-05 Thread binhnt22
Hi Bejoy, Both my tables has 65m records ( ~ 1.8-1.9GB on hadoop) and bucketized on 'calling' column into 10 buckets. As you said, hive will load only 1 bucket ~ 180-190MB into memory. That's hardly to blow the heap (1.3GB) According to wiki, I set: set hive.input.format=org.apach

Re: Why BucketJoinMap consume too much memory

2012-04-05 Thread Bejoy Ks
Hi Amit       Sorry for the delayed response, had a terrible schedule. AFAIK, there is no flags that would help you to take the hash table creation, compression and load into tmp files away from client node.        From my understanding if you use a Map side join, the small table as a whole is