Seems most likely that the CSV file block is lost or unreadable. Quick
questions?
- What is the dfs.replication in your Hadoop cluster?
- Can you share the Query used for creating the table?
- Can you load the same data from local path rather than from the HDFS?
On Fri, Apr 6, 2012 at 1:44 AM,
Are you adding commons-lang to the classpath? one thing to look out
for is hive includes a section of commons-lang in hive-exec.jar which
can make it problematic to get another commons-lang on the classpath.
Edward
On Thu, Apr 5, 2012 at 5:09 PM, Tucker, Matt wrote:
> I’m trying to convert sanit
I'm trying to convert sanitized HTML data into standard HTML for a query, and
found that the Commons Lang project has existing methods to do this. I
downloaded the jar, with the thought of using the reflect() UDF to call the
unescapeHtml function, but it doesn't work (java.lang.ClassNotFoundExc
Thanks very much . This is great !!
On Apr 5, 2012, at 12:39 PM, Nitin Pawar wrote:
> you can have a common a hiverc which everyone has access to and then alias
> your hive command to use that hiverc
>
> ex: store hiverc in /etc/ with read permission for all along with the udfs
> then alias h
When i am creating table from comma delimited csv file then i am getting below
error:
Failed to open file '/user/manish/minfo.csv': Could not read block
Block(genStamp=1094, blockId=6807603852292275080L, numBytes=44429,
token='AA', startOffset=0, path=u'/user/manish/minfo.csv',
nodes=[Da
you can have a common a hiverc which everyone has access to and then alias
your hive command to use that hiverc
ex: store hiverc in /etc/ with read permission for all along with the udfs
then alias hive command with hive -i /etc/hiverc
On Fri, Apr 6, 2012 at 1:05 AM, Sam William wrote:
> Hi,
>
Use hiverc files https://issues.apache.org/jira/browse/HIVE-1414
Or patch the function into hive.
On Thu, Apr 5, 2012 at 3:35 PM, Sam William wrote:
> Hi,
> I have this external jar with UDFs . I do not want to everyone in the
> company using these functions to run add jar blah.jar; crea
Hi,
I have this external jar with UDFs . I do not want to everyone in the
company using these functions to run add jar blah.jar; create temporary
function blah as 'com.foo.bar.blahblah'; . I want every session to be
preinitialized with these 2 statements ?what are my options ?
Your hive jars in hadoop and restart the hadoop then try.
On Fri, Apr 6, 2012 at 12:49 AM, Carl Steinbach wrote:
> HI Buddhika,
>
> Which version of Hive are you running?
>
> Thanks.
>
> Carl
>
> On Thu, Apr 5, 2012 at 10:12 AM, buddhika chamith > wrote:
>
>> Hi All,
>>
>> I am trying to use hi
HI Buddhika,
Which version of Hive are you running?
Thanks.
Carl
On Thu, Apr 5, 2012 at 10:12 AM, buddhika chamith
wrote:
> Hi All,
>
> I am trying to use hive inside an OSGi enviornment. So far I faced several
> OSGi issues in hive and some issues related to JDO implementation used
> (datanuc
I think the additions you make to the Configuration object in the client
are done after the job.xml is written to disk and sent to the rest of the
cluster, but you could add the external resource URL in hive-site.xml, add
a SET external.resource.url=... in the hive query or on your hiverc, or
pass
Hello Shashwat
Thanks for your reply
Infact yes , its started :)
I thought i would say some message in console
Hi , Thrift has been started :)
Anyways my bad didnt not check it . Thanks Edward and to you also Shashwat :)
- Original Message -
From: shashwat shriparv
Sent: 04/05/12
Hi Shin,
Not sure what do you mean by hangs here. I think this is about what the
server would output on proper startup. (it is at least for me. :)) Did you
try invoking hive sever using hive jdbc client [1] or equivalent? If not
can you try and see whether it works.
Regards
Buddhika
[1] https:/
It does not hangs rather it is started, just open another shell and type
netstat -nl | grep 1 if something shows up means the server is running,
and also in new terminal windows type bin/hive and the start createing your
creating tables.
On Thu, Apr 5, 2012 at 10:18 PM, Shin Chan wrote:
> He
Hello Edward
Thank You for your reply
I have dowloaded hadoop 0.20 version.
Setup hadoop for pesudo mode
exported HADOOP_HOME , HIVE_HOME variables
Now when in start thrift server it still hangs
$hive --service hiveserver -verbose
Starting Hive Thrift Server
Starting hive server on por
I am working on a hive SerDe where both SerDe and RecordReader need to have
access to an external resource with information.
This external resource could be on hdfs, in hbase, or on a http server.
This situation is very similar to what haivvreo does.
The way i go about it right now is that i store
In a nutshell the hive community is trying to catch up with all the
versions of hadoop. 0.20.205, 1.0, 0.23.1, I have lost track myself.
Every time I consider a major like 0.20.205 I find it does not work
with hive. Then I watch issues open up to get hive to work with
0.20.205, before those issues
Hello All
I am trying to start Thrift server on HIVE
My configuration for Hadoop is Pseudo mode
When i say
hive --service hiveserver
It just prints
Starting Hive Thrift Server
WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please use
org.apache.hadoop.log.metrics.E
Can you tell me the size of table b?
If you are doing bucketing and still size b table is huge then it will
reach this problem
On Thu, Apr 5, 2012 at 4:22 PM, binhnt22 wrote:
> Thank Nitin,
>
> ** **
>
> I tried but no luck. Here’s hive log, please spend a little time to view
> it.
>
>
Thank Nitin,
I tried but no luck. Here's hive log, please spend a little time to view it.
hive> set hive.optimize.bucketmapjoin = true;
hive> set hive.enforce.bucketing=true;
hive> set
hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
hive> select /*+ MAPJOIN(b) */ *
can you try adding these settings
set hive.enforce.bucketing=true;
hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
I have tried bucketing with 1000 buckets and with more than 1TB data tables
.. they do go through fine
On Thu, Apr 5, 2012 at 3:37 PM, binhnt22 wrote:
> H
Hi Bejoy,
Both my tables has 65m records ( ~ 1.8-1.9GB on hadoop) and bucketized on
'calling' column into 10 buckets.
As you said, hive will load only 1 bucket ~ 180-190MB into memory. That's
hardly to blow the heap (1.3GB)
According to wiki, I set:
set
hive.input.format=org.apach
Hi Amit
Sorry for the delayed response, had a terrible schedule. AFAIK, there is
no flags that would help you to take the hash table creation, compression and
load into tmp files away from client node.
From my understanding if you use a Map side join, the small table as a
whole is
23 matches
Mail list logo