Bah. *similar error.

On 28 September 2014 12:39, Oliver Keyes <oke...@wikimedia.org> wrote:

> I run into that error, or a simpler error, rather a lot; I consider it
> pretty normal (although experiences may differ).
>
> A more direct way of increasing the heapsize would be export
> HADOOP_HEAPSIZE=N, where N is...well, the heapsize. HADOOP_HEAPSIZE=1024
> handles the queries I run, but I don't know how big your data store is and
> how big your operations are.
>
> On 28 September 2014 03:47, sunww <spe...@outlook.com> wrote:
>
>> Hi:
>>    I'm using hive0.11, When I run a sql with 2 union all, hive client
>> throw exception:
>> Exception in thread "main" java.lang.OutOfMemoryError: Java heap space
>> at java.lang.StringCoding$StringEncoder.encode(StringCoding.java:232)
>> at java.lang.StringCoding.encode(StringCoding.java:272)
>> at java.lang.String.getBytes(String.java:946)
>> at
>> org.apache.hadoop.hive.ql.optimizer.physical.CommonJoinResolver$CommonJoinTaskDispatcher.processCurrentTask(CommonJoinResolver.java:589)
>>
>> at
>> org.apache.hadoop.hive.ql.optimizer.physical.CommonJoinResolver$CommonJoinTaskDispatcher.dispatch(CommonJoinResolver.java:743)
>>
>> at
>> org.apache.hadoop.hive.ql.lib.TaskGraphWalker.dispatch(TaskGraphWalker.java:111)
>>
>> at
>> org.apache.hadoop.hive.ql.lib.TaskGraphWalker.walk(TaskGraphWalker.java:194)
>>
>> at
>> org.apache.hadoop.hive.ql.lib.TaskGraphWalker.startWalking(TaskGraphWalker.java:139)
>>
>> at
>> org.apache.hadoop.hive.ql.optimizer.physical.CommonJoinResolver.resolve(CommonJoinResolver.java:112)
>>
>> at
>> org.apache.hadoop.hive.ql.optimizer.physical.PhysicalOptimizer.optimize(PhysicalOptimizer.java:79)
>>
>> at
>> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapRedTasks(SemanticAnalyzer.java:8399)
>>
>> at
>> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:8741)
>>
>> at
>> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:278)
>>
>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:433)
>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:337)
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
>> at
>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:259)
>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:216)
>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:413)
>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:756)
>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> at java.lang.reflect.Method.invoke(Method.java:597)
>> at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
>>
>>
>> I use "export HADOOP_CLIENT_OPTS=-Xmx1024m" to increase hive client
>> heapsize, then the sql is ok.
>> I saw CommonJoinResolver  line 589:
>> InputStream in = new ByteArrayInputStream(xml.getBytes("UTF-8"));
>>
>> hive client need so many heapsize, is this normal
>> And where Can I find this large xml
>>
>>
>> Thanks
>>
>
>
>
> --
> Oliver Keyes
> Research Analyst
> Wikimedia Foundation
>



-- 
Oliver Keyes
Research Analyst
Wikimedia Foundation

Reply via email to