[ 
https://issues.apache.org/jira/browse/HIVE-10837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15123523#comment-15123523
 ] 

David Watzke commented on HIVE-10837:
-------------------------------------

Cloudera recommends increasing the HiveServer2's heap size

http://www.cloudera.com/documentation/enterprise/latest/topics/admin_hos_troubleshooting.html

so that's what I've done (from 256M to 8G) and it seems that it helped.

BTW: I noticed that Cloudera increased the default heap size for hive roles in 
CDH 5.5 which is maybe related to this.

> Running large queries (inserts) fails and crashes hiveserver2
> -------------------------------------------------------------
>
>                 Key: HIVE-10837
>                 URL: https://issues.apache.org/jira/browse/HIVE-10837
>             Project: Hive
>          Issue Type: Bug
>         Environment: Hive 1.1.0 on RHEL with Cloudera (cdh5.4.0)
>            Reporter: Patrick McAnneny
>            Priority: Critical
>
> When running a large insert statement through beeline or pyhs2, a thrift 
> error is returned and hiveserver2 crashes.
> I ran into this with large insert statements -- my initial failing query was 
> around 6million characters. After further testing however it seems like the 
> failure threshold is based on number of inserted rows rather than the query's 
> size in characters. My testing shows the failure threshold between 199,000 
> and 230,000 inserted rows.
> The thrift error is as follows:
> Error: org.apache.thrift.transport.TTransportException: 
> java.net.SocketException: Broken pipe (state=08S01,code=0)
> Also note for anyone that tests this issue - when testing different queries I 
> ran into https://issues.apache.org/jira/browse/HIVE-10836



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to