[ https://issues.apache.org/jira/browse/HIVE-10837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15123523#comment-15123523 ]
David Watzke commented on HIVE-10837: ------------------------------------- Cloudera recommends increasing the HiveServer2's heap size http://www.cloudera.com/documentation/enterprise/latest/topics/admin_hos_troubleshooting.html so that's what I've done (from 256M to 8G) and it seems that it helped. BTW: I noticed that Cloudera increased the default heap size for hive roles in CDH 5.5 which is maybe related to this. > Running large queries (inserts) fails and crashes hiveserver2 > ------------------------------------------------------------- > > Key: HIVE-10837 > URL: https://issues.apache.org/jira/browse/HIVE-10837 > Project: Hive > Issue Type: Bug > Environment: Hive 1.1.0 on RHEL with Cloudera (cdh5.4.0) > Reporter: Patrick McAnneny > Priority: Critical > > When running a large insert statement through beeline or pyhs2, a thrift > error is returned and hiveserver2 crashes. > I ran into this with large insert statements -- my initial failing query was > around 6million characters. After further testing however it seems like the > failure threshold is based on number of inserted rows rather than the query's > size in characters. My testing shows the failure threshold between 199,000 > and 230,000 inserted rows. > The thrift error is as follows: > Error: org.apache.thrift.transport.TTransportException: > java.net.SocketException: Broken pipe (state=08S01,code=0) > Also note for anyone that tests this issue - when testing different queries I > ran into https://issues.apache.org/jira/browse/HIVE-10836 -- This message was sent by Atlassian JIRA (v6.3.4#6332)