[ https://issues.apache.org/jira/browse/HIVE-13657?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Mohit Sabharwal updated HIVE-13657: ----------------------------------- Attachment: HIVE-13657.patch > Spark driver stderr logs should appear in hive client logs > ---------------------------------------------------------- > > Key: HIVE-13657 > URL: https://issues.apache.org/jira/browse/HIVE-13657 > Project: Hive > Issue Type: Bug > Reporter: Mohit Sabharwal > Assignee: Mohit Sabharwal > Attachments: HIVE-13657.patch > > > Currently, spark driver exceptions are not getting logged in beeline. > Instead, the users sees the not-so-useful: > {code} > ERROR : Failed to execute spark task, with exception > 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create spark > client.)' > <huge stack trace ommitted> > {code} > The user has to look at HS2 logs to discover the root cause: > {code} > 2015-04-01 11:33:16,048 INFO org.apache.hive.spark.client.SparkClientImpl: > 15/04/01 11:33:16 WARN UserGroupInformation: PriviledgedActionException > as:foo (auth:PROXY) via hive (auth:SIMPLE) > cause:org.apache.hadoop.security.AccessControlException: Permission denied: > user=foo, access=WRITE, inode="/user":hdfs:supergroup:drwxr-xr-x > ... > {code} > We should surface these critical errors in hive client. -- This message was sent by Atlassian JIRA (v6.3.4#6332)