Mike Zhao created ZEPPELIN-4180: ----------------------------------- Summary: Error running hive paragraph due to lost credential Key: ZEPPELIN-4180 URL: https://issues.apache.org/jira/browse/ZEPPELIN-4180 Project: Zeppelin Issue Type: Bug Components: zeppelin-interpreter, zeppelin-server Affects Versions: 0.8.0 Environment: *Server:*
Zeppelin 0.8.0 on RHEL 6.9 2.6.32-754.9.1.el6.x86_64 *Client:* |*Google Chrome*|74.0.3729.169 (Official Build) (32-bit) (cohort: Stable)| |*OS*|Windows 7 Service Pack 1 Build 7601.24443| Reporter: Mike Zhao User has random but frequent issue when running hive paragraph due to sudden loss of credential although user didn't change credential in the user profile. Trying to update user credential will come to the following error: {{Username \ Entity can not be empty.}} The workaround now is to: * Restart Zeppelin * Remove the jdbc.hive credential from the user profile * Add the jdbc.hive credential in the user profile The Error: {{INFO [2019-06-05 09:13:51,029] (\{pool-2-thread-55} SchedulerFactory.java[jobStarted]:109) - Job 20190520-132926_1580001585 started by scheduler org.apache.zeppelin.interpreter.remote.RemoteInterpreter-hive:d286131:-shared_session}} {{ INFO [2019-06-05 09:13:51,031] (\{pool-2-thread-55} Paragraph.java[jobRun]:380) - Run paragraph [paragraph_id: 20190520-132926_1580001585, interpreter: , note_id: 2EDZEYDMU, user: d286131]}} {{ WARN [2019-06-05 09:13:51,694] (\{pool-2-thread-55} NotebookServer.java[afterStatusChange]:2302) - Job 20190520-132926_1580001585 is finished, status: ERROR, exception: null, result: %text org.apache.hive.service.cli.HiveSQLException: Error while compiling statement: FAILED: SemanticException Unable to fetch table s_order. org.apache.hadoop.ipc.RemoteException(org.apache.ranger.authorization.hadoop.exceptions.RangerAccessControlException): Permission denied: user=anonymous, access=EXECUTE, inode="/data/prod/historic/rcrm/s_order"}} {{ at org.apache.ranger.authorization.hadoop.RangerHdfsAuthorizer$RangerAccessControlEnforcer.checkPermission(RangerHdfsAuthorizer.java:383)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1950)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:108)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:4146)}} {{ at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1137)}} {{ at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:866)}} {{ at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)}} {{ at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)}} {{ at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)}} {{ at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351)}} {{ at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347)}} {{ at java.security.AccessController.doPrivileged(Native Method)}} {{ at javax.security.auth.Subject.doAs(Subject.java:422)}} {{ at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)}} {{ at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2347)}}{{at org.apache.hive.jdbc.Utils.verifySuccess(Utils.java:279)}} {{ at org.apache.hive.jdbc.Utils.verifySuccessWithInfo(Utils.java:265)}} {{ at org.apache.hive.jdbc.HiveStatement.runAsyncOnServer(HiveStatement.java:303)}} {{ at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:244)}} {{ at org.apache.commons.dbcp2.DelegatingStatement.execute(DelegatingStatement.java:291)}} {{at org.apache.commons.dbcp2.DelegatingStatement.execute(DelegatingStatement.java:291)}} {{ at org.apache.zeppelin.jdbc.JDBCInterpreter.executeSql(JDBCInterpreter.java:737)}} {{ at org.apache.zeppelin.jdbc.JDBCInterpreter.interpret(JDBCInterpreter.java:820)}} {{ at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:103)}} {{ at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:633)}} {{ at org.apache.zeppelin.scheduler.Job.run(Job.java:188)}} {{ at org.apache.zeppelin.scheduler.ParallelScheduler$JobRunner.run(ParallelScheduler.java:162)}} {{ at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)}} {{ at java.util.concurrent.FutureTask.run(FutureTask.java:266)}} {{ at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)}} {{ at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)}} {{ at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)}} {{ at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)}} {{ at java.lang.Thread.run(Thread.java:745)}} {{Caused by: org.apache.hive.service.cli.HiveSQLException: Error while compiling statement: FAILED: SemanticException Unable to fetch table s_order. org.apache.hadoop.ipc.RemoteException(org.apache.ranger.authorization.hadoop.exceptions.RangerAccessControlException): Permission denied: user=anonymous, access=EXECUTE, inode="/data/prod/historic/rcrm/s_order"}} {{ at org.apache.ranger.authorization.hadoop.RangerHdfsAuthorizer$RangerAccessControlEnforcer.checkPermission(RangerHdfsAuthorizer.java:383)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1950)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:108)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:4146)}} {{ at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1137)}} {{ at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:866)}} {{ at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)}} {{ at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)}} {{ at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)}} {{ at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351)}} {{ at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347)}} {{ at java.security.AccessController.doPrivileged(Native Method)}} {{ at javax.security.auth.Subject.doAs(Subject.java:422)}} {{ at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)}} {{ at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2347)}}{{at org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:324)}} {{ at org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:148)}} {{ at org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:228)}} {{ at org.apache.hive.service.cli.operation.Operation.run(Operation.java:264)}} {{ at org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:479)}} {{ at org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:466)}} {{ at sun.reflect.GeneratedMethodAccessor105.invoke(Unknown Source)}} {{ at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)}} {{ at java.lang.reflect.Method.invoke(Method.java:498)}} {{ at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)}} {{ at org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)}} {{ at org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)}} {{ at java.security.AccessController.doPrivileged(Native Method)}} {{ at javax.security.auth.Subject.doAs(Subject.java:422)}} {{ at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)}} {{ at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)}} {{ at com.sun.proxy.$Proxy46.executeStatementAsync(Unknown Source)}} {{ at org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:315)}} {{ at org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:509)}} {{ at org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1377)}} {{ at org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1362)}} {{ at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)}} {{ at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)}} {{ at org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)}} {{ at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)}} {{ ... 3 more}} {{Caused by: java.lang.RuntimeException: org.apache.hadoop.hive.ql.parse.SemanticException:Unable to fetch table s_order. org.apache.hadoop.ipc.RemoteException(org.apache.ranger.authorization.hadoop.exceptions.RangerAccessControlException): Permission denied: user=anonymous, access=EXECUTE, inode="/data/prod/historic/rcrm/s_order"}} {{ at org.apache.ranger.authorization.hadoop.RangerHdfsAuthorizer$RangerAccessControlEnforcer.checkPermission(RangerHdfsAuthorizer.java:383)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1950)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:108)}} {{ at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:4146)}} {{ at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1137)}} {{ at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:866)}} {{ at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)}} {{ at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)}} {{ at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)}} {{ at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351)}} {{ at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347)}} {{ at java.security.AccessController.doPrivileged(Native Method)}} {{ at javax.security.auth.Subject.doAs(Subject.java:422)}} {{ at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)}} {{ at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2347)}} -- This message was sent by Atlassian JIRA (v7.6.3#76005)