maybe relevant (maybe not) https://github.com/apache/zeppelin/pull/1323
Had some issues hitting a cloud HDFS instance a while back- you may be able to hack a solution out relevant to your problem. tg Trevor Grant Data Scientist https://github.com/rawkintrevo http://stackexchange.com/users/3002022/rawkintrevo http://trevorgrant.org *"Fortunate is he, who is able to know the causes of things." -Virgil* On Wed, May 31, 2017 at 12:52 PM, BigData Consultant < bigdata.consultant2...@gmail.com> wrote: > How to set the access for zeppelin user to the HDFS? > > On Tue, May 30, 2017 at 2:16 AM, Felix Cheung <felixcheun...@hotmail.com> > wrote: > >> Seems to be with hdfs ACL - does the service user Zeppelin have access to >> your storage? >> >> ------------------------------ >> *From:* BigData Consultant <bigdata.consultant2...@gmail.com> >> *Sent:* Friday, May 26, 2017 10:56:31 PM >> *To:* dev@zeppelin.apache.org; us...@zeppelin.apache.org >> *Subject:* Permission denied: user=zeppelin while using %spark.pyspark >> interpreter in AWS EMR cluster >> >> Hi Team, >> >> I have created pyspark structure streaming program and trying to execute >> in >> the Zeppelin notebook, I am getting the following error: >> >> Py4JJavaError: An error occurred while calling o191.start. >> : org.apache.hadoop.security.AccessControlException: Permission denied: >> user=zeppelin, access=WRITE, >> inode="/mnt/tmp/temporary-e0cf0f09-a6f4-44d6-9a72-3246600856 >> 08/metadata":hdfs:hadoop:drwxr-xr-x >> >> >> I am using Zeppelin Notebook Version 0.7.1 in AWS EMR cluster. >> >> Help would be much appreciated. >> >> *Full stacktrace:* >> >> >> Traceback (most recent call last): >> File "/tmp/zeppelin_pyspark-8165971491474576109.py", line 349, in >> <module> >> raise Exception(traceback.format_exc()) >> Exception: Traceback (most recent call last): >> File "/tmp/zeppelin_pyspark-8165971491474576109.py", line 342, in >> <module> >> exec(code) >> File "<stdin>", line 5, in <module> >> File "/usr/lib/spark/python/pyspark/sql/streaming.py", line 816, in start >> return self._sq(self._jwrite.start()) >> File "/usr/lib/spark/python/lib/py4j-0.10.4-src.zip/py4j/java_gat >> eway.py", >> line 1133, in __call__ >> answer, self.gateway_client, self.target_id, self.name) >> File "/usr/lib/spark/python/pyspark/sql/utils.py", line 63, in deco >> return f(*a, **kw) >> File "/usr/lib/spark/python/lib/py4j-0.10.4-src.zip/py4j/protocol.py", >> line >> 319, in get_return_value >> format(target_id, ".", name), value) >> Py4JJavaError: An error occurred while calling o191.start. >> : org.apache.hadoop.security.AccessControlException: Permission denied: >> user=zeppelin, access=WRITE, >> inode="/mnt/tmp/temporary-e0cf0f09-a6f4-44d6-9a72-3246600856 >> 08/metadata":hdfs:hadoop:drwxr-xr-x >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c >> heck(FSPermissionChecker.java:320) >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c >> heck(FSPermissionChecker.java:292) >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c >> heckPermission(FSPermissionChecker.java:213) >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c >> heckPermission(FSPermissionChecker.java:190) >> at >> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPerm >> ission(FSDirectory.java:1728) >> at >> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPerm >> ission(FSDirectory.java:1712) >> at >> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAnce >> storAccess(FSDirectory.java:1695) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil >> eInternal(FSNamesystem.java:2515) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil >> eInt(FSNamesystem.java:2450) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil >> e(FSNamesystem.java:2334) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.cre >> ate(NameNodeRpcServer.java:624) >> at >> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ >> erSideTranslatorPB.create(ClientNamenodeProtocolServerSideTr >> anslatorPB.java:397) >> at >> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol >> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam >> enodeProtocolProtos.java) >> at >> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn >> voker.call(ProtobufRpcEngine.java:616) >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:422) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGro >> upInformation.java:1698) >> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2045) >> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) >> at >> sun.reflect.NativeConstructorAccessorImpl.newInstance(Native >> ConstructorAccessorImpl.java:62) >> at >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(De >> legatingConstructorAccessorImpl.java:45) >> at java.lang.reflect.Constructor.newInstance(Constructor.java:423) >> at >> org.apache.hadoop.ipc.RemoteException.instantiateException(R >> emoteException.java:106) >> at >> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException( >> RemoteException.java:73) >> at >> org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DF >> SOutputStream.java:1653) >> at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1689) >> at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1624) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall( >> DistributedFileSystem.java:448) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall( >> DistributedFileSystem.java:444) >> at >> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSyst >> emLinkResolver.java:81) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.create(Distribu >> tedFileSystem.java:459) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.create(Distribu >> tedFileSystem.java:387) >> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:915) >> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:896) >> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:793) >> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:782) >> at >> org.apache.spark.sql.execution.streaming.StreamMetadata$. >> write(StreamMetadata.scala:76) >> at >> org.apache.spark.sql.execution.streaming.StreamExecution$$ >> anonfun$5.apply(StreamExecution.scala:102) >> at >> org.apache.spark.sql.execution.streaming.StreamExecution$$ >> anonfun$5.apply(StreamExecution.scala:100) >> at scala.Option.getOrElse(Option.scala:121) >> at >> org.apache.spark.sql.execution.streaming.StreamExecution.< >> init>(StreamExecution.scala:100) >> at >> org.apache.spark.sql.streaming.StreamingQueryManager.createQ >> uery(StreamingQueryManager.scala:232) >> at >> org.apache.spark.sql.streaming.StreamingQueryManager.startQu >> ery(StreamingQueryManager.scala:269) >> at >> org.apache.spark.sql.streaming.DataStreamWriter.start( >> DataStreamWriter.scala:227) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >> ssorImpl.java:62) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >> thodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) >> at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) >> at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) >> at py4j.Gateway.invoke(Gateway.java:280) >> at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) >> at py4j.commands.CallCommand.execute(CallCommand.java:79) >> at py4j.GatewayConnection.run(GatewayConnection.java:214) >> at java.lang.Thread.run(Thread.java:745) >> Caused by: >> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop. >> security.AccessControlException): >> Permission denied: user=zeppelin, access=WRITE, >> inode="/mnt/tmp/temporary-e0cf0f09-a6f4-44d6-9a72-3246600856 >> 08/metadata":hdfs:hadoop:drwxr-xr-x >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c >> heck(FSPermissionChecker.java:320) >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c >> heck(FSPermissionChecker.java:292) >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c >> heckPermission(FSPermissionChecker.java:213) >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.c >> heckPermission(FSPermissionChecker.java:190) >> at >> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPerm >> ission(FSDirectory.java:1728) >> at >> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPerm >> ission(FSDirectory.java:1712) >> at >> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAnce >> storAccess(FSDirectory.java:1695) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil >> eInternal(FSNamesystem.java:2515) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil >> eInt(FSNamesystem.java:2450) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFil >> e(FSNamesystem.java:2334) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.cre >> ate(NameNodeRpcServer.java:624) >> at >> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ >> erSideTranslatorPB.create(ClientNamenodeProtocolServerSideTr >> anslatorPB.java:397) >> at >> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol >> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam >> enodeProtocolProtos.java) >> at >> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn >> voker.call(ProtobufRpcEngine.java:616) >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:422) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGro >> upInformation.java:1698) >> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2045) >> at org.apache.hadoop.ipc.Client.call(Client.java:1475) >> at org.apache.hadoop.ipc.Client.call(Client.java:1412) >> at >> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke( >> ProtobufRpcEngine.java:229) >> at com.sun.proxy.$Proxy12.create(Unknown Source) >> at >> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTran >> slatorPB.create(ClientNamenodeProtocolTranslatorPB.java:296) >> at sun.reflect.GeneratedMethodAccessor14.invoke(Unknown Source) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >> thodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) >> at >> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMeth >> od(RetryInvocationHandler.java:191) >> at >> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(Ret >> ryInvocationHandler.java:102) >> at com.sun.proxy.$Proxy13.create(Unknown Source) >> at >> org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DF >> SOutputStream.java:1648) >> ... 30 more >> > >