[ https://issues.apache.org/jira/browse/FLINK-30793?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17680915#comment-17680915 ]
Matthias Pohl commented on FLINK-30793: --------------------------------------- I'm linking FLINK-24434 because it has the same stacktrace. But with FLINK-24434 the cause seems to be disk usage of 100% which is not the case here. > PyFlink YARN per-job on Docker test fails on Azure due to permission issues > --------------------------------------------------------------------------- > > Key: FLINK-30793 > URL: https://issues.apache.org/jira/browse/FLINK-30793 > Project: Flink > Issue Type: Bug > Components: Connectors / Hadoop Compatibility, Deployment / YARN > Affects Versions: 1.17.0 > Reporter: Matthias Pohl > Priority: Major > Labels: test-stability > > The following build failed due to some hdfs/yarn permission issues in > PyFlink YARN per-job on Docker e2e test: > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=45202&view=logs&j=af184cdd-c6d8-5084-0b69-7e9c67b35f7a&t=160c9ae5-96fd-516e-1c91-deb81f59292a&l=10587 > {code} > [...] > Jan 26 02:17:31 23/01/26 02:12:20 FATAL hs.JobHistoryServer: Error starting > JobHistoryServer > Jan 26 02:17:31 org.apache.hadoop.yarn.exceptions.YarnRuntimeException: Error > creating done directory: > [hdfs://master.docker-hadoop-cluster-network:9000/tmp/hadoop-yarn/staging/history/done] > Jan 26 02:17:31 at > org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager.tryCreatingHistoryDirs(HistoryFileManager.java:698) > Jan 26 02:17:31 at > org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager.createHistoryDirs(HistoryFileManager.java:634) > Jan 26 02:17:31 at > org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager.serviceInit(HistoryFileManager.java:595) > Jan 26 02:17:31 at > org.apache.hadoop.service.AbstractService.init(AbstractService.java:164) > Jan 26 02:17:31 at > org.apache.hadoop.mapreduce.v2.hs.JobHistory.serviceInit(JobHistory.java:96) > Jan 26 02:17:31 at > org.apache.hadoop.service.AbstractService.init(AbstractService.java:164) > Jan 26 02:17:31 at > org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:108) > Jan 26 02:17:31 at > org.apache.hadoop.mapreduce.v2.hs.JobHistoryServer.serviceInit(JobHistoryServer.java:152) > Jan 26 02:17:31 at > org.apache.hadoop.service.AbstractService.init(AbstractService.java:164) > Jan 26 02:17:31 at > org.apache.hadoop.mapreduce.v2.hs.JobHistoryServer.launchJobHistoryServer(JobHistoryServer.java:228) > Jan 26 02:17:31 at > org.apache.hadoop.mapreduce.v2.hs.JobHistoryServer.main(JobHistoryServer.java:238) > Jan 26 02:17:31 Caused by: org.apache.hadoop.security.AccessControlException: > Permission denied: user=mapred, access=WRITE, inode="/":hdfs:hadoop:drwxr-xr-x > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:350) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:251) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:189) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1756) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1740) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1699) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:60) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3007) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:1141) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:659) > Jan 26 02:17:31 at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > [...] > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)