I'm not an expert on S3. If it is not a credential issue, have you finish the checklist of this doc[1]?
[1] https://aws.amazon.com/premiumsupport/knowledge-center/emr-s3-403-access-denied/?nc1=h_ls Best, Yangze Guo On Wed, Sep 22, 2021 at 3:39 PM Dhiru <userdh...@yahoo.com> wrote: > > > Not sure @yangze ... but other services which are deployed in same places we > are able to access s3 bucket, the link which you share are recommended way, > if we have access to s3 then we should not pass credentials ? > > On Wednesday, September 22, 2021, 02:59:05 AM EDT, Yangze Guo > <karma...@gmail.com> wrote: > > > You might need to configure the access credential. [1] > > [1] > https://ci.apache.org/projects/flink/flink-docs-master/docs/deployment/filesystems/s3/#configure-access-credentials > > Best, > Yangze Guo > > On Wed, Sep 22, 2021 at 2:17 PM Dhiru <userdh...@yahoo.com> wrote: > > > > > > i see org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:2326) plugin > > is not able to create folder , not sure if I need to change something > > Whereas when We are trying to pass from the local laptop and passing aws > > credentails its able to create a folder and running as expected > > On Wednesday, September 22, 2021, 01:39:04 AM EDT, Dhiru > > <userdh...@yahoo.com> wrote: > > > > > > flink image I have added both s3 plugin > > FROM flink:1.11.3-scala_2.12-java11 > > RUN mkdir ./plugins/flink-s3-fs-presto > > RUN cp ./opt/flink-s3-fs-presto-1.11.3.jar ./plugins/flink-s3-fs-presto/ > > RUN mkdir ./plugins/flink-s3-fs-hadoop > > RUN cp ./opt/flink-s3-fs-hadoop-1.11.3.jar ./plugins/flink-s3-fs-hadoop/ > > > > some part of flink-conf.yaml ( I tried with both s3a and s3 ) > > # REQUIRED: set storage location for job metadata in remote storage > > state.backend: filesystem > > state.backend.fs.checkpointdir: > > s3a://msc-actigraph-test-bucket/flink-checkpointing/checkpoints > > state.checkpoints.dir: > > s3a://msc-actigraph-test-bucket/flink-checkpointing/externalized-checkpoints > > state.savepoints.dir: > > s3a://msc-actigraph-test-bucket/flink-checkpointing/savepoints > > high-availability.storageDir: > > s3a://msc-actigraph-test-bucket/flink-checkpointing/storagedir > > s3.path.style.access: true > > > > org.apache.flink.runtime.rest.handler.RestHandlerException: Could not > > execute application. at > > org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$1(JarRunHandler.java:103) > > at java.base/java.util.concurrent.CompletableFuture.uniHandle(Unknown > > Source) at > > java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(Unknown > > Source) at > > java.base/java.util.concurrent.CompletableFuture.postComplete(Unknown > > Source) at > > java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(Unknown > > Source) at > > java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown > > Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at > > java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown > > Source) at > > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) > > at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown > > Source) at java.base/java.lang.Thread.run(Unknown Source) Caused by: > > java.util.concurrent.CompletionException: > > org.apache.flink.util.FlinkRuntimeException: Could not execute application. > > at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(Unknown > > Source) at > > java.base/java.util.concurrent.CompletableFuture.completeThrowable(Unknown > > Source) ... 7 more Caused by: org.apache.flink.util.FlinkRuntimeException: > > Could not execute application. at > > org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:81) > > at > > org.apache.flink.client.deployment.application.DetachedApplicationRunner.run(DetachedApplicationRunner.java:67) > > at > > org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$0(JarRunHandler.java:100) > > ... 7 more Caused by: > > org.apache.flink.client.program.ProgramInvocationException: The main method > > caused an error: Failed to execute job 'DeduplicationJob'. at > > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:302) > > at > > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:198) > > at > > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:149) at > > org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:78) > > ... 9 more Caused by: org.apache.flink.util.FlinkException: Failed to > > execute job 'DeduplicationJob'. at > > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1829) > > at > > org.apache.flink.client.program.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:128) > > at > > org.apache.flink.client.program.StreamContextEnvironment.execute(StreamContextEnvironment.java:76) > > at > > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1700) > > at > > org.apache.flink.streaming.api.scala.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.scala:699) > > at io.epiphanous.flinkrunner.flink.BaseFlinkJob.run(BaseFlinkJob.scala:45) > > at io.epiphanous.flinkrunner.FlinkRunner.process1(FlinkRunner.scala:56) at > > io.epiphanous.flinkrunner.FlinkRunner.process(FlinkRunner.scala:33) at > > com.mdsol.flink.delivery_streams.Runner$.run(Runner.scala:25) at > > com.mdsol.flink.delivery_streams.Runner$.main(Runner.scala:7) at > > com.mdsol.flink.delivery_streams.Runner.main(Runner.scala) at > > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > > Method) at > > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown > > Source) at > > java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown > > Source) at java.base/java.lang.reflect.Method.invoke(Unknown Source) at > > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:288) > > ... 12 more Caused by: > > org.apache.flink.runtime.client.JobSubmissionException: Failed to submit > > job. at > > org.apache.flink.runtime.dispatcher.Dispatcher.lambda$internalSubmitJob$3(Dispatcher.java:362) > > at java.base/java.util.concurrent.CompletableFuture.uniHandle(Unknown > > Source) at > > java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(Unknown > > Source) at > > java.base/java.util.concurrent.CompletableFuture$Completion.run(Unknown > > Source) at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40) at > > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:44) > > at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at > > akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > > at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > > at > > akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > Caused by: org.apache.flink.runtime.client.JobExecutionException: Could > > not instantiate JobManager. at > > org.apache.flink.runtime.dispatcher.Dispatcher.lambda$createJobManagerRunner$6(Dispatcher.java:427) > > at > > java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(Unknown > > Source) ... 6 more Caused by: org.apache.flink.util.FlinkRuntimeException: > > Failed to create checkpoint storage at checkpoint coordinator side. at > > org.apache.flink.runtime.checkpoint.CheckpointCoordinator.(CheckpointCoordinator.java:307) > > at > > org.apache.flink.runtime.checkpoint.CheckpointCoordinator.(CheckpointCoordinator.java:226) > > at > > org.apache.flink.runtime.executiongraph.ExecutionGraph.enableCheckpointing(ExecutionGraph.java:483) > > at > > org.apache.flink.runtime.executiongraph.ExecutionGraphBuilder.buildGraph(ExecutionGraphBuilder.java:338) > > at > > org.apache.flink.runtime.scheduler.SchedulerBase.createExecutionGraph(SchedulerBase.java:270) > > at > > org.apache.flink.runtime.scheduler.SchedulerBase.createAndRestoreExecutionGraph(SchedulerBase.java:244) > > at > > org.apache.flink.runtime.scheduler.SchedulerBase.(SchedulerBase.java:231) > > at > > org.apache.flink.runtime.scheduler.DefaultScheduler.(DefaultScheduler.java:119) > > at > > org.apache.flink.runtime.scheduler.DefaultSchedulerFactory.createInstance(DefaultSchedulerFactory.java:103) > > at > > org.apache.flink.runtime.jobmaster.JobMaster.createScheduler(JobMaster.java:290) > > at org.apache.flink.runtime.jobmaster.JobMaster.(JobMaster.java:278) at > > org.apache.flink.runtime.jobmaster.factories.DefaultJobMasterServiceFactory.createJobMasterService(DefaultJobMasterServiceFactory.java:98) > > at > > org.apache.flink.runtime.jobmaster.factories.DefaultJobMasterServiceFactory.createJobMasterService(DefaultJobMasterServiceFactory.java:40) > > at > > org.apache.flink.runtime.jobmaster.JobManagerRunnerImpl.(JobManagerRunnerImpl.java:140) > > at > > org.apache.flink.runtime.dispatcher.DefaultJobManagerRunnerFactory.createJobManagerRunner(DefaultJobManagerRunnerFactory.java:84) > > at > > org.apache.flink.runtime.dispatcher.Dispatcher.lambda$createJobManagerRunner$6(Dispatcher.java:417) > > ... 7 more Caused by: java.nio.file.AccessDeniedException: > > s3a://msc-actigraph-test-bucket/flink-checkpointing/b3b19b338b6c2b2bf022b219051d6d1a/shared: > > getFileStatus on > > s3a://msc-actigraph-test-bucket/flink-checkpointing/b3b19b338b6c2b2bf022b219051d6d1a/shared: > > com.amazonaws.services.s3.model.AmazonS3Exception: Forbidden (Service: > > Amazon S3; Status Code: 403; Error Code: 403 Forbidden; Request ID: > > B99FWA0KE3JYKGN9; S3 Extended Request ID: > > OL+2LER+8Pofuv5sM7G6qvjHiTzmSce66URzMRvlw4VwkS4jKeU2/INZRj9UEAsGASjl8Ohn3OE=), > > S3 Extended Request ID: > > OL+2LER+8Pofuv5sM7G6qvjHiTzmSce66URzMRvlw4VwkS4jKeU2/INZRj9UEAsGASjl8Ohn3OE=:403 > > Forbidden at > > org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:218) at > > org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:145) at > > org.apache.hadoop.fs.s3a.S3AFileSystem.s3GetFileStatus(S3AFileSystem.java:2184) > > at > > org.apache.hadoop.fs.s3a.S3AFileSystem.innerGetFileStatus(S3AFileSystem.java:2149) > > at > > org.apache.hadoop.fs.s3a.S3AFileSystem.getFileStatus(S3AFileSystem.java:2088) > > at > > org.apache.hadoop.fs.s3a.S3AFileSystem.innerMkdirs(S3AFileSystem.java:2037) > > at org.apache.hadoop.fs.s3a.S3AFileSystem.mkdirs(S3AFileSystem.java:2007) > > at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:2326) at > > org.apache.flink.fs.s3hadoop.common.HadoopFileSystem.mkdirs(HadoopFileSystem.java:170) > > at > > org.apache.flink.core.fs.PluginFileSystemFactory$ClassLoaderFixingFileSystem.mkdirs(PluginFileSystemFactory.java:162) > > at > > org.apache.flink.runtime.state.filesystem.FsCheckpointStorage.initializeBaseLocations(FsCheckpointStorage.java:111) > > at > > org.apache.flink.runtime.checkpoint.CheckpointCoordinator.(CheckpointCoordinator.java:305) > > ... 22 more Caused by: com.amazonaws.services.s3.model.AmazonS3Exception: > > Forbidden (Service: Amazon S3; Status Code: 403; Error Code: 403 Forbidden; > > Request ID: B99FWA0KE3JYKGN9; S3 Extended Request ID: > > OL+2LER+8Pofuv5sM7G6qvjHiTzmSce66URzMRvlw4VwkS4jKeU2/INZRj9UEAsGASjl8Ohn3OE=), > > S3 Extended Request ID: > > OL+2LER+8Pofuv5sM7G6qvjHiTzmSce66URzMRvlw4VwkS4jKeU2/INZRj9UEAsGASjl8Ohn3OE= > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutor.handleErrorResponse(AmazonHttpClient.java:1799) > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutor.handleServiceErrorResponse(AmazonHttpClient.java:1383) > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeOneRequest(AmazonHttpClient.java:1359) > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1139) > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:796) > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:764) > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:738) > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:698) > > at > > com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:680) > > at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:544) > > at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:524) > > at > > com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:5054) > > at > > com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:5000) > > at > > com.amazonaws.services.s3.AmazonS3Client.getObjectMetadata(AmazonS3Client.java:1335) > > at > > org.apache.hadoop.fs.s3a.S3AFileSystem.lambda$getObjectMetadata$4(S3AFileSystem.java:1235) > > at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:317) at > > org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:280) at > > org.apache.hadoop.fs.s3a.S3AFileSystem.getObjectMetadata(S3AFileSystem.java:1232) > > at > > org.apache.hadoop.fs.s3a.S3AFileSystem.s3GetFileStatus(S3AFileSystem.java:2169) > > ... 31 more > > > > > > I have tried with some other sample code, I am able to read/write/create > > folder access of s3 bucket , but when I try from flink I see some new > > folder try to create > > s3a://msc-actigraph-test-bucket/flink-checkpointing/b3b19b338b6c2b2bf022b219051d6d1a/ > > > > do not see (b3b19b338b6c2b2bf022b219051d6d1a) folder created to s3, I am > > not sure if this is the right way we are trying to write to the s3 bucket > > using flink ? > > > >