Hi Alexandre, This seems to be complaining about the python script loading. It seems that the local file system is using `file` file prefix not `local`[1]. FYI inside your python script you can add more dependencies like connectors and so using python dependency management[2] which differs from Java's maven/gradle dependency management systems.
1- https://nightlies.apache.org/flink/flink-docs-release-1.20/docs/deployment/filesystems/overview/#local-file-system 2- https://nightlies.apache.org/flink/flink-docs-master/docs/dev/python/dependency_management Best Regards Ahmed Hamdy On Mon, 2 Sept 2024 at 15:42, Alexandre KY <alexandre...@magellium.fr> wrote: > Hello, > > I am trying to deploy my application on Amazon EMR on EKS. My application > is in Python and from what I have read, I must create my own custom image. > I tried to run a simple test with a very simple Flink job. Here is the > Flink application: > > ```python > # main.py > > import logging > import sys > > from pyflink.datastream import StreamExecutionEnvironment > > from pyflink_proto.docker.aws.test.Person import Person > > > def main(): > env = StreamExecutionEnvironment.get_execution_environment() > test = [ > Person("Fred", 35), > Person("Wilma", 35), > Person("Pebbles", 2) > ] > > flintstones = env.from_collection(test) > > adults = flintstones.filter( > lambda p: p.age >= 18 > ) > > adults.print() > > # submit for execution > env.execute() > > > if __name__ == '__main__': > logging.basicConfig(stream=sys.stdout, level=logging.INFO, > format="%(message)s") > > main() > ``` > > ```python > # Person.py > > class Person: > > def __init__(self, name, age): > self.name = name > self.age = age > > def __str__(self): > return '{}: age {}'.format(self.name, str(self.age)) > ``` > > Here is the Dockerfile used: > > ```dockerfile > FROM public.ecr.aws/emr-on-eks/flink/emr-7.0.0-flink:latest > USER root > RUN pip3 install numpy apache-flink==1.18.1 > USER hadoop:hadoop > ``` > > When I try to execute the Flink job in application mode using native > Kubernetes: > > ```shell > $FLINK_HOME/bin/flink run-application \ > --target kubernetes-application \ > -Dkubernetes.namespace=$NAMESPACE \ > -Dkubernetes.cluster-id=$EMR_CLUSTER \ > -Dkubernetes.container.image.ref=$IMAGE \ > -Dkubernetes.service-account=$FLINK_SERVICE_ACCOUNT \ > -Djobmanager.heap.size=1024m \ > -Dtaskmanager.memory.process.size=2048m \ > -py local:///opt/flink_job/test.py > ``` > > I get the following error: > > ```shell > ERROR org.apache.flink.client.python.PythonDriver [] - > Run python process failed > org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not > find a file system implementation for scheme 'local'. The scheme is not > directly supported by Flink and no Hadoop file system to support this > scheme could be loaded. For a full list of supported file systems, please > see https://nightlies.apache.org/flink/flink-docs-stable/ops/filesystems/. > at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:543) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:409) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.Path.getFileSystem(Path.java:274) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.python.PythonEnvUtils.addToPythonPath(PythonEnvUtils.java:291) > ~[flink-python-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.python.PythonEnvUtils.preparePythonEnvironment(PythonEnvUtils.java:226) > ~[flink-python-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.python.PythonEnvUtils.launchPy4jPythonClient(PythonEnvUtils.java:487) > ~[flink-python-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:92) > ~[flink-python-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) ~[?:?] > at > jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:?] > at > jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:?] > at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:301) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.lambda$runApplicationAsync$2(ApplicationDispatcherBootstrap.java:254) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) > ~[?:?] > at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] > at > org.apache.flink.runtime.concurrent.pekko.ActorSystemScheduledExecutorAdapter$ScheduledFutureTask.run(ActorSystemScheduledExecutorAdapter.java:172) > ~[?:?] > at > org.apache.flink.runtime.concurrent.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:68) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.runtime.concurrent.ClassLoadingUtils.lambda$withContextClassLoader$0(ClassLoadingUtils.java:41) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.pekko.dispatch.TaskInvocation.run(AbstractDispatcher.scala:59) > [flink-rpc-akka73f86dfe-dc62-4aea-8884-6e4fccebb026.jar:1.18.0-amzn-0] > at > org.apache.pekko.dispatch.ForkJoinExecutorConfigurator$PekkoForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:57) > [flink-rpc-akka73f86dfe-dc62-4aea-8884-6e4fccebb026.jar:1.18.0-amzn-0] > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) > [?:?] > at > java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) > [?:?] > at java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) [?:?] > at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) > [?:?] > at > java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183) > [?:?] > Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: > Hadoop is not in the classpath/dependencies. > at > org.apache.flink.core.fs.UnsupportedSchemeFactory.create(UnsupportedSchemeFactory.java:55) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:526) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > ... 27 more > 2024-09-02 12:31:06,606 WARN > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap > [] - Application failed unexpectedly: > java.util.concurrent.CompletionException: > org.apache.flink.client.deployment.application.ApplicationExecutionException: > Could not execute application. > at > java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) > ~[?:?] > at > java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) > ~[?:?] > at > java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) > ~[?:?] > at > java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) > ~[?:?] > at > java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2094) > ~[?:?] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:337) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.lambda$runApplicationAsync$2(ApplicationDispatcherBootstrap.java:254) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) > ~[?:?] > at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] > at > org.apache.flink.runtime.concurrent.pekko.ActorSystemScheduledExecutorAdapter$ScheduledFutureTask.run(ActorSystemScheduledExecutorAdapter.java:172) > ~[?:?] > at > org.apache.flink.runtime.concurrent.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:68) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.runtime.concurrent.ClassLoadingUtils.lambda$withContextClassLoader$0(ClassLoadingUtils.java:41) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.pekko.dispatch.TaskInvocation.run(AbstractDispatcher.scala:59) > [flink-rpc-akka73f86dfe-dc62-4aea-8884-6e4fccebb026.jar:1.18.0-amzn-0] > at > org.apache.pekko.dispatch.ForkJoinExecutorConfigurator$PekkoForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:57) > [flink-rpc-akka73f86dfe-dc62-4aea-8884-6e4fccebb026.jar:1.18.0-amzn-0] > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) > [?:?] > at > java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) > [?:?] > at java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) [?:?] > at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) > [?:?] > at > java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183) > [?:?] > Caused by: > org.apache.flink.client.deployment.application.ApplicationExecutionException: > Could not execute application. > ... 14 more > Caused by: org.apache.flink.client.program.ProgramAbortException: > org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not > find a file system implementation for scheme 'local'. The scheme is not > directly supported by Flink and no Hadoop file system to support this > scheme could be loaded. For a full list of supported file systems, please > see https://nightlies.apache.org/flink/flink-docs-stable/ops/filesystems/. > at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:140) > ~[?:?] > at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) ~[?:?] > at > jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:?] > at > jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:?] > at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:301) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > ... 13 more > Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: > Could not find a file system implementation for scheme 'local'. The scheme > is not directly supported by Flink and no Hadoop file system to support > this scheme could be loaded. For a full list of supported file systems, > please see > https://nightlies.apache.org/flink/flink-docs-stable/ops/filesystems/. > at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:543) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:409) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.Path.getFileSystem(Path.java:274) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.python.PythonEnvUtils.addToPythonPath(PythonEnvUtils.java:291) > ~[?:?] > at > org.apache.flink.client.python.PythonEnvUtils.preparePythonEnvironment(PythonEnvUtils.java:226) > ~[?:?] > at > org.apache.flink.client.python.PythonEnvUtils.launchPy4jPythonClient(PythonEnvUtils.java:487) > ~[?:?] > at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:92) > ~[?:?] > at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) ~[?:?] > at > jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:?] > at > jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:?] > at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:301) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > ... 13 more > Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: > Hadoop is not in the classpath/dependencies. > at > org.apache.flink.core.fs.UnsupportedSchemeFactory.create(UnsupportedSchemeFactory.java:55) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:526) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:409) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.Path.getFileSystem(Path.java:274) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.python.PythonEnvUtils.addToPythonPath(PythonEnvUtils.java:291) > ~[?:?] > at > org.apache.flink.client.python.PythonEnvUtils.preparePythonEnvironment(PythonEnvUtils.java:226) > ~[?:?] > at > org.apache.flink.client.python.PythonEnvUtils.launchPy4jPythonClient(PythonEnvUtils.java:487) > ~[?:?] > at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:92) > ~[?:?] > at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) ~[?:?] > at > jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:?] > at > jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:?] > at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:301) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > ... 13 more > 2024-09-02 12:31:06,615 ERROR > org.apache.flink.runtime.entrypoint.ClusterEntrypoint [] - Fatal > error occurred in the cluster entrypoint. > java.util.concurrent.CompletionException: > org.apache.flink.client.deployment.application.ApplicationExecutionException: > Could not execute application. > at > java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) > ~[?:?] > at > java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) > ~[?:?] > at > java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) > ~[?:?] > at > java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) > ~[?:?] > at > java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2094) > ~[?:?] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:337) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.lambda$runApplicationAsync$2(ApplicationDispatcherBootstrap.java:254) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) > ~[?:?] > at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] > at > org.apache.flink.runtime.concurrent.pekko.ActorSystemScheduledExecutorAdapter$ScheduledFutureTask.run(ActorSystemScheduledExecutorAdapter.java:172) > ~[?:?] > at > org.apache.flink.runtime.concurrent.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:68) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.runtime.concurrent.ClassLoadingUtils.lambda$withContextClassLoader$0(ClassLoadingUtils.java:41) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.pekko.dispatch.TaskInvocation.run(AbstractDispatcher.scala:59) > [flink-rpc-akka73f86dfe-dc62-4aea-8884-6e4fccebb026.jar:1.18.0-amzn-0] > at > org.apache.pekko.dispatch.ForkJoinExecutorConfigurator$PekkoForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:57) > [flink-rpc-akka73f86dfe-dc62-4aea-8884-6e4fccebb026.jar:1.18.0-amzn-0] > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) > [?:?] > at > java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) > [?:?] > at java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) [?:?] > at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) > [?:?] > at > java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183) > [?:?] > Caused by: > org.apache.flink.client.deployment.application.ApplicationExecutionException: > Could not execute application. > ... 14 more > Caused by: org.apache.flink.client.program.ProgramAbortException: > org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not > find a file system implementation for scheme 'local'. The scheme is not > directly supported by Flink and no Hadoop file system to support this > scheme could be loaded. For a full list of supported file systems, please > see https://nightlies.apache.org/flink/flink-docs-stable/ops/filesystems/. > at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:140) > ~[?:?] > at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) ~[?:?] > at > jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:?] > at > jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:?] > at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:301) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > ... 13 more > Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: > Could not find a file system implementation for scheme 'local'. The scheme > is not directly supported by Flink and no Hadoop file system to support > this scheme could be loaded. For a full list of supported file systems, > please see > https://nightlies.apache.org/flink/flink-docs-stable/ops/filesystems/. > at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:543) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:409) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.Path.getFileSystem(Path.java:274) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.python.PythonEnvUtils.addToPythonPath(PythonEnvUtils.java:291) > ~[?:?] > at > org.apache.flink.client.python.PythonEnvUtils.preparePythonEnvironment(PythonEnvUtils.java:226) > ~[?:?] > at > org.apache.flink.client.python.PythonEnvUtils.launchPy4jPythonClient(PythonEnvUtils.java:487) > ~[?:?] > at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:92) > ~[?:?] > at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) ~[?:?] > at > jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:?] > at > jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:?] > at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:301) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > ... 13 more > Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: > Hadoop is not in the classpath/dependencies. > at > org.apache.flink.core.fs.UnsupportedSchemeFactory.create(UnsupportedSchemeFactory.java:55) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:526) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:409) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at org.apache.flink.core.fs.Path.getFileSystem(Path.java:274) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.python.PythonEnvUtils.addToPythonPath(PythonEnvUtils.java:291) > ~[?:?] > at > org.apache.flink.client.python.PythonEnvUtils.preparePythonEnvironment(PythonEnvUtils.java:226) > ~[?:?] > at > org.apache.flink.client.python.PythonEnvUtils.launchPy4jPythonClient(PythonEnvUtils.java:487) > ~[?:?] > at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:92) > ~[?:?] > at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) ~[?:?] > at > jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:?] > at > jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:?] > at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:105) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > at > org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap.runApplicationEntryPoint(ApplicationDispatcherBootstrap.java:301) > ~[flink-dist-1.18.0-amzn-0.jar:1.18.0-amzn-0] > > ``` > > I tested the Flink job locally and it perfectly worker. I even tried to > translate it in Java and run using the AWS EMR_on_EKS_Flin container and it > also workerd. So I do not understand why it does not work in PyFlink. > Moreover, this error is related to the file system, but I never make any > call to it. And even if I did, why does it work in Java and not in Python ? > > > Best regards, > > KY Alexandre >