I was able to get more info to output on jobmanager. 2020-09-10 19:50:17,722 INFO org.apache.flink.client.cli.CliFrontend [] - --------------------------------------------------------------------------------
2020-09-10 19:50:17,731 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: taskmanager.numberOfTaskSlots, 2 2020-09-10 19:50:17,732 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: blob.server.port, 6124 2020-09-10 19:50:17,732 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: taskmanager.rpc.port, 6122 2020-09-10 19:50:17,732 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: jobmanager.heap.size, 1g 2020-09-10 19:50:17,732 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: taskmanager.memory.process.size, 1g 2020-09-10 19:50:17,733 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend, rocksdb 2020-09-10 19:50:17,733 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.checkpoints.dir, file:///flink_state/checkpoints 2020-09-10 19:50:17,733 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.savepoints.dir, file:///flink_state/savepoints 2020-09-10 19:50:17,733 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.async, true 2020-09-10 19:50:17,733 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.fs.memory-threshold, 1024 2020-09-10 19:50:17,734 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.fs.write-buffer-size, 4096 2020-09-10 19:50:17,734 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.incremental, true 2020-09-10 19:50:17,734 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.local-recovery, true 2020-09-10 19:50:17,734 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.checkpoints.num-retained, 1 2020-09-10 19:50:17,734 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: taskmanager.state.local.root-dirs, file:///flink_state/local-recovery 2020-09-10 19:50:17,735 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.rocksdb.checkpoint.transfer.thread.num, 1 2020-09-10 19:50:17,735 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.rocksdb.localdir, /flink_state/rocksdb 2020-09-10 19:50:17,735 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.rocksdb.options-factory, org.apache.flink.contrib.streaming.state.DefaultConfigurableOptionsFactory 2020-09-10 19:50:17,735 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.rocksdb.predefined-options, DEFAULT 2020-09-10 19:50:17,735 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.rocksdb.timer-service.factory, HEAP 2020-09-10 19:50:17,735 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: state.backend.rocksdb.ttl.compaction.filter.enabled, false 2020-09-10 19:50:17,736 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: jobmanager.rpc.address, flink-jobmanager 2020-09-10 19:50:17,736 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: jobmanager.rpc.port, 6123 2020-09-10 19:50:17,736 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: taskmanager.memory.jvm-metaspace.size, 256mb 2020-09-10 19:50:17,736 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: s3a.endpoint, http://minio:9000 2020-09-10 19:50:17,736 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: s3a.path.style.access, true 2020-09-10 19:50:17,736 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: s3a.access-key, YOURACCESSKEY 2020-09-10 19:50:17,737 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: s3a.secret-key, ****** 2020-09-10 19:50:17,737 INFO org.apache.flink.configuration.GlobalConfiguration [] - Loading configuration property: s3a.aws.credentials.provider, org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider 2020-09-10 19:50:17,802 INFO org.apache.flink.client.cli.CliFrontend [] - Loading FallbackYarnSessionCli 2020-09-10 19:50:17,929 INFO org.apache.flink.core.fs.FileSystem [] - Hadoop is not in the classpath/dependencies. The extended set of supported File Systems via Hadoop is not available. 2020-09-10 19:50:18,102 INFO org.apache.flink.runtime.security.modules.HadoopModuleFactory [] - Cannot create Hadoop Security Module because Hadoop cannot be found in the Classpath. 2020-09-10 19:50:18,126 INFO org.apache.flink.runtime.security.modules.JaasModule [] - Jaas file will be created as /tmp/jaas-1506212733867615019.conf. 2020-09-10 19:50:18,161 INFO org.apache.flink.runtime.security.contexts.HadoopSecurityContextFactory [] - Cannot install HadoopSecurityContext because Hadoop cannot be found in the Classpath. 2020-09-10 19:50:18,163 INFO org.apache.flink.client.cli.CliFrontend [] - Running 'list' command. 2020-09-10 19:50:18,226 INFO org.apache.flink.client.deployment.DefaultClusterClientServiceLoader [] - Could not load factory due to missing dependencies. 2020-09-10 19:50:19,107 INFO org.apache.flink.client.cli.CliFrontend [] - Waiting for response... 2020-09-10 19:50:19,414 INFO org.apache.flink.client.cli.CliFrontend [] - Successfully retrieved list of jobs On Thu, Sep 10, 2020 at 10:50 AM Dan Hill <quietgol...@gmail.com> wrote: > I changed the levels to DEBUG. I don't see useful data in the logs. > > > https://drive.google.com/file/d/1ua1zsr3BInY_8xdsWwA__F0uloAqy-vG/view?usp=sharing > > On Thu, Sep 10, 2020 at 8:45 AM Arvid Heise <ar...@ververica.com> wrote: > >> Could you try 1) or 2) and enable debug logging* and share the log with >> us? >> >> *Usually by adjusting FLINK_HOME/conf/log4j.properties. >> >> On Thu, Sep 10, 2020 at 5:38 PM Dan Hill <quietgol...@gmail.com> wrote: >> >>> Ah, sorry, it's a copy/paste issue with this email. I've tried both: >>> 1) using s3a uri with flink-s3-fs-hadoop jar >>> in /opt/flink/plugins/s3-fs-hadoop. >>> 2) using s3p uri with flink-s3-fs-presto jar >>> in /opt/flink/plugins/s3-fs-presto. >>> 3) loading both 1 and 2 >>> 4) trying s3 uri. >>> >>> When doing 1) >>> >>> Caused by: >>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not >>> find a file system implementation for scheme 's3a'. The scheme is directly >>> supported by Flink through the following plugin: flink-s3-fs-hadoop. Please >>> ensure that each plugin resides within its own subfolder within the plugins >>> directory. See >>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>> for more information. If you want to use a Hadoop file system for that >>> scheme, please add the scheme to the configuration >>> fs.allowed-fallback-filesystems. For a full list of supported file systems, >>> please see >>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/. >>> >>> When doing 2) >>> >>> Caused by: >>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not >>> find a file system implementation for scheme 's3p'. The scheme is directly >>> supported by Flink through the following plugin: flink-s3-fs-presto. Please >>> ensure that each plugin resides within its own subfolder within the plugins >>> directory. See >>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>> for more information. If you want to use a Hadoop file system for that >>> scheme, please add the scheme to the configuration >>> fs.allowed-fallback-filesystems. For a full list of supported file systems, >>> please see >>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/. >>> >>> etc >>> >>> On Thu, Sep 10, 2020 at 8:15 AM Arvid Heise <ar...@ververica.com> wrote: >>> >>>> Hi Dan, >>>> >>>> s3p is only provided by flink-s3-fs-presto plugin. The plugin you used >>>> provides s3a. >>>> (and both provide s3, but it's good to use the more specific prefix). >>>> >>>> Best, >>>> >>>> Arvid >>>> >>>> On Thu, Sep 10, 2020 at 9:24 AM Dan Hill <quietgol...@gmail.com> wrote: >>>> >>>>> *Background* >>>>> I'm converting some prototype Flink v1.11.1 code that uses >>>>> DataSet/DataTable APIs to use the Table API. >>>>> >>>>> *Problem* >>>>> When switching to using the Table API, my s3 plugins stopped working. >>>>> I don't know why. I've added the required maven table dependencies to the >>>>> job. >>>>> >>>>> I've tried us moving both the presto and/or the hadoop s3 jars to >>>>> plugin subfolders. No luck. >>>>> >>>>> Any ideas what is wrong? I'm guessing I'm missing something simple. >>>>> >>>>> >>>>> *Error* >>>>> >>>>> Caused by: >>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not >>>>> find a file system implementation for scheme 's3p'. The scheme is directly >>>>> supported by Flink through the following plugin: flink-s3-fs-presto. >>>>> Please >>>>> ensure that each plugin resides within its own subfolder within the >>>>> plugins >>>>> directory. See >>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>>>> for more information. If you want to use a Hadoop file system for that >>>>> scheme, please add the scheme to the configuration >>>>> fs.allowed-fallback-filesystems. For a full list of supported file >>>>> systems, >>>>> please see >>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/ >>>>> . >>>>> >>>>> at >>>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:473) >>>>> >>>>> at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389) >>>>> >>>>> at org.apache.flink.core.fs.Path.getFileSystem(Path.java:292) >>>>> >>>>> at >>>>> org.apache.flink.table.filesystem.FileSystemTableSink.toStagingPath(FileSystemTableSink.java:232) >>>>> >>>>> ... 35 more >>>>> >>>>> *ls of plugins directory (same for taskmanager)* >>>>> >>>>> kubectl exec pod/flink-jobmanager-0 -- ls -l >>>>> /opt/flink/plugins/s3-fs-hadoop >>>>> >>>>> total 19520 >>>>> >>>>> -rw-r--r-- 1 root root 19985452 Sep 10 06:27 >>>>> flink-s3-fs-hadoop-1.11.1.jar >>>>> >>>>> >>>>> >>>> >>>> -- >>>> >>>> Arvid Heise | Senior Java Developer >>>> >>>> <https://www.ververica.com/> >>>> >>>> Follow us @VervericaData >>>> >>>> -- >>>> >>>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink >>>> Conference >>>> >>>> Stream Processing | Event Driven | Real Time >>>> >>>> -- >>>> >>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >>>> >>>> -- >>>> Ververica GmbH >>>> Registered at Amtsgericht Charlottenburg: HRB 158244 B >>>> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji >>>> (Toni) Cheng >>>> >>> >> >> -- >> >> Arvid Heise | Senior Java Developer >> >> <https://www.ververica.com/> >> >> Follow us @VervericaData >> >> -- >> >> Join Flink Forward <https://flink-forward.org/> - The Apache Flink >> Conference >> >> Stream Processing | Event Driven | Real Time >> >> -- >> >> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >> >> -- >> Ververica GmbH >> Registered at Amtsgericht Charlottenburg: HRB 158244 B >> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji >> (Toni) Cheng >> >