This is running on my local minikube and is trying to hit minio. On Thu, Sep 10, 2020 at 1:10 PM Dan Hill <quietgol...@gmail.com> wrote:
> I'm using this Helm chart > <https://github.com/riskfocus/helm-charts-public/tree/master/flink>. I > start the job by building an image with the job jar and using kubectl apply > to do a flink run with the jar. > > The log4j.properties on jobmanager and taskmanager have debug level set > and are pretty embedded into the Helm chart. My log4j-cli.properties is > hacked on the CLI side. > > I thought I just needed the s3 plugins in the jobmanager and taskmanager. > Do I need to have a similar plugin structure from the image where I run > 'flink run'? > > > On Thu, Sep 10, 2020 at 1:03 PM Dan Hill <quietgol...@gmail.com> wrote: > >> Copying more of the log >> >> 2020-09-10 19:50:17,712 INFO org.apache.flink.client.cli.CliFrontend >> [] - >> -------------------------------------------------------------------------------- >> >> 2020-09-10 19:50:17,718 INFO org.apache.flink.client.cli.CliFrontend >> [] - Starting Command Line Client (Version: 1.11.1, >> Scala: 2.12, Rev:7eb514a, Date:2020-07-15T07:02:09+02:00) >> >> 2020-09-10 19:50:17,719 INFO org.apache.flink.client.cli.CliFrontend >> [] - OS current user: root >> >> 2020-09-10 19:50:17,719 INFO org.apache.flink.client.cli.CliFrontend >> [] - Current Hadoop/Kerberos user: <no hadoop >> dependency found> >> >> 2020-09-10 19:50:17,719 INFO org.apache.flink.client.cli.CliFrontend >> [] - JVM: OpenJDK 64-Bit Server VM - Oracle >> Corporation - 1.8/25.265-b01 >> >> 2020-09-10 19:50:17,719 INFO org.apache.flink.client.cli.CliFrontend >> [] - Maximum heap size: 2167 MiBytes >> >> tail: log/flink--client-flink-jobmanager-0.log: file truncated >> >> 2020-09-10 19:50:17,720 INFO org.apache.flink.client.cli.CliFrontend >> [] - JAVA_HOME: /usr/local/openjdk-8 >> >> 2020-09-10 19:50:17,720 INFO org.apache.flink.client.cli.CliFrontend >> [] - No Hadoop Dependency available >> >> 2020-09-10 19:50:17,720 INFO org.apache.flink.client.cli.CliFrontend >> [] - JVM Options: >> >> 2020-09-10 19:50:17,720 INFO org.apache.flink.client.cli.CliFrontend >> [] - >> -Djava.security.properties=/opt/flink/conf/security.properties >> >> 2020-09-10 19:50:17,720 INFO org.apache.flink.client.cli.CliFrontend >> [] - >> -Dlog.file=/opt/flink/log/flink--client-flink-jobmanager-0.log >> >> 2020-09-10 19:50:17,720 INFO org.apache.flink.client.cli.CliFrontend >> [] - >> -Dlog4j.configuration=file:/opt/flink/conf/log4j-cli.properties >> >> 2020-09-10 19:50:17,721 INFO org.apache.flink.client.cli.CliFrontend >> [] - >> -Dlog4j.configurationFile=file:/opt/flink/conf/log4j-cli.properties >> >> 2020-09-10 19:50:17,721 INFO org.apache.flink.client.cli.CliFrontend >> [] - >> -Dlogback.configurationFile=file:/opt/flink/conf/logback.xml >> >> 2020-09-10 19:50:17,721 INFO org.apache.flink.client.cli.CliFrontend >> [] - Program Arguments: >> >> 2020-09-10 19:50:17,721 INFO org.apache.flink.client.cli.CliFrontend >> [] - list >> >> 2020-09-10 19:50:17,721 INFO org.apache.flink.client.cli.CliFrontend >> [] - --jobmanager >> >> 2020-09-10 19:50:17,721 INFO org.apache.flink.client.cli.CliFrontend >> [] - localhost:8081 >> >> 2020-09-10 19:50:17,721 INFO org.apache.flink.client.cli.CliFrontend >> [] - Classpath: >> /opt/flink/lib/flink-csv-1.11.1.jar:/opt/flink/lib/flink-json-1.11.1.jar:/opt/flink/lib/flink-shaded-zookeeper-3.4.14.jar:/opt/flink/lib/flink-table-blink_2.12-1.11.1.jar:/opt/flink/lib/flink-table_2.12-1.11.1.jar:/opt/flink/lib/jna-5.4.0.jar:/opt/flink/lib/jna-platform-5.4.0.jar:/opt/flink/lib/log4j-1.2-api-2.12.1.jar:/opt/flink/lib/log4j-api-2.12.1.jar:/opt/flink/lib/log4j-core-2.12.1.jar:/opt/flink/lib/log4j-slf4j-impl-2.12.1.jar:/opt/flink/lib/oshi-core-3.4.0.jar:/opt/flink/lib/flink-dist_2.12-1.11.1.jar::: >> >> 2020-09-10 19:50:17,722 INFO org.apache.flink.client.cli.CliFrontend >> [] - >> -------------------------------------------------------------------------------- >> >> 2020-09-10 19:50:17,731 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> taskmanager.numberOfTaskSlots, 2 >> >> 2020-09-10 19:50:17,732 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: blob.server.port, 6124 >> >> 2020-09-10 19:50:17,732 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: taskmanager.rpc.port, 6122 >> >> 2020-09-10 19:50:17,732 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: jobmanager.heap.size, 1g >> >> 2020-09-10 19:50:17,732 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> taskmanager.memory.process.size, 1g >> >> 2020-09-10 19:50:17,733 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: state.backend, rocksdb >> >> 2020-09-10 19:50:17,733 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: state.checkpoints.dir, >> file:///flink_state/checkpoints >> >> 2020-09-10 19:50:17,733 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: state.savepoints.dir, >> file:///flink_state/savepoints >> >> 2020-09-10 19:50:17,733 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: state.backend.async, true >> >> 2020-09-10 19:50:17,733 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.fs.memory-threshold, 1024 >> >> 2020-09-10 19:50:17,734 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.fs.write-buffer-size, 4096 >> >> 2020-09-10 19:50:17,734 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.incremental, true >> >> 2020-09-10 19:50:17,734 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.local-recovery, true >> >> 2020-09-10 19:50:17,734 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.checkpoints.num-retained, 1 >> >> 2020-09-10 19:50:17,734 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> taskmanager.state.local.root-dirs, file:///flink_state/local-recovery >> >> 2020-09-10 19:50:17,735 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.rocksdb.checkpoint.transfer.thread.num, 1 >> >> 2020-09-10 19:50:17,735 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.rocksdb.localdir, /flink_state/rocksdb >> >> 2020-09-10 19:50:17,735 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.rocksdb.options-factory, >> org.apache.flink.contrib.streaming.state.DefaultConfigurableOptionsFactory >> >> 2020-09-10 19:50:17,735 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.rocksdb.predefined-options, DEFAULT >> >> 2020-09-10 19:50:17,735 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.rocksdb.timer-service.factory, HEAP >> >> 2020-09-10 19:50:17,735 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> state.backend.rocksdb.ttl.compaction.filter.enabled, false >> >> 2020-09-10 19:50:17,736 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: jobmanager.rpc.address, >> flink-jobmanager >> >> 2020-09-10 19:50:17,736 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: jobmanager.rpc.port, 6123 >> >> 2020-09-10 19:50:17,736 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> taskmanager.memory.jvm-metaspace.size, 256mb >> >> 2020-09-10 19:50:17,736 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: s3a.endpoint, >> http://minio:9000 >> >> 2020-09-10 19:50:17,736 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: s3a.path.style.access, >> true >> >> 2020-09-10 19:50:17,736 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: s3a.access-key, >> YOURACCESSKEY >> >> 2020-09-10 19:50:17,737 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: s3a.secret-key, ****** >> >> 2020-09-10 19:50:17,737 INFO >> org.apache.flink.configuration.GlobalConfiguration >> [] - Loading configuration property: >> s3a.aws.credentials.provider, >> org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider >> >> 2020-09-10 19:50:17,802 INFO org.apache.flink.client.cli.CliFrontend >> [] - Loading FallbackYarnSessionCli >> >> 2020-09-10 19:50:17,929 INFO org.apache.flink.core.fs.FileSystem >> [] - Hadoop is not in the classpath/dependencies. The >> extended set of supported File Systems via Hadoop is not available. >> >> 2020-09-10 19:50:18,102 INFO >> org.apache.flink.runtime.security.modules.HadoopModuleFactory >> [] - Cannot create Hadoop Security Module because Hadoop cannot be found in >> the Classpath. >> >> 2020-09-10 19:50:18,126 INFO >> org.apache.flink.runtime.security.modules.JaasModule >> [] - Jaas file will be created as >> /tmp/jaas-1506212733867615019.conf. >> >> 2020-09-10 19:50:18,161 INFO >> org.apache.flink.runtime.security.contexts.HadoopSecurityContextFactory >> [] - Cannot install HadoopSecurityContext because Hadoop cannot be found in >> the Classpath. >> >> 2020-09-10 19:50:18,163 INFO org.apache.flink.client.cli.CliFrontend >> [] - Running 'list' command. >> >> 2020-09-10 19:50:18,226 INFO >> org.apache.flink.client.deployment.DefaultClusterClientServiceLoader >> [] - Could not load factory due to missing dependencies. >> >> 2020-09-10 19:50:19,107 INFO org.apache.flink.client.cli.CliFrontend >> [] - Waiting for response... >> >> 2020-09-10 19:50:19,414 INFO org.apache.flink.client.cli.CliFrontend >> [] - Successfully retrieved list of jobs >> >> >> On Thu, Sep 10, 2020 at 1:02 PM Arvid Heise <ar...@ververica.com> wrote: >> >>> Hi Dan, >>> >>> somehow enabling debug statements did not work. >>> >>> However, the logs helps to narrow down the issue. The exception occurs >>> neither on jobmanager nor on taskmanager. It occurs wherever you execute >>> the command line interface. >>> >>> How do you execute the job? Do you start it from your machine? Can you >>> try out to also add the respective s3 plugin there? >>> >>> Best, >>> >>> Arvid >>> >>> On Thu, Sep 10, 2020 at 7:50 PM Dan Hill <quietgol...@gmail.com> wrote: >>> >>>> I changed the levels to DEBUG. I don't see useful data in the logs. >>>> >>>> >>>> https://drive.google.com/file/d/1ua1zsr3BInY_8xdsWwA__F0uloAqy-vG/view?usp=sharing >>>> >>>> On Thu, Sep 10, 2020 at 8:45 AM Arvid Heise <ar...@ververica.com> >>>> wrote: >>>> >>>>> Could you try 1) or 2) and enable debug logging* and share the log >>>>> with us? >>>>> >>>>> *Usually by adjusting FLINK_HOME/conf/log4j.properties. >>>>> >>>>> On Thu, Sep 10, 2020 at 5:38 PM Dan Hill <quietgol...@gmail.com> >>>>> wrote: >>>>> >>>>>> Ah, sorry, it's a copy/paste issue with this email. I've tried both: >>>>>> 1) using s3a uri with flink-s3-fs-hadoop jar >>>>>> in /opt/flink/plugins/s3-fs-hadoop. >>>>>> 2) using s3p uri with flink-s3-fs-presto jar >>>>>> in /opt/flink/plugins/s3-fs-presto. >>>>>> 3) loading both 1 and 2 >>>>>> 4) trying s3 uri. >>>>>> >>>>>> When doing 1) >>>>>> >>>>>> Caused by: >>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not >>>>>> find a file system implementation for scheme 's3a'. The scheme is >>>>>> directly >>>>>> supported by Flink through the following plugin: flink-s3-fs-hadoop. >>>>>> Please >>>>>> ensure that each plugin resides within its own subfolder within the >>>>>> plugins >>>>>> directory. See >>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>>>>> for more information. If you want to use a Hadoop file system for that >>>>>> scheme, please add the scheme to the configuration >>>>>> fs.allowed-fallback-filesystems. For a full list of supported file >>>>>> systems, >>>>>> please see >>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/ >>>>>> . >>>>>> >>>>>> When doing 2) >>>>>> >>>>>> Caused by: >>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not >>>>>> find a file system implementation for scheme 's3p'. The scheme is >>>>>> directly >>>>>> supported by Flink through the following plugin: flink-s3-fs-presto. >>>>>> Please >>>>>> ensure that each plugin resides within its own subfolder within the >>>>>> plugins >>>>>> directory. See >>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>>>>> for more information. If you want to use a Hadoop file system for that >>>>>> scheme, please add the scheme to the configuration >>>>>> fs.allowed-fallback-filesystems. For a full list of supported file >>>>>> systems, >>>>>> please see >>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/ >>>>>> . >>>>>> >>>>>> etc >>>>>> >>>>>> On Thu, Sep 10, 2020 at 8:15 AM Arvid Heise <ar...@ververica.com> >>>>>> wrote: >>>>>> >>>>>>> Hi Dan, >>>>>>> >>>>>>> s3p is only provided by flink-s3-fs-presto plugin. The plugin you >>>>>>> used provides s3a. >>>>>>> (and both provide s3, but it's good to use the more specific prefix). >>>>>>> >>>>>>> Best, >>>>>>> >>>>>>> Arvid >>>>>>> >>>>>>> On Thu, Sep 10, 2020 at 9:24 AM Dan Hill <quietgol...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> *Background* >>>>>>>> I'm converting some prototype Flink v1.11.1 code that uses >>>>>>>> DataSet/DataTable APIs to use the Table API. >>>>>>>> >>>>>>>> *Problem* >>>>>>>> When switching to using the Table API, my s3 plugins stopped >>>>>>>> working. I don't know why. I've added the required maven table >>>>>>>> dependencies to the job. >>>>>>>> >>>>>>>> I've tried us moving both the presto and/or the hadoop s3 jars to >>>>>>>> plugin subfolders. No luck. >>>>>>>> >>>>>>>> Any ideas what is wrong? I'm guessing I'm missing something simple. >>>>>>>> >>>>>>>> >>>>>>>> *Error* >>>>>>>> >>>>>>>> Caused by: >>>>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could >>>>>>>> not >>>>>>>> find a file system implementation for scheme 's3p'. The scheme is >>>>>>>> directly >>>>>>>> supported by Flink through the following plugin: flink-s3-fs-presto. >>>>>>>> Please >>>>>>>> ensure that each plugin resides within its own subfolder within the >>>>>>>> plugins >>>>>>>> directory. See >>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html >>>>>>>> for more information. If you want to use a Hadoop file system for that >>>>>>>> scheme, please add the scheme to the configuration >>>>>>>> fs.allowed-fallback-filesystems. For a full list of supported file >>>>>>>> systems, >>>>>>>> please see >>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/ >>>>>>>> . >>>>>>>> >>>>>>>> at >>>>>>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:473) >>>>>>>> >>>>>>>> at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389) >>>>>>>> >>>>>>>> at org.apache.flink.core.fs.Path.getFileSystem(Path.java:292) >>>>>>>> >>>>>>>> at >>>>>>>> org.apache.flink.table.filesystem.FileSystemTableSink.toStagingPath(FileSystemTableSink.java:232) >>>>>>>> >>>>>>>> ... 35 more >>>>>>>> >>>>>>>> *ls of plugins directory (same for taskmanager)* >>>>>>>> >>>>>>>> kubectl exec pod/flink-jobmanager-0 -- ls -l >>>>>>>> /opt/flink/plugins/s3-fs-hadoop >>>>>>>> >>>>>>>> total 19520 >>>>>>>> >>>>>>>> -rw-r--r-- 1 root root 19985452 Sep 10 06:27 >>>>>>>> flink-s3-fs-hadoop-1.11.1.jar >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> Arvid Heise | Senior Java Developer >>>>>>> >>>>>>> <https://www.ververica.com/> >>>>>>> >>>>>>> Follow us @VervericaData >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink >>>>>>> Conference >>>>>>> >>>>>>> Stream Processing | Event Driven | Real Time >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >>>>>>> >>>>>>> -- >>>>>>> Ververica GmbH >>>>>>> Registered at Amtsgericht Charlottenburg: HRB 158244 B >>>>>>> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, >>>>>>> Ji (Toni) Cheng >>>>>>> >>>>>> >>>>> >>>>> -- >>>>> >>>>> Arvid Heise | Senior Java Developer >>>>> >>>>> <https://www.ververica.com/> >>>>> >>>>> Follow us @VervericaData >>>>> >>>>> -- >>>>> >>>>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink >>>>> Conference >>>>> >>>>> Stream Processing | Event Driven | Real Time >>>>> >>>>> -- >>>>> >>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >>>>> >>>>> -- >>>>> Ververica GmbH >>>>> Registered at Amtsgericht Charlottenburg: HRB 158244 B >>>>> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, >>>>> Ji (Toni) Cheng >>>>> >>>> >>> >>> -- >>> >>> Arvid Heise | Senior Java Developer >>> >>> <https://www.ververica.com/> >>> >>> Follow us @VervericaData >>> >>> -- >>> >>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink >>> Conference >>> >>> Stream Processing | Event Driven | Real Time >>> >>> -- >>> >>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany >>> >>> -- >>> Ververica GmbH >>> Registered at Amtsgericht Charlottenburg: HRB 158244 B >>> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji >>> (Toni) Cheng >>> >>