I'm using this Helm chart
<https://github.com/riskfocus/helm-charts-public/tree/master/flink>.  I
start the job by building an image with the job jar and using kubectl apply
to do a flink run with the jar.

The log4j.properties on jobmanager and taskmanager have debug level set and
are pretty embedded into the Helm chart.  My log4j-cli.properties is hacked
on the CLI side.

I thought I just needed the s3 plugins in the jobmanager and taskmanager.
Do I need to have a similar plugin structure from the image where I run
'flink run'?


On Thu, Sep 10, 2020 at 1:03 PM Dan Hill <quietgol...@gmail.com> wrote:

> Copying more of the log
>
> 2020-09-10 19:50:17,712 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -
> --------------------------------------------------------------------------------
>
> 2020-09-10 19:50:17,718 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  Starting Command Line Client (Version: 1.11.1,
> Scala: 2.12, Rev:7eb514a, Date:2020-07-15T07:02:09+02:00)
>
> 2020-09-10 19:50:17,719 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  OS current user: root
>
> 2020-09-10 19:50:17,719 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  Current Hadoop/Kerberos user: <no hadoop
> dependency found>
>
> 2020-09-10 19:50:17,719 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  JVM: OpenJDK 64-Bit Server VM - Oracle
> Corporation - 1.8/25.265-b01
>
> 2020-09-10 19:50:17,719 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  Maximum heap size: 2167 MiBytes
>
> tail: log/flink--client-flink-jobmanager-0.log: file truncated
>
> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  JAVA_HOME: /usr/local/openjdk-8
>
> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  No Hadoop Dependency available
>
> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  JVM Options:
>
> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -
> -Djava.security.properties=/opt/flink/conf/security.properties
>
> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -
> -Dlog.file=/opt/flink/log/flink--client-flink-jobmanager-0.log
>
> 2020-09-10 19:50:17,720 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -
> -Dlog4j.configuration=file:/opt/flink/conf/log4j-cli.properties
>
> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -
> -Dlog4j.configurationFile=file:/opt/flink/conf/log4j-cli.properties
>
> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -
> -Dlogback.configurationFile=file:/opt/flink/conf/logback.xml
>
> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  Program Arguments:
>
> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -     list
>
> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -     --jobmanager
>
> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -     localhost:8081
>
> 2020-09-10 19:50:17,721 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -  Classpath:
> /opt/flink/lib/flink-csv-1.11.1.jar:/opt/flink/lib/flink-json-1.11.1.jar:/opt/flink/lib/flink-shaded-zookeeper-3.4.14.jar:/opt/flink/lib/flink-table-blink_2.12-1.11.1.jar:/opt/flink/lib/flink-table_2.12-1.11.1.jar:/opt/flink/lib/jna-5.4.0.jar:/opt/flink/lib/jna-platform-5.4.0.jar:/opt/flink/lib/log4j-1.2-api-2.12.1.jar:/opt/flink/lib/log4j-api-2.12.1.jar:/opt/flink/lib/log4j-core-2.12.1.jar:/opt/flink/lib/log4j-slf4j-impl-2.12.1.jar:/opt/flink/lib/oshi-core-3.4.0.jar:/opt/flink/lib/flink-dist_2.12-1.11.1.jar:::
>
> 2020-09-10 19:50:17,722 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] -
> --------------------------------------------------------------------------------
>
> 2020-09-10 19:50:17,731 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> taskmanager.numberOfTaskSlots, 2
>
> 2020-09-10 19:50:17,732 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: blob.server.port, 6124
>
> 2020-09-10 19:50:17,732 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: taskmanager.rpc.port, 6122
>
> 2020-09-10 19:50:17,732 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: jobmanager.heap.size, 1g
>
> 2020-09-10 19:50:17,732 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> taskmanager.memory.process.size, 1g
>
> 2020-09-10 19:50:17,733 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: state.backend, rocksdb
>
> 2020-09-10 19:50:17,733 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: state.checkpoints.dir,
> file:///flink_state/checkpoints
>
> 2020-09-10 19:50:17,733 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: state.savepoints.dir,
> file:///flink_state/savepoints
>
> 2020-09-10 19:50:17,733 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: state.backend.async, true
>
> 2020-09-10 19:50:17,733 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.fs.memory-threshold, 1024
>
> 2020-09-10 19:50:17,734 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.fs.write-buffer-size, 4096
>
> 2020-09-10 19:50:17,734 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: state.backend.incremental,
> true
>
> 2020-09-10 19:50:17,734 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.local-recovery, true
>
> 2020-09-10 19:50:17,734 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.checkpoints.num-retained, 1
>
> 2020-09-10 19:50:17,734 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> taskmanager.state.local.root-dirs, file:///flink_state/local-recovery
>
> 2020-09-10 19:50:17,735 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.rocksdb.checkpoint.transfer.thread.num, 1
>
> 2020-09-10 19:50:17,735 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.rocksdb.localdir, /flink_state/rocksdb
>
> 2020-09-10 19:50:17,735 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.rocksdb.options-factory,
> org.apache.flink.contrib.streaming.state.DefaultConfigurableOptionsFactory
>
> 2020-09-10 19:50:17,735 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.rocksdb.predefined-options, DEFAULT
>
> 2020-09-10 19:50:17,735 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.rocksdb.timer-service.factory, HEAP
>
> 2020-09-10 19:50:17,735 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> state.backend.rocksdb.ttl.compaction.filter.enabled, false
>
> 2020-09-10 19:50:17,736 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: jobmanager.rpc.address,
> flink-jobmanager
>
> 2020-09-10 19:50:17,736 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: jobmanager.rpc.port, 6123
>
> 2020-09-10 19:50:17,736 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> taskmanager.memory.jvm-metaspace.size, 256mb
>
> 2020-09-10 19:50:17,736 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: s3a.endpoint,
> http://minio:9000
>
> 2020-09-10 19:50:17,736 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: s3a.path.style.access, true
>
> 2020-09-10 19:50:17,736 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: s3a.access-key,
> YOURACCESSKEY
>
> 2020-09-10 19:50:17,737 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property: s3a.secret-key, ******
>
> 2020-09-10 19:50:17,737 INFO  
> org.apache.flink.configuration.GlobalConfiguration
>           [] - Loading configuration property:
> s3a.aws.credentials.provider,
> org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider
>
> 2020-09-10 19:50:17,802 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] - Loading FallbackYarnSessionCli
>
> 2020-09-10 19:50:17,929 INFO  org.apache.flink.core.fs.FileSystem
>                   [] - Hadoop is not in the classpath/dependencies. The
> extended set of supported File Systems via Hadoop is not available.
>
> 2020-09-10 19:50:18,102 INFO  
> org.apache.flink.runtime.security.modules.HadoopModuleFactory
> [] - Cannot create Hadoop Security Module because Hadoop cannot be found in
> the Classpath.
>
> 2020-09-10 19:50:18,126 INFO  
> org.apache.flink.runtime.security.modules.JaasModule
>         [] - Jaas file will be created as
> /tmp/jaas-1506212733867615019.conf.
>
> 2020-09-10 19:50:18,161 INFO  
> org.apache.flink.runtime.security.contexts.HadoopSecurityContextFactory
> [] - Cannot install HadoopSecurityContext because Hadoop cannot be found in
> the Classpath.
>
> 2020-09-10 19:50:18,163 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] - Running 'list' command.
>
> 2020-09-10 19:50:18,226 INFO  
> org.apache.flink.client.deployment.DefaultClusterClientServiceLoader
> [] - Could not load factory due to missing dependencies.
>
> 2020-09-10 19:50:19,107 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] - Waiting for response...
>
> 2020-09-10 19:50:19,414 INFO  org.apache.flink.client.cli.CliFrontend
>                   [] - Successfully retrieved list of jobs
>
>
> On Thu, Sep 10, 2020 at 1:02 PM Arvid Heise <ar...@ververica.com> wrote:
>
>> Hi Dan,
>>
>> somehow enabling debug statements did not work.
>>
>> However, the logs helps to narrow down the issue. The exception occurs
>> neither on jobmanager nor on taskmanager. It occurs wherever you execute
>> the command line interface.
>>
>> How do you execute the job? Do you start it from your machine? Can you
>> try out to also add the respective s3 plugin there?
>>
>> Best,
>>
>> Arvid
>>
>> On Thu, Sep 10, 2020 at 7:50 PM Dan Hill <quietgol...@gmail.com> wrote:
>>
>>> I changed the levels to DEBUG.  I don't see useful data in the logs.
>>>
>>>
>>> https://drive.google.com/file/d/1ua1zsr3BInY_8xdsWwA__F0uloAqy-vG/view?usp=sharing
>>>
>>> On Thu, Sep 10, 2020 at 8:45 AM Arvid Heise <ar...@ververica.com> wrote:
>>>
>>>> Could you try 1) or 2) and enable debug logging* and share the log with
>>>> us?
>>>>
>>>> *Usually by adjusting FLINK_HOME/conf/log4j.properties.
>>>>
>>>> On Thu, Sep 10, 2020 at 5:38 PM Dan Hill <quietgol...@gmail.com> wrote:
>>>>
>>>>> Ah, sorry, it's a copy/paste issue with this email.  I've tried both:
>>>>> 1) using s3a uri with flink-s3-fs-hadoop jar
>>>>> in /opt/flink/plugins/s3-fs-hadoop.
>>>>> 2) using s3p uri with flink-s3-fs-presto jar
>>>>> in /opt/flink/plugins/s3-fs-presto.
>>>>> 3) loading both 1 and 2
>>>>> 4) trying s3 uri.
>>>>>
>>>>> When doing 1)
>>>>>
>>>>> Caused by:
>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not
>>>>> find a file system implementation for scheme 's3a'. The scheme is directly
>>>>> supported by Flink through the following plugin: flink-s3-fs-hadoop. 
>>>>> Please
>>>>> ensure that each plugin resides within its own subfolder within the 
>>>>> plugins
>>>>> directory. See
>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html
>>>>> for more information. If you want to use a Hadoop file system for that
>>>>> scheme, please add the scheme to the configuration
>>>>> fs.allowed-fallback-filesystems. For a full list of supported file 
>>>>> systems,
>>>>> please see
>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/
>>>>> .
>>>>>
>>>>> When doing 2)
>>>>>
>>>>> Caused by:
>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not
>>>>> find a file system implementation for scheme 's3p'. The scheme is directly
>>>>> supported by Flink through the following plugin: flink-s3-fs-presto. 
>>>>> Please
>>>>> ensure that each plugin resides within its own subfolder within the 
>>>>> plugins
>>>>> directory. See
>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html
>>>>> for more information. If you want to use a Hadoop file system for that
>>>>> scheme, please add the scheme to the configuration
>>>>> fs.allowed-fallback-filesystems. For a full list of supported file 
>>>>> systems,
>>>>> please see
>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/
>>>>> .
>>>>>
>>>>> etc
>>>>>
>>>>> On Thu, Sep 10, 2020 at 8:15 AM Arvid Heise <ar...@ververica.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Dan,
>>>>>>
>>>>>> s3p is only provided by flink-s3-fs-presto plugin. The plugin you
>>>>>> used provides s3a.
>>>>>> (and both provide s3, but it's good to use the more specific prefix).
>>>>>>
>>>>>> Best,
>>>>>>
>>>>>> Arvid
>>>>>>
>>>>>> On Thu, Sep 10, 2020 at 9:24 AM Dan Hill <quietgol...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> *Background*
>>>>>>> I'm converting some prototype Flink v1.11.1 code that uses
>>>>>>> DataSet/DataTable APIs to use the Table API.
>>>>>>>
>>>>>>> *Problem*
>>>>>>> When switching to using the Table API, my s3 plugins stopped
>>>>>>> working.  I don't know why.  I've added the required maven table
>>>>>>> dependencies to the job.
>>>>>>>
>>>>>>> I've tried us moving both the presto and/or the hadoop s3 jars to
>>>>>>> plugin subfolders.  No luck.
>>>>>>>
>>>>>>> Any ideas what is wrong?  I'm guessing I'm missing something simple.
>>>>>>>
>>>>>>>
>>>>>>> *Error*
>>>>>>>
>>>>>>> Caused by:
>>>>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not
>>>>>>> find a file system implementation for scheme 's3p'. The scheme is 
>>>>>>> directly
>>>>>>> supported by Flink through the following plugin: flink-s3-fs-presto. 
>>>>>>> Please
>>>>>>> ensure that each plugin resides within its own subfolder within the 
>>>>>>> plugins
>>>>>>> directory. See
>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html
>>>>>>> for more information. If you want to use a Hadoop file system for that
>>>>>>> scheme, please add the scheme to the configuration
>>>>>>> fs.allowed-fallback-filesystems. For a full list of supported file 
>>>>>>> systems,
>>>>>>> please see
>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/
>>>>>>> .
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:473)
>>>>>>>
>>>>>>> at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389)
>>>>>>>
>>>>>>> at org.apache.flink.core.fs.Path.getFileSystem(Path.java:292)
>>>>>>>
>>>>>>> at
>>>>>>> org.apache.flink.table.filesystem.FileSystemTableSink.toStagingPath(FileSystemTableSink.java:232)
>>>>>>>
>>>>>>> ... 35 more
>>>>>>>
>>>>>>> *ls of plugins directory (same for taskmanager)*
>>>>>>>
>>>>>>> kubectl exec pod/flink-jobmanager-0  -- ls -l
>>>>>>> /opt/flink/plugins/s3-fs-hadoop
>>>>>>>
>>>>>>> total 19520
>>>>>>>
>>>>>>> -rw-r--r-- 1 root root 19985452 Sep 10 06:27
>>>>>>> flink-s3-fs-hadoop-1.11.1.jar
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Arvid Heise | Senior Java Developer
>>>>>>
>>>>>> <https://www.ververica.com/>
>>>>>>
>>>>>> Follow us @VervericaData
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink
>>>>>> Conference
>>>>>>
>>>>>> Stream Processing | Event Driven | Real Time
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany
>>>>>>
>>>>>> --
>>>>>> Ververica GmbH
>>>>>> Registered at Amtsgericht Charlottenburg: HRB 158244 B
>>>>>> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason,
>>>>>> Ji (Toni) Cheng
>>>>>>
>>>>>
>>>>
>>>> --
>>>>
>>>> Arvid Heise | Senior Java Developer
>>>>
>>>> <https://www.ververica.com/>
>>>>
>>>> Follow us @VervericaData
>>>>
>>>> --
>>>>
>>>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink
>>>> Conference
>>>>
>>>> Stream Processing | Event Driven | Real Time
>>>>
>>>> --
>>>>
>>>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany
>>>>
>>>> --
>>>> Ververica GmbH
>>>> Registered at Amtsgericht Charlottenburg: HRB 158244 B
>>>> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji
>>>> (Toni) Cheng
>>>>
>>>
>>
>> --
>>
>> Arvid Heise | Senior Java Developer
>>
>> <https://www.ververica.com/>
>>
>> Follow us @VervericaData
>>
>> --
>>
>> Join Flink Forward <https://flink-forward.org/> - The Apache Flink
>> Conference
>>
>> Stream Processing | Event Driven | Real Time
>>
>> --
>>
>> Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany
>>
>> --
>> Ververica GmbH
>> Registered at Amtsgericht Charlottenburg: HRB 158244 B
>> Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji
>> (Toni) Cheng
>>
>

Reply via email to