Hello,
     I have been trying to Use StreamingFileSink to write to parquetFiles into azure blob storage. I am getting the following error. I did see in the ticket https://issues.apache.org/jira/browse/FLINK-17444 that support for StreamingFileSink is not yet provided.

Attachment: code.java
Description: Binary data

WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by org.apache.flink.api.java.ClosureCleaner 
(file:/Users/user/.m2/repository/org/apache/flink/flink-core/1.13.0/flink-core-1.13.0.jar)
 to field java.util.Properties.serialVersionUID
WARNING: Please consider reporting this to the maintainers of 
org.apache.flink.api.java.ClosureCleaner
WARNING: Use --illegal-access=warn to enable warnings of further illegal 
reflective access operations
WARNING: All illegal access operations will be denied in a future release
11:43:23,780 INFO  org.apache.flink.fs.azurefs.AzureFSFactory                   
[] - Trying to load and instantiate Azure File System
11:43:23,907 INFO  org.apache.flink.fs.azure.shaded.org.eclipse.jetty.util.log  
[] - Logging initialized @6378ms
log4j:WARN No appenders could be found for logger 
(org.apache.flink.fs.shaded.hadoop3.org.apache.commons.beanutils.converters.BooleanConverter).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more 
info.
11:43:24,034 WARN  
org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.metrics2.impl.MetricsConfig
 [] - Cannot locate configuration: tried 
hadoop-metrics2-azure-file-system.properties,hadoop-metrics2.properties
11:43:24,076 INFO  
org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.metrics2.impl.MetricsSystemImpl
 [] - Scheduled Metric snapshot period at 10 second(s).
11:43:24,076 INFO  
org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.metrics2.impl.MetricsSystemImpl
 [] - azure-file-system metrics system started
Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/flink/fs/azure/common/hadoop/HadoopRecoverableWriter
        at 
org.apache.flink.fs.azure.common.hadoop.HadoopFileSystem.createRecoverableWriter(HadoopFileSystem.java:215)
        at 
org.apache.flink.connector.file.sink.FileSink$BulkFormatBuilder.createBucketWriter(FileSink.java:466)
        at 
org.apache.flink.connector.file.sink.FileSink$BulkFormatBuilder.getWriterStateSerializer(FileSink.java:447)
        at 
org.apache.flink.connector.file.sink.FileSink.getWriterStateSerializer(FileSink.java:128)
        at 
org.apache.flink.streaming.runtime.translators.SinkTransformationTranslator.addWriter(SinkTransformationTranslator.java:178)
        at 
org.apache.flink.streaming.runtime.translators.SinkTransformationTranslator.internalTranslate(SinkTransformationTranslator.java:146)
        at 
org.apache.flink.streaming.runtime.translators.SinkTransformationTranslator.translateForStreaming(SinkTransformationTranslator.java:103)
        at 
org.apache.flink.streaming.runtime.translators.SinkTransformationTranslator.translateForStreaming(SinkTransformationTranslator.java:58)
        at 
org.apache.flink.streaming.api.graph.StreamGraphGenerator.translate(StreamGraphGenerator.java:714)
        at 
org.apache.flink.streaming.api.graph.StreamGraphGenerator.transform(StreamGraphGenerator.java:446)
        at 
org.apache.flink.streaming.api.graph.StreamGraphGenerator.generate(StreamGraphGenerator.java:301)
        at 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2010)
        at 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:1995)
        at 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1834)
        at com.grpId.StreamingJob.main(StreamingJob.java:113)
Caused by: java.lang.ClassNotFoundException: 
org.apache.flink.fs.azure.common.hadoop.HadoopRecoverableWriter
        at 
java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:606)
        at 
java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:168)
        at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:522)
        ... 15 more

Process finished with exit code 1

I am trying to run flink in IntelliJ and created the project using the maven archetype as mentioned in the docs and added flink-azure-fs-hadoop Dependency

Attachment: pom.xml
Description: XML document

Kindly let me know what I am missing also Is there a boiler plate code I can refer to?
Thank you.
Regards,
Sudhanva

Reply via email to