Simya Jose created FLINK-14955:
----------------------------------

             Summary: Not able to write to swift via 
StreamingFileSink.forBulkFormat
                 Key: FLINK-14955
                 URL: https://issues.apache.org/jira/browse/FLINK-14955
             Project: Flink
          Issue Type: Bug
            Reporter: Simya Jose


not able to use StreamingFileSink to write to swift file storage

 

*Code*:

flink version: 1.9.1.

 scala 2.11

build tool : maven

main part of the code:

val eligibleItems: DataStream[EligibleItem] = env.fromCollection(Seq(
 EligibleItem("pencil"),
 EligibleItem("rubber"),
 EligibleItem("beer")))(TypeInformation.of(classOf[EligibleItem]))

val factory2: ParquetWriterFactory[EligibleItem] = 
ParquetAvroWriters.forReflectRecord(classOf[EligibleItem])
val sink: StreamingFileSink[EligibleItem] = StreamingFileSink
 .forBulkFormat(new Path(capHadoopPath),factory2)
 .build()

eligibleItems.addSink(sink)
 .setParallelism(1)
 .uid("TEST_1")
 .name("TEST")

*scenario* : when path is set to point to swift ( capHadoopPath = 
"swift://<path>" ) , getting exception - 
_java.lang.UnsupportedOperationException: Recoverable writers on Hadoop are 
only supported for HDFS and for Hadoop version 2.7 or 
newerjava.lang.UnsupportedOperationException: Recoverable writers on Hadoop are 
only supported for HDFS and for Hadoop version 2.7 or newer at 
org.apache.flink.fs.openstackhadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopRecoverableWriter.<init>(HadoopRecoverableWriter.java:57)_
 

 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to