Simya Jose created FLINK-14955: ---------------------------------- Summary: Not able to write to swift via StreamingFileSink.forBulkFormat Key: FLINK-14955 URL: https://issues.apache.org/jira/browse/FLINK-14955 Project: Flink Issue Type: Bug Reporter: Simya Jose
not able to use StreamingFileSink to write to swift file storage *Code*: flink version: 1.9.1. scala 2.11 build tool : maven main part of the code: val eligibleItems: DataStream[EligibleItem] = env.fromCollection(Seq( EligibleItem("pencil"), EligibleItem("rubber"), EligibleItem("beer")))(TypeInformation.of(classOf[EligibleItem])) val factory2: ParquetWriterFactory[EligibleItem] = ParquetAvroWriters.forReflectRecord(classOf[EligibleItem]) val sink: StreamingFileSink[EligibleItem] = StreamingFileSink .forBulkFormat(new Path(capHadoopPath),factory2) .build() eligibleItems.addSink(sink) .setParallelism(1) .uid("TEST_1") .name("TEST") *scenario* : when path is set to point to swift ( capHadoopPath = "swift://<path>" ) , getting exception - _java.lang.UnsupportedOperationException: Recoverable writers on Hadoop are only supported for HDFS and for Hadoop version 2.7 or newerjava.lang.UnsupportedOperationException: Recoverable writers on Hadoop are only supported for HDFS and for Hadoop version 2.7 or newer at org.apache.flink.fs.openstackhadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopRecoverableWriter.<init>(HadoopRecoverableWriter.java:57)_ -- This message was sent by Atlassian Jira (v8.3.4#803005)