[ https://issues.apache.org/jira/browse/FLINK-9411?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519405#comment-16519405 ]
Triones Deng edited comment on FLINK-9411 at 6/21/18 2:29 PM: -------------------------------------------------------------- [~StephanEwen] sure a design is necessary, here there may be two ways to do it i think in production. Orc and Parquet file is popupar because of columnar storage. there are two method to support parquet writer. 1. just write a ParquetStreamWriter which is a subclass of StreamWriterBase. which looks the same as FLINK-9407 2. with a HdfsWriterWrapper which own one delegate writer, when end user want to use one format, just simply to specify the format like orc, parquet and let the wrapper create a suitable writer like OrcStreamWriter or ParquetStreamWriter and so on. sample code for the HdfsWriterWrapper {code:java} public class HdfsWriterWrapper<T> implements Writer<T> { private Writer<T> delegate; private String format; private Configuration configuration; private TableSchema tableSchema; public HdfsWriterWrapper(Configuration configuration, String format,Class<T> tableClass,String[] columnFields){ } {code} which one is better? was (Author: triones): [~StephanEwen] sure a design is necessary, here there may be two ways to do it i think in production. Orc and Parquet file is popupar because of columnar storage. there are two method to support parquet writer. 1. just write a ParquetStreamWriter which is a subclass of StreamWriterBase. which looks the same as FLINK-9407 2. with a HdfsWriterWrapper which own one delegate writer, when end user want to use one format, just simply to specify the format like orc, parquet and let the wrapper create a suitable writer like OrcStreamWriter or ParquetStreamWriter and so on. sample code for the HdfsWriterWrapper {code:java} public class HdfsWriterWrapper<T> implements Writer<T> { private Writer<T> delegate; private String format; private Configuration configuration; private TableSchema tableSchema; public HdfsWriterWrapper(Configuration configuration, String format,Class<T> tableClass,String[] columnFields){ } {code} what do you think will be better? > Support parquet rolling sink writer > ----------------------------------- > > Key: FLINK-9411 > URL: https://issues.apache.org/jira/browse/FLINK-9411 > Project: Flink > Issue Type: New Feature > Components: filesystem-connector > Reporter: mingleizhang > Assignee: Triones Deng > Priority: Major > > Like support orc rolling sink writer in FLINK-9407 , we should also support > parquet rolling sink writer. -- This message was sent by Atlassian JIRA (v7.6.3#76005)