Hi Mariano, thanks a lot for your question. The resolution on StackOverflow seems to be that Azure Datalake is not yet ( https://issues.apache.org/jira/browse/FLINK-18568) supported by the StreamingFileSink.
On Thu, Jul 30, 2020 at 5:34 PM Mariano González Núñez < mariano....@hotmail.com> wrote: > Hi Flink Team, > I'm Mariano & I'm working with Apache Flink to process data and sink from > Kafka to Azure Datalake (ADLS Gen1). > We are having problems with the sink in parquet format in the ADLS Gen1, > also don't work with the gen2. > We try to do the StreamingFileSink to store in parquet but we can't sink > in adls because the hadoop doesn't work fine with the library and adls > prefix. (HDFS problem... > https://stackoverflow.com/questions/62884450/problem-with-flink-streamingfilesinkgenericrecord-azure-datalake-gen-2 > ) > > We change and use the deprecated BucketingSink (Works for adls without > .setWriter) but we can't sink in Parquet format using the .setWriter. > > You have some suggestion for do the sink to ADLS Gen1 or Gen2 or you have > a new feature in the future to use the FileStreaming. > > > Thank you very much > Best Regards > >