Hello folks,
                 We see threads from
https://github.com/aws/aws-sdk-java/blob/master/aws-java-sdk-s3/src/main/java/com/amazonaws/services/s3/transfer/internal/TransferManagerUtils.java#L49
outlive a batch job that writes Parquet Files to S3, causing  a ClassLoader
Leak. Is this a known issue ?  Logically a close on the TransferManager
should close the ExecutorService ( and thus the threads ),

The code is fairly straightforward,

    val job = new Job()    val hadoopOutFormat = new
HadoopOutputFormat[Void, GenericRecord](      new
AvroParquetOutputFormat(),      job    )
AvroParquetOutputFormat.setSchema(job, schema)
FileOutputFormat.setOutputPath(job, new
org.apache.hadoop.fs.Path(path))
ParquetOutputFormat.setCompression(job, CompressionCodecName.SNAPPY)
 ParquetOutputFormat.setEnableDictionary(job, true)  // do we need
this?

     and then an output

This is  using


scalaVersion := "2.12.12"flinkVersion = "1.11.2"hadoopVersion = "2.8.3"



Regards

Vishal

Reply via email to