I just had a look at Hadoop's TextInputFormat.
In hadoop-common-2.2.0.jar there are the following compression codecs
contained:

org.apache.hadoop.io.compress.BZip2Codec
org.apache.hadoop.io.compress.DefaultCodec
org.apache.hadoop.io.compress.DeflateCodec
org.apache.hadoop.io.compress.GzipCodec
org.apache.hadoop.io.compress.Lz4Codec
org.apache.hadoop.io.compress.SnappyCodec

(See also CompressionCodecFactory). So you should be good to go.


On Thu, Feb 19, 2015 at 9:31 PM, Robert Metzger <rmetz...@apache.org> wrote:

> Hi,
>
> right now Flink itself has only support for reading ".deflate" files. Its
> basically the same algorithm as gzip but gzip files seem to have some
> header which makes the two formats incompatible.
>
> But you can easily use HadoopInputFormats with Flink. I'm sure there is a
> Hadoop IF for reading gzip'ed files.
>
>
> Best,
> Robert
>
>
> On Thu, Feb 19, 2015 at 9:25 PM, Sebastian <ssc.o...@googlemail.com>
> wrote:
>
>> Hi,
>>
>> does flink support reading gzipped files? Haven't found any info about
>> this on the website.
>>
>> Best,
>> Sebastian
>>
>
>

Reply via email to