It's not a Flume native solution, but an alternative I used in the past was
Kafka Connect using the HDFS connector plugin. That plugin provides
configuration regarding how often to roll over Avro files.

On 20 April 2018 at 13:49, Nitin Kumar <nitin.kumar2...@gmail.com> wrote:

> Hi All,
>
> I am using Flume v1.8 in which Flume agent comprises of Kafka Channel &
> HDFS Sink.
> I am able to write data in Avro file on HDFS into a external HIVE table,
> but the problem is whenever Flume gets restarted it closes that file and
> open a new file because of which I can see many small files. (Data is
> partition on the basis of date)
>
> Can't Flume append to existing file to avoid creation of new file?
> Also, how can I solve this problem which leads to creation of too many
> small files?
>
> Any help would be appreciated.
>
> --
>
> *Regards,Nitin Kumar*
>



-- 
Matt Sicker <boa...@gmail.com>

Reply via email to