Why don’t you get an S3 notification on SQS and do the actions from there?

You will probably need to write the content of the files to a no sql database .

Alternatively send the s3 notification to Kafka and read flink from there.


https://docs.aws.amazon.com/AmazonS3/latest/dev/NotificationHowTo.html


> Am 01.09.2020 um 16:46 schrieb orionemail <orionem...@protonmail.com>:
> 
> 
> Hi,
> 
> I have a S3 bucket that is continuously written to by millions of devices.  
> These upload small compressed archives.
> 
> What I want to do is treat the tar gzipped (.tgz) files as a streaming source 
> and process each archive.  The archive contains three files that each might 
> need to be processed.
> 
> I see that 
> env.readFile(f, bucket, FileProcessingMode.PROCESS_CONTINUOUSLY, 
> 10000L).print();
> might do what I need, but I am unsure how best to implement 'f' - the 
> InputFileFormat.  Is there a similar example for me to reference?
> 
> Or is this idea not workable with this method? I need to ensure exactly once, 
> and also trigger removal of the files after processing.
> 
> Thanks,
> 
> 
> Sent with ProtonMail Secure Email.
> 

Reply via email to