Hi,

Maybe you could use streamingContext.fileStream like in the example from
https://spark.apache.org/docs/latest/streaming-programming-guide.html#input-dstreams-and-receivers,
you can read "from files on any file system compatible with the HDFS API
(that is, HDFS, S3, NFS, etc.)". You could split the file into several
smaller files, and move them to the target folder one by one with some
sleep time in between to simulate a stream of data with custom granularity.

Hope that helps,

Greetings,

Juan

2015-05-01 9:30 GMT+02:00 anshu shukla <anshushuk...@gmail.com>:

>
>
>
>
> I have the real DEBS-TAxi data in csv file , in order to operate over it
> how to simulate a "Spout" kind  of thing as event generator using the
> timestamps in CSV file.
>
>
>
>
> --
> Thanks & Regards,
> Anshu Shukla
>

Reply via email to