Hi, Maybe you could use streamingContext.fileStream like in the example from https://spark.apache.org/docs/latest/streaming-programming-guide.html#input-dstreams-and-receivers, you can read "from files on any file system compatible with the HDFS API (that is, HDFS, S3, NFS, etc.)". You could split the file into several smaller files, and move them to the target folder one by one with some sleep time in between to simulate a stream of data with custom granularity.
Hope that helps, Greetings, Juan 2015-05-01 9:30 GMT+02:00 anshu shukla <anshushuk...@gmail.com>: > > > > > I have the real DEBS-TAxi data in csv file , in order to operate over it > how to simulate a "Spout" kind of thing as event generator using the > timestamps in CSV file. > > > > > -- > Thanks & Regards, > Anshu Shukla >