Hi Fabian Thanks for the reply. Yes my json is separated by new lines. It would have been great if you had explained the function that goes inside the map. I tried to use the 'scala.util.parsing.json._' library but got no luck.
On Tue, Apr 26, 2016 at 1:11 PM, Fabian Hueske <fhue...@gmail.com> wrote: > Hi Punit, > > JSON can be hard to parse in parallel due to its nested structure. It > depends on the schema and (textual) representation of the JSON whether and > how it can be done. The problem is that a parallel input format needs to be > able to identify record boundaries without context information. This can be > very easy, if your JSON data is a list of JSON objects which are separated > by a new line character. However, this is hard to generalize. That's why > Flink does not offer tooling for it (yet). > > If your JSON objects are separated by new line characters, the easiest way > is to read it as text file, where each line results in a String and parse > each object using a standard JSON parser. This would look like: > > ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment(); > > DataSet<String> text = env.readTextFile("/path/to/jsonfile"); > DataSet<YourObject> json = text.map(new YourMapFunctionWhichParsesJSON()); > > Best, Fabian > > 2016-04-26 8:06 GMT+02:00 Punit Naik <naik.puni...@gmail.com>: > > > Hi > > > > I am new to Flink. I was experimenting with the Dataset API and found out > > that there is no explicit method for loading a JSON file as input. Can > > anyone please suggest me a workaround? > > > > -- > > Thank You > > > > Regards > > > > Punit Naik > > > -- Thank You Regards Punit Naik