Hi Fabian

Thanks for the reply. Yes my json is separated by new lines. It would have
been great if you had explained the function that goes inside the map. I
tried to use the 'scala.util.parsing.json._' library but got no luck.

On Tue, Apr 26, 2016 at 1:11 PM, Fabian Hueske <fhue...@gmail.com> wrote:

> Hi Punit,
>
> JSON can be hard to parse in parallel due to its nested structure. It
> depends on the schema and (textual) representation of the JSON whether and
> how it can be done. The problem is that a parallel input format needs to be
> able to identify record boundaries without context information. This can be
> very easy, if your JSON data is a list of JSON objects which are separated
> by a new line character. However, this is hard to generalize. That's why
> Flink does not offer tooling for it (yet).
>
> If your JSON objects are separated by new line characters, the easiest way
> is to read it as text file, where each line results in a String and parse
> each object using a standard JSON parser. This would look like:
>
> ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
>
> DataSet<String> text = env.readTextFile("/path/to/jsonfile");
> DataSet<YourObject> json = text.map(new YourMapFunctionWhichParsesJSON());
>
> Best, Fabian
>
> 2016-04-26 8:06 GMT+02:00 Punit Naik <naik.puni...@gmail.com>:
>
> > Hi
> >
> > I am new to Flink. I was experimenting with the Dataset API and found out
> > that there is no explicit method for loading a JSON file as input. Can
> > anyone please suggest me a workaround?
> >
> > --
> > Thank You
> >
> > Regards
> >
> > Punit Naik
> >
>



-- 
Thank You

Regards

Punit Naik

Reply via email to