We are currently pulling out the JSON columns, passing them through
read.json, and then joining them back onto the initial DF so something like
from_json would be a nice quality of life improvement for us.

On Wed, Sep 28, 2016 at 10:52 AM, Michael Armbrust <mich...@databricks.com>
wrote:

> Spark SQL has great support for reading text files that contain JSON data.
> However, in many cases the JSON data is just one column amongst others.
> This is particularly true when reading from sources such as Kafka. This PR
> <https://github.com/apache/spark/pull/15274> adds a new functions
> from_json that converts a string column into a nested StructType with a
> user specified schema, using the same internal logic as the json Data
> Source.
>
> Would love to hear any comments / suggestions.
>
> Michael
>

Reply via email to