Hi, could not agree more with Molotch :)
Regards, Gourav Sengupta On Thu, May 27, 2021 at 7:08 PM Molotch <ma...@kth.se> wrote: > You can specify the line separator to make spark split your records into > separate rows. > > df = spark.read.option("lineSep","^^^").text("path") > > Then you need to df.select(split("value", "***").as("arrayColumn")) the > column into an array and map over it with getItem to create a column for > each property. > > df.select((0 until 8).map(i => $"arrayColumn".getItem(i).as(s"col$i")): _* > ) > > Then you should have a DataFrame with each record on a row and each > property > in a column. > > > > -- > Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ > > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >