sure i can do that On Tue, May 3, 2016 at 1:21 PM, Reynold Xin <r...@databricks.com> wrote:
> Probably not. Want to submit a pull request? > > > On Tuesday, May 3, 2016, Koert Kuipers <ko...@tresata.com> wrote: > >> yes it works fine if i switch to using the implicits on the SparkSession >> (which is a val) >> >> but do we want to break the old way of doing the import? >> >> On Tue, May 3, 2016 at 12:56 PM, Ted Yu <yuzhih...@gmail.com> wrote: >> >>> Have you tried the following ? >>> >>> scala> import spark.implicits._ >>> import spark.implicits._ >>> >>> scala> spark >>> res0: org.apache.spark.sql.SparkSession = >>> org.apache.spark.sql.SparkSession@323d1fa2 >>> >>> Cheers >>> >>> On Tue, May 3, 2016 at 9:16 AM, Koert Kuipers <ko...@tresata.com> wrote: >>> >>>> with the introduction of SparkSession SQLContext changed from being a >>>> lazy val to a def. >>>> however this is troublesome if you want to do: >>>> >>>> import someDataset.sqlContext.implicits._ >>>> >>>> because it is no longer a stable identifier, i think? i get: >>>> stable identifier required, but someDataset.sqlContext.implicits found. >>>> >>>> anyone else seen this? >>>> >>>> >>> >>