[ 
https://issues.apache.org/jira/browse/SPARK-14146?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15320357#comment-15320357
 ] 

Prashant Sharma commented on SPARK-14146:
-----------------------------------------

So I tried that option, looks like it does not help either. 
here is the branch. 
https://github.com/ScrapCodes/spark/tree/SPARK-14146/import-fix

Unexplored ideas that can may be fix this issue.
-Yunused-imports is one unexplored territory.
I am not sure, if replacing semicolon in the input with \n can work. Because 
sometimes input can be an XML(or even scala-xml) literal. 


And I will think of more options.

Thanks,

> Imported implicits can't be found in Spark REPL in some cases
> -------------------------------------------------------------
>
>                 Key: SPARK-14146
>                 URL: https://issues.apache.org/jira/browse/SPARK-14146
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core, SQL
>    Affects Versions: 2.0.0
>            Reporter: Wenchen Fan
>
> {code}
> class I(i: Int) {
>   def double: Int = i * 2
> }
> class Context {
>   implicit def toI(i: Int): I = new I(i)
> }
> val c = new Context
> import c._
> // OK
> 1.double
> // Fail
> class A; 1.double
> {code}
> The above code snippets can work in Scala REPL however.
> This will affect our Dataset functionality, for example:
> {code}
> class A; Seq(1 -> "a").toDS() // fail
> {code}
> or in paste mode:
> {code}
> :paste
> class A
> Seq(1 -> "a").toDS() // fail
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to