I've seen this recently as well. Seems to be an issue with the Scala REPL after running and rerunning notebooks with a lot of code.
Only solution I found was too restart the interpreter. Even Databricks cloud seems to have this issue: https://forums.databricks.com/questions/427/why-do-i-see-this-error-when-i-run-my-notebook-jav.html Thanks, Silvio ________________________________ From: Randy Gelhausen<mailto:[email protected]> Sent: 8/22/2015 6:33 PM To: [email protected]<mailto:[email protected]> Subject: "File name too long" error in Spark paragraphs Hi All, Anyone see something similar to this: %spark import org.apache.spark.sql._ import org.apache.phoenix.spark._ val input = "/user/root/crimes/atlanta" val df = sqlContext.read.format("com.databricks.spark.csv").option("header", "true").option("DROPMALFORMED", "true").load(input) val columns = df.columns.map(x => x.toUpperCase + " varchar,\n") column The result is an error: File name too long I tried commenting out various lines, and then ALL lines, but everything (even in new paragraphs) passed to the interpreter results in "File name too long". Am I doing something silly? Thanks, -Randy
