Hi folks,
I have just upgraded to Spark 1.1.0, and try some examples like:
./run-example SparkPageRank pagerank_data.txt 5
It turns out that Spark keeps trying to connect to my name node and read the
file from HDFS other than local FS:
Client: Retrying connect to server: Node1/192.168.0.101:9000. Already
tried 0 time(s)
Even if I use "file://" in my data file path, the issue still comes. This
does not happen in spark-shell.
Is there anything I am missing in configurations or the way I specified the
path?
Thanks,
Max
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Running-Example-in-local-mode-with-input-files-tp16186.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]