I'm trying to run Spark ALS using the netflix dataset but failed due to "No space on device" exception. It seems the exception is thrown after the training phase. It's not clear to me what is being written and where is the output directory.
I was able to run the same code on the provided test.data dataset. I'm new to Spark and I'd like to get some hints for resolving this problem. The code I ran was got from https://spark.apache.org/docs/latest/mllib-collaborative-filtering.html (the Java version). Relevant info: Spark version: 1.0.2 (Standalone deployment) # slaves/workers/exectuors: 8 Core per worker: 64 memory per executor: 100g Application parameters are left as default. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Failed-running-Spark-ALS-tp14704.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org