Since the cluster is standalone, you are better off reading/writing to hdfs instead of local filesystem.
Thanks Best Regards On Wed, Feb 18, 2015 at 2:32 PM, bit1...@163.com <bit1...@163.com> wrote: > But I am able to run the SparkPi example: > ./run-example SparkPi 1000 --master spark://192.168.26.131:7077 > > Result:Pi is roughly 3.14173708 > > ------------------------------ > bit1...@163.com > > > *From:* bit1...@163.com > *Date:* 2015-02-18 16:29 > *To:* user <user@spark.apache.org> > *Subject:* Problem with 1 master + 2 slaves cluster > Hi sparkers, > I setup a spark(1.2.1) cluster with 1 master and 2 slaves, and then > startup them, everything looks running normally. > In the master node, I run the spark-shell, with the following steps: > > bin/spark-shell --master spark://192.168.26.131:7077 > scala> var rdd = > sc.textFile("file:///home/hadoop/history.txt.used.byspark", 7) > rdd.flatMap(_.split(" ")).map((_, 1)).reduceByKey(_ + _,5).map(x => (x._2, > x._1)).sortByKey(false).map(x => (x._2, > x._1)).saveAsTextFile("file:///home/hadoop/output") > > After finishing running the application, there is no word count related > output, there does exist an output directory appear on each slave node, > but there is only a "_temporary" subdirectory > > Any ideas? Thanks! > > ------------------------------ > >