I am trying to access directory in Hadoop from my Spark code on local machine.Hadoop is HA enabled .
val conf = new SparkConf().setAppName("LDA Sample").setMaster("local[2]") val sc=new SparkContext(conf) val distFile = sc.textFile("hdfs://hdpha/mini_newsgroups/") println(distFile.count()) but getting error java.net.UnknownHostException: hdpha As hdpha not resolves to a particular machine it is the name I have chosen for my HA Hadoop.I have already copied all hadoop configuration on my local machine and have set the env. variable HADOOP_CONF_DIR But still no success. Any suggestion will be of a great help Note:- Hadoop HA is working properly as i have tried uploading file to hadoop and it works -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Unable-to-Access-files-in-Hadoop-HA-enabled-from-using-Spark-tp26768.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org