Folks,
I have the following program :
SparkConf conf = new
SparkConf().setMaster("local").setAppName("Indexer").set("spark.driver.maxResultSize",
"2g");conf.set("es.index.auto.create", "true");conf.set("es.nodes",
"localhost");conf.set("es.port", "9200");conf.set("es.write.operation",
"index");JavaSparkContext sc = new JavaSparkContext(conf);
. .
JavaEsSpark.saveToEs(filteredFields, "foo");
I get an error saying cannot find storage. Looks like the driver program cannot
the Elastic Search Server. Seeing the program, I have not associated
JavaEsSpark to the SparkConf.
Question: How do I associate JavaEsSpark to SparkConf?