Hi, All, Although there are lots of discussions related to logging in this news group, I did not find an answer to my specific question so I am posting mine with the hope that this will not cause a duplicated question.
Here is my simplified Java testing Spark app: public class SparkJobEntry { public static void main(String[] args) { // Following line is in stdout from JobTracker UI System.out.println("argc=" + args.length); SparkConf conf = new SparkConf().setAppName("TestSparkApp"); JavaSparkContext sc = new JavaSparkContext(conf); JavaRDD<String> fileRDD = sc.textFile(args[0]); fileRDD.mapToPair(new PairFunction<String, String, String>() { private static final long serialVersionUID = 1L; @Override public Tuple2<String, String> call(String input) throws Exception { // Following line is not in stdout from JobTracker UI System.out.println("This line should be printed in stdout"); // Other code removed from here to make things simple return new Tuple2<String, String>("1", "Testing data"); }}).saveAsTextFile(args[0] + ".results"); } } What I expected from JobTracker UI is to see both stdout lines: first line is "argc=2" and second line is "This line should be printed in stdout". But I only see the first line which is outside of the 'mapToPair'. I actually have verified my 'mapToPair' is called and the statements after the second logging line were executed. The only issue for me is why the second logging is not in JobTracker UI. Appreciate your help. Thanks John -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Logging-in-RDD-mapToPair-of-Java-Spark-application-tp29007.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org