Running as Standalone Cluster. From my monitoring console:
[spark-logo-77x50px-hd.png] Spark Master at spark://101.73.54.149:7077
* URL: spark://101.73.54.149:7077 * Workers: 1 * Cores: 2 Total, 0
Used * Memory: 2.4 GB Total, 0.0 B Used * Applications: 0 Running, 24
Completed
Thanks Sean! I got that working last night similar to how you solved it. Any
ideas about how to monitor that same folder in another script by creating a
stream? I can use sc.sequenceFile() to read in the RDD, but how do I get the
name of the file that got added since there is no sequenceFileStre