Say I have a spark job that looks like following:
def loadTable1() {
val table1 = sqlContext.jsonFile(s"s3://textfiledirectory/")
table1.cache().registerTempTable("table1")}
def loadTable2() {
val table2 = sqlContext.jsonFile(s"s3://testfiledirectory2/")
table2.cache().registerTempTable("table2")}
def loadAllTables() {
loadTable1()
loadTable2()}
loadAllTables()
How do I parallelize this Spark job so that both tables are created at the
same time or in parallel?
