Re: Refreshing a persisted RDD

2017-05-19 Thread Sudhir Menon
; >> >> >> Now some time later while the query is running we do >> >> >> >> val dfRefreshedBlackList = spark.read.csv(….) >> dfRefreshedBlackList.createOrReplaceTempView(“blacklist”) >> >> >> >> Now, will dfBlackListedAc

Re: Refreshing a persisted RDD

2017-05-03 Thread Tathagata Das
ewly created blacklist? Or > will it continue to hold the reference to the old dataframe? What if we had > done RDD operations instead of using Spark SQL to join the dataframes? > > > > *From: *Tathagata Das > *Date: *Wednesday, May 3, 2017 at 6:32 PM > *To: *"Lalwani, Jayes

Re: Refreshing a persisted RDD

2017-05-03 Thread Lalwani, Jayesh
operations instead of using Spark SQL to join the dataframes? From: Tathagata Das Date: Wednesday, May 3, 2017 at 6:32 PM To: "Lalwani, Jayesh" Cc: user Subject: Re: Refreshing a persisted RDD If you want to always get the latest data in files, its best to always recreate the DataFrame. O

Re: Refreshing a persisted RDD

2017-05-03 Thread Tathagata Das
o refresh cached data > without restarting the Spark application? > > > > -- > View this message in context: http://apache-spark-user-list. > 1001560.n3.nabble.com/Refreshing-a-persisted-RDD-tp28642.html > Sent from the Apache Spark User List mailing list archive at Nabble.com.

Refreshing a persisted RDD

2017-05-03 Thread JayeshLalwani
Is there a better way to refresh cached data without restarting the Spark application? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Refreshing-a-persisted-RDD-tp28642.html Sent from the Apache Spark User List mailing list archive at