t; Take a look at this article
>>
>>
>>
>>
>> https://jaceklaskowski.gitbooks.io/mastering-apache-spark/spark-rdd-caching.html
>>
>>
>>
>> *From:* Tzahi File
>> *Sent:* Wednesday, August 28, 2019 5:18 AM
>> *To:* user
>> *Subject:* Caching t
;
>
> *From:* Tzahi File
> *Sent:* Wednesday, August 28, 2019 5:18 AM
> *To:* user
> *Subject:* Caching tables in spark
>
>
>
> Hi,
>
>
>
> Looking for your knowledge with some question.
>
> I have 2 different processes that read from the same raw data t
Take a look at this article
https://jaceklaskowski.gitbooks.io/mastering-apache-spark/spark-rdd-caching.html
From: Tzahi File
Sent: Wednesday, August 28, 2019 5:18 AM
To: user
Subject: Caching tables in spark
Hi,
Looking for your knowledge with some question.
I have 2
Hi,
Looking for your knowledge with some question.
I have 2 different processes that read from the same raw data table (around
1.5 TB).
Is there a way to read this data once and cache it somehow and to use this
data in both processes?
Thanks
--
Tzahi File
Data Engineer
[image: ironSource]