I dint mean that. When you try the above approach only one client will have
access to the cached data.
But when you expose your data through a thrift server the case is quite
different.
In the case of thrift server all the request goes to the thrift server and
spark will be able to take the advan
Hi Ashish,
In order to answer your question , I assume that you are planning to
process data and cache them in the memory.If you are using a thrift server
that comes with Spark then you can query on top of it. And multiple
applications can use the cached data as internally all the requests go to
t
I am implementing this approach currently.
A
1.Create data tables in spark-sql and cache them.
2. Configure the hive metastore to read the cached tables and share the
same metastore as spark-sql (You get the spark caching advantage)
3.Run spark code to fetch form the cached tables. In the spark co