Spark users,
We really would want to get an input here about how the results from a
Spark Query will be accessible to a web-application. Given Spark is a well
used in the industry I would have thought that this part would have lots of
answers/tutorials about it, but I didnt find anything.

Here are a few options that come to mind

1) Spark results are saved in another DB ( perhaps a traditional one) and a
request for query returns the new table name for access through a paginated
query. That seems doable, although a bit convoluted as we need to handle
the completion of the query.

2) Spark results are pumped into a messaging queue from which a socket
server like connection is made.

What confuses me is that other connectors to spark, like those for Tableau,
using something like JDBC should have all the data (not the top 500 that we
typically can get via Livy or other REST interfaces to Spark). How do those
connectors get all the data through a single connection?


Can someone with expertise help in bringing clarity.

Thank you.

Affan
ᐧ
ᐧ

Reply via email to