This is exactly what the Spark SQL Thrift server does, if you just want to access it using JDBC.
Matei > On Oct 6, 2016, at 4:27 PM, Benjamin Kim <bbuil...@gmail.com> wrote: > > Has anyone tried to integrate Spark with a server farm of RESTful API > endpoints or even HTTP web-servers for that matter? I know it’s typically > done using a web farm as the presentation interface, then data flows through > a firewall/router to direct calls to a JDBC listener that will SELECT, > INSERT, UPDATE and, at times, DELETE data in a database. Can the same be done > using Spark SQL Thriftserver on top of, say, HBase, Kudu, Parquet, etc.? Or > can Kafka be used somewhere? Spark would be an ideal solution as the > intermediary because it can talk to any data store underneath; so, swapping > out a technology at any time would be possible. > > Just want some ideas. > > Thank, > Ben > > > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org