Thanks for the advice Koert: when everything is in the same essential data-store (HDFS), can't I just run whatever complex tools I'm whichever paradigm they like?
E.g.: GraphX, Mahout &etc. Also, what about Tajo or Drill? Best, Samuel Marks http://linkedin.com/in/samuelmarks PS: Spark-SQL is read-only IIRC, right? On 31 Jan 2015 03:39, "Koert Kuipers" <ko...@tresata.com> wrote: > since you require high-powered analytics, and i assume you want to stay > sane while doing so, you require the ability to "drop out of sql" when > needed. so spark-sql and lingual would be my choices. > > low latency indicates phoenix or spark-sql to me. > > so i would say spark-sql > > On Fri, Jan 30, 2015 at 7:56 AM, Samuel Marks <samuelma...@gmail.com> > wrote: > >> HAWQ is pretty nifty due to its full SQL compliance (ANSI 92) and >> exposing both JDBC and ODBC interfaces. However, although Pivotal does >> open-source >> a lot of software <http://www.pivotal.io/oss>, I don't believe they open >> source Pivotal HD: HAWQ. >> >> So that doesn't meet my requirements. I should note that the project I am >> building will also be open-source, which heightens the importance of having >> all components also being open-source. >> >> Cheers, >> >> Samuel Marks >> http://linkedin.com/in/samuelmarks >> >> On Fri, Jan 30, 2015 at 11:35 PM, Siddharth Tiwari < >> siddharth.tiw...@live.com> wrote: >> >>> Have you looked at HAWQ from Pivotal ? >>> >>> Sent from my iPhone >>> >>> On Jan 30, 2015, at 4:27 AM, Samuel Marks <samuelma...@gmail.com> wrote: >>> >>> Since Hadoop <https://hive.apache.org> came out, there have been >>> various commercial and/or open-source attempts to expose some compatibility >>> with SQL <http://drill.apache.org>. Obviously by posting here I am not >>> expecting an unbiased answer. >>> >>> Seeking an SQL-on-Hadoop offering which provides: low-latency querying, >>> and supports the most common CRUD <https://spark.apache.org>, including >>> [the basics!] along these lines: CREATE TABLE, INSERT INTO, SELECT * >>> FROM, UPDATE Table SET C1=2 WHERE, DELETE FROM, and DROP TABLE. >>> Transactional support would be nice also, but is not a must-have. >>> >>> Essentially I want a full replacement for the more traditional RDBMS, >>> one which can scale from 1 node to a serious Hadoop cluster. >>> >>> Python is my language of choice for interfacing, however there does seem >>> to be a Python JDBC wrapper <https://spark.apache.org/sql>. >>> >>> Here is what I've found thus far: >>> >>> - Apache Hive <https://hive.apache.org> (SQL-like, with interactive >>> SQL thanks to the Stinger initiative) >>> - Apache Drill <http://drill.apache.org> (ANSI SQL support) >>> - Apache Spark <https://spark.apache.org> (Spark SQL >>> <https://spark.apache.org/sql>, queries only, add data via Hive, RDD >>> >>> <https://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.sql.SchemaRDD> >>> or Paraquet <http://parquet.io/>) >>> - Apache Phoenix <http://phoenix.apache.org> (built atop Apache HBase >>> <http://hbase.apache.org>, lacks full transaction >>> <http://en.wikipedia.org/wiki/Database_transaction> support, relational >>> operators <http://en.wikipedia.org/wiki/Relational_operators> and >>> some built-in functions) >>> - Cloudera Impala >>> >>> <http://www.cloudera.com/content/cloudera/en/products-and-services/cdh/impala.html> >>> (significant HiveQL support, some SQL language support, no support for >>> indexes on its tables, importantly missing DELETE, UPDATE and INTERSECT; >>> amongst others) >>> - Presto <https://github.com/facebook/presto> from Facebook (can >>> query Hive, Cassandra <http://cassandra.apache.org>, relational DBs >>> &etc. Doesn't seem to be designed for low-latency responses across small >>> clusters, or support UPDATE operations. It is optimized for data >>> warehousing or analytics¹ >>> <http://prestodb.io/docs/current/overview/use-cases.html>) >>> - SQL-Hadoop <https://www.mapr.com/why-hadoop/sql-hadoop> via MapR >>> community edition <https://www.mapr.com/products/hadoop-download> >>> (seems to be a packaging of Hive, HP Vertica >>> <http://www.vertica.com/hp-vertica-products/sqlonhadoop>, SparkSQL, >>> Drill and a native ODBC wrapper >>> <http://package.mapr.com/tools/MapR-ODBC/MapR_ODBC>) >>> - Apache Kylin <http://www.kylin.io> from Ebay (provides an SQL >>> interface and multi-dimensional analysis [OLAP >>> <http://en.wikipedia.org/wiki/OLAP>], "… offers ANSI SQL on Hadoop >>> and supports most ANSI SQL query functions". It depends on HDFS, >>> MapReduce, >>> Hive and HBase; and seems targeted at very large data-sets though >>> maintains >>> low query latency) >>> - Apache Tajo <http://tajo.apache.org> (ANSI/ISO SQL standard >>> compliance with JDBC <http://en.wikipedia.org/wiki/JDBC> driver >>> support [benchmarks against Hive and Impala >>> >>> <http://blogs.gartner.com/nick-heudecker/apache-tajo-enters-the-sql-on-hadoop-space> >>> ]) >>> - Cascading <http://en.wikipedia.org/wiki/Cascading_%28software%29>'s >>> Lingual <http://docs.cascading.org/lingual/1.0/>² >>> <http://docs.cascading.org/lingual/1.0/#sql-support> ("Lingual >>> provides JDBC Drivers, a SQL command shell, and a catalog manager for >>> publishing files [or any resource] as schemas and tables.") >>> >>> Which—from this list or elsewhere—would you recommend, and why? >>> Thanks for all suggestions, >>> >>> Samuel Marks >>> http://linkedin.com/in/samuelmarks >>> >>> >> >