Thanks Mohammed,

I was aware of Calliope, but haven't used it since with since the
spark-cassandra-connector project got released.  I was not aware of the
CalliopeServer2; cool thanks for sharing that one.

I would appreciate it if you could lmk how you decide to proceed with this;
I can see this coming up on my radar in the next few months; thanks.

-Todd

On Fri, Apr 3, 2015 at 5:53 PM, Mohammed Guller <moham...@glassbeam.com>
wrote:

>  Thanks, Todd.
>
>
>
> It is an interesting idea; worth trying.
>
>
>
> I think the cash project is old. The tuplejump guy has created another
> project called CalliopeServer2, which works like a charm with BI tools that
> use JDBC, but unfortunately Tableau throws an error when it connects to it.
>
>
>
> Mohammed
>
>
>
> *From:* Todd Nist [mailto:tsind...@gmail.com]
> *Sent:* Friday, April 3, 2015 11:39 AM
> *To:* pawan kumar
> *Cc:* Mohammed Guller; user@spark.apache.org
>
> *Subject:* Re: Tableau + Spark SQL Thrift Server + Cassandra
>
>
>
> Hi Mohammed,
>
>
>
> Not sure if you have tried this or not.  You could try using the below api
> to start the thriftserver with an existing context.
>
>
> https://github.com/apache/spark/blob/master/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2.scala#L42
>
> The one thing that Michael Ambrust @ databrick recommended was this:
>
> You can start a JDBC server with an existing context.  See my answer here:
> http://apache-spark-user-list.1001560.n3.nabble.com/Standard-SQL-tool-access-to-SchemaRDD-td20197.html
>
> So something like this based on example from Cheng Lian:
>
>
> * Server*
>
> import  org.apache.spark.sql.hive.HiveContext
>
> import  org.apache.spark.sql.catalyst.types._
>
>
>
> val  sparkContext  =  sc
>
> import  sparkContext._
>
> val  sqlContext  =  new  HiveContext(sparkContext)
>
> import  sqlContext._
>
> makeRDD((1,"hello") :: (2,"world") 
> ::Nil).toSchemaRDD.cache().registerTempTable("t")
>
> // replace the above with the C* + spark-casandra-connectore to generate 
> SchemaRDD and registerTempTable
>
>
>
> import  org.apache.spark.sql.hive.thriftserver._
>
> HiveThriftServer2.startWithContext(sqlContext)
>
>   Then Startup
>
> ./bin/beeline -u jdbc:hive2://localhost:10000/default
>
> 0: jdbc:hive2://localhost:10000/default> select * from t;
>
>
>
>   I have not tried this yet from Tableau.   My understanding is that the
> tempTable is only valid as long as the sqlContext is, so if one terminates
> the code representing the *Server*, and then restarts the standard thrift
> server, sbin/start-thriftserver ..., the table won't be available.
>
>
>
> Another possibility is to perhaps use the tuplejump cash project,
> https://github.com/tuplejump/cash.
>
>
>
> HTH.
>
>
>
> -Todd
>
>
>
> On Fri, Apr 3, 2015 at 11:11 AM, pawan kumar <pkv...@gmail.com> wrote:
>
> Thanks mohammed. Will give it a try today. We would also need the
> sparksSQL piece as we are migrating our data store from oracle to C* and it
> would be easier to maintain all the reports rather recreating each one from
> scratch.
>
> Thanks,
> Pawan Venugopal.
>
> On Apr 3, 2015 7:59 AM, "Mohammed Guller" <moham...@glassbeam.com> wrote:
>
> Hi Todd,
>
>
>
> We are using Apache C* 2.1.3, not DSE. We got Tableau to work directly
> with C* using the ODBC driver, but now would like to add Spark SQL to the
> mix. I haven’t been able to find any documentation for how to make this
> combination work.
>
>
>
> We are using the Spark-Cassandra-Connector in our applications, but
> haven’t been able to figure out how to get the Spark SQL Thrift Server to
> use it and connect to C*. That is the missing piece. Once we solve that
> piece of the puzzle then Tableau should be able to see the tables in C*.
>
>
>
> Hi Pawan,
>
> Tableau + C* is pretty straight forward, especially if you are using DSE.
> Create a new DSN in Tableau using the ODBC driver that comes with DSE. Once
> you connect, Tableau allows to use C* keyspace as schema and column
> families as tables.
>
>
>
> Mohammed
>
>
>
> *From:* pawan kumar [mailto:pkv...@gmail.com]
> *Sent:* Friday, April 3, 2015 7:41 AM
> *To:* Todd Nist
> *Cc:* user@spark.apache.org; Mohammed Guller
> *Subject:* Re: Tableau + Spark SQL Thrift Server + Cassandra
>
>
>
> Hi Todd,
>
> Thanks for the link. I would be interested in this solution. I am using
> DSE for cassandra. Would you provide me with info on connecting with DSE
> either through Tableau or zeppelin. The goal here is query cassandra
> through spark sql so that I could perform joins and groupby on my queries.
> Are you able to perform spark sql queries with tableau?
>
> Thanks,
> Pawan Venugopal
>
> On Apr 3, 2015 5:03 AM, "Todd Nist" <tsind...@gmail.com> wrote:
>
> What version of Cassandra are you using?  Are you using DSE or the stock
> Apache Cassandra version?  I have connected it with DSE, but have not
> attempted it with the standard Apache Cassandra version.
>
>
>
> FWIW,
> http://www.datastax.com/dev/blog/datastax-odbc-cql-connector-apache-cassandra-datastax-enterprise,
> provides an ODBC driver tor accessing C* from Tableau.  Granted it does not
> provide all the goodness of Spark.  Are you attempting to leverage the
> spark-cassandra-connector for this?
>
>
>
>
>
>
>
> On Thu, Apr 2, 2015 at 10:20 PM, Mohammed Guller <moham...@glassbeam.com>
> wrote:
>
> Hi –
>
>
>
> Is anybody using Tableau to analyze data in Cassandra through the Spark
> SQL Thrift Server?
>
>
>
> Thanks!
>
>
>
> Mohammed
>
>
>
>
>
>
>

Reply via email to