Hello, Xin. Thank you for an answer.
Is there any plans to make catalog API public? Any specific release versions or dates? 2017-09-25 20:54 GMT+03:00 Reynold Xin <r...@databricks.com>: > It's probably just an indication of lack of interest (or at least there > isn't a substantial overlap between Ignite users and Spark users). A new > catalog implementation is also pretty fundamental to Spark and the bar for > that would be pretty high. See my comment in SPARK-17767. > > Guys - while I think this is very useful to do, I'm going to mark this as > "later" for now. The reason is that there are a lot of things to consider > before making this switch, including: > > - The ExternalCatalog API is currently internal, and we can't just > make it public without thinking about the consequences and whether this API > is maintainable in the long run. > - SPARK-15777 <https://issues.apache.org/jira/browse/SPARK-15777> We > need to design this in the context of catalog federation and persistence. > - SPARK-15691 <https://issues.apache.org/jira/browse/SPARK-15691> > Refactoring > of how we integrate with Hive. > > This is not as simple as just submitting a PR to make it pluggable. > > On Mon, Sep 25, 2017 at 10:50 AM, Николай Ижиков <nizhikov....@gmail.com> > wrote: > >> Guys. >> >> Am I miss something and wrote a fully wrong mail? >> Can you give me some feedback? >> What I have missed with my propositions? >> >> 2017-09-19 15:39 GMT+03:00 Nikolay Izhikov <nizhikov....@gmail.com>: >> >>> Guys, >>> >>> Anyone had a chance to look at my message? >>> >>> 15.09.2017 15:50, Nikolay Izhikov пишет: >>> >>> Hello, guys. >>>> >>>> I’m contributor of Apache Ignite project which is self-described as an >>>> in-memory computing platform. >>>> >>>> It has Data Grid features: distribute, transactional key-value store >>>> [1], Distributed SQL support [2], etc…[3] >>>> >>>> Currently, I’m working on integration between Ignite and Spark [4] >>>> I want to add support of Spark Data Frame API for Ignite. >>>> >>>> As far as Ignite is distributed store it would be useful to create >>>> implementation of Catalog [5] for an Apache Ignite. >>>> >>>> I see two ways to implement this feature: >>>> >>>> 1. Spark can provide API for any custom catalog implementation. As >>>> far as I can see there is a ticket for it [6]. It is closed with resolution >>>> “Later”. Is it suitable time to continue working on the ticket? How can I >>>> help with it? >>>> >>>> 2. I can provide an implementation of Catalog and other required >>>> API in the form of pull request in Spark, as it was implemented for Hive >>>> [7]. Can such pull request be acceptable? >>>> >>>> Which way is more convenient for Spark community? >>>> >>>> [1] https://ignite.apache.org/features/datagrid.html >>>> [2] https://ignite.apache.org/features/sql.html >>>> [3] https://ignite.apache.org/features.html >>>> [4] https://issues.apache.org/jira/browse/IGNITE-3084 >>>> [5] https://github.com/apache/spark/blob/master/sql/catalyst/src >>>> /main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala >>>> [6] https://issues.apache.org/jira/browse/SPARK-17767 >>>> [7] https://github.com/apache/spark/blob/master/sql/hive/src/mai >>>> n/scala/org/apache/spark/sql/hive/HiveExternalCatalog.scala >>>> >>> >> >> >> -- >> Nikolay Izhikov >> nizhikov....@gmail.com >> > > -- Nikolay Izhikov nizhikov....@gmail.com