> I'm wondering if you'd have any recommendations of how to approach this barring maintaining our own spark fork?
You can probably get everything working in Spark SQL extensions. That allows you to add rules to certain batches in the analyzer, which is all you need. The implementation that John has been trying to get into Spark is nice because it moves all view resolution to its own batch before the main analysis batch, which I'm not sure you'd be able to do. But that's the approach I'd take. Ryan On Tue, Nov 15, 2022 at 3:21 PM Marc Laforet <mlafor...@gmail.com> wrote: > Hey guys, > > Thanks for the responses. > > Ryan - Thanks for confirming the behaviour. I'm wondering if you'd have > any recommendations of how to approach this barring maintaining our own > spark fork? > > Walaa - I tried creating the view using spark sql's standard `create view > as select` statement (trying with the fully qualified table name as well as > first setting catalog & namespace). Our iceberg tables are backed by a HMS > so it would presumably be stored there? > > Thanks again for your responses! > > On Tue, Nov 15, 2022 at 5:38 PM Walaa Eldin Moustafa < > wa.moust...@gmail.com> wrote: > >> Hi Marc, >> >> Could you clarify where you store the view definitions in this case, and >> how the syntax looks like? >> >> Thanks, >> Walaa. >> >> >> On Tue, Nov 15, 2022 at 2:34 PM Ryan Blue <b...@tabular.io> wrote: >> >>> Hi Marc, >>> >>> This is expected. Although the ViewCatalog SPIP was approved by the >>> Spark community, the implementation hasn't made it in yet for v2. >>> >>> Ryan >>> >>> On Tue, Nov 15, 2022 at 11:38 AM Marc Laforet <mlafor...@gmail.com> >>> wrote: >>> >>>> Hi Iceberg folks, >>>> >>>> I'm working on a project where we're migrating tables from hive to >>>> iceberg. We are revamping our ingestion pipeline in parallel from batch to >>>> stream. Originally, our plan was to have two separate tables, a backfill >>>> table and a live table, that would be stitched together via a view for >>>> downstream consumers. This is proving rather difficult. In the absence of >>>> engine agnostic views we were going to prepend views with the engine type >>>> (ie trino_my_table and spark_my_table) but I receive a >>>> org.apache.spark.sql.AnalysisException: >>>> Catalog iceberg_catalog does not support views error when trying to >>>> create the spark view. With the ongoing work towards engine agnostic views >>>> I'm unsure if this limitation is expected or easily surpassed with some >>>> config/spark change? >>>> >>>> Thank you for your time, >>>> >>>> Marc >>>> >>> >>> >>> -- >>> Ryan Blue >>> Tabular >>> >> -- Ryan Blue Tabular