+1 (non-binding)

El jue, 10 abr 2025, 1:50, Burak Yavuz <brk...@gmail.com> escribió:

> +1
>
> On Wed, Apr 9, 2025 at 4:33 PM Szehon Ho <szehon.apa...@gmail.com> wrote:
>
>> +1 really excited to finally see Materialized View finally make its way
>> to Spark, as many other ecosystem projects (Trino, Starrocks, soon Iceberg)
>> already supporting it.
>>
>> Thanks
>> Szehon
>>
>> On Wed, Apr 9, 2025 at 2:33 AM Martin Grund <mar...@databricks.com.invalid>
>> wrote:
>>
>>> +1
>>>
>>> On Wed, Apr 9, 2025 at 9:37 AM Mich Talebzadeh <
>>> mich.talebza...@gmail.com> wrote:
>>>
>>>> +1
>>>>
>>>> Dr Mich Talebzadeh,
>>>> Architect | Data Science | Financial Crime | Forensic Analysis | GDPR
>>>>
>>>>    view my Linkedin profile
>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Wed, 9 Apr 2025 at 08:07, Peter Toth <peter.t...@gmail.com> wrote:
>>>>
>>>>> +1
>>>>>
>>>>> On Wed, Apr 9, 2025 at 8:51 AM Cheng Pan <pan3...@gmail.com> wrote:
>>>>>
>>>>>> +1 (non-binding)
>>>>>>
>>>>>> Glad to see Spark SQL extended to streaming use cases.
>>>>>>
>>>>>> Thanks,
>>>>>> Cheng Pan
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Apr 9, 2025, at 14:43, Anton Okolnychyi <aokolnyc...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>> +1
>>>>>>
>>>>>> вт, 8 квіт. 2025 р. о 23:36 Jacky Lee <qcsd2...@gmail.com> пише:
>>>>>>
>>>>>>> +1 I'm delighted that it will be open-sourced, enabling greater
>>>>>>> integration with Iceberg/Delta to unlock more value.
>>>>>>>
>>>>>>> Jungtaek Lim <kabhwan.opensou...@gmail.com> 于2025年4月9日周三 10:47写道:
>>>>>>> >
>>>>>>> > +1 looking forward to seeing this make progress!
>>>>>>> >
>>>>>>> > On Wed, Apr 9, 2025 at 11:32 AM Yang Jie <yangji...@apache.org>
>>>>>>> wrote:
>>>>>>> >>
>>>>>>> >> +1
>>>>>>> >>
>>>>>>> >> On 2025/04/09 01:07:57 Hyukjin Kwon wrote:
>>>>>>> >> > +1
>>>>>>> >> >
>>>>>>> >> > I am actually pretty excited to have this. Happy to see this
>>>>>>> being proposed.
>>>>>>> >> >
>>>>>>> >> > On Wed, 9 Apr 2025 at 01:55, Chao Sun <sunc...@apache.org>
>>>>>>> wrote:
>>>>>>> >> >
>>>>>>> >> > > +1. Super excited about this effort!
>>>>>>> >> > >
>>>>>>> >> > > On Tue, Apr 8, 2025 at 9:47 AM huaxin gao <
>>>>>>> huaxin.ga...@gmail.com> wrote:
>>>>>>> >> > >
>>>>>>> >> > >> +1 I support this SPIP because it simplifies data pipeline
>>>>>>> management and
>>>>>>> >> > >> enhances error detection.
>>>>>>> >> > >>
>>>>>>> >> > >>
>>>>>>> >> > >> On Tue, Apr 8, 2025 at 9:33 AM Dilip Biswal <
>>>>>>> dkbis...@gmail.com> wrote:
>>>>>>> >> > >>
>>>>>>> >> > >>> Excited to see this heading toward open source —
>>>>>>> materialized views and
>>>>>>> >> > >>> other features will bring a lot of value.
>>>>>>> >> > >>> +1 (non-binding)
>>>>>>> >> > >>>
>>>>>>> >> > >>> On Mon, Apr 7, 2025 at 10:37 AM Sandy Ryza <
>>>>>>> sa...@apache.org> wrote:
>>>>>>> >> > >>>
>>>>>>> >> > >>>> Hi Khalid – the CLI in the current proposal will need to
>>>>>>> be built on
>>>>>>> >> > >>>> top of internal APIs for constructing and launching
>>>>>>> pipeline executions.
>>>>>>> >> > >>>> We'll have the option to expose these in the future.
>>>>>>> >> > >>>>
>>>>>>> >> > >>>> It would be worthwhile to understand the use cases in more
>>>>>>> depth before
>>>>>>> >> > >>>> exposing these, because APIs are one-way doors and can be
>>>>>>> costly to
>>>>>>> >> > >>>> maintain.
>>>>>>> >> > >>>>
>>>>>>> >> > >>>> On Sat, Apr 5, 2025 at 11:59 PM Khalid Mammadov <
>>>>>>> >> > >>>> khalidmammad...@gmail.com> wrote:
>>>>>>> >> > >>>>
>>>>>>> >> > >>>>> Looks great!
>>>>>>> >> > >>>>> QQ: will user able to run this pipeline from normal code?
>>>>>>> I.e. can I
>>>>>>> >> > >>>>> trigger a pipeline from *driver* code based on some
>>>>>>> condition etc. or
>>>>>>> >> > >>>>> it must be executed via separate shell command ?
>>>>>>> >> > >>>>> As a background Databricks imposes similar limitation
>>>>>>> where as you
>>>>>>> >> > >>>>> cannot run normal Spark code and DLT on the same cluster
>>>>>>> for some reason
>>>>>>> >> > >>>>> and forces to use two clusters increasing the cost and
>>>>>>> latency.
>>>>>>> >> > >>>>>
>>>>>>> >> > >>>>> On Sat, 5 Apr 2025 at 23:03, Sandy Ryza <sa...@apache.org>
>>>>>>> wrote:
>>>>>>> >> > >>>>>
>>>>>>> >> > >>>>>> Hi all – starting a discussion thread for a SPIP that
>>>>>>> I've been
>>>>>>> >> > >>>>>> working on with Chao Sun, Kent Yao, Yuming Wang, and Jie
>>>>>>> Yang: [JIRA
>>>>>>> >> > >>>>>> <https://issues.apache.org/jira/browse/SPARK-51727>]
>>>>>>> [Doc
>>>>>>> >> > >>>>>> <
>>>>>>> https://docs.google.com/document/d/1PsSTngFuRVEOvUGzp_25CQL1yfzFHFr02XdMfQ7jOM4/edit?tab=t.0
>>>>>>> >
>>>>>>> >> > >>>>>> ].
>>>>>>> >> > >>>>>>
>>>>>>> >> > >>>>>> The SPIP proposes extending Spark's lazy, declarative
>>>>>>> execution model
>>>>>>> >> > >>>>>> beyond single queries, to pipelines that keep multiple
>>>>>>> datasets up to date.
>>>>>>> >> > >>>>>> It introduces the ability to compose multiple
>>>>>>> transformations into a single
>>>>>>> >> > >>>>>> declarative dataflow graph.
>>>>>>> >> > >>>>>>
>>>>>>> >> > >>>>>> Declarative pipelines aim to simplify the development
>>>>>>> and management
>>>>>>> >> > >>>>>> of data pipelines, by  removing the need for manual
>>>>>>> orchestration of
>>>>>>> >> > >>>>>> dependencies and making it possible to catch many errors
>>>>>>> before any
>>>>>>> >> > >>>>>> execution steps are launched.
>>>>>>> >> > >>>>>>
>>>>>>> >> > >>>>>> Declarative pipelines can include both batch and
>>>>>>> streaming
>>>>>>> >> > >>>>>> computations, leveraging Structured Streaming for stream
>>>>>>> processing and new
>>>>>>> >> > >>>>>> materialized view syntax for batch processing. Tight
>>>>>>> integration with Spark
>>>>>>> >> > >>>>>> SQL's analyzer enables deeper analysis and earlier error
>>>>>>> detection than is
>>>>>>> >> > >>>>>> achievable with more generic frameworks.
>>>>>>> >> > >>>>>>
>>>>>>> >> > >>>>>> Let us know what you think!
>>>>>>> >> > >>>>>>
>>>>>>> >> > >>>>>>
>>>>>>> >> >
>>>>>>> >>
>>>>>>> >>
>>>>>>> ---------------------------------------------------------------------
>>>>>>> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>> >>
>>>>>>>
>>>>>>> ---------------------------------------------------------------------
>>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>>
>>>>>>>
>>>>>>

Reply via email to