That's kind of solution Ed, can you elaborate how can I do this on Spark side? Or do I need to update table configuration in the DB
Siavash On Wed, Aug 12, 2020 at 5:55 PM ed elliott <ed.elli...@outlook.com> wrote: > You’ll need to do an insert and use a trigger on the table to change it > into an upsert, also make sure your mode is append rather than overwrite. > > Ed > > ------------------------------ > *From:* Siavash Namvar <sns...@gmail.com> > *Sent:* Wednesday, August 12, 2020 4:09:07 PM > *To:* Sean Owen <sro...@gmail.com> > *Cc:* User <user@spark.apache.org> > *Subject:* Re: How can I use pyspark to upsert one row without replacing > entire table > > Thanks Sean, > > Do you have any URL or reference to help me how to upsert in Spark? I need > to update Sybase db > > On Wed, Aug 12, 2020 at 11:06 AM Sean Owen <sro...@gmail.com> wrote: > > It's not so much Spark but the data format, whether it supports > upserts. Parquet, CSV, JSON, etc would not. > That is what Delta, Hudi et al are for, and yes you can upsert them in > Spark. > > On Wed, Aug 12, 2020 at 9:57 AM Siavash Namvar <sns...@gmail.com> wrote: > > > > Hi, > > > > I have a use case, and read data from a db table and need to update few > rows based on primary key without replacing the entire table. > > > > for instance if I have 3 following rows > > > > ------------------- > > id | fname > > ------------------- > > 1 | john > > ------------------- > > 2 | Steve > > ------------------- > > 3 | Jack > > ------------------- > > > > And I would like to update the row with id=2 from Steve to Michael > without replacing the entire table and the outpur looks like > > > > ------------------- > > id | fname > > ------------------- > > 1 | john > > ------------------- > > 2 | Michael > > ------------------- > > 3 | Jack > > ------------------- > > > > Keep in mind the actual db table is so huge and database is old and > cannot read and replace entire table > > > > Thanks > >