We could support EXPECT statements in proposal 2 as long as we restricted it to known fields.
We are getting into implementation details now. Making unknown fields just a normal column introduces a number of problems. ZetaSQL doesn't support Map type. All our IOs would need to explicitly deal with that special column. There would be a lack of consistency between the various types (Avro, Proto, Json) which should all support this. We might also want something even more invasive: everything is an unknown field unless it is referenced in the SQL query. All of these options are possible. I guess we need someone who has time to work on it to write a proposal. On Tue, Dec 8, 2020 at 10:03 AM Reuven Lax <re...@google.com> wrote: > I'm not sure that we could support EXCEPT statements, as that would > require introspecting the unknown fields (what if the EXCEPT statement > matches a field that later is added as an unknown field?). IMO this sort of > behavior only makes sense on true pass-through queries. Anything that > modifies the input record would be tricky to support. > > Nested rows would work for proposal 2. You would need to make sure that > the unknown-fields map is recursively added to all nested rows, and you > would do this when you infer a schema from the avro schema. > > On Tue, Dec 8, 2020 at 9:58 AM Andrew Pilloud <apill...@google.com> wrote: > >> Proposal 1 would also interact poorly with SELECT * EXCEPT ... >> statements, which returns all columns except specific ones. Adding an >> unknown field does seem like a reasonable way to handle this. It probably >> needs to be something that is native to the Row type, so columns added to >> nested rows also work. >> >> Andrew >> >> On Tue, Dec 8, 2020 at 9:50 AM Reuven Lax <re...@google.com> wrote: >> >>> There's a difference between a fully dynamic schema and simply being >>> able to forward "unknown" fields to the output. >>> >>> A fully-dynamic schema is not really necessary unless we also had >>> dynamic SQL statements. Since the existing SQL statements do not reference >>> the new fields by name, there's no reason to add them to the main schema. >>> >>> However, if you have a SELECT * FROM WHERE XXXX statement that does no >>> aggregation, there's fundamentally no reason we couldn't forward the >>> messages exactly. In theory we could forward the exact bytes that are in >>> the input PCollection, which would necessarily forward the new fields. In >>> practice I believe that we convert the input messages to Beam Row objects >>> in order to evaluate the WHERE clause, and then convert back to Avro to >>> output those messages. I believe this is where we "lose" the unknown >>> messages,but this is an implementation artifact - in theory we could output >>> the original bytes whenever we see a SELECT *. This is not truly a dynamic >>> schema, since you can't really do anything with these extra fields except >>> forward them to your output. >>> >>> I see two possible ways to address this. >>> >>> 1. As I mentioned above, in the case of a SELECT * we could output the >>> original bytes, and only use the Beam Row for evaluating the WHERE clause. >>> This might be very expensive though - we risk having to keep two copies of >>> every message around, one in the original Avro format and one in Row format. >>> >>> 2. The other way would be to do what protocol buffers do. We could add >>> one extra field to the inferred Beam schema to store new, unknown fields >>> (probably this would be a map-valued field). This extra field would simply >>> store the raw bytes of these unknown fields, and then when converting back >>> to Avro they would be added to the output message. This might also add some >>> overhead to the pipeline, so might be best to make this behavior opt in. >>> >>> Reuven >>> >>> On Tue, Dec 8, 2020 at 9:33 AM Brian Hulette <bhule...@google.com> >>> wrote: >>> >>>> Reuven, could you clarify what you have in mind? I know multiple times >>>> we've discussed the possibility of adding update compatibility support to >>>> SchemaCoder, including support for certain schema changes (field >>>> additions/deletions) - I think the most recent discussion was here [1]. >>>> >>>> But it sounds like Talat is asking for something a little beyond that, >>>> effectively a dynamic schema. Is that something you think we can support? >>>> >>>> [1] >>>> https://lists.apache.org/thread.html/ref73a8c40e24e0b038b4e5b065cd502f4c5df2e5e15af6f7ea1cdaa7%40%3Cdev.beam.apache.org%3E >>>> >>>> On Tue, Dec 8, 2020 at 9:20 AM Reuven Lax <re...@google.com> wrote: >>>> >>>>> Thanks. It might be theoretically possible to do this (at least for >>>>> the case where existing fields do not change). Whether anyone currently >>>>> has >>>>> available time to do this is a different question, but it's something that >>>>> can be looked into. >>>>> >>>>> On Mon, Dec 7, 2020 at 9:29 PM Talat Uyarer < >>>>> tuya...@paloaltonetworks.com> wrote: >>>>> >>>>>> Adding new fields is more common than modifying existing fields. But >>>>>> type change is also possible for existing fields, such as regular >>>>>> mandatory >>>>>> field(string,integer) to union(nullable field). No field deletion. >>>>>> >>>>>> On Mon, Dec 7, 2020 at 9:22 PM Reuven Lax <re...@google.com> wrote: >>>>>> >>>>>>> And when you say schema changes, are these new fields being added to >>>>>>> the schema? Or are you making changes to the existing fields? >>>>>>> >>>>>>> On Mon, Dec 7, 2020 at 9:02 PM Talat Uyarer < >>>>>>> tuya...@paloaltonetworks.com> wrote: >>>>>>> >>>>>>>> Hi, >>>>>>>> For sure let me explain a little bit about my pipeline. >>>>>>>> My Pipeline is actually simple >>>>>>>> Read Kafka -> Convert Avro Bytes to Beam Row(DoFn<KV<byte[], byte[]>, >>>>>>>> Row>) -> Apply Filter(SqlTransform.query(sql)) -> Convert back >>>>>>>> from Row to Avro (DoFn<Row, byte[]>)-> Write DB/GCS/GRPC etc >>>>>>>> >>>>>>>> On our jobs We have three type sqls >>>>>>>> - SELECT * FROM PCOLLECTION >>>>>>>> - SELECT * FROM PCOLLECTION <with Where Condition> >>>>>>>> - SQL Projection with or without Where clause SELECT col1, col2 >>>>>>>> FROM PCOLLECTION >>>>>>>> >>>>>>>> We know writerSchema for each message. While deserializing avro >>>>>>>> binary we use writer schema and reader schema on Convert Avro Bytes to >>>>>>>> Beam >>>>>>>> Row step. It always produces a reader schema's generic record and we >>>>>>>> convert that generic record to Row. >>>>>>>> While submitting DF job we use latest schema to generate beamSchema. >>>>>>>> >>>>>>>> In the current scenario When we have schema changes first we >>>>>>>> restart all 15k jobs with the latest updated schema then whenever we >>>>>>>> are >>>>>>>> done we turn on the latest schema for writers. Because of Avro's >>>>>>>> GrammerResolver[1] we read different versions of the schema and we >>>>>>>> always >>>>>>>> produce the latest schema's record. Without breaking our pipeline we >>>>>>>> are >>>>>>>> able to handle multiple versions of data in the same streaming >>>>>>>> pipeline. If >>>>>>>> we can generate SQL's java code when we get notified wirth latest >>>>>>>> schema we >>>>>>>> will handle all schema changes. The only remaining obstacle is Beam's >>>>>>>> SQL >>>>>>>> Java code. That's why I am looking for some solution. We dont need >>>>>>>> multiple >>>>>>>> versions of SQL. We only need to regenerate SQL schema with the latest >>>>>>>> schema on the fly. >>>>>>>> >>>>>>>> I hope I can explain it :) >>>>>>>> >>>>>>>> Thanks >>>>>>>> >>>>>>>> [1] >>>>>>>> https://avro.apache.org/docs/1.7.2/api/java/org/apache/avro/io/parsing/doc-files/parsing.html >>>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__avro.apache.org_docs_1.7.2_api_java_org_apache_avro_io_parsing_doc-2Dfiles_parsing.html&d=DwMFaQ&c=V9IgWpI5PvzTw83UyHGVSoW3Uc1MFWe5J8PTfkrzVSo&r=BkW1L6EF7ergAVYDXCo-3Vwkpy6qjsWAz7_GD7pAR8g&m=0qahAe7vDisJq_hMYGY8F-Bp7-_5lOwOKzNoQ3r3-IQ&s=lwwIMsJO9nmj6_xZcSG_7qkBIaxOwyUXry4st1q70Rc&e=> >>>>>>>> >>>>>>>> On Mon, Dec 7, 2020 at 7:49 PM Reuven Lax <re...@google.com> wrote: >>>>>>>> >>>>>>>>> Can you explain the use case some more? Are you wanting to change >>>>>>>>> your SQL statement as well when the schema changes? If not, what are >>>>>>>>> those >>>>>>>>> new fields doing in the pipeline? What I mean is that your old SQL >>>>>>>>> statement clearly didn't reference those fields in a SELECT statement >>>>>>>>> since >>>>>>>>> they didn't exist, so what are you missing by not having them unless >>>>>>>>> you >>>>>>>>> are also changing the SQL statement? >>>>>>>>> >>>>>>>>> Is this a case where you have a SELECT *, and just want to make >>>>>>>>> sure those fields are included? >>>>>>>>> >>>>>>>>> Reuven >>>>>>>>> >>>>>>>>> On Mon, Dec 7, 2020 at 6:31 PM Talat Uyarer < >>>>>>>>> tuya...@paloaltonetworks.com> wrote: >>>>>>>>> >>>>>>>>>> Hi Andrew, >>>>>>>>>> >>>>>>>>>> I assume SQL query is not going to change. Changing things is the >>>>>>>>>> Row schema by adding new columns or rename columns. if we keep a >>>>>>>>>> version >>>>>>>>>> information on somewhere for example a KV pair. Key is schema >>>>>>>>>> information, >>>>>>>>>> value is Row. Can not we generate SQL code ? Why I am asking We have >>>>>>>>>> 15k >>>>>>>>>> pipelines. When we have a schema change we restart a 15k DF job >>>>>>>>>> which is >>>>>>>>>> pain. I am looking for a possible way to avoid job restart. Dont you >>>>>>>>>> think >>>>>>>>>> it is not still doable ? >>>>>>>>>> >>>>>>>>>> Thanks >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Dec 7, 2020 at 6:10 PM Andrew Pilloud < >>>>>>>>>> apill...@google.com> wrote: >>>>>>>>>> >>>>>>>>>>> Unfortunately we don't have a way to generate the SQL Java code >>>>>>>>>>> on the fly, even if we did, that wouldn't solve your problem. I >>>>>>>>>>> believe our >>>>>>>>>>> recommended practice is to run both the old and new pipeline for >>>>>>>>>>> some time, >>>>>>>>>>> then pick a window boundary to transition the output from the old >>>>>>>>>>> pipeline >>>>>>>>>>> to the new one. >>>>>>>>>>> >>>>>>>>>>> Beam doesn't handle changing the format of data sent between >>>>>>>>>>> intermediate steps in a running pipeline. Beam uses "coders" to >>>>>>>>>>> serialize >>>>>>>>>>> data between steps of the pipeline. The builtin coders (including >>>>>>>>>>> the >>>>>>>>>>> Schema Row Coder used by SQL) have a fixed data format and don't >>>>>>>>>>> handle >>>>>>>>>>> schema evolution. They are optimized for performance at all costs. >>>>>>>>>>> >>>>>>>>>>> If you worked around this, the Beam model doesn't support >>>>>>>>>>> changing the structure of the pipeline graph. This would >>>>>>>>>>> significantly >>>>>>>>>>> limit the changes you can make. It would also require some changes >>>>>>>>>>> to SQL >>>>>>>>>>> to try to produce the same plan for an updated SQL query. >>>>>>>>>>> >>>>>>>>>>> Andrew >>>>>>>>>>> >>>>>>>>>>> On Mon, Dec 7, 2020 at 5:44 PM Talat Uyarer < >>>>>>>>>>> tuya...@paloaltonetworks.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> We are using Beamsql on our pipeline. Our Data is written in >>>>>>>>>>>> Avro format. We generate our rows based on our Avro schema. Over >>>>>>>>>>>> time the >>>>>>>>>>>> schema is changing. I believe Beam SQL generates Java code based >>>>>>>>>>>> on what we >>>>>>>>>>>> define as BeamSchema while submitting the pipeline. Do you have >>>>>>>>>>>> any idea >>>>>>>>>>>> How can we handle schema changes with resubmitting our beam job. >>>>>>>>>>>> Is it >>>>>>>>>>>> possible to generate SQL java code on the fly ? >>>>>>>>>>>> >>>>>>>>>>>> Thanks >>>>>>>>>>>> >>>>>>>>>>>