Hi Manu,

Just to clarify:
- Are you proposing to create a user facing locking feature in Iceberg, or
just something something for internal use?

I think we shouldn't add locking to Iceberg's user facing scope in this
stage. A fully featured locking system has many more features that we need
(priorities, fairness, timeouts etc). I could be tempted when we are
talking about the REST catalog, but I think that should be further down the
road, if ever...

About using the tags:
- I whole-heartedly agree that using tags is not intuitive, and I see your
points in most of your arguments. OTOH, introducing new requirement
(locking mechanism) seems like a wrong direction to me.
- We already defined a requirement (atomic changes on the table) for the
Catalog implementations which could be used to archive our goal here.
- We also already store technical data in snapshot properties in Flink jobs
(JobId, OperatorId, CheckpointId). Maybe technical tags/table properties is
not a big stretch.

Or we can look at these tags or metadata as 'technical data' which is
internal to Iceberg, and shouldn't expressed on the external API. My
concern is:
- Would it be used often enough to worth the additional complexity?

Knowing that Spark compaction is struggling with the same issue is a good
indicator, but probably we would need more use cases for introducing a new
feature with this complexity, or simpler solution.

Thanks, Peter


On Mon, Apr 1, 2024, 10:18 Manu Zhang <owenzhang1...@gmail.com> wrote:

> What would the community think of exploiting tags for preventing
>> concurrent maintenance loop executions.
>
>
> This issue is not specific to Flink maintenance jobs. We have a service
> scheduling Spark maintenance jobs by watching table commits. When we don't
> check in-progress maintenance jobs for the same table, multiple jobs will
> run concurrently and have conflicts.
>
> Basically, I think we need a lock mechanism like the metastore lock
> <https://iceberg.apache.org/docs/nightly/configuration/#hadoop-configuration>
> if we want to handle it for users. However, using TAG doesn't look
> intuitive to me. We are also mixing user data with system metadata.
> Maybe we can define some general interfaces and leave the implementation
> to users in the first version.
>
> Regards,
> Manu
>
>
>
> On Fri, Mar 29, 2024 at 1:59 PM Péter Váry <peter.vary.apa...@gmail.com>
> wrote:
>
>> What would the community think of exploiting tags for preventing
>> concurrent maintenance loop executions.
>>
>> The issue:
>> Some maintenance tasks couldn't run parallel, like DeleteOrphanFiles vs.
>> ExpireSnapshots, or RewriteDataFiles vs. RewriteManifestFiles. We make
>> sure, not to run tasks started by a single trigger concurrently by
>> serializing them, but there are no loops in Flink, so we can't synchronize
>> tasks started by the next trigger.
>>
>> In the document, I describe that we need to rely on the user to ensure
>> that the rate limit is high enough to prevent concurrent triggers.
>>
>> Proposal:
>> When firing a trigger, RateLimiter could check and create an Iceberg
>> table tag [1] for the current table snapshot, with the name:
>> '__flink_maitenance'. When the execution finishes we remove this tag. The
>> RateLimiter keep accumulating changes, and doesn't fire new triggers until
>> it finds this tag on the table.
>> The solution relies on Flink 'forceNonParallel' to prevent concurrent
>> execution of placing the tag, and on Iceberg to store it.
>>
>> This not uses the tags as intended, but seems like a better solution than
>> adding/removing table properties which would clutter the table history with
>> technical data.
>>
>> Your thoughts? Any other suggestions/solutions would be welcome.
>>
>> Thanks,
>> Peter
>>
>> [1]
>> https://iceberg.apache.org/docs/latest/java-api-quickstart/#branching-and-tagging
>>
>> On Thu, Mar 28, 2024, 14:44 Péter Váry <peter.vary.apa...@gmail.com>
>> wrote:
>>
>>> Hi Team,
>>>
>>> As discussed on yesterday's community sync, I am working on adding a
>>> possibility to the Flink Iceberg connector to run maintenance tasks on the
>>> Iceberg tables. This will fix the small files issues and in the long run
>>> help compacting the high number of positional and equality deletes created
>>> by Flink tasks writing CDC data to Iceberg tables without the need of Spark
>>> in the infrastructure.
>>>
>>> I did some planning, prototyping and currently trying out the solution
>>> on a larger scale.
>>>
>>> I put together a document how my current solution looks like:
>>>
>>> https://docs.google.com/document/d/16g3vR18mVBy8jbFaLjf2JwAANuYOmIwr15yDDxovdnA/edit?usp=sharing
>>>
>>> I would love to hear your thoughts and feedback on this to find a good
>>> final solution.
>>>
>>> Thanks,
>>> Peter
>>>
>>

Reply via email to