Pucheng,

Rather than using the changelog, I'd just look at the metadata tables. You
should be able to query the all_entries metadata table to see file
additions or deletions for a given snapshot. Then from there you can join
to the snapshots table for timestamps and aggregate to the partition level.

Ryan

On Fri, Apr 28, 2023 at 12:49 PM Pucheng Yang <py...@pinterest.com.invalid>
wrote:

> Hi Ajantha and the community,
>
> I am interested and I am wondering where we can see the latest progress of
> this feature?
>
> Regarding the partition stats in Iceberg, I am specifically curious if we
> can consider a new field called "last modified time" to be included for the
> partitions stats (or have a plugable way to allow users to
> configure partition stats they need). My use case is to find out if a
> partition is changed or not given two snapshots (old and new) with a
> quick and light way process. I previously was suggested by the community to
> use the change log (CDC) but I think that is too heavy (I guess, since it
> requires to run SparkSQL procedure) and it is over do the work (I don't
> need what rows are changed, I just need true or false for whether a
> partition is changed).
>
> Thanks
>
> On Tue, Feb 7, 2023 at 11:36 AM Mayur Srivastava <
> mayur.srivast...@twosigma.com> wrote:
>
>> Thanks Ajantha.
>>
>>
>>
>> > It should be very easy to add a few more fields to it like the latest
>> sequence number or last modified time per partition.
>>
>>
>>
>> Among sequence number and modified time, which one do you think is more
>> likely to be available in Iceberg partition stats? Note that we would like
>> to avoid compaction change the sequence number or modified time stats.
>>
>>
>>
>> Thanks,
>>
>> Mayur
>>
>>
>>
>> *From:* Ajantha Bhat <ajanthab...@gmail.com>
>> *Sent:* Tuesday, February 7, 2023 10:02 AM
>> *To:* dev@iceberg.apache.org
>> *Subject:* Re: [Proposal] Partition stats in Iceberg
>>
>>
>>
>> Hi Hrishi and Mayur, thanks for the inputs.
>>
>> To get things moving I have frozen the scope of phase 1 implementation.
>> (Recently added the delete file stats to phase 1 too). You can find the
>> scope in the "Design for approval" section of the design doc.
>>
>> That said, once we have phase 1 implemented, It should be very easy to
>> add a few more fields to it like the latest sequence number or last
>> modified time per partition.
>> I will be opening up the discussion about phase 2 schema again once phase
>> 1 implementation is done.
>>
>> Thanks,
>> Ajantha
>>
>>
>>
>> On Tue, Feb 7, 2023 at 8:15 PM Mayur Srivastava <
>> mayur.srivast...@twosigma.com> wrote:
>>
>> +1 for the initiative.
>>
>>
>>
>> We’ve been exploring options for storing last-modified-time per
>> partition. It an important building block for data pipelines – especially
>> if there is a dependency between jobs with strong consistency requirements.
>>
>>
>>
>> Is partition stats a good place for storing last-modified-time per
>> partition?
>>
>>
>>
>> Thanks,
>>
>> Mayur
>>
>>
>>
>> *From:* Ajantha Bhat <ajanthab...@gmail.com>
>> *Sent:* Monday, January 23, 2023 11:56 AM
>> *To:* dev@iceberg.apache.org
>> *Subject:* Re: [Proposal] Partition stats in Iceberg
>>
>>
>>
>> Hi All,
>>
>> In the same design document (
>> https://docs.google.com/document/d/1vaufuD47kMijz97LxM67X8OX-W2Wq7nmlz3jRo8J5Qk/edit?usp=sharing
>> ),
>> I have added a section called
>> *"Design for approval".  *It also contains a potential PR breakdown for
>> the phase 1 implementation and future development scope.
>> Please take a look and please vote if you think the design is ok.
>>
>> Thanks,
>> Ajantha
>>
>>
>>
>> On Mon, Dec 5, 2022 at 8:37 PM Ajantha Bhat <ajanthab...@gmail.com>
>> wrote:
>>
>> A big thanks to everyone who was involved in the review and the
>> discussions so far.
>>
>> Please find the meeting minutes from the last iceberg sync about the
>> partition stats.
>>     a. Writers should not write the partition stats or any stats as of
>> now.
>>         Because it requires bumping the spec to V3. (We can have it as
>> part of the v3 spec later on. But not anytime soon).
>>     b. So, there can be an async way of generating the stats like ANALYZE
>> table or call procedure.
>>         Which will compute the stats till the current snapshot and store
>> it as a partition stats file.
>>     c. In phase 1, partition stats will just store the row_count and
>> file_count per partition value as mentioned in the design document.
>>         Later it can be enhanced to store puffin file location and other
>> metrics per partition value.
>>     d. These tuples are stored in a single sorted Avro/parquet file (we
>> need to finalize this).
>>     e. Each time "analyze table" will rewrite the whole stats file as
>> keeping multiple delta files will just make the read path messy.
>>         Also, even with million rows, it can be of a few MB size.
>>         Once the writers start writing the stats (V3 spec), we can
>> revisit storing as the delta files if there are any performance issues.
>>
>> The next immediate plan is to
>>     a. Get these PRs merged (open points in existing StatictisFile
>> interface added during Puffin)
>>         #6267 <https://github.com/apache/iceberg/pull/6267>, #6090
>> <https://github.com/apache/iceberg/pull/6090>, #6091
>> <https://github.com/apache/iceberg/pull/6091>
>>     b. Figure out how to give accurate stats with row-level deletes and
>> how to mask dropped partition values from stats.
>>         https://github.com/apache/iceberg/issues/6042
>>     c. Standardize the `StatictisFile` interface to hold the parquet/Avro
>> stats file (instead of always assuming it as a Puffin file)
>>         and introduce a `StatisticsType` enum.
>>     d. Conclude the storage format and get approval for the design.
>>
>> I will wait another week or two for some more people to take a look at
>> the document
>>
>> before jumping into the implementation.
>>
>> Thanks,
>> Ajantha.
>>
>>
>>
>> On Sat, Nov 26, 2022 at 8:25 AM Ajantha Bhat <ajanthab...@gmail.com>
>> wrote:
>>
>> Hi Ryan,
>>
>> are you saying that you think the partition-level stats should not be
>> required? I think that would be best.
>>
>> I think there is some confusion here. Partition-level stats are
>> required (hence the proposal).
>> But does the writer always write it? (with the append/delete/replace
>> operation)
>> or writer skips writing it and then the user generates it using DML like
>> "Analyze table" was the point of discussion.
>> I think we can have both options with the writer stats writing controlled
>> by a table property "write.stats.enabled"
>>
>>
>>
>> I’m all for improving the interface for retrieving stats. It’s a separate
>> issue
>>
>> Agree. Let us discuss it in a separate thread.
>>
>> Thanks,
>> Ajantha
>>
>>
>>
>> On Sat, Nov 26, 2022 at 12:12 AM Ryan Blue <b...@tabular.io> wrote:
>>
>> Ajantha, are you saying that you think the partition-level stats should
>> not be required? I think that would be best.
>>
>> I’m all for improving the interface for retrieving stats. It’s a separate
>> issue, but I think that Iceberg should provide both access to the Puffin
>> files and metadata as well as a higher-level interface for retrieving
>> information like a column’s NDV. Something like this:
>>
>> int ndv = 
>> table.findStat(Statistics.NDV).limitSnapshotDistance(3).forColumn("x");
>>
>>
>>
>> On Thu, Nov 24, 2022 at 2:31 AM Ajantha Bhat <ajanthab...@gmail.com>
>> wrote:
>>
>> Hi Ryan,
>> Thanks a lot for the review and suggestions.
>>
>> but I think there is also a decision that we need to make before that:
>> Should Iceberg require writers to maintain the partition stats?
>>
>> I think I would prefer to take a lazy approach and not assume that
>> writers will keep the partition stats up to date,
>>
>> in which case we need a way to know which parts of a table are newer than
>> the most recent stats.
>>
>>
>>
>> This is a common problem for existing table-level puffin stats too.  Not
>> just for partition stats.
>> As mentioned in the "integration with the current code" section point 8),
>> I was planning to introduce a table property "write.stats.enabled" with a
>> default value set to false.
>>
>> And as per point 7), I was planning to introduce an "ANALYZE table" or
>> "CALL procedure" SQL (maybe table-level API too) to asynchronously
>> compute the stats on demand from the previous checkpoints.
>>
>> But currently, `TableMetadata` doesn't have a clean Interface to provide
>> the statistics file for the current snapshot.
>> If stats are not present, we need another interface to provide a last
>> successful snapshot id for which stats was computed.
>> Also, there is some confusion around reusing the statistics file (because
>> the spec only has a computed snapshot id, not the referenced snapshot id).
>> I am planning to open up a PR to handle these interface updates
>> this week. (same things as you suggested in the last Iceberg sync).
>> This should serve as a good foundation to get insights for lazy &
>> incremental stats computing.
>>
>>
>> Thanks,
>> Ajantha
>>
>>
>>
>> On Thu, Nov 24, 2022 at 12:50 AM Ryan Blue <b...@tabular.io> wrote:
>>
>> Thanks for writing this up, Ajantha! I think that we have all the
>> upstream pieces in place to work on this so it's great to have a proposal.
>>
>>
>>
>> The proposal does a good job of summarizing the choices for how to store
>> the data, but I think there is also a decision that we need to make before
>> that: Should Iceberg require writers to maintain the partition stats?
>>
>>
>>
>> If we do want writers to participate, then we may want to make choices
>> that are easier for writers. But I think that is going to be a challenge.
>> Adding requirements for writers would mean that we need to bump the spec
>> version. Otherwise, we aren't guaranteed that writers will update the files
>> correctly. I think I would prefer to take a lazy approach and not assume
>> that writers will keep the partition stats up to date, in which case we
>> need a way to know which parts of a table are newer than the most recent
>> stats.
>>
>>
>>
>> Ryan
>>
>>
>>
>> On Wed, Nov 23, 2022 at 4:36 AM Ajantha Bhat <ajanthab...@gmail.com>
>> wrote:
>>
>> Thanks Piotr for taking a look at it.
>> I have replied to all the comments in the document.
>> I might need your support in standardising the existing `StatisticsFile`
>> interface to adopt partition stats as mentioned in the design.
>>
>>
>>
>> *We do need more eyes on the design. Once I get approval for the design,
>> I can start the implementation.  *
>> Thanks,
>> Ajantha
>>
>>
>>
>> On Wed, Nov 23, 2022 at 3:28 PM Piotr Findeisen <pi...@starburstdata.com>
>> wrote:
>>
>> Hi Ajantha,
>>
>>
>>
>> this is very interesting document, thank you for your work on this!
>>
>> I've added a few comments there.
>>
>>
>>
>> I have one high-level design comment so I thought it would be nicer to
>> everyone if I re-post it here
>>
>>
>>
>> is "partition" the right level of keeping the stats?
>> We do this in Hive, but was it an accidental choice? or just the only
>> thing that was possible to be implemented many years ago?
>>
>>
>>
>> Iceberg allows to have higher number of partitions compared to Hive,
>> because it scales better. But that means partition-level may or may not be
>> the right granularity.
>>
>>
>> A self-optimizing system would gather stats on "per query unit" basis --
>> for example if i partition by [ day x country ], but usually query by day,
>> the days are the "query unit" and from stats perspective country can be
>> ignored.
>> Having more fine-grained partitions may lead to expensive planning time,
>> so it's not theoretical problem.
>>
>>
>> I am not saying we should implement all this logic right now, but I think
>> we should decouple partitioning scheme from stats partitions, to allow
>>  query engine to become smarter.
>>
>>
>>
>>
>>
>> cc @Alexander Jo <alex...@starburstdata.com>
>>
>>
>>
>> Best
>>
>> PF
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> On Mon, Nov 14, 2022 at 12:47 PM Ajantha Bhat <ajanthab...@gmail.com>
>> wrote:
>>
>> Hi Community,
>> I did a proposal write-up for the partition stats in Iceberg.
>> Please have a look and let me know what you think. I would like to work
>> on it.
>>
>>
>> https://docs.google.com/document/d/1vaufuD47kMijz97LxM67X8OX-W2Wq7nmlz3jRo8J5Qk/edit?usp=sharing
>>
>> Requirement background snippet from the above document.
>>
>> For some query engines that use cost-based-optimizer instead or along
>> with rule-based-optimizer (like Dremio, Trino, etc), at the planning time,
>> it is good to know the partition level stats like total rows per
>> partition and total files per partition to take decisions for CBO (
>> like deciding on the join reordering and join type, identifying the
>> parallelism).
>> Currently, the only way to do this is to read the partition info from 
>> data_file
>> in manifest_entry of the manifest file and compute partition-level
>> statistics (the same thing that ‘partitions’ metadata table is doing *[see
>> **Appendix A*
>> <https://docs.google.com/document/d/1vaufuD47kMijz97LxM67X8OX-W2Wq7nmlz3jRo8J5Qk/edit#heading=h.s8iywtu7x8m6>
>> *]*).
>> Doing this on each query is expensive. *Hence, this is a proposal for
>> computing and storing partition-level stats for Iceberg tables and using
>> them during queries.*
>>
>>
>>
>> Thanks,
>> Ajantha
>>
>>
>>
>>
>> --
>>
>> Ryan Blue
>>
>> Tabular
>>
>>
>>
>>
>> --
>>
>> Ryan Blue
>>
>> Tabular
>>
>>

-- 
Ryan Blue
Tabular

Reply via email to