On Tue, Sep 20, 2022 at 03:12:00PM -0700, Peter Geoghegan wrote: > On Mon, Sep 12, 2022 at 2:01 PM Peter Geoghegan <p...@bowt.ie> wrote: >> I'd like to talk about one such technique on this thread. The attached >> WIP patch reduces the size of xl_heap_freeze_page records by applying >> a simple deduplication process. > > Attached is v2, which I'm just posting to keep CFTester happy. No real > changes here.
This idea seems promising. I see that you called this patch a work-in-progress, so I'm curious what else you are planning to do with it. As I'm reading this thread and the patch, I'm finding myself wondering if it's worth exploring using wal_compression for these records instead. I think you've essentially created an efficient compression mechanism for this one type of record, but I'm assuming that lz4/zstd would also yield some rather substantial improvements for this kind of data. Presumably a generic WAL record compression mechanism could be reused for other large records, too. That could be much easier than devising a deduplication strategy for every record type. -- Nathan Bossart Amazon Web Services: https://aws.amazon.com