Hi list, I have a misbehaving query which uses all available disk space and then terminates with a "cannot write block" error. To prevent other processes from running into trouble I've set the following:
temp_file_limit = 100GB The query does parallelize and uses one parallel worker while executing, but it does not abort when the temp file limit is reached: 345G pgsql_tmp It does abort way later, after using around 300+ GB: [53400] ERROR: temporary file size exceeds temp_file_limit (104857600kB) Where: parallel worker The comment in the file states that this is a per-session parameter, so what is going wrong here? I am using Postgres 14 on Ubuntu. Regards, Frits