Hi, Applications often face an "out of memory" condition as they try to fetch "N rows" from the database. If N is small, then the execution becomes inefficient due to many roundtrips. If N is high, there's a risk that many rows would overflow the client's memory.
Note: the client can't stop reading the data at arbitrary times as they might want to execute another query while processing the resultset, so "fetch all rows at once and read them from the socket as needed" is not really an option. The problem is the clients can't tell how much memory it would take to buffer the results of "fetch 100 rows". The row size might vary. Currently, the only case the database returns PortalSuspended seems to be when max_rows is reached when executing a select. I wonder if the database can suspend portals earlier in case it knows it produced a lot of data. In other words: 1) If the user calls execute(fetch all rows), then follow the current behavior. 2) If the user provides non-zero max_rows, then optionally suspend the portal as the result exceeds a pre-configured amount (GUC?). Then clients could configure the connection like "ok, please suspend portals if a single fetch exceeds 50MiB". To my understanding, * it should be a backward-compatible change * it would require no protocol changes, * it would make applications more robust when it comes to "out of memory while fetching many rows" * it might improve the performance overall as the applications could increase their pessimistic "max_rows" fetch size Is it ok to contribute a patch like that? Vladimir