bnetzi commented on PR #40771: URL: https://github.com/apache/spark/pull/40771#issuecomment-2619648271
I know this PR is closed for a while but I think the decision not to merge it was not correct. As mentioned in this issue: https://issues.apache.org/jira/browse/SPARK-35723 And in https://issues.apache.org/jira/browse/SPARK-37358 Although the default of using limits same as request make some sense, in some cases it is much wiser to have a larger limit. Let's say many executors / drivers are sharing the same instance - if the limit and the request are the same - you must allocate to each pod the **max** memory it would use, causing a huge over allocation. It is very typical to have a momentarily high memory usage, every user should have the power to decide if it makes sense for him to have a large limit. it has its risks, but in many use cases the cost saving might worth it. For now in our env we are actually using (for almost a year) webhooks to override this behavior, I think it should be allowed natively. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org