beliefer commented on PR #10335: URL: https://github.com/apache/incubator-gluten/pull/10335#issuecomment-3157399953
``` [2025-08-05T10:15:43.063Z] 10:15:42.955 ERROR org.apache.spark.sql.execution.exchange.BroadcastExchangeExec: Could not execute broadcast in 300 secs. [2025-08-05T10:15:43.066Z] java.util.concurrent.TimeoutException [2025-08-05T10:15:43.066Z] at java.base/java.util.concurrent.FutureTask.get(FutureTask.java:204) [2025-08-05T10:15:43.066Z] at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast(BroadcastExchangeExec.scala:212) [2025-08-05T10:15:43.066Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeBroadcast$1(SparkPlan.scala:208) [2025-08-05T10:15:43.066Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:246) [2025-08-05T10:15:43.066Z] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) [2025-08-05T10:15:43.066Z] at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243) [2025-08-05T10:15:43.066Z] at org.apache.spark.sql.execution.SparkPlan.executeBroadcast(SparkPlan.scala:204) [2025-08-05T10:15:43.066Z] 'unsafe broadcast left semi join updates peak execution memory' offload to gluten [2025-08-05T10:15:43.319Z] - unsafe broadcast left semi join updates peak execution memory *** FAILED *** [2025-08-05T10:15:43.323Z] org.apache.spark.SparkException: Could not execute broadcast in 300 secs. You can increase the timeout for broadcasts via "spark.sql.broadcastTimeout" or disable broadcast join by setting "spark.sql.autoBroadcastJoinThreshold" to -1. [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.errors.QueryExecutionErrors$.executeBroadcastTimeoutError(QueryExecutionErrors.scala:1099) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast(BroadcastExchangeExec.scala:220) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeBroadcast$1(SparkPlan.scala:208) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:246) [2025-08-05T10:15:43.323Z] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.executeBroadcast(SparkPlan.scala:204) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doExecute(BroadcastHashJoinExec.scala:126) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:246) [2025-08-05T10:15:43.323Z] ... [2025-08-05T10:15:43.323Z] Cause: java.util.concurrent.TimeoutException: [2025-08-05T10:15:43.323Z] at java.base/java.util.concurrent.FutureTask.get(FutureTask.java:204) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast(BroadcastExchangeExec.scala:212) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeBroadcast$1(SparkPlan.scala:208) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:246) [2025-08-05T10:15:43.323Z] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.executeBroadcast(SparkPlan.scala:204) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doExecute(BroadcastHashJoinExec.scala:126) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195) [2025-08-05T10:15:43.323Z] at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:246) ``` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
