ivoson commented on PR #52336:
URL: https://github.com/apache/spark/pull/52336#issuecomment-3336646076

   > > Currently there are two related configs for the feature:
   > > spark.sql.shuffle.orderIndependentChecksum.enabled: whether compute 
order independent checksum for shuffle output;
   > > spark.scheduler.checksumMismatchFullRetry.enabled: whether retry all 
tasks for a succeeding stages when shuffle checksum mismatch detected;
   > 
   > Thank you for correcting me. In that case, 
`spark.sql.shuffle.orderIndependentChecksum.*` seems to be the parent name 
space for this feature. If 
`spark.sql.shuffle.orderIndependentChecksum.enabled=false` disabled this PR's 
configuration, this should be under the same namespace. The revised config name 
might be the following. WDYT, @ivoson ?
   > 
   > ```
   > - spark.scheduler.checksumMismatchFullRetry.enabled
   > + spark.sql.shuffle.orderIndependentChecksum.enableFullRetryOnMismatch
   > ```
   > 
   > The basic idea is the dependency among the configurations. Please let me 
know your hierarchy for new set of configurations for this feature.
   
   Thanks @dongjoon-hyun for the suggestion. Updated. For the new configs:
   
   `spark.sql.shuffle.orderIndependentChecksum.enabled` -> when it's true, 
we'll compute the shuffle checksum and only log detected checksum mismatch if 
`spark.sql.shuffle.orderIndependentChecksum.enableFullRetryOnMismatch` is false;
   `spark.sql.shuffle.orderIndependentChecksum.enableFullRetryOnMismatch` -> 
when it's true, we'll compute the shuffle checksum and fully retry consumer 
stages once mismatch happens. 
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to