yihua commented on issue #6015:
URL: https://github.com/apache/hudi/issues/6015#issuecomment-1171646296

   @rohit-m-99 there might be compatibility issues between hudi-spark3.2-bundle 
and hudi-utilities-bundle_2.12 in Spark 3.2.  However, that may not be the root 
cause here.  A few things to check:
   (1) the average file sizes and number of files in the target table
   (2) the memory usage and disk spill of the Spark tasks
   (3) any warnings in the logs of the executors that fail the stage
   
   Here is the tuning guide for Hudi Spark jobs that is handy: 
https://hudi.apache.org/docs/tuning-guide


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to