Hi,

Few details will help

   1. Spark version
   2. Spark SQL, Scala or PySpark
   3. size of tables in join.
   4. What does explain() or the joining operation show?


HTH


   view my Linkedin profile
<https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>


 https://en.everybodywiki.com/Mich_Talebzadeh



*Disclaimer:* Use it at your own risk. Any and all responsibility for any
loss, damage or destruction of data or any other property which may arise
from relying on this email's technical content is explicitly disclaimed.
The author will in no case be liable for any monetary damages arising from
such loss, damage or destruction.




On Wed, 22 Feb 2023 at 15:42, Prem Sahoo <prem.re...@gmail.com> wrote:

> Hello Team,
> We are observing Spark Union performance issues when unioning big tables
> with lots of rows. Do we have any option apart from the Union ?
>

Reply via email to