Do you get some error on passing the master option to your spark connect command?
On Tue, 6 Aug, 2024, 15:36 Ilango, <elango...@gmail.com> wrote: > > > > Thanks Prabodh. I'm having an issue with the Spark Connect connection as > the `spark.master` value is set to `local[*]` in Spark Connect UI, whereas > the actual master node for our Spark standalone cluster is different. I am > passing that master node ip in the Spark Connect Connection. But still it > is not set correctly. Could you please help me update this configuration to > reflect the correct master node value? > > > > This is my spark connect connection > > > > spark = SparkSession.builder\ > > .remote("sc://<spark-stand-alone-master-node-ip>:15002")\ > > .getOrCreate() > > > Thanks, > Elango > > > On Tue, 6 Aug 2024 at 5:45 PM, Prabodh Agarwal <prabodh1...@gmail.com> > wrote: > >> There is an executors tab on spark connect. It's contents are generally >> similar to the workers section of the spark master ui. >> >> You might need to specify --master option in your spark connect command >> if you haven't done so yet. >> >> On Tue, 6 Aug, 2024, 14:19 Ilango, <elango...@gmail.com> wrote: >> >>> >>> Hi all, >>> >>> I am evaluating the use of Spark Connect with my Spark stand-alone >>> cluster, which has a master node and 3 worker nodes. I have successfully >>> created a Spark Connect connection. However, when submitting Spark SQL >>> queries, the jobs are being executed only on the master node, and I do not >>> observe any executors running on the worker nodes, despite requesting 4 >>> executors. >>> >>> >>> >>> I would appreciate clarification on whether Spark stand-alone cluster is >>> supported for use with Spark Connect. >>> >>> If so, how can I leverage the existing Spark stand-alone cluster's >>> worker nodes? >>> >>> >>> >>> >>> >>> >>> Thanks, >>> Elango >>> >>