... using spark 3.2.1
________________________________
From: Shay Elbaz <shay.el...@gm.com>
Sent: Tuesday, July 19, 2022 1:26 PM
To: user@spark.apache.org <user@spark.apache.org>
Cc: Jeffrey O'Donoghue <jeffrey.odonog...@gm.com>
Subject: [EXTERNAL] spark.executor.pyspark.memory not added to the executor 
resource request on Kubernetes


ATTENTION: This email originated from outside of GM.


Hi,

We are trying tune executor memory on Kubernetes. Specifically, 8g for the jvm, 
8g for the python process, and additional 500m overhead:
--conf spark.executor.memory=8g
--conf spark.executor.pyspark.memory=8g
--conf spark.executor.memoryOverhead=500m

 According the docs, the executor pods should have 8+8+0.5 requested memory 
(spark.executor.pyspark.memory: "... When PySpark is run in YARN or Kubernetes, 
this memory is added to executor resource requests").
On Spark UI we can see the right configuration:

Executor Reqs:
      cores: [amount: 1]
      offHeap: [amount: 0]
      memoryOverhead: [amount: 500]
      pyspark.memory: [amount: 8192]
      memory: [amount: 8192]
Task Reqs:
      cpus: [amount: 1.0]


However, the running pod spec is different:
    Limits:
      memory:  8692Mi
    Requests:
      cpu:     1
      memory:  8692Mi



Looks like pyspark.memory value was not added to the resource request.
What are we missing?

Thanks,
Shay

Reply via email to