I believe it’s a spark Ui issue which do not display correct value. I
believe it is resolved for spark 3.0.

Thanks
Amit

On Fri, Jan 8, 2021 at 4:00 PM Luca Canali <luca.can...@cern.ch> wrote:

> You report 'Storage Memory': 3.3TB/ 598.5 GB -> The first number is the
> memory used for storage, the second one is the available memory (for
> storage) in the unified memory pool.
>
> The used memory shown in your webui snippet is indeed quite high (higher
> than the available memory!? ), you can probably profit by drilling down on
> that to understand better what is happening.
>
> For example look at the details per executor (the numbers you reported are
> aggregated values), then also look at the “storage tab” for a list of
> cached RDDs with details.
>
> In case, Spark 3.0 has improved memory instrumentation and improved
> instrumentation for streaming, so you can you profit from testing there too.
>
>
>
>
>
> *From:* Eric Beabes <mailinglist...@gmail.com>
> *Sent:* Friday, January 8, 2021 04:23
> *To:* Luca Canali <luca.can...@cern.ch>
> *Cc:* spark-user <user@spark.apache.org>
> *Subject:* Re: Understanding Executors UI
>
>
>
> So when I see this for 'Storage Memory': *3.3TB/ 598.5 GB* *- it's
> telling me that Spark is using 3.3 TB of memory & 598.5 GB is used for
> caching data, correct?* What I am surprised about is that these numbers
> don't change at all throughout the day even though the load on the system
> is low after 5pm PST.
>
>
>
> I would expect the "Memory used" to be lower than 3.3Tb after 5pm PST.
>
>
>
> Does Spark 3.0 do a better job of memory management? Wondering if
> upgrading to Spark 3.0 would improve performance?
>
>
>
>
>
> On Wed, Jan 6, 2021 at 2:29 PM Luca Canali <luca.can...@cern.ch> wrote:
>
> Hi Eric,
>
>
>
> A few links, in case they can be useful for your troubleshooting:
>
>
>
> The Spark Web UI is documented in Spark 3.x documentation, although you
> can use most of it for Spark 2.4 too:
> https://spark.apache.org/docs/latest/web-ui.html
>
>
>
> Spark memory management is documented at
> https://spark.apache.org/docs/latest/tuning.html#memory-management-overview
>
>
> Additional resource: see also this diagram
> https://canali.web.cern.ch/docs/SparkExecutorMemory.png  and
> https://db-blog.web.cern.ch/blog/luca-canali/2020-08-spark3-memory-monitoring
>
>
>
> Best,
>
> Luca
>
>
>
> *From:* Eric Beabes <mailinglist...@gmail.com>
> *Sent:* Wednesday, January 6, 2021 00:20
> *To:* spark-user <user@spark.apache.org>
> *Subject:* Understanding Executors UI
>
>
>
> [image: image.png]
>
>
>
>
>
> Not sure if this image will go through. (Never sent an email to this
> mailing list with an image).
>
>
>
> I am trying to understand this 'Executors' UI in Spark 2.4. I have a
> Stateful Structured Streaming job with 'State timeout' set to 10 minutes.
> When the load on the system is low a message gets written to Kafka
> immediately after the State times out BUT under heavy load it takes over 40
> minutes to get a message on the output topic. Trying to debug this issue &
> see if performance can be improved.
>
>
>
> Questions:
>
>
>
> 1) I am requesting 3.2 TB of memory but it seems the job keeps using only
> 598.5 GB as per the values in 'Storage Memory' as well as 'On Heap Storage
> Memory'. Wondering if this is a Cluster issue OR am I not setting values
> correctly?
>
> 2) Where can I find documentation to understand different 'Tabs' in the
> Spark UI? (Sorry, Googling didn't help. I will keep searching.)
>
>
>
> Any pointers would be appreciated. Thanks.
>
>
>
>

Reply via email to