----- On May 28, 2024, at 1:55 AM, Dmitry Kozlyuk dmitry.kozl...@gmail.com 
wrote:

> Hi Lewis,
> 
> Memory reserved by eal_get_virtual_area() is not yet useful,
> but it is very large, so by excluding it from dumps,
> DPDK prevents dumps from including large zero-filled parts.
> 
> It also makes sense to call eal_mem_set_dump(..., false)
> from eal_memalloc.c:free_seg(), because of --huge-unlink=never:
> in this mode (Linux-only), freed segments are not cleared,
> so if they were included into dump, it would be a lot of garbage data.
> 
> Newly allocated hugepages are not included into dumps
> because this would make dumps very large by default.
> However, this could be an opt-in as a runtime option if need be.

Thanks for the clarification.  I agree that not including freed segments makes 
perfect sense.

But when debugging a core dump, it's sometimes really helpful to be able to see 
what's in the mbuf that was being processed at the time.  Perhaps it would be a 
useful option to be able to tell the allocator not to disable core dumps.

In the mean time, my experiments to get around this have not been fruitful.

I wondered if we could enable core dumps for mbufs by calling 
rte_mempool_mem_iter() on the pool returned by rte_pktmbuf_pool_create(), and 
have the callback function call madvise(memhdr->addr, memhdr->len, MADV_CORE).  
But that didn't help, or at least the size of the core file didn't increase.

I then tried disabling the call to madvise() in the DPDK source code, and that 
didn't make any difference either.

Note that this is on FreeBSD, so I wonder if there's some fundamental reason 
that the contigmem memory doesn't get included in a core dump?

Reply via email to