Please ignore, sent to wrong list.
> -----Original Message-----
> From: Dylan Yip <dylan....@xilinx.com>
> Sent: Tuesday, September 17, 2019 1:21 PM
> To: linux-media@vger.kernel.org; Satish Kumar Nagireddy
> <satis...@xilinx.com>
> Cc: Dylan Yip <dyl...@xilinx.com>
> Subject: [LINUX PATCH] dma-mapping: Control memset operation using gfp
> flags
>
> In case of 4k video buffer, the allocation from a reserved memory is taking a
> long time, ~500ms. This is root caused to the memset() operations on the
> allocated memory which is consuming more cpu cycles.
> Due to this delay, we see that initial frames are being dropped.
>
> To fix this, we have wrapped the default memset, done when allocating
> coherent memory, under the __GFP_ZERO flag. So, we only clear allocated
> memory if __GFP_ZERO flag is enabled. We believe this should be safe as the
> video decoder always writes before reading.
> This optimizes decoder initialization as we do not set the __GFP_ZERO flag
> when allocating memory for decoder. With this optimization, we don't see
> initial frame drops and decoder initialization time is ~100ms.
>
> This patch adds plumbing through dma_alloc functions to pass gfp flag set by
> user to __dma_alloc_from_coherent(). Here gfp flag is checked for
> __GFP_ZERO. If present, we memset the buffer to 0 otherwise we skip
> memset.
>
> Signed-off-by: Dylan Yip <dylan....@xilinx.com>
> ---
> arch/arm/mm/dma-mapping-nommu.c | 2 +-
> include/linux/dma-mapping.h | 11 +++++++----
> kernel/dma/coherent.c | 15 +++++++++------
> kernel/dma/mapping.c | 2 +-
> 4 files changed, 18 insertions(+), 12 deletions(-)
>
> diff --git a/arch/arm/mm/dma-mapping-nommu.c b/arch/arm/mm/dma-
> mapping-nommu.c index 52b8255..242b2c3 100644
> --- a/arch/arm/mm/dma-mapping-nommu.c
> +++ b/arch/arm/mm/dma-mapping-nommu.c
> @@ -35,7 +35,7 @@ static void *arm_nommu_dma_alloc(struct device *dev,
> size_t size,
> unsigned long attrs)
>
> {
> - void *ret = dma_alloc_from_global_coherent(size, dma_handle);
> + void *ret = dma_alloc_from_global_coherent(size, dma_handle,
> gfp);
>
> /*
> * dma_alloc_from_global_coherent() may fail because:
> diff --git a/include/linux/dma-mapping.h b/include/linux/dma-mapping.h
> index f7d1eea..b715c9f 100644
> --- a/include/linux/dma-mapping.h
> +++ b/include/linux/dma-mapping.h
> @@ -160,24 +160,27 @@ static inline int is_device_dma_capable(struct
> device *dev)
> * Don't use them in device drivers.
> */
> int dma_alloc_from_dev_coherent(struct device *dev, ssize_t size,
> - dma_addr_t *dma_handle, void **ret);
> + dma_addr_t *dma_handle, void **ret,
> + gfp_t flag);
> int dma_release_from_dev_coherent(struct device *dev, int order, void
> *vaddr);
>
> int dma_mmap_from_dev_coherent(struct device *dev, struct
> vm_area_struct *vma,
> void *cpu_addr, size_t size, int *ret);
>
> -void *dma_alloc_from_global_coherent(ssize_t size, dma_addr_t
> *dma_handle);
> +void *dma_alloc_from_global_coherent(ssize_t size, dma_addr_t
> *dma_handle,
> + gfp_t flag);
> int dma_release_from_global_coherent(int order, void *vaddr); int
> dma_mmap_from_global_coherent(struct vm_area_struct *vma, void
> *cpu_addr,
> size_t size, int *ret);
>
> #else
> -#define dma_alloc_from_dev_coherent(dev, size, handle, ret) (0)
> +#define dma_alloc_from_dev_coherent(dev, size, handle, ret, flag) (0)
> #define dma_release_from_dev_coherent(dev, order, vaddr) (0) #define
> dma_mmap_from_dev_coherent(dev, vma, vaddr, order, ret) (0)
>
> static inline void *dma_alloc_from_global_coherent(ssize_t size,
> - dma_addr_t *dma_handle)
> + dma_addr_t *dma_handle,
> + gfp_t flag)
> {
> return NULL;
> }
> diff --git a/kernel/dma/coherent.c b/kernel/dma/coherent.c index
> 29fd659..d85fab5 100644
> --- a/kernel/dma/coherent.c
> +++ b/kernel/dma/coherent.c
> @@ -136,7 +136,7 @@ void dma_release_declared_memory(struct device
> *dev) EXPORT_SYMBOL(dma_release_declared_memory);
>
> static void *__dma_alloc_from_coherent(struct dma_coherent_mem
> *mem,
> - ssize_t size, dma_addr_t *dma_handle)
> + ssize_t size, dma_addr_t *dma_handle, gfp_t gfp_flag)
> {
> int order = get_order(size);
> unsigned long flags;
> @@ -158,7 +158,8 @@ static void *__dma_alloc_from_coherent(struct
> dma_coherent_mem *mem,
> *dma_handle = mem->device_base + (pageno << PAGE_SHIFT);
> ret = mem->virt_base + (pageno << PAGE_SHIFT);
> spin_unlock_irqrestore(&mem->spinlock, flags);
> - memset(ret, 0, size);
> + if (gfp_flag & __GFP_ZERO)
> + memset(ret, 0, size);
> return ret;
> err:
> spin_unlock_irqrestore(&mem->spinlock, flags); @@ -172,6 +173,7
> @@ static void *__dma_alloc_from_coherent(struct dma_coherent_mem
> *mem,
> * @dma_handle: This will be filled with the correct dma handle
> * @ret: This pointer will be filled with the virtual address
> * to allocated area.
> + * @flag: gfp flag set by user
> *
> * This function should be only called from per-arch dma_alloc_coherent()
> * to support allocation from per-device coherent memory pools.
> @@ -180,24 +182,25 @@ static void *__dma_alloc_from_coherent(struct
> dma_coherent_mem *mem,
> * generic memory areas, or !0 if dma_alloc_coherent should return @ret.
> */
> int dma_alloc_from_dev_coherent(struct device *dev, ssize_t size,
> - dma_addr_t *dma_handle, void **ret)
> + dma_addr_t *dma_handle, void **ret, gfp_t flag)
> {
> struct dma_coherent_mem *mem =
> dev_get_coherent_memory(dev);
>
> if (!mem)
> return 0;
>
> - *ret = __dma_alloc_from_coherent(mem, size, dma_handle);
> + *ret = __dma_alloc_from_coherent(mem, size, dma_handle, flag);
> return 1;
> }
>
> -void *dma_alloc_from_global_coherent(ssize_t size, dma_addr_t
> *dma_handle)
> +void *dma_alloc_from_global_coherent(ssize_t size, dma_addr_t
> *dma_handle,
> + gfp_t flag)
> {
> if (!dma_coherent_default_memory)
> return NULL;
>
> return
> __dma_alloc_from_coherent(dma_coherent_default_memory, size,
> - dma_handle);
> + dma_handle, flag);
> }
>
> static int __dma_release_from_coherent(struct dma_coherent_mem
> *mem, diff --git a/kernel/dma/mapping.c b/kernel/dma/mapping.c index
> b0038ca..bfea1d2 100644
> --- a/kernel/dma/mapping.c
> +++ b/kernel/dma/mapping.c
> @@ -272,7 +272,7 @@ void *dma_alloc_attrs(struct device *dev, size_t size,
> dma_addr_t *dma_handle,
>
> WARN_ON_ONCE(!dev->coherent_dma_mask);
>
> - if (dma_alloc_from_dev_coherent(dev, size, dma_handle,
> &cpu_addr))
> + if (dma_alloc_from_dev_coherent(dev, size, dma_handle,
> &cpu_addr,
> +flag))
> return cpu_addr;
>
> /* let the implementation decide on the zone to allocate from: */
> --
> 2.7.4