> -----Original Message-----
> From: Maxime Coquelin <maxime.coque...@redhat.com>
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> <chenbo....@intel.com>; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming <cunming.li...@intel.com>; Xie, Yongji
> <xieyon...@bytedance.com>; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin <maxime.coque...@redhat.com>
> Subject: [RFC 04/27] vhost: add helper of IOTLB entries coredump
> 
> This patch reworks IOTLB code to extract madvise-related
> bits into dedicated helper. This refactoring improves code
> sharing.
> 
> Signed-off-by: Maxime Coquelin <maxime.coque...@redhat.com>
> ---
>  lib/vhost/iotlb.c | 77 +++++++++++++++++++++++++----------------------
>  1 file changed, 41 insertions(+), 36 deletions(-)
> 
> diff --git a/lib/vhost/iotlb.c b/lib/vhost/iotlb.c
> index 870c8acb88..e8f1cb661e 100644
> --- a/lib/vhost/iotlb.c
> +++ b/lib/vhost/iotlb.c
> @@ -23,6 +23,34 @@ struct vhost_iotlb_entry {
> 
>  #define IOTLB_CACHE_SIZE 2048
> 
> +static void
> +vhost_user_iotlb_set_dump(struct virtio_net *dev, struct
> vhost_iotlb_entry *node)
> +{
> +     uint64_t align;
> +
> +     align = hua_to_alignment(dev->mem, (void *)(uintptr_t)node->uaddr);
> +
> +     mem_set_dump((void *)(uintptr_t)node->uaddr, node->size, false,
> align);
> +}
> +
> +static void
> +vhost_user_iotlb_clear_dump(struct virtio_net *dev, struct
> vhost_iotlb_entry *node,
> +             struct vhost_iotlb_entry *prev, struct vhost_iotlb_entry *next)
> +{
> +     uint64_t align, mask;
> +
> +     align = hua_to_alignment(dev->mem, (void *)(uintptr_t)node->uaddr);
> +     mask = ~(align - 1);
> +
> +     /* Don't disable coredump if the previous node is in the same page
> */
> +     if (prev == NULL || (node->uaddr & mask) != ((prev->uaddr + prev-
> >size - 1) & mask)) {
> +             /* Don't disable coredump if the next node is in the same page
> */
> +             if (next == NULL ||
> +                             ((node->uaddr + node->size - 1) & mask) != 
> (next-
> >uaddr & mask))
> +                     mem_set_dump((void *)(uintptr_t)node->uaddr, node->size,
> false, align);
> +     }
> +}
> +
>  static struct vhost_iotlb_entry *
>  vhost_user_iotlb_pool_get(struct vhost_virtqueue *vq)
>  {
> @@ -149,8 +177,8 @@ vhost_user_iotlb_cache_remove_all(struct virtio_net
> *dev, struct vhost_virtqueue
>       rte_rwlock_write_lock(&vq->iotlb_lock);
> 
>       RTE_TAILQ_FOREACH_SAFE(node, &vq->iotlb_list, next, temp_node) {
> -             mem_set_dump((void *)(uintptr_t)node->uaddr, node->size, false,
> -                     hua_to_alignment(dev->mem, (void *)(uintptr_t)node-
> >uaddr));
> +             vhost_user_iotlb_set_dump(dev, node);
> +
>               TAILQ_REMOVE(&vq->iotlb_list, node, next);
>               vhost_user_iotlb_pool_put(vq, node);
>       }
> @@ -164,7 +192,6 @@ static void
>  vhost_user_iotlb_cache_random_evict(struct virtio_net *dev, struct
> vhost_virtqueue *vq)
>  {
>       struct vhost_iotlb_entry *node, *temp_node, *prev_node = NULL;
> -     uint64_t alignment, mask;
>       int entry_idx;
> 
>       rte_rwlock_write_lock(&vq->iotlb_lock);
> @@ -173,20 +200,10 @@ vhost_user_iotlb_cache_random_evict(struct
> virtio_net *dev, struct vhost_virtque
> 
>       RTE_TAILQ_FOREACH_SAFE(node, &vq->iotlb_list, next, temp_node) {
>               if (!entry_idx) {
> -                     struct vhost_iotlb_entry *next_node;
> -                     alignment = hua_to_alignment(dev->mem, (void
> *)(uintptr_t)node->uaddr);
> -                     mask = ~(alignment - 1);
> -
> -                     /* Don't disable coredump if the previous node is in the
> same page */
> -                     if (prev_node == NULL || (node->uaddr & mask) !=
> -                                     ((prev_node->uaddr + prev_node->size - 
> 1) &
> mask)) {
> -                             next_node = RTE_TAILQ_NEXT(node, next);
> -                             /* Don't disable coredump if the next node is in
> the same page */
> -                             if (next_node == NULL || ((node->uaddr + node-
> >size - 1) & mask) !=
> -                                             (next_node->uaddr & mask))
> -                                     mem_set_dump((void 
> *)(uintptr_t)node->uaddr,
> node->size,
> -                                                     false, alignment);
> -                     }
> +                     struct vhost_iotlb_entry *next_node =
> RTE_TAILQ_NEXT(node, next);
> +
> +                     vhost_user_iotlb_clear_dump(dev, node, prev_node,
> next_node);
> +
>                       TAILQ_REMOVE(&vq->iotlb_list, node, next);
>                       vhost_user_iotlb_pool_put(vq, node);
>                       vq->iotlb_cache_nr--;
> @@ -240,16 +257,16 @@ vhost_user_iotlb_cache_insert(struct virtio_net *dev,
> struct vhost_virtqueue *vq
>                       vhost_user_iotlb_pool_put(vq, new_node);
>                       goto unlock;
>               } else if (node->iova > new_node->iova) {
> -                     mem_set_dump((void *)(uintptr_t)new_node->uaddr,
> new_node->size, true,
> -                             hua_to_alignment(dev->mem, (void
> *)(uintptr_t)new_node->uaddr));
> +                     vhost_user_iotlb_set_dump(dev, new_node);
> +
>                       TAILQ_INSERT_BEFORE(node, new_node, next);
>                       vq->iotlb_cache_nr++;
>                       goto unlock;
>               }
>       }
> 
> -     mem_set_dump((void *)(uintptr_t)new_node->uaddr, new_node->size,
> true,
> -             hua_to_alignment(dev->mem, (void *)(uintptr_t)new_node-
> >uaddr));
> +     vhost_user_iotlb_set_dump(dev, new_node);
> +
>       TAILQ_INSERT_TAIL(&vq->iotlb_list, new_node, next);
>       vq->iotlb_cache_nr++;
> 
> @@ -265,7 +282,6 @@ vhost_user_iotlb_cache_remove(struct virtio_net *dev,
> struct vhost_virtqueue *vq
>                                       uint64_t iova, uint64_t size)
>  {
>       struct vhost_iotlb_entry *node, *temp_node, *prev_node = NULL;
> -     uint64_t alignment, mask;
> 
>       if (unlikely(!size))
>               return;
> @@ -278,20 +294,9 @@ vhost_user_iotlb_cache_remove(struct virtio_net *dev,
> struct vhost_virtqueue *vq
>                       break;
> 
>               if (iova < node->iova + node->size) {
> -                     struct vhost_iotlb_entry *next_node;
> -                     alignment = hua_to_alignment(dev->mem, (void
> *)(uintptr_t)node->uaddr);
> -                     mask = ~(alignment-1);
> -
> -                     /* Don't disable coredump if the previous node is in the
> same page */
> -                     if (prev_node == NULL || (node->uaddr & mask) !=
> -                                     ((prev_node->uaddr + prev_node->size - 
> 1) &
> mask)) {
> -                             next_node = RTE_TAILQ_NEXT(node, next);
> -                             /* Don't disable coredump if the next node is in
> the same page */
> -                             if (next_node == NULL || ((node->uaddr + node-
> >size - 1) & mask) !=
> -                                             (next_node->uaddr & mask))
> -                                     mem_set_dump((void 
> *)(uintptr_t)node->uaddr,
> node->size,
> -                                                     false, alignment);
> -                     }
> +                     struct vhost_iotlb_entry *next_node =
> RTE_TAILQ_NEXT(node, next);
> +
> +                     vhost_user_iotlb_clear_dump(dev, node, prev_node,
> next_node);
> 
>                       TAILQ_REMOVE(&vq->iotlb_list, node, next);
>                       vhost_user_iotlb_pool_put(vq, node);
> --
> 2.39.2

Reviewed-by: Chenbo Xia <chenbo....@intel.com>

Reply via email to