[...]

> +static int
> +dlb_pf_ldb_port_create(struct dlb_hw_dev *handle,
> +                    struct dlb_create_ldb_port_args *cfg,
> +                    enum dlb_cq_poll_modes poll_mode)
> +{
> +     struct dlb_dev *dlb_dev = (struct dlb_dev *)handle->pf_dev;
> +     struct dlb_cmd_response response = {0};
> +     int ret;
> +     uint8_t *port_base;
> +     const struct rte_memzone *mz;
> +     int alloc_sz, qe_sz, cq_alloc_depth;
> +     rte_iova_t pp_dma_base;
> +     rte_iova_t pc_dma_base;
> +     rte_iova_t cq_dma_base;
> +     int is_dir = false;
> +
> +     DLB_INFO(dev->dlb_device, "Entering %s()\n", __func__);
> +
> +     if (poll_mode == DLB_CQ_POLL_MODE_STD)
> +             qe_sz = sizeof(struct dlb_dequeue_qe);
> +     else
> +             qe_sz = RTE_CACHE_LINE_SIZE;
> +
> +     /* The hardware always uses a CQ depth of at least
> +      * DLB_MIN_HARDWARE_CQ_DEPTH, even though from the user
> +      * perspective we support a depth as low as 1 for LDB ports.
> +      */
> +     cq_alloc_depth = RTE_MAX(cfg->cq_depth,
> DLB_MIN_HARDWARE_CQ_DEPTH);
> +
> +     /* Calculate the port memory required, including two cache lines for
> +      * credit pop counts. Round up to the nearest cache line.
> +      */
> +     alloc_sz = 2 * RTE_CACHE_LINE_SIZE + cq_alloc_depth * qe_sz;
> +     alloc_sz = RTE_CACHE_LINE_ROUNDUP(alloc_sz);
> +
> +     port_base = dlb_alloc_coherent_aligned(&mz, &pc_dma_base,
> +                                            alloc_sz, PAGE_SIZE);
> +     if (port_base == NULL)
> +             return -ENOMEM;
> +
> +     /* Lock the page in memory */
> +     ret = rte_mem_lock_page(port_base);
> +     if (ret < 0) {
> +             DLB_LOG_ERR("dlb pf pmd could not lock page for device i/o\n");
> +             goto create_port_err;
> +     }
> +
> +     memset(port_base, 0, alloc_sz);
> +     cq_dma_base = (uintptr_t)(pc_dma_base + (2 *
> RTE_CACHE_LINE_SIZE));
> +
> +     ret = dlb_hw_create_ldb_port(&dlb_dev->hw,
> +                                  handle->domain_id,
> +                                  cfg,
> +                                  pc_dma_base,
> +                                  cq_dma_base,
> +                                  &response);
> +     if (ret)
> +             goto create_port_err;
> +
> +     pp_dma_base = (uintptr_t)dlb_dev->hw.func_kva + PP_BASE(is_dir);
> +     dlb_port[response.id][DLB_LDB].pp_addr =
> +             (void *)(uintptr_t)(pp_dma_base + (PAGE_SIZE * response.id));
> +
> +     dlb_port[response.id][DLB_LDB].cq_base =
> +             (void *)(uintptr_t)(port_base + (2 * RTE_CACHE_LINE_SIZE));
> +
> +     dlb_port[response.id][DLB_LDB].ldb_popcount =
> +             (void *)(uintptr_t)port_base;
> +     dlb_port[response.id][DLB_LDB].dir_popcount = (void *)(uintptr_t)
> +             (port_base + RTE_CACHE_LINE_SIZE);
> +     dlb_port[response.id][DLB_LDB].mz = mz;
> +
> +     *(struct dlb_cmd_response *)cfg->response = response;
> +
> +     DLB_INFO(dev->dlb_device, "Exiting %s() with ret=%d\n", __func__,
> ret);
> +
> +create_port_err:

Need to free the memzone if the PMD jumps to this label

> +
> +     return ret;
> +}
> +
> +static int
> +dlb_pf_dir_port_create(struct dlb_hw_dev *handle,
> +                    struct dlb_create_dir_port_args *cfg,
> +                    enum dlb_cq_poll_modes poll_mode)
> +{
> +     struct dlb_dev *dlb_dev = (struct dlb_dev *)handle->pf_dev;
> +     struct dlb_cmd_response response = {0};
> +     int ret;
> +     uint8_t *port_base;
> +     const struct rte_memzone *mz;
> +     int alloc_sz, qe_sz;
> +     rte_iova_t pp_dma_base;
> +     rte_iova_t pc_dma_base;
> +     rte_iova_t cq_dma_base;
> +     int is_dir = true;
> +
> +     DLB_INFO(dev->dlb_device, "Entering %s()\n", __func__);
> +
> +     if (poll_mode == DLB_CQ_POLL_MODE_STD)
> +             qe_sz = sizeof(struct dlb_dequeue_qe);
> +     else
> +             qe_sz = RTE_CACHE_LINE_SIZE;
> +
> +     /* Calculate the port memory required, including two cache lines for
> +      * credit pop counts. Round up to the nearest cache line.
> +      */
> +     alloc_sz = 2 * RTE_CACHE_LINE_SIZE + cfg->cq_depth * qe_sz;
> +     alloc_sz = RTE_CACHE_LINE_ROUNDUP(alloc_sz);
> +
> +     port_base = dlb_alloc_coherent_aligned(&mz, &pc_dma_base,
> +                                            alloc_sz, PAGE_SIZE);
> +     if (port_base == NULL)
> +             return -ENOMEM;
> +
> +     /* Lock the page in memory */
> +     ret = rte_mem_lock_page(port_base);
> +     if (ret < 0) {
> +             DLB_LOG_ERR("dlb pf pmd could not lock page for device i/o\n");
> +             goto create_port_err;
> +     }
> +
> +     memset(port_base, 0, alloc_sz);
> +     cq_dma_base = (uintptr_t)(pc_dma_base + (2 *
> RTE_CACHE_LINE_SIZE));
> +
> +     ret = dlb_hw_create_dir_port(&dlb_dev->hw,
> +                                  handle->domain_id,
> +                                  cfg,
> +                                  pc_dma_base,
> +                                  cq_dma_base,
> +                                  &response);
> +     if (ret)
> +             goto create_port_err;
> +
> +     pp_dma_base = (uintptr_t)dlb_dev->hw.func_kva + PP_BASE(is_dir);
> +     dlb_port[response.id][DLB_DIR].pp_addr =
> +             (void *)(uintptr_t)(pp_dma_base + (PAGE_SIZE * response.id));
> +
> +     dlb_port[response.id][DLB_DIR].cq_base =
> +             (void *)(uintptr_t)(port_base + (2 * RTE_CACHE_LINE_SIZE));
> +
> +     dlb_port[response.id][DLB_DIR].ldb_popcount =
> +             (void *)(uintptr_t)port_base;
> +     dlb_port[response.id][DLB_DIR].dir_popcount = (void *)(uintptr_t)
> +             (port_base + RTE_CACHE_LINE_SIZE);
> +     dlb_port[response.id][DLB_DIR].mz = mz;
> +
> +     *(struct dlb_cmd_response *)cfg->response = response;
> +
> +     DLB_INFO(dev->dlb_device, "Exiting %s() with ret=%d\n", __func__,
> ret);
> +
> +create_port_err:
> +

Ditto

Thanks,
Gage

Reply via email to