[...] > +static int > +dlb_pf_ldb_port_create(struct dlb_hw_dev *handle, > + struct dlb_create_ldb_port_args *cfg, > + enum dlb_cq_poll_modes poll_mode) > +{ > + struct dlb_dev *dlb_dev = (struct dlb_dev *)handle->pf_dev; > + struct dlb_cmd_response response = {0}; > + int ret; > + uint8_t *port_base; > + const struct rte_memzone *mz; > + int alloc_sz, qe_sz, cq_alloc_depth; > + rte_iova_t pp_dma_base; > + rte_iova_t pc_dma_base; > + rte_iova_t cq_dma_base; > + int is_dir = false; > + > + DLB_INFO(dev->dlb_device, "Entering %s()\n", __func__); > + > + if (poll_mode == DLB_CQ_POLL_MODE_STD) > + qe_sz = sizeof(struct dlb_dequeue_qe); > + else > + qe_sz = RTE_CACHE_LINE_SIZE; > + > + /* The hardware always uses a CQ depth of at least > + * DLB_MIN_HARDWARE_CQ_DEPTH, even though from the user > + * perspective we support a depth as low as 1 for LDB ports. > + */ > + cq_alloc_depth = RTE_MAX(cfg->cq_depth, > DLB_MIN_HARDWARE_CQ_DEPTH); > + > + /* Calculate the port memory required, including two cache lines for > + * credit pop counts. Round up to the nearest cache line. > + */ > + alloc_sz = 2 * RTE_CACHE_LINE_SIZE + cq_alloc_depth * qe_sz; > + alloc_sz = RTE_CACHE_LINE_ROUNDUP(alloc_sz); > + > + port_base = dlb_alloc_coherent_aligned(&mz, &pc_dma_base, > + alloc_sz, PAGE_SIZE); > + if (port_base == NULL) > + return -ENOMEM; > + > + /* Lock the page in memory */ > + ret = rte_mem_lock_page(port_base); > + if (ret < 0) { > + DLB_LOG_ERR("dlb pf pmd could not lock page for device i/o\n"); > + goto create_port_err; > + } > + > + memset(port_base, 0, alloc_sz); > + cq_dma_base = (uintptr_t)(pc_dma_base + (2 * > RTE_CACHE_LINE_SIZE)); > + > + ret = dlb_hw_create_ldb_port(&dlb_dev->hw, > + handle->domain_id, > + cfg, > + pc_dma_base, > + cq_dma_base, > + &response); > + if (ret) > + goto create_port_err; > + > + pp_dma_base = (uintptr_t)dlb_dev->hw.func_kva + PP_BASE(is_dir); > + dlb_port[response.id][DLB_LDB].pp_addr = > + (void *)(uintptr_t)(pp_dma_base + (PAGE_SIZE * response.id)); > + > + dlb_port[response.id][DLB_LDB].cq_base = > + (void *)(uintptr_t)(port_base + (2 * RTE_CACHE_LINE_SIZE)); > + > + dlb_port[response.id][DLB_LDB].ldb_popcount = > + (void *)(uintptr_t)port_base; > + dlb_port[response.id][DLB_LDB].dir_popcount = (void *)(uintptr_t) > + (port_base + RTE_CACHE_LINE_SIZE); > + dlb_port[response.id][DLB_LDB].mz = mz; > + > + *(struct dlb_cmd_response *)cfg->response = response; > + > + DLB_INFO(dev->dlb_device, "Exiting %s() with ret=%d\n", __func__, > ret); > + > +create_port_err:
Need to free the memzone if the PMD jumps to this label > + > + return ret; > +} > + > +static int > +dlb_pf_dir_port_create(struct dlb_hw_dev *handle, > + struct dlb_create_dir_port_args *cfg, > + enum dlb_cq_poll_modes poll_mode) > +{ > + struct dlb_dev *dlb_dev = (struct dlb_dev *)handle->pf_dev; > + struct dlb_cmd_response response = {0}; > + int ret; > + uint8_t *port_base; > + const struct rte_memzone *mz; > + int alloc_sz, qe_sz; > + rte_iova_t pp_dma_base; > + rte_iova_t pc_dma_base; > + rte_iova_t cq_dma_base; > + int is_dir = true; > + > + DLB_INFO(dev->dlb_device, "Entering %s()\n", __func__); > + > + if (poll_mode == DLB_CQ_POLL_MODE_STD) > + qe_sz = sizeof(struct dlb_dequeue_qe); > + else > + qe_sz = RTE_CACHE_LINE_SIZE; > + > + /* Calculate the port memory required, including two cache lines for > + * credit pop counts. Round up to the nearest cache line. > + */ > + alloc_sz = 2 * RTE_CACHE_LINE_SIZE + cfg->cq_depth * qe_sz; > + alloc_sz = RTE_CACHE_LINE_ROUNDUP(alloc_sz); > + > + port_base = dlb_alloc_coherent_aligned(&mz, &pc_dma_base, > + alloc_sz, PAGE_SIZE); > + if (port_base == NULL) > + return -ENOMEM; > + > + /* Lock the page in memory */ > + ret = rte_mem_lock_page(port_base); > + if (ret < 0) { > + DLB_LOG_ERR("dlb pf pmd could not lock page for device i/o\n"); > + goto create_port_err; > + } > + > + memset(port_base, 0, alloc_sz); > + cq_dma_base = (uintptr_t)(pc_dma_base + (2 * > RTE_CACHE_LINE_SIZE)); > + > + ret = dlb_hw_create_dir_port(&dlb_dev->hw, > + handle->domain_id, > + cfg, > + pc_dma_base, > + cq_dma_base, > + &response); > + if (ret) > + goto create_port_err; > + > + pp_dma_base = (uintptr_t)dlb_dev->hw.func_kva + PP_BASE(is_dir); > + dlb_port[response.id][DLB_DIR].pp_addr = > + (void *)(uintptr_t)(pp_dma_base + (PAGE_SIZE * response.id)); > + > + dlb_port[response.id][DLB_DIR].cq_base = > + (void *)(uintptr_t)(port_base + (2 * RTE_CACHE_LINE_SIZE)); > + > + dlb_port[response.id][DLB_DIR].ldb_popcount = > + (void *)(uintptr_t)port_base; > + dlb_port[response.id][DLB_DIR].dir_popcount = (void *)(uintptr_t) > + (port_base + RTE_CACHE_LINE_SIZE); > + dlb_port[response.id][DLB_DIR].mz = mz; > + > + *(struct dlb_cmd_response *)cfg->response = response; > + > + DLB_INFO(dev->dlb_device, "Exiting %s() with ret=%d\n", __func__, > ret); > + > +create_port_err: > + Ditto Thanks, Gage