>> IOVA mode in DPDK is either PA or VA. The new build option iova_as_va >> configures the mode to VA at compile time and prevents setting it to PA >> at runtime. For now, all drivers which are not always enabled are >> disabled with this option. Supported driver can set the flag >> pmd_iova_as_va in its build file to enable build. >> >> mbuf structure holds the physical (PA) and virtual address (VA) of a >> buffer. if IOVA mode is set to VA, PA is redundant as it is the same as >> VA. So PA field need not be updated and marked invalid if the build is >> configured to use only VA. >> >> Signed-off-by: Shijith Thotton <sthot...@marvell.com> >> --- >> app/test-bbdev/test_bbdev_perf.c | 2 +- >> app/test-crypto-perf/cperf_test_common.c | 5 +-- >> app/test/test_bpf.c | 2 +- >> app/test/test_dmadev.c | 33 ++++++--------- >> app/test/test_mbuf.c | 12 +++--- >> app/test/test_pcapng.c | 2 +- >> config/meson.build | 1 + >> drivers/meson.build | 6 +++ >> lib/eal/linux/eal.c | 7 +++ >> lib/mbuf/rte_mbuf.c | 8 ++-- >> lib/mbuf/rte_mbuf.h | 17 +++++--- >> lib/mbuf/rte_mbuf_core.h | 10 +++++ >> lib/vhost/vhost.h | 2 +- >> lib/vhost/vhost_crypto.c | 54 ++++++++++++++++++------ >> meson_options.txt | 2 + >> 15 files changed, 109 insertions(+), 54 deletions(-) >> >> diff --git a/app/test-bbdev/test_bbdev_perf.c b/app/test- >bbdev/test_bbdev_perf.c >> index 8fab52d821..f6aa25b67d 100644 >> --- a/app/test-bbdev/test_bbdev_perf.c >> +++ b/app/test-bbdev/test_bbdev_perf.c >> @@ -1001,7 +1001,7 @@ init_op_data_objs(struct rte_bbdev_op_data *bufs, >> seg->length); >> memcpy(data, seg->addr, seg->length); >> m_head->buf_addr = data; >> - m_head->buf_iova = rte_malloc_virt2iova(data); >> + rte_mbuf_iova_set(m_head, >rte_malloc_virt2iova(data)); > >Wouldn't it be better to have a preliminary patch that replaces direct >accesses to m->buf_iova by rte_mbuf_iova_*() functions in app and libs? >This would make this commit smaller to read. Yes. I will add this change in v4.
> >If I understand properly, the drivers/ part has to be done at the same time >than setting "pmd_iova_as_va" in the meson config. > This approach was taken as per previous discussions. Also removing buf_iova from a PMD would require proper testing and performance checks. Current approach would give ample time for this. >> m_head->data_off = 0; >> m_head->data_len = seg->length; >> } else { >> diff --git a/app/test-crypto-perf/cperf_test_common.c b/app/test-crypto- >perf/cperf_test_common.c >> index 00aadc9a47..27646cd619 100644 >> --- a/app/test-crypto-perf/cperf_test_common.c >> +++ b/app/test-crypto-perf/cperf_test_common.c >> @@ -26,8 +26,7 @@ fill_single_seg_mbuf(struct rte_mbuf *m, struct >rte_mempool *mp, >> /* start of buffer is after mbuf structure and priv data */ >> m->priv_size = 0; >> m->buf_addr = (char *)m + mbuf_hdr_size; >> - m->buf_iova = rte_mempool_virt2iova(obj) + >> - mbuf_offset + mbuf_hdr_size; >> + rte_mbuf_iova_set(m, rte_mempool_virt2iova(obj) + mbuf_offset + >mbuf_hdr_size); >> m->buf_len = segment_sz; >> m->data_len = data_len; >> m->pkt_len = data_len; >> @@ -58,7 +57,7 @@ fill_multi_seg_mbuf(struct rte_mbuf *m, struct >rte_mempool *mp, >> /* start of buffer is after mbuf structure and priv data */ >> m->priv_size = 0; >> m->buf_addr = (char *)m + mbuf_hdr_size; >> - m->buf_iova = next_seg_phys_addr; >> + rte_mbuf_iova_set(m, next_seg_phys_addr); >> next_seg_phys_addr += mbuf_hdr_size + segment_sz; >> m->buf_len = segment_sz; >> m->data_len = data_len; >> diff --git a/app/test/test_bpf.c b/app/test/test_bpf.c >> index 97f500809e..f5af5e8a3f 100644 >> --- a/app/test/test_bpf.c >> +++ b/app/test/test_bpf.c >> @@ -2600,7 +2600,7 @@ dummy_mbuf_prep(struct rte_mbuf *mb, uint8_t >buf[], uint32_t buf_len, >> uint8_t *db; >> >> mb->buf_addr = buf; >> - mb->buf_iova = (uintptr_t)buf; >> + rte_mbuf_iova_set(mb, (uintptr_t)buf); >> mb->buf_len = buf_len; >> rte_mbuf_refcnt_set(mb, 1); >> >> diff --git a/app/test/test_dmadev.c b/app/test/test_dmadev.c >> index 9e8e101f40..8306947eda 100644 >> --- a/app/test/test_dmadev.c >> +++ b/app/test/test_dmadev.c >> @@ -110,8 +110,8 @@ do_multi_copies(int16_t dev_id, uint16_t vchan, >> for (j = 0; j < COPY_LEN/sizeof(uint64_t); j++) >> src_data[j] = rte_rand(); >> >> - if (rte_dma_copy(dev_id, vchan, srcs[i]->buf_iova + srcs[i]- >>data_off, >> - dsts[i]->buf_iova + dsts[i]->data_off, COPY_LEN, >0) != id_count++) >> + if (rte_dma_copy(dev_id, vchan, rte_pktmbuf_iova_offset(srcs[i], >0), >> + rte_pktmbuf_iova_offset(dsts[i], 0), COPY_LEN, >0) != id_count++) >> ERR_RETURN("Error with rte_dma_copy for buffer %u\n", >i); >> } >> rte_dma_submit(dev_id, vchan); >> @@ -317,9 +317,8 @@ test_failure_in_full_burst(int16_t dev_id, uint16_t >vchan, bool fence, >> rte_dma_stats_get(dev_id, vchan, &baseline); /* get a baseline set of >stats */ >> for (i = 0; i < COMP_BURST_SZ; i++) { >> int id = rte_dma_copy(dev_id, vchan, >> - (i == fail_idx ? 0 : (srcs[i]->buf_iova + >> srcs[i]- >>data_off)), >> - dsts[i]->buf_iova + dsts[i]->data_off, >> - COPY_LEN, OPT_FENCE(i)); >> + (i == fail_idx ? 0 : >rte_pktmbuf_iova_offset(srcs[i], 0)), >> + rte_pktmbuf_iova_offset(dsts[i], 0), >COPY_LEN, OPT_FENCE(i)); >> if (id < 0) >> ERR_RETURN("Error with rte_dma_copy for buffer %u\n", >i); >> if (i == fail_idx) >> @@ -407,9 +406,8 @@ test_individual_status_query_with_failure(int16_t >dev_id, uint16_t vchan, bool f >> >> for (j = 0; j < COMP_BURST_SZ; j++) { >> int id = rte_dma_copy(dev_id, vchan, >> - (j == fail_idx ? 0 : (srcs[j]->buf_iova + >> srcs[j]- >>data_off)), >> - dsts[j]->buf_iova + dsts[j]->data_off, >> - COPY_LEN, OPT_FENCE(j)); >> + (j == fail_idx ? 0 : >rte_pktmbuf_iova_offset(srcs[j], 0)), >> + rte_pktmbuf_iova_offset(dsts[j], 0), >COPY_LEN, OPT_FENCE(j)); >> if (id < 0) >> ERR_RETURN("Error with rte_dma_copy for buffer %u\n", >j); >> if (j == fail_idx) >> @@ -470,9 +468,8 @@ test_single_item_status_query_with_failure(int16_t >dev_id, uint16_t vchan, >> >> for (j = 0; j < COMP_BURST_SZ; j++) { >> int id = rte_dma_copy(dev_id, vchan, >> - (j == fail_idx ? 0 : (srcs[j]->buf_iova + >> srcs[j]- >>data_off)), >> - dsts[j]->buf_iova + dsts[j]->data_off, >> - COPY_LEN, 0); >> + (j == fail_idx ? 0 : >rte_pktmbuf_iova_offset(srcs[j], 0)), >> + rte_pktmbuf_iova_offset(dsts[j], 0), >COPY_LEN, 0); >> if (id < 0) >> ERR_RETURN("Error with rte_dma_copy for buffer %u\n", >j); >> if (j == fail_idx) >> @@ -529,15 +526,14 @@ test_multi_failure(int16_t dev_id, uint16_t vchan, >struct rte_mbuf **srcs, struc >> >> /* enqueue and gather completions in one go */ >> for (j = 0; j < COMP_BURST_SZ; j++) { >> - uintptr_t src = srcs[j]->buf_iova + srcs[j]->data_off; >> + uintptr_t src = rte_pktmbuf_iova_offset(srcs[j], 0); >> /* set up for failure if the current index is anywhere is the >> fails >array */ >> for (i = 0; i < num_fail; i++) >> if (j == fail[i]) >> src = 0; >> >> - int id = rte_dma_copy(dev_id, vchan, >> - src, dsts[j]->buf_iova + dsts[j]->data_off, >> - COPY_LEN, 0); >> + int id = rte_dma_copy(dev_id, vchan, src, >rte_pktmbuf_iova_offset(dsts[j], 0), >> + COPY_LEN, 0); >> if (id < 0) >> ERR_RETURN("Error with rte_dma_copy for buffer %u\n", >j); >> } >> @@ -565,15 +561,14 @@ test_multi_failure(int16_t dev_id, uint16_t vchan, >struct rte_mbuf **srcs, struc >> >> /* enqueue and gather completions in bursts, but getting errors one at a >time */ >> for (j = 0; j < COMP_BURST_SZ; j++) { >> - uintptr_t src = srcs[j]->buf_iova + srcs[j]->data_off; >> + uintptr_t src = rte_pktmbuf_iova_offset(srcs[j], 0); >> /* set up for failure if the current index is anywhere is the >> fails >array */ >> for (i = 0; i < num_fail; i++) >> if (j == fail[i]) >> src = 0; >> >> - int id = rte_dma_copy(dev_id, vchan, >> - src, dsts[j]->buf_iova + dsts[j]->data_off, >> - COPY_LEN, 0); >> + int id = rte_dma_copy(dev_id, vchan, src, >rte_pktmbuf_iova_offset(dsts[j], 0), >> + COPY_LEN, 0); >> if (id < 0) >> ERR_RETURN("Error with rte_dma_copy for buffer %u\n", >j); >> } >> diff --git a/app/test/test_mbuf.c b/app/test/test_mbuf.c >> index e09b2549ca..45431f2c9c 100644 >> --- a/app/test/test_mbuf.c >> +++ b/app/test/test_mbuf.c >> @@ -1232,11 +1232,13 @@ test_failing_mbuf_sanity_check(struct >rte_mempool *pktmbuf_pool) >> return -1; >> } >> >> - badbuf = *buf; >> - badbuf.buf_iova = 0; >> - if (verify_mbuf_check_panics(&badbuf)) { >> - printf("Error with bad-physaddr mbuf test\n"); >> - return -1; >> + if (!RTE_IOVA_AS_VA) { >> + badbuf = *buf; >> + rte_mbuf_iova_set(&badbuf, 0); >> + if (verify_mbuf_check_panics(&badbuf)) { >> + printf("Error with bad-physaddr mbuf test\n"); >> + return -1; >> + } >> } >> >> badbuf = *buf; >> diff --git a/app/test/test_pcapng.c b/app/test/test_pcapng.c >> index 320dacea34..abbf00f6da 100644 >> --- a/app/test/test_pcapng.c >> +++ b/app/test/test_pcapng.c >> @@ -40,7 +40,7 @@ dummy_mbuf_prep(struct rte_mbuf *mb, uint8_t buf[], >uint32_t buf_len, >> uint8_t *db; >> >> mb->buf_addr = buf; >> - mb->buf_iova = (uintptr_t)buf; >> + rte_mbuf_iova_set(mb, (uintptr_t)buf); >> mb->buf_len = buf_len; >> rte_mbuf_refcnt_set(mb, 1); >> >> diff --git a/config/meson.build b/config/meson.build >> index 7f7b6c92fd..6b6c3e7eb6 100644 >> --- a/config/meson.build >> +++ b/config/meson.build >> @@ -309,6 +309,7 @@ endif >> if get_option('mbuf_refcnt_atomic') >> dpdk_conf.set('RTE_MBUF_REFCNT_ATOMIC', true) >> endif >> +dpdk_conf.set10('RTE_IOVA_AS_VA', get_option('iova_as_va')) >> >> compile_time_cpuflags = [] >> subdir(arch_subdir) >> diff --git a/drivers/meson.build b/drivers/meson.build >> index 376a64f4da..989770cffd 100644 >> --- a/drivers/meson.build >> +++ b/drivers/meson.build >> @@ -105,6 +105,7 @@ foreach subpath:subdirs >> ext_deps = [] >> pkgconfig_extra_libs = [] >> testpmd_sources = [] >> + pmd_iova_as_va = false > >This option should be documented, however I don't know where is the proper >place. A comment here would be a good start I think. > Will add. >I'm trying to find a more explicit name, but it's not easy. >What do you think about pmd_supports_disable_iova_as_pa? > Makes sense. I will change to pmd_supports_disable_iova_as_pa. >Explicit is always better, it could avoid someone adding a new driver to >blindly copy the flag from a template driver. > >> >> if not enable_drivers.contains(drv_path) >> build = false >> @@ -122,6 +123,11 @@ foreach subpath:subdirs >> # pull in driver directory which should update all the local >> variables >> subdir(drv_path) >> >> + if dpdk_conf.get('RTE_IOVA_AS_VA') == 1 and not pmd_iova_as_va >> and >not always_enable.contains(drv_path) >> + build = false >> + reason = 'driver does not support IOVA as VA mode' >> + endif >> + >> # get dependency objs from strings >> shared_deps = ext_deps >> static_deps = ext_deps >> diff --git a/lib/eal/linux/eal.c b/lib/eal/linux/eal.c >> index 37d29643a5..b70c4dcc5f 100644 >> --- a/lib/eal/linux/eal.c >> +++ b/lib/eal/linux/eal.c >> @@ -1127,6 +1127,13 @@ rte_eal_init(int argc, char **argv) >> return -1; >> } >> >> + if (rte_eal_iova_mode() == RTE_IOVA_PA && RTE_IOVA_AS_VA) { >> + rte_eal_init_alert( >> + "Cannot use IOVA as 'PA' since build is configured to >> use >only 'VA'"); >> + rte_errno = EINVAL; >> + return -1; >> + } >> + >> RTE_LOG(INFO, EAL, "Selected IOVA mode '%s'\n", >> rte_eal_iova_mode() == RTE_IOVA_PA ? "PA" : "VA"); >> >> diff --git a/lib/mbuf/rte_mbuf.c b/lib/mbuf/rte_mbuf.c >> index a2307cebe6..5af290c53a 100644 >> --- a/lib/mbuf/rte_mbuf.c >> +++ b/lib/mbuf/rte_mbuf.c >> @@ -89,7 +89,7 @@ rte_pktmbuf_init(struct rte_mempool *mp, >> /* start of buffer is after mbuf structure and priv data */ >> m->priv_size = priv_size; >> m->buf_addr = (char *)m + mbuf_size; >> - m->buf_iova = rte_mempool_virt2iova(m) + mbuf_size; >> + rte_mbuf_iova_set(m, rte_mempool_virt2iova(m) + mbuf_size); >> m->buf_len = (uint16_t)buf_len; >> >> /* keep some headroom between start of buffer and data */ >> @@ -187,8 +187,8 @@ __rte_pktmbuf_init_extmem(struct rte_mempool *mp, >> RTE_ASSERT(ctx->off + ext_mem->elt_size <= ext_mem->buf_len); >> >> m->buf_addr = RTE_PTR_ADD(ext_mem->buf_ptr, ctx->off); >> - m->buf_iova = ext_mem->buf_iova == RTE_BAD_IOVA ? >> - RTE_BAD_IOVA : (ext_mem->buf_iova + ctx->off); >> + rte_mbuf_iova_set(m, ext_mem->buf_iova == RTE_BAD_IOVA ? >RTE_BAD_IOVA : >> + (ext_mem- >>buf_iova + ctx->off)); >> >> ctx->off += ext_mem->elt_size; >> if (ctx->off + ext_mem->elt_size > ext_mem->buf_len) { >> @@ -388,7 +388,7 @@ int rte_mbuf_check(const struct rte_mbuf *m, int >is_header, >> *reason = "bad mbuf pool"; >> return -1; >> } >> - if (m->buf_iova == 0) { >> + if (m->buf_iova == 0 && !RTE_IOVA_AS_VA) { >> *reason = "bad IO addr"; >> return -1; >> } >> diff --git a/lib/mbuf/rte_mbuf.h b/lib/mbuf/rte_mbuf.h >> index 9811e8c760..05be146bc2 100644 >> --- a/lib/mbuf/rte_mbuf.h >> +++ b/lib/mbuf/rte_mbuf.h >> @@ -146,7 +146,7 @@ static inline uint16_t rte_pktmbuf_priv_size(struct >rte_mempool *mp); >> static inline rte_iova_t >> rte_mbuf_data_iova(const struct rte_mbuf *mb) >> { >> - return mb->buf_iova + mb->data_off; >> + return (RTE_IOVA_AS_VA ? (uint64_t)mb->buf_addr : mb->buf_iova) + >mb->data_off; > >nit: cast should be rte_iova_t instead of uint64_t Will change. > >> } >> >> /** >> @@ -164,7 +164,7 @@ rte_mbuf_data_iova(const struct rte_mbuf *mb) >> static inline rte_iova_t >> rte_mbuf_data_iova_default(const struct rte_mbuf *mb) >> { >> - return mb->buf_iova + RTE_PKTMBUF_HEADROOM; >> + return (RTE_IOVA_AS_VA ? (uint64_t)mb->buf_addr : mb->buf_iova) + >RTE_PKTMBUF_HEADROOM; >> } > >same here > Will change. >> >> /** >> @@ -469,6 +469,13 @@ rte_mbuf_ext_refcnt_update(struct >rte_mbuf_ext_shared_info *shinfo, >> __ATOMIC_ACQ_REL); >> } >> >> +static inline void >> +rte_mbuf_iova_set(struct rte_mbuf *m, rte_iova_t iova) >> +{ >> + if (!RTE_IOVA_AS_VA) >> + m->buf_iova = iova; >> +} >> + >> /** Mbuf prefetch */ >> #define RTE_MBUF_PREFETCH_TO_FREE(m) do { \ >> if ((m) != NULL) \ >> @@ -1056,7 +1063,7 @@ rte_pktmbuf_attach_extbuf(struct rte_mbuf *m, void >*buf_addr, >> RTE_ASSERT(shinfo->free_cb != NULL); >> >> m->buf_addr = buf_addr; >> - m->buf_iova = buf_iova; >> + rte_mbuf_iova_set(m, buf_iova); >> m->buf_len = buf_len; >> >> m->data_len = 0; >> @@ -1143,7 +1150,7 @@ static inline void rte_pktmbuf_attach(struct rte_mbuf >*mi, struct rte_mbuf *m) >> >> mi->data_off = m->data_off; >> mi->data_len = m->data_len; >> - mi->buf_iova = m->buf_iova; >> + rte_mbuf_iova_set(mi, m->buf_iova); >> mi->buf_addr = m->buf_addr; >> mi->buf_len = m->buf_len; >> >> @@ -1245,7 +1252,7 @@ static inline void rte_pktmbuf_detach(struct rte_mbuf >*m) >> >> m->priv_size = priv_size; >> m->buf_addr = (char *)m + mbuf_size; >> - m->buf_iova = rte_mempool_virt2iova(m) + mbuf_size; >> + rte_mbuf_iova_set(m, rte_mempool_virt2iova(m) + mbuf_size); >> m->buf_len = (uint16_t)buf_len; >> rte_pktmbuf_reset_headroom(m); >> m->data_len = 0; >> diff --git a/lib/mbuf/rte_mbuf_core.h b/lib/mbuf/rte_mbuf_core.h >> index 3d6ddd6773..c6292e7252 100644 >> --- a/lib/mbuf/rte_mbuf_core.h >> +++ b/lib/mbuf/rte_mbuf_core.h >> @@ -581,6 +581,8 @@ struct rte_mbuf { >> void *buf_addr; /**< Virtual address of segment buffer. */ >> /** >> * Physical address of segment buffer. >> + * This field is invalid if the build is configured to use only >> + * virtual address as IOVA (i.e. RTE_IOVA_AS_VA is 1). >> * Force alignment to 8-bytes, so as to ensure we have the exact >> * same mbuf cacheline0 layout for 32-bit and 64-bit. This makes >> * working on vector drivers easier. > >If the field is invalid, can't we add an #if condition ? I mean: > >#if !RTE_IOVA_AS_VA > rte_iova_t buf_iova; >#else > uint64_t dummy; >#endif > >I think it is preferable, because it would ensure that we never use >buf_iova when RTE_IOVA_AS_VA is set (especially useful when compiling >out-of-tree drivers). > >This would certainly require to change some of the static inline >functions to use #if instead of if(), but I think it's worth the effort. > > Agree. I will change in v4. >> @@ -848,8 +850,12 @@ struct rte_mbuf_ext_shared_info { >> * @param o >> * The offset into the data to calculate address from. >> */ >> +#if RTE_IOVA_AS_VA >> +#define rte_pktmbuf_iova_offset(m, o) rte_pktmbuf_mtod_offset(m, >rte_iova_t, o) >> +#else >> #define rte_pktmbuf_iova_offset(m, o) \ >> (rte_iova_t)((m)->buf_iova + (m)->data_off + (o)) >> +#endif >> >> /** >> * A macro that returns the IO address that points to the start of the >> @@ -858,7 +864,11 @@ struct rte_mbuf_ext_shared_info { >> * @param m >> * The packet mbuf. >> */ >> +#if RTE_IOVA_AS_VA >> +#define rte_pktmbuf_iova(m) rte_pktmbuf_mtod(m, rte_iova_t) >> +#else >> #define rte_pktmbuf_iova(m) rte_pktmbuf_iova_offset(m, 0) >> +#endif >> >> #ifdef __cplusplus >> } >> diff --git a/lib/vhost/vhost.h b/lib/vhost/vhost.h >> index 782d916ae0..05cde6e118 100644 >> --- a/lib/vhost/vhost.h >> +++ b/lib/vhost/vhost.h >> @@ -967,7 +967,7 @@ restore_mbuf(struct rte_mbuf *m) >> /* start of buffer is after mbuf structure and priv data */ >> >> m->buf_addr = (char *)m + mbuf_size; >> - m->buf_iova = rte_mempool_virt2iova(m) + mbuf_size; >> + rte_mbuf_iova_set(m, rte_mempool_virt2iova(m) + mbuf_size); >> m = m->next; >> } >> } >> diff --git a/lib/vhost/vhost_crypto.c b/lib/vhost/vhost_crypto.c >> index 54946f46d9..7b50735796 100644 >> --- a/lib/vhost/vhost_crypto.c >> +++ b/lib/vhost/vhost_crypto.c >> @@ -823,11 +823,17 @@ prepare_sym_cipher_op(struct vhost_crypto >*vcrypto, struct rte_crypto_op *op, >> switch (vcrypto->option) { >> case RTE_VHOST_CRYPTO_ZERO_COPY_ENABLE: >> m_src->data_len = cipher->para.src_data_len; >> - m_src->buf_iova = gpa_to_hpa(vcrypto->dev, desc->addr, >> - cipher->para.src_data_len); >> + if (!RTE_IOVA_AS_VA) { >> + m_src->buf_iova = >> + gpa_to_hpa(vcrypto->dev, desc->addr, cipher- >>para.src_data_len); >> + if (unlikely(m_src->buf_iova == 0)) { >> + VC_LOG_ERR("zero_copy may fail due to cross >page data"); >> + ret = VIRTIO_CRYPTO_ERR; >> + goto error_exit; >> + } >> + } >> m_src->buf_addr = get_data_ptr(vc_req, desc, >VHOST_ACCESS_RO); >> - if (unlikely(m_src->buf_iova == 0 || >> - m_src->buf_addr == NULL)) { >> + if (unlikely(m_src->buf_addr == NULL)) { >> VC_LOG_ERR("zero_copy may fail due to cross page >data"); >> ret = VIRTIO_CRYPTO_ERR; >> goto error_exit; >> @@ -867,10 +873,17 @@ prepare_sym_cipher_op(struct vhost_crypto >*vcrypto, struct rte_crypto_op *op, >> >> switch (vcrypto->option) { >> case RTE_VHOST_CRYPTO_ZERO_COPY_ENABLE: >> - m_dst->buf_iova = gpa_to_hpa(vcrypto->dev, >> - desc->addr, cipher->para.dst_data_len); >> + if (!RTE_IOVA_AS_VA) { >> + m_dst->buf_iova = >> + gpa_to_hpa(vcrypto->dev, desc->addr, cipher- >>para.dst_data_len); >> + if (unlikely(m_dst->buf_iova == 0)) { >> + VC_LOG_ERR("zero_copy may fail due to cross >page data"); >> + ret = VIRTIO_CRYPTO_ERR; >> + goto error_exit; >> + } >> + } >> m_dst->buf_addr = get_data_ptr(vc_req, desc, >VHOST_ACCESS_RW); >> - if (unlikely(m_dst->buf_iova == 0 || m_dst->buf_addr == NULL)) { >> + if (unlikely(m_dst->buf_addr == NULL)) { >> VC_LOG_ERR("zero_copy may fail due to cross page >data"); >> ret = VIRTIO_CRYPTO_ERR; >> goto error_exit; >> @@ -980,11 +993,17 @@ prepare_sym_chain_op(struct vhost_crypto *vcrypto, >struct rte_crypto_op *op, >> case RTE_VHOST_CRYPTO_ZERO_COPY_ENABLE: >> m_src->data_len = chain->para.src_data_len; >> m_dst->data_len = chain->para.dst_data_len; >> - >> - m_src->buf_iova = gpa_to_hpa(vcrypto->dev, desc->addr, >> - chain->para.src_data_len); >> + if (!RTE_IOVA_AS_VA) { >> + m_src->buf_iova = >> + gpa_to_hpa(vcrypto->dev, desc->addr, chain- >>para.src_data_len); >> + if (unlikely(m_src->buf_iova == 0)) { >> + VC_LOG_ERR("zero_copy may fail due to cross >page data"); >> + ret = VIRTIO_CRYPTO_ERR; >> + goto error_exit; >> + } >> + } >> m_src->buf_addr = get_data_ptr(vc_req, desc, >VHOST_ACCESS_RO); >> - if (unlikely(m_src->buf_iova == 0 || m_src->buf_addr == NULL)) { >> + if (unlikely(m_src->buf_addr == NULL)) { >> VC_LOG_ERR("zero_copy may fail due to cross page >data"); >> ret = VIRTIO_CRYPTO_ERR; >> goto error_exit; >> @@ -1024,10 +1043,17 @@ prepare_sym_chain_op(struct vhost_crypto >*vcrypto, struct rte_crypto_op *op, >> >> switch (vcrypto->option) { >> case RTE_VHOST_CRYPTO_ZERO_COPY_ENABLE: >> - m_dst->buf_iova = gpa_to_hpa(vcrypto->dev, >> - desc->addr, chain->para.dst_data_len); >> + if (!RTE_IOVA_AS_VA) { >> + m_dst->buf_iova = >> + gpa_to_hpa(vcrypto->dev, desc->addr, chain- >>para.dst_data_len); >> + if (unlikely(m_dst->buf_iova == 0)) { >> + VC_LOG_ERR("zero_copy may fail due to cross >page data"); >> + ret = VIRTIO_CRYPTO_ERR; >> + goto error_exit; >> + } >> + } >> m_dst->buf_addr = get_data_ptr(vc_req, desc, >VHOST_ACCESS_RW); >> - if (unlikely(m_dst->buf_iova == 0 || m_dst->buf_addr == NULL)) { >> + if (unlikely(m_dst->buf_addr == NULL)) { >> VC_LOG_ERR("zero_copy may fail due to cross page >data"); >> ret = VIRTIO_CRYPTO_ERR; >> goto error_exit; >> diff --git a/meson_options.txt b/meson_options.txt >> index 7c220ad68d..f0fa6cf04c 100644 >> --- a/meson_options.txt >> +++ b/meson_options.txt >> @@ -44,6 +44,8 @@ option('platform', type: 'string', value: 'native', >> description: >> 'Platform to build, either "native", "generic" or a SoC. Please >> refer to the >Linux build guide for more information.') >> option('enable_trace_fp', type: 'boolean', value: false, description: >> 'enable fast path trace points.') >> +option('iova_as_va', type: 'boolean', value: false, description: >> + 'Build which only supports IOVA as VA mode. Unsupported drivers are >disabled.') > >I wonder if we can find a better name for the option. Currently, it is a bit >confusing to me, because iova_as_va=false does not mean that iova_as_va is >disabled. > >What about iova_as_pa=true|false, or enable_iova_as_pa=true|false, or >disable_iova_as_pa=true|false? > Agree. Will go with the option enable_iova_as_pa and will rename macro to RTE_IOVA_AS_PA. >The help string is maybe easier to find, something like >"Enable or disable support for IOVA as PA mode." > >We can also explain that enabling this option removes the buf_iova field from >the mbuf. > >> option('tests', type: 'boolean', value: true, description: >> 'build unit tests') >> option('use_hpet', type: 'boolean', value: false, description: >> -- >> 2.25.1 >>