From: Vidya Sagar Velumuri <vvelum...@marvell.com> Add ODM copy and copy SG ops.
Signed-off-by: Anoob Joseph <ano...@marvell.com> Signed-off-by: Gowrishankar Muthukrishnan <gmuthukri...@marvell.com> Signed-off-by: Vidya Sagar Velumuri <vvelum...@marvell.com> --- drivers/dma/odm/odm_dmadev.c | 236 +++++++++++++++++++++++++++++++++++ 1 file changed, 236 insertions(+) diff --git a/drivers/dma/odm/odm_dmadev.c b/drivers/dma/odm/odm_dmadev.c index 13b2588246..b21be83a89 100644 --- a/drivers/dma/odm/odm_dmadev.c +++ b/drivers/dma/odm/odm_dmadev.c @@ -9,6 +9,7 @@ #include <rte_common.h> #include <rte_dmadev.h> #include <rte_dmadev_pmd.h> +#include <rte_memcpy.h> #include <rte_pci.h> #include "odm.h" @@ -87,6 +88,238 @@ odm_dmadev_close(struct rte_dma_dev *dev) return 0; } +static int +odm_dmadev_copy(void *dev_private, uint16_t vchan, rte_iova_t src, rte_iova_t dst, uint32_t length, + uint64_t flags) +{ + uint16_t pending_submit_len, pending_submit_cnt, iring_sz_available, iring_head; + const int num_words = ODM_IRING_ENTRY_SIZE_MIN; + struct odm_dev *odm = dev_private; + uint64_t *iring_head_ptr; + struct odm_queue *vq; + uint64_t h; + + const union odm_instr_hdr_s hdr = { + .s.ct = ODM_HDR_CT_CW_NC, + .s.xtype = ODM_XTYPE_INTERNAL, + .s.nfst = 1, + .s.nlst = 1, + }; + + vq = &odm->vq[vchan]; + + h = length; + h |= ((uint64_t)length << 32); + + const uint16_t max_iring_words = vq->iring_max_words; + + iring_sz_available = vq->iring_sz_available; + pending_submit_len = vq->pending_submit_len; + pending_submit_cnt = vq->pending_submit_cnt; + iring_head_ptr = vq->iring_mz->addr; + iring_head = vq->iring_head; + + if (iring_sz_available < num_words) + return -ENOSPC; + + if ((iring_head + num_words) >= max_iring_words) { + + iring_head_ptr[iring_head] = hdr.u; + iring_head = (iring_head + 1) % max_iring_words; + + iring_head_ptr[iring_head] = h; + iring_head = (iring_head + 1) % max_iring_words; + + iring_head_ptr[iring_head] = src; + iring_head = (iring_head + 1) % max_iring_words; + + iring_head_ptr[iring_head] = dst; + iring_head = (iring_head + 1) % max_iring_words; + } else { + iring_head_ptr[iring_head++] = hdr.u; + iring_head_ptr[iring_head++] = h; + iring_head_ptr[iring_head++] = src; + iring_head_ptr[iring_head++] = dst; + } + + pending_submit_len += num_words; + + if (flags & RTE_DMA_OP_FLAG_SUBMIT) { + rte_wmb(); + odm_write64(pending_submit_len, odm->rbase + ODM_VDMA_DBELL(vchan)); + vq->stats.submitted += pending_submit_cnt + 1; + vq->pending_submit_len = 0; + vq->pending_submit_cnt = 0; + } else { + vq->pending_submit_len = pending_submit_len; + vq->pending_submit_cnt++; + } + + vq->iring_head = iring_head; + + vq->iring_sz_available = iring_sz_available - num_words; + + /* No extra space to save. Skip entry in extra space ring. */ + vq->ins_ring_head = (vq->ins_ring_head + 1) % vq->cring_max_entry; + + return vq->desc_idx++; +} + +static inline void +odm_dmadev_fill_sg(uint64_t *cmd, const struct rte_dma_sge *src, const struct rte_dma_sge *dst, + uint16_t nb_src, uint16_t nb_dst, union odm_instr_hdr_s *hdr) +{ + int i = 0, j = 0; + uint64_t h = 0; + + cmd[j++] = hdr->u; + /* When nb_src is even */ + if (!(nb_src & 0x1)) { + /* Fill the iring with src pointers */ + for (i = 1; i < nb_src; i += 2) { + h = ((uint64_t)src[i].length << 32) | src[i - 1].length; + cmd[j++] = h; + cmd[j++] = src[i - 1].addr; + cmd[j++] = src[i].addr; + } + + /* Fill the iring with dst pointers */ + for (i = 1; i < nb_dst; i += 2) { + h = ((uint64_t)dst[i].length << 32) | dst[i - 1].length; + cmd[j++] = h; + cmd[j++] = dst[i - 1].addr; + cmd[j++] = dst[i].addr; + } + + /* Handle the last dst pointer when nb_dst is odd */ + if (nb_dst & 0x1) { + h = dst[nb_dst - 1].length; + cmd[j++] = h; + cmd[j++] = dst[nb_dst - 1].addr; + cmd[j++] = 0; + } + } else { + /* When nb_src is odd */ + + /* Fill the iring with src pointers */ + for (i = 1; i < nb_src; i += 2) { + h = ((uint64_t)src[i].length << 32) | src[i - 1].length; + cmd[j++] = h; + cmd[j++] = src[i - 1].addr; + cmd[j++] = src[i].addr; + } + + /* Handle the last src pointer */ + h = ((uint64_t)dst[0].length << 32) | src[nb_src - 1].length; + cmd[j++] = h; + cmd[j++] = src[nb_src - 1].addr; + cmd[j++] = dst[0].addr; + + /* Fill the iring with dst pointers */ + for (i = 2; i < nb_dst; i += 2) { + h = ((uint64_t)dst[i].length << 32) | dst[i - 1].length; + cmd[j++] = h; + cmd[j++] = dst[i - 1].addr; + cmd[j++] = dst[i].addr; + } + + /* Handle the last dst pointer when nb_dst is even */ + if (!(nb_dst & 0x1)) { + h = dst[nb_dst - 1].length; + cmd[j++] = h; + cmd[j++] = dst[nb_dst - 1].addr; + cmd[j++] = 0; + } + } +} + +static int +odm_dmadev_copy_sg(void *dev_private, uint16_t vchan, const struct rte_dma_sge *src, + const struct rte_dma_sge *dst, uint16_t nb_src, uint16_t nb_dst, uint64_t flags) +{ + uint16_t pending_submit_len, pending_submit_cnt, iring_head, ins_ring_head; + uint16_t iring_sz_available, i, nb, num_words; + uint64_t cmd[ODM_IRING_ENTRY_SIZE_MAX]; + struct odm_dev *odm = dev_private; + uint32_t s_sz = 0, d_sz = 0; + uint64_t *iring_head_ptr; + struct odm_queue *vq; + union odm_instr_hdr_s hdr = { + .s.ct = ODM_HDR_CT_CW_NC, + .s.xtype = ODM_XTYPE_INTERNAL, + }; + + vq = &odm->vq[vchan]; + const uint16_t max_iring_words = vq->iring_max_words; + + iring_head_ptr = vq->iring_mz->addr; + iring_head = vq->iring_head; + iring_sz_available = vq->iring_sz_available; + ins_ring_head = vq->ins_ring_head; + pending_submit_len = vq->pending_submit_len; + pending_submit_cnt = vq->pending_submit_cnt; + + if (unlikely(nb_src > 4 || nb_dst > 4)) + return -EINVAL; + + for (i = 0; i < nb_src; i++) + s_sz += src[i].length; + + for (i = 0; i < nb_dst; i++) + d_sz += dst[i].length; + + if (s_sz != d_sz) + return -EINVAL; + + nb = nb_src + nb_dst; + hdr.s.nfst = nb_src; + hdr.s.nlst = nb_dst; + num_words = 1 + 3 * (nb / 2 + (nb & 0x1)); + + if (iring_sz_available < num_words) + return -ENOSPC; + + if ((iring_head + num_words) >= max_iring_words) { + uint16_t words_avail = max_iring_words - iring_head; + uint16_t words_pend = num_words - words_avail; + + if (unlikely(words_avail + words_pend > ODM_IRING_ENTRY_SIZE_MAX)) + return -ENOSPC; + + odm_dmadev_fill_sg(cmd, src, dst, nb_src, nb_dst, &hdr); + rte_memcpy((void *)&iring_head_ptr[iring_head], (void *)cmd, words_avail * 8); + rte_memcpy((void *)iring_head_ptr, (void *)&cmd[words_avail], words_pend * 8); + iring_head = words_pend; + } else { + odm_dmadev_fill_sg(&iring_head_ptr[iring_head], src, dst, nb_src, nb_dst, &hdr); + iring_head += num_words; + } + + pending_submit_len += num_words; + + if (flags & RTE_DMA_OP_FLAG_SUBMIT) { + rte_wmb(); + odm_write64(pending_submit_len, odm->rbase + ODM_VDMA_DBELL(vchan)); + vq->stats.submitted += pending_submit_cnt + 1; + vq->pending_submit_len = 0; + vq->pending_submit_cnt = 0; + } else { + vq->pending_submit_len = pending_submit_len; + vq->pending_submit_cnt++; + } + + vq->iring_head = iring_head; + + vq->iring_sz_available = iring_sz_available - num_words; + + /* Save extra space used for the instruction. */ + vq->extra_ins_sz[ins_ring_head] = num_words - 4; + + vq->ins_ring_head = (ins_ring_head + 1) % vq->cring_max_entry; + + return vq->desc_idx++; +} + static int odm_stats_get(const struct rte_dma_dev *dev, uint16_t vchan, struct rte_dma_stats *rte_stats, uint32_t size) @@ -184,6 +417,9 @@ odm_dmadev_probe(struct rte_pci_driver *pci_drv __rte_unused, struct rte_pci_dev dmadev->fp_obj->dev_private = odm; dmadev->dev_ops = &odm_dmadev_ops; + dmadev->fp_obj->copy = odm_dmadev_copy; + dmadev->fp_obj->copy_sg = odm_dmadev_copy_sg; + odm->pci_dev = pci_dev; rc = odm_dev_init(odm); -- 2.25.1