Hi Shameer, On 7/14/25 5:59 PM, Shameer Kolothum wrote: > From: Nicolin Chen <nicol...@nvidia.com> > > Use the provided smmuv3-accel helper functions to issue the > invalidation commands to host SMMUv3. > > Signed-off-by: Nicolin Chen <nicol...@nvidia.com> > Signed-off-by: Shameer Kolothum <shameerali.kolothum.th...@huawei.com> > --- > hw/arm/smmuv3-internal.h | 11 +++++++++++ > hw/arm/smmuv3.c | 28 ++++++++++++++++++++++++++++ > 2 files changed, 39 insertions(+) > > diff --git a/hw/arm/smmuv3-internal.h b/hw/arm/smmuv3-internal.h > index 8cb6a9238a..f3aeaf6375 100644 > --- a/hw/arm/smmuv3-internal.h > +++ b/hw/arm/smmuv3-internal.h > @@ -233,6 +233,17 @@ static inline bool smmuv3_gerror_irq_enabled(SMMUv3State > *s) > #define Q_CONS_WRAP(q) (((q)->cons & WRAP_MASK(q)) >> (q)->log2size) > #define Q_PROD_WRAP(q) (((q)->prod & WRAP_MASK(q)) >> (q)->log2size) > > +static inline int smmuv3_q_ncmds(SMMUQueue *q) > +{ > + uint32_t prod = Q_PROD(q); > + uint32_t cons = Q_CONS(q); > + > + if (Q_PROD_WRAP(q) == Q_CONS_WRAP(q)) > + return prod - cons; > + else > + return WRAP_MASK(q) - cons + prod; > +} > + > static inline bool smmuv3_q_full(SMMUQueue *q) > { > return ((q->cons ^ q->prod) & WRAP_INDEX_MASK(q)) == WRAP_MASK(q); > diff --git a/hw/arm/smmuv3.c b/hw/arm/smmuv3.c > index c94bfe6564..97ecca0764 100644 > --- a/hw/arm/smmuv3.c > +++ b/hw/arm/smmuv3.c > @@ -1285,10 +1285,17 @@ static int smmuv3_cmdq_consume(SMMUv3State *s) > SMMUCmdError cmd_error = SMMU_CERROR_NONE; > SMMUQueue *q = &s->cmdq; > SMMUCommandType type = 0; > + SMMUCommandBatch batch = {}; > + uint32_t ncmds; > > if (!smmuv3_cmdq_enabled(s)) { > return 0; > } > + > + ncmds = smmuv3_q_ncmds(q); > + batch.cmds = g_new0(Cmd, ncmds); > + batch.cons = g_new0(uint32_t, ncmds); so you are provisionning space for n commands found in the queue, independently on knowing whether they will be batched, ie. only invalidation commands are. Then commands are added in the batch one by one and you increment batch->ncmds in smmuv3_accel_batch_cmd. I agree with Jonathan. This looks weird. AT least I would introduce a kelper that inits a Back of ncmds and I would make all the batch fields private. You you end up with the init + smmuv3_accel_add_cmd_to_batch(batch, cmd). Then independently on the ncmds you can issue a smmuv3_accel_issue_cmd_batch that would return if there is nothing in the batch. You also need a batch deallocation helper. I remember I expressed in the past my concern about having commands executed out of order. I don't remember out conclusion on that but this shall be clearly studied and conclusion shall be put in the commit message. > + > /* > * some commands depend on register values, typically CR0. In case those > * register values change while handling the command, spec says it > @@ -1383,6 +1390,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s) > > trace_smmuv3_cmdq_cfgi_cd(sid); > smmuv3_flush_config(sdev); > + smmuv3_accel_batch_cmd(sdev->smmu, sdev, &batch, &cmd, &q->cons); > break; > } > case SMMU_CMD_TLBI_NH_ASID: > @@ -1406,6 +1414,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s) > trace_smmuv3_cmdq_tlbi_nh_asid(asid); > smmu_inv_notifiers_all(&s->smmu_state); > smmu_iotlb_inv_asid_vmid(bs, asid, vmid); > + smmuv3_accel_batch_cmd(bs, NULL, &batch, &cmd, &q->cons); > break; > } > case SMMU_CMD_TLBI_NH_ALL: > @@ -1433,6 +1442,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s) > trace_smmuv3_cmdq_tlbi_nsnh(); > smmu_inv_notifiers_all(&s->smmu_state); > smmu_iotlb_inv_all(bs); > + smmuv3_accel_batch_cmd(bs, NULL, &batch, &cmd, &q->cons); > break; > case SMMU_CMD_TLBI_NH_VAA: > case SMMU_CMD_TLBI_NH_VA: > @@ -1441,6 +1451,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s) > break; > } > smmuv3_range_inval(bs, &cmd, SMMU_STAGE_1); > + smmuv3_accel_batch_cmd(bs, NULL, &batch, &cmd, &q->cons); > break; > case SMMU_CMD_TLBI_S12_VMALL: > { > @@ -1499,12 +1510,29 @@ static int smmuv3_cmdq_consume(SMMUv3State *s) > queue_cons_incr(q); > } > > + qemu_mutex_lock(&s->mutex); > + if (!cmd_error && batch.ncmds) { > + if (!smmuv3_accel_issue_cmd_batch(bs, &batch)) { > + if (batch.ncmds) { > + q->cons = batch.cons[batch.ncmds - 1]; > + } else { > + q->cons = batch.cons[0]; /* FIXME: Check */ > + } > + qemu_log_mask(LOG_GUEST_ERROR, "Illegal command type: %d\n", > + CMD_TYPE(&batch.cmds[batch.ncmds])); Can't you have other error types returned?
Thanks Eric > + cmd_error = SMMU_CERROR_ILL; > + } > + } > + qemu_mutex_unlock(&s->mutex); > + > if (cmd_error) { > trace_smmuv3_cmdq_consume_error(smmu_cmd_string(type), cmd_error); > smmu_write_cmdq_err(s, cmd_error); > smmuv3_trigger_irq(s, SMMU_IRQ_GERROR, R_GERROR_CMDQ_ERR_MASK); > } > > + g_free(batch.cmds); > + g_free(batch.cons); > trace_smmuv3_cmdq_consume_out(Q_PROD(q), Q_CONS(q), > Q_PROD_WRAP(q), Q_CONS_WRAP(q)); >