11/03/2023 16:09, Srikanth Yalavarthi: > @@ -528,8 +533,8 @@ ml_request_initialize(struct rte_mempool *mp, void > *opaque, void *obj, unsigned > req->niters = 0; > > /* quantize data */ > - rte_ml_io_quantize(t->cmn.opt->dev_id, t->model[t->fid].id, > - t->model[t->fid].info.batch_size, > t->model[t->fid].input, req->input); > + rte_ml_io_quantize(t->cmn.opt->dev_id, t->model[t->fid].id, > t->model[t->fid].nb_batches, > + t->model[t->fid].input, req->input); > } > > int > @@ -547,7 +552,7 @@ ml_inference_iomem_setup(struct ml_test *test, struct > ml_options *opt, uint16_t > int ret; > > /* get input buffer size */ > - ret = rte_ml_io_input_size_get(opt->dev_id, t->model[fid].id, > t->model[fid].info.batch_size, > + ret = rte_ml_io_input_size_get(opt->dev_id, t->model[fid].id, > t->model[fid].nb_batches, > &t->model[fid].inp_qsize, > &t->model[fid].inp_dsize); > if (ret != 0) { > ml_err("Failed to get input size, model : %s\n", > opt->filelist[fid].model); > @@ -555,9 +560,8 @@ ml_inference_iomem_setup(struct ml_test *test, struct > ml_options *opt, uint16_t > } > > /* get output buffer size */ > - ret = rte_ml_io_output_size_get(opt->dev_id, t->model[fid].id, > - t->model[fid].info.batch_size, > &t->model[fid].out_qsize, > - &t->model[fid].out_dsize); > + ret = rte_ml_io_output_size_get(opt->dev_id, t->model[fid].id, > t->model[fid].nb_batches, > + &t->model[fid].out_qsize, > &t->model[fid].out_dsize); > if (ret != 0) { > ml_err("Failed to get input size, model : %s\n", > opt->filelist[fid].model); > return ret; > @@ -702,7 +706,7 @@ ml_request_finish(struct rte_mempool *mp, void *opaque, > void *obj, unsigned int > return; > > t->nb_used++; > - rte_ml_io_dequantize(t->cmn.opt->dev_id, model->id, > t->model[req->fid].info.batch_size, > + rte_ml_io_dequantize(t->cmn.opt->dev_id, model->id, > t->model[req->fid].nb_batches, > req->output, model->output);
These changes look unrelated with the topic of the patch. You should probably fix it when adding those lines at first.