11/03/2023 16:09, Srikanth Yalavarthi:
> @@ -528,8 +533,8 @@ ml_request_initialize(struct rte_mempool *mp, void 
> *opaque, void *obj, unsigned
>       req->niters = 0;
>  
>       /* quantize data */
> -     rte_ml_io_quantize(t->cmn.opt->dev_id, t->model[t->fid].id,
> -                        t->model[t->fid].info.batch_size, 
> t->model[t->fid].input, req->input);
> +     rte_ml_io_quantize(t->cmn.opt->dev_id, t->model[t->fid].id, 
> t->model[t->fid].nb_batches,
> +                        t->model[t->fid].input, req->input);
>  }
>  
>  int
> @@ -547,7 +552,7 @@ ml_inference_iomem_setup(struct ml_test *test, struct 
> ml_options *opt, uint16_t
>       int ret;
>  
>       /* get input buffer size */
> -     ret = rte_ml_io_input_size_get(opt->dev_id, t->model[fid].id, 
> t->model[fid].info.batch_size,
> +     ret = rte_ml_io_input_size_get(opt->dev_id, t->model[fid].id, 
> t->model[fid].nb_batches,
>                                      &t->model[fid].inp_qsize, 
> &t->model[fid].inp_dsize);
>       if (ret != 0) {
>               ml_err("Failed to get input size, model : %s\n", 
> opt->filelist[fid].model);
> @@ -555,9 +560,8 @@ ml_inference_iomem_setup(struct ml_test *test, struct 
> ml_options *opt, uint16_t
>       }
>  
>       /* get output buffer size */
> -     ret = rte_ml_io_output_size_get(opt->dev_id, t->model[fid].id,
> -                                     t->model[fid].info.batch_size, 
> &t->model[fid].out_qsize,
> -                                     &t->model[fid].out_dsize);
> +     ret = rte_ml_io_output_size_get(opt->dev_id, t->model[fid].id, 
> t->model[fid].nb_batches,
> +                                     &t->model[fid].out_qsize, 
> &t->model[fid].out_dsize);
>       if (ret != 0) {
>               ml_err("Failed to get input size, model : %s\n", 
> opt->filelist[fid].model);
>               return ret;
> @@ -702,7 +706,7 @@ ml_request_finish(struct rte_mempool *mp, void *opaque, 
> void *obj, unsigned int
>               return;
>  
>       t->nb_used++;
> -     rte_ml_io_dequantize(t->cmn.opt->dev_id, model->id, 
> t->model[req->fid].info.batch_size,
> +     rte_ml_io_dequantize(t->cmn.opt->dev_id, model->id, 
> t->model[req->fid].nb_batches,
>                            req->output, model->output);

These changes look unrelated with the topic of the patch.
You should probably fix it when adding those lines at first.


Reply via email to