Hi Zhi,

Thanks, this patch has become super easy to review.

On Thu Feb 27, 2025 at 10:35 AM JST, Zhi Wang wrote:
> There can be multiple cases of handling the GSP RPC messages, which are
> the reply of GSP RPC commands according to the requirement of the
> callers and the nature of the GSP RPC commands.
>
> The current supported reply policies are "callers don't care" and "receive
> the entire message" according to the requirement of the callers. To
> introduce a new policy, factor out the current RPC command reply polices.
> Also, centralize the handling of the reply in a single function.
>
> Factor out NVKM_GSP_RPC_REPLY_NOWAIT as "callers don't care" and
> NVKM_GSP_RPC_REPLY_RECV as "receive the entire message". Introduce a
> kernel doc to document the policies. Factor out
> r535_gsp_rpc_handle_reply().
>
> No functional change is intended for small GSP RPC commands. For large GSP
> commands, the caller decides the policy of how to handle the returned GSP
> RPC message.
>
> Cc: Ben Skeggs <bske...@nvidia.com>
> Cc: Alexandre Courbot <acour...@nvidia.com>
> Signed-off-by: Zhi Wang <z...@nvidia.com>
> ---
>  Documentation/gpu/nouveau.rst                 |  3 +
>  .../gpu/drm/nouveau/include/nvkm/subdev/gsp.h | 34 +++++++--
>  .../gpu/drm/nouveau/nvkm/subdev/bar/r535.c    |  2 +-
>  .../gpu/drm/nouveau/nvkm/subdev/gsp/r535.c    | 75 ++++++++++---------
>  .../drm/nouveau/nvkm/subdev/instmem/r535.c    |  2 +-
>  5 files changed, 72 insertions(+), 44 deletions(-)
>
> diff --git a/Documentation/gpu/nouveau.rst b/Documentation/gpu/nouveau.rst
> index 0f34131ccc27..b8c801e0068c 100644
> --- a/Documentation/gpu/nouveau.rst
> +++ b/Documentation/gpu/nouveau.rst
> @@ -27,3 +27,6 @@ GSP Support
>  
>  .. kernel-doc:: drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c
>     :doc: GSP message queue element
> +
> +.. kernel-doc:: drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
> +   :doc: GSP message handling policy
> diff --git a/drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h 
> b/drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
> index 746e126c3ecf..e5fe44589bbd 100644
> --- a/drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
> +++ b/drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
> @@ -31,6 +31,25 @@ typedef int (*nvkm_gsp_msg_ntfy_func)(void *priv, u32 fn, 
> void *repv, u32 repc);
>  struct nvkm_gsp_event;
>  typedef void (*nvkm_gsp_event_func)(struct nvkm_gsp_event *, void *repv, u32 
> repc);
>  
> +/**
> + * DOC: GSP message handling policy
> + *
> + * When sending a GSP RPC command, there can be multiple cases of handling
> + * the GSP RPC messages, which are the reply of GSP RPC commands, according
> + * to the requirement of the callers and the nature of the GSP RPC commands.
> + *
> + * NVKM_GSP_RPC_REPLY_NOWAIT - If specified, immediately return to the
> + * caller after the GSP RPC command is issued.
> + *
> + * NVKM_GSP_RPC_REPLY_RECV - If specified, wait and receive the entire GSP
> + * RPC message after the GSP RPC command is issued.
> + *
> + */
> +enum nvkm_gsp_rpc_reply_policy {
> +     NVKM_GSP_RPC_REPLY_NOWAIT = 0,
> +     NVKM_GSP_RPC_REPLY_RECV,
> +};
> +
>  struct nvkm_gsp {
>       const struct nvkm_gsp_func *func;
>       struct nvkm_subdev subdev;
> @@ -188,7 +207,8 @@ struct nvkm_gsp {
>  
>       const struct nvkm_gsp_rm {
>               void *(*rpc_get)(struct nvkm_gsp *, u32 fn, u32 argc);
> -             void *(*rpc_push)(struct nvkm_gsp *, void *argv, bool wait, u32 
> repc);
> +             void *(*rpc_push)(struct nvkm_gsp *gsp, void *argv,
> +                               enum nvkm_gsp_rpc_reply_policy policy, u32 
> repc);
>               void (*rpc_done)(struct nvkm_gsp *gsp, void *repv);
>  
>               void *(*rm_ctrl_get)(struct nvkm_gsp_object *, u32 cmd, u32 
> argc);
> @@ -255,9 +275,10 @@ nvkm_gsp_rpc_get(struct nvkm_gsp *gsp, u32 fn, u32 argc)
>  }
>  
>  static inline void *
> -nvkm_gsp_rpc_push(struct nvkm_gsp *gsp, void *argv, bool wait, u32 repc)
> +nvkm_gsp_rpc_push(struct nvkm_gsp *gsp, void *argv,
> +               enum nvkm_gsp_rpc_reply_policy policy, u32 repc)
>  {
> -     return gsp->rm->rpc_push(gsp, argv, wait, repc);
> +     return gsp->rm->rpc_push(gsp, argv, policy, repc);
>  }
>  
>  static inline void *
> @@ -268,13 +289,14 @@ nvkm_gsp_rpc_rd(struct nvkm_gsp *gsp, u32 fn, u32 argc)
>       if (IS_ERR_OR_NULL(argv))
>               return argv;
>  
> -     return nvkm_gsp_rpc_push(gsp, argv, true, argc);
> +     return nvkm_gsp_rpc_push(gsp, argv, NVKM_GSP_RPC_REPLY_RECV, argc);
>  }
>  
>  static inline int
> -nvkm_gsp_rpc_wr(struct nvkm_gsp *gsp, void *argv, bool wait)
> +nvkm_gsp_rpc_wr(struct nvkm_gsp *gsp, void *argv,
> +             enum nvkm_gsp_rpc_reply_policy policy)
>  {
> -     void *repv = nvkm_gsp_rpc_push(gsp, argv, wait, 0);
> +     void *repv = nvkm_gsp_rpc_push(gsp, argv, policy, 0);
>  
>       if (IS_ERR(repv))
>               return PTR_ERR(repv);
> diff --git a/drivers/gpu/drm/nouveau/nvkm/subdev/bar/r535.c 
> b/drivers/gpu/drm/nouveau/nvkm/subdev/bar/r535.c
> index 3a30bea30e36..90186f98065c 100644
> --- a/drivers/gpu/drm/nouveau/nvkm/subdev/bar/r535.c
> +++ b/drivers/gpu/drm/nouveau/nvkm/subdev/bar/r535.c
> @@ -56,7 +56,7 @@ r535_bar_bar2_update_pde(struct nvkm_gsp *gsp, u64 addr)
>       rpc->info.entryValue = addr ? ((addr >> 4) | 2) : 0; /* PD3 entry 
> format! */
>       rpc->info.entryLevelShift = 47; //XXX: probably fetch this from mmu!
>  
> -     return nvkm_gsp_rpc_wr(gsp, rpc, true);
> +     return nvkm_gsp_rpc_wr(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV);
>  }
>  
>  static void
> diff --git a/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c 
> b/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c
> index db2602e88006..f73dcc3e1c0d 100644
> --- a/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c
> +++ b/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c
> @@ -585,13 +585,34 @@ r535_gsp_rpc_poll(struct nvkm_gsp *gsp, u32 fn)
>  }
>  
>  static void *
> -r535_gsp_rpc_send(struct nvkm_gsp *gsp, void *payload, bool wait,
> -               u32 gsp_rpc_len)
> +r535_gsp_rpc_handle_reply(struct nvkm_gsp *gsp, u32 fn,
> +                       enum nvkm_gsp_rpc_reply_policy policy,
> +                       u32 gsp_rpc_len)
> +{
> +     struct nvfw_gsp_rpc *reply;
> +     void *repv = NULL;
> +
> +     switch (policy) {
> +     case NVKM_GSP_RPC_REPLY_NOWAIT:
> +             break;
> +     case NVKM_GSP_RPC_REPLY_RECV:
> +             reply = r535_gsp_msg_recv(gsp, fn, gsp_rpc_len);
> +             if (!IS_ERR_OR_NULL(reply))
> +                     repv = reply->data;
> +             else
> +                     repv = reply;
> +             break;
> +     }
> +
> +     return repv;
> +}
> +
> +static void *
> +r535_gsp_rpc_send(struct nvkm_gsp *gsp, void *payload,
> +               enum nvkm_gsp_rpc_reply_policy policy, u32 gsp_rpc_len)
>  {
>       struct nvfw_gsp_rpc *rpc = to_gsp_hdr(payload, rpc);
> -     struct nvfw_gsp_rpc *msg;
>       u32 fn = rpc->function;
> -     void *repv = NULL;
>       int ret;
>  
>       if (gsp->subdev.debug >= NV_DBG_TRACE) {
> @@ -605,15 +626,7 @@ r535_gsp_rpc_send(struct nvkm_gsp *gsp, void *payload, 
> bool wait,
>       if (ret)
>               return ERR_PTR(ret);
>  
> -     if (wait) {
> -             msg = r535_gsp_msg_recv(gsp, fn, gsp_rpc_len);
> -             if (!IS_ERR_OR_NULL(msg))
> -                     repv = msg->data;
> -             else
> -                     repv = msg;
> -     }
> -
> -     return repv;
> +     return r535_gsp_rpc_handle_reply(gsp, fn, policy, gsp_rpc_len);
>  }
>  
>  static void
> @@ -797,7 +810,7 @@ r535_gsp_rpc_rm_free(struct nvkm_gsp_object *object)
>       rpc->params.hRoot = client->object.handle;
>       rpc->params.hObjectParent = 0;
>       rpc->params.hObjectOld = object->handle;
> -     return nvkm_gsp_rpc_wr(gsp, rpc, true);
> +     return nvkm_gsp_rpc_wr(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV);
>  }
>  
>  static void
> @@ -815,7 +828,7 @@ r535_gsp_rpc_rm_alloc_push(struct nvkm_gsp_object 
> *object, void *params)
>       struct nvkm_gsp *gsp = object->client->gsp;
>       void *ret = NULL;
>  
> -     rpc = nvkm_gsp_rpc_push(gsp, rpc, true, sizeof(*rpc));
> +     rpc = nvkm_gsp_rpc_push(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV, 
> sizeof(*rpc));
>       if (IS_ERR_OR_NULL(rpc))
>               return rpc;
>  
> @@ -876,7 +889,7 @@ r535_gsp_rpc_rm_ctrl_push(struct nvkm_gsp_object *object, 
> void **params, u32 rep
>       struct nvkm_gsp *gsp = object->client->gsp;
>       int ret = 0;
>  
> -     rpc = nvkm_gsp_rpc_push(gsp, rpc, true, repc);
> +     rpc = nvkm_gsp_rpc_push(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV, repc);
>       if (IS_ERR_OR_NULL(rpc)) {
>               *params = NULL;
>               return PTR_ERR(rpc);
> @@ -948,8 +961,8 @@ r535_gsp_rpc_get(struct nvkm_gsp *gsp, u32 fn, u32 
> payload_size)
>  }
>  
>  static void *
> -r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, bool wait,
> -               u32 gsp_rpc_len)
> +r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload,
> +               enum nvkm_gsp_rpc_reply_policy policy, u32 gsp_rpc_len)
>  {
>       struct nvfw_gsp_rpc *rpc = to_gsp_hdr(payload, rpc);
>       struct r535_gsp_msg *msg = to_gsp_hdr(rpc, msg);
> @@ -967,7 +980,7 @@ r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, 
> bool wait,
>               rpc->length = sizeof(*rpc) + max_payload_size;
>               msg->checksum = rpc->length;
>  
> -             repv = r535_gsp_rpc_send(gsp, payload, false, 0);
> +             repv = r535_gsp_rpc_send(gsp, payload, 
> NVKM_GSP_RPC_REPLY_NOWAIT, 0);
>               if (IS_ERR(repv))
>                       goto done;
>  
> @@ -988,7 +1001,7 @@ r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, 
> bool wait,
>  
>                       memcpy(next, payload, size);
>  
> -                     repv = r535_gsp_rpc_send(gsp, next, false, 0);
> +                     repv = r535_gsp_rpc_send(gsp, next, 
> NVKM_GSP_RPC_REPLY_NOWAIT, 0);
>                       if (IS_ERR(repv))
>                               goto done;
>  
> @@ -997,20 +1010,10 @@ r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, 
> bool wait,
>               }
>  
>               /* Wait for reply. */
> -             rpc = r535_gsp_msg_recv(gsp, fn, payload_size +
> -                                     sizeof(*rpc));
> -             if (!IS_ERR_OR_NULL(rpc)) {
> -                     if (wait) {
> -                             repv = rpc->data;
> -                     } else {
> -                             nvkm_gsp_rpc_done(gsp, rpc);
> -                             repv = NULL;
> -                     }
> -             } else {
> -                     repv = wait ? rpc : NULL;
> -             }

This bit of code seems to have a slightly different flow from what
r535_gsp_rpc_handle_reply() does - it receives the response before
checking for the wait flag, while rpc_handle_reply() does things in the
opposite order. The new code also drops the call to nvkm_gsp_rpc_done().
Can you double-check and confirm this is ok?

Reply via email to