Hi Zhi, Thanks, this patch has become super easy to review.
On Thu Feb 27, 2025 at 10:35 AM JST, Zhi Wang wrote: > There can be multiple cases of handling the GSP RPC messages, which are > the reply of GSP RPC commands according to the requirement of the > callers and the nature of the GSP RPC commands. > > The current supported reply policies are "callers don't care" and "receive > the entire message" according to the requirement of the callers. To > introduce a new policy, factor out the current RPC command reply polices. > Also, centralize the handling of the reply in a single function. > > Factor out NVKM_GSP_RPC_REPLY_NOWAIT as "callers don't care" and > NVKM_GSP_RPC_REPLY_RECV as "receive the entire message". Introduce a > kernel doc to document the policies. Factor out > r535_gsp_rpc_handle_reply(). > > No functional change is intended for small GSP RPC commands. For large GSP > commands, the caller decides the policy of how to handle the returned GSP > RPC message. > > Cc: Ben Skeggs <bske...@nvidia.com> > Cc: Alexandre Courbot <acour...@nvidia.com> > Signed-off-by: Zhi Wang <z...@nvidia.com> > --- > Documentation/gpu/nouveau.rst | 3 + > .../gpu/drm/nouveau/include/nvkm/subdev/gsp.h | 34 +++++++-- > .../gpu/drm/nouveau/nvkm/subdev/bar/r535.c | 2 +- > .../gpu/drm/nouveau/nvkm/subdev/gsp/r535.c | 75 ++++++++++--------- > .../drm/nouveau/nvkm/subdev/instmem/r535.c | 2 +- > 5 files changed, 72 insertions(+), 44 deletions(-) > > diff --git a/Documentation/gpu/nouveau.rst b/Documentation/gpu/nouveau.rst > index 0f34131ccc27..b8c801e0068c 100644 > --- a/Documentation/gpu/nouveau.rst > +++ b/Documentation/gpu/nouveau.rst > @@ -27,3 +27,6 @@ GSP Support > > .. kernel-doc:: drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c > :doc: GSP message queue element > + > +.. kernel-doc:: drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h > + :doc: GSP message handling policy > diff --git a/drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h > b/drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h > index 746e126c3ecf..e5fe44589bbd 100644 > --- a/drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h > +++ b/drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h > @@ -31,6 +31,25 @@ typedef int (*nvkm_gsp_msg_ntfy_func)(void *priv, u32 fn, > void *repv, u32 repc); > struct nvkm_gsp_event; > typedef void (*nvkm_gsp_event_func)(struct nvkm_gsp_event *, void *repv, u32 > repc); > > +/** > + * DOC: GSP message handling policy > + * > + * When sending a GSP RPC command, there can be multiple cases of handling > + * the GSP RPC messages, which are the reply of GSP RPC commands, according > + * to the requirement of the callers and the nature of the GSP RPC commands. > + * > + * NVKM_GSP_RPC_REPLY_NOWAIT - If specified, immediately return to the > + * caller after the GSP RPC command is issued. > + * > + * NVKM_GSP_RPC_REPLY_RECV - If specified, wait and receive the entire GSP > + * RPC message after the GSP RPC command is issued. > + * > + */ > +enum nvkm_gsp_rpc_reply_policy { > + NVKM_GSP_RPC_REPLY_NOWAIT = 0, > + NVKM_GSP_RPC_REPLY_RECV, > +}; > + > struct nvkm_gsp { > const struct nvkm_gsp_func *func; > struct nvkm_subdev subdev; > @@ -188,7 +207,8 @@ struct nvkm_gsp { > > const struct nvkm_gsp_rm { > void *(*rpc_get)(struct nvkm_gsp *, u32 fn, u32 argc); > - void *(*rpc_push)(struct nvkm_gsp *, void *argv, bool wait, u32 > repc); > + void *(*rpc_push)(struct nvkm_gsp *gsp, void *argv, > + enum nvkm_gsp_rpc_reply_policy policy, u32 > repc); > void (*rpc_done)(struct nvkm_gsp *gsp, void *repv); > > void *(*rm_ctrl_get)(struct nvkm_gsp_object *, u32 cmd, u32 > argc); > @@ -255,9 +275,10 @@ nvkm_gsp_rpc_get(struct nvkm_gsp *gsp, u32 fn, u32 argc) > } > > static inline void * > -nvkm_gsp_rpc_push(struct nvkm_gsp *gsp, void *argv, bool wait, u32 repc) > +nvkm_gsp_rpc_push(struct nvkm_gsp *gsp, void *argv, > + enum nvkm_gsp_rpc_reply_policy policy, u32 repc) > { > - return gsp->rm->rpc_push(gsp, argv, wait, repc); > + return gsp->rm->rpc_push(gsp, argv, policy, repc); > } > > static inline void * > @@ -268,13 +289,14 @@ nvkm_gsp_rpc_rd(struct nvkm_gsp *gsp, u32 fn, u32 argc) > if (IS_ERR_OR_NULL(argv)) > return argv; > > - return nvkm_gsp_rpc_push(gsp, argv, true, argc); > + return nvkm_gsp_rpc_push(gsp, argv, NVKM_GSP_RPC_REPLY_RECV, argc); > } > > static inline int > -nvkm_gsp_rpc_wr(struct nvkm_gsp *gsp, void *argv, bool wait) > +nvkm_gsp_rpc_wr(struct nvkm_gsp *gsp, void *argv, > + enum nvkm_gsp_rpc_reply_policy policy) > { > - void *repv = nvkm_gsp_rpc_push(gsp, argv, wait, 0); > + void *repv = nvkm_gsp_rpc_push(gsp, argv, policy, 0); > > if (IS_ERR(repv)) > return PTR_ERR(repv); > diff --git a/drivers/gpu/drm/nouveau/nvkm/subdev/bar/r535.c > b/drivers/gpu/drm/nouveau/nvkm/subdev/bar/r535.c > index 3a30bea30e36..90186f98065c 100644 > --- a/drivers/gpu/drm/nouveau/nvkm/subdev/bar/r535.c > +++ b/drivers/gpu/drm/nouveau/nvkm/subdev/bar/r535.c > @@ -56,7 +56,7 @@ r535_bar_bar2_update_pde(struct nvkm_gsp *gsp, u64 addr) > rpc->info.entryValue = addr ? ((addr >> 4) | 2) : 0; /* PD3 entry > format! */ > rpc->info.entryLevelShift = 47; //XXX: probably fetch this from mmu! > > - return nvkm_gsp_rpc_wr(gsp, rpc, true); > + return nvkm_gsp_rpc_wr(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV); > } > > static void > diff --git a/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c > b/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c > index db2602e88006..f73dcc3e1c0d 100644 > --- a/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c > +++ b/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/r535.c > @@ -585,13 +585,34 @@ r535_gsp_rpc_poll(struct nvkm_gsp *gsp, u32 fn) > } > > static void * > -r535_gsp_rpc_send(struct nvkm_gsp *gsp, void *payload, bool wait, > - u32 gsp_rpc_len) > +r535_gsp_rpc_handle_reply(struct nvkm_gsp *gsp, u32 fn, > + enum nvkm_gsp_rpc_reply_policy policy, > + u32 gsp_rpc_len) > +{ > + struct nvfw_gsp_rpc *reply; > + void *repv = NULL; > + > + switch (policy) { > + case NVKM_GSP_RPC_REPLY_NOWAIT: > + break; > + case NVKM_GSP_RPC_REPLY_RECV: > + reply = r535_gsp_msg_recv(gsp, fn, gsp_rpc_len); > + if (!IS_ERR_OR_NULL(reply)) > + repv = reply->data; > + else > + repv = reply; > + break; > + } > + > + return repv; > +} > + > +static void * > +r535_gsp_rpc_send(struct nvkm_gsp *gsp, void *payload, > + enum nvkm_gsp_rpc_reply_policy policy, u32 gsp_rpc_len) > { > struct nvfw_gsp_rpc *rpc = to_gsp_hdr(payload, rpc); > - struct nvfw_gsp_rpc *msg; > u32 fn = rpc->function; > - void *repv = NULL; > int ret; > > if (gsp->subdev.debug >= NV_DBG_TRACE) { > @@ -605,15 +626,7 @@ r535_gsp_rpc_send(struct nvkm_gsp *gsp, void *payload, > bool wait, > if (ret) > return ERR_PTR(ret); > > - if (wait) { > - msg = r535_gsp_msg_recv(gsp, fn, gsp_rpc_len); > - if (!IS_ERR_OR_NULL(msg)) > - repv = msg->data; > - else > - repv = msg; > - } > - > - return repv; > + return r535_gsp_rpc_handle_reply(gsp, fn, policy, gsp_rpc_len); > } > > static void > @@ -797,7 +810,7 @@ r535_gsp_rpc_rm_free(struct nvkm_gsp_object *object) > rpc->params.hRoot = client->object.handle; > rpc->params.hObjectParent = 0; > rpc->params.hObjectOld = object->handle; > - return nvkm_gsp_rpc_wr(gsp, rpc, true); > + return nvkm_gsp_rpc_wr(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV); > } > > static void > @@ -815,7 +828,7 @@ r535_gsp_rpc_rm_alloc_push(struct nvkm_gsp_object > *object, void *params) > struct nvkm_gsp *gsp = object->client->gsp; > void *ret = NULL; > > - rpc = nvkm_gsp_rpc_push(gsp, rpc, true, sizeof(*rpc)); > + rpc = nvkm_gsp_rpc_push(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV, > sizeof(*rpc)); > if (IS_ERR_OR_NULL(rpc)) > return rpc; > > @@ -876,7 +889,7 @@ r535_gsp_rpc_rm_ctrl_push(struct nvkm_gsp_object *object, > void **params, u32 rep > struct nvkm_gsp *gsp = object->client->gsp; > int ret = 0; > > - rpc = nvkm_gsp_rpc_push(gsp, rpc, true, repc); > + rpc = nvkm_gsp_rpc_push(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV, repc); > if (IS_ERR_OR_NULL(rpc)) { > *params = NULL; > return PTR_ERR(rpc); > @@ -948,8 +961,8 @@ r535_gsp_rpc_get(struct nvkm_gsp *gsp, u32 fn, u32 > payload_size) > } > > static void * > -r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, bool wait, > - u32 gsp_rpc_len) > +r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, > + enum nvkm_gsp_rpc_reply_policy policy, u32 gsp_rpc_len) > { > struct nvfw_gsp_rpc *rpc = to_gsp_hdr(payload, rpc); > struct r535_gsp_msg *msg = to_gsp_hdr(rpc, msg); > @@ -967,7 +980,7 @@ r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, > bool wait, > rpc->length = sizeof(*rpc) + max_payload_size; > msg->checksum = rpc->length; > > - repv = r535_gsp_rpc_send(gsp, payload, false, 0); > + repv = r535_gsp_rpc_send(gsp, payload, > NVKM_GSP_RPC_REPLY_NOWAIT, 0); > if (IS_ERR(repv)) > goto done; > > @@ -988,7 +1001,7 @@ r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, > bool wait, > > memcpy(next, payload, size); > > - repv = r535_gsp_rpc_send(gsp, next, false, 0); > + repv = r535_gsp_rpc_send(gsp, next, > NVKM_GSP_RPC_REPLY_NOWAIT, 0); > if (IS_ERR(repv)) > goto done; > > @@ -997,20 +1010,10 @@ r535_gsp_rpc_push(struct nvkm_gsp *gsp, void *payload, > bool wait, > } > > /* Wait for reply. */ > - rpc = r535_gsp_msg_recv(gsp, fn, payload_size + > - sizeof(*rpc)); > - if (!IS_ERR_OR_NULL(rpc)) { > - if (wait) { > - repv = rpc->data; > - } else { > - nvkm_gsp_rpc_done(gsp, rpc); > - repv = NULL; > - } > - } else { > - repv = wait ? rpc : NULL; > - } This bit of code seems to have a slightly different flow from what r535_gsp_rpc_handle_reply() does - it receives the response before checking for the wait flag, while rpc_handle_reply() does things in the opposite order. The new code also drops the call to nvkm_gsp_rpc_done(). Can you double-check and confirm this is ok?