Apply a consistent style for comments in the lnet selftest
code.

Signed-off-by: James Simmons <jsimm...@infradead.org>
---
 drivers/staging/lustre/lnet/selftest/brw_test.c  |    8 ++--
 drivers/staging/lustre/lnet/selftest/conctl.c    |   50 +++++++++++-----------
 drivers/staging/lustre/lnet/selftest/conrpc.c    |   23 +++++-----
 drivers/staging/lustre/lnet/selftest/console.c   |   11 +++--
 drivers/staging/lustre/lnet/selftest/framework.c |   20 ++++----
 drivers/staging/lustre/lnet/selftest/ping_test.c |    2 +-
 drivers/staging/lustre/lnet/selftest/rpc.c       |   46 ++++++++++----------
 drivers/staging/lustre/lnet/selftest/rpc.h       |    2 +-
 drivers/staging/lustre/lnet/selftest/selftest.h  |    3 +-
 drivers/staging/lustre/lnet/selftest/timer.c     |    6 +-
 10 files changed, 87 insertions(+), 84 deletions(-)

diff --git a/drivers/staging/lustre/lnet/selftest/brw_test.c 
b/drivers/staging/lustre/lnet/selftest/brw_test.c
index eebc924..6ac4d02 100644
--- a/drivers/staging/lustre/lnet/selftest/brw_test.c
+++ b/drivers/staging/lustre/lnet/selftest/brw_test.c
@@ -86,7 +86,7 @@ brw_client_init(sfw_test_instance_t *tsi)
                opc = breq->blk_opc;
                flags = breq->blk_flags;
                npg = breq->blk_npg;
-               /*
+               /**
                 * NB: this is not going to work for variable page size,
                 * but we have to keep it for compatibility
                 */
@@ -95,7 +95,7 @@ brw_client_init(sfw_test_instance_t *tsi)
        } else {
                test_bulk_req_v1_t *breq = &tsi->tsi_u.bulk_v1;
 
-               /*
+               /**
                 * I should never get this step if it's unknown feature
                 * because make_session will reject unknown feature
                 */
@@ -283,7 +283,7 @@ brw_client_prep_rpc(sfw_test_unit_t *tsu,
        } else {
                test_bulk_req_v1_t *breq = &tsi->tsi_u.bulk_v1;
 
-               /*
+               /**
                 * I should never get this step if it's unknown feature
                 * because make_session will reject unknown feature
                 */
@@ -329,7 +329,7 @@ brw_client_done_rpc(sfw_test_unit_t *tsu, srpc_client_rpc_t 
*rpc)
        if (rpc->crpc_status) {
                CERROR("BRW RPC to %s failed with %d\n",
                       libcfs_id2str(rpc->crpc_dest), rpc->crpc_status);
-               if (!tsi->tsi_stopping) /* rpc could have been aborted */
+               if (!tsi->tsi_stopping) /* rpc could have been aborted */
                        atomic_inc(&sn->sn_brw_errors);
                return;
        }
diff --git a/drivers/staging/lustre/lnet/selftest/conctl.c 
b/drivers/staging/lustre/lnet/selftest/conctl.c
index 872df72..d045ac5 100644
--- a/drivers/staging/lustre/lnet/selftest/conctl.c
+++ b/drivers/staging/lustre/lnet/selftest/conctl.c
@@ -51,9 +51,9 @@ lst_session_new_ioctl(lstio_session_new_args_t *args)
        char *name;
        int rc;
 
-       if (!args->lstio_ses_idp || /* address for output sid */
-           !args->lstio_ses_key ||    /* no key is specified */
-           !args->lstio_ses_namep || /* session name */
+       if (!args->lstio_ses_idp ||     /* address for output sid */
+           !args->lstio_ses_key ||     /* no key is specified */
+           !args->lstio_ses_namep ||   /* session name */
            args->lstio_ses_nmlen <= 0 ||
            args->lstio_ses_nmlen > LST_NAME_SIZE)
                return -EINVAL;
@@ -95,11 +95,11 @@ lst_session_info_ioctl(lstio_session_info_args_t *args)
 {
        /* no checking of key */
 
-       if (!args->lstio_ses_idp || /* address for output sid */
-           !args->lstio_ses_keyp || /* address for output key */
-           !args->lstio_ses_featp || /* address for output features */
-           !args->lstio_ses_ndinfo || /* address for output ndinfo */
-           !args->lstio_ses_namep || /* address for output name */
+       if (!args->lstio_ses_idp ||     /* address for output sid */
+           !args->lstio_ses_keyp ||    /* address for output key */
+           !args->lstio_ses_featp ||   /* address for output features */
+           !args->lstio_ses_ndinfo ||  /* address for output ndinfo */
+           !args->lstio_ses_namep ||   /* address for output name */
            args->lstio_ses_nmlen <= 0 ||
            args->lstio_ses_nmlen > LST_NAME_SIZE)
                return -EINVAL;
@@ -125,7 +125,7 @@ lst_debug_ioctl(lstio_debug_args_t *args)
        if (!args->lstio_dbg_resultp)
                return -EINVAL;
 
-       if (args->lstio_dbg_namep && /* name of batch/group */
+       if (args->lstio_dbg_namep &&    /* name of batch/group */
            (args->lstio_dbg_nmlen <= 0 ||
             args->lstio_dbg_nmlen > LST_NAME_SIZE))
                return -EINVAL;
@@ -327,7 +327,7 @@ lst_nodes_add_ioctl(lstio_group_nodes_args_t *args)
        if (args->lstio_grp_key != console_session.ses_key)
                return -EACCES;
 
-       if (!args->lstio_grp_idsp || /* array of ids */
+       if (!args->lstio_grp_idsp ||    /* array of ids */
            args->lstio_grp_count <= 0 ||
            !args->lstio_grp_resultp ||
            !args->lstio_grp_featp ||
@@ -395,13 +395,13 @@ lst_group_info_ioctl(lstio_group_info_args_t *args)
            args->lstio_grp_nmlen > LST_NAME_SIZE)
                return -EINVAL;
 
-       if (!args->lstio_grp_entp &&  /* output: group entry */
-           !args->lstio_grp_dentsp)  /* output: node entry */
+       if (!args->lstio_grp_entp &&    /* output: group entry */
+           !args->lstio_grp_dentsp)    /* output: node entry */
                return -EINVAL;
 
-       if (args->lstio_grp_dentsp) { /* have node entry */
-               if (!args->lstio_grp_idxp || /* node index */
-                   !args->lstio_grp_ndentp) /* # of node entry */
+       if (args->lstio_grp_dentsp) {           /* have node entry */
+               if (!args->lstio_grp_idxp ||    /* node index */
+                   !args->lstio_grp_ndentp)    /* # of node entry */
                        return -EINVAL;
 
                if (copy_from_user(&ndent, args->lstio_grp_ndentp,
@@ -613,18 +613,18 @@ lst_batch_info_ioctl(lstio_batch_info_args_t *args)
        if (args->lstio_bat_key != console_session.ses_key)
                return -EACCES;
 
-       if (!args->lstio_bat_namep || /* batch name */
+       if (!args->lstio_bat_namep ||   /* batch name */
            args->lstio_bat_nmlen <= 0 ||
            args->lstio_bat_nmlen > LST_NAME_SIZE)
                return -EINVAL;
 
-       if (!args->lstio_bat_entp && /* output: batch entry */
-           !args->lstio_bat_dentsp) /* output: node entry */
+       if (!args->lstio_bat_entp &&    /* output: batch entry */
+           !args->lstio_bat_dentsp)    /* output: node entry */
                return -EINVAL;
 
-       if (args->lstio_bat_dentsp) { /* have node entry */
-               if (!args->lstio_bat_idxp || /* node index */
-                   !args->lstio_bat_ndentp) /* # of node entry */
+       if (args->lstio_bat_dentsp) {           /* have node entry */
+               if (!args->lstio_bat_idxp ||    /* node index */
+                   !args->lstio_bat_ndentp)    /* # of node entry */
                        return -EINVAL;
 
                if (copy_from_user(&index, args->lstio_bat_idxp,
@@ -723,18 +723,18 @@ static int lst_test_add_ioctl(lstio_test_args_t *args)
 
        if (!args->lstio_tes_resultp ||
            !args->lstio_tes_retp ||
-           !args->lstio_tes_bat_name || /* no specified batch */
+           !args->lstio_tes_bat_name ||        /* no specified batch */
            args->lstio_tes_bat_nmlen <= 0 ||
            args->lstio_tes_bat_nmlen > LST_NAME_SIZE ||
-           !args->lstio_tes_sgrp_name || /* no source group */
+           !args->lstio_tes_sgrp_name ||       /* no source group */
            args->lstio_tes_sgrp_nmlen <= 0 ||
            args->lstio_tes_sgrp_nmlen > LST_NAME_SIZE ||
-           !args->lstio_tes_dgrp_name || /* no target group */
+           !args->lstio_tes_dgrp_name ||       /* no target group */
            args->lstio_tes_dgrp_nmlen <= 0 ||
            args->lstio_tes_dgrp_nmlen > LST_NAME_SIZE)
                return -EINVAL;
 
-       if (!args->lstio_tes_loop || /* negative is infinite */
+       if (!args->lstio_tes_loop ||            /* negative is infinite */
            args->lstio_tes_concur <= 0 ||
            args->lstio_tes_dist <= 0 ||
            args->lstio_tes_span <= 0)
diff --git a/drivers/staging/lustre/lnet/selftest/conrpc.c 
b/drivers/staging/lustre/lnet/selftest/conrpc.c
index cdb660e..9401e1a 100644
--- a/drivers/staging/lustre/lnet/selftest/conrpc.c
+++ b/drivers/staging/lustre/lnet/selftest/conrpc.c
@@ -60,7 +60,7 @@ lstcon_rpc_done(srpc_client_rpc_t *rpc)
        spin_lock(&rpc->crpc_lock);
 
        if (!crpc->crp_trans) {
-               /*
+               /**
                 * Orphan RPC is not in any transaction,
                 * I'm just a poor body and nobody loves me
                 */
@@ -242,7 +242,7 @@ lstcon_rpc_trans_prep(struct list_head *translist,
 
        if (translist) {
                list_for_each_entry(trans, translist, tas_link) {
-                       /*
+                       /**
                         * Can't enqueue two private transaction on
                         * the same object
                         */
@@ -296,8 +296,8 @@ lstcon_rpc_trans_abort(lstcon_rpc_trans_t *trans, int error)
 
                spin_lock(&rpc->crpc_lock);
 
-               if (!crpc->crp_posted || /* not posted */
-                   crpc->crp_stamp) { /* rpc done or aborted already */
+               if (!crpc->crp_posted ||        /* not posted */
+                   crpc->crp_stamp) {          /* rpc done or aborted already 
*/
                        if (!crpc->crp_stamp) {
                                crpc->crp_stamp = cfs_time_current();
                                crpc->crp_status = -EINTR;
@@ -563,11 +563,11 @@ lstcon_rpc_trans_destroy(lstcon_rpc_trans_t *trans)
                        continue;
                }
 
-               /*
-                * rpcs can be still not callbacked (even LNetMDUnlink is 
called)
-                * because huge timeout for inaccessible network, don't make
-                * user wait for them, just abandon them, they will be recycled
-                * in callback
+               /**
+                * rpcs can be still not callbacked (even LNetMDUnlink is
+                * called) because huge timeout for inaccessible network,
+                * don't make user wait for them, just abandon them, they
+                * will be recycled in callback
                 */
                LASSERT(crpc->crp_status);
 
@@ -941,7 +941,7 @@ lstcon_sesnew_stat_reply(lstcon_rpc_trans_t *trans,
 
        if (!trans->tas_feats_updated) {
                spin_lock(&console_session.ses_rpc_lock);
-               if (!trans->tas_feats_updated) { /* recheck with lock */
+               if (!trans->tas_feats_updated) {        /* recheck with lock */
                        trans->tas_feats_updated = 1;
                        trans->tas_features = reply->msg_ses_feats;
                }
@@ -1181,7 +1181,8 @@ lstcon_rpc_pinger(void *arg)
        int count = 0;
        int rc;
 
-       /* RPC pinger is a special case of transaction,
+       /**
+        * RPC pinger is a special case of transaction,
         * it's called by timer at 8 seconds interval.
         */
        mutex_lock(&console_session.ses_mutex);
diff --git a/drivers/staging/lustre/lnet/selftest/console.c 
b/drivers/staging/lustre/lnet/selftest/console.c
index 6017f21..25917ac 100644
--- a/drivers/staging/lustre/lnet/selftest/console.c
+++ b/drivers/staging/lustre/lnet/selftest/console.c
@@ -104,7 +104,7 @@ lstcon_node_find(lnet_process_id_t id, lstcon_node_t 
**ndpp, int create)
        ndl->ndl_node->nd_timeout = 0;
        memset(&ndl->ndl_node->nd_ping, 0, sizeof(lstcon_rpc_t));
 
-       /*
+       /**
         * queued in global hash & list, no refcount is taken by
         * global hash & list, if caller release his refcount,
         * node will be released
@@ -276,7 +276,7 @@ lstcon_group_find(const char *name, lstcon_group_t **grpp)
                if (strncmp(grp->grp_name, name, LST_NAME_SIZE))
                        continue;
 
-               lstcon_group_addref(grp);  /* +1 ref for caller */
+               lstcon_group_addref(grp); /* +1 ref for caller */
                *grpp = grp;
                return 0;
        }
@@ -608,7 +608,7 @@ lstcon_group_del(char *name)
        lstcon_rpc_trans_destroy(trans);
 
        lstcon_group_decref(grp);
-       /*
+       /**
         * -ref for session, it's destroyed,
         * status can't be rolled back, destroy group anyway
         */
@@ -1289,7 +1289,7 @@ lstcon_test_add(char *batch_name, int type, int loop,
        lstcon_group_t *dst_grp = NULL;
        lstcon_batch_t *batch = NULL;
 
-       /*
+       /**
         * verify that a batch of the given name exists, and the groups
         * that will be part of the batch exist and have at least one
         * active node
@@ -1447,7 +1447,8 @@ lstcon_test_batch_query(char *name, int testidx, int 
client,
 
        lstcon_rpc_trans_postwait(trans, timeout);
 
-       if (!testidx && /* query a batch, not a test */
+       /* query a batch, not a test */
+       if (!testidx &&
            !lstcon_rpc_stat_failure(lstcon_trans_stat(), 0) &&
            !lstcon_tsbqry_stat_run(lstcon_trans_stat(), 0)) {
                /* all RPCs finished, and no active test */
diff --git a/drivers/staging/lustre/lnet/selftest/framework.c 
b/drivers/staging/lustre/lnet/selftest/framework.c
index 2ff47d2..ed2f7e9 100644
--- a/drivers/staging/lustre/lnet/selftest/framework.c
+++ b/drivers/staging/lustre/lnet/selftest/framework.c
@@ -226,7 +226,7 @@ __must_hold(&sfw_data.fw_lock)
        }
 
        if (nactive)
-               return;   /* wait for active batches to stop */
+               return; /* wait for active batches to stop */
 
        list_del_init(&sn->sn_list);
        spin_unlock(&sfw_data.fw_lock);
@@ -382,7 +382,7 @@ sfw_get_stats(srpc_stat_reqst_t *request, srpc_stat_reply_t 
*reply)
        lnet_counters_get(&reply->str_lnet);
        srpc_get_counters(&reply->str_rpc);
 
-       /*
+       /**
         * send over the msecs since the session was started
         * with 32 bits to send, this is ~49 days
         */
@@ -435,7 +435,7 @@ sfw_make_session(srpc_mksn_reqst_t *request, 
srpc_mksn_reply_t *reply)
                }
        }
 
-       /*
+       /**
         * reject the request if it requires unknown features
         * NB: old version will always accept all features because it's not
         * aware of srpc_msg_t::msg_ses_feats, it's a defect but it's also
@@ -576,7 +576,7 @@ sfw_load_test(struct sfw_test_instance *tsi)
        if (rc) {
                CWARN("Failed to reserve enough buffers: service %s, %d needed: 
%d\n",
                      svc->sv_name, nbuf, rc);
-               /*
+               /**
                 * NB: this error handler is not strictly correct, because
                 * it may release more buffers than already allocated,
                 * but it doesn't matter because request portal should
@@ -604,7 +604,7 @@ sfw_unload_test(struct sfw_test_instance *tsi)
        if (tsi->tsi_is_client)
                return;
 
-       /*
+       /**
         * shrink buffers, because request portal is lazy portal
         * which can grow buffers at runtime so we may leave
         * some buffers behind, but never mind...
@@ -693,7 +693,7 @@ sfw_unpack_addtest_req(srpc_msg_t *msg)
        LASSERT(req->tsr_is_client);
 
        if (msg->msg_magic == SRPC_MSG_MAGIC)
-               return; /* no flipping needed */
+               return; /* no flipping needed */
 
        LASSERT(msg->msg_magic == __swab32(SRPC_MSG_MAGIC));
 
@@ -789,7 +789,7 @@ sfw_add_test_instance(sfw_batch_t *tsb, struct 
srpc_server_rpc *rpc)
                int j;
 
                dests = page_address(bk->bk_iovs[i / 
SFW_ID_PER_PAGE].kiov_page);
-               LASSERT(dests);  /* my pages are within KVM always */
+               LASSERT(dests);         /* my pages are within KVM always */
                id = dests[i % SFW_ID_PER_PAGE];
                if (msg->msg_magic != SRPC_MSG_MAGIC)
                        sfw_unpack_id(id);
@@ -844,8 +844,8 @@ sfw_test_unit_done(sfw_test_unit_t *tsu)
 
        spin_lock(&sfw_data.fw_lock);
 
-       if (!atomic_dec_and_test(&tsb->bat_nactive) ||/* tsb still active */
-           sn == sfw_data.fw_session) {                  /* sn also active */
+       if (!atomic_dec_and_test(&tsb->bat_nactive) ||  /* tsb still active */
+           sn == sfw_data.fw_session) {                /* sn also active */
                spin_unlock(&sfw_data.fw_lock);
                return;
        }
@@ -978,7 +978,7 @@ sfw_run_test(swi_workitem_t *wi)
        return 0;
 
 test_done:
-       /*
+       /**
         * No one can schedule me now since:
         * - previous RPC, if any, has done and
         * - no new RPC is initiated.
diff --git a/drivers/staging/lustre/lnet/selftest/ping_test.c 
b/drivers/staging/lustre/lnet/selftest/ping_test.c
index 81a4504..e05acce 100644
--- a/drivers/staging/lustre/lnet/selftest/ping_test.c
+++ b/drivers/staging/lustre/lnet/selftest/ping_test.c
@@ -129,7 +129,7 @@ ping_client_done_rpc(sfw_test_unit_t *tsu, 
srpc_client_rpc_t *rpc)
        LASSERT(sn);
 
        if (rpc->crpc_status) {
-               if (!tsi->tsi_stopping) /* rpc could have been aborted */
+               if (!tsi->tsi_stopping) /* rpc could have been aborted */
                        atomic_inc(&sn->sn_ping_errors);
                CERROR("Unable to ping %s (%d): %d\n",
                       libcfs_id2str(rpc->crpc_dest),
diff --git a/drivers/staging/lustre/lnet/selftest/rpc.c 
b/drivers/staging/lustre/lnet/selftest/rpc.c
index 69be7d6..83216aa 100644
--- a/drivers/staging/lustre/lnet/selftest/rpc.c
+++ b/drivers/staging/lustre/lnet/selftest/rpc.c
@@ -276,7 +276,7 @@ srpc_service_init(struct srpc_service *svc)
                scd->scd_ev.ev_data = scd;
                scd->scd_ev.ev_type = SRPC_REQUEST_RCVD;
 
-               /*
+               /**
                 * NB: don't use lst_sched_serial for adding buffer,
                 * see details in srpc_service_add_buffers()
                 */
@@ -284,7 +284,7 @@ srpc_service_init(struct srpc_service *svc)
                                  srpc_add_buffer, lst_sched_test[i]);
 
                if (i && srpc_serv_is_framework(svc)) {
-                       /*
+                       /**
                         * NB: framework service only needs srpc_service_cd for
                         * one partition, but we allocate for all to make
                         * it easier to implement, it will waste a little
@@ -415,7 +415,7 @@ srpc_post_active_rdma(int portal, __u64 matchbits, void 
*buf, int len,
                return -ENOMEM;
        }
 
-       /*
+       /**
         * this is kind of an abuse of the LNET_MD_OP_{PUT,GET} options.
         * they're only meaningful for MDs attached to an ME (i.e. passive
         * buffers...
@@ -434,7 +434,7 @@ srpc_post_active_rdma(int portal, __u64 matchbits, void 
*buf, int len,
                       options & LNET_MD_OP_PUT ? "Put" : "Get",
                       libcfs_id2str(peer), portal, matchbits, rc);
 
-               /*
+               /**
                 * The forthcoming unlink event will complete this operation
                 * with failure, so fall through and return success here.
                 */
@@ -479,7 +479,7 @@ __must_hold(&scd->scd_lock)
                                      msg, sizeof(*msg), &buf->buf_mdh,
                                      &scd->scd_ev);
 
-       /*
+       /**
         * At this point, a RPC (new or delayed) may have arrived in
         * msg and its event handler has been called. So we must add
         * buf to scd_buf_posted _before_ dropping scd_lock
@@ -491,7 +491,7 @@ __must_hold(&scd->scd_lock)
                        return 0;
 
                spin_unlock(&scd->scd_lock);
-               /*
+               /**
                 * srpc_shutdown_service might have tried to unlink me
                 * when my buf_mdh was still invalid
                 */
@@ -520,7 +520,7 @@ srpc_add_buffer(struct swi_workitem *wi)
        struct srpc_buffer *buf;
        int rc = 0;
 
-       /*
+       /**
         * it's called by workitem scheduler threads, these threads
         * should have been set CPT affinity, so buffers will be posted
         * on CPT local list of Portal
@@ -602,7 +602,7 @@ srpc_service_add_buffers(struct srpc_service *sv, int 
nbuffer)
 
        cfs_percpt_for_each(scd, i, sv->sv_cpt_data) {
                spin_lock(&scd->scd_lock);
-               /*
+               /**
                 * NB: srpc_service_add_buffers() can be called inside
                 * thread context of lst_sched_serial, and we don't normally
                 * allow to sleep inside thread context of WI scheduler
@@ -740,7 +740,7 @@ srpc_abort_service(struct srpc_service *sv)
        cfs_percpt_for_each(scd, i, sv->sv_cpt_data) {
                spin_lock(&scd->scd_lock);
 
-               /*
+               /**
                 * schedule in-flight RPCs to notice the abort, NB:
                 * racing with incoming RPCs; complete fix should make test
                 * RPCs carry session ID in its headers
@@ -782,7 +782,7 @@ srpc_shutdown_service(srpc_service_t *sv)
 
                spin_unlock(&scd->scd_lock);
 
-               /*
+               /**
                 * OK to traverse scd_buf_posted without lock, since no one
                 * touches scd_buf_posted now
                 */
@@ -927,7 +927,7 @@ srpc_server_rpc_done(struct srpc_server_rpc *rpc, int 
status)
        spin_lock(&scd->scd_lock);
 
        if (rpc->srpc_reqstbuf) {
-               /*
+               /**
                 * NB might drop sv_lock in srpc_service_recycle_buffer, but
                 * sv won't go away for scd_rpc_active must not be empty
                 */
@@ -937,7 +937,7 @@ srpc_server_rpc_done(struct srpc_server_rpc *rpc, int 
status)
 
        list_del(&rpc->srpc_list); /* from scd->scd_rpc_active */
 
-       /*
+       /**
         * No one can schedule me now since:
         * - I'm not on scd_rpc_active.
         * - all LNet events have been fired.
@@ -1110,7 +1110,7 @@ srpc_add_client_rpc_timer(srpc_client_rpc_t *rpc)
        stt_add_timer(timer);
 }
 
-/*
+/**
  * Called with rpc->crpc_lock held.
  *
  * Upon exit the RPC expiry timer is not queued and the handler is not
@@ -1157,7 +1157,7 @@ srpc_client_rpc_done(srpc_client_rpc_t *rpc, int status)
                     rpc->crpc_service, libcfs_id2str(rpc->crpc_dest),
                     swi_state2str(wi->swi_state), rpc->crpc_aborted, status);
 
-       /*
+       /**
         * No one can schedule me now since:
         * - RPC timer has been defused.
         * - all LNet events have been fired.
@@ -1222,7 +1222,7 @@ srpc_send_rpc(swi_workitem_t *wi)
                break;
 
        case SWI_STATE_REQUEST_SUBMITTED:
-               /*
+               /**
                 * CAVEAT EMPTOR: rqtev, rpyev, and bulkev may come in any
                 * order; however, they're processed in a strict order:
                 * rqt, rpy, and bulk.
@@ -1273,7 +1273,7 @@ srpc_send_rpc(swi_workitem_t *wi)
 
                rc = do_bulk ? rpc->crpc_bulkev.ev_status : 0;
 
-               /*
+               /**
                 * Bulk buffer was unlinked due to remote error. Clear error
                 * since reply buffer still contains valid data.
                 * NB rpc->crpc_done shouldn't look into bulk data in case of
@@ -1332,8 +1332,8 @@ srpc_abort_rpc(srpc_client_rpc_t *rpc, int why)
 {
        LASSERT(why);
 
-       if (rpc->crpc_aborted || /* already aborted */
-           rpc->crpc_closed)    /* callback imminent */
+       if (rpc->crpc_aborted ||        /* already aborted */
+           rpc->crpc_closed)           /* callback imminent */
                return;
 
        CDEBUG(D_NET, "Aborting RPC: service %d, peer %s, state %s, why %d\n",
@@ -1377,7 +1377,7 @@ srpc_send_reply(struct srpc_server_rpc *rpc)
        spin_lock(&scd->scd_lock);
 
        if (!sv->sv_shuttingdown && !srpc_serv_is_framework(sv)) {
-               /*
+               /**
                 * Repost buffer before replying since test client
                 * might send me another RPC once it gets the reply
                 */
@@ -1401,7 +1401,7 @@ srpc_send_reply(struct srpc_server_rpc *rpc)
                                   rpc->srpc_peer, rpc->srpc_self,
                                   &rpc->srpc_replymdh, ev);
        if (rc)
-               ev->ev_fired = 1;  /* no more event expected */
+               ev->ev_fired = 1; /* no more event expected */
        return rc;
 }
 
@@ -1494,7 +1494,7 @@ srpc_lnet_ev_handler(lnet_event_t *ev)
                scd->scd_buf_nposted--;
 
                if (sv->sv_shuttingdown) {
-                       /*
+                       /**
                         * Leave buffer on scd->scd_buf_nposted since
                         * srpc_finish_service needs to traverse it.
                         */
@@ -1509,7 +1509,7 @@ srpc_lnet_ev_handler(lnet_event_t *ev)
                        scd->scd_buf_err = 0;
                }
 
-               if (!scd->scd_buf_err && /* adding buffer is enabled */
+               if (!scd->scd_buf_err &&        /* adding buffer is enabled */
                    !scd->scd_buf_adjust &&
                    scd->scd_buf_nposted < scd->scd_buf_low) {
                        scd->scd_buf_adjust = max(scd->scd_buf_total / 2,
@@ -1531,7 +1531,7 @@ srpc_lnet_ev_handler(lnet_event_t *ev)
                               ev->status, ev->mlength,
                               msg->msg_type, msg->msg_magic);
 
-                       /*
+                       /**
                         * NB can't call srpc_service_recycle_buffer here since
                         * it may call LNetM[DE]Attach. The invalid magic tells
                         * srpc_handle_rpc to drop this RPC
diff --git a/drivers/staging/lustre/lnet/selftest/rpc.h 
b/drivers/staging/lustre/lnet/selftest/rpc.h
index a79c315..7cbf19d 100644
--- a/drivers/staging/lustre/lnet/selftest/rpc.h
+++ b/drivers/staging/lustre/lnet/selftest/rpc.h
@@ -281,7 +281,7 @@ srpc_unpack_msg_hdr(srpc_msg_t *msg)
        if (msg->msg_magic == SRPC_MSG_MAGIC)
                return; /* no flipping needed */
 
-       /*
+       /**
         * We do not swap the magic number here as it is needed to
         * determine whether the body needs to be swapped.
         */
diff --git a/drivers/staging/lustre/lnet/selftest/selftest.h 
b/drivers/staging/lustre/lnet/selftest/selftest.h
index 288522d..b1460f3 100644
--- a/drivers/staging/lustre/lnet/selftest/selftest.h
+++ b/drivers/staging/lustre/lnet/selftest/selftest.h
@@ -71,7 +71,8 @@ struct srpc_service_cd;
 struct sfw_test_unit;
 struct sfw_test_instance;
 
-/* services below SRPC_FRAMEWORK_SERVICE_MAX_ID are framework
+/**
+ * services below SRPC_FRAMEWORK_SERVICE_MAX_ID are framework
  * services, e.g. create/modify session.
  */
 #define SRPC_SERVICE_DEBUG             0
diff --git a/drivers/staging/lustre/lnet/selftest/timer.c 
b/drivers/staging/lustre/lnet/selftest/timer.c
index 8be5252..3e83442 100644
--- a/drivers/staging/lustre/lnet/selftest/timer.c
+++ b/drivers/staging/lustre/lnet/selftest/timer.c
@@ -42,14 +42,14 @@
 
 #include "selftest.h"
 
-/*
+/**
  * Timers are implemented as a sorted queue of expiry times. The queue
  * is slotted, with each slot holding timers which expire in a
  * 2**STTIMER_MINPOLL (8) second period. The timers in each slot are
  * sorted by increasing expiry time. The number of slots is 2**7 (128),
  * to cover a time period of 1024 seconds into the future before wrapping.
  */
-#define STTIMER_MINPOLL        3   /* log2 min poll interval (8 s) */
+#define STTIMER_MINPOLL        3       /* log2 min poll interval (8 s) */
 #define STTIMER_SLOTTIME       (1 << STTIMER_MINPOLL)
 #define STTIMER_SLOTTIMEMASK   (~(STTIMER_SLOTTIME - 1))
 #define STTIMER_NSLOTS        (1 << 7)
@@ -92,7 +92,7 @@ stt_add_timer(struct stt_timer *timer)
        spin_unlock(&stt_data.stt_lock);
 }
 
-/*
+/**
  * The function returns whether it has deactivated a pending timer or not.
  * (ie. del_timer() of an inactive timer returns 0, del_timer() of an
  * active timer returns 1.)
-- 
1.7.1

_______________________________________________
devel mailing list
de...@linuxdriverproject.org
http://driverdev.linuxdriverproject.org/mailman/listinfo/driverdev-devel

Reply via email to