Hi Hongbo, kernel test robot noticed the following build errors:
[auto build test ERROR on next-20240902] url: https://github.com/intel-lab-lkp/linux/commits/Hongbo-Li/soc-fsl-qbman-make-use-of-the-helper-function-kthread_run_on_cpu/20240903-060257 base: next-20240902 patch link: https://lore.kernel.org/r/20240902133125.3089560-1-lihongbo22%40huawei.com patch subject: [PATCH -next] soc/fsl/qbman: make use of the helper function kthread_run_on_cpu() config: x86_64-allyesconfig (https://download.01.org/0day-ci/archive/20240903/202409032300.9u9g0c8j-...@intel.com/config) compiler: clang version 18.1.5 (https://github.com/llvm/llvm-project 617a15a9eac96088ae5e9134248d8236e34b91b1) reproduce (this is a W=1 build): (https://download.01.org/0day-ci/archive/20240903/202409032300.9u9g0c8j-...@intel.com/reproduce) If you fix the issue in a separate patch/commit (i.e. not just a new version of the same patch/commit), kindly add following tags | Reported-by: kernel test robot <l...@intel.com> | Closes: https://lore.kernel.org/oe-kbuild-all/202409032300.9u9g0c8j-...@intel.com/ All errors (new ones prefixed by >>): >> drivers/soc/fsl/qbman/qman_test_stash.c:112:27: error: too many arguments to >> function call, expected 4, have 5 111 | struct task_struct *k = kthread_run_on_cpu(bstrap_fn, &bstrap, | ~~~~~~~~~~~~~~~~~~ 112 | cpu, "hotpotato%d", cpu); | ^~~ include/linux/kthread.h:73:1: note: 'kthread_run_on_cpu' declared here 73 | kthread_run_on_cpu(int (*threadfn)(void *data), void *data, | ^ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 74 | unsigned int cpu, const char *namefmt) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 1 error generated. Kconfig warnings: (for reference only) WARNING: unmet direct dependencies detected for OMAP2PLUS_MBOX Depends on [n]: MAILBOX [=y] && (ARCH_OMAP2PLUS || ARCH_K3) Selected by [y]: - TI_K3_M4_REMOTEPROC [=y] && REMOTEPROC [=y] && (ARCH_K3 || COMPILE_TEST [=y]) vim +112 drivers/soc/fsl/qbman/qman_test_stash.c 35 36 /* 37 * Algorithm: 38 * 39 * Each cpu will have HP_PER_CPU "handlers" set up, each of which incorporates 40 * an rx/tx pair of FQ objects (both of which are stashed on dequeue). The 41 * organisation of FQIDs is such that the HP_PER_CPU*NUM_CPUS handlers will 42 * shuttle a "hot potato" frame around them such that every forwarding action 43 * moves it from one cpu to another. (The use of more than one handler per cpu 44 * is to allow enough handlers/FQs to truly test the significance of caching - 45 * ie. when cache-expiries are occurring.) 46 * 47 * The "hot potato" frame content will be HP_NUM_WORDS*4 bytes in size, and the 48 * first and last words of the frame data will undergo a transformation step on 49 * each forwarding action. To achieve this, each handler will be assigned a 50 * 32-bit "mixer", that is produced using a 32-bit LFSR. When a frame is 51 * received by a handler, the mixer of the expected sender is XOR'd into all 52 * words of the entire frame, which is then validated against the original 53 * values. Then, before forwarding, the entire frame is XOR'd with the mixer of 54 * the current handler. Apart from validating that the frame is taking the 55 * expected path, this also provides some quasi-realistic overheads to each 56 * forwarding action - dereferencing *all* the frame data, computation, and 57 * conditional branching. There is a "special" handler designated to act as the 58 * instigator of the test by creating an enqueuing the "hot potato" frame, and 59 * to determine when the test has completed by counting HP_LOOPS iterations. 60 * 61 * Init phases: 62 * 63 * 1. prepare each cpu's 'hp_cpu' struct using on_each_cpu(,,1) and link them 64 * into 'hp_cpu_list'. Specifically, set processor_id, allocate HP_PER_CPU 65 * handlers and link-list them (but do no other handler setup). 66 * 67 * 2. scan over 'hp_cpu_list' HP_PER_CPU times, the first time sets each 68 * hp_cpu's 'iterator' to point to its first handler. With each loop, 69 * allocate rx/tx FQIDs and mixer values to the hp_cpu's iterator handler 70 * and advance the iterator for the next loop. This includes a final fixup, 71 * which connects the last handler to the first (and which is why phase 2 72 * and 3 are separate). 73 * 74 * 3. scan over 'hp_cpu_list' HP_PER_CPU times, the first time sets each 75 * hp_cpu's 'iterator' to point to its first handler. With each loop, 76 * initialise FQ objects and advance the iterator for the next loop. 77 * Moreover, do this initialisation on the cpu it applies to so that Rx FQ 78 * initialisation targets the correct cpu. 79 */ 80 81 /* 82 * helper to run something on all cpus (can't use on_each_cpu(), as that invokes 83 * the fn from irq context, which is too restrictive). 84 */ 85 struct bstrap { 86 int (*fn)(void); 87 atomic_t started; 88 }; 89 static int bstrap_fn(void *bs) 90 { 91 struct bstrap *bstrap = bs; 92 int err; 93 94 atomic_inc(&bstrap->started); 95 err = bstrap->fn(); 96 if (err) 97 return err; 98 while (!kthread_should_stop()) 99 msleep(20); 100 return 0; 101 } 102 static int on_all_cpus(int (*fn)(void)) 103 { 104 int cpu; 105 106 for_each_cpu(cpu, cpu_online_mask) { 107 struct bstrap bstrap = { 108 .fn = fn, 109 .started = ATOMIC_INIT(0) 110 }; 111 struct task_struct *k = kthread_run_on_cpu(bstrap_fn, &bstrap, > 112 cpu, "hotpotato%d", > cpu); 113 int ret; 114 115 if (IS_ERR(k)) 116 return -ENOMEM; 117 /* 118 * If we call kthread_stop() before the "wake up" has had an 119 * effect, then the thread may exit with -EINTR without ever 120 * running the function. So poll until it's started before 121 * requesting it to stop. 122 */ 123 while (!atomic_read(&bstrap.started)) 124 msleep(20); 125 ret = kthread_stop(k); 126 if (ret) 127 return ret; 128 } 129 return 0; 130 } 131 -- 0-DAY CI Kernel Test Service https://github.com/intel/lkp-tests/wiki