From: Pierce Griffiths <pierceagriffi...@gmail.com> *Condense ssequential if statements into a single if statement when they share an outcome *Eliminate a jump instruction by replacing a goto with a return *Eliminate an unnecessary local variable *Replace "if(function or boolean expression) return true else return false" with "return (function or boolean expression);" *Remove null pointer checks before calls to kfree on the grounds that kfree(NULL) results in a no-op
Signed-off-by: Pierce Griffiths <pierceagriffi...@gmail.com> --- Please tell me if I've made any mistakes in the submission process, this is the first time I've submitted a patch to the Linux kernel. kernel/sched/core.c | 8 ++------ kernel/sched/cpufreq.c | 5 +---- kernel/sched/cpupri.c | 22 ++++++++-------------- kernel/sched/rt.c | 31 ++++++++++++------------------- 4 files changed, 23 insertions(+), 43 deletions(-) diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 625bc9897f62..443a1f235cfd 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -617,12 +617,8 @@ bool sched_can_stop_tick(struct rq *rq) * If there are more than one RR tasks, we need the tick to effect the * actual RR behaviour. */ - if (rq->rt.rr_nr_running) { - if (rq->rt.rr_nr_running == 1) - return true; - else - return false; - } + if (rq->rt.rr_nr_running) + return rq->rt.rr_nr_running == 1; /* * If there's no RR tasks, but FIFO tasks, we can skip the tick, no diff --git a/kernel/sched/cpufreq.c b/kernel/sched/cpufreq.c index 5e54cbcae673..a8fd4bd68954 100644 --- a/kernel/sched/cpufreq.c +++ b/kernel/sched/cpufreq.c @@ -34,10 +34,7 @@ void cpufreq_add_update_util_hook(int cpu, struct update_util_data *data, void (*func)(struct update_util_data *data, u64 time, unsigned int flags)) { - if (WARN_ON(!data || !func)) - return; - - if (WARN_ON(per_cpu(cpufreq_update_util_data, cpu))) + if (WARN_ON(!data || !func || per_cpu(cpufreq_update_util_data, cpu))) return; data->func = func; diff --git a/kernel/sched/cpupri.c b/kernel/sched/cpupri.c index daaadf939ccb..152c133e8247 100644 --- a/kernel/sched/cpupri.c +++ b/kernel/sched/cpupri.c @@ -29,20 +29,16 @@ #include "sched.h" /* Convert between a 140 based task->prio, and our 102 based cpupri */ -static int convert_prio(int prio) +static int convert_prio(const int prio) { - int cpupri; - if (prio == CPUPRI_INVALID) - cpupri = CPUPRI_INVALID; + return CPUPRI_INVALID; else if (prio == MAX_PRIO) - cpupri = CPUPRI_IDLE; + return CPUPRI_IDLE; else if (prio >= MAX_RT_PRIO) - cpupri = CPUPRI_NORMAL; + return CPUPRI_NORMAL; else - cpupri = MAX_RT_PRIO - prio + 1; - - return cpupri; + return MAX_RT_PRIO - prio + 1; } /** @@ -95,10 +91,8 @@ int cpupri_find(struct cpupri *cp, struct task_struct *p, smp_rmb(); /* Need to do the rmb for every iteration */ - if (skip) - continue; - - if (cpumask_any_and(&p->cpus_allowed, vec->mask) >= nr_cpu_ids) + if (skip || cpumask_any_and(&p->cpus_allowed, vec->mask) + >= nr_cpu_ids) continue; if (lowest_mask) { @@ -222,7 +216,7 @@ int cpupri_init(struct cpupri *cp) return 0; cleanup: - for (i--; i >= 0; i--) + while (--i >= 0) free_cpumask_var(cp->pri_to_cpu[i].mask); return -ENOMEM; } diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c index 2e2955a8cf8f..acf1b94669ad 100644 --- a/kernel/sched/rt.c +++ b/kernel/sched/rt.c @@ -142,10 +142,12 @@ void free_rt_sched_group(struct task_group *tg) destroy_rt_bandwidth(&tg->rt_bandwidth); for_each_possible_cpu(i) { - if (tg->rt_rq) - kfree(tg->rt_rq[i]); - if (tg->rt_se) - kfree(tg->rt_se[i]); + /* Don't need to check if tg->rt_rq[i] + * or tg->rt_se[i] are NULL, since kfree(NULL) + * simply performs no operation + */ + kfree(tg->rt_rq[i]); + kfree(tg->rt_se[i]); } kfree(tg->rt_rq); @@ -1015,10 +1017,7 @@ enqueue_top_rt_rq(struct rt_rq *rt_rq) BUG_ON(&rq->rt != rt_rq); - if (rt_rq->rt_queued) - return; - - if (rt_rq_throttled(rt_rq)) + if (rt_rq->rt_queued || rt_rq_throttled(rt_rq)) return; if (rt_rq->rt_nr_running) { @@ -1211,10 +1210,7 @@ void dec_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) */ static inline bool move_entity(unsigned int flags) { - if ((flags & (DEQUEUE_SAVE | DEQUEUE_MOVE)) == DEQUEUE_SAVE) - return false; - - return true; + return !((flags & (DEQUEUE_SAVE | DEQUEUE_MOVE)) == DEQUEUE_SAVE) } static void __delist_rt_entity(struct sched_rt_entity *rt_se, struct rt_prio_array *array) @@ -1393,7 +1389,7 @@ select_task_rq_rt(struct task_struct *p, int cpu, int sd_flag, int flags) /* For anything but wake ups, just return the task_cpu */ if (sd_flag != SD_BALANCE_WAKE && sd_flag != SD_BALANCE_FORK) - goto out; + return cpu; rq = cpu_rq(cpu); @@ -1437,7 +1433,6 @@ select_task_rq_rt(struct task_struct *p, int cpu, int sd_flag, int flags) } rcu_read_unlock(); -out: return cpu; } @@ -2518,12 +2513,10 @@ static int tg_set_rt_bandwidth(struct task_group *tg, /* * Disallowing the root group RT runtime is BAD, it would disallow the * kernel creating (and or operating) RT threads. + * + * No period doesn't make any sense. */ - if (tg == &root_task_group && rt_runtime == 0) - return -EINVAL; - - /* No period doesn't make any sense. */ - if (rt_period == 0) + if ((tg == &root_task_group && !rt_runtime) || !rt_period) return -EINVAL; mutex_lock(&rt_constraints_mutex); -- 2.19.0