Get rid of the superfluous dst_cpu, move the cpu_mask inside the search
function, and collapse the two redundant pick-next-rt() functions.

Signed-off-by: Gregory Haskins <[EMAIL PROTECTED]>
---

 kernel/sched.c    |   27 ++++++++++++---------------
 kernel/sched_rt.c |   44 --------------------------------------------
 2 files changed, 12 insertions(+), 59 deletions(-)

diff --git a/kernel/sched.c b/kernel/sched.c
index 821f115..de5fc1c 100644
--- a/kernel/sched.c
+++ b/kernel/sched.c
@@ -1471,9 +1471,10 @@ next_in_queue:
         * Return the highest-prio non-running RT task (if task
         * may run on this CPU):
         */
-       if (!task_running(src_rq, tmp) &&
-                               cpu_isset(this_cpu, tmp->cpus_allowed))
-               return tmp;
+       if (!task_running(src_rq, tmp)) {
+               if ((this_cpu == -1) || cpu_isset(this_cpu, tmp->cpus_allowed))
+                       return tmp;
+       }
 
        curr = curr->next;
        if (curr != head)
@@ -1489,20 +1490,21 @@ static int double_lock_balance(struct rq *this_rq, 
struct rq *busiest);
 #define RT_PUSH_MAX_TRIES 3
 
 /* Will lock the rq it finds */
-static struct rq *find_lock_lowest_rq(cpumask_t *cpu_mask,
-                                     struct task_struct *task,
+static struct rq *find_lock_lowest_rq(struct task_struct *task,
                                      struct rq *this_rq)
 {
        struct rq *lowest_rq = NULL;
-       int dst_cpu = -1;
        int cpu;
        int tries;
+       cpumask_t cpu_mask;
+
+       cpus_and(cpu_mask, cpu_online_map, task->cpus_allowed);
 
        for (tries = 0; tries < RT_PUSH_MAX_TRIES; tries++) {
                /*
                 * Scan each rq for the lowest prio.
                 */
-               for_each_cpu_mask(cpu, *cpu_mask) {
+               for_each_cpu_mask(cpu, cpu_mask) {
                        struct rq *rq = &per_cpu(runqueues, cpu);
 
                        if (cpu == this_rq->cpu)
@@ -1511,7 +1513,6 @@ static struct rq *find_lock_lowest_rq(cpumask_t *cpu_mask,
                        /* We look for lowest RT prio or non-rt CPU */
                        if (rq->curr_prio >= MAX_RT_PRIO) {
                                lowest_rq = rq;
-                               dst_cpu = cpu;
                                break;
                        }
 
@@ -1519,7 +1520,6 @@ static struct rq *find_lock_lowest_rq(cpumask_t *cpu_mask,
                        if (rq->curr_prio > task->prio &&
                            (!lowest_rq || rq->curr_prio < 
lowest_rq->curr_prio)) {
                                lowest_rq = rq;
-                               dst_cpu = cpu;
                        }
                }
 
@@ -1535,7 +1535,7 @@ static struct rq *find_lock_lowest_rq(cpumask_t *cpu_mask,
                         */
                        if (unlikely(task_running(this_rq, task) ||
                                     task_rq(task) != this_rq ||
-                                    !cpu_isset(dst_cpu, task->cpus_allowed))) {
+                                    !cpu_isset(lowest_rq->cpu, 
task->cpus_allowed))) {
                                spin_unlock(&lowest_rq->lock);
                                lowest_rq = NULL;
                                break;
@@ -1566,21 +1566,18 @@ static int push_rt_task(struct rq *this_rq)
        struct rq *lowest_rq;
        int dst_cpu;
        int ret = 0;
-       cpumask_t cpu_mask;
 
        assert_spin_locked(&this_rq->lock);
 
-       next_task = rt_next_highest_task(this_rq);
+       next_task = pick_rt_task(this_rq, -1);
        if (!next_task)
                return 0;
 
-       cpus_and(cpu_mask, cpu_online_map, next_task->cpus_allowed);
-
        /* We might release this_rq lock */
        get_task_struct(next_task);
 
        /* find_lock_lowest_rq locks the rq if found */
-       lowest_rq = find_lock_lowest_rq(&cpu_mask, next_task, this_rq);
+       lowest_rq = find_lock_lowest_rq(next_task, this_rq);
        if (!lowest_rq)
                goto out;
 
diff --git a/kernel/sched_rt.c b/kernel/sched_rt.c
index 546526b..9b677c1 100644
--- a/kernel/sched_rt.c
+++ b/kernel/sched_rt.c
@@ -96,50 +96,6 @@ static struct task_struct *pick_next_task_rt(struct rq *rq)
        return next;
 }
 
-#ifdef CONFIG_PREEMPT_RT
-static struct task_struct *rt_next_highest_task(struct rq *rq)
-{
-       struct rt_prio_array *array = &rq->rt.active;
-       struct task_struct *next;
-       struct list_head *queue;
-       int idx;
-
-       if (likely (rq->rt_nr_running < 2))
-               return NULL;
-
-       idx = sched_find_first_bit(array->bitmap);
-       if (idx >= MAX_RT_PRIO) {
-               WARN_ON(1); /* rt_nr_running is bad */
-               return NULL;
-       }
-
-       queue = array->queue + idx;
-       next = list_entry(queue->next, struct task_struct, run_list);
-       if (unlikely(next != current))
-               return next;
-
-       if (queue->next->next != queue) {
-               /* same prio task */
-               next = list_entry(queue->next->next, struct task_struct, 
run_list);
-               goto out;
-       }
-
-       /* slower, but more flexible */
-       idx = find_next_bit(array->bitmap, MAX_RT_PRIO, idx+1);
-       if (idx >= MAX_RT_PRIO) {
-               WARN_ON(1); /* rt_nr_running was 2 and above! */
-               return NULL;
-       }
-
-       queue = array->queue + idx;
-       next = list_entry(queue->next, struct task_struct, run_list);
-
- out:
-       return next;
-
-}
-#endif /* CONFIG_PREEMPT_RT */
-
 static void put_prev_task_rt(struct rq *rq, struct task_struct *p)
 {
        update_curr_rt(rq);

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to