[PATCH] sched: make idlest_group/cpu cpus_allowed-aware

Add relevant checks into find_idlest_group() and find_idlest_cpu() to make
them return only the groups that have allowed CPUs and allowed CPUs
respectively.

Signed-off-by: M.Baris Demiray <baris@labristeknoloji.com>
Signed-off-by: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
M.Baris Demiray 2005-09-10 00:26:09 -07:00 committed by Linus Torvalds
parent fc38ed7531
commit da5a552270
1 changed files with 13 additions and 4 deletions

View File

@ -966,8 +966,11 @@ find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu)
int local_group; int local_group;
int i; int i;
/* Skip over this group if it has no CPUs allowed */
if (!cpus_intersects(group->cpumask, p->cpus_allowed))
goto nextgroup;
local_group = cpu_isset(this_cpu, group->cpumask); local_group = cpu_isset(this_cpu, group->cpumask);
/* XXX: put a cpus allowed check */
/* Tally up the load of all CPUs in the group */ /* Tally up the load of all CPUs in the group */
avg_load = 0; avg_load = 0;
@ -992,6 +995,7 @@ find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu)
min_load = avg_load; min_load = avg_load;
idlest = group; idlest = group;
} }
nextgroup:
group = group->next; group = group->next;
} while (group != sd->groups); } while (group != sd->groups);
@ -1003,13 +1007,18 @@ find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu)
/* /*
* find_idlest_queue - find the idlest runqueue among the cpus in group. * find_idlest_queue - find the idlest runqueue among the cpus in group.
*/ */
static int find_idlest_cpu(struct sched_group *group, int this_cpu) static int find_idlest_cpu(struct sched_group *group,
struct task_struct *p, int this_cpu)
{ {
cpumask_t tmp;
unsigned long load, min_load = ULONG_MAX; unsigned long load, min_load = ULONG_MAX;
int idlest = -1; int idlest = -1;
int i; int i;
for_each_cpu_mask(i, group->cpumask) { /* Traverse only the allowed CPUs */
cpus_and(tmp, group->cpumask, p->cpus_allowed);
for_each_cpu_mask(i, tmp) {
load = source_load(i, 0); load = source_load(i, 0);
if (load < min_load || (load == min_load && i == this_cpu)) { if (load < min_load || (load == min_load && i == this_cpu)) {
@ -1052,7 +1061,7 @@ static int sched_balance_self(int cpu, int flag)
if (!group) if (!group)
goto nextlevel; goto nextlevel;
new_cpu = find_idlest_cpu(group, cpu); new_cpu = find_idlest_cpu(group, t, cpu);
if (new_cpu == -1 || new_cpu == cpu) if (new_cpu == -1 || new_cpu == cpu)
goto nextlevel; goto nextlevel;