void (*put_prev_task) (struct rq *rq, struct task_struct *p);
#ifdef CONFIG_SMP
- int (*select_task_rq)(struct task_struct *p, int flag, int sync);
+ int (*select_task_rq)(struct task_struct *p, int sd_flag, int sync);
unsigned long (*load_balance) (struct rq *this_rq, int this_cpu,
struct rq *busiest, unsigned long max_load_move,
*
* preempt must be disabled.
*/
-static int select_task_rq_fair(struct task_struct *p, int flag, int sync)
+static int select_task_rq_fair(struct task_struct *p, int sd_flag, int sync)
{
struct sched_domain *tmp, *sd = NULL;
int cpu = smp_processor_id();
int new_cpu = cpu;
int want_affine = 0;
- if (flag & SD_BALANCE_WAKE) {
+ if (sd_flag & SD_BALANCE_WAKE) {
if (sched_feat(AFFINE_WAKEUPS))
want_affine = 1;
new_cpu = prev_cpu;
break;
}
- switch (flag) {
+ switch (sd_flag) {
case SD_BALANCE_WAKE:
if (!sched_feat(LB_WAKEUP_UPDATE))
break;
want_affine = 0;
}
- if (!(tmp->flags & flag))
+ if (!(tmp->flags & sd_flag))
continue;
sd = tmp;
struct sched_group *group;
int weight;
- if (!(sd->flags & flag)) {
+ if (!(sd->flags & sd_flag)) {
sd = sd->child;
continue;
}
- group = find_idlest_group(sd, p, cpu, flag);
+ group = find_idlest_group(sd, p, cpu, sd_flag);
if (!group) {
sd = sd->child;
continue;
for_each_domain(cpu, tmp) {
if (weight <= cpumask_weight(sched_domain_span(tmp)))
break;
- if (tmp->flags & flag)
+ if (tmp->flags & sd_flag)
sd = tmp;
}
/* while loop will break here if sd == NULL */
*/
#ifdef CONFIG_SMP
-static int select_task_rq_idle(struct task_struct *p, int flag, int sync)
+static int select_task_rq_idle(struct task_struct *p, int sd_flag, int sync)
{
return task_cpu(p); /* IDLE tasks as never migrated */
}
#ifdef CONFIG_SMP
static int find_lowest_rq(struct task_struct *task);
-static int select_task_rq_rt(struct task_struct *p, int flag, int sync)
+static int select_task_rq_rt(struct task_struct *p, int sd_flag, int sync)
{
struct rq *rq = task_rq(p);
- if (flag != SD_BALANCE_WAKE)
+ if (sd_flag != SD_BALANCE_WAKE)
return smp_processor_id();
/*