this_rq 909 include/linux/sched.h unsigned long (*load_balance) (struct rq *this_rq, int this_cpu,
this_rq 914 include/linux/sched.h int (*move_one_task) (struct rq *this_rq, int this_cpu,
this_rq 917 include/linux/sched.h void (*pre_schedule) (struct rq *this_rq, struct task_struct *task);
this_rq 918 include/linux/sched.h void (*post_schedule) (struct rq *this_rq);
this_rq 919 include/linux/sched.h void (*task_wake_up) (struct rq *this_rq, struct task_struct *task);
this_rq 931 include/linux/sched.h void (*switched_from) (struct rq *this_rq, struct task_struct *task,
this_rq 933 include/linux/sched.h void (*switched_to) (struct rq *this_rq, struct task_struct *task,
this_rq 935 include/linux/sched.h void (*prio_changed) (struct rq *this_rq, struct task_struct *task,
this_rq 986 kernel/sched.c rq = this_rq();
this_rq 1068 kernel/sched.c if (rq == this_rq()) {
this_rq 2588 kernel/sched.c struct rq *rq = this_rq();
this_rq 2654 kernel/sched.c finish_task_switch(this_rq(), prev);
this_rq 2733 kernel/sched.c unsigned long this_load = this_rq->load.weight;
this_rq 2736 kernel/sched.c this_rq->nr_load_updates++;
this_rq 2744 kernel/sched.c old_load = this_rq->cpu_load[i];
this_rq 2753 kernel/sched.c this_rq->cpu_load[i] = (old_load*(scale-1) + new_load) >> i;
this_rq 2815 kernel/sched.c spin_unlock(&this_rq->lock);
this_rq 2819 kernel/sched.c if (busiest < this_rq) {
this_rq 2820 kernel/sched.c spin_unlock(&this_rq->lock);
this_rq 2822 kernel/sched.c spin_lock_nested(&this_rq->lock, SINGLE_DEPTH_NESTING);
this_rq 2834 kernel/sched.c lock_set_subclass(&this_rq->lock.dep_map, 0, _RET_IP_);
this_rq 2893 kernel/sched.c activate_task(this_rq, p, 0);
this_rq 2898 kernel/sched.c check_preempt_curr(this_rq, p, 0);
this_rq 2979 kernel/sched.c pull_task(busiest, p, this_rq, this_cpu);
this_rq 3020 kernel/sched.c int this_best_prio = this_rq->curr->prio;
this_rq 3024 kernel/sched.c class->load_balance(this_rq, this_cpu, busiest,
this_rq 3029 kernel/sched.c if (idle == CPU_NEWLY_IDLE && this_rq->nr_running)
this_rq 3047 kernel/sched.c pull_task(busiest, p, this_rq, this_cpu);
this_rq 3076 kernel/sched.c if (class->move_one_task(this_rq, this_cpu, busiest, sd, idle))
this_rq 3491 kernel/sched.c BUG_ON(busiest == this_rq);
this_rq 3504 kernel/sched.c double_rq_lock(this_rq, busiest);
this_rq 3505 kernel/sched.c ld_moved = move_tasks(this_rq, this_cpu, busiest,
this_rq 3507 kernel/sched.c double_rq_unlock(this_rq, busiest);
this_rq 3634 kernel/sched.c update_shares_locked(this_rq, sd);
this_rq 3648 kernel/sched.c BUG_ON(busiest == this_rq);
this_rq 3655 kernel/sched.c double_lock_balance(this_rq, busiest);
this_rq 3658 kernel/sched.c ld_moved = move_tasks(this_rq, this_cpu, busiest,
this_rq 3661 kernel/sched.c double_unlock_balance(this_rq, busiest);
this_rq 3678 kernel/sched.c update_shares_locked(this_rq, sd);
this_rq 3710 kernel/sched.c pulled_task = load_balance_newidle(this_cpu, this_rq,
this_rq 3719 kernel/sched.c if (pulled_task || time_after(jiffies, this_rq->next_balance)) {
this_rq 3724 kernel/sched.c this_rq->next_balance = next_balance;
this_rq 3940 kernel/sched.c struct rq *this_rq = cpu_rq(this_cpu);
this_rq 3941 kernel/sched.c enum cpu_idle_type idle = this_rq->idle_at_tick ?
this_rq 3952 kernel/sched.c if (this_rq->idle_at_tick &&
this_rq 3971 kernel/sched.c if (time_after(this_rq->next_balance, rq->next_balance))
this_rq 3972 kernel/sched.c this_rq->next_balance = rq->next_balance;
this_rq 4138 kernel/sched.c struct rq *rq = this_rq();
this_rq 4184 kernel/sched.c struct rq *rq = this_rq();
this_rq 4377 kernel/sched.c schedstat_inc(this_rq(), sched_count);
this_rq 4380 kernel/sched.c schedstat_inc(this_rq(), bkl_count);
this_rq 1096 kernel/sched_fair.c struct task_struct *curr = this_rq->curr;
this_rq 1161 kernel/sched_fair.c struct rq *this_rq;
this_rq 1167 kernel/sched_fair.c this_rq = cpu_rq(this_cpu);
this_rq 1199 kernel/sched_fair.c if (wake_affine(this_sd, this_rq, p, prev_cpu, this_cpu, sync, idx,
this_rq 1373 kernel/sched_fair.c return balance_tasks(this_rq, this_cpu, busiest,
this_rq 1407 kernel/sched_fair.c moved_load = __load_balance_fair(this_rq, this_cpu, busiest,
this_rq 1432 kernel/sched_fair.c return __load_balance_fair(this_rq, this_cpu, busiest,
this_rq 1454 kernel/sched_fair.c if (iter_move_one_task(this_rq, this_cpu, busiest, sd, idle,
this_rq 1171 kernel/sched_rt.c int this_cpu = this_rq->cpu, ret = 0, cpu;
this_rq 1175 kernel/sched_rt.c if (likely(!rt_overloaded(this_rq)))
this_rq 1178 kernel/sched_rt.c next = pick_next_task_rt(this_rq);
this_rq 1180 kernel/sched_rt.c for_each_cpu_mask_nr(cpu, this_rq->rd->rto_mask) {
this_rq 1192 kernel/sched_rt.c if (double_lock_balance(this_rq, src_rq)) {
this_rq 1195 kernel/sched_rt.c next = pick_next_task_rt(this_rq);
this_rq 1234 kernel/sched_rt.c activate_task(this_rq, p, 0);
this_rq 1249 kernel/sched_rt.c double_unlock_balance(this_rq, src_rq);