task_cpu 482 fs/proc/array.c task_cpu(task), task_cpu 180 kernel/cpu.c if (task_cpu(p) == cpu && task_cpu 1175 kernel/fork.c if (unlikely(!cpu_isset(task_cpu(p), p->cpus_allowed) || task_cpu 1176 kernel/fork.c !cpu_online(task_cpu(p)))) task_cpu 638 kernel/sched.c #define task_rq(p) cpu_rq(task_cpu(p)) task_cpu 1167 kernel/sched.c cpu = task_cpu(p); task_cpu 1755 kernel/sched.c return cpu_curr(task_cpu(p)) == p; task_cpu 1822 kernel/sched.c int old_cpu = task_cpu(p); task_cpu 2006 kernel/sched.c cpu = task_cpu(p); task_cpu 2225 kernel/sched.c cpu = task_cpu(p); task_cpu 2245 kernel/sched.c cpu = task_cpu(p); task_cpu 2266 kernel/sched.c cpu = task_cpu(p); task_cpu 5974 kernel/sched.c if (cpu_isset(task_cpu(p), *new_mask)) task_cpu 6016 kernel/sched.c if (task_cpu(p) != src_cpu) task_cpu 6194 kernel/sched.c if (task_cpu(p) == src_cpu) task_cpu 8723 kernel/sched.c set_task_rq(tsk, task_cpu(tsk)); task_cpu 9355 kernel/sched.c u64 *cpuusage = percpu_ptr(ca->cpuusage, task_cpu(tsk)); task_cpu 105 kernel/sched_debug.c if (!p->se.on_rq || task_cpu(p) != rq_cpu) task_cpu 982 kernel/sched_fair.c if (i != task_cpu(p)) { task_cpu 1165 kernel/sched_fair.c prev_cpu = task_cpu(p); task_cpu 1498 kernel/sched_fair.c if (sysctl_sched_child_runs_first && this_cpu == task_cpu(p) && task_cpu 11 kernel/sched_idletask.c return task_cpu(p); /* IDLE tasks as never migrated */ task_cpu 794 kernel/sched_rt.c return (cpu == -1) ? task_cpu(p) : cpu; task_cpu 801 kernel/sched_rt.c return task_cpu(p); task_cpu 979 kernel/sched_rt.c int cpu = task_cpu(task); task_cpu 70 kernel/trace/trace_sched_wakeup.c if (task_cpu(wakeup_task) != cpu) task_cpu 269 kernel/trace/trace_sched_wakeup.c wakeup_cpu = task_cpu(p); task_cpu 302 mm/oom_kill.c get_mm_rss(p->mm), (int)task_cpu(p), p->oomkilladj,