CPU_NEWLY_IDLE 3029 kernel/sched.c if (idle == CPU_NEWLY_IDLE && this_rq->nr_running) CPU_NEWLY_IDLE 3111 kernel/sched.c else if (idle == CPU_NEWLY_IDLE) CPU_NEWLY_IDLE 3178 kernel/sched.c if (idle != CPU_NEWLY_IDLE && local_group && CPU_NEWLY_IDLE 3632 kernel/sched.c schedstat_inc(sd, lb_count[CPU_NEWLY_IDLE]); CPU_NEWLY_IDLE 3635 kernel/sched.c group = find_busiest_group(sd, this_cpu, &imbalance, CPU_NEWLY_IDLE, CPU_NEWLY_IDLE 3638 kernel/sched.c schedstat_inc(sd, lb_nobusyg[CPU_NEWLY_IDLE]); CPU_NEWLY_IDLE 3642 kernel/sched.c busiest = find_busiest_queue(group, CPU_NEWLY_IDLE, imbalance, cpus); CPU_NEWLY_IDLE 3644 kernel/sched.c schedstat_inc(sd, lb_nobusyq[CPU_NEWLY_IDLE]); CPU_NEWLY_IDLE 3650 kernel/sched.c schedstat_add(sd, lb_imbalance[CPU_NEWLY_IDLE], imbalance); CPU_NEWLY_IDLE 3659 kernel/sched.c imbalance, sd, CPU_NEWLY_IDLE, CPU_NEWLY_IDLE 3671 kernel/sched.c schedstat_inc(sd, lb_failed[CPU_NEWLY_IDLE]); CPU_NEWLY_IDLE 3682 kernel/sched.c schedstat_inc(sd, lb_balanced[CPU_NEWLY_IDLE]);