CPU_NEWLY_IDLE   3029 kernel/sched.c 		if (idle == CPU_NEWLY_IDLE && this_rq->nr_running)
CPU_NEWLY_IDLE   3111 kernel/sched.c 	else if (idle == CPU_NEWLY_IDLE)
CPU_NEWLY_IDLE   3178 kernel/sched.c 		if (idle != CPU_NEWLY_IDLE && local_group &&
CPU_NEWLY_IDLE   3632 kernel/sched.c 	schedstat_inc(sd, lb_count[CPU_NEWLY_IDLE]);
CPU_NEWLY_IDLE   3635 kernel/sched.c 	group = find_busiest_group(sd, this_cpu, &imbalance, CPU_NEWLY_IDLE,
CPU_NEWLY_IDLE   3638 kernel/sched.c 		schedstat_inc(sd, lb_nobusyg[CPU_NEWLY_IDLE]);
CPU_NEWLY_IDLE   3642 kernel/sched.c 	busiest = find_busiest_queue(group, CPU_NEWLY_IDLE, imbalance, cpus);
CPU_NEWLY_IDLE   3644 kernel/sched.c 		schedstat_inc(sd, lb_nobusyq[CPU_NEWLY_IDLE]);
CPU_NEWLY_IDLE   3650 kernel/sched.c 	schedstat_add(sd, lb_imbalance[CPU_NEWLY_IDLE], imbalance);
CPU_NEWLY_IDLE   3659 kernel/sched.c 					imbalance, sd, CPU_NEWLY_IDLE,
CPU_NEWLY_IDLE   3671 kernel/sched.c 		schedstat_inc(sd, lb_failed[CPU_NEWLY_IDLE]);
CPU_NEWLY_IDLE   3682 kernel/sched.c 	schedstat_inc(sd, lb_balanced[CPU_NEWLY_IDLE]);