rq_of             432 kernel/sched_fair.c 	u64 now = rq_of(cfs_rq)->clock;
rq_of             458 kernel/sched_fair.c 	schedstat_set(se->wait_start, rq_of(cfs_rq)->clock);
rq_of             478 kernel/sched_fair.c 			rq_of(cfs_rq)->clock - se->wait_start));
rq_of             481 kernel/sched_fair.c 			rq_of(cfs_rq)->clock - se->wait_start);
rq_of             505 kernel/sched_fair.c 	se->exec_start = rq_of(cfs_rq)->clock;
rq_of             530 kernel/sched_fair.c 		inc_cpu_load(rq_of(cfs_rq), se->load.weight);
rq_of             544 kernel/sched_fair.c 		dec_cpu_load(rq_of(cfs_rq), se->load.weight);
rq_of             557 kernel/sched_fair.c 		u64 delta = rq_of(cfs_rq)->clock - se->sleep_start;
rq_of             572 kernel/sched_fair.c 		u64 delta = rq_of(cfs_rq)->clock - se->block_start;
rq_of             688 kernel/sched_fair.c 				se->sleep_start = rq_of(cfs_rq)->clock;
rq_of             690 kernel/sched_fair.c 				se->block_start = rq_of(cfs_rq)->clock;
rq_of             711 kernel/sched_fair.c 		resched_task(rq_of(cfs_rq)->curr);
rq_of             736 kernel/sched_fair.c 	if (rq_of(cfs_rq)->load.weight >= 2*se->load.weight) {
rq_of             747 kernel/sched_fair.c 	struct rq *rq = rq_of(cfs_rq);
rq_of             803 kernel/sched_fair.c 		resched_task(rq_of(cfs_rq)->curr);
rq_of             810 kernel/sched_fair.c 			hrtimer_active(&rq_of(cfs_rq)->hrtick_timer))