rq_of 432 kernel/sched_fair.c u64 now = rq_of(cfs_rq)->clock; rq_of 458 kernel/sched_fair.c schedstat_set(se->wait_start, rq_of(cfs_rq)->clock); rq_of 478 kernel/sched_fair.c rq_of(cfs_rq)->clock - se->wait_start)); rq_of 481 kernel/sched_fair.c rq_of(cfs_rq)->clock - se->wait_start); rq_of 505 kernel/sched_fair.c se->exec_start = rq_of(cfs_rq)->clock; rq_of 530 kernel/sched_fair.c inc_cpu_load(rq_of(cfs_rq), se->load.weight); rq_of 544 kernel/sched_fair.c dec_cpu_load(rq_of(cfs_rq), se->load.weight); rq_of 557 kernel/sched_fair.c u64 delta = rq_of(cfs_rq)->clock - se->sleep_start; rq_of 572 kernel/sched_fair.c u64 delta = rq_of(cfs_rq)->clock - se->block_start; rq_of 688 kernel/sched_fair.c se->sleep_start = rq_of(cfs_rq)->clock; rq_of 690 kernel/sched_fair.c se->block_start = rq_of(cfs_rq)->clock; rq_of 711 kernel/sched_fair.c resched_task(rq_of(cfs_rq)->curr); rq_of 736 kernel/sched_fair.c if (rq_of(cfs_rq)->load.weight >= 2*se->load.weight) { rq_of 747 kernel/sched_fair.c struct rq *rq = rq_of(cfs_rq); rq_of 803 kernel/sched_fair.c resched_task(rq_of(cfs_rq)->curr); rq_of 810 kernel/sched_fair.c hrtimer_active(&rq_of(cfs_rq)->hrtick_timer))