rq 148 block/as-iosched.c #define RQ_IOC(rq) ((struct io_context *) (rq)->elevator_private)
rq 149 block/as-iosched.c #define RQ_STATE(rq) ((enum arq_state)(rq)->elevator_private2)
rq 150 block/as-iosched.c #define RQ_SET_STATE(rq, state) ((rq)->elevator_private2 = (void *) state)
rq 244 block/as-iosched.c if (unlikely(!RQ_IOC(rq)))
rq 247 block/as-iosched.c aic = RQ_IOC(rq)->aic;
rq 249 block/as-iosched.c if (rq_is_sync(rq) && aic) {
rq 258 block/as-iosched.c put_io_context(RQ_IOC(rq));
rq 264 block/as-iosched.c #define RQ_RB_ROOT(ad, rq) (&(ad)->sort_list[rq_is_sync((rq))])
rq 270 block/as-iosched.c while ((unlikely(alias = elv_rb_add(RQ_RB_ROOT(ad, rq), rq)))) {
rq 278 block/as-iosched.c elv_rb_del(RQ_RB_ROOT(ad, rq), rq);
rq 550 block/as-iosched.c int data_dir = rq_is_sync(rq);
rq 572 block/as-iosched.c if (aic->last_request_pos < rq->sector)
rq 573 block/as-iosched.c seek_dist = rq->sector - aic->last_request_pos;
rq 575 block/as-iosched.c seek_dist = aic->last_request_pos - rq->sector;
rq 578 block/as-iosched.c aic->last_request_pos = rq->sector + rq->nr_sectors;
rq 593 block/as-iosched.c sector_t next = rq->sector;
rq 659 block/as-iosched.c if (rq && ioc == RQ_IOC(rq)) {
rq 692 block/as-iosched.c if (rq && rq_is_sync(rq) && as_close_req(ad, aic, rq)) {
rq 708 block/as-iosched.c as_update_iohist(ad, aic, rq);
rq 768 block/as-iosched.c if (as_can_break_anticipation(ad, rq))
rq 792 block/as-iosched.c const int data_dir = rq_is_sync(rq);
rq 795 block/as-iosched.c ad->next_rq[data_dir] = as_choose_req(ad, rq, ad->next_rq[data_dir]);
rq 804 block/as-iosched.c if (as_can_break_anticipation(ad, rq))
rq 845 block/as-iosched.c WARN_ON(!list_empty(&rq->queuelist));
rq 847 block/as-iosched.c if (RQ_STATE(rq) != AS_RQ_REMOVED) {
rq 848 block/as-iosched.c WARN(1, "rq->state %d\n", RQ_STATE(rq));
rq 869 block/as-iosched.c if (ad->new_batch && ad->batch_data_dir == rq_is_sync(rq)) {
rq 876 block/as-iosched.c if (ad->io_context == RQ_IOC(rq) && ad->io_context) {
rq 888 block/as-iosched.c as_put_io_context(rq);
rq 890 block/as-iosched.c RQ_SET_STATE(rq, AS_RQ_POSTSCHED);
rq 902 block/as-iosched.c const int data_dir = rq_is_sync(rq);
rq 906 block/as-iosched.c WARN_ON(RQ_STATE(rq) != AS_RQ_QUEUED);
rq 908 block/as-iosched.c ioc = RQ_IOC(rq);
rq 918 block/as-iosched.c if (ad->next_rq[data_dir] == rq)
rq 919 block/as-iosched.c ad->next_rq[data_dir] = as_find_next_rq(ad, rq);
rq 921 block/as-iosched.c rq_fifo_clear(rq);
rq 922 block/as-iosched.c as_del_rq_rb(ad, rq);
rq 935 block/as-iosched.c struct request *rq;
rq 949 block/as-iosched.c rq = rq_entry_fifo(ad->fifo_list[adir].next);
rq 951 block/as-iosched.c return time_after(jiffies, rq_fifo_time(rq));
rq 976 block/as-iosched.c const int data_dir = rq_is_sync(rq);
rq 978 block/as-iosched.c BUG_ON(RB_EMPTY_NODE(&rq->rb_node));
rq 987 block/as-iosched.c ad->last_sector[data_dir] = rq->sector + rq->nr_sectors;
rq 990 block/as-iosched.c struct io_context *ioc = RQ_IOC(rq);
rq 1004 block/as-iosched.c ad->next_rq[data_dir] = as_find_next_rq(ad, rq);
rq 1009 block/as-iosched.c as_remove_queued_request(ad->q, rq);
rq 1010 block/as-iosched.c WARN_ON(RQ_STATE(rq) != AS_RQ_QUEUED);
rq 1012 block/as-iosched.c elv_dispatch_sort(ad->q, rq);
rq 1014 block/as-iosched.c RQ_SET_STATE(rq, AS_RQ_DISPATCHED);
rq 1015 block/as-iosched.c if (RQ_IOC(rq) && RQ_IOC(rq)->aic)
rq 1016 block/as-iosched.c atomic_inc(&RQ_IOC(rq)->aic->nr_dispatched);
rq 1030 block/as-iosched.c struct request *rq;
rq 1077 block/as-iosched.c rq = ad->next_rq[ad->batch_data_dir];
rq 1083 block/as-iosched.c if (as_can_anticipate(ad, rq)) {
rq 1089 block/as-iosched.c if (rq) {
rq 1117 block/as-iosched.c rq = rq_entry_fifo(ad->fifo_list[REQ_SYNC].next);
rq 1143 block/as-iosched.c rq = rq_entry_fifo(ad->fifo_list[REQ_ASYNC].next);
rq 1158 block/as-iosched.c rq = rq_entry_fifo(ad->fifo_list[ad->batch_data_dir].next);
rq 1179 block/as-iosched.c as_move_to_dispatch(ad, rq);
rq 1192 block/as-iosched.c RQ_SET_STATE(rq, AS_RQ_NEW);
rq 1194 block/as-iosched.c data_dir = rq_is_sync(rq);
rq 1196 block/as-iosched.c rq->elevator_private = as_get_io_context(q->node);
rq 1198 block/as-iosched.c if (RQ_IOC(rq)) {
rq 1199 block/as-iosched.c as_update_iohist(ad, RQ_IOC(rq)->aic, rq);
rq 1200 block/as-iosched.c atomic_inc(&RQ_IOC(rq)->aic->nr_queued);
rq 1203 block/as-iosched.c as_add_rq_rb(ad, rq);
rq 1208 block/as-iosched.c rq_set_fifo_time(rq, jiffies + ad->fifo_expire[data_dir]);
rq 1209 block/as-iosched.c list_add_tail(&rq->queuelist, &ad->fifo_list[data_dir]);
rq 1211 block/as-iosched.c as_update_rq(ad, rq); /* keep state machine up to date */
rq 1212 block/as-iosched.c RQ_SET_STATE(rq, AS_RQ_QUEUED);
rq 1217 block/as-iosched.c WARN_ON(RQ_STATE(rq) != AS_RQ_DISPATCHED);
rq 1218 block/as-iosched.c RQ_SET_STATE(rq, AS_RQ_REMOVED);
rq 1219 block/as-iosched.c if (RQ_IOC(rq) && RQ_IOC(rq)->aic)
rq 1220 block/as-iosched.c atomic_dec(&RQ_IOC(rq)->aic->nr_dispatched);
rq 1225 block/as-iosched.c WARN_ON(RQ_STATE(rq) != AS_RQ_REMOVED);
rq 1226 block/as-iosched.c RQ_SET_STATE(rq, AS_RQ_DISPATCHED);
rq 1227 block/as-iosched.c if (RQ_IOC(rq) && RQ_IOC(rq)->aic)
rq 1228 block/as-iosched.c atomic_inc(&RQ_IOC(rq)->aic->nr_dispatched);
rq 64 block/blk-barrier.c struct request_queue *q = rq->q;
rq 68 block/blk-barrier.c if (rq == &q->pre_flush_rq)
rq 70 block/blk-barrier.c if (rq == &q->bar_rq)
rq 72 block/blk-barrier.c if (rq == &q->post_flush_rq)
rq 81 block/blk-barrier.c if (!blk_fs_request(rq))
rq 84 block/blk-barrier.c if ((rq->cmd_flags & REQ_ORDERED_COLOR) ==
rq 93 block/blk-barrier.c struct request *rq;
rq 108 block/blk-barrier.c rq = q->orig_bar_rq;
rq 110 block/blk-barrier.c if (__blk_end_request(rq, q->orderr, blk_rq_bytes(rq)))
rq 116 block/blk-barrier.c elv_completed_request(rq->q, rq);
rq 117 block/blk-barrier.c blk_ordered_complete_seq(rq->q, QUEUE_ORDSEQ_PREFLUSH, error);
rq 122 block/blk-barrier.c elv_completed_request(rq->q, rq);
rq 123 block/blk-barrier.c blk_ordered_complete_seq(rq->q, QUEUE_ORDSEQ_BAR, error);
rq 128 block/blk-barrier.c elv_completed_request(rq->q, rq);
rq 129 block/blk-barrier.c blk_ordered_complete_seq(rq->q, QUEUE_ORDSEQ_POSTFLUSH, error);
rq 134 block/blk-barrier.c struct request *rq;
rq 138 block/blk-barrier.c rq = &q->pre_flush_rq;
rq 141 block/blk-barrier.c rq = &q->post_flush_rq;
rq 145 block/blk-barrier.c blk_rq_init(q, rq);
rq 146 block/blk-barrier.c rq->cmd_flags = REQ_HARDBARRIER;
rq 147 block/blk-barrier.c rq->rq_disk = q->bar_rq.rq_disk;
rq 148 block/blk-barrier.c rq->end_io = end_io;
rq 149 block/blk-barrier.c q->prepare_flush_fn(q, rq);
rq 151 block/blk-barrier.c elv_insert(q, rq, ELEVATOR_INSERT_FRONT);
rq 164 block/blk-barrier.c blkdev_dequeue_request(rq);
rq 165 block/blk-barrier.c q->orig_bar_rq = rq;
rq 166 block/blk-barrier.c rq = &q->bar_rq;
rq 167 block/blk-barrier.c blk_rq_init(q, rq);
rq 169 block/blk-barrier.c rq->cmd_flags |= REQ_RW;
rq 171 block/blk-barrier.c rq->cmd_flags |= REQ_FUA;
rq 172 block/blk-barrier.c init_request_from_bio(rq, q->orig_bar_rq->bio);
rq 173 block/blk-barrier.c rq->end_io = bar_end_io;
rq 184 block/blk-barrier.c if ((q->ordered & QUEUE_ORDERED_POSTFLUSH) && !blk_empty_barrier(rq))
rq 189 block/blk-barrier.c elv_insert(q, rq, ELEVATOR_INSERT_FRONT);
rq 193 block/blk-barrier.c rq = &q->pre_flush_rq;
rq 200 block/blk-barrier.c rq = NULL;
rq 202 block/blk-barrier.c return rq;
rq 207 block/blk-barrier.c struct request *rq = *rqp;
rq 208 block/blk-barrier.c const int is_barrier = blk_fs_request(rq) && blk_barrier_rq(rq);
rq 215 block/blk-barrier.c *rqp = start_ordered(q, rq);
rq 222 block/blk-barrier.c blkdev_dequeue_request(rq);
rq 223 block/blk-barrier.c if (__blk_end_request(rq, -EOPNOTSUPP,
rq 224 block/blk-barrier.c blk_rq_bytes(rq)))
rq 236 block/blk-barrier.c if (!blk_fs_request(rq) &&
rq 237 block/blk-barrier.c rq != &q->pre_flush_rq && rq != &q->post_flush_rq)
rq 242 block/blk-barrier.c if (is_barrier && rq != &q->bar_rq)
rq 246 block/blk-barrier.c WARN_ON(blk_ordered_req_seq(rq) < blk_ordered_cur_seq(q));
rq 247 block/blk-barrier.c if (blk_ordered_req_seq(rq) > blk_ordered_cur_seq(q))
rq 54 block/blk-core.c int rw = rq_data_dir(rq);
rq 57 block/blk-core.c if (!blk_fs_request(rq) || !rq->rq_disk)
rq 61 block/blk-core.c part = disk_map_sector_rcu(rq->rq_disk, rq->sector);
rq 110 block/blk-core.c memset(rq, 0, sizeof(*rq));
rq 112 block/blk-core.c INIT_LIST_HEAD(&rq->queuelist);
rq 113 block/blk-core.c INIT_LIST_HEAD(&rq->timeout_list);
rq 114 block/blk-core.c rq->cpu = -1;
rq 115 block/blk-core.c rq->q = q;
rq 116 block/blk-core.c rq->sector = rq->hard_sector = (sector_t) -1;
rq 117 block/blk-core.c INIT_HLIST_NODE(&rq->hash);
rq 118 block/blk-core.c RB_CLEAR_NODE(&rq->rb_node);
rq 119 block/blk-core.c rq->cmd = rq->__cmd;
rq 120 block/blk-core.c rq->tag = -1;
rq 121 block/blk-core.c rq->ref_count = 1;
rq 128 block/blk-core.c struct request_queue *q = rq->q;
rq 130 block/blk-core.c if (&q->bar_rq != rq) {
rq 166 block/blk-core.c rq->rq_disk ? rq->rq_disk->disk_name : "?", rq->cmd_type,
rq 167 block/blk-core.c rq->cmd_flags);
rq 170 block/blk-core.c (unsigned long long)rq->sector,
rq 171 block/blk-core.c rq->nr_sectors,
rq 172 block/blk-core.c rq->current_nr_sectors);
rq 174 block/blk-core.c rq->bio, rq->biotail,
rq 175 block/blk-core.c rq->buffer, rq->data,
rq 176 block/blk-core.c rq->data_len);
rq 178 block/blk-core.c if (blk_pc_request(rq)) {
rq 181 block/blk-core.c printk("%02x ", rq->cmd[bit]);
rq 297 block/blk-core.c q->rq.count[READ] + q->rq.count[WRITE]);
rq 307 block/blk-core.c q->rq.count[READ] + q->rq.count[WRITE]);
rq 319 block/blk-core.c q->rq.count[READ] + q->rq.count[WRITE]);
rq 460 block/blk-core.c struct request_list *rl = &q->rq;
rq 618 block/blk-core.c if (rq->cmd_flags & REQ_ELVPRIV)
rq 619 block/blk-core.c elv_put_request(q, rq);
rq 620 block/blk-core.c mempool_free(rq, q->rq.rq_pool);
rq 626 block/blk-core.c struct request *rq = mempool_alloc(q->rq.rq_pool, gfp_mask);
rq 628 block/blk-core.c if (!rq)
rq 631 block/blk-core.c blk_rq_init(q, rq);
rq 633 block/blk-core.c rq->cmd_flags = rw | REQ_ALLOCED;
rq 636 block/blk-core.c if (unlikely(elv_set_request(q, rq, gfp_mask))) {
rq 637 block/blk-core.c mempool_free(rq, q->rq.rq_pool);
rq 640 block/blk-core.c rq->cmd_flags |= REQ_ELVPRIV;
rq 643 block/blk-core.c return rq;
rq 682 block/blk-core.c struct request_list *rl = &q->rq;
rq 701 block/blk-core.c struct request_list *rl = &q->rq;
rq 722 block/blk-core.c struct request *rq = NULL;
rq 723 block/blk-core.c struct request_list *rl = &q->rq;
rq 776 block/blk-core.c rq = blk_alloc_request(q, rw_flags, priv, gfp_mask);
rq 777 block/blk-core.c if (unlikely(!rq)) {
rq 813 block/blk-core.c return rq;
rq 826 block/blk-core.c struct request *rq;
rq 828 block/blk-core.c rq = get_request(q, rw_flags, bio, GFP_NOIO);
rq 829 block/blk-core.c while (!rq) {
rq 832 block/blk-core.c struct request_list *rl = &q->rq;
rq 855 block/blk-core.c rq = get_request(q, rw_flags, bio, GFP_NOIO);
rq 858 block/blk-core.c return rq;
rq 863 block/blk-core.c struct request *rq;
rq 869 block/blk-core.c rq = get_request_wait(q, rw, NULL);
rq 871 block/blk-core.c rq = get_request(q, rw, NULL, gfp_mask);
rq 872 block/blk-core.c if (!rq)
rq 877 block/blk-core.c return rq;
rq 914 block/blk-core.c blk_delete_timer(rq);
rq 915 block/blk-core.c blk_clear_rq_complete(rq);
rq 916 block/blk-core.c blk_add_trace_rq(q, rq, BLK_TA_REQUEUE);
rq 918 block/blk-core.c if (blk_rq_tagged(rq))
rq 919 block/blk-core.c blk_queue_end_tag(q, rq);
rq 921 block/blk-core.c elv_requeue_request(q, rq);
rq 955 block/blk-core.c rq->cmd_type = REQ_TYPE_SPECIAL;
rq 956 block/blk-core.c rq->cmd_flags |= REQ_SOFTBARRIER;
rq 958 block/blk-core.c rq->special = data;
rq 965 block/blk-core.c if (blk_rq_tagged(rq))
rq 966 block/blk-core.c blk_queue_end_tag(q, rq);
rq 968 block/blk-core.c drive_stat_acct(rq, 1);
rq 969 block/blk-core.c __elv_add_request(q, rq, where, 0);
rq 1558 block/blk-core.c if (rq->nr_sectors > q->max_sectors ||
rq 1559 block/blk-core.c rq->data_len > q->max_hw_sectors << 9) {
rq 1570 block/blk-core.c blk_recalc_rq_segments(rq);
rq 1571 block/blk-core.c if (rq->nr_phys_segments > q->max_phys_segments ||
rq 1572 block/blk-core.c rq->nr_phys_segments > q->max_hw_segments) {
rq 1590 block/blk-core.c if (blk_rq_check_limits(q, rq))
rq 1594 block/blk-core.c if (rq->rq_disk && rq->rq_disk->part0.make_it_fail &&
rq 1595 block/blk-core.c should_fail(&fail_make_request, blk_rq_bytes(rq)))
rq 1605 block/blk-core.c BUG_ON(blk_queued_rq(rq));
rq 1607 block/blk-core.c drive_stat_acct(rq, 1);
rq 1608 block/blk-core.c __elv_add_request(q, rq, ELEVATOR_INSERT_BACK, 0);
rq 1799 block/blk-core.c if (blk_fs_request(rq))
rq 1800 block/blk-core.c return rq->hard_nr_sectors << 9;
rq 1802 block/blk-core.c return rq->data_len;
rq 1812 block/blk-core.c if (blk_fs_request(rq))
rq 1813 block/blk-core.c return rq->current_nr_sectors << 9;
rq 1815 block/blk-core.c if (rq->bio)
rq 1816 block/blk-core.c return rq->bio->bi_size;
rq 1818 block/blk-core.c return rq->data_len;
rq 1852 block/blk-core.c if (rq->bio) {
rq 1853 block/blk-core.c if (__end_that_request_first(rq, error, nr_bytes))
rq 1857 block/blk-core.c if (blk_bidi_rq(rq) &&
rq 1858 block/blk-core.c __end_that_request_first(rq->next_rq, error, bidi_bytes))
rq 1888 block/blk-core.c struct request_queue *q = rq->q;
rq 1891 block/blk-core.c if (end_that_request_data(rq, error, nr_bytes, bidi_bytes))
rq 1895 block/blk-core.c if (drv_callback && drv_callback(rq))
rq 1898 block/blk-core.c add_disk_randomness(rq->rq_disk);
rq 1901 block/blk-core.c end_that_request_last(rq, error);
rq 1923 block/blk-core.c return blk_end_io(rq, error, nr_bytes, 0, NULL);
rq 1942 block/blk-core.c if (rq->bio && __end_that_request_first(rq, error, nr_bytes))
rq 1945 block/blk-core.c add_disk_randomness(rq->rq_disk);
rq 1947 block/blk-core.c end_that_request_last(rq, error);
rq 1970 block/blk-core.c return blk_end_io(rq, error, nr_bytes, bidi_bytes, NULL);
rq 1991 block/blk-core.c if (!end_that_request_data(rq, error, nr_bytes, 0)) {
rq 1998 block/blk-core.c rq->nr_sectors = rq->hard_nr_sectors = 0;
rq 1999 block/blk-core.c rq->current_nr_sectors = rq->hard_cur_sectors = 0;
rq 2033 block/blk-core.c return blk_end_io(rq, error, nr_bytes, 0, drv_callback);
rq 2042 block/blk-core.c rq->cmd_flags |= (bio->bi_rw & 3);
rq 2045 block/blk-core.c rq->nr_phys_segments = bio_phys_segments(q, bio);
rq 2046 block/blk-core.c rq->buffer = bio_data(bio);
rq 2048 block/blk-core.c rq->current_nr_sectors = bio_cur_sectors(bio);
rq 2049 block/blk-core.c rq->hard_cur_sectors = rq->current_nr_sectors;
rq 2050 block/blk-core.c rq->hard_nr_sectors = rq->nr_sectors = bio_sectors(bio);
rq 2051 block/blk-core.c rq->data_len = bio->bi_size;
rq 2053 block/blk-core.c rq->bio = rq->biotail = bio;
rq 2056 block/blk-core.c rq->rq_disk = bio->bi_bdev->bd_disk;
rq 23 block/blk-exec.c struct completion *waiting = rq->end_io_data;
rq 25 block/blk-exec.c rq->end_io_data = NULL;
rq 26 block/blk-exec.c __blk_put_request(rq->q, rq);
rq 53 block/blk-exec.c rq->rq_disk = bd_disk;
rq 54 block/blk-exec.c rq->cmd_flags |= REQ_NOMERGE;
rq 55 block/blk-exec.c rq->end_io = done;
rq 58 block/blk-exec.c __elv_add_request(q, rq, where, 1);
rq 61 block/blk-exec.c if (blk_pm_resume_request(rq))
rq 89 block/blk-exec.c rq->ref_count++;
rq 91 block/blk-exec.c if (!rq->sense) {
rq 93 block/blk-exec.c rq->sense = sense;
rq 94 block/blk-exec.c rq->sense_len = 0;
rq 97 block/blk-exec.c rq->end_io_data = &wait;
rq 98 block/blk-exec.c blk_execute_rq_nowait(q, bd_disk, rq, at_head, blk_end_sync_rq);
rq 101 block/blk-exec.c if (rq->errors)
rq 48 block/blk-integrity.c rq_for_each_integrity_segment(iv, rq, iter) {
rq 80 block/blk-integrity.c rq_for_each_integrity_segment(iv, rq, iter) {
rq 15 block/blk-map.c if (!rq->bio)
rq 16 block/blk-map.c blk_rq_bio_prep(q, rq, bio);
rq 17 block/blk-map.c else if (!ll_back_merge_fn(q, rq, bio))
rq 20 block/blk-map.c rq->biotail->bi_next = bio;
rq 21 block/blk-map.c rq->biotail = bio;
rq 23 block/blk-map.c rq->data_len += bio->bi_size;
rq 51 block/blk-map.c reading = rq_data_dir(rq) == READ;
rq 78 block/blk-map.c ret = blk_rq_append_bio(q, rq, bio);
rq 124 block/blk-map.c if (!map_data || rq_data_dir(rq) != READ)
rq 145 block/blk-map.c ret = __blk_rq_map_user(q, rq, map_data, ubuf, map_len,
rq 150 block/blk-map.c bio = rq->bio;
rq 156 block/blk-map.c rq->cmd_flags |= REQ_COPY_USER;
rq 158 block/blk-map.c rq->buffer = rq->data = NULL;
rq 162 block/blk-map.c rq->bio = NULL;
rq 195 block/blk-map.c int i, read = rq_data_dir(rq) == READ;
rq 226 block/blk-map.c rq->cmd_flags |= REQ_COPY_USER;
rq 230 block/blk-map.c blk_rq_bio_prep(q, rq, bio);
rq 231 block/blk-map.c rq->buffer = rq->data = NULL;
rq 283 block/blk-map.c int reading = rq_data_dir(rq) == READ;
rq 301 block/blk-map.c if (rq_data_dir(rq) == WRITE)
rq 305 block/blk-map.c rq->cmd_flags |= REQ_COPY_USER;
rq 307 block/blk-map.c blk_rq_bio_prep(q, rq, bio);
rq 308 block/blk-map.c blk_queue_bounce(q, &rq->bio);
rq 309 block/blk-map.c rq->buffer = rq->data = NULL;
rq 14 block/blk-merge.c if (blk_fs_request(rq) || blk_discard_rq(rq)) {
rq 15 block/blk-merge.c rq->hard_sector += nsect;
rq 16 block/blk-merge.c rq->hard_nr_sectors -= nsect;
rq 21 block/blk-merge.c if ((rq->nr_sectors >= rq->hard_nr_sectors) &&
rq 22 block/blk-merge.c (rq->sector <= rq->hard_sector)) {
rq 23 block/blk-merge.c rq->sector = rq->hard_sector;
rq 24 block/blk-merge.c rq->nr_sectors = rq->hard_nr_sectors;
rq 25 block/blk-merge.c rq->hard_cur_sectors = bio_cur_sectors(rq->bio);
rq 26 block/blk-merge.c rq->current_nr_sectors = rq->hard_cur_sectors;
rq 27 block/blk-merge.c rq->buffer = bio_data(rq->bio);
rq 34 block/blk-merge.c if (rq->nr_sectors < rq->current_nr_sectors) {
rq 36 block/blk-merge.c rq->nr_sectors = rq->current_nr_sectors;
rq 50 block/blk-merge.c struct request_queue *q = rq->q;
rq 52 block/blk-merge.c if (!rq->bio)
rq 58 block/blk-merge.c rq_for_each_segment(bv, rq, iter) {
rq 86 block/blk-merge.c rq->nr_phys_segments = nr_phys_segs;
rq 91 block/blk-merge.c struct request rq;
rq 93 block/blk-merge.c rq.q = q;
rq 94 block/blk-merge.c rq.bio = rq.biotail = bio;
rq 96 block/blk-merge.c blk_recalc_rq_segments(&rq);
rq 98 block/blk-merge.c bio->bi_phys_segments = rq.nr_phys_segments;
rq 148 block/blk-merge.c rq_for_each_segment(bvec, rq, iter) {
rq 187 block/blk-merge.c if (unlikely(rq->cmd_flags & REQ_COPY_USER) &&
rq 188 block/blk-merge.c (rq->data_len & q->dma_pad_mask)) {
rq 189 block/blk-merge.c unsigned int pad_len = (q->dma_pad_mask & ~rq->data_len) + 1;
rq 192 block/blk-merge.c rq->extra_len += pad_len;
rq 195 block/blk-merge.c if (q->dma_drain_size && q->dma_drain_needed(rq)) {
rq 196 block/blk-merge.c if (rq->cmd_flags & REQ_RW)
rq 206 block/blk-merge.c rq->extra_len += q->dma_drain_size;
rq 412 block/blk-merge.c struct request *next = elv_latter_request(q, rq);
rq 415 block/blk-merge.c return attempt_merge(q, rq, next);
rq 422 block/blk-merge.c struct request *prev = elv_former_request(q, rq);
rq 425 block/blk-merge.c return attempt_merge(q, prev, rq);
rq 30 block/blk-softirq.c struct request *rq;
rq 32 block/blk-softirq.c rq = list_entry(local_list.next, struct request, csd.list);
rq 33 block/blk-softirq.c list_del_init(&rq->csd.list);
rq 34 block/blk-softirq.c rq->q->softirq_done_fn(rq);
rq 41 block/blk-softirq.c struct request *rq = data;
rq 47 block/blk-softirq.c list_add_tail(&rq->csd.list, list);
rq 49 block/blk-softirq.c if (list->next == &rq->csd.list)
rq 61 block/blk-softirq.c struct call_single_data *data = &rq->csd;
rq 64 block/blk-softirq.c data->info = rq;
rq 41 block/blk-sysfs.c struct request_list *rl = &q->rq;
rq 305 block/blk-sysfs.c struct request_list *rl = &q->rq;
rq 285 block/blk-tag.c int tag = rq->tag;
rq 296 block/blk-tag.c list_del_init(&rq->queuelist);
rq 297 block/blk-tag.c rq->cmd_flags &= ~REQ_QUEUED;
rq 298 block/blk-tag.c rq->tag = -1;
rq 343 block/blk-tag.c if (unlikely((rq->cmd_flags & REQ_QUEUED))) {
rq 346 block/blk-tag.c __func__, rq,
rq 347 block/blk-tag.c rq->rq_disk ? rq->rq_disk->disk_name : "?", rq->tag);
rq 359 block/blk-tag.c if (rq_is_sync(rq))
rq 375 block/blk-tag.c rq->cmd_flags |= REQ_QUEUED;
rq 376 block/blk-tag.c rq->tag = tag;
rq 377 block/blk-tag.c bqt->tag_index[tag] = rq;
rq 378 block/blk-tag.c blkdev_dequeue_request(rq);
rq 379 block/blk-tag.c list_add(&rq->queuelist, &q->tag_busy_list);
rq 122 block/blk-timeout.c struct request *rq, *tmp;
rq 126 block/blk-timeout.c list_for_each_entry_safe(rq, tmp, &q->timeout_list, timeout_list) {
rq 127 block/blk-timeout.c if (time_after_eq(jiffies, rq->deadline)) {
rq 128 block/blk-timeout.c list_del_init(&rq->timeout_list);
rq 133 block/blk-timeout.c if (blk_mark_rq_complete(rq))
rq 135 block/blk-timeout.c blk_rq_timed_out(rq);
rq 138 block/blk-timeout.c next = rq->deadline;
rq 140 block/blk-timeout.c } else if (time_after(next, rq->deadline))
rq 141 block/blk-timeout.c next = rq->deadline;
rq 226 block/blk-timeout.c struct request *rq, *tmp;
rq 232 block/blk-timeout.c list_for_each_entry_safe(rq, tmp, &q->timeout_list, timeout_list)
rq 233 block/blk-timeout.c blk_abort_request(rq);
rq 37 block/blk.h return test_and_set_bit(REQ_ATOM_COMPLETE, &rq->atomic_flags);
rq 42 block/blk.h clear_bit(REQ_ATOM_COMPLETE, &rq->atomic_flags);
rq 82 block/bsg.c struct request *rq;
rq 179 block/bsg.c rq->cmd = kzalloc(hdr->request_len, GFP_KERNEL);
rq 180 block/bsg.c if (!rq->cmd)
rq 184 block/bsg.c if (copy_from_user(rq->cmd, (void *)(unsigned long)hdr->request,
rq 189 block/bsg.c if (blk_verify_command(&q->cmd_filter, rq->cmd, has_write_perm))
rq 197 block/bsg.c rq->cmd_len = hdr->request_len;
rq 198 block/bsg.c rq->cmd_type = REQ_TYPE_BLOCK_PC;
rq 200 block/bsg.c rq->timeout = (hdr->timeout * HZ) / 1000;
rq 201 block/bsg.c if (!rq->timeout)
rq 202 block/bsg.c rq->timeout = q->sg_timeout;
rq 203 block/bsg.c if (!rq->timeout)
rq 204 block/bsg.c rq->timeout = BLK_DEFAULT_SG_TIMEOUT;
rq 248 block/bsg.c struct request *rq, *next_rq = NULL;
rq 264 block/bsg.c rq = blk_get_request(q, rw, GFP_KERNEL);
rq 265 block/bsg.c if (!rq)
rq 267 block/bsg.c ret = blk_fill_sgv4_hdr_rq(q, rq, hdr, bd, has_write_perm);
rq 282 block/bsg.c rq->next_rq = next_rq;
rq 283 block/bsg.c next_rq->cmd_type = rq->cmd_type;
rq 302 block/bsg.c ret = blk_rq_map_user(q, rq, NULL, dxferp, dxfer_len,
rq 307 block/bsg.c return rq;
rq 309 block/bsg.c if (rq->cmd != rq->__cmd)
rq 310 block/bsg.c kfree(rq->cmd);
rq 311 block/bsg.c blk_put_request(rq);
rq 325 block/bsg.c struct bsg_command *bc = rq->end_io_data;
rq 330 block/bsg.c bd->name, rq, bc, bc->bio, uptodate);
rq 349 block/bsg.c rq->sense = bc->sense;
rq 350 block/bsg.c rq->sense_len = 0;
rq 355 block/bsg.c bc->rq = rq;
rq 356 block/bsg.c bc->bio = rq->bio;
rq 357 block/bsg.c if (rq->next_rq)
rq 358 block/bsg.c bc->bidi_bio = rq->next_rq->bio;
rq 364 block/bsg.c dprintk("%s: queueing rq %p, bc %p\n", bd->name, rq, bc);
rq 366 block/bsg.c rq->end_io_data = bc;
rq 367 block/bsg.c blk_execute_rq_nowait(q, NULL, rq, 1, bsg_rq_end_io);
rq 420 block/bsg.c dprintk("rq %p bio %p %u\n", rq, bio, rq->errors);
rq 424 block/bsg.c hdr->device_status = status_byte(rq->errors);
rq 425 block/bsg.c hdr->transport_status = host_byte(rq->errors);
rq 426 block/bsg.c hdr->driver_status = driver_byte(rq->errors);
rq 432 block/bsg.c if (rq->sense_len && hdr->response) {
rq 434 block/bsg.c rq->sense_len);
rq 437 block/bsg.c rq->sense, len);
rq 444 block/bsg.c if (rq->next_rq) {
rq 445 block/bsg.c hdr->dout_resid = rq->data_len;
rq 446 block/bsg.c hdr->din_resid = rq->next_rq->data_len;
rq 448 block/bsg.c blk_put_request(rq->next_rq);
rq 449 block/bsg.c } else if (rq_data_dir(rq) == READ)
rq 450 block/bsg.c hdr->din_resid = rq->data_len;
rq 452 block/bsg.c hdr->dout_resid = rq->data_len;
rq 460 block/bsg.c if (!ret && rq->errors < 0)
rq 461 block/bsg.c ret = rq->errors;
rq 464 block/bsg.c if (rq->cmd != rq->__cmd)
rq 465 block/bsg.c kfree(rq->cmd);
rq 466 block/bsg.c blk_put_request(rq);
rq 509 block/bsg.c tret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio,
rq 544 block/bsg.c ret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio,
rq 607 block/bsg.c struct request *rq;
rq 614 block/bsg.c rq = NULL;
rq 635 block/bsg.c rq = bsg_map_hdr(bd, &bc->hdr, has_write_perm);
rq 636 block/bsg.c if (IS_ERR(rq)) {
rq 637 block/bsg.c ret = PTR_ERR(rq);
rq 638 block/bsg.c rq = NULL;
rq 642 block/bsg.c bsg_add_command(bd, q, bc, rq);
rq 644 block/bsg.c rq = NULL;
rq 763 block/bsg.c ret = blk_get_queue(rq);
rq 769 block/bsg.c blk_put_queue(rq);
rq 773 block/bsg.c bd->queue = rq;
rq 781 block/bsg.c strncpy(bd->name, rq->bsg_dev.class_dev->bus_id, sizeof(bd->name) - 1);
rq 919 block/bsg.c struct request *rq;
rq 926 block/bsg.c rq = bsg_map_hdr(bd, &hdr, file->f_mode & FMODE_WRITE);
rq 927 block/bsg.c if (IS_ERR(rq))
rq 928 block/bsg.c return PTR_ERR(rq);
rq 930 block/bsg.c bio = rq->bio;
rq 931 block/bsg.c if (rq->next_rq)
rq 932 block/bsg.c bidi_bio = rq->next_rq->bio;
rq 933 block/bsg.c blk_execute_rq(bd->queue, NULL, rq, 0);
rq 934 block/bsg.c ret = blk_complete_sgv4_hdr_rq(rq, &hdr, bio, bidi_bio);
rq 45 block/cfq-iosched.c ((struct cfq_io_context *) (rq)->elevator_private)
rq 46 block/cfq-iosched.c #define RQ_CFQQ(rq) (struct cfq_queue *) ((rq)->elevator_private2)
rq 591 block/cfq-iosched.c struct cfq_queue *cfqq = RQ_CFQQ(rq);
rq 593 block/cfq-iosched.c const int sync = rq_is_sync(rq);
rq 598 block/cfq-iosched.c elv_rb_del(&cfqq->sort_list, rq);
rq 606 block/cfq-iosched.c struct cfq_queue *cfqq = RQ_CFQQ(rq);
rq 610 block/cfq-iosched.c cfqq->queued[rq_is_sync(rq)]++;
rq 616 block/cfq-iosched.c while ((__alias = elv_rb_add(&cfqq->sort_list, rq)) != NULL)
rq 625 block/cfq-iosched.c cfqq->next_rq = cfq_choose_req(cfqd, cfqq->next_rq, rq);
rq 631 block/cfq-iosched.c elv_rb_del(&cfqq->sort_list, rq);
rq 632 block/cfq-iosched.c cfqq->queued[rq_is_sync(rq)]--;
rq 633 block/cfq-iosched.c cfq_add_rq_rb(rq);
rq 662 block/cfq-iosched.c cfq_log_cfqq(cfqd, RQ_CFQQ(rq), "activate rq, drv=%d",
rq 665 block/cfq-iosched.c cfqd->last_position = rq->hard_sector + rq->hard_nr_sectors;
rq 674 block/cfq-iosched.c cfq_log_cfqq(cfqd, RQ_CFQQ(rq), "deactivate rq, drv=%d",
rq 680 block/cfq-iosched.c struct cfq_queue *cfqq = RQ_CFQQ(rq);
rq 682 block/cfq-iosched.c if (cfqq->next_rq == rq)
rq 683 block/cfq-iosched.c cfqq->next_rq = cfq_find_next_rq(cfqq->cfqd, cfqq, rq);
rq 685 block/cfq-iosched.c list_del_init(&rq->queuelist);
rq 686 block/cfq-iosched.c cfq_del_rq_rb(rq);
rq 689 block/cfq-iosched.c if (rq_is_meta(rq)) {
rq 727 block/cfq-iosched.c if (!list_empty(&rq->queuelist) && !list_empty(&next->queuelist) &&
rq 728 block/cfq-iosched.c time_before(next->start_time, rq->start_time))
rq 729 block/cfq-iosched.c list_move(&rq->queuelist, &next->queuelist);
rq 744 block/cfq-iosched.c if (cfq_bio_sync(bio) && !rq_is_sync(rq))
rq 756 block/cfq-iosched.c if (cfqq == RQ_CFQQ(rq))
rq 846 block/cfq-iosched.c if (rq->sector >= cfqd->last_position)
rq 847 block/cfq-iosched.c return rq->sector - cfqd->last_position;
rq 849 block/cfq-iosched.c return cfqd->last_position - rq->sector;
rq 859 block/cfq-iosched.c return cfq_dist_from_last(cfqd, rq) <= cic->seek_mean;
rq 940 block/cfq-iosched.c struct cfq_queue *cfqq = RQ_CFQQ(rq);
rq 944 block/cfq-iosched.c cfq_remove_request(rq);
rq 946 block/cfq-iosched.c elv_dispatch_sort(q, rq);
rq 958 block/cfq-iosched.c struct request *rq;
rq 970 block/cfq-iosched.c rq = rq_entry_fifo(cfqq->fifo.next);
rq 972 block/cfq-iosched.c if (time_before(jiffies, rq->start_time + cfqd->cfq_fifo_expire[fifo]))
rq 973 block/cfq-iosched.c rq = NULL;
rq 975 block/cfq-iosched.c cfq_log_cfqq(cfqd, cfqq, "fifo=%p", rq);
rq 976 block/cfq-iosched.c return rq;
rq 1046 block/cfq-iosched.c struct request *rq;
rq 1051 block/cfq-iosched.c rq = cfq_check_fifo(cfqq);
rq 1052 block/cfq-iosched.c if (rq == NULL)
rq 1053 block/cfq-iosched.c rq = cfqq->next_rq;
rq 1058 block/cfq-iosched.c cfq_dispatch_insert(cfqd->queue, rq);
rq 1063 block/cfq-iosched.c atomic_inc(&RQ_CIC(rq)->ioc->refcount);
rq 1064 block/cfq-iosched.c cfqd->active_cic = RQ_CIC(rq);
rq 1705 block/cfq-iosched.c if (cic->last_request_pos < rq->sector)
rq 1706 block/cfq-iosched.c sdist = rq->sector - cic->last_request_pos;
rq 1708 block/cfq-iosched.c sdist = cic->last_request_pos - rq->sector;
rq 1790 block/cfq-iosched.c if (rq_is_sync(rq) && !cfq_cfqq_sync(cfqq))
rq 1797 block/cfq-iosched.c if (rq_is_meta(rq) && !cfqq->meta_pending)
rq 1807 block/cfq-iosched.c if (cfq_rq_close(cfqd, rq))
rq 1842 block/cfq-iosched.c struct cfq_io_context *cic = RQ_CIC(rq);
rq 1845 block/cfq-iosched.c if (rq_is_meta(rq))
rq 1849 block/cfq-iosched.c cfq_update_io_seektime(cfqd, cic, rq);
rq 1852 block/cfq-iosched.c cic->last_request_pos = rq->sector + rq->nr_sectors;
rq 1865 block/cfq-iosched.c } else if (cfq_should_preempt(cfqd, cfqq, rq)) {
rq 1880 block/cfq-iosched.c struct cfq_queue *cfqq = RQ_CFQQ(rq);
rq 1883 block/cfq-iosched.c cfq_init_prio_data(cfqq, RQ_CIC(rq)->ioc);
rq 1885 block/cfq-iosched.c cfq_add_rq_rb(rq);
rq 1887 block/cfq-iosched.c list_add_tail(&rq->queuelist, &cfqq->fifo);
rq 1889 block/cfq-iosched.c cfq_rq_enqueued(cfqd, cfqq, rq);
rq 1919 block/cfq-iosched.c struct cfq_queue *cfqq = RQ_CFQQ(rq);
rq 1921 block/cfq-iosched.c const int sync = rq_is_sync(rq);
rq 1941 block/cfq-iosched.c RQ_CIC(rq)->last_end_request = now;
rq 2032 block/cfq-iosched.c struct cfq_queue *cfqq = RQ_CFQQ(rq);
rq 2035 block/cfq-iosched.c const int rw = rq_data_dir(rq);
rq 2040 block/cfq-iosched.c put_io_context(RQ_CIC(rq)->ioc);
rq 2042 block/cfq-iosched.c rq->elevator_private = NULL;
rq 2043 block/cfq-iosched.c rq->elevator_private2 = NULL;
rq 2057 block/cfq-iosched.c const int rw = rq_data_dir(rq);
rq 2058 block/cfq-iosched.c const int is_sync = rq_is_sync(rq);
rq 2087 block/cfq-iosched.c rq->elevator_private = cic;
rq 2088 block/cfq-iosched.c rq->elevator_private2 = cfqq;
rq 59 block/deadline-iosched.c return &dd->sort_list[rq_data_dir(rq)];
rq 68 block/deadline-iosched.c struct rb_node *node = rb_next(&rq->rb_node);
rq 79 block/deadline-iosched.c struct rb_root *root = deadline_rb_root(dd, rq);
rq 82 block/deadline-iosched.c while (unlikely(__alias = elv_rb_add(root, rq)))
rq 89 block/deadline-iosched.c const int data_dir = rq_data_dir(rq);
rq 91 block/deadline-iosched.c if (dd->next_rq[data_dir] == rq)
rq 92 block/deadline-iosched.c dd->next_rq[data_dir] = deadline_latter_request(rq);
rq 94 block/deadline-iosched.c elv_rb_del(deadline_rb_root(dd, rq), rq);
rq 104 block/deadline-iosched.c const int data_dir = rq_data_dir(rq);
rq 106 block/deadline-iosched.c deadline_add_rq_rb(dd, rq);
rq 111 block/deadline-iosched.c rq_set_fifo_time(rq, jiffies + dd->fifo_expire[data_dir]);
rq 112 block/deadline-iosched.c list_add_tail(&rq->queuelist, &dd->fifo_list[data_dir]);
rq 122 block/deadline-iosched.c rq_fifo_clear(rq);
rq 123 block/deadline-iosched.c deadline_del_rq_rb(dd, rq);
rq 197 block/deadline-iosched.c struct request_queue *q = rq->q;
rq 199 block/deadline-iosched.c deadline_remove_request(q, rq);
rq 200 block/deadline-iosched.c elv_dispatch_add_tail(q, rq);
rq 209 block/deadline-iosched.c const int data_dir = rq_data_dir(rq);
rq 213 block/deadline-iosched.c dd->next_rq[data_dir] = deadline_latter_request(rq);
rq 215 block/deadline-iosched.c dd->last_sector = rq_end_sector(rq);
rq 221 block/deadline-iosched.c deadline_move_to_dispatch(dd, rq);
rq 230 block/deadline-iosched.c struct request *rq = rq_entry_fifo(dd->fifo_list[ddir].next);
rq 235 block/deadline-iosched.c if (time_after(jiffies, rq_fifo_time(rq)))
rq 250 block/deadline-iosched.c struct request *rq;
rq 257 block/deadline-iosched.c rq = dd->next_rq[WRITE];
rq 259 block/deadline-iosched.c rq = dd->next_rq[READ];
rq 261 block/deadline-iosched.c if (rq && dd->batching < dd->fifo_batch)
rq 308 block/deadline-iosched.c rq = rq_entry_fifo(dd->fifo_list[data_dir].next);
rq 314 block/deadline-iosched.c rq = dd->next_rq[data_dir];
rq 324 block/deadline-iosched.c deadline_move_request(dd, rq);
rq 52 block/elevator.c #define rq_hash_key(rq) ((rq)->sector + (rq)->nr_sectors)
rq 53 block/elevator.c #define ELV_ON_HASH(rq) (!hlist_unhashed(&(rq)->hash))
rq 61 block/elevator.c struct request_queue *q = rq->q;
rq 65 block/elevator.c return e->ops->elevator_allow_merge_fn(q, rq, bio);
rq 75 block/elevator.c if (!rq_mergeable(rq))
rq 81 block/elevator.c if (bio_discard(bio) != bio_discard(rq->bio))
rq 87 block/elevator.c if (bio_data_dir(bio) != rq_data_dir(rq))
rq 93 block/elevator.c if (rq->rq_disk != bio->bi_bdev->bd_disk || rq->special)
rq 99 block/elevator.c if (bio_integrity(bio) != blk_integrity_rq(rq))
rq 102 block/elevator.c if (!elv_iosched_allow_merge(rq, bio))
rq 311 block/elevator.c e->ops->elevator_activate_req_fn(q, rq);
rq 319 block/elevator.c e->ops->elevator_deactivate_req_fn(q, rq);
rq 324 block/elevator.c hlist_del_init(&rq->hash);
rq 329 block/elevator.c if (ELV_ON_HASH(rq))
rq 330 block/elevator.c __elv_rqhash_del(rq);
rq 337 block/elevator.c BUG_ON(ELV_ON_HASH(rq));
rq 338 block/elevator.c hlist_add_head(&rq->hash, &e->hash[ELV_HASH_FN(rq_hash_key(rq))]);
rq 343 block/elevator.c __elv_rqhash_del(rq);
rq 344 block/elevator.c elv_rqhash_add(q, rq);
rq 352 block/elevator.c struct request *rq;
rq 354 block/elevator.c hlist_for_each_entry_safe(rq, entry, next, hash_list, hash) {
rq 355 block/elevator.c BUG_ON(!ELV_ON_HASH(rq));
rq 357 block/elevator.c if (unlikely(!rq_mergeable(rq))) {
rq 358 block/elevator.c __elv_rqhash_del(rq);
rq 362 block/elevator.c if (rq_hash_key(rq) == offset)
rq 363 block/elevator.c return rq;
rq 383 block/elevator.c if (rq->sector < __rq->sector)
rq 385 block/elevator.c else if (rq->sector > __rq->sector)
rq 391 block/elevator.c rb_link_node(&rq->rb_node, parent, p);
rq 392 block/elevator.c rb_insert_color(&rq->rb_node, root);
rq 399 block/elevator.c BUG_ON(RB_EMPTY_NODE(&rq->rb_node));
rq 400 block/elevator.c rb_erase(&rq->rb_node, root);
rq 401 block/elevator.c RB_CLEAR_NODE(&rq->rb_node);
rq 408 block/elevator.c struct request *rq;
rq 411 block/elevator.c rq = rb_entry(n, struct request, rb_node);
rq 413 block/elevator.c if (sector < rq->sector)
rq 415 block/elevator.c else if (sector > rq->sector)
rq 418 block/elevator.c return rq;
rq 436 block/elevator.c if (q->last_merge == rq)
rq 439 block/elevator.c elv_rqhash_del(q, rq);
rq 448 block/elevator.c if (blk_discard_rq(rq) != blk_discard_rq(pos))
rq 450 block/elevator.c if (rq_data_dir(rq) != rq_data_dir(pos))
rq 454 block/elevator.c if (rq->sector >= boundary) {
rq 461 block/elevator.c if (rq->sector >= pos->sector)
rq 465 block/elevator.c list_add(&rq->queuelist, entry);
rq 476 block/elevator.c if (q->last_merge == rq)
rq 479 block/elevator.c elv_rqhash_del(q, rq);
rq 483 block/elevator.c q->end_sector = rq_end_sector(rq);
rq 484 block/elevator.c q->boundary_rq = rq;
rq 485 block/elevator.c list_add_tail(&rq->queuelist, &q->queue_head);
rq 529 block/elevator.c e->ops->elevator_merged_fn(q, rq, type);
rq 532 block/elevator.c elv_rqhash_reposition(q, rq);
rq 534 block/elevator.c q->last_merge = rq;
rq 543 block/elevator.c e->ops->elevator_merge_req_fn(q, rq, next);
rq 545 block/elevator.c elv_rqhash_reposition(q, rq);
rq 549 block/elevator.c q->last_merge = rq;
rq 558 block/elevator.c if (blk_account_rq(rq)) {
rq 560 block/elevator.c if (blk_sorted_rq(rq))
rq 561 block/elevator.c elv_deactivate_rq(q, rq);
rq 564 block/elevator.c rq->cmd_flags &= ~REQ_STARTED;
rq 566 block/elevator.c elv_insert(q, rq, ELEVATOR_INSERT_REQUEUE);
rq 589 block/elevator.c blk_add_trace_rq(q, rq, BLK_TA_INSERT);
rq 591 block/elevator.c rq->q = q;
rq 595 block/elevator.c rq->cmd_flags |= REQ_SOFTBARRIER;
rq 597 block/elevator.c list_add(&rq->queuelist, &q->queue_head);
rq 601 block/elevator.c rq->cmd_flags |= REQ_SOFTBARRIER;
rq 603 block/elevator.c list_add_tail(&rq->queuelist, &q->queue_head);
rq 619 block/elevator.c BUG_ON(!blk_fs_request(rq) && !blk_discard_rq(rq));
rq 620 block/elevator.c rq->cmd_flags |= REQ_SORTED;
rq 622 block/elevator.c if (rq_mergeable(rq)) {
rq 623 block/elevator.c elv_rqhash_add(q, rq);
rq 625 block/elevator.c q->last_merge = rq;
rq 633 block/elevator.c q->elevator->ops->elevator_add_req_fn(q, rq);
rq 642 block/elevator.c rq->cmd_flags |= REQ_SOFTBARRIER;
rq 651 block/elevator.c list_add(&rq->queuelist, &q->queue_head);
rq 655 block/elevator.c ordseq = blk_ordered_req_seq(rq);
rq 663 block/elevator.c list_add_tail(&rq->queuelist, pos);
rq 673 block/elevator.c int nrq = q->rq.count[READ] + q->rq.count[WRITE]
rq 685 block/elevator.c rq->cmd_flags |= REQ_ORDERED_COLOR;
rq 687 block/elevator.c if (rq->cmd_flags & (REQ_SOFTBARRIER | REQ_HARDBARRIER)) {
rq 691 block/elevator.c if (blk_barrier_rq(rq))
rq 704 block/elevator.c if (blk_fs_request(rq) || blk_discard_rq(rq)) {
rq 705 block/elevator.c q->end_sector = rq_end_sector(rq);
rq 706 block/elevator.c q->boundary_rq = rq;
rq 708 block/elevator.c } else if (!(rq->cmd_flags & REQ_ELVPRIV) &&
rq 715 block/elevator.c elv_insert(q, rq, where);
rq 725 block/elevator.c __elv_add_request(q, rq, where, plug);
rq 732 block/elevator.c struct request *rq;
rq 736 block/elevator.c rq = list_entry_rq(q->queue_head.next);
rq 737 block/elevator.c if (blk_do_ordered(q, &rq))
rq 738 block/elevator.c return rq;
rq 748 block/elevator.c struct request *rq;
rq 751 block/elevator.c while ((rq = __elv_next_request(q)) != NULL) {
rq 756 block/elevator.c if (blk_empty_barrier(rq)) {
rq 757 block/elevator.c __blk_end_request(rq, 0, blk_rq_bytes(rq));
rq 760 block/elevator.c if (!(rq->cmd_flags & REQ_STARTED)) {
rq 766 block/elevator.c if (blk_sorted_rq(rq))
rq 767 block/elevator.c elv_activate_rq(q, rq);
rq 774 block/elevator.c rq->cmd_flags |= REQ_STARTED;
rq 775 block/elevator.c blk_add_trace_rq(q, rq, BLK_TA_ISSUE);
rq 781 block/elevator.c blk_add_timer(rq);
rq 784 block/elevator.c if (!q->boundary_rq || q->boundary_rq == rq) {
rq 785 block/elevator.c q->end_sector = rq_end_sector(rq);
rq 789 block/elevator.c if (rq->cmd_flags & REQ_DONTPREP)
rq 792 block/elevator.c if (q->dma_drain_size && rq->data_len) {
rq 799 block/elevator.c rq->nr_phys_segments++;
rq 805 block/elevator.c ret = q->prep_rq_fn(q, rq);
rq 815 block/elevator.c if (q->dma_drain_size && rq->data_len &&
rq 816 block/elevator.c !(rq->cmd_flags & REQ_DONTPREP)) {
rq 821 block/elevator.c --rq->nr_phys_segments;
rq 824 block/elevator.c rq = NULL;
rq 827 block/elevator.c rq->cmd_flags |= REQ_QUIET;
rq 828 block/elevator.c __blk_end_request(rq, -EIO, blk_rq_bytes(rq));
rq 835 block/elevator.c return rq;
rq 841 block/elevator.c BUG_ON(list_empty(&rq->queuelist));
rq 842 block/elevator.c BUG_ON(ELV_ON_HASH(rq));
rq 844 block/elevator.c list_del_init(&rq->queuelist);
rq 851 block/elevator.c if (blk_account_rq(rq))
rq 875 block/elevator.c return e->ops->elevator_latter_req_fn(q, rq);
rq 884 block/elevator.c return e->ops->elevator_former_req_fn(q, rq);
rq 893 block/elevator.c return e->ops->elevator_set_req_fn(q, rq, gfp_mask);
rq 895 block/elevator.c rq->elevator_private = NULL;
rq 904 block/elevator.c e->ops->elevator_put_req_fn(rq);
rq 919 block/elevator.c struct request *rq;
rq 922 block/elevator.c rq = list_entry_rq(q->queue_head.next);
rq 923 block/elevator.c rq->cmd_flags |= REQ_QUIET;
rq 924 block/elevator.c blk_add_trace_rq(q, rq, BLK_TA_ABORT);
rq 925 block/elevator.c __blk_end_request(rq, -EIO, blk_rq_bytes(rq));
rq 937 block/elevator.c if (blk_account_rq(rq)) {
rq 939 block/elevator.c if (blk_sorted_rq(rq) && e->ops->elevator_completed_req_fn)
rq 940 block/elevator.c e->ops->elevator_completed_req_fn(q, rq);
rq 1111 block/elevator.c while (q->rq.elvpriv) {
rq 1219 block/elevator.c struct rb_node *rbprev = rb_prev(&rq->rb_node);
rq 1231 block/elevator.c struct rb_node *rbnext = rb_next(&rq->rb_node);
rq 25 block/noop-iosched.c struct request *rq;
rq 26 block/noop-iosched.c rq = list_entry(nd->queue.next, struct request, queuelist);
rq 27 block/noop-iosched.c list_del_init(&rq->queuelist);
rq 28 block/noop-iosched.c elv_dispatch_sort(q, rq);
rq 38 block/noop-iosched.c list_add_tail(&rq->queuelist, &nd->queue);
rq 53 block/noop-iosched.c if (rq->queuelist.prev == &nd->queue)
rq 55 block/noop-iosched.c return list_entry(rq->queuelist.prev, struct request, queuelist);
rq 63 block/noop-iosched.c if (rq->queuelist.next == &nd->queue)
rq 65 block/noop-iosched.c return list_entry(rq->queuelist.next, struct request, queuelist);
rq 195 block/scsi_ioctl.c if (copy_from_user(rq->cmd, hdr->cmdp, hdr->cmd_len))
rq 197 block/scsi_ioctl.c if (blk_verify_command(&q->cmd_filter, rq->cmd,
rq 204 block/scsi_ioctl.c rq->cmd_len = hdr->cmd_len;
rq 205 block/scsi_ioctl.c rq->cmd_type = REQ_TYPE_BLOCK_PC;
rq 207 block/scsi_ioctl.c rq->timeout = msecs_to_jiffies(hdr->timeout);
rq 208 block/scsi_ioctl.c if (!rq->timeout)
rq 209 block/scsi_ioctl.c rq->timeout = q->sg_timeout;
rq 210 block/scsi_ioctl.c if (!rq->timeout)
rq 211 block/scsi_ioctl.c rq->timeout = BLK_DEFAULT_SG_TIMEOUT;
rq 222 block/scsi_ioctl.c blk_rq_unmap_user(rq->bio);
rq 223 block/scsi_ioctl.c blk_put_request(rq);
rq 235 block/scsi_ioctl.c hdr->status = rq->errors & 0xff;
rq 236 block/scsi_ioctl.c hdr->masked_status = status_byte(rq->errors);
rq 237 block/scsi_ioctl.c hdr->msg_status = msg_byte(rq->errors);
rq 238 block/scsi_ioctl.c hdr->host_status = host_byte(rq->errors);
rq 239 block/scsi_ioctl.c hdr->driver_status = driver_byte(rq->errors);
rq 243 block/scsi_ioctl.c hdr->resid = rq->data_len;
rq 246 block/scsi_ioctl.c if (rq->sense_len && hdr->sbp) {
rq 247 block/scsi_ioctl.c int len = min((unsigned int) hdr->mx_sb_len, rq->sense_len);
rq 249 block/scsi_ioctl.c if (!copy_to_user(hdr->sbp, rq->sense, len))
rq 255 block/scsi_ioctl.c rq->bio = bio;
rq 256 block/scsi_ioctl.c r = blk_unmap_sghdr_rq(rq, hdr);
rq 268 block/scsi_ioctl.c struct request *rq;
rq 292 block/scsi_ioctl.c rq = blk_get_request(q, writing ? WRITE : READ, GFP_KERNEL);
rq 293 block/scsi_ioctl.c if (!rq)
rq 296 block/scsi_ioctl.c if (blk_fill_sghdr_rq(q, rq, hdr, file)) {
rq 297 block/scsi_ioctl.c blk_put_request(rq);
rq 317 block/scsi_ioctl.c ret = blk_rq_map_user_iov(q, rq, NULL, iov, hdr->iovec_count,
rq 321 block/scsi_ioctl.c ret = blk_rq_map_user(q, rq, NULL, hdr->dxferp, hdr->dxfer_len,
rq 327 block/scsi_ioctl.c bio = rq->bio;
rq 329 block/scsi_ioctl.c rq->sense = sense;
rq 330 block/scsi_ioctl.c rq->sense_len = 0;
rq 331 block/scsi_ioctl.c rq->retries = 0;
rq 339 block/scsi_ioctl.c blk_execute_rq(q, bd_disk, rq, 0);
rq 343 block/scsi_ioctl.c return blk_complete_sghdr_rq(rq, hdr, bio);
rq 345 block/scsi_ioctl.c blk_put_request(rq);
rq 386 block/scsi_ioctl.c struct request *rq;
rq 414 block/scsi_ioctl.c rq = blk_get_request(q, in_len ? WRITE : READ, __GFP_WAIT);
rq 422 block/scsi_ioctl.c rq->cmd_len = cmdlen;
rq 423 block/scsi_ioctl.c if (copy_from_user(rq->cmd, sic->data, cmdlen))
rq 433 block/scsi_ioctl.c err = blk_verify_command(&q->cmd_filter, rq->cmd, write_perm);
rq 438 block/scsi_ioctl.c rq->retries = 5;
rq 443 block/scsi_ioctl.c rq->timeout = FORMAT_UNIT_TIMEOUT;
rq 444 block/scsi_ioctl.c rq->retries = 1;
rq 447 block/scsi_ioctl.c rq->timeout = START_STOP_TIMEOUT;
rq 450 block/scsi_ioctl.c rq->timeout = MOVE_MEDIUM_TIMEOUT;
rq 453 block/scsi_ioctl.c rq->timeout = READ_ELEMENT_STATUS_TIMEOUT;
rq 456 block/scsi_ioctl.c rq->timeout = READ_DEFECT_DATA_TIMEOUT;
rq 457 block/scsi_ioctl.c rq->retries = 1;
rq 460 block/scsi_ioctl.c rq->timeout = BLK_DEFAULT_SG_TIMEOUT;
rq 464 block/scsi_ioctl.c if (bytes && blk_rq_map_kern(q, rq, buffer, bytes, __GFP_WAIT)) {
rq 470 block/scsi_ioctl.c rq->sense = sense;
rq 471 block/scsi_ioctl.c rq->sense_len = 0;
rq 472 block/scsi_ioctl.c rq->cmd_type = REQ_TYPE_BLOCK_PC;
rq 474 block/scsi_ioctl.c blk_execute_rq(q, disk, rq, 0);
rq 477 block/scsi_ioctl.c err = rq->errors & 0xff; /* only 8 bit SCSI status */
rq 479 block/scsi_ioctl.c if (rq->sense_len && rq->sense) {
rq 480 block/scsi_ioctl.c bytes = (OMAX_SB_LEN > rq->sense_len) ?
rq 481 block/scsi_ioctl.c rq->sense_len : OMAX_SB_LEN;
rq 482 block/scsi_ioctl.c if (copy_to_user(sic->data, rq->sense, bytes))
rq 492 block/scsi_ioctl.c blk_put_request(rq);
rq 501 block/scsi_ioctl.c struct request *rq;
rq 504 block/scsi_ioctl.c rq = blk_get_request(q, WRITE, __GFP_WAIT);
rq 505 block/scsi_ioctl.c rq->cmd_type = REQ_TYPE_BLOCK_PC;
rq 506 block/scsi_ioctl.c rq->data = NULL;
rq 507 block/scsi_ioctl.c rq->data_len = 0;
rq 508 block/scsi_ioctl.c rq->extra_len = 0;
rq 509 block/scsi_ioctl.c rq->timeout = BLK_DEFAULT_SG_TIMEOUT;
rq 510 block/scsi_ioctl.c rq->cmd[0] = cmd;
rq 511 block/scsi_ioctl.c rq->cmd[4] = data;
rq 512 block/scsi_ioctl.c rq->cmd_len = 6;
rq 513 block/scsi_ioctl.c err = blk_execute_rq(q, bd_disk, rq, 0);
rq 514 block/scsi_ioctl.c blk_put_request(rq);
rq 134 fs/dlm/lock.c __dlm_compat_matrix[(gr)->lkb_grmode + 1][(rq)->lkb_rqmode + 1]
rq 1798 fs/dlm/lock.c if ((gr->lkb_grmode == DLM_LOCK_PR && rq->lkb_rqmode == DLM_LOCK_CW) ||
rq 1799 fs/dlm/lock.c (gr->lkb_grmode == DLM_LOCK_CW && rq->lkb_rqmode == DLM_LOCK_PR)) {
rq 1805 fs/dlm/lock.c if (gr->lkb_highbast < rq->lkb_rqmode && !modes_compat(gr, rq))
rq 215 fs/ncpfs/sock.c struct ncp_request_reply *rq;
rq 220 fs/ncpfs/sock.c rq = server->tx.creq;
rq 221 fs/ncpfs/sock.c if (!rq)
rq 225 fs/ncpfs/sock.c memcpy(iovc, rq->tx_ciov, rq->tx_iovlen * sizeof(iov[0]));
rq 226 fs/ncpfs/sock.c result = do_send(server->ncp_sock, iovc, rq->tx_iovlen,
rq 227 fs/ncpfs/sock.c rq->tx_totallen, MSG_NOSIGNAL | MSG_DONTWAIT);
rq 234 fs/ncpfs/sock.c __ncp_abort_request(server, rq, result);
rq 237 fs/ncpfs/sock.c if (result >= rq->tx_totallen) {
rq 238 fs/ncpfs/sock.c server->rcv.creq = rq;
rq 242 fs/ncpfs/sock.c rq->tx_totallen -= result;
rq 243 fs/ncpfs/sock.c iov = rq->tx_ciov;
rq 247 fs/ncpfs/sock.c rq->tx_iovlen--;
rq 251 fs/ncpfs/sock.c rq->tx_ciov = iov;
rq 317 include/linux/blkdev.h struct request_list rq;
rq 558 include/linux/blkdev.h #define blk_fs_request(rq) ((rq)->cmd_type == REQ_TYPE_FS)
rq 559 include/linux/blkdev.h #define blk_pc_request(rq) ((rq)->cmd_type == REQ_TYPE_BLOCK_PC)
rq 560 include/linux/blkdev.h #define blk_special_request(rq) ((rq)->cmd_type == REQ_TYPE_SPECIAL)
rq 561 include/linux/blkdev.h #define blk_sense_request(rq) ((rq)->cmd_type == REQ_TYPE_SENSE)
rq 563 include/linux/blkdev.h #define blk_noretry_request(rq) ((rq)->cmd_flags & REQ_FAILFAST)
rq 564 include/linux/blkdev.h #define blk_rq_started(rq) ((rq)->cmd_flags & REQ_STARTED)
rq 566 include/linux/blkdev.h #define blk_account_rq(rq) (blk_rq_started(rq) && (blk_fs_request(rq) || blk_discard_rq(rq)))
rq 568 include/linux/blkdev.h #define blk_pm_suspend_request(rq) ((rq)->cmd_type == REQ_TYPE_PM_SUSPEND)
rq 569 include/linux/blkdev.h #define blk_pm_resume_request(rq) ((rq)->cmd_type == REQ_TYPE_PM_RESUME)
rq 571 include/linux/blkdev.h (blk_pm_suspend_request(rq) || blk_pm_resume_request(rq))
rq 573 include/linux/blkdev.h #define blk_rq_cpu_valid(rq) ((rq)->cpu != -1)
rq 574 include/linux/blkdev.h #define blk_sorted_rq(rq) ((rq)->cmd_flags & REQ_SORTED)
rq 575 include/linux/blkdev.h #define blk_barrier_rq(rq) ((rq)->cmd_flags & REQ_HARDBARRIER)
rq 576 include/linux/blkdev.h #define blk_fua_rq(rq) ((rq)->cmd_flags & REQ_FUA)
rq 577 include/linux/blkdev.h #define blk_discard_rq(rq) ((rq)->cmd_flags & REQ_DISCARD)
rq 578 include/linux/blkdev.h #define blk_bidi_rq(rq) ((rq)->next_rq != NULL)
rq 579 include/linux/blkdev.h #define blk_empty_barrier(rq) (blk_barrier_rq(rq) && blk_fs_request(rq) && !(rq)->hard_nr_sectors)
rq 581 include/linux/blkdev.h #define blk_queued_rq(rq) (!list_empty(&(rq)->queuelist))
rq 585 include/linux/blkdev.h #define rq_data_dir(rq) ((rq)->cmd_flags & 1)
rq 590 include/linux/blkdev.h #define rq_is_sync(rq) (rq_data_dir((rq)) == READ || (rq)->cmd_flags & REQ_RW_SYNC)
rq 591 include/linux/blkdev.h #define rq_is_meta(rq) ((rq)->cmd_flags & REQ_RW_META)
rq 624 include/linux/blkdev.h (!((rq)->cmd_flags & RQ_NOMERGE_FLAGS) && \
rq 625 include/linux/blkdev.h (blk_discard_rq(rq) || blk_fs_request((rq))))
rq 683 include/linux/blkdev.h if ((rq->bio)) \
rq 684 include/linux/blkdev.h for (_bio = (rq)->bio; _bio; _bio = _bio->bi_next)
rq 870 include/linux/blkdev.h #define blk_rq_tagged(rq) ((rq)->cmd_flags & REQ_QUEUED)
rq 1032 include/linux/blkdev.h if (rq->bio == NULL)
rq 1035 include/linux/blkdev.h return bio_integrity(rq->bio);
rq 200 include/linux/blktrace_api.h int rw = rq->cmd_flags & 0x03;
rq 205 include/linux/blktrace_api.h if (blk_discard_rq(rq))
rq 208 include/linux/blktrace_api.h if (blk_pc_request(rq)) {
rq 210 include/linux/blktrace_api.h __blk_add_trace(bt, 0, rq->data_len, rw, what, rq->errors, sizeof(rq->cmd), rq->cmd);
rq 213 include/linux/blktrace_api.h __blk_add_trace(bt, rq->hard_sector, rq->hard_nr_sectors << 9, rw, what, rq->errors, 0, NULL);
rq 173 include/linux/elevator.h #define rq_end_sector(rq) ((rq)->sector + (rq)->nr_sectors)
rq 180 include/linux/elevator.h #define rq_fifo_time(rq) ((unsigned long) (rq)->csd.list.next)
rq 181 include/linux/elevator.h #define rq_set_fifo_time(rq,exp) ((rq)->csd.list.next = (void *) (exp))
rq 184 include/linux/elevator.h list_del_init(&(rq)->queuelist); \
rq 185 include/linux/elevator.h INIT_LIST_HEAD(&(rq)->csd.list); \
rq 387 include/linux/ide.h struct request *rq; /* copy of request */
rq 432 include/linux/ide.h struct request *rq;
rq 590 include/linux/ide.h struct request *rq; /* current request */
rq 885 include/linux/ide.h struct request *rq;
rq 1119 include/linux/ide.h ide_startstop_t (*error)(ide_drive_t *, struct request *rq, u8, u8);
rq 214 include/linux/isdn_ppp.h struct ippp_buf_queue rq[NUM_RCV_BUFFS]; /* packet queue for isdn_ppp_read() */
rq 183 include/linux/mii.h return (struct mii_ioctl_data *) &rq->ifr_ifru;
rq 892 include/linux/sched.h struct rq;
rq 898 include/linux/sched.h void (*enqueue_task) (struct rq *rq, struct task_struct *p, int wakeup);
rq 899 include/linux/sched.h void (*dequeue_task) (struct rq *rq, struct task_struct *p, int sleep);
rq 900 include/linux/sched.h void (*yield_task) (struct rq *rq);
rq 903 include/linux/sched.h void (*check_preempt_curr) (struct rq *rq, struct task_struct *p, int sync);
rq 905 include/linux/sched.h struct task_struct * (*pick_next_task) (struct rq *rq);
rq 906 include/linux/sched.h void (*put_prev_task) (struct rq *rq, struct task_struct *p);
rq 909 include/linux/sched.h unsigned long (*load_balance) (struct rq *this_rq, int this_cpu,
rq 910 include/linux/sched.h struct rq *busiest, unsigned long max_load_move,
rq 914 include/linux/sched.h int (*move_one_task) (struct rq *this_rq, int this_cpu,
rq 915 include/linux/sched.h struct rq *busiest, struct sched_domain *sd,
rq 917 include/linux/sched.h void (*pre_schedule) (struct rq *this_rq, struct task_struct *task);
rq 918 include/linux/sched.h void (*post_schedule) (struct rq *this_rq);
rq 919 include/linux/sched.h void (*task_wake_up) (struct rq *this_rq, struct task_struct *task);
rq 922 include/linux/sched.h void (*set_curr_task) (struct rq *rq);
rq 923 include/linux/sched.h void (*task_tick) (struct rq *rq, struct task_struct *p, int queued);
rq 924 include/linux/sched.h void (*task_new) (struct rq *rq, struct task_struct *p);
rq 928 include/linux/sched.h void (*rq_online)(struct rq *rq);
rq 929 include/linux/sched.h void (*rq_offline)(struct rq *rq);
rq 931 include/linux/sched.h void (*switched_from) (struct rq *this_rq, struct task_struct *task,
rq 933 include/linux/sched.h void (*switched_to) (struct rq *this_rq, struct task_struct *task,
rq 935 include/linux/sched.h void (*prio_changed) (struct rq *this_rq, struct task_struct *task,
rq 99 include/linux/sunrpc/svcauth.h int (*accept)(struct svc_rqst *rq, __be32 *authp);
rq 100 include/linux/sunrpc/svcauth.h int (*release)(struct svc_rqst *rq);
rq 102 include/linux/sunrpc/svcauth.h int (*set_client)(struct svc_rqst *rq);
rq 405 kernel/sched.c struct rq *rq; /* cpu runqueue to which this cfs_rq is attached */
rq 465 kernel/sched.c struct rq *rq;
rq 614 kernel/sched.c rq->curr->sched_class->check_preempt_curr(rq, p, sync);
rq 620 kernel/sched.c return rq->cpu;
rq 643 kernel/sched.c rq->clock = sched_clock_cpu(cpu_of(rq));
rq 665 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 668 kernel/sched.c ret = spin_is_locked(&rq->lock);
rq 856 kernel/sched.c return rq->curr == p;
rq 862 kernel/sched.c return task_current(rq, p);
rq 873 kernel/sched.c rq->lock.owner = current;
rq 880 kernel/sched.c spin_acquire(&rq->lock.dep_map, 0, 0, _THIS_IP_);
rq 882 kernel/sched.c spin_unlock_irq(&rq->lock);
rq 891 kernel/sched.c return task_current(rq, p);
rq 906 kernel/sched.c spin_unlock_irq(&rq->lock);
rq 908 kernel/sched.c spin_unlock(&rq->lock);
rq 933 kernel/sched.c static inline struct rq *__task_rq_lock(struct task_struct *p)
rq 937 kernel/sched.c struct rq *rq = task_rq(p);
rq 938 kernel/sched.c spin_lock(&rq->lock);
rq 939 kernel/sched.c if (likely(rq == task_rq(p)))
rq 940 kernel/sched.c return rq;
rq 941 kernel/sched.c spin_unlock(&rq->lock);
rq 950 kernel/sched.c static struct rq *task_rq_lock(struct task_struct *p, unsigned long *flags)
rq 953 kernel/sched.c struct rq *rq;
rq 957 kernel/sched.c rq = task_rq(p);
rq 958 kernel/sched.c spin_lock(&rq->lock);
rq 959 kernel/sched.c if (likely(rq == task_rq(p)))
rq 960 kernel/sched.c return rq;
rq 961 kernel/sched.c spin_unlock_irqrestore(&rq->lock, *flags);
rq 968 kernel/sched.c spin_unlock(&rq->lock);
rq 974 kernel/sched.c spin_unlock_irqrestore(&rq->lock, *flags);
rq 980 kernel/sched.c static struct rq *this_rq_lock(void)
rq 983 kernel/sched.c struct rq *rq;
rq 986 kernel/sched.c rq = this_rq();
rq 987 kernel/sched.c spin_lock(&rq->lock);
rq 989 kernel/sched.c return rq;
rq 1013 kernel/sched.c if (!cpu_active(cpu_of(rq)))
rq 1015 kernel/sched.c return hrtimer_is_hres_active(&rq->hrtick_timer);
rq 1020 kernel/sched.c if (hrtimer_active(&rq->hrtick_timer))
rq 1021 kernel/sched.c hrtimer_cancel(&rq->hrtick_timer);
rq 1030 kernel/sched.c struct rq *rq = container_of(timer, struct rq, hrtick_timer);
rq 1032 kernel/sched.c WARN_ON_ONCE(cpu_of(rq) != smp_processor_id());
rq 1034 kernel/sched.c spin_lock(&rq->lock);
rq 1035 kernel/sched.c update_rq_clock(rq);
rq 1036 kernel/sched.c rq->curr->sched_class->task_tick(rq, rq->curr, 1);
rq 1037 kernel/sched.c spin_unlock(&rq->lock);
rq 1048 kernel/sched.c struct rq *rq = arg;
rq 1050 kernel/sched.c spin_lock(&rq->lock);
rq 1051 kernel/sched.c hrtimer_restart(&rq->hrtick_timer);
rq 1052 kernel/sched.c rq->hrtick_csd_pending = 0;
rq 1053 kernel/sched.c spin_unlock(&rq->lock);
rq 1063 kernel/sched.c struct hrtimer *timer = &rq->hrtick_timer;
rq 1068 kernel/sched.c if (rq == this_rq()) {
rq 1070 kernel/sched.c } else if (!rq->hrtick_csd_pending) {
rq 1071 kernel/sched.c __smp_call_function_single(cpu_of(rq), &rq->hrtick_csd);
rq 1072 kernel/sched.c rq->hrtick_csd_pending = 1;
rq 1107 kernel/sched.c hrtimer_start(&rq->hrtick_timer, ns_to_ktime(delay), HRTIMER_MODE_REL);
rq 1118 kernel/sched.c rq->hrtick_csd_pending = 0;
rq 1120 kernel/sched.c rq->hrtick_csd.flags = 0;
rq 1121 kernel/sched.c rq->hrtick_csd.func = __hrtick_start;
rq 1122 kernel/sched.c rq->hrtick_csd.info = rq;
rq 1125 kernel/sched.c hrtimer_init(&rq->hrtick_timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
rq 1126 kernel/sched.c rq->hrtick_timer.function = hrtick;
rq 1127 kernel/sched.c rq->hrtick_timer.cb_mode = HRTIMER_CB_IRQSAFE_PERCPU;
rq 1179 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 1182 kernel/sched.c if (!spin_trylock_irqsave(&rq->lock, flags))
rq 1185 kernel/sched.c spin_unlock_irqrestore(&rq->lock, flags);
rq 1201 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 1213 kernel/sched.c if (rq->curr != rq->idle)
rq 1221 kernel/sched.c set_tsk_thread_flag(rq->idle, TIF_NEED_RESCHED);
rq 1225 kernel/sched.c if (!tsk_is_polling(rq->idle))
rq 1380 kernel/sched.c update_load_add(&rq->load, load);
rq 1385 kernel/sched.c update_load_sub(&rq->load, load);
rq 1440 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 1442 kernel/sched.c if (rq->nr_running)
rq 1443 kernel/sched.c rq->avg_load_per_task = rq->load.weight / rq->nr_running;
rq 1445 kernel/sched.c return rq->avg_load_per_task;
rq 1530 kernel/sched.c struct rq *rq = cpu_rq(i);
rq 1533 kernel/sched.c spin_lock_irqsave(&rq->lock, flags);
rq 1535 kernel/sched.c spin_unlock_irqrestore(&rq->lock, flags);
rq 1577 kernel/sched.c spin_unlock(&rq->lock);
rq 1579 kernel/sched.c spin_lock(&rq->lock);
rq 1624 kernel/sched.c rq->nr_running++;
rq 1629 kernel/sched.c rq->nr_running--;
rq 1662 kernel/sched.c p->sched_class->enqueue_task(rq, p, wakeup);
rq 1675 kernel/sched.c p->sched_class->dequeue_task(rq, p, sleep);
rq 1731 kernel/sched.c rq->nr_uninterruptible--;
rq 1733 kernel/sched.c enqueue_task(rq, p, wakeup);
rq 1734 kernel/sched.c inc_nr_running(rq);
rq 1743 kernel/sched.c rq->nr_uninterruptible++;
rq 1745 kernel/sched.c dequeue_task(rq, p, sleep);
rq 1746 kernel/sched.c dec_nr_running(rq);
rq 1778 kernel/sched.c prev_class->switched_from(rq, p, running);
rq 1779 kernel/sched.c p->sched_class->switched_to(rq, p, running);
rq 1781 kernel/sched.c p->sched_class->prio_changed(rq, p, oldprio, running);
rq 1823 kernel/sched.c struct rq *old_rq = cpu_rq(old_cpu), *new_rq = cpu_rq(new_cpu);
rq 1865 kernel/sched.c struct rq *rq = task_rq(p);
rq 1871 kernel/sched.c if (!p->se.on_rq && !task_running(rq, p)) {
rq 1879 kernel/sched.c list_add(&req->list, &rq->migration_queue);
rq 1905 kernel/sched.c struct rq *rq;
rq 1914 kernel/sched.c rq = task_rq(p);
rq 1927 kernel/sched.c while (task_running(rq, p)) {
rq 1938 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 1939 kernel/sched.c running = task_running(rq, p);
rq 1944 kernel/sched.c task_rq_unlock(rq, &flags);
rq 2021 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 2027 kernel/sched.c return min(rq->cpu_load[type-1], total);
rq 2036 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 2042 kernel/sched.c return max(rq->cpu_load[type-1], total);
rq 2215 kernel/sched.c struct rq *rq;
rq 2237 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 2250 kernel/sched.c if (unlikely(task_running(rq, p)))
rq 2256 kernel/sched.c task_rq_unlock(rq, &flags);
rq 2258 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 2270 kernel/sched.c schedstat_inc(rq, ttwu_count);
rq 2272 kernel/sched.c schedstat_inc(rq, ttwu_local);
rq 2295 kernel/sched.c update_rq_clock(rq);
rq 2296 kernel/sched.c activate_task(rq, p, 1);
rq 2302 kernel/sched.c p->pid, p->state, rq, p, rq->curr);
rq 2303 kernel/sched.c check_preempt_curr(rq, p, sync);
rq 2308 kernel/sched.c p->sched_class->task_wake_up(rq, p);
rq 2313 kernel/sched.c task_rq_unlock(rq, &flags);
rq 2417 kernel/sched.c struct rq *rq;
rq 2419 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 2421 kernel/sched.c update_rq_clock(rq);
rq 2426 kernel/sched.c activate_task(rq, p, 0);
rq 2432 kernel/sched.c p->sched_class->task_new(rq, p);
rq 2433 kernel/sched.c inc_nr_running(rq);
rq 2437 kernel/sched.c p->pid, p->state, rq, p, rq->curr);
rq 2438 kernel/sched.c check_preempt_curr(rq, p, 0);
rq 2441 kernel/sched.c p->sched_class->task_wake_up(rq, p);
rq 2443 kernel/sched.c task_rq_unlock(rq, &flags);
rq 2522 kernel/sched.c prepare_lock_switch(rq, next);
rq 2544 kernel/sched.c struct mm_struct *mm = rq->prev_mm;
rq 2547 kernel/sched.c rq->prev_mm = NULL;
rq 2562 kernel/sched.c finish_lock_switch(rq, prev);
rq 2565 kernel/sched.c current->sched_class->post_schedule(rq);
rq 2588 kernel/sched.c struct rq *rq = this_rq();
rq 2590 kernel/sched.c finish_task_switch(rq, prev);
rq 2609 kernel/sched.c prepare_task_switch(rq, prev, next);
rq 2614 kernel/sched.c rq, prev, next);
rq 2633 kernel/sched.c rq->prev_mm = oldmm;
rq 2642 kernel/sched.c spin_release(&rq->lock.dep_map, 1, _THIS_IP_);
rq 2847 kernel/sched.c struct rq *rq;
rq 2849 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 2857 kernel/sched.c struct task_struct *mt = rq->migration_thread;
rq 2860 kernel/sched.c task_rq_unlock(rq, &flags);
rq 2868 kernel/sched.c task_rq_unlock(rq, &flags);
rq 2921 kernel/sched.c if (task_running(rq, p)) {
rq 2932 kernel/sched.c if (!task_hot(p, rq->clock, sd) ||
rq 2935 kernel/sched.c if (task_hot(p, rq->clock, sd)) {
rq 2943 kernel/sched.c if (task_hot(p, rq->clock, sd)) {
rq 3139 kernel/sched.c struct rq *rq;
rq 3144 kernel/sched.c rq = cpu_rq(i);
rq 3146 kernel/sched.c if (*sd_idle && rq->nr_running)
rq 3166 kernel/sched.c sum_nr_running += rq->nr_running;
rq 3409 kernel/sched.c static struct rq *
rq 3413 kernel/sched.c struct rq *busiest = NULL, *rq;
rq 3423 kernel/sched.c rq = cpu_rq(i);
rq 3426 kernel/sched.c if (rq->nr_running == 1 && wl > imbalance)
rq 3431 kernel/sched.c busiest = rq;
rq 3455 kernel/sched.c struct rq *busiest;
rq 3614 kernel/sched.c struct rq *busiest = NULL;
rq 3740 kernel/sched.c struct rq *target_rq;
rq 3864 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 3896 kernel/sched.c if (load_balance(cpu, rq, sd, idle, &balance, &tmp)) {
rq 3929 kernel/sched.c rq->next_balance = next_balance;
rq 3940 kernel/sched.c struct rq *this_rq = cpu_rq(this_cpu);
rq 3955 kernel/sched.c struct rq *rq;
rq 3970 kernel/sched.c rq = cpu_rq(balance_cpu);
rq 3971 kernel/sched.c if (time_after(this_rq->next_balance, rq->next_balance))
rq 3972 kernel/sched.c this_rq->next_balance = rq->next_balance;
rq 3993 kernel/sched.c if (rq->in_nohz_recently && !rq->idle_at_tick) {
rq 3994 kernel/sched.c rq->in_nohz_recently = 0;
rq 4021 kernel/sched.c if (rq->idle_at_tick && atomic_read(&nohz.load_balancer) == cpu &&
rq 4031 kernel/sched.c if (rq->idle_at_tick && atomic_read(&nohz.load_balancer) != cpu &&
rq 4035 kernel/sched.c if (time_after_eq(jiffies, rq->next_balance))
rq 4062 kernel/sched.c struct rq *rq;
rq 4064 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 4066 kernel/sched.c if (task_current(rq, p)) {
rq 4067 kernel/sched.c update_rq_clock(rq);
rq 4068 kernel/sched.c delta_exec = rq->clock - p->se.exec_start;
rq 4072 kernel/sched.c task_rq_unlock(rq, &flags);
rq 4138 kernel/sched.c struct rq *rq = this_rq();
rq 4154 kernel/sched.c else if (p != rq->idle)
rq 4156 kernel/sched.c else if (atomic_read(&rq->nr_iowait) > 0)
rq 4184 kernel/sched.c struct rq *rq = this_rq();
rq 4186 kernel/sched.c if (p == rq->idle) {
rq 4188 kernel/sched.c if (atomic_read(&rq->nr_iowait) > 0)
rq 4265 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 4266 kernel/sched.c struct task_struct *curr = rq->curr;
rq 4270 kernel/sched.c spin_lock(&rq->lock);
rq 4271 kernel/sched.c update_rq_clock(rq);
rq 4272 kernel/sched.c update_cpu_load(rq);
rq 4273 kernel/sched.c curr->sched_class->task_tick(rq, curr, 0);
rq 4274 kernel/sched.c spin_unlock(&rq->lock);
rq 4277 kernel/sched.c rq->idle_at_tick = idle_cpu(cpu);
rq 4278 kernel/sched.c trigger_load_balance(rq, cpu);
rq 4399 kernel/sched.c if (likely(rq->nr_running == rq->cfs.nr_running)) {
rq 4400 kernel/sched.c p = fair_sched_class.pick_next_task(rq);
rq 4407 kernel/sched.c p = class->pick_next_task(rq);
rq 4425 kernel/sched.c struct rq *rq;
rq 4431 kernel/sched.c rq = cpu_rq(cpu);
rq 4433 kernel/sched.c prev = rq->curr;
rq 4442 kernel/sched.c hrtick_clear(rq);
rq 4448 kernel/sched.c update_rq_clock(rq);
rq 4449 kernel/sched.c spin_lock(&rq->lock);
rq 4456 kernel/sched.c deactivate_task(rq, prev, 1);
rq 4462 kernel/sched.c prev->sched_class->pre_schedule(rq, prev);
rq 4465 kernel/sched.c if (unlikely(!rq->nr_running))
rq 4466 kernel/sched.c idle_balance(cpu, rq);
rq 4468 kernel/sched.c prev->sched_class->put_prev_task(rq, prev);
rq 4469 kernel/sched.c next = pick_next_task(rq, prev);
rq 4474 kernel/sched.c rq->nr_switches++;
rq 4475 kernel/sched.c rq->curr = next;
rq 4478 kernel/sched.c context_switch(rq, prev, next); /* unlocks the rq */
rq 4484 kernel/sched.c rq = cpu_rq(cpu);
rq 4486 kernel/sched.c spin_unlock_irq(&rq->lock);
rq 4907 kernel/sched.c struct rq *rq;
rq 4912 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 4913 kernel/sched.c update_rq_clock(rq);
rq 4917 kernel/sched.c running = task_current(rq, p);
rq 4919 kernel/sched.c dequeue_task(rq, p, 0);
rq 4921 kernel/sched.c p->sched_class->put_prev_task(rq, p);
rq 4931 kernel/sched.c p->sched_class->set_curr_task(rq);
rq 4933 kernel/sched.c enqueue_task(rq, p, 0);
rq 4935 kernel/sched.c check_class_changed(rq, p, prev_class, oldprio, running);
rq 4937 kernel/sched.c task_rq_unlock(rq, &flags);
rq 4946 kernel/sched.c struct rq *rq;
rq 4954 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 4955 kernel/sched.c update_rq_clock(rq);
rq 4968 kernel/sched.c dequeue_task(rq, p, 0);
rq 4977 kernel/sched.c enqueue_task(rq, p, 0);
rq 4982 kernel/sched.c if (delta < 0 || (delta > 0 && task_running(rq, p)))
rq 4983 kernel/sched.c resched_task(rq->curr);
rq 4986 kernel/sched.c task_rq_unlock(rq, &flags);
rq 5128 kernel/sched.c struct rq *rq;
rq 5211 kernel/sched.c rq = __task_rq_lock(p);
rq 5215 kernel/sched.c __task_rq_unlock(rq);
rq 5219 kernel/sched.c update_rq_clock(rq);
rq 5221 kernel/sched.c running = task_current(rq, p);
rq 5223 kernel/sched.c deactivate_task(rq, p, 0);
rq 5225 kernel/sched.c p->sched_class->put_prev_task(rq, p);
rq 5228 kernel/sched.c __setscheduler(rq, p, policy, param->sched_priority);
rq 5231 kernel/sched.c p->sched_class->set_curr_task(rq);
rq 5233 kernel/sched.c activate_task(rq, p, 0);
rq 5235 kernel/sched.c check_class_changed(rq, p, prev_class, oldprio, running);
rq 5237 kernel/sched.c __task_rq_unlock(rq);
rq 5534 kernel/sched.c struct rq *rq = this_rq_lock();
rq 5536 kernel/sched.c schedstat_inc(rq, yld_count);
rq 5537 kernel/sched.c current->sched_class->yield_task(rq);
rq 5543 kernel/sched.c __release(rq->lock);
rq 5544 kernel/sched.c spin_release(&rq->lock.dep_map, 1, _THIS_IP_);
rq 5545 kernel/sched.c _raw_spin_unlock(&rq->lock);
rq 5643 kernel/sched.c struct rq *rq = &__raw_get_cpu_var(runqueues);
rq 5646 kernel/sched.c atomic_inc(&rq->nr_iowait);
rq 5648 kernel/sched.c atomic_dec(&rq->nr_iowait);
rq 5655 kernel/sched.c struct rq *rq = &__raw_get_cpu_var(runqueues);
rq 5659 kernel/sched.c atomic_inc(&rq->nr_iowait);
rq 5661 kernel/sched.c atomic_dec(&rq->nr_iowait);
rq 5754 kernel/sched.c struct rq *rq;
rq 5756 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 5757 kernel/sched.c if (rq->cfs.load.weight)
rq 5758 kernel/sched.c time_slice = NS_TO_JIFFIES(sched_slice(&rq->cfs, se));
rq 5759 kernel/sched.c task_rq_unlock(rq, &flags);
rq 5856 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 5866 kernel/sched.c spin_lock_irqsave(&rq->lock, flags);
rq 5867 kernel/sched.c rq->curr = rq->idle = idle;
rq 5871 kernel/sched.c spin_unlock_irqrestore(&rq->lock, flags);
rq 5951 kernel/sched.c struct rq *rq;
rq 5954 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 5979 kernel/sched.c task_rq_unlock(rq, &flags);
rq 5980 kernel/sched.c wake_up_process(rq->migration_thread);
rq 5986 kernel/sched.c task_rq_unlock(rq, &flags);
rq 6005 kernel/sched.c struct rq *rq_dest, *rq_src;
rq 6046 kernel/sched.c struct rq *rq;
rq 6048 kernel/sched.c rq = cpu_rq(cpu);
rq 6049 kernel/sched.c BUG_ON(rq->migration_thread != current);
rq 6056 kernel/sched.c spin_lock_irq(&rq->lock);
rq 6059 kernel/sched.c spin_unlock_irq(&rq->lock);
rq 6063 kernel/sched.c if (rq->active_balance) {
rq 6064 kernel/sched.c active_load_balance(rq, cpu);
rq 6065 kernel/sched.c rq->active_balance = 0;
rq 6068 kernel/sched.c head = &rq->migration_queue;
rq 6071 kernel/sched.c spin_unlock_irq(&rq->lock);
rq 6079 kernel/sched.c spin_unlock(&rq->lock);
rq 6119 kernel/sched.c struct rq *rq;
rq 6144 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 6147 kernel/sched.c task_rq_unlock(rq, &flags);
rq 6172 kernel/sched.c struct rq *rq_dest = cpu_rq(any_online_cpu(*CPU_MASK_ALL_PTR));
rq 6209 kernel/sched.c struct rq *rq = cpu_rq(this_cpu);
rq 6210 kernel/sched.c struct task_struct *p = rq->idle;
rq 6220 kernel/sched.c spin_lock_irqsave(&rq->lock, flags);
rq 6222 kernel/sched.c __setscheduler(rq, p, SCHED_FIFO, MAX_RT_PRIO-1);
rq 6224 kernel/sched.c update_rq_clock(rq);
rq 6225 kernel/sched.c activate_task(rq, p, 0);
rq 6227 kernel/sched.c spin_unlock_irqrestore(&rq->lock, flags);
rq 6248 kernel/sched.c struct rq *rq = cpu_rq(dead_cpu);
rq 6263 kernel/sched.c spin_unlock_irq(&rq->lock);
rq 6265 kernel/sched.c spin_lock_irq(&rq->lock);
rq 6273 kernel/sched.c struct rq *rq = cpu_rq(dead_cpu);
rq 6277 kernel/sched.c if (!rq->nr_running)
rq 6279 kernel/sched.c update_rq_clock(rq);
rq 6280 kernel/sched.c next = pick_next_task(rq, rq->curr);
rq 6283 kernel/sched.c next->sched_class->put_prev_task(rq, next);
rq 6459 kernel/sched.c if (!rq->online) {
rq 6462 kernel/sched.c cpu_set(rq->cpu, rq->rd->online);
rq 6463 kernel/sched.c rq->online = 1;
rq 6467 kernel/sched.c class->rq_online(rq);
rq 6474 kernel/sched.c if (rq->online) {
rq 6479 kernel/sched.c class->rq_offline(rq);
rq 6482 kernel/sched.c cpu_clear(rq->cpu, rq->rd->online);
rq 6483 kernel/sched.c rq->online = 0;
rq 6497 kernel/sched.c struct rq *rq;
rq 6508 kernel/sched.c rq = task_rq_lock(p, &flags);
rq 6509 kernel/sched.c __setscheduler(rq, p, SCHED_FIFO, MAX_RT_PRIO-1);
rq 6510 kernel/sched.c task_rq_unlock(rq, &flags);
rq 6520 kernel/sched.c rq = cpu_rq(cpu);
rq 6521 kernel/sched.c spin_lock_irqsave(&rq->lock, flags);
rq 6522 kernel/sched.c if (rq->rd) {
rq 6523 kernel/sched.c BUG_ON(!cpu_isset(cpu, rq->rd->span));
rq 6525 kernel/sched.c set_rq_online(rq);
rq 6527 kernel/sched.c spin_unlock_irqrestore(&rq->lock, flags);
rq 6546 kernel/sched.c rq = cpu_rq(cpu);
rq 6547 kernel/sched.c kthread_stop(rq->migration_thread);
rq 6548 kernel/sched.c rq->migration_thread = NULL;
rq 6550 kernel/sched.c spin_lock_irq(&rq->lock);
rq 6551 kernel/sched.c update_rq_clock(rq);
rq 6552 kernel/sched.c deactivate_task(rq, rq->idle, 0);
rq 6553 kernel/sched.c rq->idle->static_prio = MAX_PRIO;
rq 6554 kernel/sched.c __setscheduler(rq, rq->idle, SCHED_NORMAL, 0);
rq 6555 kernel/sched.c rq->idle->sched_class = &idle_sched_class;
rq 6557 kernel/sched.c spin_unlock_irq(&rq->lock);
rq 6559 kernel/sched.c migrate_nr_uninterruptible(rq);
rq 6560 kernel/sched.c BUG_ON(rq->nr_running != 0);
rq 6567 kernel/sched.c spin_lock_irq(&rq->lock);
rq 6568 kernel/sched.c while (!list_empty(&rq->migration_queue)) {
rq 6571 kernel/sched.c req = list_entry(rq->migration_queue.next,
rq 6576 kernel/sched.c spin_unlock_irq(&rq->lock);
rq 6582 kernel/sched.c rq = cpu_rq(cpu);
rq 6583 kernel/sched.c spin_lock_irqsave(&rq->lock, flags);
rq 6584 kernel/sched.c if (rq->rd) {
rq 6585 kernel/sched.c BUG_ON(!cpu_isset(cpu, rq->rd->span));
rq 6586 kernel/sched.c set_rq_offline(rq);
rq 6588 kernel/sched.c spin_unlock_irqrestore(&rq->lock, flags);
rq 6812 kernel/sched.c spin_lock_irqsave(&rq->lock, flags);
rq 6814 kernel/sched.c if (rq->rd) {
rq 6815 kernel/sched.c struct root_domain *old_rd = rq->rd;
rq 6817 kernel/sched.c if (cpu_isset(rq->cpu, old_rd->online))
rq 6818 kernel/sched.c set_rq_offline(rq);
rq 6820 kernel/sched.c cpu_clear(rq->cpu, old_rd->span);
rq 6827 kernel/sched.c rq->rd = rd;
rq 6829 kernel/sched.c cpu_set(rq->cpu, rd->span);
rq 6830 kernel/sched.c if (cpu_isset(rq->cpu, cpu_online_map))
rq 6831 kernel/sched.c set_rq_online(rq);
rq 6833 kernel/sched.c spin_unlock_irqrestore(&rq->lock, flags);
rq 6872 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 6895 kernel/sched.c rq_attach_root(rq, rd);
rq 6896 kernel/sched.c rcu_assign_pointer(rq->sd, sd);
rq 8021 kernel/sched.c cfs_rq->rq = rq;
rq 8054 kernel/sched.c rt_rq->rq = rq;
rq 8063 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 8065 kernel/sched.c init_cfs_rq(cfs_rq, rq);
rq 8068 kernel/sched.c list_add(&cfs_rq->leaf_cfs_rq_list, &rq->leaf_cfs_rq_list);
rq 8076 kernel/sched.c se->cfs_rq = &rq->cfs;
rq 8092 kernel/sched.c struct rq *rq = cpu_rq(cpu);
rq 8095 kernel/sched.c init_rt_rq(rt_rq, rq);
rq 8100 kernel/sched.c list_add(&rt_rq->leaf_rt_rq_list, &rq->leaf_rt_rq_list);
rq 8107 kernel/sched.c rt_se->rt_rq = &rq->rt;
rq 8198 kernel/sched.c struct rq *rq;
rq 8200 kernel/sched.c rq = cpu_rq(i);
rq 8201 kernel/sched.c spin_lock_init(&rq->lock);
rq 8202 kernel/sched.c rq->nr_running = 0;
rq 8203 kernel/sched.c init_cfs_rq(&rq->cfs, rq);
rq 8204 kernel/sched.c init_rt_rq(&rq->rt, rq);
rq 8207 kernel/sched.c INIT_LIST_HEAD(&rq->leaf_cfs_rq_list);
rq 8228 kernel/sched.c init_tg_cfs_entry(&init_task_group, &rq->cfs, NULL, i, 1, NULL);
rq 8231 kernel/sched.c init_tg_cfs_entry(&root_task_group, &rq->cfs, NULL, i, 0, NULL);
rq 8251 kernel/sched.c rq->rt.rt_runtime = def_rt_bandwidth.rt_runtime;
rq 8253 kernel/sched.c INIT_LIST_HEAD(&rq->leaf_rt_rq_list);
rq 8255 kernel/sched.c init_tg_rt_entry(&init_task_group, &rq->rt, NULL, i, 1, NULL);
rq 8257 kernel/sched.c init_tg_rt_entry(&root_task_group, &rq->rt, NULL, i, 0, NULL);
rq 8266 kernel/sched.c rq->cpu_load[j] = 0;
rq 8268 kernel/sched.c rq->sd = NULL;
rq 8269 kernel/sched.c rq->rd = NULL;
rq 8270 kernel/sched.c rq->active_balance = 0;
rq 8271 kernel/sched.c rq->next_balance = jiffies;
rq 8272 kernel/sched.c rq->push_cpu = 0;
rq 8273 kernel/sched.c rq->cpu = i;
rq 8274 kernel/sched.c rq->online = 0;
rq 8275 kernel/sched.c rq->migration_thread = NULL;
rq 8276 kernel/sched.c INIT_LIST_HEAD(&rq->migration_queue);
rq 8277 kernel/sched.c rq_attach_root(rq, &def_root_domain);
rq 8279 kernel/sched.c init_rq_hrtick(rq);
rq 8280 kernel/sched.c atomic_set(&rq->nr_iowait, 0);
rq 8353 kernel/sched.c update_rq_clock(rq);
rq 8356 kernel/sched.c deactivate_task(rq, p, 0);
rq 8357 kernel/sched.c __setscheduler(rq, p, SCHED_NORMAL, 0);
rq 8359 kernel/sched.c activate_task(rq, p, 0);
rq 8360 kernel/sched.c resched_task(rq->curr);
rq 8368 kernel/sched.c struct rq *rq;
rq 8396 kernel/sched.c rq = __task_rq_lock(p);
rq 8398 kernel/sched.c normalize_task(rq, p);
rq 8400 kernel/sched.c __task_rq_unlock(rq);
rq 8474 kernel/sched.c struct rq *rq;
rq 8487 kernel/sched.c rq = cpu_rq(i);
rq 8562 kernel/sched.c struct rq *rq;
rq 8576 kernel/sched.c rq = cpu_rq(i);
rq 8709 kernel/sched.c struct rq *rq;
rq 8711 kernel/sched.c rq = task_rq_lock(tsk, &flags);
rq 8713 kernel/sched.c update_rq_clock(rq);
rq 8715 kernel/sched.c running = task_current(rq, tsk);
rq 8719 kernel/sched.c dequeue_task(rq, tsk, 0);
rq 8721 kernel/sched.c tsk->sched_class->put_prev_task(rq, tsk);
rq 8731 kernel/sched.c tsk->sched_class->set_curr_task(rq);
rq 8733 kernel/sched.c enqueue_task(rq, tsk, 0);
rq 8735 kernel/sched.c task_rq_unlock(rq, &flags);
rq 8759 kernel/sched.c struct rq *rq = cfs_rq->rq;
rq 8762 kernel/sched.c spin_lock_irqsave(&rq->lock, flags);
rq 8764 kernel/sched.c spin_unlock_irqrestore(&rq->lock, flags);
rq 59 kernel/sched_debug.c if (rq->curr == p)
rq 108 kernel/sched_debug.c print_task(m, rq, p);
rq 118 kernel/sched_debug.c struct rq *rq = &per_cpu(runqueues, cpu);
rq 141 kernel/sched_debug.c spin_lock_irqsave(&rq->lock, flags);
rq 147 kernel/sched_debug.c min_vruntime = rq->cfs.min_vruntime;
rq 149 kernel/sched_debug.c spin_unlock_irqrestore(&rq->lock, flags);
rq 165 kernel/sched_debug.c #define P(n) SEQ_printf(m, " .%-30s: %d\n", #n, rq->n);
rq 227 kernel/sched_debug.c struct rq *rq = &per_cpu(runqueues, cpu);
rq 241 kernel/sched_debug.c SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rq->x))
rq 243 kernel/sched_debug.c SEQ_printf(m, " .%-30s: %Ld.%06ld\n", #x, SPLIT_NS(rq->x))
rq 247 kernel/sched_debug.c rq->load.weight);
rq 266 kernel/sched_debug.c print_rq(m, rq, cpu);
rq 88 kernel/sched_fair.c static inline struct rq *rq_of(struct cfs_rq *cfs_rq)
rq 90 kernel/sched_fair.c return cfs_rq->rq;
rq 127 kernel/sched_fair.c list_for_each_entry_rcu(cfs_rq, &rq->leaf_cfs_rq_list, leaf_cfs_rq_list)
rq 146 kernel/sched_fair.c static inline struct rq *rq_of(struct cfs_rq *cfs_rq)
rq 148 kernel/sched_fair.c return container_of(cfs_rq, struct rq, cfs);
rq 164 kernel/sched_fair.c struct rq *rq = task_rq(p);
rq 166 kernel/sched_fair.c return &rq->cfs;
rq 181 kernel/sched_fair.c for (cfs_rq = &rq->cfs; cfs_rq; cfs_rq = NULL)
rq 747 kernel/sched_fair.c struct rq *rq = rq_of(cfs_rq);
rq 748 kernel/sched_fair.c u64 pair_slice = rq->clock - cfs_rq->pair_start;
rq 751 kernel/sched_fair.c cfs_rq->pair_start = rq->clock;
rq 828 kernel/sched_fair.c WARN_ON(task_rq(p) != rq);
rq 830 kernel/sched_fair.c if (hrtick_enabled(rq) && cfs_rq->nr_running > 1) {
rq 836 kernel/sched_fair.c if (rq->curr == p)
rq 845 kernel/sched_fair.c if (rq->curr != p)
rq 848 kernel/sched_fair.c hrtick_start(rq, delta);
rq 876 kernel/sched_fair.c hrtick_start_fair(rq, rq->curr);
rq 898 kernel/sched_fair.c hrtick_start_fair(rq, rq->curr);
rq 908 kernel/sched_fair.c struct task_struct *curr = rq->curr;
rq 919 kernel/sched_fair.c update_rq_clock(rq);
rq 1161 kernel/sched_fair.c struct rq *this_rq;
rq 1239 kernel/sched_fair.c struct task_struct *curr = rq->curr;
rq 1245 kernel/sched_fair.c update_rq_clock(rq);
rq 1288 kernel/sched_fair.c struct cfs_rq *cfs_rq = &rq->cfs;
rq 1300 kernel/sched_fair.c hrtick_start_fair(rq, p);
rq 1505 kernel/sched_fair.c resched_task(rq->curr);
rq 1508 kernel/sched_fair.c enqueue_task_fair(rq, p, 0);
rq 1525 kernel/sched_fair.c resched_task(rq->curr);
rq 1527 kernel/sched_fair.c check_preempt_curr(rq, p, 0);
rq 1542 kernel/sched_fair.c resched_task(rq->curr);
rq 1544 kernel/sched_fair.c check_preempt_curr(rq, p, 0);
rq 1554 kernel/sched_fair.c struct sched_entity *se = &rq->curr->se;
rq 19 kernel/sched_idletask.c resched_task(rq->idle);
rq 24 kernel/sched_idletask.c schedstat_inc(rq, sched_goidle);
rq 26 kernel/sched_idletask.c return rq->idle;
rq 36 kernel/sched_idletask.c spin_unlock_irq(&rq->lock);
rq 39 kernel/sched_idletask.c spin_lock_irq(&rq->lock);
rq 77 kernel/sched_idletask.c resched_task(rq->curr);
rq 79 kernel/sched_idletask.c check_preempt_curr(rq, p, 0);
rq 94 kernel/sched_idletask.c resched_task(rq->curr);
rq 96 kernel/sched_idletask.c check_preempt_curr(rq, p, 0);
rq 10 kernel/sched_rt.c return atomic_read(&rq->rd->rto_count);
rq 15 kernel/sched_rt.c if (!rq->online)
rq 18 kernel/sched_rt.c cpu_set(rq->cpu, rq->rd->rto_mask);
rq 27 kernel/sched_rt.c atomic_inc(&rq->rd->rto_count);
rq 32 kernel/sched_rt.c if (!rq->online)
rq 36 kernel/sched_rt.c atomic_dec(&rq->rd->rto_count);
rq 37 kernel/sched_rt.c cpu_clear(rq->cpu, rq->rd->rto_mask);
rq 42 kernel/sched_rt.c if (rq->rt.rt_nr_migratory && (rq->rt.rt_nr_running > 1)) {
rq 43 kernel/sched_rt.c if (!rq->rt.overloaded) {
rq 44 kernel/sched_rt.c rt_set_overload(rq);
rq 45 kernel/sched_rt.c rq->rt.overloaded = 1;
rq 47 kernel/sched_rt.c } else if (rq->rt.overloaded) {
rq 48 kernel/sched_rt.c rt_clear_overload(rq);
rq 49 kernel/sched_rt.c rq->rt.overloaded = 0;
rq 80 kernel/sched_rt.c list_for_each_entry(rt_rq, &rq->leaf_rt_rq_list, leaf_rt_rq_list)
rq 82 kernel/sched_rt.c static inline struct rq *rq_of_rt_rq(struct rt_rq *rt_rq)
rq 84 kernel/sched_rt.c return rt_rq->rq;
rq 177 kernel/sched_rt.c for (rt_rq = &rq->rt; rt_rq; rt_rq = NULL)
rq 179 kernel/sched_rt.c static inline struct rq *rq_of_rt_rq(struct rt_rq *rt_rq)
rq 181 kernel/sched_rt.c return container_of(rt_rq, struct rq, rt);
rq 187 kernel/sched_rt.c struct rq *rq = task_rq(p);
rq 189 kernel/sched_rt.c return &rq->rt;
rq 294 kernel/sched_rt.c struct root_domain *rd = rq->rd;
rq 300 kernel/sched_rt.c for_each_leaf_rt_rq(rt_rq, rq) {
rq 373 kernel/sched_rt.c spin_lock_irqsave(&rq->lock, flags);
rq 374 kernel/sched_rt.c __disable_runtime(rq);
rq 375 kernel/sched_rt.c spin_unlock_irqrestore(&rq->lock, flags);
rq 388 kernel/sched_rt.c for_each_leaf_rt_rq(rt_rq, rq) {
rq 405 kernel/sched_rt.c spin_lock_irqsave(&rq->lock, flags);
rq 406 kernel/sched_rt.c __enable_runtime(rq);
rq 407 kernel/sched_rt.c spin_unlock_irqrestore(&rq->lock, flags);
rq 441 kernel/sched_rt.c struct rq *rq = rq_of_rt_rq(rt_rq);
rq 443 kernel/sched_rt.c spin_lock(&rq->lock);
rq 464 kernel/sched_rt.c spin_unlock(&rq->lock);
rq 514 kernel/sched_rt.c struct task_struct *curr = rq->curr;
rq 522 kernel/sched_rt.c delta_exec = rq->clock - curr->se.exec_start;
rq 529 kernel/sched_rt.c curr->se.exec_start = rq->clock;
rq 556 kernel/sched_rt.c struct rq *rq = rq_of_rt_rq(rt_rq);
rq 561 kernel/sched_rt.c if (rq->online)
rq 562 kernel/sched_rt.c cpupri_set(&rq->rd->cpupri, rq->cpu,
rq 569 kernel/sched_rt.c struct rq *rq = rq_of_rt_rq(rt_rq);
rq 571 kernel/sched_rt.c rq->rt.rt_nr_migratory++;
rq 613 kernel/sched_rt.c struct rq *rq = rq_of_rt_rq(rt_rq);
rq 614 kernel/sched_rt.c rq->rt.rt_nr_migratory--;
rq 618 kernel/sched_rt.c struct rq *rq = rq_of_rt_rq(rt_rq);
rq 620 kernel/sched_rt.c if (rq->online)
rq 621 kernel/sched_rt.c cpupri_set(&rq->rd->cpupri, rq->cpu,
rq 719 kernel/sched_rt.c inc_cpu_load(rq, p->se.load.weight);
rq 726 kernel/sched_rt.c update_curr_rt(rq);
rq 729 kernel/sched_rt.c dec_cpu_load(rq, p->se.load.weight);
rq 763 kernel/sched_rt.c requeue_task_rt(rq, rq->curr, 0);
rq 771 kernel/sched_rt.c struct rq *rq = task_rq(p);
rq 790 kernel/sched_rt.c if (unlikely(rt_task(rq->curr)) &&
rq 808 kernel/sched_rt.c if (rq->curr->rt.nr_cpus_allowed == 1)
rq 812 kernel/sched_rt.c && cpupri_find(&rq->rd->cpupri, p, &mask))
rq 815 kernel/sched_rt.c if (!cpupri_find(&rq->rd->cpupri, rq->curr, &mask))
rq 823 kernel/sched_rt.c requeue_task_rt(rq, p, 1);
rq 824 kernel/sched_rt.c resched_task(rq->curr);
rq 834 kernel/sched_rt.c if (p->prio < rq->curr->prio) {
rq 835 kernel/sched_rt.c resched_task(rq->curr);
rq 852 kernel/sched_rt.c if (p->prio == rq->curr->prio && !need_resched())
rq 853 kernel/sched_rt.c check_preempt_equal_prio(rq, p);
rq 880 kernel/sched_rt.c rt_rq = &rq->rt;
rq 889 kernel/sched_rt.c rt_se = pick_next_rt_entity(rq, rt_rq);
rq 895 kernel/sched_rt.c p->se.exec_start = rq->clock;
rq 901 kernel/sched_rt.c update_curr_rt(rq);
rq 917 kernel/sched_rt.c if (!task_running(rq, p) &&
rq 933 kernel/sched_rt.c for_each_leaf_rt_rq(rt_rq, rq) {
rq 943 kernel/sched_rt.c if (pick_rt_task(rq, p, cpu)) {
rq 1035 kernel/sched_rt.c static struct rq *find_lock_lowest_rq(struct task_struct *task, struct rq *rq)
rq 1037 kernel/sched_rt.c struct rq *lowest_rq = NULL;
rq 1044 kernel/sched_rt.c if ((cpu == -1) || (cpu == rq->cpu))
rq 1050 kernel/sched_rt.c if (double_lock_balance(rq, lowest_rq)) {
rq 1057 kernel/sched_rt.c if (unlikely(task_rq(task) != rq ||
rq 1060 kernel/sched_rt.c task_running(rq, task) ||
rq 1074 kernel/sched_rt.c double_unlock_balance(rq, lowest_rq);
rq 1089 kernel/sched_rt.c struct rq *lowest_rq;
rq 1093 kernel/sched_rt.c if (!rq->rt.overloaded)
rq 1096 kernel/sched_rt.c next_task = pick_next_highest_task_rt(rq, -1);
rq 1101 kernel/sched_rt.c if (unlikely(next_task == rq->curr)) {
rq 1111 kernel/sched_rt.c if (unlikely(next_task->prio < rq->curr->prio)) {
rq 1112 kernel/sched_rt.c resched_task(rq->curr);
rq 1120 kernel/sched_rt.c lowest_rq = find_lock_lowest_rq(next_task, rq);
rq 1128 kernel/sched_rt.c task = pick_next_highest_task_rt(rq, -1);
rq 1137 kernel/sched_rt.c deactivate_task(rq, next_task, 0);
rq 1143 kernel/sched_rt.c double_unlock_balance(rq, lowest_rq);
rq 1165 kernel/sched_rt.c while (push_rt_task(rq))
rq 1173 kernel/sched_rt.c struct rq *src_rq;
rq 1258 kernel/sched_rt.c if (unlikely(rt_task(prev)) && rq->rt.highest_prio > prev->prio)
rq 1259 kernel/sched_rt.c pull_rt_task(rq);
rq 1271 kernel/sched_rt.c if (unlikely(rq->rt.overloaded)) {
rq 1272 kernel/sched_rt.c spin_lock_irq(&rq->lock);
rq 1273 kernel/sched_rt.c push_rt_tasks(rq);
rq 1274 kernel/sched_rt.c spin_unlock_irq(&rq->lock);
rq 1284 kernel/sched_rt.c if (!task_running(rq, p) &&
rq 1285 kernel/sched_rt.c !test_tsk_need_resched(rq->curr) &&
rq 1286 kernel/sched_rt.c rq->rt.overloaded)
rq 1287 kernel/sched_rt.c push_rt_tasks(rq);
rq 1320 kernel/sched_rt.c struct rq *rq = task_rq(p);
rq 1323 kernel/sched_rt.c rq->rt.rt_nr_migratory++;
rq 1325 kernel/sched_rt.c BUG_ON(!rq->rt.rt_nr_migratory);
rq 1326 kernel/sched_rt.c rq->rt.rt_nr_migratory--;
rq 1329 kernel/sched_rt.c update_rt_migration(rq);
rq 1339 kernel/sched_rt.c if (rq->rt.overloaded)
rq 1340 kernel/sched_rt.c rt_set_overload(rq);
rq 1342 kernel/sched_rt.c __enable_runtime(rq);
rq 1344 kernel/sched_rt.c cpupri_set(&rq->rd->cpupri, rq->cpu, rq->rt.highest_prio);
rq 1350 kernel/sched_rt.c if (rq->rt.overloaded)
rq 1351 kernel/sched_rt.c rt_clear_overload(rq);
rq 1353 kernel/sched_rt.c __disable_runtime(rq);
rq 1355 kernel/sched_rt.c cpupri_set(&rq->rd->cpupri, rq->cpu, CPUPRI_INVALID);
rq 1372 kernel/sched_rt.c if (!rq->rt.rt_nr_running)
rq 1373 kernel/sched_rt.c pull_rt_task(rq);
rq 1396 kernel/sched_rt.c if (rq->rt.overloaded && push_rt_task(rq) &&
rq 1398 kernel/sched_rt.c rq != task_rq(p))
rq 1401 kernel/sched_rt.c if (check_resched && p->prio < rq->curr->prio)
rq 1402 kernel/sched_rt.c resched_task(rq->curr);
rq 1420 kernel/sched_rt.c pull_rt_task(rq);
rq 1427 kernel/sched_rt.c if (p->prio > rq->rt.highest_prio && rq->curr == p)
rq 1440 kernel/sched_rt.c if (p->prio < rq->curr->prio)
rq 1441 kernel/sched_rt.c resched_task(rq->curr);
rq 1467 kernel/sched_rt.c update_curr_rt(rq);
rq 1469 kernel/sched_rt.c watchdog(rq, p);
rq 1488 kernel/sched_rt.c requeue_task_rt(rq, p, 0);
rq 1495 kernel/sched_rt.c struct task_struct *p = rq->curr;
rq 1497 kernel/sched_rt.c p->se.exec_start = rq->clock;
rq 21 kernel/sched_stats.h struct rq *rq = cpu_rq(cpu);
rq 30 kernel/sched_stats.h cpu, rq->yld_both_empty,
rq 31 kernel/sched_stats.h rq->yld_act_empty, rq->yld_exp_empty, rq->yld_count,
rq 32 kernel/sched_stats.h rq->sched_switch, rq->sched_count, rq->sched_goidle,
rq 33 kernel/sched_stats.h rq->ttwu_count, rq->ttwu_local,
rq 34 kernel/sched_stats.h rq->rq_sched_info.cpu_time,
rq 35 kernel/sched_stats.h rq->rq_sched_info.run_delay, rq->rq_sched_info.pcount);
rq 106 kernel/sched_stats.h if (rq) {
rq 107 kernel/sched_stats.h rq->rq_sched_info.run_delay += delta;
rq 108 kernel/sched_stats.h rq->rq_sched_info.pcount++;
rq 118 kernel/sched_stats.h if (rq)
rq 119 kernel/sched_stats.h rq->rq_sched_info.cpu_time += delta;
rq 125 kernel/sched_stats.h if (rq)
rq 126 kernel/sched_stats.h rq->rq_sched_info.run_delay += delta;
rq 128 kernel/sched_stats.h # define schedstat_inc(rq, field) do { (rq)->field++; } while (0)
rq 129 kernel/sched_stats.h # define schedstat_add(rq, field, amt) do { (rq)->field += (amt); } while (0)
rq 247 kernel/sched_stats.h struct rq *rq = task_rq(prev);
rq 254 kernel/sched_stats.h if (prev != rq->idle)
rq 257 kernel/sched_stats.h if (next != rq->idle)
rq 56 kernel/trace/trace_sched_switch.c struct rq *__rq;
rq 110 kernel/trace/trace_sched_switch.c struct rq *__rq;
rq 193 kernel/trace/trace_sched_wakeup.c struct rq *__rq;
rq 295 kernel/trace/trace_sched_wakeup.c struct rq *__rq;
rq 191 net/9p/trans_fd.c struct work_struct rq;
rq 415 net/9p/trans_fd.c INIT_WORK(&m->rq, p9_read_work);
rq 457 net/9p/trans_fd.c cancel_work_sync(&m->rq);
rq 530 net/9p/trans_fd.c queue_work(p9_mux_wq, &m->rq);
rq 705 net/9p/trans_fd.c m = container_of(work, struct p9_conn, rq);
rq 825 net/9p/trans_fd.c queue_work(p9_mux_wq, &m->rq);
rq 458 net/atm/br2684.c struct sk_buff_head *rq;
rq 515 net/atm/br2684.c rq = &sk_atm(atmvcc)->sk_receive_queue;
rq 517 net/atm/br2684.c spin_lock_irqsave(&rq->lock, flags);
rq 518 net/atm/br2684.c if (skb_queue_empty(rq)) {
rq 522 net/atm/br2684.c rq->prev->next = NULL;
rq 523 net/atm/br2684.c skb = rq->next;
rq 525 net/atm/br2684.c rq->prev = rq->next = (struct sk_buff *)rq;
rq 526 net/atm/br2684.c rq->qlen = 0;
rq 527 net/atm/br2684.c spin_unlock_irqrestore(&rq->lock, flags);
rq 455 net/atm/clip.c struct sk_buff_head *rq;
rq 477 net/atm/clip.c rq = &sk_atm(vcc)->sk_receive_queue;
rq 479 net/atm/clip.c spin_lock_irqsave(&rq->lock, flags);
rq 480 net/atm/clip.c if (skb_queue_empty(rq)) {
rq 484 net/atm/clip.c rq->prev->next = NULL;
rq 485 net/atm/clip.c skb = rq->next;
rq 487 net/atm/clip.c rq->prev = rq->next = (struct sk_buff *)rq;
rq 488 net/atm/clip.c rq->qlen = 0;
rq 489 net/atm/clip.c spin_unlock_irqrestore(&rq->lock, flags);
rq 115 net/bridge/br_ioctl.c if (copy_from_user(args, rq->ifr_data, sizeof(args)))
rq 406 net/bridge/br_ioctl.c return old_dev_ioctl(dev, rq, cmd);
rq 410 net/bridge/br_ioctl.c return add_del_if(br, rq->ifr_ifindex, cmd == SIOCBRADDIF);
rq 1866 net/key/af_key.c if (rq->sadb_x_ipsecrequest_mode == 0)
rq 1869 net/key/af_key.c t->id.proto = rq->sadb_x_ipsecrequest_proto; /* XXX check proto */
rq 1870 net/key/af_key.c if ((mode = pfkey_mode_to_xfrm(rq->sadb_x_ipsecrequest_mode)) < 0)
rq 1873 net/key/af_key.c if (rq->sadb_x_ipsecrequest_level == IPSEC_LEVEL_USE)
rq 1875 net/key/af_key.c else if (rq->sadb_x_ipsecrequest_level == IPSEC_LEVEL_UNIQUE) {
rq 1876 net/key/af_key.c t->reqid = rq->sadb_x_ipsecrequest_reqid;
rq 1885 net/key/af_key.c u8 *sa = (u8 *) (rq + 1);
rq 1912 net/key/af_key.c struct sadb_x_ipsecrequest *rq = (void*)(pol+1);
rq 1915 net/key/af_key.c if ((err = parse_ipsecrequest(xp, rq)) < 0)
rq 1917 net/key/af_key.c len -= rq->sadb_x_ipsecrequest_len;
rq 1918 net/key/af_key.c rq = (void*)((u8*)rq + rq->sadb_x_ipsecrequest_len);
rq 2067 net/key/af_key.c struct sadb_x_ipsecrequest *rq;
rq 2080 net/key/af_key.c rq = (void*)skb_put(skb, req_size);
rq 2082 net/key/af_key.c memset(rq, 0, sizeof(*rq));
rq 2083 net/key/af_key.c rq->sadb_x_ipsecrequest_len = req_size;
rq 2084 net/key/af_key.c rq->sadb_x_ipsecrequest_proto = t->id.proto;
rq 2087 net/key/af_key.c rq->sadb_x_ipsecrequest_mode = mode;
rq 2088 net/key/af_key.c rq->sadb_x_ipsecrequest_level = IPSEC_LEVEL_REQUIRE;
rq 2090 net/key/af_key.c rq->sadb_x_ipsecrequest_level = IPSEC_LEVEL_UNIQUE;
rq 2092 net/key/af_key.c rq->sadb_x_ipsecrequest_level = IPSEC_LEVEL_USE;
rq 2093 net/key/af_key.c rq->sadb_x_ipsecrequest_reqid = t->reqid;
rq 2096 net/key/af_key.c u8 *sa = (void *)(rq + 1);
rq 2468 net/key/af_key.c struct sadb_x_ipsecrequest *rq;
rq 2521 net/key/af_key.c rq = (struct sadb_x_ipsecrequest *)(pol + 1);
rq 2528 net/key/af_key.c ret = ipsecrequests_to_migrate(rq, len, &m[i]);
rq 2533 net/key/af_key.c rq = (struct sadb_x_ipsecrequest *)((u8 *)rq + ret);
rq 3368 net/key/af_key.c struct sadb_x_ipsecrequest *rq;
rq 3376 net/key/af_key.c rq = (struct sadb_x_ipsecrequest *)skb_put(skb, size_req);
rq 3377 net/key/af_key.c memset(rq, 0, size_req);
rq 3378 net/key/af_key.c rq->sadb_x_ipsecrequest_len = size_req;
rq 3379 net/key/af_key.c rq->sadb_x_ipsecrequest_proto = proto;
rq 3380 net/key/af_key.c rq->sadb_x_ipsecrequest_mode = mode;
rq 3381 net/key/af_key.c rq->sadb_x_ipsecrequest_level = level;
rq 3382 net/key/af_key.c rq->sadb_x_ipsecrequest_reqid = reqid;
rq 3384 net/key/af_key.c sa = (u8 *) (rq + 1);
rq 698 net/sunrpc/cache.c struct cache_request *rq;
rq 722 net/sunrpc/cache.c rq = container_of(rp->q.list.next, struct cache_request, q.list);
rq 723 net/sunrpc/cache.c BUG_ON(rq->q.reader);
rq 725 net/sunrpc/cache.c rq->readers++;
rq 728 net/sunrpc/cache.c if (rp->offset == 0 && !test_bit(CACHE_PENDING, &rq->item->flags)) {
rq 731 net/sunrpc/cache.c list_move(&rp->q.list, &rq->q.list);
rq 734 net/sunrpc/cache.c if (rp->offset + count > rq->len)
rq 735 net/sunrpc/cache.c count = rq->len - rp->offset;
rq 737 net/sunrpc/cache.c if (copy_to_user(buf, rq->buf + rp->offset, count))
rq 740 net/sunrpc/cache.c if (rp->offset >= rq->len) {
rq 743 net/sunrpc/cache.c list_move(&rp->q.list, &rq->q.list);
rq 752 net/sunrpc/cache.c rq->readers--;
rq 753 net/sunrpc/cache.c if (rq->readers == 0 &&
rq 754 net/sunrpc/cache.c !test_bit(CACHE_PENDING, &rq->item->flags)) {
rq 755 net/sunrpc/cache.c list_del(&rq->q.list);
rq 757 net/sunrpc/cache.c cache_put(rq->item, cd);
rq 758 net/sunrpc/cache.c kfree(rq->buf);
rq 759 net/sunrpc/cache.c kfree(rq);
rq 227 net/sunrpc/xprtrdma/xprt_rdma.h (rpcx_to_rdmad(rq->rq_task->tk_xprt).inline_rsize)
rq 230 net/sunrpc/xprtrdma/xprt_rdma.h (rpcx_to_rdmad(rq->rq_task->tk_xprt).inline_wsize)
rq 233 net/sunrpc/xprtrdma/xprt_rdma.h rpcx_to_rdmad(rq->rq_task->tk_xprt).padding