bio 1249 block/as-iosched.c sector_t rb_key = bio->bi_sector + bio_sectors(bio);
bio 1255 block/as-iosched.c __rq = elv_rb_find(&ad->sort_list[bio_data_dir(bio)], rb_key);
bio 1256 block/as-iosched.c if (__rq && elv_rq_merge_ok(__rq, bio)) {
bio 168 block/blk-barrier.c if (bio_data_dir(q->orig_bar_rq->bio) == WRITE)
bio 172 block/blk-barrier.c init_request_from_bio(rq, q->orig_bar_rq->bio);
bio 258 block/blk-barrier.c set_bit(BIO_EOPNOTSUPP, &bio->bi_flags);
bio 259 block/blk-barrier.c clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio 262 block/blk-barrier.c complete(bio->bi_private);
bio 279 block/blk-barrier.c struct bio *bio;
bio 289 block/blk-barrier.c bio = bio_alloc(GFP_KERNEL, 0);
bio 290 block/blk-barrier.c if (!bio)
bio 293 block/blk-barrier.c bio->bi_end_io = bio_end_empty_barrier;
bio 294 block/blk-barrier.c bio->bi_private = &wait;
bio 295 block/blk-barrier.c bio->bi_bdev = bdev;
bio 296 block/blk-barrier.c submit_bio(WRITE_BARRIER, bio);
bio 306 block/blk-barrier.c *error_sector = bio->bi_sector;
bio 309 block/blk-barrier.c if (bio_flagged(bio, BIO_EOPNOTSUPP))
bio 311 block/blk-barrier.c else if (!bio_flagged(bio, BIO_UPTODATE))
bio 314 block/blk-barrier.c bio_put(bio);
bio 323 block/blk-barrier.c set_bit(BIO_EOPNOTSUPP, &bio->bi_flags);
bio 324 block/blk-barrier.c clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio 327 block/blk-barrier.c bio_put(bio);
bio 344 block/blk-barrier.c struct bio *bio;
bio 358 block/blk-barrier.c bio = bio_alloc(gfp_mask, 0);
bio 359 block/blk-barrier.c if (!bio)
bio 362 block/blk-barrier.c bio->bi_end_io = blkdev_discard_end_io;
bio 363 block/blk-barrier.c bio->bi_bdev = bdev;
bio 365 block/blk-barrier.c bio->bi_sector = sector;
bio 368 block/blk-barrier.c bio->bi_size = q->max_hw_sectors << 9;
bio 372 block/blk-barrier.c bio->bi_size = nr_sects << 9;
bio 375 block/blk-barrier.c bio_get(bio);
bio 376 block/blk-barrier.c submit_bio(DISCARD_BARRIER, bio);
bio 379 block/blk-barrier.c if (bio_flagged(bio, BIO_EOPNOTSUPP))
bio 381 block/blk-barrier.c else if (!bio_flagged(bio, BIO_UPTODATE))
bio 383 block/blk-barrier.c bio_put(bio);
bio 132 block/blk-core.c clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio 133 block/blk-core.c else if (!test_bit(BIO_UPTODATE, &bio->bi_flags))
bio 136 block/blk-core.c if (unlikely(nbytes > bio->bi_size)) {
bio 138 block/blk-core.c __func__, nbytes, bio->bi_size);
bio 139 block/blk-core.c nbytes = bio->bi_size;
bio 142 block/blk-core.c bio->bi_size -= nbytes;
bio 143 block/blk-core.c bio->bi_sector += (nbytes >> 9);
bio 145 block/blk-core.c if (bio_integrity(bio))
bio 146 block/blk-core.c bio_integrity_advance(bio, nbytes);
bio 148 block/blk-core.c if (bio->bi_size == 0)
bio 149 block/blk-core.c bio_endio(bio, error);
bio 174 block/blk-core.c rq->bio, rq->biotail,
bio 811 block/blk-core.c blk_add_trace_generic(q, bio, rw, BLK_TA_GETRQ);
bio 828 block/blk-core.c rq = get_request(q, rw_flags, bio, GFP_NOIO);
bio 837 block/blk-core.c blk_add_trace_generic(q, bio, rw, BLK_TA_SLEEPRQ);
bio 855 block/blk-core.c rq = get_request(q, rw_flags, bio, GFP_NOIO);
bio 1072 block/blk-core.c req->cpu = bio->bi_comp_cpu;
bio 1078 block/blk-core.c if (bio_rw_ahead(bio) || bio_failfast(bio))
bio 1084 block/blk-core.c if (unlikely(bio_discard(bio))) {
bio 1086 block/blk-core.c if (bio_barrier(bio))
bio 1089 block/blk-core.c } else if (unlikely(bio_barrier(bio)))
bio 1092 block/blk-core.c if (bio_sync(bio))
bio 1094 block/blk-core.c if (bio_rw_meta(bio))
bio 1098 block/blk-core.c req->hard_sector = req->sector = bio->bi_sector;
bio 1099 block/blk-core.c req->ioprio = bio_prio(bio);
bio 1101 block/blk-core.c blk_rq_bio_prep(req->q, req, bio);
bio 1108 block/blk-core.c const unsigned short prio = bio_prio(bio);
bio 1109 block/blk-core.c const int sync = bio_sync(bio);
bio 1112 block/blk-core.c nr_sectors = bio_sectors(bio);
bio 1119 block/blk-core.c blk_queue_bounce(q, &bio);
bio 1121 block/blk-core.c barrier = bio_barrier(bio);
bio 1122 block/blk-core.c if (unlikely(barrier) && bio_has_data(bio) &&
bio 1128 block/blk-core.c discard = bio_discard(bio);
bio 1139 block/blk-core.c el_ret = elv_merge(q, &req, bio);
bio 1144 block/blk-core.c if (!ll_back_merge_fn(q, req, bio))
bio 1147 block/blk-core.c blk_add_trace_bio(q, bio, BLK_TA_BACKMERGE);
bio 1149 block/blk-core.c req->biotail->bi_next = bio;
bio 1150 block/blk-core.c req->biotail = bio;
bio 1154 block/blk-core.c req->cpu = bio->bi_comp_cpu;
bio 1163 block/blk-core.c if (!ll_front_merge_fn(q, req, bio))
bio 1166 block/blk-core.c blk_add_trace_bio(q, bio, BLK_TA_FRONTMERGE);
bio 1168 block/blk-core.c bio->bi_next = req->bio;
bio 1169 block/blk-core.c req->bio = bio;
bio 1176 block/blk-core.c req->buffer = bio_data(bio);
bio 1177 block/blk-core.c req->current_nr_sectors = bio_cur_sectors(bio);
bio 1179 block/blk-core.c req->sector = req->hard_sector = bio->bi_sector;
bio 1183 block/blk-core.c req->cpu = bio->bi_comp_cpu;
bio 1200 block/blk-core.c rw_flags = bio_data_dir(bio);
bio 1208 block/blk-core.c req = get_request_wait(q, rw_flags, bio);
bio 1216 block/blk-core.c init_request_from_bio(req, bio);
bio 1220 block/blk-core.c bio_flagged(bio, BIO_CPU_AFFINE))
bio 1232 block/blk-core.c bio_endio(bio, err);
bio 1241 block/blk-core.c struct block_device *bdev = bio->bi_bdev;
bio 1243 block/blk-core.c if (bio_sectors(bio) && bdev != bdev->bd_contains) {
bio 1246 block/blk-core.c bio->bi_sector += p->start_sect;
bio 1247 block/blk-core.c bio->bi_bdev = bdev->bd_contains;
bio 1249 block/blk-core.c blk_add_trace_remap(bdev_get_queue(bio->bi_bdev), bio,
bio 1250 block/blk-core.c bdev->bd_dev, bio->bi_sector,
bio 1251 block/blk-core.c bio->bi_sector - p->start_sect);
bio 1261 block/blk-core.c bdevname(bio->bi_bdev, b),
bio 1262 block/blk-core.c bio->bi_rw,
bio 1263 block/blk-core.c (unsigned long long)bio->bi_sector + bio_sectors(bio),
bio 1264 block/blk-core.c (long long)(bio->bi_bdev->bd_inode->i_size >> 9));
bio 1266 block/blk-core.c set_bit(BIO_EOF, &bio->bi_flags);
bio 1281 block/blk-core.c struct hd_struct *part = bio->bi_bdev->bd_part;
bio 1284 block/blk-core.c return should_fail(&fail_make_request, bio->bi_size);
bio 1317 block/blk-core.c maxsector = bio->bi_bdev->bd_inode->i_size >> 9;
bio 1319 block/blk-core.c sector_t sector = bio->bi_sector;
bio 1327 block/blk-core.c handle_bad_sector(bio);
bio 1363 block/blk-core.c int ret, nr_sectors = bio_sectors(bio);
bio 1369 block/blk-core.c if (bio_check_eod(bio, nr_sectors))
bio 1385 block/blk-core.c q = bdev_get_queue(bio->bi_bdev);
bio 1390 block/blk-core.c bdevname(bio->bi_bdev, b),
bio 1391 block/blk-core.c (long long) bio->bi_sector);
bio 1393 block/blk-core.c bio_endio(bio, err);
bio 1399 block/blk-core.c bdevname(bio->bi_bdev, b),
bio 1400 block/blk-core.c bio_sectors(bio),
bio 1408 block/blk-core.c if (should_fail_request(bio))
bio 1415 block/blk-core.c blk_partition_remap(bio);
bio 1417 block/blk-core.c if (bio_integrity_enabled(bio) && bio_integrity_prep(bio))
bio 1421 block/blk-core.c blk_add_trace_remap(q, bio, old_dev, bio->bi_sector,
bio 1424 block/blk-core.c blk_add_trace_bio(q, bio, BLK_TA_QUEUE);
bio 1426 block/blk-core.c old_sector = bio->bi_sector;
bio 1427 block/blk-core.c old_dev = bio->bi_bdev->bd_dev;
bio 1429 block/blk-core.c if (bio_check_eod(bio, nr_sectors))
bio 1431 block/blk-core.c if ((bio_empty_barrier(bio) && !q->prepare_flush_fn) ||
bio 1432 block/blk-core.c (bio_discard(bio) && !q->prepare_discard_fn)) {
bio 1437 block/blk-core.c ret = q->make_request_fn(q, bio);
bio 1456 block/blk-core.c *(current->bio_tail) = bio;
bio 1457 block/blk-core.c bio->bi_next = NULL;
bio 1458 block/blk-core.c current->bio_tail = &bio->bi_next;
bio 1479 block/blk-core.c BUG_ON(bio->bi_next);
bio 1481 block/blk-core.c current->bio_list = bio->bi_next;
bio 1482 block/blk-core.c if (bio->bi_next == NULL)
bio 1485 block/blk-core.c bio->bi_next = NULL;
bio 1486 block/blk-core.c __generic_make_request(bio);
bio 1487 block/blk-core.c bio = current->bio_list;
bio 1488 block/blk-core.c } while (bio);
bio 1505 block/blk-core.c int count = bio_sectors(bio);
bio 1507 block/blk-core.c bio->bi_rw |= rw;
bio 1513 block/blk-core.c if (bio_has_data(bio)) {
bio 1517 block/blk-core.c task_io_account_read(bio->bi_size);
bio 1526 block/blk-core.c (unsigned long long)bio->bi_sector,
bio 1527 block/blk-core.c bdevname(bio->bi_bdev, b));
bio 1531 block/blk-core.c generic_make_request(bio);
bio 1634 block/blk-core.c struct bio *bio;
bio 1663 block/blk-core.c while ((bio = req->bio) != NULL) {
bio 1672 block/blk-core.c bio->bi_sector = req->sector;
bio 1674 block/blk-core.c if (nr_bytes >= bio->bi_size) {
bio 1675 block/blk-core.c req->bio = bio->bi_next;
bio 1676 block/blk-core.c nbytes = bio->bi_size;
bio 1677 block/blk-core.c req_bio_endio(req, bio, nbytes, error);
bio 1681 block/blk-core.c int idx = bio->bi_idx + next_idx;
bio 1683 block/blk-core.c if (unlikely(bio->bi_idx >= bio->bi_vcnt)) {
bio 1686 block/blk-core.c __func__, bio->bi_idx, bio->bi_vcnt);
bio 1690 block/blk-core.c nbytes = bio_iovec_idx(bio, idx)->bv_len;
bio 1691 block/blk-core.c BIO_BUG_ON(nbytes > bio->bi_size);
bio 1712 block/blk-core.c bio = req->bio;
bio 1713 block/blk-core.c if (bio) {
bio 1725 block/blk-core.c if (!req->bio)
bio 1732 block/blk-core.c req_bio_endio(req, bio, bio_nbytes, error);
bio 1733 block/blk-core.c bio->bi_idx += next_idx;
bio 1734 block/blk-core.c bio_iovec(bio)->bv_offset += nr_bytes;
bio 1735 block/blk-core.c bio_iovec(bio)->bv_len -= nr_bytes;
bio 1815 block/blk-core.c if (rq->bio)
bio 1816 block/blk-core.c return rq->bio->bi_size;
bio 1852 block/blk-core.c if (rq->bio) {
bio 1942 block/blk-core.c if (rq->bio && __end_that_request_first(rq, error, nr_bytes))
bio 2042 block/blk-core.c rq->cmd_flags |= (bio->bi_rw & 3);
bio 2044 block/blk-core.c if (bio_has_data(bio)) {
bio 2045 block/blk-core.c rq->nr_phys_segments = bio_phys_segments(q, bio);
bio 2046 block/blk-core.c rq->buffer = bio_data(bio);
bio 2048 block/blk-core.c rq->current_nr_sectors = bio_cur_sectors(bio);
bio 2050 block/blk-core.c rq->hard_nr_sectors = rq->nr_sectors = bio_sectors(bio);
bio 2051 block/blk-core.c rq->data_len = bio->bi_size;
bio 2053 block/blk-core.c rq->bio = rq->biotail = bio;
bio 2055 block/blk-core.c if (bio->bi_bdev)
bio 2056 block/blk-core.c rq->rq_disk = bio->bi_bdev->bd_disk;
bio 15 block/blk-map.c if (!rq->bio)
bio 16 block/blk-map.c blk_rq_bio_prep(q, rq, bio);
bio 17 block/blk-map.c else if (!ll_back_merge_fn(q, rq, bio))
bio 20 block/blk-map.c rq->biotail->bi_next = bio;
bio 21 block/blk-map.c rq->biotail = bio;
bio 23 block/blk-map.c rq->data_len += bio->bi_size;
bio 33 block/blk-map.c if (bio) {
bio 34 block/blk-map.c if (bio_flagged(bio, BIO_USER_MAPPED))
bio 35 block/blk-map.c bio_unmap_user(bio);
bio 37 block/blk-map.c ret = bio_uncopy_user(bio);
bio 48 block/blk-map.c struct bio *bio, *orig_bio;
bio 59 block/blk-map.c bio = bio_map_user(q, NULL, uaddr, len, reading, gfp_mask);
bio 61 block/blk-map.c bio = bio_copy_user(q, map_data, uaddr, len, reading, gfp_mask);
bio 63 block/blk-map.c if (IS_ERR(bio))
bio 64 block/blk-map.c return PTR_ERR(bio);
bio 67 block/blk-map.c bio->bi_flags |= (1 << BIO_NULL_MAPPED);
bio 69 block/blk-map.c orig_bio = bio;
bio 70 block/blk-map.c blk_queue_bounce(q, &bio);
bio 76 block/blk-map.c bio_get(bio);
bio 78 block/blk-map.c ret = blk_rq_append_bio(q, rq, bio);
bio 80 block/blk-map.c return bio->bi_size;
bio 83 block/blk-map.c bio_endio(bio, 0);
bio 85 block/blk-map.c bio_put(bio);
bio 116 block/blk-map.c struct bio *bio = NULL;
bio 149 block/blk-map.c if (!bio)
bio 150 block/blk-map.c bio = rq->bio;
bio 155 block/blk-map.c if (!bio_flagged(bio, BIO_USER_MAPPED))
bio 161 block/blk-map.c blk_rq_unmap_user(bio);
bio 162 block/blk-map.c rq->bio = NULL;
bio 194 block/blk-map.c struct bio *bio;
bio 211 block/blk-map.c bio = bio_copy_user_iov(q, map_data, iov, iov_count, read,
bio 214 block/blk-map.c bio = bio_map_user_iov(q, NULL, iov, iov_count, read, gfp_mask);
bio 216 block/blk-map.c if (IS_ERR(bio))
bio 217 block/blk-map.c return PTR_ERR(bio);
bio 219 block/blk-map.c if (bio->bi_size != len) {
bio 220 block/blk-map.c bio_endio(bio, 0);
bio 221 block/blk-map.c bio_unmap_user(bio);
bio 225 block/blk-map.c if (!bio_flagged(bio, BIO_USER_MAPPED))
bio 228 block/blk-map.c blk_queue_bounce(q, &bio);
bio 229 block/blk-map.c bio_get(bio);
bio 230 block/blk-map.c blk_rq_bio_prep(q, rq, bio);
bio 247 block/blk-map.c struct bio *mapped_bio;
bio 250 block/blk-map.c while (bio) {
bio 251 block/blk-map.c mapped_bio = bio;
bio 252 block/blk-map.c if (unlikely(bio_flagged(bio, BIO_BOUNCED)))
bio 253 block/blk-map.c mapped_bio = bio->bi_private;
bio 259 block/blk-map.c mapped_bio = bio;
bio 260 block/blk-map.c bio = bio->bi_next;
bio 285 block/blk-map.c struct bio *bio;
bio 294 block/blk-map.c bio = bio_copy_kern(q, kbuf, len, gfp_mask, reading);
bio 296 block/blk-map.c bio = bio_map_kern(q, kbuf, len, gfp_mask);
bio 298 block/blk-map.c if (IS_ERR(bio))
bio 299 block/blk-map.c return PTR_ERR(bio);
bio 302 block/blk-map.c bio->bi_rw |= (1 << BIO_RW);
bio 307 block/blk-map.c blk_rq_bio_prep(q, rq, bio);
bio 308 block/blk-map.c blk_queue_bounce(q, &rq->bio);
bio 25 block/blk-merge.c rq->hard_cur_sectors = bio_cur_sectors(rq->bio);
bio 27 block/blk-merge.c rq->buffer = bio_data(rq->bio);
bio 52 block/blk-merge.c if (!rq->bio)
bio 92 block/blk-merge.c struct bio *nxt = bio->bi_next;
bio 94 block/blk-merge.c rq.bio = rq.biotail = bio;
bio 95 block/blk-merge.c bio->bi_next = NULL;
bio 97 block/blk-merge.c bio->bi_next = nxt;
bio 98 block/blk-merge.c bio->bi_phys_segments = rq.nr_phys_segments;
bio 99 block/blk-merge.c bio->bi_flags |= (1 << BIO_SEG_VALID);
bio 109 block/blk-merge.c if (bio->bi_size + nxt->bi_size > q->max_segment_size)
bio 112 block/blk-merge.c if (!bio_has_data(bio))
bio 115 block/blk-merge.c if (!BIOVEC_PHYS_MERGEABLE(__BVEC_END(bio), __BVEC_START(nxt)))
bio 122 block/blk-merge.c if (BIO_SEG_BOUNDARY(q, bio, nxt))
bio 220 block/blk-merge.c int nr_phys_segs = bio_phys_segments(q, bio);
bio 241 block/blk-merge.c int nr_phys_segs = bio_phys_segments(q, bio);
bio 269 block/blk-merge.c if (req->nr_sectors + bio_sectors(bio) > max_sectors) {
bio 277 block/blk-merge.c if (!bio_flagged(bio, BIO_SEG_VALID))
bio 278 block/blk-merge.c blk_recount_segments(q, bio);
bio 280 block/blk-merge.c return ll_new_hw_segment(q, req, bio);
bio 294 block/blk-merge.c if (req->nr_sectors + bio_sectors(bio) > max_sectors) {
bio 300 block/blk-merge.c if (!bio_flagged(bio, BIO_SEG_VALID))
bio 301 block/blk-merge.c blk_recount_segments(q, bio);
bio 302 block/blk-merge.c if (!bio_flagged(req->bio, BIO_SEG_VALID))
bio 303 block/blk-merge.c blk_recount_segments(q, req->bio);
bio 305 block/blk-merge.c return ll_new_hw_segment(q, req, bio);
bio 327 block/blk-merge.c if (blk_phys_contig_segment(q, req->biotail, next->bio))
bio 382 block/blk-merge.c req->biotail->bi_next = next->bio;
bio 93 block/blk.h __rq_for_each_bio(_iter.bio, _rq) \
bio 94 block/blk.h bip_for_each_vec(bvl, _iter.bio->bi_integrity, _iter.i)
bio 83 block/bsg.c struct bio *bio;
bio 84 block/bsg.c struct bio *bidi_bio;
bio 313 block/bsg.c blk_rq_unmap_user(next_rq->bio);
bio 330 block/bsg.c bd->name, rq, bc, bc->bio, uptodate);
bio 356 block/bsg.c bc->bio = rq->bio;
bio 358 block/bsg.c bc->bidi_bio = rq->next_rq->bio;
bio 420 block/bsg.c dprintk("rq %p bio %p %u\n", rq, bio, rq->errors);
bio 463 block/bsg.c blk_rq_unmap_user(bio);
bio 509 block/bsg.c tret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio,
bio 544 block/bsg.c ret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio,
bio 920 block/bsg.c struct bio *bio, *bidi_bio = NULL;
bio 930 block/bsg.c bio = rq->bio;
bio 932 block/bsg.c bidi_bio = rq->next_rq->bio;
bio 934 block/bsg.c ret = blk_complete_sgv4_hdr_rq(rq, &hdr, bio, bidi_bio);
bio 241 block/cfq-iosched.c if (bio_data_dir(bio) == READ || bio_sync(bio))
bio 647 block/cfq-iosched.c cfqq = cic_to_cfqq(cic, cfq_bio_sync(bio));
bio 649 block/cfq-iosched.c sector_t sector = bio->bi_sector + bio_sectors(bio);
bio 701 block/cfq-iosched.c __rq = cfq_find_rq_fmerge(cfqd, bio);
bio 702 block/cfq-iosched.c if (__rq && elv_rq_merge_ok(__rq, bio)) {
bio 744 block/cfq-iosched.c if (cfq_bio_sync(bio) && !rq_is_sync(rq))
bio 755 block/cfq-iosched.c cfqq = cic_to_cfqq(cic, cfq_bio_sync(bio));
bio 137 block/deadline-iosched.c sector_t sector = bio->bi_sector + bio_sectors(bio);
bio 139 block/deadline-iosched.c __rq = elv_rb_find(&dd->sort_list[bio_data_dir(bio)], sector);
bio 143 block/deadline-iosched.c if (elv_rq_merge_ok(__rq, bio)) {
bio 65 block/elevator.c return e->ops->elevator_allow_merge_fn(q, rq, bio);
bio 81 block/elevator.c if (bio_discard(bio) != bio_discard(rq->bio))
bio 87 block/elevator.c if (bio_data_dir(bio) != rq_data_dir(rq))
bio 93 block/elevator.c if (rq->rq_disk != bio->bi_bdev->bd_disk || rq->special)
bio 99 block/elevator.c if (bio_integrity(bio) != blk_integrity_rq(rq))
bio 102 block/elevator.c if (!elv_iosched_allow_merge(rq, bio))
bio 116 block/elevator.c if (elv_rq_merge_ok(__rq, bio)) {
bio 117 block/elevator.c if (__rq->sector + __rq->nr_sectors == bio->bi_sector)
bio 119 block/elevator.c else if (__rq->sector - bio_sectors(bio) == bio->bi_sector)
bio 499 block/elevator.c ret = elv_try_merge(q->last_merge, bio);
bio 512 block/elevator.c __rq = elv_rqhash_find(q, bio->bi_sector);
bio 513 block/elevator.c if (__rq && elv_rq_merge_ok(__rq, bio)) {
bio 519 block/elevator.c return e->ops->elevator_merge_fn(q, req, bio);
bio 120 block/ioctl.c set_bit(BIO_EOPNOTSUPP, &bio->bi_flags);
bio 121 block/ioctl.c clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio 123 block/ioctl.c complete(bio->bi_private);
bio 147 block/ioctl.c struct bio *bio;
bio 149 block/ioctl.c bio = bio_alloc(GFP_KERNEL, 0);
bio 150 block/ioctl.c if (!bio)
bio 153 block/ioctl.c bio->bi_end_io = blk_ioc_discard_endio;
bio 154 block/ioctl.c bio->bi_bdev = bdev;
bio 155 block/ioctl.c bio->bi_private = &wait;
bio 156 block/ioctl.c bio->bi_sector = start;
bio 159 block/ioctl.c bio->bi_size = q->max_hw_sectors << 9;
bio 163 block/ioctl.c bio->bi_size = len << 9;
bio 166 block/ioctl.c submit_bio(DISCARD_NOBARRIER, bio);
bio 170 block/ioctl.c if (bio_flagged(bio, BIO_EOPNOTSUPP))
bio 172 block/ioctl.c else if (!bio_flagged(bio, BIO_UPTODATE))
bio 174 block/ioctl.c bio_put(bio);
bio 222 block/scsi_ioctl.c blk_rq_unmap_user(rq->bio);
bio 255 block/scsi_ioctl.c rq->bio = bio;
bio 270 block/scsi_ioctl.c struct bio *bio;
bio 327 block/scsi_ioctl.c bio = rq->bio;
bio 343 block/scsi_ioctl.c return blk_complete_sghdr_rq(rq, hdr, bio);
bio 51 fs/bio-integrity.c BUG_ON(bio == NULL);
bio 70 fs/bio-integrity.c bip->bip_bio = bio;
bio 71 fs/bio-integrity.c bio->bi_integrity = bip;
bio 91 fs/bio-integrity.c return bio_integrity_alloc_bioset(bio, gfp_mask, nr_vecs, fs_bio_set);
bio 105 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity;
bio 110 fs/bio-integrity.c if (!bio_flagged(bio, BIO_CLONED) && !bio_flagged(bio, BIO_FS_INTEGRITY)
bio 117 fs/bio-integrity.c bio->bi_integrity = NULL;
bio 133 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity;
bio 184 fs/bio-integrity.c if (bio_integrity(bio))
bio 187 fs/bio-integrity.c return bdev_integrity_enabled(bio->bi_bdev, bio_data_dir(bio));
bio 221 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio 223 fs/bio-integrity.c BUG_ON(bio->bi_size == 0);
bio 225 fs/bio-integrity.c return bi->tag_size * (bio->bi_size / bi->sector_size);
bio 231 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity;
bio 232 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio 270 fs/bio-integrity.c BUG_ON(bio_data_dir(bio) != WRITE);
bio 272 fs/bio-integrity.c return bio_integrity_tag(bio, tag_buf, len, 1);
bio 288 fs/bio-integrity.c BUG_ON(bio_data_dir(bio) != READ);
bio 290 fs/bio-integrity.c return bio_integrity_tag(bio, tag_buf, len, 0);
bio 305 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio 308 fs/bio-integrity.c sector_t sector = bio->bi_sector;
bio 310 fs/bio-integrity.c void *prot_buf = bio->bi_integrity->bip_buf;
bio 313 fs/bio-integrity.c bix.disk_name = bio->bi_bdev->bd_disk->disk_name;
bio 316 fs/bio-integrity.c bio_for_each_segment(bv, bio, i) {
bio 329 fs/bio-integrity.c BUG_ON(total > bio->bi_integrity->bip_size);
bio 365 fs/bio-integrity.c bi = bdev_get_integrity(bio->bi_bdev);
bio 366 fs/bio-integrity.c q = bdev_get_queue(bio->bi_bdev);
bio 368 fs/bio-integrity.c BUG_ON(bio_integrity(bio));
bio 370 fs/bio-integrity.c sectors = bio_integrity_hw_sectors(bi, bio_sectors(bio));
bio 385 fs/bio-integrity.c bip = bio_integrity_alloc(bio, GFP_NOIO, nr_pages);
bio 394 fs/bio-integrity.c bip->bip_sector = bio->bi_sector;
bio 408 fs/bio-integrity.c ret = bio_integrity_add_page(bio, virt_to_page(buf),
bio 423 fs/bio-integrity.c if (bio_data_dir(bio) == READ) {
bio 424 fs/bio-integrity.c bip->bip_end_io = bio->bi_end_io;
bio 425 fs/bio-integrity.c bio->bi_end_io = bio_integrity_endio;
bio 429 fs/bio-integrity.c if (bio_data_dir(bio) == WRITE)
bio 430 fs/bio-integrity.c bio_integrity_generate(bio);
bio 446 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio 449 fs/bio-integrity.c sector_t sector = bio->bi_integrity->bip_sector;
bio 451 fs/bio-integrity.c void *prot_buf = bio->bi_integrity->bip_buf;
bio 454 fs/bio-integrity.c bix.disk_name = bio->bi_bdev->bd_disk->disk_name;
bio 457 fs/bio-integrity.c bio_for_each_segment(bv, bio, i) {
bio 475 fs/bio-integrity.c BUG_ON(total > bio->bi_integrity->bip_size);
bio 495 fs/bio-integrity.c struct bio *bio = bip->bip_bio;
bio 498 fs/bio-integrity.c if (bio_integrity_verify(bio)) {
bio 499 fs/bio-integrity.c clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio 504 fs/bio-integrity.c bio->bi_end_io = bip->bip_end_io;
bio 506 fs/bio-integrity.c if (bio->bi_end_io)
bio 507 fs/bio-integrity.c bio->bi_end_io(bio, error);
bio 524 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity;
bio 526 fs/bio-integrity.c BUG_ON(bip->bip_bio != bio);
bio 595 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity;
bio 596 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio 621 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity;
bio 622 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio 627 fs/bio-integrity.c BUG_ON(!bio_flagged(bio, BIO_CLONED));
bio 647 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity;
bio 650 fs/bio-integrity.c if (bio_integrity(bio) == 0)
bio 653 fs/bio-integrity.c bi = bdev_get_integrity(bio->bi_bdev);
bio 672 fs/bio-integrity.c bp->bip1.bip_sector = bio->bi_integrity->bip_sector;
bio 673 fs/bio-integrity.c bp->bip2.bip_sector = bio->bi_integrity->bip_sector + nr_sectors;
bio 696 fs/bio-integrity.c bip = bio_integrity_alloc_bioset(bio, GFP_NOIO, bip_src->bip_vcnt, bs);
bio 109 fs/bio.c if (bio->bi_io_vec) {
bio 110 fs/bio.c const int pool_idx = BIO_POOL_IDX(bio);
bio 114 fs/bio.c mempool_free(bio->bi_io_vec, bio_set->bvec_pools[pool_idx]);
bio 117 fs/bio.c if (bio_integrity(bio))
bio 118 fs/bio.c bio_integrity_free(bio, bio_set);
bio 120 fs/bio.c mempool_free(bio, bio_set->bio_pool);
bio 128 fs/bio.c bio_free(bio, fs_bio_set);
bio 133 fs/bio.c kfree(bio->bi_io_vec);
bio 134 fs/bio.c kfree(bio);
bio 139 fs/bio.c memset(bio, 0, sizeof(*bio));
bio 140 fs/bio.c bio->bi_flags = 1 << BIO_UPTODATE;
bio 141 fs/bio.c bio->bi_comp_cpu = -1;
bio 142 fs/bio.c atomic_set(&bio->bi_cnt, 1);
bio 160 fs/bio.c struct bio *bio_alloc_bioset(gfp_t gfp_mask, int nr_iovecs, struct bio_set *bs)
bio 162 fs/bio.c struct bio *bio;
bio 165 fs/bio.c bio = mempool_alloc(bs->bio_pool, gfp_mask);
bio 167 fs/bio.c bio = kmalloc(sizeof(*bio), gfp_mask);
bio 169 fs/bio.c if (likely(bio)) {
bio 172 fs/bio.c bio_init(bio);
bio 179 fs/bio.c mempool_free(bio, bs->bio_pool);
bio 181 fs/bio.c kfree(bio);
bio 182 fs/bio.c bio = NULL;
bio 185 fs/bio.c bio->bi_flags |= idx << BIO_POOL_OFFSET;
bio 186 fs/bio.c bio->bi_max_vecs = bvec_nr_vecs(idx);
bio 188 fs/bio.c bio->bi_io_vec = bvl;
bio 191 fs/bio.c return bio;
bio 194 fs/bio.c struct bio *bio_alloc(gfp_t gfp_mask, int nr_iovecs)
bio 196 fs/bio.c struct bio *bio = bio_alloc_bioset(gfp_mask, nr_iovecs, fs_bio_set);
bio 198 fs/bio.c if (bio)
bio 199 fs/bio.c bio->bi_destructor = bio_fs_destructor;
bio 201 fs/bio.c return bio;
bio 211 fs/bio.c struct bio *bio_kmalloc(gfp_t gfp_mask, int nr_iovecs)
bio 213 fs/bio.c struct bio *bio = bio_alloc_bioset(gfp_mask, nr_iovecs, NULL);
bio 215 fs/bio.c if (bio)
bio 216 fs/bio.c bio->bi_destructor = bio_kmalloc_destructor;
bio 218 fs/bio.c return bio;
bio 227 fs/bio.c bio_for_each_segment(bv, bio, i) {
bio 246 fs/bio.c BIO_BUG_ON(!atomic_read(&bio->bi_cnt));
bio 251 fs/bio.c if (atomic_dec_and_test(&bio->bi_cnt)) {
bio 252 fs/bio.c bio->bi_next = NULL;
bio 253 fs/bio.c bio->bi_destructor(bio);
bio 259 fs/bio.c if (unlikely(!bio_flagged(bio, BIO_SEG_VALID)))
bio 260 fs/bio.c blk_recount_segments(q, bio);
bio 262 fs/bio.c return bio->bi_phys_segments;
bio 276 fs/bio.c memcpy(bio->bi_io_vec, bio_src->bi_io_vec,
bio 283 fs/bio.c bio->bi_sector = bio_src->bi_sector;
bio 284 fs/bio.c bio->bi_bdev = bio_src->bi_bdev;
bio 285 fs/bio.c bio->bi_flags |= 1 << BIO_CLONED;
bio 286 fs/bio.c bio->bi_rw = bio_src->bi_rw;
bio 287 fs/bio.c bio->bi_vcnt = bio_src->bi_vcnt;
bio 288 fs/bio.c bio->bi_size = bio_src->bi_size;
bio 289 fs/bio.c bio->bi_idx = bio_src->bi_idx;
bio 299 fs/bio.c struct bio *bio_clone(struct bio *bio, gfp_t gfp_mask)
bio 301 fs/bio.c struct bio *b = bio_alloc_bioset(gfp_mask, bio->bi_max_vecs, fs_bio_set);
bio 307 fs/bio.c __bio_clone(b, bio);
bio 309 fs/bio.c if (bio_integrity(bio)) {
bio 312 fs/bio.c ret = bio_integrity_clone(b, bio, fs_bio_set);
bio 354 fs/bio.c if (unlikely(bio_flagged(bio, BIO_CLONED)))
bio 357 fs/bio.c if (((bio->bi_size + len) >> 9) > max_sectors)
bio 365 fs/bio.c if (bio->bi_vcnt > 0) {
bio 366 fs/bio.c struct bio_vec *prev = &bio->bi_io_vec[bio->bi_vcnt - 1];
bio 374 fs/bio.c .bi_bdev = bio->bi_bdev,
bio 375 fs/bio.c .bi_sector = bio->bi_sector,
bio 376 fs/bio.c .bi_size = bio->bi_size,
bio 377 fs/bio.c .bi_rw = bio->bi_rw,
bio 390 fs/bio.c if (bio->bi_vcnt >= bio->bi_max_vecs)
bio 398 fs/bio.c while (bio->bi_phys_segments >= q->max_phys_segments
bio 399 fs/bio.c || bio->bi_phys_segments >= q->max_hw_segments) {
bio 405 fs/bio.c blk_recount_segments(q, bio);
bio 412 fs/bio.c bvec = &bio->bi_io_vec[bio->bi_vcnt];
bio 424 fs/bio.c .bi_bdev = bio->bi_bdev,
bio 425 fs/bio.c .bi_sector = bio->bi_sector,
bio 426 fs/bio.c .bi_size = bio->bi_size,
bio 427 fs/bio.c .bi_rw = bio->bi_rw,
bio 443 fs/bio.c if (bio->bi_vcnt && (BIOVEC_PHYS_MERGEABLE(bvec-1, bvec)))
bio 444 fs/bio.c bio->bi_flags &= ~(1 << BIO_SEG_VALID);
bio 446 fs/bio.c bio->bi_vcnt++;
bio 447 fs/bio.c bio->bi_phys_segments++;
bio 449 fs/bio.c bio->bi_size += len;
bio 470 fs/bio.c return __bio_add_page(q, bio, page, len, offset, q->max_hw_sectors);
bio 489 fs/bio.c struct request_queue *q = bdev_get_queue(bio->bi_bdev);
bio 490 fs/bio.c return __bio_add_page(q, bio, page, len, offset, q->max_sectors);
bio 504 fs/bio.c memcpy(bmd->iovecs, bio->bi_io_vec, sizeof(struct bio_vec) * bio->bi_vcnt);
bio 508 fs/bio.c bio->bi_private = bmd;
bio 549 fs/bio.c int read = bio_data_dir(bio) == READ;
bio 551 fs/bio.c __bio_for_each_segment(bvec, bio, i, 0) {
bio 602 fs/bio.c struct bio_map_data *bmd = bio->bi_private;
bio 605 fs/bio.c if (!bio_flagged(bio, BIO_NULL_MAPPED))
bio 606 fs/bio.c ret = __bio_copy_iov(bio, bmd->iovecs, bmd->sgvecs,
bio 609 fs/bio.c bio_put(bio);
bio 626 fs/bio.c struct bio *bio_copy_user_iov(struct request_queue *q,
bio 634 fs/bio.c struct bio *bio;
bio 657 fs/bio.c bio = bio_alloc(gfp_mask, nr_pages);
bio 658 fs/bio.c if (!bio)
bio 661 fs/bio.c bio->bi_rw |= (!write_to_vm << BIO_RW);
bio 689 fs/bio.c if (bio_add_pc_page(q, bio, page, bytes, 0) < bytes)
bio 702 fs/bio.c ret = __bio_copy_iov(bio, bio->bi_io_vec, iov, iov_count, 0, 0);
bio 707 fs/bio.c bio_set_map_data(bmd, bio, iov, iov_count, map_data ? 0 : 1);
bio 708 fs/bio.c return bio;
bio 711 fs/bio.c bio_for_each_segment(bvec, bio, i)
bio 714 fs/bio.c bio_put(bio);
bio 733 fs/bio.c struct bio *bio_copy_user(struct request_queue *q, struct rq_map_data *map_data,
bio 745 fs/bio.c static struct bio *__bio_map_user_iov(struct request_queue *q,
bio 753 fs/bio.c struct bio *bio;
bio 774 fs/bio.c bio = bio_alloc(gfp_mask, nr_pages);
bio 775 fs/bio.c if (!bio)
bio 811 fs/bio.c if (bio_add_pc_page(q, bio, pages[j], bytes, offset) <
bio 833 fs/bio.c bio->bi_rw |= (1 << BIO_RW);
bio 835 fs/bio.c bio->bi_bdev = bdev;
bio 836 fs/bio.c bio->bi_flags |= (1 << BIO_USER_MAPPED);
bio 837 fs/bio.c return bio;
bio 847 fs/bio.c bio_put(bio);
bio 863 fs/bio.c struct bio *bio_map_user(struct request_queue *q, struct block_device *bdev,
bio 887 fs/bio.c struct bio *bio_map_user_iov(struct request_queue *q, struct block_device *bdev,
bio 891 fs/bio.c struct bio *bio;
bio 893 fs/bio.c bio = __bio_map_user_iov(q, bdev, iov, iov_count, write_to_vm,
bio 895 fs/bio.c if (IS_ERR(bio))
bio 896 fs/bio.c return bio;
bio 904 fs/bio.c bio_get(bio);
bio 906 fs/bio.c return bio;
bio 917 fs/bio.c __bio_for_each_segment(bvec, bio, i, 0) {
bio 918 fs/bio.c if (bio_data_dir(bio) == READ)
bio 924 fs/bio.c bio_put(bio);
bio 938 fs/bio.c __bio_unmap_user(bio);
bio 939 fs/bio.c bio_put(bio);
bio 944 fs/bio.c bio_put(bio);
bio 948 fs/bio.c static struct bio *__bio_map_kern(struct request_queue *q, void *data,
bio 956 fs/bio.c struct bio *bio;
bio 958 fs/bio.c bio = bio_alloc(gfp_mask, nr_pages);
bio 959 fs/bio.c if (!bio)
bio 972 fs/bio.c if (bio_add_pc_page(q, bio, virt_to_page(data), bytes,
bio 981 fs/bio.c bio->bi_end_io = bio_map_kern_endio;
bio 982 fs/bio.c return bio;
bio 995 fs/bio.c struct bio *bio_map_kern(struct request_queue *q, void *data, unsigned int len,
bio 998 fs/bio.c struct bio *bio;
bio 1000 fs/bio.c bio = __bio_map_kern(q, data, len, gfp_mask);
bio 1001 fs/bio.c if (IS_ERR(bio))
bio 1002 fs/bio.c return bio;
bio 1004 fs/bio.c if (bio->bi_size == len)
bio 1005 fs/bio.c return bio;
bio 1010 fs/bio.c bio_put(bio);
bio 1017 fs/bio.c const int read = bio_data_dir(bio) == READ;
bio 1018 fs/bio.c struct bio_map_data *bmd = bio->bi_private;
bio 1022 fs/bio.c __bio_for_each_segment(bvec, bio, i, 0) {
bio 1034 fs/bio.c bio_put(bio);
bio 1048 fs/bio.c struct bio *bio_copy_kern(struct request_queue *q, void *data, unsigned int len,
bio 1051 fs/bio.c struct bio *bio;
bio 1055 fs/bio.c bio = bio_copy_user(q, NULL, (unsigned long)data, len, 1, gfp_mask);
bio 1056 fs/bio.c if (IS_ERR(bio))
bio 1057 fs/bio.c return bio;
bio 1062 fs/bio.c bio_for_each_segment(bvec, bio, i) {
bio 1070 fs/bio.c bio->bi_end_io = bio_copy_kern_endio;
bio 1072 fs/bio.c return bio;
bio 1106 fs/bio.c struct bio_vec *bvec = bio->bi_io_vec;
bio 1109 fs/bio.c for (i = 0; i < bio->bi_vcnt; i++) {
bio 1119 fs/bio.c struct bio_vec *bvec = bio->bi_io_vec;
bio 1122 fs/bio.c for (i = 0; i < bio->bi_vcnt; i++) {
bio 1145 fs/bio.c static struct bio *bio_dirty_list;
bio 1153 fs/bio.c struct bio *bio;
bio 1156 fs/bio.c bio = bio_dirty_list;
bio 1160 fs/bio.c while (bio) {
bio 1161 fs/bio.c struct bio *next = bio->bi_private;
bio 1163 fs/bio.c bio_set_pages_dirty(bio);
bio 1164 fs/bio.c bio_release_pages(bio);
bio 1165 fs/bio.c bio_put(bio);
bio 1166 fs/bio.c bio = next;
bio 1172 fs/bio.c struct bio_vec *bvec = bio->bi_io_vec;
bio 1176 fs/bio.c for (i = 0; i < bio->bi_vcnt; i++) {
bio 1191 fs/bio.c bio->bi_private = bio_dirty_list;
bio 1192 fs/bio.c bio_dirty_list = bio;
bio 1196 fs/bio.c bio_put(bio);
bio 1217 fs/bio.c clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio 1218 fs/bio.c else if (!test_bit(BIO_UPTODATE, &bio->bi_flags))
bio 1221 fs/bio.c if (bio->bi_end_io)
bio 1222 fs/bio.c bio->bi_end_io(bio, error);
bio 1228 fs/bio.c struct bio *master = bp->bio1.bi_private;
bio 1316 fs/bio.c unsigned int sector_sz = queue_hardsect_size(bio->bi_bdev->bd_disk->queue);
bio 1323 fs/bio.c if (index >= bio->bi_idx)
bio 1324 fs/bio.c index = bio->bi_vcnt - 1;
bio 1326 fs/bio.c __bio_for_each_segment(bv, bio, i, 0) {
bio 1421 fs/bio.c bio_slab = KMEM_CACHE(bio, SLAB_HWCACHE_ALIGN|SLAB_PANIC);
bio 2909 fs/buffer.c struct buffer_head *bh = bio->bi_private;
bio 2912 fs/buffer.c set_bit(BIO_EOPNOTSUPP, &bio->bi_flags);
bio 2916 fs/buffer.c bh->b_end_io(bh, test_bit(BIO_UPTODATE, &bio->bi_flags));
bio 2917 fs/buffer.c bio_put(bio);
bio 2922 fs/buffer.c struct bio *bio;
bio 2946 fs/buffer.c bio = bio_alloc(GFP_NOIO, 1);
bio 2948 fs/buffer.c bio->bi_sector = bh->b_blocknr * (bh->b_size >> 9);
bio 2949 fs/buffer.c bio->bi_bdev = bh->b_bdev;
bio 2950 fs/buffer.c bio->bi_io_vec[0].bv_page = bh->b_page;
bio 2951 fs/buffer.c bio->bi_io_vec[0].bv_len = bh->b_size;
bio 2952 fs/buffer.c bio->bi_io_vec[0].bv_offset = bh_offset(bh);
bio 2954 fs/buffer.c bio->bi_vcnt = 1;
bio 2955 fs/buffer.c bio->bi_idx = 0;
bio 2956 fs/buffer.c bio->bi_size = bh->b_size;
bio 2958 fs/buffer.c bio->bi_end_io = end_bio_bh_io_sync;
bio 2959 fs/buffer.c bio->bi_private = bh;
bio 2961 fs/buffer.c bio_get(bio);
bio 2962 fs/buffer.c submit_bio(rw, bio);
bio 2964 fs/buffer.c if (bio_flagged(bio, BIO_EOPNOTSUPP))
bio 2967 fs/buffer.c bio_put(bio);
bio 67 fs/direct-io.c struct bio *bio; /* bio under assembly */
bio 126 fs/direct-io.c struct bio *bio_list; /* singly linked via bi_private */
bio 263 fs/direct-io.c struct dio *dio = bio->bi_private;
bio 268 fs/direct-io.c dio_bio_complete(dio, bio);
bio 292 fs/direct-io.c struct dio *dio = bio->bi_private;
bio 296 fs/direct-io.c bio->bi_private = dio->bio_list;
bio 297 fs/direct-io.c dio->bio_list = bio;
bio 307 fs/direct-io.c struct bio *bio;
bio 309 fs/direct-io.c bio = bio_alloc(GFP_KERNEL, nr_vecs);
bio 310 fs/direct-io.c if (bio == NULL)
bio 313 fs/direct-io.c bio->bi_bdev = bdev;
bio 314 fs/direct-io.c bio->bi_sector = first_sector;
bio 316 fs/direct-io.c bio->bi_end_io = dio_bio_end_aio;
bio 318 fs/direct-io.c bio->bi_end_io = dio_bio_end_io;
bio 320 fs/direct-io.c dio->bio = bio;
bio 333 fs/direct-io.c struct bio *bio = dio->bio;
bio 336 fs/direct-io.c bio->bi_private = dio;
bio 343 fs/direct-io.c bio_set_pages_dirty(bio);
bio 345 fs/direct-io.c submit_bio(dio->rw, bio);
bio 347 fs/direct-io.c dio->bio = NULL;
bio 366 fs/direct-io.c static struct bio *dio_await_one(struct dio *dio)
bio 369 fs/direct-io.c struct bio *bio = NULL;
bio 389 fs/direct-io.c bio = dio->bio_list;
bio 390 fs/direct-io.c dio->bio_list = bio->bi_private;
bio 393 fs/direct-io.c return bio;
bio 401 fs/direct-io.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio 402 fs/direct-io.c struct bio_vec *bvec = bio->bi_io_vec;
bio 409 fs/direct-io.c bio_check_pages_dirty(bio); /* transfers ownership */
bio 411 fs/direct-io.c for (page_no = 0; page_no < bio->bi_vcnt; page_no++) {
bio 418 fs/direct-io.c bio_put(bio);
bio 432 fs/direct-io.c struct bio *bio;
bio 434 fs/direct-io.c bio = dio_await_one(dio);
bio 435 fs/direct-io.c if (bio)
bio 436 fs/direct-io.c dio_bio_complete(dio, bio);
bio 437 fs/direct-io.c } while (bio);
bio 454 fs/direct-io.c struct bio *bio;
bio 458 fs/direct-io.c bio = dio->bio_list;
bio 459 fs/direct-io.c dio->bio_list = bio->bi_private;
bio 461 fs/direct-io.c ret2 = dio_bio_complete(dio, bio);
bio 572 fs/direct-io.c ret = bio_add_page(dio->bio, dio->cur_page,
bio 604 fs/direct-io.c if (dio->bio) {
bio 618 fs/direct-io.c if (dio->bio == NULL) {
bio 1030 fs/direct-io.c if (dio->bio)
bio 2295 fs/ext4/extents.c complete((struct completion *)bio->bi_private);
bio 2302 fs/ext4/extents.c struct bio *bio;
bio 2324 fs/ext4/extents.c bio = bio_alloc(GFP_NOIO, len);
bio 2325 fs/ext4/extents.c if (!bio)
bio 2327 fs/ext4/extents.c bio->bi_sector = ee_pblock;
bio 2328 fs/ext4/extents.c bio->bi_bdev = inode->i_sb->s_bdev;
bio 2333 fs/ext4/extents.c ret = bio_add_page(bio, ZERO_PAGE(0),
bio 2349 fs/ext4/extents.c bio->bi_private = &event;
bio 2350 fs/ext4/extents.c bio->bi_end_io = bi_complete;
bio 2351 fs/ext4/extents.c submit_bio(WRITE, bio);
bio 2354 fs/ext4/extents.c if (test_bit(BIO_UPTODATE, &bio->bi_flags))
bio 2360 fs/ext4/extents.c bio_put(bio);
bio 219 fs/gfs2/ops_fstype.c struct page *page = bio->bi_private;
bio 272 fs/gfs2/ops_fstype.c struct bio *bio;
bio 282 fs/gfs2/ops_fstype.c bio = bio_alloc(GFP_NOFS, 1);
bio 283 fs/gfs2/ops_fstype.c if (unlikely(!bio)) {
bio 288 fs/gfs2/ops_fstype.c bio->bi_sector = sector * (sb->s_blocksize >> 9);
bio 289 fs/gfs2/ops_fstype.c bio->bi_bdev = sb->s_bdev;
bio 290 fs/gfs2/ops_fstype.c bio_add_page(bio, page, PAGE_SIZE, 0);
bio 292 fs/gfs2/ops_fstype.c bio->bi_end_io = end_bio_io_page;
bio 293 fs/gfs2/ops_fstype.c bio->bi_private = page;
bio 294 fs/gfs2/ops_fstype.c submit_bio(READ_SYNC | (1 << BIO_RW_META), bio);
bio 296 fs/gfs2/ops_fstype.c bio_put(bio);
bio 1994 fs/jfs/jfs_logmgr.c struct bio *bio;
bio 2005 fs/jfs/jfs_logmgr.c bio = bio_alloc(GFP_NOFS, 1);
bio 2007 fs/jfs/jfs_logmgr.c bio->bi_sector = bp->l_blkno << (log->l2bsize - 9);
bio 2008 fs/jfs/jfs_logmgr.c bio->bi_bdev = log->bdev;
bio 2009 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_page = bp->l_page;
bio 2010 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_len = LOGPSIZE;
bio 2011 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_offset = bp->l_offset;
bio 2013 fs/jfs/jfs_logmgr.c bio->bi_vcnt = 1;
bio 2014 fs/jfs/jfs_logmgr.c bio->bi_idx = 0;
bio 2015 fs/jfs/jfs_logmgr.c bio->bi_size = LOGPSIZE;
bio 2017 fs/jfs/jfs_logmgr.c bio->bi_end_io = lbmIODone;
bio 2018 fs/jfs/jfs_logmgr.c bio->bi_private = bp;
bio 2019 fs/jfs/jfs_logmgr.c submit_bio(READ_SYNC, bio);
bio 2142 fs/jfs/jfs_logmgr.c struct bio *bio;
bio 2147 fs/jfs/jfs_logmgr.c bio = bio_alloc(GFP_NOFS, 1);
bio 2148 fs/jfs/jfs_logmgr.c bio->bi_sector = bp->l_blkno << (log->l2bsize - 9);
bio 2149 fs/jfs/jfs_logmgr.c bio->bi_bdev = log->bdev;
bio 2150 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_page = bp->l_page;
bio 2151 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_len = LOGPSIZE;
bio 2152 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_offset = bp->l_offset;
bio 2154 fs/jfs/jfs_logmgr.c bio->bi_vcnt = 1;
bio 2155 fs/jfs/jfs_logmgr.c bio->bi_idx = 0;
bio 2156 fs/jfs/jfs_logmgr.c bio->bi_size = LOGPSIZE;
bio 2158 fs/jfs/jfs_logmgr.c bio->bi_end_io = lbmIODone;
bio 2159 fs/jfs/jfs_logmgr.c bio->bi_private = bp;
bio 2163 fs/jfs/jfs_logmgr.c bio->bi_size = 0;
bio 2164 fs/jfs/jfs_logmgr.c lbmIODone(bio, 0);
bio 2166 fs/jfs/jfs_logmgr.c submit_bio(WRITE_SYNC, bio);
bio 2204 fs/jfs/jfs_logmgr.c struct lbuf *bp = bio->bi_private;
bio 2218 fs/jfs/jfs_logmgr.c if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) {
bio 2224 fs/jfs/jfs_logmgr.c bio_put(bio);
bio 287 fs/jfs/jfs_metapage.c struct page *page = bio->bi_private;
bio 289 fs/jfs/jfs_metapage.c if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) {
bio 295 fs/jfs/jfs_metapage.c bio_put(bio);
bio 342 fs/jfs/jfs_metapage.c struct page *page = bio->bi_private;
bio 346 fs/jfs/jfs_metapage.c if (! test_bit(BIO_UPTODATE, &bio->bi_flags)) {
bio 351 fs/jfs/jfs_metapage.c bio_put(bio);
bio 356 fs/jfs/jfs_metapage.c struct bio *bio = NULL;
bio 399 fs/jfs/jfs_metapage.c if (bio) {
bio 409 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, bio_bytes, bio_offset) <
bio 417 fs/jfs/jfs_metapage.c if (!bio->bi_size)
bio 419 fs/jfs/jfs_metapage.c submit_bio(WRITE, bio);
bio 421 fs/jfs/jfs_metapage.c bio = NULL;
bio 435 fs/jfs/jfs_metapage.c bio = bio_alloc(GFP_NOFS, 1);
bio 436 fs/jfs/jfs_metapage.c bio->bi_bdev = inode->i_sb->s_bdev;
bio 437 fs/jfs/jfs_metapage.c bio->bi_sector = pblock << (inode->i_blkbits - 9);
bio 438 fs/jfs/jfs_metapage.c bio->bi_end_io = metapage_write_end_io;
bio 439 fs/jfs/jfs_metapage.c bio->bi_private = page;
bio 448 fs/jfs/jfs_metapage.c if (bio) {
bio 449 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, bio_bytes, bio_offset) < bio_bytes)
bio 451 fs/jfs/jfs_metapage.c if (!bio->bi_size)
bio 454 fs/jfs/jfs_metapage.c submit_bio(WRITE, bio);
bio 472 fs/jfs/jfs_metapage.c 4, bio, sizeof(*bio), 0);
bio 474 fs/jfs/jfs_metapage.c bio_put(bio);
bio 484 fs/jfs/jfs_metapage.c struct bio *bio = NULL;
bio 506 fs/jfs/jfs_metapage.c if (bio)
bio 507 fs/jfs/jfs_metapage.c submit_bio(READ, bio);
bio 509 fs/jfs/jfs_metapage.c bio = bio_alloc(GFP_NOFS, 1);
bio 510 fs/jfs/jfs_metapage.c bio->bi_bdev = inode->i_sb->s_bdev;
bio 511 fs/jfs/jfs_metapage.c bio->bi_sector = pblock << (inode->i_blkbits - 9);
bio 512 fs/jfs/jfs_metapage.c bio->bi_end_io = metapage_read_end_io;
bio 513 fs/jfs/jfs_metapage.c bio->bi_private = page;
bio 516 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, len, offset) < len)
bio 522 fs/jfs/jfs_metapage.c if (bio)
bio 523 fs/jfs/jfs_metapage.c submit_bio(READ, bio);
bio 531 fs/jfs/jfs_metapage.c bio_put(bio);
bio 44 fs/mpage.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio 45 fs/mpage.c struct bio_vec *bvec = bio->bi_io_vec + bio->bi_vcnt - 1;
bio 50 fs/mpage.c if (--bvec >= bio->bi_io_vec)
bio 60 fs/mpage.c } while (bvec >= bio->bi_io_vec);
bio 61 fs/mpage.c bio_put(bio);
bio 66 fs/mpage.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio 67 fs/mpage.c struct bio_vec *bvec = bio->bi_io_vec + bio->bi_vcnt - 1;
bio 72 fs/mpage.c if (--bvec >= bio->bi_io_vec)
bio 81 fs/mpage.c } while (bvec >= bio->bi_io_vec);
bio 82 fs/mpage.c bio_put(bio);
bio 85 fs/mpage.c struct bio *mpage_bio_submit(int rw, struct bio *bio)
bio 87 fs/mpage.c bio->bi_end_io = mpage_end_io_read;
bio 89 fs/mpage.c bio->bi_end_io = mpage_end_io_write;
bio 90 fs/mpage.c submit_bio(rw, bio);
bio 95 fs/mpage.c static struct bio *
bio 100 fs/mpage.c struct bio *bio;
bio 102 fs/mpage.c bio = bio_alloc(gfp_flags, nr_vecs);
bio 104 fs/mpage.c if (bio == NULL && (current->flags & PF_MEMALLOC)) {
bio 105 fs/mpage.c while (!bio && (nr_vecs /= 2))
bio 106 fs/mpage.c bio = bio_alloc(gfp_flags, nr_vecs);
bio 109 fs/mpage.c if (bio) {
bio 110 fs/mpage.c bio->bi_bdev = bdev;
bio 111 fs/mpage.c bio->bi_sector = first_sector;
bio 113 fs/mpage.c return bio;
bio 168 fs/mpage.c static struct bio *
bio 293 fs/mpage.c if (bio && (*last_block_in_bio != blocks[0] - 1))
bio 294 fs/mpage.c bio = mpage_bio_submit(READ, bio);
bio 297 fs/mpage.c if (bio == NULL) {
bio 298 fs/mpage.c bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9),
bio 301 fs/mpage.c if (bio == NULL)
bio 306 fs/mpage.c if (bio_add_page(bio, page, length, 0) < length) {
bio 307 fs/mpage.c bio = mpage_bio_submit(READ, bio);
bio 312 fs/mpage.c bio = mpage_bio_submit(READ, bio);
bio 316 fs/mpage.c return bio;
bio 319 fs/mpage.c if (bio)
bio 320 fs/mpage.c bio = mpage_bio_submit(READ, bio);
bio 375 fs/mpage.c struct bio *bio = NULL;
bio 389 fs/mpage.c bio = do_mpage_readpage(bio, page,
bio 398 fs/mpage.c if (bio)
bio 399 fs/mpage.c mpage_bio_submit(READ, bio);
bio 409 fs/mpage.c struct bio *bio = NULL;
bio 415 fs/mpage.c bio = do_mpage_readpage(bio, page, 1, &last_block_in_bio,
bio 417 fs/mpage.c if (bio)
bio 418 fs/mpage.c mpage_bio_submit(READ, bio);
bio 444 fs/mpage.c struct bio *bio = mpd->bio;
bio 570 fs/mpage.c if (bio && mpd->last_block_in_bio != blocks[0] - 1)
bio 571 fs/mpage.c bio = mpage_bio_submit(WRITE, bio);
bio 574 fs/mpage.c if (bio == NULL) {
bio 575 fs/mpage.c bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9),
bio 577 fs/mpage.c if (bio == NULL)
bio 587 fs/mpage.c if (bio_add_page(bio, page, length, 0) < length) {
bio 588 fs/mpage.c bio = mpage_bio_submit(WRITE, bio);
bio 621 fs/mpage.c bio = mpage_bio_submit(WRITE, bio);
bio 632 fs/mpage.c if (bio)
bio 633 fs/mpage.c bio = mpage_bio_submit(WRITE, bio);
bio 646 fs/mpage.c mpd->bio = bio;
bio 680 fs/mpage.c .bio = NULL,
bio 687 fs/mpage.c if (mpd.bio)
bio 688 fs/mpage.c mpage_bio_submit(WRITE, mpd.bio);
bio 698 fs/mpage.c .bio = NULL,
bio 704 fs/mpage.c if (mpd.bio)
bio 705 fs/mpage.c mpage_bio_submit(WRITE, mpd.bio);
bio 222 fs/ocfs2/cluster/heartbeat.c struct o2hb_bio_wait_ctxt *wc = bio->bi_private;
bio 230 fs/ocfs2/cluster/heartbeat.c bio_put(bio);
bio 235 fs/ocfs2/cluster/heartbeat.c static struct bio *o2hb_setup_one_bio(struct o2hb_region *reg,
bio 245 fs/ocfs2/cluster/heartbeat.c struct bio *bio;
bio 252 fs/ocfs2/cluster/heartbeat.c bio = bio_alloc(GFP_ATOMIC, 16);
bio 253 fs/ocfs2/cluster/heartbeat.c if (!bio) {
bio 255 fs/ocfs2/cluster/heartbeat.c bio = ERR_PTR(-ENOMEM);
bio 260 fs/ocfs2/cluster/heartbeat.c bio->bi_sector = (reg->hr_start_block + cs) << (bits - 9);
bio 261 fs/ocfs2/cluster/heartbeat.c bio->bi_bdev = reg->hr_bdev;
bio 262 fs/ocfs2/cluster/heartbeat.c bio->bi_private = wc;
bio 263 fs/ocfs2/cluster/heartbeat.c bio->bi_end_io = o2hb_bio_end_io;
bio 276 fs/ocfs2/cluster/heartbeat.c len = bio_add_page(bio, page, vec_len, vec_start);
bio 285 fs/ocfs2/cluster/heartbeat.c return bio;
bio 294 fs/ocfs2/cluster/heartbeat.c struct bio *bio;
bio 299 fs/ocfs2/cluster/heartbeat.c bio = o2hb_setup_one_bio(reg, &wc, ¤t_slot, max_slots);
bio 300 fs/ocfs2/cluster/heartbeat.c if (IS_ERR(bio)) {
bio 301 fs/ocfs2/cluster/heartbeat.c status = PTR_ERR(bio);
bio 307 fs/ocfs2/cluster/heartbeat.c submit_bio(READ, bio);
bio 325 fs/ocfs2/cluster/heartbeat.c struct bio *bio;
bio 331 fs/ocfs2/cluster/heartbeat.c bio = o2hb_setup_one_bio(reg, write_wc, &slot, slot+1);
bio 332 fs/ocfs2/cluster/heartbeat.c if (IS_ERR(bio)) {
bio 333 fs/ocfs2/cluster/heartbeat.c status = PTR_ERR(bio);
bio 339 fs/ocfs2/cluster/heartbeat.c submit_bio(WRITE, bio);
bio 347 fs/xfs/linux-2.6/xfs_aops.c xfs_ioend_t *ioend = bio->bi_private;
bio 349 fs/xfs/linux-2.6/xfs_aops.c ASSERT(atomic_read(&bio->bi_cnt) >= 1);
bio 350 fs/xfs/linux-2.6/xfs_aops.c ioend->io_error = test_bit(BIO_UPTODATE, &bio->bi_flags) ? 0 : error;
bio 353 fs/xfs/linux-2.6/xfs_aops.c bio->bi_private = NULL;
bio 354 fs/xfs/linux-2.6/xfs_aops.c bio->bi_end_io = NULL;
bio 355 fs/xfs/linux-2.6/xfs_aops.c bio_put(bio);
bio 367 fs/xfs/linux-2.6/xfs_aops.c bio->bi_private = ioend;
bio 368 fs/xfs/linux-2.6/xfs_aops.c bio->bi_end_io = xfs_end_bio;
bio 370 fs/xfs/linux-2.6/xfs_aops.c submit_bio(WRITE, bio);
bio 371 fs/xfs/linux-2.6/xfs_aops.c ASSERT(!bio_flagged(bio, BIO_EOPNOTSUPP));
bio 372 fs/xfs/linux-2.6/xfs_aops.c bio_put(bio);
bio 375 fs/xfs/linux-2.6/xfs_aops.c STATIC struct bio *
bio 379 fs/xfs/linux-2.6/xfs_aops.c struct bio *bio;
bio 383 fs/xfs/linux-2.6/xfs_aops.c bio = bio_alloc(GFP_NOIO, nvecs);
bio 385 fs/xfs/linux-2.6/xfs_aops.c } while (!bio);
bio 387 fs/xfs/linux-2.6/xfs_aops.c ASSERT(bio->bi_private == NULL);
bio 388 fs/xfs/linux-2.6/xfs_aops.c bio->bi_sector = bh->b_blocknr * (bh->b_size >> 9);
bio 389 fs/xfs/linux-2.6/xfs_aops.c bio->bi_bdev = bh->b_bdev;
bio 390 fs/xfs/linux-2.6/xfs_aops.c bio_get(bio);
bio 391 fs/xfs/linux-2.6/xfs_aops.c return bio;
bio 427 fs/xfs/linux-2.6/xfs_aops.c return bio_add_page(bio, bh->b_page, bh->b_size, bh_offset(bh));
bio 454 fs/xfs/linux-2.6/xfs_aops.c struct bio *bio;
bio 469 fs/xfs/linux-2.6/xfs_aops.c bio = NULL;
bio 473 fs/xfs/linux-2.6/xfs_aops.c if (!bio) {
bio 475 fs/xfs/linux-2.6/xfs_aops.c bio = xfs_alloc_ioend_bio(bh);
bio 477 fs/xfs/linux-2.6/xfs_aops.c xfs_submit_ioend_bio(ioend, bio);
bio 481 fs/xfs/linux-2.6/xfs_aops.c if (bio_add_buffer(bio, bh) != bh->b_size) {
bio 482 fs/xfs/linux-2.6/xfs_aops.c xfs_submit_ioend_bio(ioend, bio);
bio 488 fs/xfs/linux-2.6/xfs_aops.c if (bio)
bio 489 fs/xfs/linux-2.6/xfs_aops.c xfs_submit_ioend_bio(ioend, bio);
bio 1113 fs/xfs/linux-2.6/xfs_buf.c xfs_buf_t *bp = (xfs_buf_t *)bio->bi_private;
bio 1115 fs/xfs/linux-2.6/xfs_buf.c struct bio_vec *bvec = bio->bi_io_vec + bio->bi_vcnt - 1;
bio 1117 fs/xfs/linux-2.6/xfs_buf.c if (!test_bit(BIO_UPTODATE, &bio->bi_flags))
bio 1134 fs/xfs/linux-2.6/xfs_buf.c if (--bvec >= bio->bi_io_vec)
bio 1139 fs/xfs/linux-2.6/xfs_buf.c } while (bvec >= bio->bi_io_vec);
bio 1142 fs/xfs/linux-2.6/xfs_buf.c bio_put(bio);
bio 1150 fs/xfs/linux-2.6/xfs_buf.c struct bio *bio;
bio 1180 fs/xfs/linux-2.6/xfs_buf.c bio = bio_alloc(GFP_NOIO, 1);
bio 1182 fs/xfs/linux-2.6/xfs_buf.c bio->bi_bdev = bp->b_target->bt_bdev;
bio 1183 fs/xfs/linux-2.6/xfs_buf.c bio->bi_sector = sector - (offset >> BBSHIFT);
bio 1184 fs/xfs/linux-2.6/xfs_buf.c bio->bi_end_io = xfs_buf_bio_end_io;
bio 1185 fs/xfs/linux-2.6/xfs_buf.c bio->bi_private = bp;
bio 1187 fs/xfs/linux-2.6/xfs_buf.c bio_add_page(bio, bp->b_pages[0], PAGE_CACHE_SIZE, 0);
bio 1201 fs/xfs/linux-2.6/xfs_buf.c bio = bio_alloc(GFP_NOIO, nr_pages);
bio 1202 fs/xfs/linux-2.6/xfs_buf.c bio->bi_bdev = bp->b_target->bt_bdev;
bio 1203 fs/xfs/linux-2.6/xfs_buf.c bio->bi_sector = sector;
bio 1204 fs/xfs/linux-2.6/xfs_buf.c bio->bi_end_io = xfs_buf_bio_end_io;
bio 1205 fs/xfs/linux-2.6/xfs_buf.c bio->bi_private = bp;
bio 1213 fs/xfs/linux-2.6/xfs_buf.c rbytes = bio_add_page(bio, bp->b_pages[map_i], nbytes, offset);
bio 1224 fs/xfs/linux-2.6/xfs_buf.c if (likely(bio->bi_size)) {
bio 1225 fs/xfs/linux-2.6/xfs_buf.c submit_bio(rw, bio);
bio 1229 fs/xfs/linux-2.6/xfs_buf.c bio_put(bio);
bio 53 include/linux/bio.h struct bio;
bio 65 include/linux/bio.h struct bio *bi_next; /* request queue link */
bio 113 include/linux/bio.h #define bio_flagged(bio, flag) ((bio)->bi_flags & (1 << (flag)))
bio 121 include/linux/bio.h #define BIO_POOL_IDX(bio) ((bio)->bi_flags >> BIO_POOL_OFFSET)
bio 156 include/linux/bio.h #define bio_prio(bio) ((bio)->bi_rw >> BIO_PRIO_SHIFT)
bio 157 include/linux/bio.h #define bio_prio_valid(bio) ioprio_valid(bio_prio(bio))
bio 161 include/linux/bio.h (bio)->bi_rw &= ((1UL << BIO_PRIO_SHIFT) - 1); \
bio 162 include/linux/bio.h (bio)->bi_rw |= ((unsigned long) (prio) << BIO_PRIO_SHIFT); \
bio 169 include/linux/bio.h #define bio_iovec_idx(bio, idx) (&((bio)->bi_io_vec[(idx)]))
bio 170 include/linux/bio.h #define bio_iovec(bio) bio_iovec_idx((bio), (bio)->bi_idx)
bio 171 include/linux/bio.h #define bio_page(bio) bio_iovec((bio))->bv_page
bio 172 include/linux/bio.h #define bio_offset(bio) bio_iovec((bio))->bv_offset
bio 173 include/linux/bio.h #define bio_segments(bio) ((bio)->bi_vcnt - (bio)->bi_idx)
bio 174 include/linux/bio.h #define bio_sectors(bio) ((bio)->bi_size >> 9)
bio 175 include/linux/bio.h #define bio_barrier(bio) ((bio)->bi_rw & (1 << BIO_RW_BARRIER))
bio 176 include/linux/bio.h #define bio_sync(bio) ((bio)->bi_rw & (1 << BIO_RW_SYNC))
bio 177 include/linux/bio.h #define bio_failfast(bio) ((bio)->bi_rw & (1 << BIO_RW_FAILFAST))
bio 178 include/linux/bio.h #define bio_rw_ahead(bio) ((bio)->bi_rw & (1 << BIO_RW_AHEAD))
bio 179 include/linux/bio.h #define bio_rw_meta(bio) ((bio)->bi_rw & (1 << BIO_RW_META))
bio 180 include/linux/bio.h #define bio_discard(bio) ((bio)->bi_rw & (1 << BIO_RW_DISCARD))
bio 181 include/linux/bio.h #define bio_empty_barrier(bio) (bio_barrier(bio) && !bio_has_data(bio) && !bio_discard(bio))
bio 185 include/linux/bio.h if (bio->bi_vcnt)
bio 186 include/linux/bio.h return bio_iovec(bio)->bv_len >> 9;
bio 188 include/linux/bio.h return bio->bi_size >> 9;
bio 193 include/linux/bio.h if (bio->bi_vcnt)
bio 194 include/linux/bio.h return page_address(bio_page(bio)) + bio_offset(bio);
bio 202 include/linux/bio.h #define bio_to_phys(bio) (page_to_phys(bio_page((bio))) + (unsigned long) bio_offset((bio)))
bio 212 include/linux/bio.h (kmap_atomic(bio_iovec_idx((bio), (idx))->bv_page, kmtype) + \
bio 213 include/linux/bio.h bio_iovec_idx((bio), (idx))->bv_offset)
bio 221 include/linux/bio.h #define __BVEC_END(bio) bio_iovec_idx((bio), (bio)->bi_vcnt - 1)
bio 222 include/linux/bio.h #define __BVEC_START(bio) bio_iovec_idx((bio), (bio)->bi_idx)
bio 239 include/linux/bio.h #define bio_io_error(bio) bio_endio((bio), -EIO)
bio 246 include/linux/bio.h for (bvl = bio_iovec_idx((bio), (start_idx)), i = (start_idx); \
bio 247 include/linux/bio.h i < (bio)->bi_vcnt; \
bio 251 include/linux/bio.h __bio_for_each_segment(bvl, bio, i, (bio)->bi_idx)
bio 267 include/linux/bio.h #define bio_get(bio) atomic_inc(&(bio)->bi_cnt)
bio 274 include/linux/bio.h struct bio *bip_bio; /* parent bio */
bio 304 include/linux/bio.h struct bio bio1, bio2;
bio 319 include/linux/bio.h extern struct bio *bio_alloc(gfp_t, int);
bio 320 include/linux/bio.h extern struct bio *bio_kmalloc(gfp_t, int);
bio 321 include/linux/bio.h extern struct bio *bio_alloc_bioset(gfp_t, int, struct bio_set *);
bio 330 include/linux/bio.h extern struct bio *bio_clone(struct bio *, gfp_t);
bio 339 include/linux/bio.h extern struct bio *bio_map_user(struct request_queue *, struct block_device *,
bio 343 include/linux/bio.h extern struct bio *bio_map_user_iov(struct request_queue *,
bio 347 include/linux/bio.h extern struct bio *bio_map_kern(struct request_queue *, void *, unsigned int,
bio 349 include/linux/bio.h extern struct bio *bio_copy_kern(struct request_queue *, void *, unsigned int,
bio 353 include/linux/bio.h extern struct bio *bio_copy_user(struct request_queue *, struct rq_map_data *,
bio 355 include/linux/bio.h extern struct bio *bio_copy_user_iov(struct request_queue *,
bio 368 include/linux/bio.h bio->bi_comp_cpu = cpu;
bio 441 include/linux/bio.h return bvec_kmap_irq(bio_iovec_idx(bio, idx), flags);
bio 446 include/linux/bio.h __bio_kmap_irq((bio), (bio)->bi_idx, (flags))
bio 454 include/linux/bio.h return bio && bio->bi_io_vec != NULL;
bio 470 include/linux/bio.h #define bio_integrity(bio) (bio->bi_integrity != NULL)
bio 167 include/linux/blkdev.h struct bio *bio;
bio 168 include/linux/blkdev.h struct bio *biotail;
bio 678 include/linux/blkdev.h struct bio *bio;
bio 683 include/linux/blkdev.h if ((rq->bio)) \
bio 684 include/linux/blkdev.h for (_bio = (rq)->bio; _bio; _bio = _bio->bi_next)
bio 687 include/linux/blkdev.h __rq_for_each_bio(_iter.bio, _rq) \
bio 688 include/linux/blkdev.h bio_for_each_segment(bvl, _iter.bio, _iter.i)
bio 691 include/linux/blkdev.h (_iter.bio->bi_next == NULL && _iter.i == _iter.bio->bi_vcnt-1)
bio 1032 include/linux/blkdev.h if (rq->bio == NULL)
bio 1035 include/linux/blkdev.h return bio_integrity(rq->bio);
bio 235 include/linux/blktrace_api.h __blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw, what, !bio_flagged(bio, BIO_UPTODATE), 0, NULL);
bio 257 include/linux/blktrace_api.h if (bio)
bio 258 include/linux/blktrace_api.h blk_add_trace_bio(q, bio, what);
bio 285 include/linux/blktrace_api.h if (bio)
bio 286 include/linux/blktrace_api.h __blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw, what, !bio_flagged(bio, BIO_UPTODATE), sizeof(rpdu), &rpdu);
bio 314 include/linux/blktrace_api.h r.device_from = cpu_to_be32(bio->bi_bdev->bd_dev);
bio 317 include/linux/blktrace_api.h __blk_add_trace(bt, from, bio->bi_size, bio->bi_rw, BLK_TA_REMAP, !bio_flagged(bio, BIO_UPTODATE), sizeof(r), &r);
bio 1787 include/linux/fs.h #define bio_rw(bio) ((bio)->bi_rw & (RW_MASK | RWA_MASK))
bio 1792 include/linux/fs.h #define bio_data_dir(bio) ((bio)->bi_rw & 1)
bio 1927 include/linux/fs.h struct bio;
bio 59 include/linux/loop.h struct bio *lo_bio;
bio 60 include/linux/loop.h struct bio *lo_biotail;
bio 15 include/linux/mpage.h struct bio *bio;
bio 23 include/linux/mpage.h struct bio *mpage_bio_submit(int rw, struct bio *bio);
bio 165 include/linux/pktcdvd.h struct bio *read_queue;
bio 166 include/linux/pktcdvd.h struct bio *read_queue_tail;
bio 167 include/linux/pktcdvd.h struct bio *write_queue;
bio 168 include/linux/pktcdvd.h struct bio *write_queue_tail;
bio 208 include/linux/pktcdvd.h struct bio *orig_bios; /* Original bios passed to pkt_make_request */
bio 209 include/linux/pktcdvd.h struct bio *orig_bios_tail;/* that will be handled by this packet */
bio 213 include/linux/pktcdvd.h struct bio *w_bio; /* The bio we will send to the real CD */
bio 228 include/linux/pktcdvd.h struct bio *r_bios[PACKET_MAX_SIZE]; /* bios to use during data gathering */
bio 241 include/linux/pktcdvd.h struct bio *bio;
bio 246 include/linux/pktcdvd.h struct bio *bio; /* Original read request bio */
bio 230 include/linux/raid/md_k.h struct bio *biolist; /* bios that need to be retried
bio 290 include/linux/raid/md_k.h int (*make_request)(struct request_queue *q, struct bio *bio);
bio 37 include/linux/raid/multipath.h struct bio *master_bio;
bio 38 include/linux/raid/multipath.h struct bio bio;
bio 95 include/linux/raid/raid1.h struct bio *master_bio;
bio 107 include/linux/raid/raid1.h struct bio *bios[0];
bio 116 include/linux/raid/raid1.h #define IO_BLOCKED ((struct bio*)1)
bio 90 include/linux/raid/raid10.h struct bio *master_bio;
bio 105 include/linux/raid/raid10.h struct bio *bio;
bio 116 include/linux/raid/raid10.h #define IO_BLOCKED ((struct bio*)1)
bio 219 include/linux/raid/raid5.h struct bio req;
bio 222 include/linux/raid/raid5.h struct bio *toread, *read, *towrite, *written;
bio 348 include/linux/raid/raid5.h struct bio *retry_read_aligned; /* currently retrying aligned bios */
bio 349 include/linux/raid/raid5.h struct bio *retry_read_aligned_list; /* aligned bios retry list */
bio 98 include/linux/sched.h struct bio;
bio 1249 include/linux/sched.h struct bio *bio_list, **bio_tail;
bio 16 include/linux/swap.h struct bio;
bio 63 kernel/power/swap.c struct bio *bio;
bio 65 kernel/power/swap.c bio = bio_alloc(__GFP_WAIT | __GFP_HIGH, 1);
bio 66 kernel/power/swap.c if (!bio)
bio 68 kernel/power/swap.c bio->bi_sector = page_off * (PAGE_SIZE >> 9);
bio 69 kernel/power/swap.c bio->bi_bdev = resume_bdev;
bio 70 kernel/power/swap.c bio->bi_end_io = end_swap_bio_read;
bio 72 kernel/power/swap.c if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) {
bio 75 kernel/power/swap.c bio_put(bio);
bio 80 kernel/power/swap.c bio_get(bio);
bio 83 kernel/power/swap.c submit_bio(rw | (1 << BIO_RW_SYNC), bio);
bio 86 kernel/power/swap.c bio_set_pages_dirty(bio);
bio 87 kernel/power/swap.c bio_put(bio);
bio 91 kernel/power/swap.c bio->bi_private = *bio_chain;
bio 92 kernel/power/swap.c *bio_chain = bio;
bio 93 kernel/power/swap.c submit_bio(rw | (1 << BIO_RW_SYNC), bio);
bio 110 kernel/power/swap.c struct bio *bio;
bio 111 kernel/power/swap.c struct bio *next_bio;
bio 117 kernel/power/swap.c bio = *bio_chain;
bio 118 kernel/power/swap.c if (bio == NULL)
bio 120 kernel/power/swap.c while (bio) {
bio 123 kernel/power/swap.c next_bio = bio->bi_private;
bio 124 kernel/power/swap.c page = bio->bi_io_vec[0].bv_page;
bio 129 kernel/power/swap.c bio_put(bio);
bio 130 kernel/power/swap.c bio = next_bio;
bio 322 kernel/power/swap.c struct bio *bio;
bio 332 kernel/power/swap.c bio = NULL;
bio 338 kernel/power/swap.c &bio);
bio 346 kernel/power/swap.c err2 = wait_on_bio_chain(&bio);
bio 505 kernel/power/swap.c struct bio *bio;
bio 515 kernel/power/swap.c bio = NULL;
bio 521 kernel/power/swap.c error = swap_read_page(handle, data_of(*snapshot), &bio);
bio 525 kernel/power/swap.c error = wait_on_bio_chain(&bio);
bio 532 kernel/power/swap.c err2 = wait_on_bio_chain(&bio);
bio 124 mm/bounce.c struct bio *bio_orig = bio->bi_private;
bio 128 mm/bounce.c if (test_bit(BIO_EOPNOTSUPP, &bio->bi_flags))
bio 134 mm/bounce.c __bio_for_each_segment(bvec, bio, i, 0) {
bio 144 mm/bounce.c bio_put(bio);
bio 149 mm/bounce.c bounce_end_io(bio, page_pool, err);
bio 155 mm/bounce.c bounce_end_io(bio, isa_page_pool, err);
bio 160 mm/bounce.c struct bio *bio_orig = bio->bi_private;
bio 162 mm/bounce.c if (test_bit(BIO_UPTODATE, &bio->bi_flags))
bio 163 mm/bounce.c copy_to_high_bio_irq(bio_orig, bio);
bio 165 mm/bounce.c bounce_end_io(bio, pool, err);
bio 170 mm/bounce.c __bounce_end_io_read(bio, page_pool, err);
bio 175 mm/bounce.c __bounce_end_io_read(bio, isa_page_pool, err);
bio 182 mm/bounce.c struct bio *bio = NULL;
bio 198 mm/bounce.c if (!bio)
bio 199 mm/bounce.c bio = bio_alloc(GFP_NOIO, (*bio_orig)->bi_vcnt);
bio 201 mm/bounce.c to = bio->bi_io_vec + i;
bio 222 mm/bounce.c if (!bio)
bio 232 mm/bounce.c to = bio_iovec_idx(bio, i);
bio 240 mm/bounce.c bio->bi_bdev = (*bio_orig)->bi_bdev;
bio 241 mm/bounce.c bio->bi_flags |= (1 << BIO_BOUNCED);
bio 242 mm/bounce.c bio->bi_sector = (*bio_orig)->bi_sector;
bio 243 mm/bounce.c bio->bi_rw = (*bio_orig)->bi_rw;
bio 245 mm/bounce.c bio->bi_vcnt = (*bio_orig)->bi_vcnt;
bio 246 mm/bounce.c bio->bi_idx = (*bio_orig)->bi_idx;
bio 247 mm/bounce.c bio->bi_size = (*bio_orig)->bi_size;
bio 250 mm/bounce.c bio->bi_end_io = bounce_end_io_write;
bio 252 mm/bounce.c bio->bi_end_io = bounce_end_io_read;
bio 254 mm/bounce.c bio->bi_end_io = bounce_end_io_write_isa;
bio 256 mm/bounce.c bio->bi_end_io = bounce_end_io_read_isa;
bio 259 mm/bounce.c bio->bi_private = *bio_orig;
bio 260 mm/bounce.c *bio_orig = bio;
bio 22 mm/page_io.c static struct bio *get_swap_bio(gfp_t gfp_flags, pgoff_t index,
bio 25 mm/page_io.c struct bio *bio;
bio 27 mm/page_io.c bio = bio_alloc(gfp_flags, 1);
bio 28 mm/page_io.c if (bio) {
bio 33 mm/page_io.c bio->bi_sector = map_swap_page(sis, swp_offset(entry)) *
bio 35 mm/page_io.c bio->bi_bdev = sis->bdev;
bio 36 mm/page_io.c bio->bi_io_vec[0].bv_page = page;
bio 37 mm/page_io.c bio->bi_io_vec[0].bv_len = PAGE_SIZE;
bio 38 mm/page_io.c bio->bi_io_vec[0].bv_offset = 0;
bio 39 mm/page_io.c bio->bi_vcnt = 1;
bio 40 mm/page_io.c bio->bi_idx = 0;
bio 41 mm/page_io.c bio->bi_size = PAGE_SIZE;
bio 42 mm/page_io.c bio->bi_end_io = end_io;
bio 44 mm/page_io.c return bio;
bio 49 mm/page_io.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio 50 mm/page_io.c struct page *page = bio->bi_io_vec[0].bv_page;
bio 64 mm/page_io.c imajor(bio->bi_bdev->bd_inode),
bio 65 mm/page_io.c iminor(bio->bi_bdev->bd_inode),
bio 66 mm/page_io.c (unsigned long long)bio->bi_sector);
bio 70 mm/page_io.c bio_put(bio);
bio 75 mm/page_io.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio 76 mm/page_io.c struct page *page = bio->bi_io_vec[0].bv_page;
bio 82 mm/page_io.c imajor(bio->bi_bdev->bd_inode),
bio 83 mm/page_io.c iminor(bio->bi_bdev->bd_inode),
bio 84 mm/page_io.c (unsigned long long)bio->bi_sector);
bio 89 mm/page_io.c bio_put(bio);
bio 98 mm/page_io.c struct bio *bio;
bio 105 mm/page_io.c bio = get_swap_bio(GFP_NOIO, page_private(page), page,
bio 107 mm/page_io.c if (bio == NULL) {
bio 118 mm/page_io.c submit_bio(rw, bio);
bio 125 mm/page_io.c struct bio *bio;
bio 130 mm/page_io.c bio = get_swap_bio(GFP_KERNEL, page_private(page), page,
bio 132 mm/page_io.c if (bio == NULL) {
bio 138 mm/page_io.c submit_bio(READ, bio);