bio 1249 block/as-iosched.c sector_t rb_key = bio->bi_sector + bio_sectors(bio); bio 1255 block/as-iosched.c __rq = elv_rb_find(&ad->sort_list[bio_data_dir(bio)], rb_key); bio 1256 block/as-iosched.c if (__rq && elv_rq_merge_ok(__rq, bio)) { bio 168 block/blk-barrier.c if (bio_data_dir(q->orig_bar_rq->bio) == WRITE) bio 172 block/blk-barrier.c init_request_from_bio(rq, q->orig_bar_rq->bio); bio 258 block/blk-barrier.c set_bit(BIO_EOPNOTSUPP, &bio->bi_flags); bio 259 block/blk-barrier.c clear_bit(BIO_UPTODATE, &bio->bi_flags); bio 262 block/blk-barrier.c complete(bio->bi_private); bio 279 block/blk-barrier.c struct bio *bio; bio 289 block/blk-barrier.c bio = bio_alloc(GFP_KERNEL, 0); bio 290 block/blk-barrier.c if (!bio) bio 293 block/blk-barrier.c bio->bi_end_io = bio_end_empty_barrier; bio 294 block/blk-barrier.c bio->bi_private = &wait; bio 295 block/blk-barrier.c bio->bi_bdev = bdev; bio 296 block/blk-barrier.c submit_bio(WRITE_BARRIER, bio); bio 306 block/blk-barrier.c *error_sector = bio->bi_sector; bio 309 block/blk-barrier.c if (bio_flagged(bio, BIO_EOPNOTSUPP)) bio 311 block/blk-barrier.c else if (!bio_flagged(bio, BIO_UPTODATE)) bio 314 block/blk-barrier.c bio_put(bio); bio 323 block/blk-barrier.c set_bit(BIO_EOPNOTSUPP, &bio->bi_flags); bio 324 block/blk-barrier.c clear_bit(BIO_UPTODATE, &bio->bi_flags); bio 327 block/blk-barrier.c bio_put(bio); bio 344 block/blk-barrier.c struct bio *bio; bio 358 block/blk-barrier.c bio = bio_alloc(gfp_mask, 0); bio 359 block/blk-barrier.c if (!bio) bio 362 block/blk-barrier.c bio->bi_end_io = blkdev_discard_end_io; bio 363 block/blk-barrier.c bio->bi_bdev = bdev; bio 365 block/blk-barrier.c bio->bi_sector = sector; bio 368 block/blk-barrier.c bio->bi_size = q->max_hw_sectors << 9; bio 372 block/blk-barrier.c bio->bi_size = nr_sects << 9; bio 375 block/blk-barrier.c bio_get(bio); bio 376 block/blk-barrier.c submit_bio(DISCARD_BARRIER, bio); bio 379 block/blk-barrier.c if (bio_flagged(bio, BIO_EOPNOTSUPP)) bio 381 block/blk-barrier.c else if (!bio_flagged(bio, BIO_UPTODATE)) bio 383 block/blk-barrier.c bio_put(bio); bio 132 block/blk-core.c clear_bit(BIO_UPTODATE, &bio->bi_flags); bio 133 block/blk-core.c else if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) bio 136 block/blk-core.c if (unlikely(nbytes > bio->bi_size)) { bio 138 block/blk-core.c __func__, nbytes, bio->bi_size); bio 139 block/blk-core.c nbytes = bio->bi_size; bio 142 block/blk-core.c bio->bi_size -= nbytes; bio 143 block/blk-core.c bio->bi_sector += (nbytes >> 9); bio 145 block/blk-core.c if (bio_integrity(bio)) bio 146 block/blk-core.c bio_integrity_advance(bio, nbytes); bio 148 block/blk-core.c if (bio->bi_size == 0) bio 149 block/blk-core.c bio_endio(bio, error); bio 174 block/blk-core.c rq->bio, rq->biotail, bio 811 block/blk-core.c blk_add_trace_generic(q, bio, rw, BLK_TA_GETRQ); bio 828 block/blk-core.c rq = get_request(q, rw_flags, bio, GFP_NOIO); bio 837 block/blk-core.c blk_add_trace_generic(q, bio, rw, BLK_TA_SLEEPRQ); bio 855 block/blk-core.c rq = get_request(q, rw_flags, bio, GFP_NOIO); bio 1072 block/blk-core.c req->cpu = bio->bi_comp_cpu; bio 1078 block/blk-core.c if (bio_rw_ahead(bio) || bio_failfast(bio)) bio 1084 block/blk-core.c if (unlikely(bio_discard(bio))) { bio 1086 block/blk-core.c if (bio_barrier(bio)) bio 1089 block/blk-core.c } else if (unlikely(bio_barrier(bio))) bio 1092 block/blk-core.c if (bio_sync(bio)) bio 1094 block/blk-core.c if (bio_rw_meta(bio)) bio 1098 block/blk-core.c req->hard_sector = req->sector = bio->bi_sector; bio 1099 block/blk-core.c req->ioprio = bio_prio(bio); bio 1101 block/blk-core.c blk_rq_bio_prep(req->q, req, bio); bio 1108 block/blk-core.c const unsigned short prio = bio_prio(bio); bio 1109 block/blk-core.c const int sync = bio_sync(bio); bio 1112 block/blk-core.c nr_sectors = bio_sectors(bio); bio 1119 block/blk-core.c blk_queue_bounce(q, &bio); bio 1121 block/blk-core.c barrier = bio_barrier(bio); bio 1122 block/blk-core.c if (unlikely(barrier) && bio_has_data(bio) && bio 1128 block/blk-core.c discard = bio_discard(bio); bio 1139 block/blk-core.c el_ret = elv_merge(q, &req, bio); bio 1144 block/blk-core.c if (!ll_back_merge_fn(q, req, bio)) bio 1147 block/blk-core.c blk_add_trace_bio(q, bio, BLK_TA_BACKMERGE); bio 1149 block/blk-core.c req->biotail->bi_next = bio; bio 1150 block/blk-core.c req->biotail = bio; bio 1154 block/blk-core.c req->cpu = bio->bi_comp_cpu; bio 1163 block/blk-core.c if (!ll_front_merge_fn(q, req, bio)) bio 1166 block/blk-core.c blk_add_trace_bio(q, bio, BLK_TA_FRONTMERGE); bio 1168 block/blk-core.c bio->bi_next = req->bio; bio 1169 block/blk-core.c req->bio = bio; bio 1176 block/blk-core.c req->buffer = bio_data(bio); bio 1177 block/blk-core.c req->current_nr_sectors = bio_cur_sectors(bio); bio 1179 block/blk-core.c req->sector = req->hard_sector = bio->bi_sector; bio 1183 block/blk-core.c req->cpu = bio->bi_comp_cpu; bio 1200 block/blk-core.c rw_flags = bio_data_dir(bio); bio 1208 block/blk-core.c req = get_request_wait(q, rw_flags, bio); bio 1216 block/blk-core.c init_request_from_bio(req, bio); bio 1220 block/blk-core.c bio_flagged(bio, BIO_CPU_AFFINE)) bio 1232 block/blk-core.c bio_endio(bio, err); bio 1241 block/blk-core.c struct block_device *bdev = bio->bi_bdev; bio 1243 block/blk-core.c if (bio_sectors(bio) && bdev != bdev->bd_contains) { bio 1246 block/blk-core.c bio->bi_sector += p->start_sect; bio 1247 block/blk-core.c bio->bi_bdev = bdev->bd_contains; bio 1249 block/blk-core.c blk_add_trace_remap(bdev_get_queue(bio->bi_bdev), bio, bio 1250 block/blk-core.c bdev->bd_dev, bio->bi_sector, bio 1251 block/blk-core.c bio->bi_sector - p->start_sect); bio 1261 block/blk-core.c bdevname(bio->bi_bdev, b), bio 1262 block/blk-core.c bio->bi_rw, bio 1263 block/blk-core.c (unsigned long long)bio->bi_sector + bio_sectors(bio), bio 1264 block/blk-core.c (long long)(bio->bi_bdev->bd_inode->i_size >> 9)); bio 1266 block/blk-core.c set_bit(BIO_EOF, &bio->bi_flags); bio 1281 block/blk-core.c struct hd_struct *part = bio->bi_bdev->bd_part; bio 1284 block/blk-core.c return should_fail(&fail_make_request, bio->bi_size); bio 1317 block/blk-core.c maxsector = bio->bi_bdev->bd_inode->i_size >> 9; bio 1319 block/blk-core.c sector_t sector = bio->bi_sector; bio 1327 block/blk-core.c handle_bad_sector(bio); bio 1363 block/blk-core.c int ret, nr_sectors = bio_sectors(bio); bio 1369 block/blk-core.c if (bio_check_eod(bio, nr_sectors)) bio 1385 block/blk-core.c q = bdev_get_queue(bio->bi_bdev); bio 1390 block/blk-core.c bdevname(bio->bi_bdev, b), bio 1391 block/blk-core.c (long long) bio->bi_sector); bio 1393 block/blk-core.c bio_endio(bio, err); bio 1399 block/blk-core.c bdevname(bio->bi_bdev, b), bio 1400 block/blk-core.c bio_sectors(bio), bio 1408 block/blk-core.c if (should_fail_request(bio)) bio 1415 block/blk-core.c blk_partition_remap(bio); bio 1417 block/blk-core.c if (bio_integrity_enabled(bio) && bio_integrity_prep(bio)) bio 1421 block/blk-core.c blk_add_trace_remap(q, bio, old_dev, bio->bi_sector, bio 1424 block/blk-core.c blk_add_trace_bio(q, bio, BLK_TA_QUEUE); bio 1426 block/blk-core.c old_sector = bio->bi_sector; bio 1427 block/blk-core.c old_dev = bio->bi_bdev->bd_dev; bio 1429 block/blk-core.c if (bio_check_eod(bio, nr_sectors)) bio 1431 block/blk-core.c if ((bio_empty_barrier(bio) && !q->prepare_flush_fn) || bio 1432 block/blk-core.c (bio_discard(bio) && !q->prepare_discard_fn)) { bio 1437 block/blk-core.c ret = q->make_request_fn(q, bio); bio 1456 block/blk-core.c *(current->bio_tail) = bio; bio 1457 block/blk-core.c bio->bi_next = NULL; bio 1458 block/blk-core.c current->bio_tail = &bio->bi_next; bio 1479 block/blk-core.c BUG_ON(bio->bi_next); bio 1481 block/blk-core.c current->bio_list = bio->bi_next; bio 1482 block/blk-core.c if (bio->bi_next == NULL) bio 1485 block/blk-core.c bio->bi_next = NULL; bio 1486 block/blk-core.c __generic_make_request(bio); bio 1487 block/blk-core.c bio = current->bio_list; bio 1488 block/blk-core.c } while (bio); bio 1505 block/blk-core.c int count = bio_sectors(bio); bio 1507 block/blk-core.c bio->bi_rw |= rw; bio 1513 block/blk-core.c if (bio_has_data(bio)) { bio 1517 block/blk-core.c task_io_account_read(bio->bi_size); bio 1526 block/blk-core.c (unsigned long long)bio->bi_sector, bio 1527 block/blk-core.c bdevname(bio->bi_bdev, b)); bio 1531 block/blk-core.c generic_make_request(bio); bio 1634 block/blk-core.c struct bio *bio; bio 1663 block/blk-core.c while ((bio = req->bio) != NULL) { bio 1672 block/blk-core.c bio->bi_sector = req->sector; bio 1674 block/blk-core.c if (nr_bytes >= bio->bi_size) { bio 1675 block/blk-core.c req->bio = bio->bi_next; bio 1676 block/blk-core.c nbytes = bio->bi_size; bio 1677 block/blk-core.c req_bio_endio(req, bio, nbytes, error); bio 1681 block/blk-core.c int idx = bio->bi_idx + next_idx; bio 1683 block/blk-core.c if (unlikely(bio->bi_idx >= bio->bi_vcnt)) { bio 1686 block/blk-core.c __func__, bio->bi_idx, bio->bi_vcnt); bio 1690 block/blk-core.c nbytes = bio_iovec_idx(bio, idx)->bv_len; bio 1691 block/blk-core.c BIO_BUG_ON(nbytes > bio->bi_size); bio 1712 block/blk-core.c bio = req->bio; bio 1713 block/blk-core.c if (bio) { bio 1725 block/blk-core.c if (!req->bio) bio 1732 block/blk-core.c req_bio_endio(req, bio, bio_nbytes, error); bio 1733 block/blk-core.c bio->bi_idx += next_idx; bio 1734 block/blk-core.c bio_iovec(bio)->bv_offset += nr_bytes; bio 1735 block/blk-core.c bio_iovec(bio)->bv_len -= nr_bytes; bio 1815 block/blk-core.c if (rq->bio) bio 1816 block/blk-core.c return rq->bio->bi_size; bio 1852 block/blk-core.c if (rq->bio) { bio 1942 block/blk-core.c if (rq->bio && __end_that_request_first(rq, error, nr_bytes)) bio 2042 block/blk-core.c rq->cmd_flags |= (bio->bi_rw & 3); bio 2044 block/blk-core.c if (bio_has_data(bio)) { bio 2045 block/blk-core.c rq->nr_phys_segments = bio_phys_segments(q, bio); bio 2046 block/blk-core.c rq->buffer = bio_data(bio); bio 2048 block/blk-core.c rq->current_nr_sectors = bio_cur_sectors(bio); bio 2050 block/blk-core.c rq->hard_nr_sectors = rq->nr_sectors = bio_sectors(bio); bio 2051 block/blk-core.c rq->data_len = bio->bi_size; bio 2053 block/blk-core.c rq->bio = rq->biotail = bio; bio 2055 block/blk-core.c if (bio->bi_bdev) bio 2056 block/blk-core.c rq->rq_disk = bio->bi_bdev->bd_disk; bio 15 block/blk-map.c if (!rq->bio) bio 16 block/blk-map.c blk_rq_bio_prep(q, rq, bio); bio 17 block/blk-map.c else if (!ll_back_merge_fn(q, rq, bio)) bio 20 block/blk-map.c rq->biotail->bi_next = bio; bio 21 block/blk-map.c rq->biotail = bio; bio 23 block/blk-map.c rq->data_len += bio->bi_size; bio 33 block/blk-map.c if (bio) { bio 34 block/blk-map.c if (bio_flagged(bio, BIO_USER_MAPPED)) bio 35 block/blk-map.c bio_unmap_user(bio); bio 37 block/blk-map.c ret = bio_uncopy_user(bio); bio 48 block/blk-map.c struct bio *bio, *orig_bio; bio 59 block/blk-map.c bio = bio_map_user(q, NULL, uaddr, len, reading, gfp_mask); bio 61 block/blk-map.c bio = bio_copy_user(q, map_data, uaddr, len, reading, gfp_mask); bio 63 block/blk-map.c if (IS_ERR(bio)) bio 64 block/blk-map.c return PTR_ERR(bio); bio 67 block/blk-map.c bio->bi_flags |= (1 << BIO_NULL_MAPPED); bio 69 block/blk-map.c orig_bio = bio; bio 70 block/blk-map.c blk_queue_bounce(q, &bio); bio 76 block/blk-map.c bio_get(bio); bio 78 block/blk-map.c ret = blk_rq_append_bio(q, rq, bio); bio 80 block/blk-map.c return bio->bi_size; bio 83 block/blk-map.c bio_endio(bio, 0); bio 85 block/blk-map.c bio_put(bio); bio 116 block/blk-map.c struct bio *bio = NULL; bio 149 block/blk-map.c if (!bio) bio 150 block/blk-map.c bio = rq->bio; bio 155 block/blk-map.c if (!bio_flagged(bio, BIO_USER_MAPPED)) bio 161 block/blk-map.c blk_rq_unmap_user(bio); bio 162 block/blk-map.c rq->bio = NULL; bio 194 block/blk-map.c struct bio *bio; bio 211 block/blk-map.c bio = bio_copy_user_iov(q, map_data, iov, iov_count, read, bio 214 block/blk-map.c bio = bio_map_user_iov(q, NULL, iov, iov_count, read, gfp_mask); bio 216 block/blk-map.c if (IS_ERR(bio)) bio 217 block/blk-map.c return PTR_ERR(bio); bio 219 block/blk-map.c if (bio->bi_size != len) { bio 220 block/blk-map.c bio_endio(bio, 0); bio 221 block/blk-map.c bio_unmap_user(bio); bio 225 block/blk-map.c if (!bio_flagged(bio, BIO_USER_MAPPED)) bio 228 block/blk-map.c blk_queue_bounce(q, &bio); bio 229 block/blk-map.c bio_get(bio); bio 230 block/blk-map.c blk_rq_bio_prep(q, rq, bio); bio 247 block/blk-map.c struct bio *mapped_bio; bio 250 block/blk-map.c while (bio) { bio 251 block/blk-map.c mapped_bio = bio; bio 252 block/blk-map.c if (unlikely(bio_flagged(bio, BIO_BOUNCED))) bio 253 block/blk-map.c mapped_bio = bio->bi_private; bio 259 block/blk-map.c mapped_bio = bio; bio 260 block/blk-map.c bio = bio->bi_next; bio 285 block/blk-map.c struct bio *bio; bio 294 block/blk-map.c bio = bio_copy_kern(q, kbuf, len, gfp_mask, reading); bio 296 block/blk-map.c bio = bio_map_kern(q, kbuf, len, gfp_mask); bio 298 block/blk-map.c if (IS_ERR(bio)) bio 299 block/blk-map.c return PTR_ERR(bio); bio 302 block/blk-map.c bio->bi_rw |= (1 << BIO_RW); bio 307 block/blk-map.c blk_rq_bio_prep(q, rq, bio); bio 308 block/blk-map.c blk_queue_bounce(q, &rq->bio); bio 25 block/blk-merge.c rq->hard_cur_sectors = bio_cur_sectors(rq->bio); bio 27 block/blk-merge.c rq->buffer = bio_data(rq->bio); bio 52 block/blk-merge.c if (!rq->bio) bio 92 block/blk-merge.c struct bio *nxt = bio->bi_next; bio 94 block/blk-merge.c rq.bio = rq.biotail = bio; bio 95 block/blk-merge.c bio->bi_next = NULL; bio 97 block/blk-merge.c bio->bi_next = nxt; bio 98 block/blk-merge.c bio->bi_phys_segments = rq.nr_phys_segments; bio 99 block/blk-merge.c bio->bi_flags |= (1 << BIO_SEG_VALID); bio 109 block/blk-merge.c if (bio->bi_size + nxt->bi_size > q->max_segment_size) bio 112 block/blk-merge.c if (!bio_has_data(bio)) bio 115 block/blk-merge.c if (!BIOVEC_PHYS_MERGEABLE(__BVEC_END(bio), __BVEC_START(nxt))) bio 122 block/blk-merge.c if (BIO_SEG_BOUNDARY(q, bio, nxt)) bio 220 block/blk-merge.c int nr_phys_segs = bio_phys_segments(q, bio); bio 241 block/blk-merge.c int nr_phys_segs = bio_phys_segments(q, bio); bio 269 block/blk-merge.c if (req->nr_sectors + bio_sectors(bio) > max_sectors) { bio 277 block/blk-merge.c if (!bio_flagged(bio, BIO_SEG_VALID)) bio 278 block/blk-merge.c blk_recount_segments(q, bio); bio 280 block/blk-merge.c return ll_new_hw_segment(q, req, bio); bio 294 block/blk-merge.c if (req->nr_sectors + bio_sectors(bio) > max_sectors) { bio 300 block/blk-merge.c if (!bio_flagged(bio, BIO_SEG_VALID)) bio 301 block/blk-merge.c blk_recount_segments(q, bio); bio 302 block/blk-merge.c if (!bio_flagged(req->bio, BIO_SEG_VALID)) bio 303 block/blk-merge.c blk_recount_segments(q, req->bio); bio 305 block/blk-merge.c return ll_new_hw_segment(q, req, bio); bio 327 block/blk-merge.c if (blk_phys_contig_segment(q, req->biotail, next->bio)) bio 382 block/blk-merge.c req->biotail->bi_next = next->bio; bio 93 block/blk.h __rq_for_each_bio(_iter.bio, _rq) \ bio 94 block/blk.h bip_for_each_vec(bvl, _iter.bio->bi_integrity, _iter.i) bio 83 block/bsg.c struct bio *bio; bio 84 block/bsg.c struct bio *bidi_bio; bio 313 block/bsg.c blk_rq_unmap_user(next_rq->bio); bio 330 block/bsg.c bd->name, rq, bc, bc->bio, uptodate); bio 356 block/bsg.c bc->bio = rq->bio; bio 358 block/bsg.c bc->bidi_bio = rq->next_rq->bio; bio 420 block/bsg.c dprintk("rq %p bio %p %u\n", rq, bio, rq->errors); bio 463 block/bsg.c blk_rq_unmap_user(bio); bio 509 block/bsg.c tret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio, bio 544 block/bsg.c ret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio, bio 920 block/bsg.c struct bio *bio, *bidi_bio = NULL; bio 930 block/bsg.c bio = rq->bio; bio 932 block/bsg.c bidi_bio = rq->next_rq->bio; bio 934 block/bsg.c ret = blk_complete_sgv4_hdr_rq(rq, &hdr, bio, bidi_bio); bio 241 block/cfq-iosched.c if (bio_data_dir(bio) == READ || bio_sync(bio)) bio 647 block/cfq-iosched.c cfqq = cic_to_cfqq(cic, cfq_bio_sync(bio)); bio 649 block/cfq-iosched.c sector_t sector = bio->bi_sector + bio_sectors(bio); bio 701 block/cfq-iosched.c __rq = cfq_find_rq_fmerge(cfqd, bio); bio 702 block/cfq-iosched.c if (__rq && elv_rq_merge_ok(__rq, bio)) { bio 744 block/cfq-iosched.c if (cfq_bio_sync(bio) && !rq_is_sync(rq)) bio 755 block/cfq-iosched.c cfqq = cic_to_cfqq(cic, cfq_bio_sync(bio)); bio 137 block/deadline-iosched.c sector_t sector = bio->bi_sector + bio_sectors(bio); bio 139 block/deadline-iosched.c __rq = elv_rb_find(&dd->sort_list[bio_data_dir(bio)], sector); bio 143 block/deadline-iosched.c if (elv_rq_merge_ok(__rq, bio)) { bio 65 block/elevator.c return e->ops->elevator_allow_merge_fn(q, rq, bio); bio 81 block/elevator.c if (bio_discard(bio) != bio_discard(rq->bio)) bio 87 block/elevator.c if (bio_data_dir(bio) != rq_data_dir(rq)) bio 93 block/elevator.c if (rq->rq_disk != bio->bi_bdev->bd_disk || rq->special) bio 99 block/elevator.c if (bio_integrity(bio) != blk_integrity_rq(rq)) bio 102 block/elevator.c if (!elv_iosched_allow_merge(rq, bio)) bio 116 block/elevator.c if (elv_rq_merge_ok(__rq, bio)) { bio 117 block/elevator.c if (__rq->sector + __rq->nr_sectors == bio->bi_sector) bio 119 block/elevator.c else if (__rq->sector - bio_sectors(bio) == bio->bi_sector) bio 499 block/elevator.c ret = elv_try_merge(q->last_merge, bio); bio 512 block/elevator.c __rq = elv_rqhash_find(q, bio->bi_sector); bio 513 block/elevator.c if (__rq && elv_rq_merge_ok(__rq, bio)) { bio 519 block/elevator.c return e->ops->elevator_merge_fn(q, req, bio); bio 120 block/ioctl.c set_bit(BIO_EOPNOTSUPP, &bio->bi_flags); bio 121 block/ioctl.c clear_bit(BIO_UPTODATE, &bio->bi_flags); bio 123 block/ioctl.c complete(bio->bi_private); bio 147 block/ioctl.c struct bio *bio; bio 149 block/ioctl.c bio = bio_alloc(GFP_KERNEL, 0); bio 150 block/ioctl.c if (!bio) bio 153 block/ioctl.c bio->bi_end_io = blk_ioc_discard_endio; bio 154 block/ioctl.c bio->bi_bdev = bdev; bio 155 block/ioctl.c bio->bi_private = &wait; bio 156 block/ioctl.c bio->bi_sector = start; bio 159 block/ioctl.c bio->bi_size = q->max_hw_sectors << 9; bio 163 block/ioctl.c bio->bi_size = len << 9; bio 166 block/ioctl.c submit_bio(DISCARD_NOBARRIER, bio); bio 170 block/ioctl.c if (bio_flagged(bio, BIO_EOPNOTSUPP)) bio 172 block/ioctl.c else if (!bio_flagged(bio, BIO_UPTODATE)) bio 174 block/ioctl.c bio_put(bio); bio 222 block/scsi_ioctl.c blk_rq_unmap_user(rq->bio); bio 255 block/scsi_ioctl.c rq->bio = bio; bio 270 block/scsi_ioctl.c struct bio *bio; bio 327 block/scsi_ioctl.c bio = rq->bio; bio 343 block/scsi_ioctl.c return blk_complete_sghdr_rq(rq, hdr, bio); bio 51 fs/bio-integrity.c BUG_ON(bio == NULL); bio 70 fs/bio-integrity.c bip->bip_bio = bio; bio 71 fs/bio-integrity.c bio->bi_integrity = bip; bio 91 fs/bio-integrity.c return bio_integrity_alloc_bioset(bio, gfp_mask, nr_vecs, fs_bio_set); bio 105 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity; bio 110 fs/bio-integrity.c if (!bio_flagged(bio, BIO_CLONED) && !bio_flagged(bio, BIO_FS_INTEGRITY) bio 117 fs/bio-integrity.c bio->bi_integrity = NULL; bio 133 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity; bio 184 fs/bio-integrity.c if (bio_integrity(bio)) bio 187 fs/bio-integrity.c return bdev_integrity_enabled(bio->bi_bdev, bio_data_dir(bio)); bio 221 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev); bio 223 fs/bio-integrity.c BUG_ON(bio->bi_size == 0); bio 225 fs/bio-integrity.c return bi->tag_size * (bio->bi_size / bi->sector_size); bio 231 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity; bio 232 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev); bio 270 fs/bio-integrity.c BUG_ON(bio_data_dir(bio) != WRITE); bio 272 fs/bio-integrity.c return bio_integrity_tag(bio, tag_buf, len, 1); bio 288 fs/bio-integrity.c BUG_ON(bio_data_dir(bio) != READ); bio 290 fs/bio-integrity.c return bio_integrity_tag(bio, tag_buf, len, 0); bio 305 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev); bio 308 fs/bio-integrity.c sector_t sector = bio->bi_sector; bio 310 fs/bio-integrity.c void *prot_buf = bio->bi_integrity->bip_buf; bio 313 fs/bio-integrity.c bix.disk_name = bio->bi_bdev->bd_disk->disk_name; bio 316 fs/bio-integrity.c bio_for_each_segment(bv, bio, i) { bio 329 fs/bio-integrity.c BUG_ON(total > bio->bi_integrity->bip_size); bio 365 fs/bio-integrity.c bi = bdev_get_integrity(bio->bi_bdev); bio 366 fs/bio-integrity.c q = bdev_get_queue(bio->bi_bdev); bio 368 fs/bio-integrity.c BUG_ON(bio_integrity(bio)); bio 370 fs/bio-integrity.c sectors = bio_integrity_hw_sectors(bi, bio_sectors(bio)); bio 385 fs/bio-integrity.c bip = bio_integrity_alloc(bio, GFP_NOIO, nr_pages); bio 394 fs/bio-integrity.c bip->bip_sector = bio->bi_sector; bio 408 fs/bio-integrity.c ret = bio_integrity_add_page(bio, virt_to_page(buf), bio 423 fs/bio-integrity.c if (bio_data_dir(bio) == READ) { bio 424 fs/bio-integrity.c bip->bip_end_io = bio->bi_end_io; bio 425 fs/bio-integrity.c bio->bi_end_io = bio_integrity_endio; bio 429 fs/bio-integrity.c if (bio_data_dir(bio) == WRITE) bio 430 fs/bio-integrity.c bio_integrity_generate(bio); bio 446 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev); bio 449 fs/bio-integrity.c sector_t sector = bio->bi_integrity->bip_sector; bio 451 fs/bio-integrity.c void *prot_buf = bio->bi_integrity->bip_buf; bio 454 fs/bio-integrity.c bix.disk_name = bio->bi_bdev->bd_disk->disk_name; bio 457 fs/bio-integrity.c bio_for_each_segment(bv, bio, i) { bio 475 fs/bio-integrity.c BUG_ON(total > bio->bi_integrity->bip_size); bio 495 fs/bio-integrity.c struct bio *bio = bip->bip_bio; bio 498 fs/bio-integrity.c if (bio_integrity_verify(bio)) { bio 499 fs/bio-integrity.c clear_bit(BIO_UPTODATE, &bio->bi_flags); bio 504 fs/bio-integrity.c bio->bi_end_io = bip->bip_end_io; bio 506 fs/bio-integrity.c if (bio->bi_end_io) bio 507 fs/bio-integrity.c bio->bi_end_io(bio, error); bio 524 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity; bio 526 fs/bio-integrity.c BUG_ON(bip->bip_bio != bio); bio 595 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity; bio 596 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev); bio 621 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity; bio 622 fs/bio-integrity.c struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev); bio 627 fs/bio-integrity.c BUG_ON(!bio_flagged(bio, BIO_CLONED)); bio 647 fs/bio-integrity.c struct bio_integrity_payload *bip = bio->bi_integrity; bio 650 fs/bio-integrity.c if (bio_integrity(bio) == 0) bio 653 fs/bio-integrity.c bi = bdev_get_integrity(bio->bi_bdev); bio 672 fs/bio-integrity.c bp->bip1.bip_sector = bio->bi_integrity->bip_sector; bio 673 fs/bio-integrity.c bp->bip2.bip_sector = bio->bi_integrity->bip_sector + nr_sectors; bio 696 fs/bio-integrity.c bip = bio_integrity_alloc_bioset(bio, GFP_NOIO, bip_src->bip_vcnt, bs); bio 109 fs/bio.c if (bio->bi_io_vec) { bio 110 fs/bio.c const int pool_idx = BIO_POOL_IDX(bio); bio 114 fs/bio.c mempool_free(bio->bi_io_vec, bio_set->bvec_pools[pool_idx]); bio 117 fs/bio.c if (bio_integrity(bio)) bio 118 fs/bio.c bio_integrity_free(bio, bio_set); bio 120 fs/bio.c mempool_free(bio, bio_set->bio_pool); bio 128 fs/bio.c bio_free(bio, fs_bio_set); bio 133 fs/bio.c kfree(bio->bi_io_vec); bio 134 fs/bio.c kfree(bio); bio 139 fs/bio.c memset(bio, 0, sizeof(*bio)); bio 140 fs/bio.c bio->bi_flags = 1 << BIO_UPTODATE; bio 141 fs/bio.c bio->bi_comp_cpu = -1; bio 142 fs/bio.c atomic_set(&bio->bi_cnt, 1); bio 160 fs/bio.c struct bio *bio_alloc_bioset(gfp_t gfp_mask, int nr_iovecs, struct bio_set *bs) bio 162 fs/bio.c struct bio *bio; bio 165 fs/bio.c bio = mempool_alloc(bs->bio_pool, gfp_mask); bio 167 fs/bio.c bio = kmalloc(sizeof(*bio), gfp_mask); bio 169 fs/bio.c if (likely(bio)) { bio 172 fs/bio.c bio_init(bio); bio 179 fs/bio.c mempool_free(bio, bs->bio_pool); bio 181 fs/bio.c kfree(bio); bio 182 fs/bio.c bio = NULL; bio 185 fs/bio.c bio->bi_flags |= idx << BIO_POOL_OFFSET; bio 186 fs/bio.c bio->bi_max_vecs = bvec_nr_vecs(idx); bio 188 fs/bio.c bio->bi_io_vec = bvl; bio 191 fs/bio.c return bio; bio 194 fs/bio.c struct bio *bio_alloc(gfp_t gfp_mask, int nr_iovecs) bio 196 fs/bio.c struct bio *bio = bio_alloc_bioset(gfp_mask, nr_iovecs, fs_bio_set); bio 198 fs/bio.c if (bio) bio 199 fs/bio.c bio->bi_destructor = bio_fs_destructor; bio 201 fs/bio.c return bio; bio 211 fs/bio.c struct bio *bio_kmalloc(gfp_t gfp_mask, int nr_iovecs) bio 213 fs/bio.c struct bio *bio = bio_alloc_bioset(gfp_mask, nr_iovecs, NULL); bio 215 fs/bio.c if (bio) bio 216 fs/bio.c bio->bi_destructor = bio_kmalloc_destructor; bio 218 fs/bio.c return bio; bio 227 fs/bio.c bio_for_each_segment(bv, bio, i) { bio 246 fs/bio.c BIO_BUG_ON(!atomic_read(&bio->bi_cnt)); bio 251 fs/bio.c if (atomic_dec_and_test(&bio->bi_cnt)) { bio 252 fs/bio.c bio->bi_next = NULL; bio 253 fs/bio.c bio->bi_destructor(bio); bio 259 fs/bio.c if (unlikely(!bio_flagged(bio, BIO_SEG_VALID))) bio 260 fs/bio.c blk_recount_segments(q, bio); bio 262 fs/bio.c return bio->bi_phys_segments; bio 276 fs/bio.c memcpy(bio->bi_io_vec, bio_src->bi_io_vec, bio 283 fs/bio.c bio->bi_sector = bio_src->bi_sector; bio 284 fs/bio.c bio->bi_bdev = bio_src->bi_bdev; bio 285 fs/bio.c bio->bi_flags |= 1 << BIO_CLONED; bio 286 fs/bio.c bio->bi_rw = bio_src->bi_rw; bio 287 fs/bio.c bio->bi_vcnt = bio_src->bi_vcnt; bio 288 fs/bio.c bio->bi_size = bio_src->bi_size; bio 289 fs/bio.c bio->bi_idx = bio_src->bi_idx; bio 299 fs/bio.c struct bio *bio_clone(struct bio *bio, gfp_t gfp_mask) bio 301 fs/bio.c struct bio *b = bio_alloc_bioset(gfp_mask, bio->bi_max_vecs, fs_bio_set); bio 307 fs/bio.c __bio_clone(b, bio); bio 309 fs/bio.c if (bio_integrity(bio)) { bio 312 fs/bio.c ret = bio_integrity_clone(b, bio, fs_bio_set); bio 354 fs/bio.c if (unlikely(bio_flagged(bio, BIO_CLONED))) bio 357 fs/bio.c if (((bio->bi_size + len) >> 9) > max_sectors) bio 365 fs/bio.c if (bio->bi_vcnt > 0) { bio 366 fs/bio.c struct bio_vec *prev = &bio->bi_io_vec[bio->bi_vcnt - 1]; bio 374 fs/bio.c .bi_bdev = bio->bi_bdev, bio 375 fs/bio.c .bi_sector = bio->bi_sector, bio 376 fs/bio.c .bi_size = bio->bi_size, bio 377 fs/bio.c .bi_rw = bio->bi_rw, bio 390 fs/bio.c if (bio->bi_vcnt >= bio->bi_max_vecs) bio 398 fs/bio.c while (bio->bi_phys_segments >= q->max_phys_segments bio 399 fs/bio.c || bio->bi_phys_segments >= q->max_hw_segments) { bio 405 fs/bio.c blk_recount_segments(q, bio); bio 412 fs/bio.c bvec = &bio->bi_io_vec[bio->bi_vcnt]; bio 424 fs/bio.c .bi_bdev = bio->bi_bdev, bio 425 fs/bio.c .bi_sector = bio->bi_sector, bio 426 fs/bio.c .bi_size = bio->bi_size, bio 427 fs/bio.c .bi_rw = bio->bi_rw, bio 443 fs/bio.c if (bio->bi_vcnt && (BIOVEC_PHYS_MERGEABLE(bvec-1, bvec))) bio 444 fs/bio.c bio->bi_flags &= ~(1 << BIO_SEG_VALID); bio 446 fs/bio.c bio->bi_vcnt++; bio 447 fs/bio.c bio->bi_phys_segments++; bio 449 fs/bio.c bio->bi_size += len; bio 470 fs/bio.c return __bio_add_page(q, bio, page, len, offset, q->max_hw_sectors); bio 489 fs/bio.c struct request_queue *q = bdev_get_queue(bio->bi_bdev); bio 490 fs/bio.c return __bio_add_page(q, bio, page, len, offset, q->max_sectors); bio 504 fs/bio.c memcpy(bmd->iovecs, bio->bi_io_vec, sizeof(struct bio_vec) * bio->bi_vcnt); bio 508 fs/bio.c bio->bi_private = bmd; bio 549 fs/bio.c int read = bio_data_dir(bio) == READ; bio 551 fs/bio.c __bio_for_each_segment(bvec, bio, i, 0) { bio 602 fs/bio.c struct bio_map_data *bmd = bio->bi_private; bio 605 fs/bio.c if (!bio_flagged(bio, BIO_NULL_MAPPED)) bio 606 fs/bio.c ret = __bio_copy_iov(bio, bmd->iovecs, bmd->sgvecs, bio 609 fs/bio.c bio_put(bio); bio 626 fs/bio.c struct bio *bio_copy_user_iov(struct request_queue *q, bio 634 fs/bio.c struct bio *bio; bio 657 fs/bio.c bio = bio_alloc(gfp_mask, nr_pages); bio 658 fs/bio.c if (!bio) bio 661 fs/bio.c bio->bi_rw |= (!write_to_vm << BIO_RW); bio 689 fs/bio.c if (bio_add_pc_page(q, bio, page, bytes, 0) < bytes) bio 702 fs/bio.c ret = __bio_copy_iov(bio, bio->bi_io_vec, iov, iov_count, 0, 0); bio 707 fs/bio.c bio_set_map_data(bmd, bio, iov, iov_count, map_data ? 0 : 1); bio 708 fs/bio.c return bio; bio 711 fs/bio.c bio_for_each_segment(bvec, bio, i) bio 714 fs/bio.c bio_put(bio); bio 733 fs/bio.c struct bio *bio_copy_user(struct request_queue *q, struct rq_map_data *map_data, bio 745 fs/bio.c static struct bio *__bio_map_user_iov(struct request_queue *q, bio 753 fs/bio.c struct bio *bio; bio 774 fs/bio.c bio = bio_alloc(gfp_mask, nr_pages); bio 775 fs/bio.c if (!bio) bio 811 fs/bio.c if (bio_add_pc_page(q, bio, pages[j], bytes, offset) < bio 833 fs/bio.c bio->bi_rw |= (1 << BIO_RW); bio 835 fs/bio.c bio->bi_bdev = bdev; bio 836 fs/bio.c bio->bi_flags |= (1 << BIO_USER_MAPPED); bio 837 fs/bio.c return bio; bio 847 fs/bio.c bio_put(bio); bio 863 fs/bio.c struct bio *bio_map_user(struct request_queue *q, struct block_device *bdev, bio 887 fs/bio.c struct bio *bio_map_user_iov(struct request_queue *q, struct block_device *bdev, bio 891 fs/bio.c struct bio *bio; bio 893 fs/bio.c bio = __bio_map_user_iov(q, bdev, iov, iov_count, write_to_vm, bio 895 fs/bio.c if (IS_ERR(bio)) bio 896 fs/bio.c return bio; bio 904 fs/bio.c bio_get(bio); bio 906 fs/bio.c return bio; bio 917 fs/bio.c __bio_for_each_segment(bvec, bio, i, 0) { bio 918 fs/bio.c if (bio_data_dir(bio) == READ) bio 924 fs/bio.c bio_put(bio); bio 938 fs/bio.c __bio_unmap_user(bio); bio 939 fs/bio.c bio_put(bio); bio 944 fs/bio.c bio_put(bio); bio 948 fs/bio.c static struct bio *__bio_map_kern(struct request_queue *q, void *data, bio 956 fs/bio.c struct bio *bio; bio 958 fs/bio.c bio = bio_alloc(gfp_mask, nr_pages); bio 959 fs/bio.c if (!bio) bio 972 fs/bio.c if (bio_add_pc_page(q, bio, virt_to_page(data), bytes, bio 981 fs/bio.c bio->bi_end_io = bio_map_kern_endio; bio 982 fs/bio.c return bio; bio 995 fs/bio.c struct bio *bio_map_kern(struct request_queue *q, void *data, unsigned int len, bio 998 fs/bio.c struct bio *bio; bio 1000 fs/bio.c bio = __bio_map_kern(q, data, len, gfp_mask); bio 1001 fs/bio.c if (IS_ERR(bio)) bio 1002 fs/bio.c return bio; bio 1004 fs/bio.c if (bio->bi_size == len) bio 1005 fs/bio.c return bio; bio 1010 fs/bio.c bio_put(bio); bio 1017 fs/bio.c const int read = bio_data_dir(bio) == READ; bio 1018 fs/bio.c struct bio_map_data *bmd = bio->bi_private; bio 1022 fs/bio.c __bio_for_each_segment(bvec, bio, i, 0) { bio 1034 fs/bio.c bio_put(bio); bio 1048 fs/bio.c struct bio *bio_copy_kern(struct request_queue *q, void *data, unsigned int len, bio 1051 fs/bio.c struct bio *bio; bio 1055 fs/bio.c bio = bio_copy_user(q, NULL, (unsigned long)data, len, 1, gfp_mask); bio 1056 fs/bio.c if (IS_ERR(bio)) bio 1057 fs/bio.c return bio; bio 1062 fs/bio.c bio_for_each_segment(bvec, bio, i) { bio 1070 fs/bio.c bio->bi_end_io = bio_copy_kern_endio; bio 1072 fs/bio.c return bio; bio 1106 fs/bio.c struct bio_vec *bvec = bio->bi_io_vec; bio 1109 fs/bio.c for (i = 0; i < bio->bi_vcnt; i++) { bio 1119 fs/bio.c struct bio_vec *bvec = bio->bi_io_vec; bio 1122 fs/bio.c for (i = 0; i < bio->bi_vcnt; i++) { bio 1145 fs/bio.c static struct bio *bio_dirty_list; bio 1153 fs/bio.c struct bio *bio; bio 1156 fs/bio.c bio = bio_dirty_list; bio 1160 fs/bio.c while (bio) { bio 1161 fs/bio.c struct bio *next = bio->bi_private; bio 1163 fs/bio.c bio_set_pages_dirty(bio); bio 1164 fs/bio.c bio_release_pages(bio); bio 1165 fs/bio.c bio_put(bio); bio 1166 fs/bio.c bio = next; bio 1172 fs/bio.c struct bio_vec *bvec = bio->bi_io_vec; bio 1176 fs/bio.c for (i = 0; i < bio->bi_vcnt; i++) { bio 1191 fs/bio.c bio->bi_private = bio_dirty_list; bio 1192 fs/bio.c bio_dirty_list = bio; bio 1196 fs/bio.c bio_put(bio); bio 1217 fs/bio.c clear_bit(BIO_UPTODATE, &bio->bi_flags); bio 1218 fs/bio.c else if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) bio 1221 fs/bio.c if (bio->bi_end_io) bio 1222 fs/bio.c bio->bi_end_io(bio, error); bio 1228 fs/bio.c struct bio *master = bp->bio1.bi_private; bio 1316 fs/bio.c unsigned int sector_sz = queue_hardsect_size(bio->bi_bdev->bd_disk->queue); bio 1323 fs/bio.c if (index >= bio->bi_idx) bio 1324 fs/bio.c index = bio->bi_vcnt - 1; bio 1326 fs/bio.c __bio_for_each_segment(bv, bio, i, 0) { bio 1421 fs/bio.c bio_slab = KMEM_CACHE(bio, SLAB_HWCACHE_ALIGN|SLAB_PANIC); bio 2909 fs/buffer.c struct buffer_head *bh = bio->bi_private; bio 2912 fs/buffer.c set_bit(BIO_EOPNOTSUPP, &bio->bi_flags); bio 2916 fs/buffer.c bh->b_end_io(bh, test_bit(BIO_UPTODATE, &bio->bi_flags)); bio 2917 fs/buffer.c bio_put(bio); bio 2922 fs/buffer.c struct bio *bio; bio 2946 fs/buffer.c bio = bio_alloc(GFP_NOIO, 1); bio 2948 fs/buffer.c bio->bi_sector = bh->b_blocknr * (bh->b_size >> 9); bio 2949 fs/buffer.c bio->bi_bdev = bh->b_bdev; bio 2950 fs/buffer.c bio->bi_io_vec[0].bv_page = bh->b_page; bio 2951 fs/buffer.c bio->bi_io_vec[0].bv_len = bh->b_size; bio 2952 fs/buffer.c bio->bi_io_vec[0].bv_offset = bh_offset(bh); bio 2954 fs/buffer.c bio->bi_vcnt = 1; bio 2955 fs/buffer.c bio->bi_idx = 0; bio 2956 fs/buffer.c bio->bi_size = bh->b_size; bio 2958 fs/buffer.c bio->bi_end_io = end_bio_bh_io_sync; bio 2959 fs/buffer.c bio->bi_private = bh; bio 2961 fs/buffer.c bio_get(bio); bio 2962 fs/buffer.c submit_bio(rw, bio); bio 2964 fs/buffer.c if (bio_flagged(bio, BIO_EOPNOTSUPP)) bio 2967 fs/buffer.c bio_put(bio); bio 67 fs/direct-io.c struct bio *bio; /* bio under assembly */ bio 126 fs/direct-io.c struct bio *bio_list; /* singly linked via bi_private */ bio 263 fs/direct-io.c struct dio *dio = bio->bi_private; bio 268 fs/direct-io.c dio_bio_complete(dio, bio); bio 292 fs/direct-io.c struct dio *dio = bio->bi_private; bio 296 fs/direct-io.c bio->bi_private = dio->bio_list; bio 297 fs/direct-io.c dio->bio_list = bio; bio 307 fs/direct-io.c struct bio *bio; bio 309 fs/direct-io.c bio = bio_alloc(GFP_KERNEL, nr_vecs); bio 310 fs/direct-io.c if (bio == NULL) bio 313 fs/direct-io.c bio->bi_bdev = bdev; bio 314 fs/direct-io.c bio->bi_sector = first_sector; bio 316 fs/direct-io.c bio->bi_end_io = dio_bio_end_aio; bio 318 fs/direct-io.c bio->bi_end_io = dio_bio_end_io; bio 320 fs/direct-io.c dio->bio = bio; bio 333 fs/direct-io.c struct bio *bio = dio->bio; bio 336 fs/direct-io.c bio->bi_private = dio; bio 343 fs/direct-io.c bio_set_pages_dirty(bio); bio 345 fs/direct-io.c submit_bio(dio->rw, bio); bio 347 fs/direct-io.c dio->bio = NULL; bio 366 fs/direct-io.c static struct bio *dio_await_one(struct dio *dio) bio 369 fs/direct-io.c struct bio *bio = NULL; bio 389 fs/direct-io.c bio = dio->bio_list; bio 390 fs/direct-io.c dio->bio_list = bio->bi_private; bio 393 fs/direct-io.c return bio; bio 401 fs/direct-io.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); bio 402 fs/direct-io.c struct bio_vec *bvec = bio->bi_io_vec; bio 409 fs/direct-io.c bio_check_pages_dirty(bio); /* transfers ownership */ bio 411 fs/direct-io.c for (page_no = 0; page_no < bio->bi_vcnt; page_no++) { bio 418 fs/direct-io.c bio_put(bio); bio 432 fs/direct-io.c struct bio *bio; bio 434 fs/direct-io.c bio = dio_await_one(dio); bio 435 fs/direct-io.c if (bio) bio 436 fs/direct-io.c dio_bio_complete(dio, bio); bio 437 fs/direct-io.c } while (bio); bio 454 fs/direct-io.c struct bio *bio; bio 458 fs/direct-io.c bio = dio->bio_list; bio 459 fs/direct-io.c dio->bio_list = bio->bi_private; bio 461 fs/direct-io.c ret2 = dio_bio_complete(dio, bio); bio 572 fs/direct-io.c ret = bio_add_page(dio->bio, dio->cur_page, bio 604 fs/direct-io.c if (dio->bio) { bio 618 fs/direct-io.c if (dio->bio == NULL) { bio 1030 fs/direct-io.c if (dio->bio) bio 2295 fs/ext4/extents.c complete((struct completion *)bio->bi_private); bio 2302 fs/ext4/extents.c struct bio *bio; bio 2324 fs/ext4/extents.c bio = bio_alloc(GFP_NOIO, len); bio 2325 fs/ext4/extents.c if (!bio) bio 2327 fs/ext4/extents.c bio->bi_sector = ee_pblock; bio 2328 fs/ext4/extents.c bio->bi_bdev = inode->i_sb->s_bdev; bio 2333 fs/ext4/extents.c ret = bio_add_page(bio, ZERO_PAGE(0), bio 2349 fs/ext4/extents.c bio->bi_private = &event; bio 2350 fs/ext4/extents.c bio->bi_end_io = bi_complete; bio 2351 fs/ext4/extents.c submit_bio(WRITE, bio); bio 2354 fs/ext4/extents.c if (test_bit(BIO_UPTODATE, &bio->bi_flags)) bio 2360 fs/ext4/extents.c bio_put(bio); bio 219 fs/gfs2/ops_fstype.c struct page *page = bio->bi_private; bio 272 fs/gfs2/ops_fstype.c struct bio *bio; bio 282 fs/gfs2/ops_fstype.c bio = bio_alloc(GFP_NOFS, 1); bio 283 fs/gfs2/ops_fstype.c if (unlikely(!bio)) { bio 288 fs/gfs2/ops_fstype.c bio->bi_sector = sector * (sb->s_blocksize >> 9); bio 289 fs/gfs2/ops_fstype.c bio->bi_bdev = sb->s_bdev; bio 290 fs/gfs2/ops_fstype.c bio_add_page(bio, page, PAGE_SIZE, 0); bio 292 fs/gfs2/ops_fstype.c bio->bi_end_io = end_bio_io_page; bio 293 fs/gfs2/ops_fstype.c bio->bi_private = page; bio 294 fs/gfs2/ops_fstype.c submit_bio(READ_SYNC | (1 << BIO_RW_META), bio); bio 296 fs/gfs2/ops_fstype.c bio_put(bio); bio 1994 fs/jfs/jfs_logmgr.c struct bio *bio; bio 2005 fs/jfs/jfs_logmgr.c bio = bio_alloc(GFP_NOFS, 1); bio 2007 fs/jfs/jfs_logmgr.c bio->bi_sector = bp->l_blkno << (log->l2bsize - 9); bio 2008 fs/jfs/jfs_logmgr.c bio->bi_bdev = log->bdev; bio 2009 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_page = bp->l_page; bio 2010 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_len = LOGPSIZE; bio 2011 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_offset = bp->l_offset; bio 2013 fs/jfs/jfs_logmgr.c bio->bi_vcnt = 1; bio 2014 fs/jfs/jfs_logmgr.c bio->bi_idx = 0; bio 2015 fs/jfs/jfs_logmgr.c bio->bi_size = LOGPSIZE; bio 2017 fs/jfs/jfs_logmgr.c bio->bi_end_io = lbmIODone; bio 2018 fs/jfs/jfs_logmgr.c bio->bi_private = bp; bio 2019 fs/jfs/jfs_logmgr.c submit_bio(READ_SYNC, bio); bio 2142 fs/jfs/jfs_logmgr.c struct bio *bio; bio 2147 fs/jfs/jfs_logmgr.c bio = bio_alloc(GFP_NOFS, 1); bio 2148 fs/jfs/jfs_logmgr.c bio->bi_sector = bp->l_blkno << (log->l2bsize - 9); bio 2149 fs/jfs/jfs_logmgr.c bio->bi_bdev = log->bdev; bio 2150 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_page = bp->l_page; bio 2151 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_len = LOGPSIZE; bio 2152 fs/jfs/jfs_logmgr.c bio->bi_io_vec[0].bv_offset = bp->l_offset; bio 2154 fs/jfs/jfs_logmgr.c bio->bi_vcnt = 1; bio 2155 fs/jfs/jfs_logmgr.c bio->bi_idx = 0; bio 2156 fs/jfs/jfs_logmgr.c bio->bi_size = LOGPSIZE; bio 2158 fs/jfs/jfs_logmgr.c bio->bi_end_io = lbmIODone; bio 2159 fs/jfs/jfs_logmgr.c bio->bi_private = bp; bio 2163 fs/jfs/jfs_logmgr.c bio->bi_size = 0; bio 2164 fs/jfs/jfs_logmgr.c lbmIODone(bio, 0); bio 2166 fs/jfs/jfs_logmgr.c submit_bio(WRITE_SYNC, bio); bio 2204 fs/jfs/jfs_logmgr.c struct lbuf *bp = bio->bi_private; bio 2218 fs/jfs/jfs_logmgr.c if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) { bio 2224 fs/jfs/jfs_logmgr.c bio_put(bio); bio 287 fs/jfs/jfs_metapage.c struct page *page = bio->bi_private; bio 289 fs/jfs/jfs_metapage.c if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) { bio 295 fs/jfs/jfs_metapage.c bio_put(bio); bio 342 fs/jfs/jfs_metapage.c struct page *page = bio->bi_private; bio 346 fs/jfs/jfs_metapage.c if (! test_bit(BIO_UPTODATE, &bio->bi_flags)) { bio 351 fs/jfs/jfs_metapage.c bio_put(bio); bio 356 fs/jfs/jfs_metapage.c struct bio *bio = NULL; bio 399 fs/jfs/jfs_metapage.c if (bio) { bio 409 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, bio_bytes, bio_offset) < bio 417 fs/jfs/jfs_metapage.c if (!bio->bi_size) bio 419 fs/jfs/jfs_metapage.c submit_bio(WRITE, bio); bio 421 fs/jfs/jfs_metapage.c bio = NULL; bio 435 fs/jfs/jfs_metapage.c bio = bio_alloc(GFP_NOFS, 1); bio 436 fs/jfs/jfs_metapage.c bio->bi_bdev = inode->i_sb->s_bdev; bio 437 fs/jfs/jfs_metapage.c bio->bi_sector = pblock << (inode->i_blkbits - 9); bio 438 fs/jfs/jfs_metapage.c bio->bi_end_io = metapage_write_end_io; bio 439 fs/jfs/jfs_metapage.c bio->bi_private = page; bio 448 fs/jfs/jfs_metapage.c if (bio) { bio 449 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, bio_bytes, bio_offset) < bio_bytes) bio 451 fs/jfs/jfs_metapage.c if (!bio->bi_size) bio 454 fs/jfs/jfs_metapage.c submit_bio(WRITE, bio); bio 472 fs/jfs/jfs_metapage.c 4, bio, sizeof(*bio), 0); bio 474 fs/jfs/jfs_metapage.c bio_put(bio); bio 484 fs/jfs/jfs_metapage.c struct bio *bio = NULL; bio 506 fs/jfs/jfs_metapage.c if (bio) bio 507 fs/jfs/jfs_metapage.c submit_bio(READ, bio); bio 509 fs/jfs/jfs_metapage.c bio = bio_alloc(GFP_NOFS, 1); bio 510 fs/jfs/jfs_metapage.c bio->bi_bdev = inode->i_sb->s_bdev; bio 511 fs/jfs/jfs_metapage.c bio->bi_sector = pblock << (inode->i_blkbits - 9); bio 512 fs/jfs/jfs_metapage.c bio->bi_end_io = metapage_read_end_io; bio 513 fs/jfs/jfs_metapage.c bio->bi_private = page; bio 516 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, len, offset) < len) bio 522 fs/jfs/jfs_metapage.c if (bio) bio 523 fs/jfs/jfs_metapage.c submit_bio(READ, bio); bio 531 fs/jfs/jfs_metapage.c bio_put(bio); bio 44 fs/mpage.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); bio 45 fs/mpage.c struct bio_vec *bvec = bio->bi_io_vec + bio->bi_vcnt - 1; bio 50 fs/mpage.c if (--bvec >= bio->bi_io_vec) bio 60 fs/mpage.c } while (bvec >= bio->bi_io_vec); bio 61 fs/mpage.c bio_put(bio); bio 66 fs/mpage.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); bio 67 fs/mpage.c struct bio_vec *bvec = bio->bi_io_vec + bio->bi_vcnt - 1; bio 72 fs/mpage.c if (--bvec >= bio->bi_io_vec) bio 81 fs/mpage.c } while (bvec >= bio->bi_io_vec); bio 82 fs/mpage.c bio_put(bio); bio 85 fs/mpage.c struct bio *mpage_bio_submit(int rw, struct bio *bio) bio 87 fs/mpage.c bio->bi_end_io = mpage_end_io_read; bio 89 fs/mpage.c bio->bi_end_io = mpage_end_io_write; bio 90 fs/mpage.c submit_bio(rw, bio); bio 95 fs/mpage.c static struct bio * bio 100 fs/mpage.c struct bio *bio; bio 102 fs/mpage.c bio = bio_alloc(gfp_flags, nr_vecs); bio 104 fs/mpage.c if (bio == NULL && (current->flags & PF_MEMALLOC)) { bio 105 fs/mpage.c while (!bio && (nr_vecs /= 2)) bio 106 fs/mpage.c bio = bio_alloc(gfp_flags, nr_vecs); bio 109 fs/mpage.c if (bio) { bio 110 fs/mpage.c bio->bi_bdev = bdev; bio 111 fs/mpage.c bio->bi_sector = first_sector; bio 113 fs/mpage.c return bio; bio 168 fs/mpage.c static struct bio * bio 293 fs/mpage.c if (bio && (*last_block_in_bio != blocks[0] - 1)) bio 294 fs/mpage.c bio = mpage_bio_submit(READ, bio); bio 297 fs/mpage.c if (bio == NULL) { bio 298 fs/mpage.c bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9), bio 301 fs/mpage.c if (bio == NULL) bio 306 fs/mpage.c if (bio_add_page(bio, page, length, 0) < length) { bio 307 fs/mpage.c bio = mpage_bio_submit(READ, bio); bio 312 fs/mpage.c bio = mpage_bio_submit(READ, bio); bio 316 fs/mpage.c return bio; bio 319 fs/mpage.c if (bio) bio 320 fs/mpage.c bio = mpage_bio_submit(READ, bio); bio 375 fs/mpage.c struct bio *bio = NULL; bio 389 fs/mpage.c bio = do_mpage_readpage(bio, page, bio 398 fs/mpage.c if (bio) bio 399 fs/mpage.c mpage_bio_submit(READ, bio); bio 409 fs/mpage.c struct bio *bio = NULL; bio 415 fs/mpage.c bio = do_mpage_readpage(bio, page, 1, &last_block_in_bio, bio 417 fs/mpage.c if (bio) bio 418 fs/mpage.c mpage_bio_submit(READ, bio); bio 444 fs/mpage.c struct bio *bio = mpd->bio; bio 570 fs/mpage.c if (bio && mpd->last_block_in_bio != blocks[0] - 1) bio 571 fs/mpage.c bio = mpage_bio_submit(WRITE, bio); bio 574 fs/mpage.c if (bio == NULL) { bio 575 fs/mpage.c bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9), bio 577 fs/mpage.c if (bio == NULL) bio 587 fs/mpage.c if (bio_add_page(bio, page, length, 0) < length) { bio 588 fs/mpage.c bio = mpage_bio_submit(WRITE, bio); bio 621 fs/mpage.c bio = mpage_bio_submit(WRITE, bio); bio 632 fs/mpage.c if (bio) bio 633 fs/mpage.c bio = mpage_bio_submit(WRITE, bio); bio 646 fs/mpage.c mpd->bio = bio; bio 680 fs/mpage.c .bio = NULL, bio 687 fs/mpage.c if (mpd.bio) bio 688 fs/mpage.c mpage_bio_submit(WRITE, mpd.bio); bio 698 fs/mpage.c .bio = NULL, bio 704 fs/mpage.c if (mpd.bio) bio 705 fs/mpage.c mpage_bio_submit(WRITE, mpd.bio); bio 222 fs/ocfs2/cluster/heartbeat.c struct o2hb_bio_wait_ctxt *wc = bio->bi_private; bio 230 fs/ocfs2/cluster/heartbeat.c bio_put(bio); bio 235 fs/ocfs2/cluster/heartbeat.c static struct bio *o2hb_setup_one_bio(struct o2hb_region *reg, bio 245 fs/ocfs2/cluster/heartbeat.c struct bio *bio; bio 252 fs/ocfs2/cluster/heartbeat.c bio = bio_alloc(GFP_ATOMIC, 16); bio 253 fs/ocfs2/cluster/heartbeat.c if (!bio) { bio 255 fs/ocfs2/cluster/heartbeat.c bio = ERR_PTR(-ENOMEM); bio 260 fs/ocfs2/cluster/heartbeat.c bio->bi_sector = (reg->hr_start_block + cs) << (bits - 9); bio 261 fs/ocfs2/cluster/heartbeat.c bio->bi_bdev = reg->hr_bdev; bio 262 fs/ocfs2/cluster/heartbeat.c bio->bi_private = wc; bio 263 fs/ocfs2/cluster/heartbeat.c bio->bi_end_io = o2hb_bio_end_io; bio 276 fs/ocfs2/cluster/heartbeat.c len = bio_add_page(bio, page, vec_len, vec_start); bio 285 fs/ocfs2/cluster/heartbeat.c return bio; bio 294 fs/ocfs2/cluster/heartbeat.c struct bio *bio; bio 299 fs/ocfs2/cluster/heartbeat.c bio = o2hb_setup_one_bio(reg, &wc, ¤t_slot, max_slots); bio 300 fs/ocfs2/cluster/heartbeat.c if (IS_ERR(bio)) { bio 301 fs/ocfs2/cluster/heartbeat.c status = PTR_ERR(bio); bio 307 fs/ocfs2/cluster/heartbeat.c submit_bio(READ, bio); bio 325 fs/ocfs2/cluster/heartbeat.c struct bio *bio; bio 331 fs/ocfs2/cluster/heartbeat.c bio = o2hb_setup_one_bio(reg, write_wc, &slot, slot+1); bio 332 fs/ocfs2/cluster/heartbeat.c if (IS_ERR(bio)) { bio 333 fs/ocfs2/cluster/heartbeat.c status = PTR_ERR(bio); bio 339 fs/ocfs2/cluster/heartbeat.c submit_bio(WRITE, bio); bio 347 fs/xfs/linux-2.6/xfs_aops.c xfs_ioend_t *ioend = bio->bi_private; bio 349 fs/xfs/linux-2.6/xfs_aops.c ASSERT(atomic_read(&bio->bi_cnt) >= 1); bio 350 fs/xfs/linux-2.6/xfs_aops.c ioend->io_error = test_bit(BIO_UPTODATE, &bio->bi_flags) ? 0 : error; bio 353 fs/xfs/linux-2.6/xfs_aops.c bio->bi_private = NULL; bio 354 fs/xfs/linux-2.6/xfs_aops.c bio->bi_end_io = NULL; bio 355 fs/xfs/linux-2.6/xfs_aops.c bio_put(bio); bio 367 fs/xfs/linux-2.6/xfs_aops.c bio->bi_private = ioend; bio 368 fs/xfs/linux-2.6/xfs_aops.c bio->bi_end_io = xfs_end_bio; bio 370 fs/xfs/linux-2.6/xfs_aops.c submit_bio(WRITE, bio); bio 371 fs/xfs/linux-2.6/xfs_aops.c ASSERT(!bio_flagged(bio, BIO_EOPNOTSUPP)); bio 372 fs/xfs/linux-2.6/xfs_aops.c bio_put(bio); bio 375 fs/xfs/linux-2.6/xfs_aops.c STATIC struct bio * bio 379 fs/xfs/linux-2.6/xfs_aops.c struct bio *bio; bio 383 fs/xfs/linux-2.6/xfs_aops.c bio = bio_alloc(GFP_NOIO, nvecs); bio 385 fs/xfs/linux-2.6/xfs_aops.c } while (!bio); bio 387 fs/xfs/linux-2.6/xfs_aops.c ASSERT(bio->bi_private == NULL); bio 388 fs/xfs/linux-2.6/xfs_aops.c bio->bi_sector = bh->b_blocknr * (bh->b_size >> 9); bio 389 fs/xfs/linux-2.6/xfs_aops.c bio->bi_bdev = bh->b_bdev; bio 390 fs/xfs/linux-2.6/xfs_aops.c bio_get(bio); bio 391 fs/xfs/linux-2.6/xfs_aops.c return bio; bio 427 fs/xfs/linux-2.6/xfs_aops.c return bio_add_page(bio, bh->b_page, bh->b_size, bh_offset(bh)); bio 454 fs/xfs/linux-2.6/xfs_aops.c struct bio *bio; bio 469 fs/xfs/linux-2.6/xfs_aops.c bio = NULL; bio 473 fs/xfs/linux-2.6/xfs_aops.c if (!bio) { bio 475 fs/xfs/linux-2.6/xfs_aops.c bio = xfs_alloc_ioend_bio(bh); bio 477 fs/xfs/linux-2.6/xfs_aops.c xfs_submit_ioend_bio(ioend, bio); bio 481 fs/xfs/linux-2.6/xfs_aops.c if (bio_add_buffer(bio, bh) != bh->b_size) { bio 482 fs/xfs/linux-2.6/xfs_aops.c xfs_submit_ioend_bio(ioend, bio); bio 488 fs/xfs/linux-2.6/xfs_aops.c if (bio) bio 489 fs/xfs/linux-2.6/xfs_aops.c xfs_submit_ioend_bio(ioend, bio); bio 1113 fs/xfs/linux-2.6/xfs_buf.c xfs_buf_t *bp = (xfs_buf_t *)bio->bi_private; bio 1115 fs/xfs/linux-2.6/xfs_buf.c struct bio_vec *bvec = bio->bi_io_vec + bio->bi_vcnt - 1; bio 1117 fs/xfs/linux-2.6/xfs_buf.c if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) bio 1134 fs/xfs/linux-2.6/xfs_buf.c if (--bvec >= bio->bi_io_vec) bio 1139 fs/xfs/linux-2.6/xfs_buf.c } while (bvec >= bio->bi_io_vec); bio 1142 fs/xfs/linux-2.6/xfs_buf.c bio_put(bio); bio 1150 fs/xfs/linux-2.6/xfs_buf.c struct bio *bio; bio 1180 fs/xfs/linux-2.6/xfs_buf.c bio = bio_alloc(GFP_NOIO, 1); bio 1182 fs/xfs/linux-2.6/xfs_buf.c bio->bi_bdev = bp->b_target->bt_bdev; bio 1183 fs/xfs/linux-2.6/xfs_buf.c bio->bi_sector = sector - (offset >> BBSHIFT); bio 1184 fs/xfs/linux-2.6/xfs_buf.c bio->bi_end_io = xfs_buf_bio_end_io; bio 1185 fs/xfs/linux-2.6/xfs_buf.c bio->bi_private = bp; bio 1187 fs/xfs/linux-2.6/xfs_buf.c bio_add_page(bio, bp->b_pages[0], PAGE_CACHE_SIZE, 0); bio 1201 fs/xfs/linux-2.6/xfs_buf.c bio = bio_alloc(GFP_NOIO, nr_pages); bio 1202 fs/xfs/linux-2.6/xfs_buf.c bio->bi_bdev = bp->b_target->bt_bdev; bio 1203 fs/xfs/linux-2.6/xfs_buf.c bio->bi_sector = sector; bio 1204 fs/xfs/linux-2.6/xfs_buf.c bio->bi_end_io = xfs_buf_bio_end_io; bio 1205 fs/xfs/linux-2.6/xfs_buf.c bio->bi_private = bp; bio 1213 fs/xfs/linux-2.6/xfs_buf.c rbytes = bio_add_page(bio, bp->b_pages[map_i], nbytes, offset); bio 1224 fs/xfs/linux-2.6/xfs_buf.c if (likely(bio->bi_size)) { bio 1225 fs/xfs/linux-2.6/xfs_buf.c submit_bio(rw, bio); bio 1229 fs/xfs/linux-2.6/xfs_buf.c bio_put(bio); bio 53 include/linux/bio.h struct bio; bio 65 include/linux/bio.h struct bio *bi_next; /* request queue link */ bio 113 include/linux/bio.h #define bio_flagged(bio, flag) ((bio)->bi_flags & (1 << (flag))) bio 121 include/linux/bio.h #define BIO_POOL_IDX(bio) ((bio)->bi_flags >> BIO_POOL_OFFSET) bio 156 include/linux/bio.h #define bio_prio(bio) ((bio)->bi_rw >> BIO_PRIO_SHIFT) bio 157 include/linux/bio.h #define bio_prio_valid(bio) ioprio_valid(bio_prio(bio)) bio 161 include/linux/bio.h (bio)->bi_rw &= ((1UL << BIO_PRIO_SHIFT) - 1); \ bio 162 include/linux/bio.h (bio)->bi_rw |= ((unsigned long) (prio) << BIO_PRIO_SHIFT); \ bio 169 include/linux/bio.h #define bio_iovec_idx(bio, idx) (&((bio)->bi_io_vec[(idx)])) bio 170 include/linux/bio.h #define bio_iovec(bio) bio_iovec_idx((bio), (bio)->bi_idx) bio 171 include/linux/bio.h #define bio_page(bio) bio_iovec((bio))->bv_page bio 172 include/linux/bio.h #define bio_offset(bio) bio_iovec((bio))->bv_offset bio 173 include/linux/bio.h #define bio_segments(bio) ((bio)->bi_vcnt - (bio)->bi_idx) bio 174 include/linux/bio.h #define bio_sectors(bio) ((bio)->bi_size >> 9) bio 175 include/linux/bio.h #define bio_barrier(bio) ((bio)->bi_rw & (1 << BIO_RW_BARRIER)) bio 176 include/linux/bio.h #define bio_sync(bio) ((bio)->bi_rw & (1 << BIO_RW_SYNC)) bio 177 include/linux/bio.h #define bio_failfast(bio) ((bio)->bi_rw & (1 << BIO_RW_FAILFAST)) bio 178 include/linux/bio.h #define bio_rw_ahead(bio) ((bio)->bi_rw & (1 << BIO_RW_AHEAD)) bio 179 include/linux/bio.h #define bio_rw_meta(bio) ((bio)->bi_rw & (1 << BIO_RW_META)) bio 180 include/linux/bio.h #define bio_discard(bio) ((bio)->bi_rw & (1 << BIO_RW_DISCARD)) bio 181 include/linux/bio.h #define bio_empty_barrier(bio) (bio_barrier(bio) && !bio_has_data(bio) && !bio_discard(bio)) bio 185 include/linux/bio.h if (bio->bi_vcnt) bio 186 include/linux/bio.h return bio_iovec(bio)->bv_len >> 9; bio 188 include/linux/bio.h return bio->bi_size >> 9; bio 193 include/linux/bio.h if (bio->bi_vcnt) bio 194 include/linux/bio.h return page_address(bio_page(bio)) + bio_offset(bio); bio 202 include/linux/bio.h #define bio_to_phys(bio) (page_to_phys(bio_page((bio))) + (unsigned long) bio_offset((bio))) bio 212 include/linux/bio.h (kmap_atomic(bio_iovec_idx((bio), (idx))->bv_page, kmtype) + \ bio 213 include/linux/bio.h bio_iovec_idx((bio), (idx))->bv_offset) bio 221 include/linux/bio.h #define __BVEC_END(bio) bio_iovec_idx((bio), (bio)->bi_vcnt - 1) bio 222 include/linux/bio.h #define __BVEC_START(bio) bio_iovec_idx((bio), (bio)->bi_idx) bio 239 include/linux/bio.h #define bio_io_error(bio) bio_endio((bio), -EIO) bio 246 include/linux/bio.h for (bvl = bio_iovec_idx((bio), (start_idx)), i = (start_idx); \ bio 247 include/linux/bio.h i < (bio)->bi_vcnt; \ bio 251 include/linux/bio.h __bio_for_each_segment(bvl, bio, i, (bio)->bi_idx) bio 267 include/linux/bio.h #define bio_get(bio) atomic_inc(&(bio)->bi_cnt) bio 274 include/linux/bio.h struct bio *bip_bio; /* parent bio */ bio 304 include/linux/bio.h struct bio bio1, bio2; bio 319 include/linux/bio.h extern struct bio *bio_alloc(gfp_t, int); bio 320 include/linux/bio.h extern struct bio *bio_kmalloc(gfp_t, int); bio 321 include/linux/bio.h extern struct bio *bio_alloc_bioset(gfp_t, int, struct bio_set *); bio 330 include/linux/bio.h extern struct bio *bio_clone(struct bio *, gfp_t); bio 339 include/linux/bio.h extern struct bio *bio_map_user(struct request_queue *, struct block_device *, bio 343 include/linux/bio.h extern struct bio *bio_map_user_iov(struct request_queue *, bio 347 include/linux/bio.h extern struct bio *bio_map_kern(struct request_queue *, void *, unsigned int, bio 349 include/linux/bio.h extern struct bio *bio_copy_kern(struct request_queue *, void *, unsigned int, bio 353 include/linux/bio.h extern struct bio *bio_copy_user(struct request_queue *, struct rq_map_data *, bio 355 include/linux/bio.h extern struct bio *bio_copy_user_iov(struct request_queue *, bio 368 include/linux/bio.h bio->bi_comp_cpu = cpu; bio 441 include/linux/bio.h return bvec_kmap_irq(bio_iovec_idx(bio, idx), flags); bio 446 include/linux/bio.h __bio_kmap_irq((bio), (bio)->bi_idx, (flags)) bio 454 include/linux/bio.h return bio && bio->bi_io_vec != NULL; bio 470 include/linux/bio.h #define bio_integrity(bio) (bio->bi_integrity != NULL) bio 167 include/linux/blkdev.h struct bio *bio; bio 168 include/linux/blkdev.h struct bio *biotail; bio 678 include/linux/blkdev.h struct bio *bio; bio 683 include/linux/blkdev.h if ((rq->bio)) \ bio 684 include/linux/blkdev.h for (_bio = (rq)->bio; _bio; _bio = _bio->bi_next) bio 687 include/linux/blkdev.h __rq_for_each_bio(_iter.bio, _rq) \ bio 688 include/linux/blkdev.h bio_for_each_segment(bvl, _iter.bio, _iter.i) bio 691 include/linux/blkdev.h (_iter.bio->bi_next == NULL && _iter.i == _iter.bio->bi_vcnt-1) bio 1032 include/linux/blkdev.h if (rq->bio == NULL) bio 1035 include/linux/blkdev.h return bio_integrity(rq->bio); bio 235 include/linux/blktrace_api.h __blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw, what, !bio_flagged(bio, BIO_UPTODATE), 0, NULL); bio 257 include/linux/blktrace_api.h if (bio) bio 258 include/linux/blktrace_api.h blk_add_trace_bio(q, bio, what); bio 285 include/linux/blktrace_api.h if (bio) bio 286 include/linux/blktrace_api.h __blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw, what, !bio_flagged(bio, BIO_UPTODATE), sizeof(rpdu), &rpdu); bio 314 include/linux/blktrace_api.h r.device_from = cpu_to_be32(bio->bi_bdev->bd_dev); bio 317 include/linux/blktrace_api.h __blk_add_trace(bt, from, bio->bi_size, bio->bi_rw, BLK_TA_REMAP, !bio_flagged(bio, BIO_UPTODATE), sizeof(r), &r); bio 1787 include/linux/fs.h #define bio_rw(bio) ((bio)->bi_rw & (RW_MASK | RWA_MASK)) bio 1792 include/linux/fs.h #define bio_data_dir(bio) ((bio)->bi_rw & 1) bio 1927 include/linux/fs.h struct bio; bio 59 include/linux/loop.h struct bio *lo_bio; bio 60 include/linux/loop.h struct bio *lo_biotail; bio 15 include/linux/mpage.h struct bio *bio; bio 23 include/linux/mpage.h struct bio *mpage_bio_submit(int rw, struct bio *bio); bio 165 include/linux/pktcdvd.h struct bio *read_queue; bio 166 include/linux/pktcdvd.h struct bio *read_queue_tail; bio 167 include/linux/pktcdvd.h struct bio *write_queue; bio 168 include/linux/pktcdvd.h struct bio *write_queue_tail; bio 208 include/linux/pktcdvd.h struct bio *orig_bios; /* Original bios passed to pkt_make_request */ bio 209 include/linux/pktcdvd.h struct bio *orig_bios_tail;/* that will be handled by this packet */ bio 213 include/linux/pktcdvd.h struct bio *w_bio; /* The bio we will send to the real CD */ bio 228 include/linux/pktcdvd.h struct bio *r_bios[PACKET_MAX_SIZE]; /* bios to use during data gathering */ bio 241 include/linux/pktcdvd.h struct bio *bio; bio 246 include/linux/pktcdvd.h struct bio *bio; /* Original read request bio */ bio 230 include/linux/raid/md_k.h struct bio *biolist; /* bios that need to be retried bio 290 include/linux/raid/md_k.h int (*make_request)(struct request_queue *q, struct bio *bio); bio 37 include/linux/raid/multipath.h struct bio *master_bio; bio 38 include/linux/raid/multipath.h struct bio bio; bio 95 include/linux/raid/raid1.h struct bio *master_bio; bio 107 include/linux/raid/raid1.h struct bio *bios[0]; bio 116 include/linux/raid/raid1.h #define IO_BLOCKED ((struct bio*)1) bio 90 include/linux/raid/raid10.h struct bio *master_bio; bio 105 include/linux/raid/raid10.h struct bio *bio; bio 116 include/linux/raid/raid10.h #define IO_BLOCKED ((struct bio*)1) bio 219 include/linux/raid/raid5.h struct bio req; bio 222 include/linux/raid/raid5.h struct bio *toread, *read, *towrite, *written; bio 348 include/linux/raid/raid5.h struct bio *retry_read_aligned; /* currently retrying aligned bios */ bio 349 include/linux/raid/raid5.h struct bio *retry_read_aligned_list; /* aligned bios retry list */ bio 98 include/linux/sched.h struct bio; bio 1249 include/linux/sched.h struct bio *bio_list, **bio_tail; bio 16 include/linux/swap.h struct bio; bio 63 kernel/power/swap.c struct bio *bio; bio 65 kernel/power/swap.c bio = bio_alloc(__GFP_WAIT | __GFP_HIGH, 1); bio 66 kernel/power/swap.c if (!bio) bio 68 kernel/power/swap.c bio->bi_sector = page_off * (PAGE_SIZE >> 9); bio 69 kernel/power/swap.c bio->bi_bdev = resume_bdev; bio 70 kernel/power/swap.c bio->bi_end_io = end_swap_bio_read; bio 72 kernel/power/swap.c if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) { bio 75 kernel/power/swap.c bio_put(bio); bio 80 kernel/power/swap.c bio_get(bio); bio 83 kernel/power/swap.c submit_bio(rw | (1 << BIO_RW_SYNC), bio); bio 86 kernel/power/swap.c bio_set_pages_dirty(bio); bio 87 kernel/power/swap.c bio_put(bio); bio 91 kernel/power/swap.c bio->bi_private = *bio_chain; bio 92 kernel/power/swap.c *bio_chain = bio; bio 93 kernel/power/swap.c submit_bio(rw | (1 << BIO_RW_SYNC), bio); bio 110 kernel/power/swap.c struct bio *bio; bio 111 kernel/power/swap.c struct bio *next_bio; bio 117 kernel/power/swap.c bio = *bio_chain; bio 118 kernel/power/swap.c if (bio == NULL) bio 120 kernel/power/swap.c while (bio) { bio 123 kernel/power/swap.c next_bio = bio->bi_private; bio 124 kernel/power/swap.c page = bio->bi_io_vec[0].bv_page; bio 129 kernel/power/swap.c bio_put(bio); bio 130 kernel/power/swap.c bio = next_bio; bio 322 kernel/power/swap.c struct bio *bio; bio 332 kernel/power/swap.c bio = NULL; bio 338 kernel/power/swap.c &bio); bio 346 kernel/power/swap.c err2 = wait_on_bio_chain(&bio); bio 505 kernel/power/swap.c struct bio *bio; bio 515 kernel/power/swap.c bio = NULL; bio 521 kernel/power/swap.c error = swap_read_page(handle, data_of(*snapshot), &bio); bio 525 kernel/power/swap.c error = wait_on_bio_chain(&bio); bio 532 kernel/power/swap.c err2 = wait_on_bio_chain(&bio); bio 124 mm/bounce.c struct bio *bio_orig = bio->bi_private; bio 128 mm/bounce.c if (test_bit(BIO_EOPNOTSUPP, &bio->bi_flags)) bio 134 mm/bounce.c __bio_for_each_segment(bvec, bio, i, 0) { bio 144 mm/bounce.c bio_put(bio); bio 149 mm/bounce.c bounce_end_io(bio, page_pool, err); bio 155 mm/bounce.c bounce_end_io(bio, isa_page_pool, err); bio 160 mm/bounce.c struct bio *bio_orig = bio->bi_private; bio 162 mm/bounce.c if (test_bit(BIO_UPTODATE, &bio->bi_flags)) bio 163 mm/bounce.c copy_to_high_bio_irq(bio_orig, bio); bio 165 mm/bounce.c bounce_end_io(bio, pool, err); bio 170 mm/bounce.c __bounce_end_io_read(bio, page_pool, err); bio 175 mm/bounce.c __bounce_end_io_read(bio, isa_page_pool, err); bio 182 mm/bounce.c struct bio *bio = NULL; bio 198 mm/bounce.c if (!bio) bio 199 mm/bounce.c bio = bio_alloc(GFP_NOIO, (*bio_orig)->bi_vcnt); bio 201 mm/bounce.c to = bio->bi_io_vec + i; bio 222 mm/bounce.c if (!bio) bio 232 mm/bounce.c to = bio_iovec_idx(bio, i); bio 240 mm/bounce.c bio->bi_bdev = (*bio_orig)->bi_bdev; bio 241 mm/bounce.c bio->bi_flags |= (1 << BIO_BOUNCED); bio 242 mm/bounce.c bio->bi_sector = (*bio_orig)->bi_sector; bio 243 mm/bounce.c bio->bi_rw = (*bio_orig)->bi_rw; bio 245 mm/bounce.c bio->bi_vcnt = (*bio_orig)->bi_vcnt; bio 246 mm/bounce.c bio->bi_idx = (*bio_orig)->bi_idx; bio 247 mm/bounce.c bio->bi_size = (*bio_orig)->bi_size; bio 250 mm/bounce.c bio->bi_end_io = bounce_end_io_write; bio 252 mm/bounce.c bio->bi_end_io = bounce_end_io_read; bio 254 mm/bounce.c bio->bi_end_io = bounce_end_io_write_isa; bio 256 mm/bounce.c bio->bi_end_io = bounce_end_io_read_isa; bio 259 mm/bounce.c bio->bi_private = *bio_orig; bio 260 mm/bounce.c *bio_orig = bio; bio 22 mm/page_io.c static struct bio *get_swap_bio(gfp_t gfp_flags, pgoff_t index, bio 25 mm/page_io.c struct bio *bio; bio 27 mm/page_io.c bio = bio_alloc(gfp_flags, 1); bio 28 mm/page_io.c if (bio) { bio 33 mm/page_io.c bio->bi_sector = map_swap_page(sis, swp_offset(entry)) * bio 35 mm/page_io.c bio->bi_bdev = sis->bdev; bio 36 mm/page_io.c bio->bi_io_vec[0].bv_page = page; bio 37 mm/page_io.c bio->bi_io_vec[0].bv_len = PAGE_SIZE; bio 38 mm/page_io.c bio->bi_io_vec[0].bv_offset = 0; bio 39 mm/page_io.c bio->bi_vcnt = 1; bio 40 mm/page_io.c bio->bi_idx = 0; bio 41 mm/page_io.c bio->bi_size = PAGE_SIZE; bio 42 mm/page_io.c bio->bi_end_io = end_io; bio 44 mm/page_io.c return bio; bio 49 mm/page_io.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); bio 50 mm/page_io.c struct page *page = bio->bi_io_vec[0].bv_page; bio 64 mm/page_io.c imajor(bio->bi_bdev->bd_inode), bio 65 mm/page_io.c iminor(bio->bi_bdev->bd_inode), bio 66 mm/page_io.c (unsigned long long)bio->bi_sector); bio 70 mm/page_io.c bio_put(bio); bio 75 mm/page_io.c const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); bio 76 mm/page_io.c struct page *page = bio->bi_io_vec[0].bv_page; bio 82 mm/page_io.c imajor(bio->bi_bdev->bd_inode), bio 83 mm/page_io.c iminor(bio->bi_bdev->bd_inode), bio 84 mm/page_io.c (unsigned long long)bio->bi_sector); bio 89 mm/page_io.c bio_put(bio); bio 98 mm/page_io.c struct bio *bio; bio 105 mm/page_io.c bio = get_swap_bio(GFP_NOIO, page_private(page), page, bio 107 mm/page_io.c if (bio == NULL) { bio 118 mm/page_io.c submit_bio(rw, bio); bio 125 mm/page_io.c struct bio *bio; bio 130 mm/page_io.c bio = get_swap_bio(GFP_KERNEL, page_private(page), page, bio 132 mm/page_io.c if (bio == NULL) { bio 138 mm/page_io.c submit_bio(READ, bio);