bio              1249 block/as-iosched.c 	sector_t rb_key = bio->bi_sector + bio_sectors(bio);
bio              1255 block/as-iosched.c 	__rq = elv_rb_find(&ad->sort_list[bio_data_dir(bio)], rb_key);
bio              1256 block/as-iosched.c 	if (__rq && elv_rq_merge_ok(__rq, bio)) {
bio               168 block/blk-barrier.c 	if (bio_data_dir(q->orig_bar_rq->bio) == WRITE)
bio               172 block/blk-barrier.c 	init_request_from_bio(rq, q->orig_bar_rq->bio);
bio               258 block/blk-barrier.c 			set_bit(BIO_EOPNOTSUPP, &bio->bi_flags);
bio               259 block/blk-barrier.c 		clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio               262 block/blk-barrier.c 	complete(bio->bi_private);
bio               279 block/blk-barrier.c 	struct bio *bio;
bio               289 block/blk-barrier.c 	bio = bio_alloc(GFP_KERNEL, 0);
bio               290 block/blk-barrier.c 	if (!bio)
bio               293 block/blk-barrier.c 	bio->bi_end_io = bio_end_empty_barrier;
bio               294 block/blk-barrier.c 	bio->bi_private = &wait;
bio               295 block/blk-barrier.c 	bio->bi_bdev = bdev;
bio               296 block/blk-barrier.c 	submit_bio(WRITE_BARRIER, bio);
bio               306 block/blk-barrier.c 		*error_sector = bio->bi_sector;
bio               309 block/blk-barrier.c 	if (bio_flagged(bio, BIO_EOPNOTSUPP))
bio               311 block/blk-barrier.c 	else if (!bio_flagged(bio, BIO_UPTODATE))
bio               314 block/blk-barrier.c 	bio_put(bio);
bio               323 block/blk-barrier.c 			set_bit(BIO_EOPNOTSUPP, &bio->bi_flags);
bio               324 block/blk-barrier.c 		clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio               327 block/blk-barrier.c 	bio_put(bio);
bio               344 block/blk-barrier.c 	struct bio *bio;
bio               358 block/blk-barrier.c 		bio = bio_alloc(gfp_mask, 0);
bio               359 block/blk-barrier.c 		if (!bio)
bio               362 block/blk-barrier.c 		bio->bi_end_io = blkdev_discard_end_io;
bio               363 block/blk-barrier.c 		bio->bi_bdev = bdev;
bio               365 block/blk-barrier.c 		bio->bi_sector = sector;
bio               368 block/blk-barrier.c 			bio->bi_size = q->max_hw_sectors << 9;
bio               372 block/blk-barrier.c 			bio->bi_size = nr_sects << 9;
bio               375 block/blk-barrier.c 		bio_get(bio);
bio               376 block/blk-barrier.c 		submit_bio(DISCARD_BARRIER, bio);
bio               379 block/blk-barrier.c 		if (bio_flagged(bio, BIO_EOPNOTSUPP))
bio               381 block/blk-barrier.c 		else if (!bio_flagged(bio, BIO_UPTODATE))
bio               383 block/blk-barrier.c 		bio_put(bio);
bio               132 block/blk-core.c 			clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio               133 block/blk-core.c 		else if (!test_bit(BIO_UPTODATE, &bio->bi_flags))
bio               136 block/blk-core.c 		if (unlikely(nbytes > bio->bi_size)) {
bio               138 block/blk-core.c 			       __func__, nbytes, bio->bi_size);
bio               139 block/blk-core.c 			nbytes = bio->bi_size;
bio               142 block/blk-core.c 		bio->bi_size -= nbytes;
bio               143 block/blk-core.c 		bio->bi_sector += (nbytes >> 9);
bio               145 block/blk-core.c 		if (bio_integrity(bio))
bio               146 block/blk-core.c 			bio_integrity_advance(bio, nbytes);
bio               148 block/blk-core.c 		if (bio->bi_size == 0)
bio               149 block/blk-core.c 			bio_endio(bio, error);
bio               174 block/blk-core.c 						rq->bio, rq->biotail,
bio               811 block/blk-core.c 	blk_add_trace_generic(q, bio, rw, BLK_TA_GETRQ);
bio               828 block/blk-core.c 	rq = get_request(q, rw_flags, bio, GFP_NOIO);
bio               837 block/blk-core.c 		blk_add_trace_generic(q, bio, rw, BLK_TA_SLEEPRQ);
bio               855 block/blk-core.c 		rq = get_request(q, rw_flags, bio, GFP_NOIO);
bio              1072 block/blk-core.c 	req->cpu = bio->bi_comp_cpu;
bio              1078 block/blk-core.c 	if (bio_rw_ahead(bio) || bio_failfast(bio))
bio              1084 block/blk-core.c 	if (unlikely(bio_discard(bio))) {
bio              1086 block/blk-core.c 		if (bio_barrier(bio))
bio              1089 block/blk-core.c 	} else if (unlikely(bio_barrier(bio)))
bio              1092 block/blk-core.c 	if (bio_sync(bio))
bio              1094 block/blk-core.c 	if (bio_rw_meta(bio))
bio              1098 block/blk-core.c 	req->hard_sector = req->sector = bio->bi_sector;
bio              1099 block/blk-core.c 	req->ioprio = bio_prio(bio);
bio              1101 block/blk-core.c 	blk_rq_bio_prep(req->q, req, bio);
bio              1108 block/blk-core.c 	const unsigned short prio = bio_prio(bio);
bio              1109 block/blk-core.c 	const int sync = bio_sync(bio);
bio              1112 block/blk-core.c 	nr_sectors = bio_sectors(bio);
bio              1119 block/blk-core.c 	blk_queue_bounce(q, &bio);
bio              1121 block/blk-core.c 	barrier = bio_barrier(bio);
bio              1122 block/blk-core.c 	if (unlikely(barrier) && bio_has_data(bio) &&
bio              1128 block/blk-core.c 	discard = bio_discard(bio);
bio              1139 block/blk-core.c 	el_ret = elv_merge(q, &req, bio);
bio              1144 block/blk-core.c 		if (!ll_back_merge_fn(q, req, bio))
bio              1147 block/blk-core.c 		blk_add_trace_bio(q, bio, BLK_TA_BACKMERGE);
bio              1149 block/blk-core.c 		req->biotail->bi_next = bio;
bio              1150 block/blk-core.c 		req->biotail = bio;
bio              1154 block/blk-core.c 			req->cpu = bio->bi_comp_cpu;
bio              1163 block/blk-core.c 		if (!ll_front_merge_fn(q, req, bio))
bio              1166 block/blk-core.c 		blk_add_trace_bio(q, bio, BLK_TA_FRONTMERGE);
bio              1168 block/blk-core.c 		bio->bi_next = req->bio;
bio              1169 block/blk-core.c 		req->bio = bio;
bio              1176 block/blk-core.c 		req->buffer = bio_data(bio);
bio              1177 block/blk-core.c 		req->current_nr_sectors = bio_cur_sectors(bio);
bio              1179 block/blk-core.c 		req->sector = req->hard_sector = bio->bi_sector;
bio              1183 block/blk-core.c 			req->cpu = bio->bi_comp_cpu;
bio              1200 block/blk-core.c 	rw_flags = bio_data_dir(bio);
bio              1208 block/blk-core.c 	req = get_request_wait(q, rw_flags, bio);
bio              1216 block/blk-core.c 	init_request_from_bio(req, bio);
bio              1220 block/blk-core.c 	    bio_flagged(bio, BIO_CPU_AFFINE))
bio              1232 block/blk-core.c 	bio_endio(bio, err);
bio              1241 block/blk-core.c 	struct block_device *bdev = bio->bi_bdev;
bio              1243 block/blk-core.c 	if (bio_sectors(bio) && bdev != bdev->bd_contains) {
bio              1246 block/blk-core.c 		bio->bi_sector += p->start_sect;
bio              1247 block/blk-core.c 		bio->bi_bdev = bdev->bd_contains;
bio              1249 block/blk-core.c 		blk_add_trace_remap(bdev_get_queue(bio->bi_bdev), bio,
bio              1250 block/blk-core.c 				    bdev->bd_dev, bio->bi_sector,
bio              1251 block/blk-core.c 				    bio->bi_sector - p->start_sect);
bio              1261 block/blk-core.c 			bdevname(bio->bi_bdev, b),
bio              1262 block/blk-core.c 			bio->bi_rw,
bio              1263 block/blk-core.c 			(unsigned long long)bio->bi_sector + bio_sectors(bio),
bio              1264 block/blk-core.c 			(long long)(bio->bi_bdev->bd_inode->i_size >> 9));
bio              1266 block/blk-core.c 	set_bit(BIO_EOF, &bio->bi_flags);
bio              1281 block/blk-core.c 	struct hd_struct *part = bio->bi_bdev->bd_part;
bio              1284 block/blk-core.c 		return should_fail(&fail_make_request, bio->bi_size);
bio              1317 block/blk-core.c 	maxsector = bio->bi_bdev->bd_inode->i_size >> 9;
bio              1319 block/blk-core.c 		sector_t sector = bio->bi_sector;
bio              1327 block/blk-core.c 			handle_bad_sector(bio);
bio              1363 block/blk-core.c 	int ret, nr_sectors = bio_sectors(bio);
bio              1369 block/blk-core.c 	if (bio_check_eod(bio, nr_sectors))
bio              1385 block/blk-core.c 		q = bdev_get_queue(bio->bi_bdev);
bio              1390 block/blk-core.c 				bdevname(bio->bi_bdev, b),
bio              1391 block/blk-core.c 				(long long) bio->bi_sector);
bio              1393 block/blk-core.c 			bio_endio(bio, err);
bio              1399 block/blk-core.c 				bdevname(bio->bi_bdev, b),
bio              1400 block/blk-core.c 				bio_sectors(bio),
bio              1408 block/blk-core.c 		if (should_fail_request(bio))
bio              1415 block/blk-core.c 		blk_partition_remap(bio);
bio              1417 block/blk-core.c 		if (bio_integrity_enabled(bio) && bio_integrity_prep(bio))
bio              1421 block/blk-core.c 			blk_add_trace_remap(q, bio, old_dev, bio->bi_sector,
bio              1424 block/blk-core.c 		blk_add_trace_bio(q, bio, BLK_TA_QUEUE);
bio              1426 block/blk-core.c 		old_sector = bio->bi_sector;
bio              1427 block/blk-core.c 		old_dev = bio->bi_bdev->bd_dev;
bio              1429 block/blk-core.c 		if (bio_check_eod(bio, nr_sectors))
bio              1431 block/blk-core.c 		if ((bio_empty_barrier(bio) && !q->prepare_flush_fn) ||
bio              1432 block/blk-core.c 		    (bio_discard(bio) && !q->prepare_discard_fn)) {
bio              1437 block/blk-core.c 		ret = q->make_request_fn(q, bio);
bio              1456 block/blk-core.c 		*(current->bio_tail) = bio;
bio              1457 block/blk-core.c 		bio->bi_next = NULL;
bio              1458 block/blk-core.c 		current->bio_tail = &bio->bi_next;
bio              1479 block/blk-core.c 	BUG_ON(bio->bi_next);
bio              1481 block/blk-core.c 		current->bio_list = bio->bi_next;
bio              1482 block/blk-core.c 		if (bio->bi_next == NULL)
bio              1485 block/blk-core.c 			bio->bi_next = NULL;
bio              1486 block/blk-core.c 		__generic_make_request(bio);
bio              1487 block/blk-core.c 		bio = current->bio_list;
bio              1488 block/blk-core.c 	} while (bio);
bio              1505 block/blk-core.c 	int count = bio_sectors(bio);
bio              1507 block/blk-core.c 	bio->bi_rw |= rw;
bio              1513 block/blk-core.c 	if (bio_has_data(bio)) {
bio              1517 block/blk-core.c 			task_io_account_read(bio->bi_size);
bio              1526 block/blk-core.c 				(unsigned long long)bio->bi_sector,
bio              1527 block/blk-core.c 				bdevname(bio->bi_bdev, b));
bio              1531 block/blk-core.c 	generic_make_request(bio);
bio              1634 block/blk-core.c 	struct bio *bio;
bio              1663 block/blk-core.c 	while ((bio = req->bio) != NULL) {
bio              1672 block/blk-core.c 			bio->bi_sector = req->sector;
bio              1674 block/blk-core.c 		if (nr_bytes >= bio->bi_size) {
bio              1675 block/blk-core.c 			req->bio = bio->bi_next;
bio              1676 block/blk-core.c 			nbytes = bio->bi_size;
bio              1677 block/blk-core.c 			req_bio_endio(req, bio, nbytes, error);
bio              1681 block/blk-core.c 			int idx = bio->bi_idx + next_idx;
bio              1683 block/blk-core.c 			if (unlikely(bio->bi_idx >= bio->bi_vcnt)) {
bio              1686 block/blk-core.c 				       __func__, bio->bi_idx, bio->bi_vcnt);
bio              1690 block/blk-core.c 			nbytes = bio_iovec_idx(bio, idx)->bv_len;
bio              1691 block/blk-core.c 			BIO_BUG_ON(nbytes > bio->bi_size);
bio              1712 block/blk-core.c 		bio = req->bio;
bio              1713 block/blk-core.c 		if (bio) {
bio              1725 block/blk-core.c 	if (!req->bio)
bio              1732 block/blk-core.c 		req_bio_endio(req, bio, bio_nbytes, error);
bio              1733 block/blk-core.c 		bio->bi_idx += next_idx;
bio              1734 block/blk-core.c 		bio_iovec(bio)->bv_offset += nr_bytes;
bio              1735 block/blk-core.c 		bio_iovec(bio)->bv_len -= nr_bytes;
bio              1815 block/blk-core.c 	if (rq->bio)
bio              1816 block/blk-core.c 		return rq->bio->bi_size;
bio              1852 block/blk-core.c 	if (rq->bio) {
bio              1942 block/blk-core.c 	if (rq->bio && __end_that_request_first(rq, error, nr_bytes))
bio              2042 block/blk-core.c 	rq->cmd_flags |= (bio->bi_rw & 3);
bio              2044 block/blk-core.c 	if (bio_has_data(bio)) {
bio              2045 block/blk-core.c 		rq->nr_phys_segments = bio_phys_segments(q, bio);
bio              2046 block/blk-core.c 		rq->buffer = bio_data(bio);
bio              2048 block/blk-core.c 	rq->current_nr_sectors = bio_cur_sectors(bio);
bio              2050 block/blk-core.c 	rq->hard_nr_sectors = rq->nr_sectors = bio_sectors(bio);
bio              2051 block/blk-core.c 	rq->data_len = bio->bi_size;
bio              2053 block/blk-core.c 	rq->bio = rq->biotail = bio;
bio              2055 block/blk-core.c 	if (bio->bi_bdev)
bio              2056 block/blk-core.c 		rq->rq_disk = bio->bi_bdev->bd_disk;
bio                15 block/blk-map.c 	if (!rq->bio)
bio                16 block/blk-map.c 		blk_rq_bio_prep(q, rq, bio);
bio                17 block/blk-map.c 	else if (!ll_back_merge_fn(q, rq, bio))
bio                20 block/blk-map.c 		rq->biotail->bi_next = bio;
bio                21 block/blk-map.c 		rq->biotail = bio;
bio                23 block/blk-map.c 		rq->data_len += bio->bi_size;
bio                33 block/blk-map.c 	if (bio) {
bio                34 block/blk-map.c 		if (bio_flagged(bio, BIO_USER_MAPPED))
bio                35 block/blk-map.c 			bio_unmap_user(bio);
bio                37 block/blk-map.c 			ret = bio_uncopy_user(bio);
bio                48 block/blk-map.c 	struct bio *bio, *orig_bio;
bio                59 block/blk-map.c 		bio = bio_map_user(q, NULL, uaddr, len, reading, gfp_mask);
bio                61 block/blk-map.c 		bio = bio_copy_user(q, map_data, uaddr, len, reading, gfp_mask);
bio                63 block/blk-map.c 	if (IS_ERR(bio))
bio                64 block/blk-map.c 		return PTR_ERR(bio);
bio                67 block/blk-map.c 		bio->bi_flags |= (1 << BIO_NULL_MAPPED);
bio                69 block/blk-map.c 	orig_bio = bio;
bio                70 block/blk-map.c 	blk_queue_bounce(q, &bio);
bio                76 block/blk-map.c 	bio_get(bio);
bio                78 block/blk-map.c 	ret = blk_rq_append_bio(q, rq, bio);
bio                80 block/blk-map.c 		return bio->bi_size;
bio                83 block/blk-map.c 	bio_endio(bio, 0);
bio                85 block/blk-map.c 	bio_put(bio);
bio               116 block/blk-map.c 	struct bio *bio = NULL;
bio               149 block/blk-map.c 		if (!bio)
bio               150 block/blk-map.c 			bio = rq->bio;
bio               155 block/blk-map.c 	if (!bio_flagged(bio, BIO_USER_MAPPED))
bio               161 block/blk-map.c 	blk_rq_unmap_user(bio);
bio               162 block/blk-map.c 	rq->bio = NULL;
bio               194 block/blk-map.c 	struct bio *bio;
bio               211 block/blk-map.c 		bio = bio_copy_user_iov(q, map_data, iov, iov_count, read,
bio               214 block/blk-map.c 		bio = bio_map_user_iov(q, NULL, iov, iov_count, read, gfp_mask);
bio               216 block/blk-map.c 	if (IS_ERR(bio))
bio               217 block/blk-map.c 		return PTR_ERR(bio);
bio               219 block/blk-map.c 	if (bio->bi_size != len) {
bio               220 block/blk-map.c 		bio_endio(bio, 0);
bio               221 block/blk-map.c 		bio_unmap_user(bio);
bio               225 block/blk-map.c 	if (!bio_flagged(bio, BIO_USER_MAPPED))
bio               228 block/blk-map.c 	blk_queue_bounce(q, &bio);
bio               229 block/blk-map.c 	bio_get(bio);
bio               230 block/blk-map.c 	blk_rq_bio_prep(q, rq, bio);
bio               247 block/blk-map.c 	struct bio *mapped_bio;
bio               250 block/blk-map.c 	while (bio) {
bio               251 block/blk-map.c 		mapped_bio = bio;
bio               252 block/blk-map.c 		if (unlikely(bio_flagged(bio, BIO_BOUNCED)))
bio               253 block/blk-map.c 			mapped_bio = bio->bi_private;
bio               259 block/blk-map.c 		mapped_bio = bio;
bio               260 block/blk-map.c 		bio = bio->bi_next;
bio               285 block/blk-map.c 	struct bio *bio;
bio               294 block/blk-map.c 		bio = bio_copy_kern(q, kbuf, len, gfp_mask, reading);
bio               296 block/blk-map.c 		bio = bio_map_kern(q, kbuf, len, gfp_mask);
bio               298 block/blk-map.c 	if (IS_ERR(bio))
bio               299 block/blk-map.c 		return PTR_ERR(bio);
bio               302 block/blk-map.c 		bio->bi_rw |= (1 << BIO_RW);
bio               307 block/blk-map.c 	blk_rq_bio_prep(q, rq, bio);
bio               308 block/blk-map.c 	blk_queue_bounce(q, &rq->bio);
bio                25 block/blk-merge.c 			rq->hard_cur_sectors = bio_cur_sectors(rq->bio);
bio                27 block/blk-merge.c 			rq->buffer = bio_data(rq->bio);
bio                52 block/blk-merge.c 	if (!rq->bio)
bio                92 block/blk-merge.c 	struct bio *nxt = bio->bi_next;
bio                94 block/blk-merge.c 	rq.bio = rq.biotail = bio;
bio                95 block/blk-merge.c 	bio->bi_next = NULL;
bio                97 block/blk-merge.c 	bio->bi_next = nxt;
bio                98 block/blk-merge.c 	bio->bi_phys_segments = rq.nr_phys_segments;
bio                99 block/blk-merge.c 	bio->bi_flags |= (1 << BIO_SEG_VALID);
bio               109 block/blk-merge.c 	if (bio->bi_size + nxt->bi_size > q->max_segment_size)
bio               112 block/blk-merge.c 	if (!bio_has_data(bio))
bio               115 block/blk-merge.c 	if (!BIOVEC_PHYS_MERGEABLE(__BVEC_END(bio), __BVEC_START(nxt)))
bio               122 block/blk-merge.c 	if (BIO_SEG_BOUNDARY(q, bio, nxt))
bio               220 block/blk-merge.c 	int nr_phys_segs = bio_phys_segments(q, bio);
bio               241 block/blk-merge.c 	int nr_phys_segs = bio_phys_segments(q, bio);
bio               269 block/blk-merge.c 	if (req->nr_sectors + bio_sectors(bio) > max_sectors) {
bio               277 block/blk-merge.c 	if (!bio_flagged(bio, BIO_SEG_VALID))
bio               278 block/blk-merge.c 		blk_recount_segments(q, bio);
bio               280 block/blk-merge.c 	return ll_new_hw_segment(q, req, bio);
bio               294 block/blk-merge.c 	if (req->nr_sectors + bio_sectors(bio) > max_sectors) {
bio               300 block/blk-merge.c 	if (!bio_flagged(bio, BIO_SEG_VALID))
bio               301 block/blk-merge.c 		blk_recount_segments(q, bio);
bio               302 block/blk-merge.c 	if (!bio_flagged(req->bio, BIO_SEG_VALID))
bio               303 block/blk-merge.c 		blk_recount_segments(q, req->bio);
bio               305 block/blk-merge.c 	return ll_new_hw_segment(q, req, bio);
bio               327 block/blk-merge.c 	if (blk_phys_contig_segment(q, req->biotail, next->bio))
bio               382 block/blk-merge.c 	req->biotail->bi_next = next->bio;
bio                93 block/blk.h    	__rq_for_each_bio(_iter.bio, _rq)			\
bio                94 block/blk.h    		bip_for_each_vec(bvl, _iter.bio->bi_integrity, _iter.i)
bio                83 block/bsg.c    	struct bio *bio;
bio                84 block/bsg.c    	struct bio *bidi_bio;
bio               313 block/bsg.c    		blk_rq_unmap_user(next_rq->bio);
bio               330 block/bsg.c    		bd->name, rq, bc, bc->bio, uptodate);
bio               356 block/bsg.c    	bc->bio = rq->bio;
bio               358 block/bsg.c    		bc->bidi_bio = rq->next_rq->bio;
bio               420 block/bsg.c    	dprintk("rq %p bio %p %u\n", rq, bio, rq->errors);
bio               463 block/bsg.c    	blk_rq_unmap_user(bio);
bio               509 block/bsg.c    		tret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio,
bio               544 block/bsg.c    		ret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio,
bio               920 block/bsg.c    		struct bio *bio, *bidi_bio = NULL;
bio               930 block/bsg.c    		bio = rq->bio;
bio               932 block/bsg.c    			bidi_bio = rq->next_rq->bio;
bio               934 block/bsg.c    		ret = blk_complete_sgv4_hdr_rq(rq, &hdr, bio, bidi_bio);
bio               241 block/cfq-iosched.c 	if (bio_data_dir(bio) == READ || bio_sync(bio))
bio               647 block/cfq-iosched.c 	cfqq = cic_to_cfqq(cic, cfq_bio_sync(bio));
bio               649 block/cfq-iosched.c 		sector_t sector = bio->bi_sector + bio_sectors(bio);
bio               701 block/cfq-iosched.c 	__rq = cfq_find_rq_fmerge(cfqd, bio);
bio               702 block/cfq-iosched.c 	if (__rq && elv_rq_merge_ok(__rq, bio)) {
bio               744 block/cfq-iosched.c 	if (cfq_bio_sync(bio) && !rq_is_sync(rq))
bio               755 block/cfq-iosched.c 	cfqq = cic_to_cfqq(cic, cfq_bio_sync(bio));
bio               137 block/deadline-iosched.c 		sector_t sector = bio->bi_sector + bio_sectors(bio);
bio               139 block/deadline-iosched.c 		__rq = elv_rb_find(&dd->sort_list[bio_data_dir(bio)], sector);
bio               143 block/deadline-iosched.c 			if (elv_rq_merge_ok(__rq, bio)) {
bio                65 block/elevator.c 		return e->ops->elevator_allow_merge_fn(q, rq, bio);
bio                81 block/elevator.c 	if (bio_discard(bio) != bio_discard(rq->bio))
bio                87 block/elevator.c 	if (bio_data_dir(bio) != rq_data_dir(rq))
bio                93 block/elevator.c 	if (rq->rq_disk != bio->bi_bdev->bd_disk || rq->special)
bio                99 block/elevator.c 	if (bio_integrity(bio) != blk_integrity_rq(rq))
bio               102 block/elevator.c 	if (!elv_iosched_allow_merge(rq, bio))
bio               116 block/elevator.c 	if (elv_rq_merge_ok(__rq, bio)) {
bio               117 block/elevator.c 		if (__rq->sector + __rq->nr_sectors == bio->bi_sector)
bio               119 block/elevator.c 		else if (__rq->sector - bio_sectors(bio) == bio->bi_sector)
bio               499 block/elevator.c 		ret = elv_try_merge(q->last_merge, bio);
bio               512 block/elevator.c 	__rq = elv_rqhash_find(q, bio->bi_sector);
bio               513 block/elevator.c 	if (__rq && elv_rq_merge_ok(__rq, bio)) {
bio               519 block/elevator.c 		return e->ops->elevator_merge_fn(q, req, bio);
bio               120 block/ioctl.c  			set_bit(BIO_EOPNOTSUPP, &bio->bi_flags);
bio               121 block/ioctl.c  		clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio               123 block/ioctl.c  	complete(bio->bi_private);
bio               147 block/ioctl.c  		struct bio *bio;
bio               149 block/ioctl.c  		bio = bio_alloc(GFP_KERNEL, 0);
bio               150 block/ioctl.c  		if (!bio)
bio               153 block/ioctl.c  		bio->bi_end_io = blk_ioc_discard_endio;
bio               154 block/ioctl.c  		bio->bi_bdev = bdev;
bio               155 block/ioctl.c  		bio->bi_private = &wait;
bio               156 block/ioctl.c  		bio->bi_sector = start;
bio               159 block/ioctl.c  			bio->bi_size = q->max_hw_sectors << 9;
bio               163 block/ioctl.c  			bio->bi_size = len << 9;
bio               166 block/ioctl.c  		submit_bio(DISCARD_NOBARRIER, bio);
bio               170 block/ioctl.c  		if (bio_flagged(bio, BIO_EOPNOTSUPP))
bio               172 block/ioctl.c  		else if (!bio_flagged(bio, BIO_UPTODATE))
bio               174 block/ioctl.c  		bio_put(bio);
bio               222 block/scsi_ioctl.c 	blk_rq_unmap_user(rq->bio);
bio               255 block/scsi_ioctl.c 	rq->bio = bio;
bio               270 block/scsi_ioctl.c 	struct bio *bio;
bio               327 block/scsi_ioctl.c 	bio = rq->bio;
bio               343 block/scsi_ioctl.c 	return blk_complete_sghdr_rq(rq, hdr, bio);
bio                51 fs/bio-integrity.c 	BUG_ON(bio == NULL);
bio                70 fs/bio-integrity.c 	bip->bip_bio = bio;
bio                71 fs/bio-integrity.c 	bio->bi_integrity = bip;
bio                91 fs/bio-integrity.c 	return bio_integrity_alloc_bioset(bio, gfp_mask, nr_vecs, fs_bio_set);
bio               105 fs/bio-integrity.c 	struct bio_integrity_payload *bip = bio->bi_integrity;
bio               110 fs/bio-integrity.c 	if (!bio_flagged(bio, BIO_CLONED) && !bio_flagged(bio, BIO_FS_INTEGRITY)
bio               117 fs/bio-integrity.c 	bio->bi_integrity = NULL;
bio               133 fs/bio-integrity.c 	struct bio_integrity_payload *bip = bio->bi_integrity;
bio               184 fs/bio-integrity.c 	if (bio_integrity(bio))
bio               187 fs/bio-integrity.c 	return bdev_integrity_enabled(bio->bi_bdev, bio_data_dir(bio));
bio               221 fs/bio-integrity.c 	struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio               223 fs/bio-integrity.c 	BUG_ON(bio->bi_size == 0);
bio               225 fs/bio-integrity.c 	return bi->tag_size * (bio->bi_size / bi->sector_size);
bio               231 fs/bio-integrity.c 	struct bio_integrity_payload *bip = bio->bi_integrity;
bio               232 fs/bio-integrity.c 	struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio               270 fs/bio-integrity.c 	BUG_ON(bio_data_dir(bio) != WRITE);
bio               272 fs/bio-integrity.c 	return bio_integrity_tag(bio, tag_buf, len, 1);
bio               288 fs/bio-integrity.c 	BUG_ON(bio_data_dir(bio) != READ);
bio               290 fs/bio-integrity.c 	return bio_integrity_tag(bio, tag_buf, len, 0);
bio               305 fs/bio-integrity.c 	struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio               308 fs/bio-integrity.c 	sector_t sector = bio->bi_sector;
bio               310 fs/bio-integrity.c 	void *prot_buf = bio->bi_integrity->bip_buf;
bio               313 fs/bio-integrity.c 	bix.disk_name = bio->bi_bdev->bd_disk->disk_name;
bio               316 fs/bio-integrity.c 	bio_for_each_segment(bv, bio, i) {
bio               329 fs/bio-integrity.c 		BUG_ON(total > bio->bi_integrity->bip_size);
bio               365 fs/bio-integrity.c 	bi = bdev_get_integrity(bio->bi_bdev);
bio               366 fs/bio-integrity.c 	q = bdev_get_queue(bio->bi_bdev);
bio               368 fs/bio-integrity.c 	BUG_ON(bio_integrity(bio));
bio               370 fs/bio-integrity.c 	sectors = bio_integrity_hw_sectors(bi, bio_sectors(bio));
bio               385 fs/bio-integrity.c 	bip = bio_integrity_alloc(bio, GFP_NOIO, nr_pages);
bio               394 fs/bio-integrity.c 	bip->bip_sector = bio->bi_sector;
bio               408 fs/bio-integrity.c 		ret = bio_integrity_add_page(bio, virt_to_page(buf),
bio               423 fs/bio-integrity.c 	if (bio_data_dir(bio) == READ) {
bio               424 fs/bio-integrity.c 		bip->bip_end_io = bio->bi_end_io;
bio               425 fs/bio-integrity.c 		bio->bi_end_io = bio_integrity_endio;
bio               429 fs/bio-integrity.c 	if (bio_data_dir(bio) == WRITE)
bio               430 fs/bio-integrity.c 		bio_integrity_generate(bio);
bio               446 fs/bio-integrity.c 	struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio               449 fs/bio-integrity.c 	sector_t sector = bio->bi_integrity->bip_sector;
bio               451 fs/bio-integrity.c 	void *prot_buf = bio->bi_integrity->bip_buf;
bio               454 fs/bio-integrity.c 	bix.disk_name = bio->bi_bdev->bd_disk->disk_name;
bio               457 fs/bio-integrity.c 	bio_for_each_segment(bv, bio, i) {
bio               475 fs/bio-integrity.c 		BUG_ON(total > bio->bi_integrity->bip_size);
bio               495 fs/bio-integrity.c 	struct bio *bio = bip->bip_bio;
bio               498 fs/bio-integrity.c 	if (bio_integrity_verify(bio)) {
bio               499 fs/bio-integrity.c 		clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio               504 fs/bio-integrity.c 	bio->bi_end_io = bip->bip_end_io;
bio               506 fs/bio-integrity.c 	if (bio->bi_end_io)
bio               507 fs/bio-integrity.c 		bio->bi_end_io(bio, error);
bio               524 fs/bio-integrity.c 	struct bio_integrity_payload *bip = bio->bi_integrity;
bio               526 fs/bio-integrity.c 	BUG_ON(bip->bip_bio != bio);
bio               595 fs/bio-integrity.c 	struct bio_integrity_payload *bip = bio->bi_integrity;
bio               596 fs/bio-integrity.c 	struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio               621 fs/bio-integrity.c 	struct bio_integrity_payload *bip = bio->bi_integrity;
bio               622 fs/bio-integrity.c 	struct blk_integrity *bi = bdev_get_integrity(bio->bi_bdev);
bio               627 fs/bio-integrity.c 	BUG_ON(!bio_flagged(bio, BIO_CLONED));
bio               647 fs/bio-integrity.c 	struct bio_integrity_payload *bip = bio->bi_integrity;
bio               650 fs/bio-integrity.c 	if (bio_integrity(bio) == 0)
bio               653 fs/bio-integrity.c 	bi = bdev_get_integrity(bio->bi_bdev);
bio               672 fs/bio-integrity.c 	bp->bip1.bip_sector = bio->bi_integrity->bip_sector;
bio               673 fs/bio-integrity.c 	bp->bip2.bip_sector = bio->bi_integrity->bip_sector + nr_sectors;
bio               696 fs/bio-integrity.c 	bip = bio_integrity_alloc_bioset(bio, GFP_NOIO, bip_src->bip_vcnt, bs);
bio               109 fs/bio.c       	if (bio->bi_io_vec) {
bio               110 fs/bio.c       		const int pool_idx = BIO_POOL_IDX(bio);
bio               114 fs/bio.c       		mempool_free(bio->bi_io_vec, bio_set->bvec_pools[pool_idx]);
bio               117 fs/bio.c       	if (bio_integrity(bio))
bio               118 fs/bio.c       		bio_integrity_free(bio, bio_set);
bio               120 fs/bio.c       	mempool_free(bio, bio_set->bio_pool);
bio               128 fs/bio.c       	bio_free(bio, fs_bio_set);
bio               133 fs/bio.c       	kfree(bio->bi_io_vec);
bio               134 fs/bio.c       	kfree(bio);
bio               139 fs/bio.c       	memset(bio, 0, sizeof(*bio));
bio               140 fs/bio.c       	bio->bi_flags = 1 << BIO_UPTODATE;
bio               141 fs/bio.c       	bio->bi_comp_cpu = -1;
bio               142 fs/bio.c       	atomic_set(&bio->bi_cnt, 1);
bio               160 fs/bio.c       struct bio *bio_alloc_bioset(gfp_t gfp_mask, int nr_iovecs, struct bio_set *bs)
bio               162 fs/bio.c       	struct bio *bio;
bio               165 fs/bio.c       		bio = mempool_alloc(bs->bio_pool, gfp_mask);
bio               167 fs/bio.c       		bio = kmalloc(sizeof(*bio), gfp_mask);
bio               169 fs/bio.c       	if (likely(bio)) {
bio               172 fs/bio.c       		bio_init(bio);
bio               179 fs/bio.c       					mempool_free(bio, bs->bio_pool);
bio               181 fs/bio.c       					kfree(bio);
bio               182 fs/bio.c       				bio = NULL;
bio               185 fs/bio.c       			bio->bi_flags |= idx << BIO_POOL_OFFSET;
bio               186 fs/bio.c       			bio->bi_max_vecs = bvec_nr_vecs(idx);
bio               188 fs/bio.c       		bio->bi_io_vec = bvl;
bio               191 fs/bio.c       	return bio;
bio               194 fs/bio.c       struct bio *bio_alloc(gfp_t gfp_mask, int nr_iovecs)
bio               196 fs/bio.c       	struct bio *bio = bio_alloc_bioset(gfp_mask, nr_iovecs, fs_bio_set);
bio               198 fs/bio.c       	if (bio)
bio               199 fs/bio.c       		bio->bi_destructor = bio_fs_destructor;
bio               201 fs/bio.c       	return bio;
bio               211 fs/bio.c       struct bio *bio_kmalloc(gfp_t gfp_mask, int nr_iovecs)
bio               213 fs/bio.c       	struct bio *bio = bio_alloc_bioset(gfp_mask, nr_iovecs, NULL);
bio               215 fs/bio.c       	if (bio)
bio               216 fs/bio.c       		bio->bi_destructor = bio_kmalloc_destructor;
bio               218 fs/bio.c       	return bio;
bio               227 fs/bio.c       	bio_for_each_segment(bv, bio, i) {
bio               246 fs/bio.c       	BIO_BUG_ON(!atomic_read(&bio->bi_cnt));
bio               251 fs/bio.c       	if (atomic_dec_and_test(&bio->bi_cnt)) {
bio               252 fs/bio.c       		bio->bi_next = NULL;
bio               253 fs/bio.c       		bio->bi_destructor(bio);
bio               259 fs/bio.c       	if (unlikely(!bio_flagged(bio, BIO_SEG_VALID)))
bio               260 fs/bio.c       		blk_recount_segments(q, bio);
bio               262 fs/bio.c       	return bio->bi_phys_segments;
bio               276 fs/bio.c       	memcpy(bio->bi_io_vec, bio_src->bi_io_vec,
bio               283 fs/bio.c       	bio->bi_sector = bio_src->bi_sector;
bio               284 fs/bio.c       	bio->bi_bdev = bio_src->bi_bdev;
bio               285 fs/bio.c       	bio->bi_flags |= 1 << BIO_CLONED;
bio               286 fs/bio.c       	bio->bi_rw = bio_src->bi_rw;
bio               287 fs/bio.c       	bio->bi_vcnt = bio_src->bi_vcnt;
bio               288 fs/bio.c       	bio->bi_size = bio_src->bi_size;
bio               289 fs/bio.c       	bio->bi_idx = bio_src->bi_idx;
bio               299 fs/bio.c       struct bio *bio_clone(struct bio *bio, gfp_t gfp_mask)
bio               301 fs/bio.c       	struct bio *b = bio_alloc_bioset(gfp_mask, bio->bi_max_vecs, fs_bio_set);
bio               307 fs/bio.c       	__bio_clone(b, bio);
bio               309 fs/bio.c       	if (bio_integrity(bio)) {
bio               312 fs/bio.c       		ret = bio_integrity_clone(b, bio, fs_bio_set);
bio               354 fs/bio.c       	if (unlikely(bio_flagged(bio, BIO_CLONED)))
bio               357 fs/bio.c       	if (((bio->bi_size + len) >> 9) > max_sectors)
bio               365 fs/bio.c       	if (bio->bi_vcnt > 0) {
bio               366 fs/bio.c       		struct bio_vec *prev = &bio->bi_io_vec[bio->bi_vcnt - 1];
bio               374 fs/bio.c       					.bi_bdev = bio->bi_bdev,
bio               375 fs/bio.c       					.bi_sector = bio->bi_sector,
bio               376 fs/bio.c       					.bi_size = bio->bi_size,
bio               377 fs/bio.c       					.bi_rw = bio->bi_rw,
bio               390 fs/bio.c       	if (bio->bi_vcnt >= bio->bi_max_vecs)
bio               398 fs/bio.c       	while (bio->bi_phys_segments >= q->max_phys_segments
bio               399 fs/bio.c       	       || bio->bi_phys_segments >= q->max_hw_segments) {
bio               405 fs/bio.c       		blk_recount_segments(q, bio);
bio               412 fs/bio.c       	bvec = &bio->bi_io_vec[bio->bi_vcnt];
bio               424 fs/bio.c       			.bi_bdev = bio->bi_bdev,
bio               425 fs/bio.c       			.bi_sector = bio->bi_sector,
bio               426 fs/bio.c       			.bi_size = bio->bi_size,
bio               427 fs/bio.c       			.bi_rw = bio->bi_rw,
bio               443 fs/bio.c       	if (bio->bi_vcnt && (BIOVEC_PHYS_MERGEABLE(bvec-1, bvec)))
bio               444 fs/bio.c       		bio->bi_flags &= ~(1 << BIO_SEG_VALID);
bio               446 fs/bio.c       	bio->bi_vcnt++;
bio               447 fs/bio.c       	bio->bi_phys_segments++;
bio               449 fs/bio.c       	bio->bi_size += len;
bio               470 fs/bio.c       	return __bio_add_page(q, bio, page, len, offset, q->max_hw_sectors);
bio               489 fs/bio.c       	struct request_queue *q = bdev_get_queue(bio->bi_bdev);
bio               490 fs/bio.c       	return __bio_add_page(q, bio, page, len, offset, q->max_sectors);
bio               504 fs/bio.c       	memcpy(bmd->iovecs, bio->bi_io_vec, sizeof(struct bio_vec) * bio->bi_vcnt);
bio               508 fs/bio.c       	bio->bi_private = bmd;
bio               549 fs/bio.c       	int read = bio_data_dir(bio) == READ;
bio               551 fs/bio.c       	__bio_for_each_segment(bvec, bio, i, 0) {
bio               602 fs/bio.c       	struct bio_map_data *bmd = bio->bi_private;
bio               605 fs/bio.c       	if (!bio_flagged(bio, BIO_NULL_MAPPED))
bio               606 fs/bio.c       		ret = __bio_copy_iov(bio, bmd->iovecs, bmd->sgvecs,
bio               609 fs/bio.c       	bio_put(bio);
bio               626 fs/bio.c       struct bio *bio_copy_user_iov(struct request_queue *q,
bio               634 fs/bio.c       	struct bio *bio;
bio               657 fs/bio.c       	bio = bio_alloc(gfp_mask, nr_pages);
bio               658 fs/bio.c       	if (!bio)
bio               661 fs/bio.c       	bio->bi_rw |= (!write_to_vm << BIO_RW);
bio               689 fs/bio.c       		if (bio_add_pc_page(q, bio, page, bytes, 0) < bytes)
bio               702 fs/bio.c       		ret = __bio_copy_iov(bio, bio->bi_io_vec, iov, iov_count, 0, 0);
bio               707 fs/bio.c       	bio_set_map_data(bmd, bio, iov, iov_count, map_data ? 0 : 1);
bio               708 fs/bio.c       	return bio;
bio               711 fs/bio.c       		bio_for_each_segment(bvec, bio, i)
bio               714 fs/bio.c       	bio_put(bio);
bio               733 fs/bio.c       struct bio *bio_copy_user(struct request_queue *q, struct rq_map_data *map_data,
bio               745 fs/bio.c       static struct bio *__bio_map_user_iov(struct request_queue *q,
bio               753 fs/bio.c       	struct bio *bio;
bio               774 fs/bio.c       	bio = bio_alloc(gfp_mask, nr_pages);
bio               775 fs/bio.c       	if (!bio)
bio               811 fs/bio.c       			if (bio_add_pc_page(q, bio, pages[j], bytes, offset) <
bio               833 fs/bio.c       		bio->bi_rw |= (1 << BIO_RW);
bio               835 fs/bio.c       	bio->bi_bdev = bdev;
bio               836 fs/bio.c       	bio->bi_flags |= (1 << BIO_USER_MAPPED);
bio               837 fs/bio.c       	return bio;
bio               847 fs/bio.c       	bio_put(bio);
bio               863 fs/bio.c       struct bio *bio_map_user(struct request_queue *q, struct block_device *bdev,
bio               887 fs/bio.c       struct bio *bio_map_user_iov(struct request_queue *q, struct block_device *bdev,
bio               891 fs/bio.c       	struct bio *bio;
bio               893 fs/bio.c       	bio = __bio_map_user_iov(q, bdev, iov, iov_count, write_to_vm,
bio               895 fs/bio.c       	if (IS_ERR(bio))
bio               896 fs/bio.c       		return bio;
bio               904 fs/bio.c       	bio_get(bio);
bio               906 fs/bio.c       	return bio;
bio               917 fs/bio.c       	__bio_for_each_segment(bvec, bio, i, 0) {
bio               918 fs/bio.c       		if (bio_data_dir(bio) == READ)
bio               924 fs/bio.c       	bio_put(bio);
bio               938 fs/bio.c       	__bio_unmap_user(bio);
bio               939 fs/bio.c       	bio_put(bio);
bio               944 fs/bio.c       	bio_put(bio);
bio               948 fs/bio.c       static struct bio *__bio_map_kern(struct request_queue *q, void *data,
bio               956 fs/bio.c       	struct bio *bio;
bio               958 fs/bio.c       	bio = bio_alloc(gfp_mask, nr_pages);
bio               959 fs/bio.c       	if (!bio)
bio               972 fs/bio.c       		if (bio_add_pc_page(q, bio, virt_to_page(data), bytes,
bio               981 fs/bio.c       	bio->bi_end_io = bio_map_kern_endio;
bio               982 fs/bio.c       	return bio;
bio               995 fs/bio.c       struct bio *bio_map_kern(struct request_queue *q, void *data, unsigned int len,
bio               998 fs/bio.c       	struct bio *bio;
bio              1000 fs/bio.c       	bio = __bio_map_kern(q, data, len, gfp_mask);
bio              1001 fs/bio.c       	if (IS_ERR(bio))
bio              1002 fs/bio.c       		return bio;
bio              1004 fs/bio.c       	if (bio->bi_size == len)
bio              1005 fs/bio.c       		return bio;
bio              1010 fs/bio.c       	bio_put(bio);
bio              1017 fs/bio.c       	const int read = bio_data_dir(bio) == READ;
bio              1018 fs/bio.c       	struct bio_map_data *bmd = bio->bi_private;
bio              1022 fs/bio.c       	__bio_for_each_segment(bvec, bio, i, 0) {
bio              1034 fs/bio.c       	bio_put(bio);
bio              1048 fs/bio.c       struct bio *bio_copy_kern(struct request_queue *q, void *data, unsigned int len,
bio              1051 fs/bio.c       	struct bio *bio;
bio              1055 fs/bio.c       	bio = bio_copy_user(q, NULL, (unsigned long)data, len, 1, gfp_mask);
bio              1056 fs/bio.c       	if (IS_ERR(bio))
bio              1057 fs/bio.c       		return bio;
bio              1062 fs/bio.c       		bio_for_each_segment(bvec, bio, i) {
bio              1070 fs/bio.c       	bio->bi_end_io = bio_copy_kern_endio;
bio              1072 fs/bio.c       	return bio;
bio              1106 fs/bio.c       	struct bio_vec *bvec = bio->bi_io_vec;
bio              1109 fs/bio.c       	for (i = 0; i < bio->bi_vcnt; i++) {
bio              1119 fs/bio.c       	struct bio_vec *bvec = bio->bi_io_vec;
bio              1122 fs/bio.c       	for (i = 0; i < bio->bi_vcnt; i++) {
bio              1145 fs/bio.c       static struct bio *bio_dirty_list;
bio              1153 fs/bio.c       	struct bio *bio;
bio              1156 fs/bio.c       	bio = bio_dirty_list;
bio              1160 fs/bio.c       	while (bio) {
bio              1161 fs/bio.c       		struct bio *next = bio->bi_private;
bio              1163 fs/bio.c       		bio_set_pages_dirty(bio);
bio              1164 fs/bio.c       		bio_release_pages(bio);
bio              1165 fs/bio.c       		bio_put(bio);
bio              1166 fs/bio.c       		bio = next;
bio              1172 fs/bio.c       	struct bio_vec *bvec = bio->bi_io_vec;
bio              1176 fs/bio.c       	for (i = 0; i < bio->bi_vcnt; i++) {
bio              1191 fs/bio.c       		bio->bi_private = bio_dirty_list;
bio              1192 fs/bio.c       		bio_dirty_list = bio;
bio              1196 fs/bio.c       		bio_put(bio);
bio              1217 fs/bio.c       		clear_bit(BIO_UPTODATE, &bio->bi_flags);
bio              1218 fs/bio.c       	else if (!test_bit(BIO_UPTODATE, &bio->bi_flags))
bio              1221 fs/bio.c       	if (bio->bi_end_io)
bio              1222 fs/bio.c       		bio->bi_end_io(bio, error);
bio              1228 fs/bio.c       		struct bio *master = bp->bio1.bi_private;
bio              1316 fs/bio.c       	unsigned int sector_sz = queue_hardsect_size(bio->bi_bdev->bd_disk->queue);
bio              1323 fs/bio.c       	if (index >= bio->bi_idx)
bio              1324 fs/bio.c       		index = bio->bi_vcnt - 1;
bio              1326 fs/bio.c       	__bio_for_each_segment(bv, bio, i, 0) {
bio              1421 fs/bio.c       	bio_slab = KMEM_CACHE(bio, SLAB_HWCACHE_ALIGN|SLAB_PANIC);
bio              2909 fs/buffer.c    	struct buffer_head *bh = bio->bi_private;
bio              2912 fs/buffer.c    		set_bit(BIO_EOPNOTSUPP, &bio->bi_flags);
bio              2916 fs/buffer.c    	bh->b_end_io(bh, test_bit(BIO_UPTODATE, &bio->bi_flags));
bio              2917 fs/buffer.c    	bio_put(bio);
bio              2922 fs/buffer.c    	struct bio *bio;
bio              2946 fs/buffer.c    	bio = bio_alloc(GFP_NOIO, 1);
bio              2948 fs/buffer.c    	bio->bi_sector = bh->b_blocknr * (bh->b_size >> 9);
bio              2949 fs/buffer.c    	bio->bi_bdev = bh->b_bdev;
bio              2950 fs/buffer.c    	bio->bi_io_vec[0].bv_page = bh->b_page;
bio              2951 fs/buffer.c    	bio->bi_io_vec[0].bv_len = bh->b_size;
bio              2952 fs/buffer.c    	bio->bi_io_vec[0].bv_offset = bh_offset(bh);
bio              2954 fs/buffer.c    	bio->bi_vcnt = 1;
bio              2955 fs/buffer.c    	bio->bi_idx = 0;
bio              2956 fs/buffer.c    	bio->bi_size = bh->b_size;
bio              2958 fs/buffer.c    	bio->bi_end_io = end_bio_bh_io_sync;
bio              2959 fs/buffer.c    	bio->bi_private = bh;
bio              2961 fs/buffer.c    	bio_get(bio);
bio              2962 fs/buffer.c    	submit_bio(rw, bio);
bio              2964 fs/buffer.c    	if (bio_flagged(bio, BIO_EOPNOTSUPP))
bio              2967 fs/buffer.c    	bio_put(bio);
bio                67 fs/direct-io.c 	struct bio *bio;		/* bio under assembly */
bio               126 fs/direct-io.c 	struct bio *bio_list;		/* singly linked via bi_private */
bio               263 fs/direct-io.c 	struct dio *dio = bio->bi_private;
bio               268 fs/direct-io.c 	dio_bio_complete(dio, bio);
bio               292 fs/direct-io.c 	struct dio *dio = bio->bi_private;
bio               296 fs/direct-io.c 	bio->bi_private = dio->bio_list;
bio               297 fs/direct-io.c 	dio->bio_list = bio;
bio               307 fs/direct-io.c 	struct bio *bio;
bio               309 fs/direct-io.c 	bio = bio_alloc(GFP_KERNEL, nr_vecs);
bio               310 fs/direct-io.c 	if (bio == NULL)
bio               313 fs/direct-io.c 	bio->bi_bdev = bdev;
bio               314 fs/direct-io.c 	bio->bi_sector = first_sector;
bio               316 fs/direct-io.c 		bio->bi_end_io = dio_bio_end_aio;
bio               318 fs/direct-io.c 		bio->bi_end_io = dio_bio_end_io;
bio               320 fs/direct-io.c 	dio->bio = bio;
bio               333 fs/direct-io.c 	struct bio *bio = dio->bio;
bio               336 fs/direct-io.c 	bio->bi_private = dio;
bio               343 fs/direct-io.c 		bio_set_pages_dirty(bio);
bio               345 fs/direct-io.c 	submit_bio(dio->rw, bio);
bio               347 fs/direct-io.c 	dio->bio = NULL;
bio               366 fs/direct-io.c static struct bio *dio_await_one(struct dio *dio)
bio               369 fs/direct-io.c 	struct bio *bio = NULL;
bio               389 fs/direct-io.c 		bio = dio->bio_list;
bio               390 fs/direct-io.c 		dio->bio_list = bio->bi_private;
bio               393 fs/direct-io.c 	return bio;
bio               401 fs/direct-io.c 	const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio               402 fs/direct-io.c 	struct bio_vec *bvec = bio->bi_io_vec;
bio               409 fs/direct-io.c 		bio_check_pages_dirty(bio);	/* transfers ownership */
bio               411 fs/direct-io.c 		for (page_no = 0; page_no < bio->bi_vcnt; page_no++) {
bio               418 fs/direct-io.c 		bio_put(bio);
bio               432 fs/direct-io.c 	struct bio *bio;
bio               434 fs/direct-io.c 		bio = dio_await_one(dio);
bio               435 fs/direct-io.c 		if (bio)
bio               436 fs/direct-io.c 			dio_bio_complete(dio, bio);
bio               437 fs/direct-io.c 	} while (bio);
bio               454 fs/direct-io.c 			struct bio *bio;
bio               458 fs/direct-io.c 			bio = dio->bio_list;
bio               459 fs/direct-io.c 			dio->bio_list = bio->bi_private;
bio               461 fs/direct-io.c 			ret2 = dio_bio_complete(dio, bio);
bio               572 fs/direct-io.c 	ret = bio_add_page(dio->bio, dio->cur_page,
bio               604 fs/direct-io.c 	if (dio->bio) {
bio               618 fs/direct-io.c 	if (dio->bio == NULL) {
bio              1030 fs/direct-io.c 	if (dio->bio)
bio              2295 fs/ext4/extents.c 	complete((struct completion *)bio->bi_private);
bio              2302 fs/ext4/extents.c 	struct bio *bio;
bio              2324 fs/ext4/extents.c 		bio = bio_alloc(GFP_NOIO, len);
bio              2325 fs/ext4/extents.c 		if (!bio)
bio              2327 fs/ext4/extents.c 		bio->bi_sector = ee_pblock;
bio              2328 fs/ext4/extents.c 		bio->bi_bdev   = inode->i_sb->s_bdev;
bio              2333 fs/ext4/extents.c 			ret = bio_add_page(bio, ZERO_PAGE(0),
bio              2349 fs/ext4/extents.c 		bio->bi_private = &event;
bio              2350 fs/ext4/extents.c 		bio->bi_end_io = bi_complete;
bio              2351 fs/ext4/extents.c 		submit_bio(WRITE, bio);
bio              2354 fs/ext4/extents.c 		if (test_bit(BIO_UPTODATE, &bio->bi_flags))
bio              2360 fs/ext4/extents.c 		bio_put(bio);
bio               219 fs/gfs2/ops_fstype.c 	struct page *page = bio->bi_private;
bio               272 fs/gfs2/ops_fstype.c 	struct bio *bio;
bio               282 fs/gfs2/ops_fstype.c 	bio = bio_alloc(GFP_NOFS, 1);
bio               283 fs/gfs2/ops_fstype.c 	if (unlikely(!bio)) {
bio               288 fs/gfs2/ops_fstype.c 	bio->bi_sector = sector * (sb->s_blocksize >> 9);
bio               289 fs/gfs2/ops_fstype.c 	bio->bi_bdev = sb->s_bdev;
bio               290 fs/gfs2/ops_fstype.c 	bio_add_page(bio, page, PAGE_SIZE, 0);
bio               292 fs/gfs2/ops_fstype.c 	bio->bi_end_io = end_bio_io_page;
bio               293 fs/gfs2/ops_fstype.c 	bio->bi_private = page;
bio               294 fs/gfs2/ops_fstype.c 	submit_bio(READ_SYNC | (1 << BIO_RW_META), bio);
bio               296 fs/gfs2/ops_fstype.c 	bio_put(bio);
bio              1994 fs/jfs/jfs_logmgr.c 	struct bio *bio;
bio              2005 fs/jfs/jfs_logmgr.c 	bio = bio_alloc(GFP_NOFS, 1);
bio              2007 fs/jfs/jfs_logmgr.c 	bio->bi_sector = bp->l_blkno << (log->l2bsize - 9);
bio              2008 fs/jfs/jfs_logmgr.c 	bio->bi_bdev = log->bdev;
bio              2009 fs/jfs/jfs_logmgr.c 	bio->bi_io_vec[0].bv_page = bp->l_page;
bio              2010 fs/jfs/jfs_logmgr.c 	bio->bi_io_vec[0].bv_len = LOGPSIZE;
bio              2011 fs/jfs/jfs_logmgr.c 	bio->bi_io_vec[0].bv_offset = bp->l_offset;
bio              2013 fs/jfs/jfs_logmgr.c 	bio->bi_vcnt = 1;
bio              2014 fs/jfs/jfs_logmgr.c 	bio->bi_idx = 0;
bio              2015 fs/jfs/jfs_logmgr.c 	bio->bi_size = LOGPSIZE;
bio              2017 fs/jfs/jfs_logmgr.c 	bio->bi_end_io = lbmIODone;
bio              2018 fs/jfs/jfs_logmgr.c 	bio->bi_private = bp;
bio              2019 fs/jfs/jfs_logmgr.c 	submit_bio(READ_SYNC, bio);
bio              2142 fs/jfs/jfs_logmgr.c 	struct bio *bio;
bio              2147 fs/jfs/jfs_logmgr.c 	bio = bio_alloc(GFP_NOFS, 1);
bio              2148 fs/jfs/jfs_logmgr.c 	bio->bi_sector = bp->l_blkno << (log->l2bsize - 9);
bio              2149 fs/jfs/jfs_logmgr.c 	bio->bi_bdev = log->bdev;
bio              2150 fs/jfs/jfs_logmgr.c 	bio->bi_io_vec[0].bv_page = bp->l_page;
bio              2151 fs/jfs/jfs_logmgr.c 	bio->bi_io_vec[0].bv_len = LOGPSIZE;
bio              2152 fs/jfs/jfs_logmgr.c 	bio->bi_io_vec[0].bv_offset = bp->l_offset;
bio              2154 fs/jfs/jfs_logmgr.c 	bio->bi_vcnt = 1;
bio              2155 fs/jfs/jfs_logmgr.c 	bio->bi_idx = 0;
bio              2156 fs/jfs/jfs_logmgr.c 	bio->bi_size = LOGPSIZE;
bio              2158 fs/jfs/jfs_logmgr.c 	bio->bi_end_io = lbmIODone;
bio              2159 fs/jfs/jfs_logmgr.c 	bio->bi_private = bp;
bio              2163 fs/jfs/jfs_logmgr.c 		bio->bi_size = 0;
bio              2164 fs/jfs/jfs_logmgr.c 		lbmIODone(bio, 0);
bio              2166 fs/jfs/jfs_logmgr.c 		submit_bio(WRITE_SYNC, bio);
bio              2204 fs/jfs/jfs_logmgr.c 	struct lbuf *bp = bio->bi_private;
bio              2218 fs/jfs/jfs_logmgr.c 	if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) {
bio              2224 fs/jfs/jfs_logmgr.c 	bio_put(bio);
bio               287 fs/jfs/jfs_metapage.c 	struct page *page = bio->bi_private;
bio               289 fs/jfs/jfs_metapage.c 	if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) {
bio               295 fs/jfs/jfs_metapage.c 	bio_put(bio);
bio               342 fs/jfs/jfs_metapage.c 	struct page *page = bio->bi_private;
bio               346 fs/jfs/jfs_metapage.c 	if (! test_bit(BIO_UPTODATE, &bio->bi_flags)) {
bio               351 fs/jfs/jfs_metapage.c 	bio_put(bio);
bio               356 fs/jfs/jfs_metapage.c 	struct bio *bio = NULL;
bio               399 fs/jfs/jfs_metapage.c 		if (bio) {
bio               409 fs/jfs/jfs_metapage.c 			if (bio_add_page(bio, page, bio_bytes, bio_offset) <
bio               417 fs/jfs/jfs_metapage.c 			if (!bio->bi_size)
bio               419 fs/jfs/jfs_metapage.c 			submit_bio(WRITE, bio);
bio               421 fs/jfs/jfs_metapage.c 			bio = NULL;
bio               435 fs/jfs/jfs_metapage.c 		bio = bio_alloc(GFP_NOFS, 1);
bio               436 fs/jfs/jfs_metapage.c 		bio->bi_bdev = inode->i_sb->s_bdev;
bio               437 fs/jfs/jfs_metapage.c 		bio->bi_sector = pblock << (inode->i_blkbits - 9);
bio               438 fs/jfs/jfs_metapage.c 		bio->bi_end_io = metapage_write_end_io;
bio               439 fs/jfs/jfs_metapage.c 		bio->bi_private = page;
bio               448 fs/jfs/jfs_metapage.c 	if (bio) {
bio               449 fs/jfs/jfs_metapage.c 		if (bio_add_page(bio, page, bio_bytes, bio_offset) < bio_bytes)
bio               451 fs/jfs/jfs_metapage.c 		if (!bio->bi_size)
bio               454 fs/jfs/jfs_metapage.c 		submit_bio(WRITE, bio);
bio               472 fs/jfs/jfs_metapage.c 		       4, bio, sizeof(*bio), 0);
bio               474 fs/jfs/jfs_metapage.c 	bio_put(bio);
bio               484 fs/jfs/jfs_metapage.c 	struct bio *bio = NULL;
bio               506 fs/jfs/jfs_metapage.c 			if (bio)
bio               507 fs/jfs/jfs_metapage.c 				submit_bio(READ, bio);
bio               509 fs/jfs/jfs_metapage.c 			bio = bio_alloc(GFP_NOFS, 1);
bio               510 fs/jfs/jfs_metapage.c 			bio->bi_bdev = inode->i_sb->s_bdev;
bio               511 fs/jfs/jfs_metapage.c 			bio->bi_sector = pblock << (inode->i_blkbits - 9);
bio               512 fs/jfs/jfs_metapage.c 			bio->bi_end_io = metapage_read_end_io;
bio               513 fs/jfs/jfs_metapage.c 			bio->bi_private = page;
bio               516 fs/jfs/jfs_metapage.c 			if (bio_add_page(bio, page, len, offset) < len)
bio               522 fs/jfs/jfs_metapage.c 	if (bio)
bio               523 fs/jfs/jfs_metapage.c 		submit_bio(READ, bio);
bio               531 fs/jfs/jfs_metapage.c 	bio_put(bio);
bio                44 fs/mpage.c     	const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio                45 fs/mpage.c     	struct bio_vec *bvec = bio->bi_io_vec + bio->bi_vcnt - 1;
bio                50 fs/mpage.c     		if (--bvec >= bio->bi_io_vec)
bio                60 fs/mpage.c     	} while (bvec >= bio->bi_io_vec);
bio                61 fs/mpage.c     	bio_put(bio);
bio                66 fs/mpage.c     	const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio                67 fs/mpage.c     	struct bio_vec *bvec = bio->bi_io_vec + bio->bi_vcnt - 1;
bio                72 fs/mpage.c     		if (--bvec >= bio->bi_io_vec)
bio                81 fs/mpage.c     	} while (bvec >= bio->bi_io_vec);
bio                82 fs/mpage.c     	bio_put(bio);
bio                85 fs/mpage.c     struct bio *mpage_bio_submit(int rw, struct bio *bio)
bio                87 fs/mpage.c     	bio->bi_end_io = mpage_end_io_read;
bio                89 fs/mpage.c     		bio->bi_end_io = mpage_end_io_write;
bio                90 fs/mpage.c     	submit_bio(rw, bio);
bio                95 fs/mpage.c     static struct bio *
bio               100 fs/mpage.c     	struct bio *bio;
bio               102 fs/mpage.c     	bio = bio_alloc(gfp_flags, nr_vecs);
bio               104 fs/mpage.c     	if (bio == NULL && (current->flags & PF_MEMALLOC)) {
bio               105 fs/mpage.c     		while (!bio && (nr_vecs /= 2))
bio               106 fs/mpage.c     			bio = bio_alloc(gfp_flags, nr_vecs);
bio               109 fs/mpage.c     	if (bio) {
bio               110 fs/mpage.c     		bio->bi_bdev = bdev;
bio               111 fs/mpage.c     		bio->bi_sector = first_sector;
bio               113 fs/mpage.c     	return bio;
bio               168 fs/mpage.c     static struct bio *
bio               293 fs/mpage.c     	if (bio && (*last_block_in_bio != blocks[0] - 1))
bio               294 fs/mpage.c     		bio = mpage_bio_submit(READ, bio);
bio               297 fs/mpage.c     	if (bio == NULL) {
bio               298 fs/mpage.c     		bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9),
bio               301 fs/mpage.c     		if (bio == NULL)
bio               306 fs/mpage.c     	if (bio_add_page(bio, page, length, 0) < length) {
bio               307 fs/mpage.c     		bio = mpage_bio_submit(READ, bio);
bio               312 fs/mpage.c     		bio = mpage_bio_submit(READ, bio);
bio               316 fs/mpage.c     	return bio;
bio               319 fs/mpage.c     	if (bio)
bio               320 fs/mpage.c     		bio = mpage_bio_submit(READ, bio);
bio               375 fs/mpage.c     	struct bio *bio = NULL;
bio               389 fs/mpage.c     			bio = do_mpage_readpage(bio, page,
bio               398 fs/mpage.c     	if (bio)
bio               399 fs/mpage.c     		mpage_bio_submit(READ, bio);
bio               409 fs/mpage.c     	struct bio *bio = NULL;
bio               415 fs/mpage.c     	bio = do_mpage_readpage(bio, page, 1, &last_block_in_bio,
bio               417 fs/mpage.c     	if (bio)
bio               418 fs/mpage.c     		mpage_bio_submit(READ, bio);
bio               444 fs/mpage.c     	struct bio *bio = mpd->bio;
bio               570 fs/mpage.c     	if (bio && mpd->last_block_in_bio != blocks[0] - 1)
bio               571 fs/mpage.c     		bio = mpage_bio_submit(WRITE, bio);
bio               574 fs/mpage.c     	if (bio == NULL) {
bio               575 fs/mpage.c     		bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9),
bio               577 fs/mpage.c     		if (bio == NULL)
bio               587 fs/mpage.c     	if (bio_add_page(bio, page, length, 0) < length) {
bio               588 fs/mpage.c     		bio = mpage_bio_submit(WRITE, bio);
bio               621 fs/mpage.c     		bio = mpage_bio_submit(WRITE, bio);
bio               632 fs/mpage.c     	if (bio)
bio               633 fs/mpage.c     		bio = mpage_bio_submit(WRITE, bio);
bio               646 fs/mpage.c     	mpd->bio = bio;
bio               680 fs/mpage.c     			.bio = NULL,
bio               687 fs/mpage.c     		if (mpd.bio)
bio               688 fs/mpage.c     			mpage_bio_submit(WRITE, mpd.bio);
bio               698 fs/mpage.c     		.bio = NULL,
bio               704 fs/mpage.c     	if (mpd.bio)
bio               705 fs/mpage.c     		mpage_bio_submit(WRITE, mpd.bio);
bio               222 fs/ocfs2/cluster/heartbeat.c 	struct o2hb_bio_wait_ctxt *wc = bio->bi_private;
bio               230 fs/ocfs2/cluster/heartbeat.c 	bio_put(bio);
bio               235 fs/ocfs2/cluster/heartbeat.c static struct bio *o2hb_setup_one_bio(struct o2hb_region *reg,
bio               245 fs/ocfs2/cluster/heartbeat.c 	struct bio *bio;
bio               252 fs/ocfs2/cluster/heartbeat.c 	bio = bio_alloc(GFP_ATOMIC, 16);
bio               253 fs/ocfs2/cluster/heartbeat.c 	if (!bio) {
bio               255 fs/ocfs2/cluster/heartbeat.c 		bio = ERR_PTR(-ENOMEM);
bio               260 fs/ocfs2/cluster/heartbeat.c 	bio->bi_sector = (reg->hr_start_block + cs) << (bits - 9);
bio               261 fs/ocfs2/cluster/heartbeat.c 	bio->bi_bdev = reg->hr_bdev;
bio               262 fs/ocfs2/cluster/heartbeat.c 	bio->bi_private = wc;
bio               263 fs/ocfs2/cluster/heartbeat.c 	bio->bi_end_io = o2hb_bio_end_io;
bio               276 fs/ocfs2/cluster/heartbeat.c 		len = bio_add_page(bio, page, vec_len, vec_start);
bio               285 fs/ocfs2/cluster/heartbeat.c 	return bio;
bio               294 fs/ocfs2/cluster/heartbeat.c 	struct bio *bio;
bio               299 fs/ocfs2/cluster/heartbeat.c 		bio = o2hb_setup_one_bio(reg, &wc, &current_slot, max_slots);
bio               300 fs/ocfs2/cluster/heartbeat.c 		if (IS_ERR(bio)) {
bio               301 fs/ocfs2/cluster/heartbeat.c 			status = PTR_ERR(bio);
bio               307 fs/ocfs2/cluster/heartbeat.c 		submit_bio(READ, bio);
bio               325 fs/ocfs2/cluster/heartbeat.c 	struct bio *bio;
bio               331 fs/ocfs2/cluster/heartbeat.c 	bio = o2hb_setup_one_bio(reg, write_wc, &slot, slot+1);
bio               332 fs/ocfs2/cluster/heartbeat.c 	if (IS_ERR(bio)) {
bio               333 fs/ocfs2/cluster/heartbeat.c 		status = PTR_ERR(bio);
bio               339 fs/ocfs2/cluster/heartbeat.c 	submit_bio(WRITE, bio);
bio               347 fs/xfs/linux-2.6/xfs_aops.c 	xfs_ioend_t		*ioend = bio->bi_private;
bio               349 fs/xfs/linux-2.6/xfs_aops.c 	ASSERT(atomic_read(&bio->bi_cnt) >= 1);
bio               350 fs/xfs/linux-2.6/xfs_aops.c 	ioend->io_error = test_bit(BIO_UPTODATE, &bio->bi_flags) ? 0 : error;
bio               353 fs/xfs/linux-2.6/xfs_aops.c 	bio->bi_private = NULL;
bio               354 fs/xfs/linux-2.6/xfs_aops.c 	bio->bi_end_io = NULL;
bio               355 fs/xfs/linux-2.6/xfs_aops.c 	bio_put(bio);
bio               367 fs/xfs/linux-2.6/xfs_aops.c 	bio->bi_private = ioend;
bio               368 fs/xfs/linux-2.6/xfs_aops.c 	bio->bi_end_io = xfs_end_bio;
bio               370 fs/xfs/linux-2.6/xfs_aops.c 	submit_bio(WRITE, bio);
bio               371 fs/xfs/linux-2.6/xfs_aops.c 	ASSERT(!bio_flagged(bio, BIO_EOPNOTSUPP));
bio               372 fs/xfs/linux-2.6/xfs_aops.c 	bio_put(bio);
bio               375 fs/xfs/linux-2.6/xfs_aops.c STATIC struct bio *
bio               379 fs/xfs/linux-2.6/xfs_aops.c 	struct bio		*bio;
bio               383 fs/xfs/linux-2.6/xfs_aops.c 		bio = bio_alloc(GFP_NOIO, nvecs);
bio               385 fs/xfs/linux-2.6/xfs_aops.c 	} while (!bio);
bio               387 fs/xfs/linux-2.6/xfs_aops.c 	ASSERT(bio->bi_private == NULL);
bio               388 fs/xfs/linux-2.6/xfs_aops.c 	bio->bi_sector = bh->b_blocknr * (bh->b_size >> 9);
bio               389 fs/xfs/linux-2.6/xfs_aops.c 	bio->bi_bdev = bh->b_bdev;
bio               390 fs/xfs/linux-2.6/xfs_aops.c 	bio_get(bio);
bio               391 fs/xfs/linux-2.6/xfs_aops.c 	return bio;
bio               427 fs/xfs/linux-2.6/xfs_aops.c 	return bio_add_page(bio, bh->b_page, bh->b_size, bh_offset(bh));
bio               454 fs/xfs/linux-2.6/xfs_aops.c 	struct bio		*bio;
bio               469 fs/xfs/linux-2.6/xfs_aops.c 		bio = NULL;
bio               473 fs/xfs/linux-2.6/xfs_aops.c 			if (!bio) {
bio               475 fs/xfs/linux-2.6/xfs_aops.c 				bio = xfs_alloc_ioend_bio(bh);
bio               477 fs/xfs/linux-2.6/xfs_aops.c 				xfs_submit_ioend_bio(ioend, bio);
bio               481 fs/xfs/linux-2.6/xfs_aops.c 			if (bio_add_buffer(bio, bh) != bh->b_size) {
bio               482 fs/xfs/linux-2.6/xfs_aops.c 				xfs_submit_ioend_bio(ioend, bio);
bio               488 fs/xfs/linux-2.6/xfs_aops.c 		if (bio)
bio               489 fs/xfs/linux-2.6/xfs_aops.c 			xfs_submit_ioend_bio(ioend, bio);
bio              1113 fs/xfs/linux-2.6/xfs_buf.c 	xfs_buf_t		*bp = (xfs_buf_t *)bio->bi_private;
bio              1115 fs/xfs/linux-2.6/xfs_buf.c 	struct bio_vec		*bvec = bio->bi_io_vec + bio->bi_vcnt - 1;
bio              1117 fs/xfs/linux-2.6/xfs_buf.c 	if (!test_bit(BIO_UPTODATE, &bio->bi_flags))
bio              1134 fs/xfs/linux-2.6/xfs_buf.c 		if (--bvec >= bio->bi_io_vec)
bio              1139 fs/xfs/linux-2.6/xfs_buf.c 	} while (bvec >= bio->bi_io_vec);
bio              1142 fs/xfs/linux-2.6/xfs_buf.c 	bio_put(bio);
bio              1150 fs/xfs/linux-2.6/xfs_buf.c 	struct bio		*bio;
bio              1180 fs/xfs/linux-2.6/xfs_buf.c 		bio = bio_alloc(GFP_NOIO, 1);
bio              1182 fs/xfs/linux-2.6/xfs_buf.c 		bio->bi_bdev = bp->b_target->bt_bdev;
bio              1183 fs/xfs/linux-2.6/xfs_buf.c 		bio->bi_sector = sector - (offset >> BBSHIFT);
bio              1184 fs/xfs/linux-2.6/xfs_buf.c 		bio->bi_end_io = xfs_buf_bio_end_io;
bio              1185 fs/xfs/linux-2.6/xfs_buf.c 		bio->bi_private = bp;
bio              1187 fs/xfs/linux-2.6/xfs_buf.c 		bio_add_page(bio, bp->b_pages[0], PAGE_CACHE_SIZE, 0);
bio              1201 fs/xfs/linux-2.6/xfs_buf.c 	bio = bio_alloc(GFP_NOIO, nr_pages);
bio              1202 fs/xfs/linux-2.6/xfs_buf.c 	bio->bi_bdev = bp->b_target->bt_bdev;
bio              1203 fs/xfs/linux-2.6/xfs_buf.c 	bio->bi_sector = sector;
bio              1204 fs/xfs/linux-2.6/xfs_buf.c 	bio->bi_end_io = xfs_buf_bio_end_io;
bio              1205 fs/xfs/linux-2.6/xfs_buf.c 	bio->bi_private = bp;
bio              1213 fs/xfs/linux-2.6/xfs_buf.c 		rbytes = bio_add_page(bio, bp->b_pages[map_i], nbytes, offset);
bio              1224 fs/xfs/linux-2.6/xfs_buf.c 	if (likely(bio->bi_size)) {
bio              1225 fs/xfs/linux-2.6/xfs_buf.c 		submit_bio(rw, bio);
bio              1229 fs/xfs/linux-2.6/xfs_buf.c 		bio_put(bio);
bio                53 include/linux/bio.h struct bio;
bio                65 include/linux/bio.h 	struct bio		*bi_next;	/* request queue link */
bio               113 include/linux/bio.h #define bio_flagged(bio, flag)	((bio)->bi_flags & (1 << (flag)))
bio               121 include/linux/bio.h #define BIO_POOL_IDX(bio)	((bio)->bi_flags >> BIO_POOL_OFFSET)	
bio               156 include/linux/bio.h #define bio_prio(bio)	((bio)->bi_rw >> BIO_PRIO_SHIFT)
bio               157 include/linux/bio.h #define bio_prio_valid(bio)	ioprio_valid(bio_prio(bio))
bio               161 include/linux/bio.h 	(bio)->bi_rw &= ((1UL << BIO_PRIO_SHIFT) - 1);		\
bio               162 include/linux/bio.h 	(bio)->bi_rw |= ((unsigned long) (prio) << BIO_PRIO_SHIFT);	\
bio               169 include/linux/bio.h #define bio_iovec_idx(bio, idx)	(&((bio)->bi_io_vec[(idx)]))
bio               170 include/linux/bio.h #define bio_iovec(bio)		bio_iovec_idx((bio), (bio)->bi_idx)
bio               171 include/linux/bio.h #define bio_page(bio)		bio_iovec((bio))->bv_page
bio               172 include/linux/bio.h #define bio_offset(bio)		bio_iovec((bio))->bv_offset
bio               173 include/linux/bio.h #define bio_segments(bio)	((bio)->bi_vcnt - (bio)->bi_idx)
bio               174 include/linux/bio.h #define bio_sectors(bio)	((bio)->bi_size >> 9)
bio               175 include/linux/bio.h #define bio_barrier(bio)	((bio)->bi_rw & (1 << BIO_RW_BARRIER))
bio               176 include/linux/bio.h #define bio_sync(bio)		((bio)->bi_rw & (1 << BIO_RW_SYNC))
bio               177 include/linux/bio.h #define bio_failfast(bio)	((bio)->bi_rw & (1 << BIO_RW_FAILFAST))
bio               178 include/linux/bio.h #define bio_rw_ahead(bio)	((bio)->bi_rw & (1 << BIO_RW_AHEAD))
bio               179 include/linux/bio.h #define bio_rw_meta(bio)	((bio)->bi_rw & (1 << BIO_RW_META))
bio               180 include/linux/bio.h #define bio_discard(bio)	((bio)->bi_rw & (1 << BIO_RW_DISCARD))
bio               181 include/linux/bio.h #define bio_empty_barrier(bio)	(bio_barrier(bio) && !bio_has_data(bio) && !bio_discard(bio))
bio               185 include/linux/bio.h 	if (bio->bi_vcnt)
bio               186 include/linux/bio.h 		return bio_iovec(bio)->bv_len >> 9;
bio               188 include/linux/bio.h 		return bio->bi_size >> 9;
bio               193 include/linux/bio.h 	if (bio->bi_vcnt)
bio               194 include/linux/bio.h 		return page_address(bio_page(bio)) + bio_offset(bio);
bio               202 include/linux/bio.h #define bio_to_phys(bio)	(page_to_phys(bio_page((bio))) + (unsigned long) bio_offset((bio)))
bio               212 include/linux/bio.h 	(kmap_atomic(bio_iovec_idx((bio), (idx))->bv_page, kmtype) +	\
bio               213 include/linux/bio.h 		bio_iovec_idx((bio), (idx))->bv_offset)
bio               221 include/linux/bio.h #define __BVEC_END(bio)		bio_iovec_idx((bio), (bio)->bi_vcnt - 1)
bio               222 include/linux/bio.h #define __BVEC_START(bio)	bio_iovec_idx((bio), (bio)->bi_idx)
bio               239 include/linux/bio.h #define bio_io_error(bio) bio_endio((bio), -EIO)
bio               246 include/linux/bio.h 	for (bvl = bio_iovec_idx((bio), (start_idx)), i = (start_idx);	\
bio               247 include/linux/bio.h 	     i < (bio)->bi_vcnt;					\
bio               251 include/linux/bio.h 	__bio_for_each_segment(bvl, bio, i, (bio)->bi_idx)
bio               267 include/linux/bio.h #define bio_get(bio)	atomic_inc(&(bio)->bi_cnt)
bio               274 include/linux/bio.h 	struct bio		*bip_bio;	/* parent bio */
bio               304 include/linux/bio.h 	struct bio			bio1, bio2;
bio               319 include/linux/bio.h extern struct bio *bio_alloc(gfp_t, int);
bio               320 include/linux/bio.h extern struct bio *bio_kmalloc(gfp_t, int);
bio               321 include/linux/bio.h extern struct bio *bio_alloc_bioset(gfp_t, int, struct bio_set *);
bio               330 include/linux/bio.h extern struct bio *bio_clone(struct bio *, gfp_t);
bio               339 include/linux/bio.h extern struct bio *bio_map_user(struct request_queue *, struct block_device *,
bio               343 include/linux/bio.h extern struct bio *bio_map_user_iov(struct request_queue *,
bio               347 include/linux/bio.h extern struct bio *bio_map_kern(struct request_queue *, void *, unsigned int,
bio               349 include/linux/bio.h extern struct bio *bio_copy_kern(struct request_queue *, void *, unsigned int,
bio               353 include/linux/bio.h extern struct bio *bio_copy_user(struct request_queue *, struct rq_map_data *,
bio               355 include/linux/bio.h extern struct bio *bio_copy_user_iov(struct request_queue *,
bio               368 include/linux/bio.h 	bio->bi_comp_cpu = cpu;
bio               441 include/linux/bio.h 	return bvec_kmap_irq(bio_iovec_idx(bio, idx), flags);
bio               446 include/linux/bio.h 	__bio_kmap_irq((bio), (bio)->bi_idx, (flags))
bio               454 include/linux/bio.h 	return bio && bio->bi_io_vec != NULL;
bio               470 include/linux/bio.h #define bio_integrity(bio) (bio->bi_integrity != NULL)
bio               167 include/linux/blkdev.h 	struct bio *bio;
bio               168 include/linux/blkdev.h 	struct bio *biotail;
bio               678 include/linux/blkdev.h 	struct bio *bio;
bio               683 include/linux/blkdev.h 	if ((rq->bio))			\
bio               684 include/linux/blkdev.h 		for (_bio = (rq)->bio; _bio; _bio = _bio->bi_next)
bio               687 include/linux/blkdev.h 	__rq_for_each_bio(_iter.bio, _rq)			\
bio               688 include/linux/blkdev.h 		bio_for_each_segment(bvl, _iter.bio, _iter.i)
bio               691 include/linux/blkdev.h 		(_iter.bio->bi_next == NULL && _iter.i == _iter.bio->bi_vcnt-1)
bio              1032 include/linux/blkdev.h 	if (rq->bio == NULL)
bio              1035 include/linux/blkdev.h 	return bio_integrity(rq->bio);
bio               235 include/linux/blktrace_api.h 	__blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw, what, !bio_flagged(bio, BIO_UPTODATE), 0, NULL);
bio               257 include/linux/blktrace_api.h 	if (bio)
bio               258 include/linux/blktrace_api.h 		blk_add_trace_bio(q, bio, what);
bio               285 include/linux/blktrace_api.h 	if (bio)
bio               286 include/linux/blktrace_api.h 		__blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw, what, !bio_flagged(bio, BIO_UPTODATE), sizeof(rpdu), &rpdu);
bio               314 include/linux/blktrace_api.h 	r.device_from = cpu_to_be32(bio->bi_bdev->bd_dev);
bio               317 include/linux/blktrace_api.h 	__blk_add_trace(bt, from, bio->bi_size, bio->bi_rw, BLK_TA_REMAP, !bio_flagged(bio, BIO_UPTODATE), sizeof(r), &r);
bio              1787 include/linux/fs.h #define bio_rw(bio)		((bio)->bi_rw & (RW_MASK | RWA_MASK))
bio              1792 include/linux/fs.h #define bio_data_dir(bio)	((bio)->bi_rw & 1)
bio              1927 include/linux/fs.h struct bio;
bio                59 include/linux/loop.h 	struct bio 		*lo_bio;
bio                60 include/linux/loop.h 	struct bio		*lo_biotail;
bio                15 include/linux/mpage.h 	struct bio *bio;
bio                23 include/linux/mpage.h struct bio *mpage_bio_submit(int rw, struct bio *bio);
bio               165 include/linux/pktcdvd.h 	struct bio		*read_queue;
bio               166 include/linux/pktcdvd.h 	struct bio		*read_queue_tail;
bio               167 include/linux/pktcdvd.h 	struct bio		*write_queue;
bio               168 include/linux/pktcdvd.h 	struct bio		*write_queue_tail;
bio               208 include/linux/pktcdvd.h 	struct bio		*orig_bios;	/* Original bios passed to pkt_make_request */
bio               209 include/linux/pktcdvd.h 	struct bio		*orig_bios_tail;/* that will be handled by this packet */
bio               213 include/linux/pktcdvd.h 	struct bio		*w_bio;		/* The bio we will send to the real CD */
bio               228 include/linux/pktcdvd.h 	struct bio		*r_bios[PACKET_MAX_SIZE]; /* bios to use during data gathering */
bio               241 include/linux/pktcdvd.h 	struct bio		*bio;
bio               246 include/linux/pktcdvd.h 	struct bio		*bio;		/* Original read request bio */
bio               230 include/linux/raid/md_k.h 	struct bio			*biolist; 	/* bios that need to be retried
bio               290 include/linux/raid/md_k.h 	int (*make_request)(struct request_queue *q, struct bio *bio);
bio                37 include/linux/raid/multipath.h 	struct bio		*master_bio;
bio                38 include/linux/raid/multipath.h 	struct bio		bio;
bio                95 include/linux/raid/raid1.h 	struct bio		*master_bio;
bio               107 include/linux/raid/raid1.h 	struct bio		*bios[0];
bio               116 include/linux/raid/raid1.h #define IO_BLOCKED ((struct bio*)1)
bio                90 include/linux/raid/raid10.h 	struct bio		*master_bio;
bio               105 include/linux/raid/raid10.h 		struct bio		*bio;
bio               116 include/linux/raid/raid10.h #define IO_BLOCKED ((struct bio*)1)
bio               219 include/linux/raid/raid5.h 		struct bio	req;
bio               222 include/linux/raid/raid5.h 		struct bio	*toread, *read, *towrite, *written;
bio               348 include/linux/raid/raid5.h 	struct bio		*retry_read_aligned; /* currently retrying aligned bios   */
bio               349 include/linux/raid/raid5.h 	struct bio		*retry_read_aligned_list; /* aligned bios retry list  */
bio                98 include/linux/sched.h struct bio;
bio              1249 include/linux/sched.h 	struct bio *bio_list, **bio_tail;
bio                16 include/linux/swap.h struct bio;
bio                63 kernel/power/swap.c 	struct bio *bio;
bio                65 kernel/power/swap.c 	bio = bio_alloc(__GFP_WAIT | __GFP_HIGH, 1);
bio                66 kernel/power/swap.c 	if (!bio)
bio                68 kernel/power/swap.c 	bio->bi_sector = page_off * (PAGE_SIZE >> 9);
bio                69 kernel/power/swap.c 	bio->bi_bdev = resume_bdev;
bio                70 kernel/power/swap.c 	bio->bi_end_io = end_swap_bio_read;
bio                72 kernel/power/swap.c 	if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) {
bio                75 kernel/power/swap.c 		bio_put(bio);
bio                80 kernel/power/swap.c 	bio_get(bio);
bio                83 kernel/power/swap.c 		submit_bio(rw | (1 << BIO_RW_SYNC), bio);
bio                86 kernel/power/swap.c 			bio_set_pages_dirty(bio);
bio                87 kernel/power/swap.c 		bio_put(bio);
bio                91 kernel/power/swap.c 		bio->bi_private = *bio_chain;
bio                92 kernel/power/swap.c 		*bio_chain = bio;
bio                93 kernel/power/swap.c 		submit_bio(rw | (1 << BIO_RW_SYNC), bio);
bio               110 kernel/power/swap.c 	struct bio *bio;
bio               111 kernel/power/swap.c 	struct bio *next_bio;
bio               117 kernel/power/swap.c 	bio = *bio_chain;
bio               118 kernel/power/swap.c 	if (bio == NULL)
bio               120 kernel/power/swap.c 	while (bio) {
bio               123 kernel/power/swap.c 		next_bio = bio->bi_private;
bio               124 kernel/power/swap.c 		page = bio->bi_io_vec[0].bv_page;
bio               129 kernel/power/swap.c 		bio_put(bio);
bio               130 kernel/power/swap.c 		bio = next_bio;
bio               322 kernel/power/swap.c 	struct bio *bio;
bio               332 kernel/power/swap.c 	bio = NULL;
bio               338 kernel/power/swap.c 						&bio);
bio               346 kernel/power/swap.c 	err2 = wait_on_bio_chain(&bio);
bio               505 kernel/power/swap.c 	struct bio *bio;
bio               515 kernel/power/swap.c 	bio = NULL;
bio               521 kernel/power/swap.c 		error = swap_read_page(handle, data_of(*snapshot), &bio);
bio               525 kernel/power/swap.c 			error = wait_on_bio_chain(&bio);
bio               532 kernel/power/swap.c 	err2 = wait_on_bio_chain(&bio);
bio               124 mm/bounce.c    	struct bio *bio_orig = bio->bi_private;
bio               128 mm/bounce.c    	if (test_bit(BIO_EOPNOTSUPP, &bio->bi_flags))
bio               134 mm/bounce.c    	__bio_for_each_segment(bvec, bio, i, 0) {
bio               144 mm/bounce.c    	bio_put(bio);
bio               149 mm/bounce.c    	bounce_end_io(bio, page_pool, err);
bio               155 mm/bounce.c    	bounce_end_io(bio, isa_page_pool, err);
bio               160 mm/bounce.c    	struct bio *bio_orig = bio->bi_private;
bio               162 mm/bounce.c    	if (test_bit(BIO_UPTODATE, &bio->bi_flags))
bio               163 mm/bounce.c    		copy_to_high_bio_irq(bio_orig, bio);
bio               165 mm/bounce.c    	bounce_end_io(bio, pool, err);
bio               170 mm/bounce.c    	__bounce_end_io_read(bio, page_pool, err);
bio               175 mm/bounce.c    	__bounce_end_io_read(bio, isa_page_pool, err);
bio               182 mm/bounce.c    	struct bio *bio = NULL;
bio               198 mm/bounce.c    		if (!bio)
bio               199 mm/bounce.c    			bio = bio_alloc(GFP_NOIO, (*bio_orig)->bi_vcnt);
bio               201 mm/bounce.c    		to = bio->bi_io_vec + i;
bio               222 mm/bounce.c    	if (!bio)
bio               232 mm/bounce.c    		to = bio_iovec_idx(bio, i);
bio               240 mm/bounce.c    	bio->bi_bdev = (*bio_orig)->bi_bdev;
bio               241 mm/bounce.c    	bio->bi_flags |= (1 << BIO_BOUNCED);
bio               242 mm/bounce.c    	bio->bi_sector = (*bio_orig)->bi_sector;
bio               243 mm/bounce.c    	bio->bi_rw = (*bio_orig)->bi_rw;
bio               245 mm/bounce.c    	bio->bi_vcnt = (*bio_orig)->bi_vcnt;
bio               246 mm/bounce.c    	bio->bi_idx = (*bio_orig)->bi_idx;
bio               247 mm/bounce.c    	bio->bi_size = (*bio_orig)->bi_size;
bio               250 mm/bounce.c    		bio->bi_end_io = bounce_end_io_write;
bio               252 mm/bounce.c    			bio->bi_end_io = bounce_end_io_read;
bio               254 mm/bounce.c    		bio->bi_end_io = bounce_end_io_write_isa;
bio               256 mm/bounce.c    			bio->bi_end_io = bounce_end_io_read_isa;
bio               259 mm/bounce.c    	bio->bi_private = *bio_orig;
bio               260 mm/bounce.c    	*bio_orig = bio;
bio                22 mm/page_io.c   static struct bio *get_swap_bio(gfp_t gfp_flags, pgoff_t index,
bio                25 mm/page_io.c   	struct bio *bio;
bio                27 mm/page_io.c   	bio = bio_alloc(gfp_flags, 1);
bio                28 mm/page_io.c   	if (bio) {
bio                33 mm/page_io.c   		bio->bi_sector = map_swap_page(sis, swp_offset(entry)) *
bio                35 mm/page_io.c   		bio->bi_bdev = sis->bdev;
bio                36 mm/page_io.c   		bio->bi_io_vec[0].bv_page = page;
bio                37 mm/page_io.c   		bio->bi_io_vec[0].bv_len = PAGE_SIZE;
bio                38 mm/page_io.c   		bio->bi_io_vec[0].bv_offset = 0;
bio                39 mm/page_io.c   		bio->bi_vcnt = 1;
bio                40 mm/page_io.c   		bio->bi_idx = 0;
bio                41 mm/page_io.c   		bio->bi_size = PAGE_SIZE;
bio                42 mm/page_io.c   		bio->bi_end_io = end_io;
bio                44 mm/page_io.c   	return bio;
bio                49 mm/page_io.c   	const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio                50 mm/page_io.c   	struct page *page = bio->bi_io_vec[0].bv_page;
bio                64 mm/page_io.c   				imajor(bio->bi_bdev->bd_inode),
bio                65 mm/page_io.c   				iminor(bio->bi_bdev->bd_inode),
bio                66 mm/page_io.c   				(unsigned long long)bio->bi_sector);
bio                70 mm/page_io.c   	bio_put(bio);
bio                75 mm/page_io.c   	const int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
bio                76 mm/page_io.c   	struct page *page = bio->bi_io_vec[0].bv_page;
bio                82 mm/page_io.c   				imajor(bio->bi_bdev->bd_inode),
bio                83 mm/page_io.c   				iminor(bio->bi_bdev->bd_inode),
bio                84 mm/page_io.c   				(unsigned long long)bio->bi_sector);
bio                89 mm/page_io.c   	bio_put(bio);
bio                98 mm/page_io.c   	struct bio *bio;
bio               105 mm/page_io.c   	bio = get_swap_bio(GFP_NOIO, page_private(page), page,
bio               107 mm/page_io.c   	if (bio == NULL) {
bio               118 mm/page_io.c   	submit_bio(rw, bio);
bio               125 mm/page_io.c   	struct bio *bio;
bio               130 mm/page_io.c   	bio = get_swap_bio(GFP_KERNEL, page_private(page), page,
bio               132 mm/page_io.c   	if (bio == NULL) {
bio               138 mm/page_io.c   	submit_bio(READ, bio);