ioc               185 block/as-iosched.c 	spin_lock_irq(&ioc->lock);
ioc               186 block/as-iosched.c 	if (ioc->aic)
ioc               187 block/as-iosched.c 		free_as_io_context(ioc->aic);
ioc               188 block/as-iosched.c 	ioc->aic = NULL;
ioc               189 block/as-iosched.c 	spin_unlock_irq(&ioc->lock);
ioc               229 block/as-iosched.c 	struct io_context *ioc = get_io_context(GFP_ATOMIC, node);
ioc               230 block/as-iosched.c 	if (ioc && !ioc->aic) {
ioc               231 block/as-iosched.c 		ioc->aic = alloc_as_io_context();
ioc               232 block/as-iosched.c 		if (!ioc->aic) {
ioc               233 block/as-iosched.c 			put_io_context(ioc);
ioc               234 block/as-iosched.c 			ioc = NULL;
ioc               237 block/as-iosched.c 	return ioc;
ioc               652 block/as-iosched.c 	struct io_context *ioc;
ioc               655 block/as-iosched.c 	ioc = ad->io_context;
ioc               656 block/as-iosched.c 	BUG_ON(!ioc);
ioc               657 block/as-iosched.c 	spin_lock(&ioc->lock);
ioc               659 block/as-iosched.c 	if (rq && ioc == RQ_IOC(rq)) {
ioc               661 block/as-iosched.c 		spin_unlock(&ioc->lock);
ioc               670 block/as-iosched.c 		spin_unlock(&ioc->lock);
ioc               674 block/as-iosched.c 	aic = ioc->aic;
ioc               676 block/as-iosched.c 		spin_unlock(&ioc->lock);
ioc               682 block/as-iosched.c 		spin_unlock(&ioc->lock);
ioc               688 block/as-iosched.c 		spin_unlock(&ioc->lock);
ioc               709 block/as-iosched.c 		spin_unlock(&ioc->lock);
ioc               719 block/as-iosched.c 			spin_unlock(&ioc->lock);
ioc               726 block/as-iosched.c 			spin_unlock(&ioc->lock);
ioc               730 block/as-iosched.c 			spin_unlock(&ioc->lock);
ioc               735 block/as-iosched.c 		spin_unlock(&ioc->lock);
ioc               738 block/as-iosched.c 	spin_unlock(&ioc->lock);
ioc               904 block/as-iosched.c 	struct io_context *ioc;
ioc               908 block/as-iosched.c 	ioc = RQ_IOC(rq);
ioc               909 block/as-iosched.c 	if (ioc && ioc->aic) {
ioc               910 block/as-iosched.c 		BUG_ON(!atomic_read(&ioc->aic->nr_queued));
ioc               911 block/as-iosched.c 		atomic_dec(&ioc->aic->nr_queued);
ioc               990 block/as-iosched.c 		struct io_context *ioc = RQ_IOC(rq);
ioc               992 block/as-iosched.c 		copy_io_context(&ad->io_context, &ioc);
ioc              1330 block/as-iosched.c 	struct io_context *ioc;
ioc              1333 block/as-iosched.c 		ioc = as_get_io_context(q->node);
ioc              1334 block/as-iosched.c 		if (ad->io_context == ioc)
ioc              1336 block/as-iosched.c 		put_io_context(ioc);
ioc               652 block/blk-core.c 	if (!ioc)
ioc               660 block/blk-core.c 	return ioc->nr_batch_requests == q->nr_batching ||
ioc               661 block/blk-core.c 		(ioc->nr_batch_requests > 0
ioc               662 block/blk-core.c 		&& time_before(jiffies, ioc->last_waited + BLK_BATCH_TIME));
ioc               673 block/blk-core.c 	if (!ioc || ioc_batching(q, ioc))
ioc               676 block/blk-core.c 	ioc->nr_batch_requests = q->nr_batching;
ioc               677 block/blk-core.c 	ioc->last_waited = jiffies;
ioc               724 block/blk-core.c 	struct io_context *ioc = NULL;
ioc               734 block/blk-core.c 			ioc = current_io_context(GFP_ATOMIC, q->node);
ioc               742 block/blk-core.c 				ioc_set_batching(q, ioc);
ioc               746 block/blk-core.c 						&& !ioc_batching(q, ioc)) {
ioc               808 block/blk-core.c 	if (ioc_batching(q, ioc))
ioc               809 block/blk-core.c 		ioc->nr_batch_requests--;
ioc               831 block/blk-core.c 		struct io_context *ioc;
ioc               849 block/blk-core.c 		ioc = current_io_context(GFP_NOIO, q->node);
ioc               850 block/blk-core.c 		ioc_set_batching(q, ioc);
ioc                20 block/blk-ioc.c 	if (!hlist_empty(&ioc->cic_list)) {
ioc                23 block/blk-ioc.c 		cic = list_entry(ioc->cic_list.first, struct cfq_io_context,
ioc                25 block/blk-ioc.c 		cic->dtor(ioc);
ioc                35 block/blk-ioc.c 	if (ioc == NULL)
ioc                38 block/blk-ioc.c 	BUG_ON(atomic_read(&ioc->refcount) == 0);
ioc                40 block/blk-ioc.c 	if (atomic_dec_and_test(&ioc->refcount)) {
ioc                42 block/blk-ioc.c 		if (ioc->aic && ioc->aic->dtor)
ioc                43 block/blk-ioc.c 			ioc->aic->dtor(ioc->aic);
ioc                44 block/blk-ioc.c 		cfq_dtor(ioc);
ioc                47 block/blk-ioc.c 		kmem_cache_free(iocontext_cachep, ioc);
ioc                58 block/blk-ioc.c 	if (!hlist_empty(&ioc->cic_list)) {
ioc                61 block/blk-ioc.c 		cic = list_entry(ioc->cic_list.first, struct cfq_io_context,
ioc                63 block/blk-ioc.c 		cic->exit(ioc);
ioc                71 block/blk-ioc.c 	struct io_context *ioc;
ioc                74 block/blk-ioc.c 	ioc = current->io_context;
ioc                78 block/blk-ioc.c 	if (atomic_dec_and_test(&ioc->nr_tasks)) {
ioc                79 block/blk-ioc.c 		if (ioc->aic && ioc->aic->exit)
ioc                80 block/blk-ioc.c 			ioc->aic->exit(ioc->aic);
ioc                81 block/blk-ioc.c 		cfq_exit(ioc);
ioc                83 block/blk-ioc.c 		put_io_context(ioc);
ioc               806 block/cfq-iosched.c 		put_io_context(cfqd->active_cic->ioc);
ioc               908 block/cfq-iosched.c 	if (!cic || !atomic_read(&cic->ioc->nr_tasks))
ioc              1063 block/cfq-iosched.c 			atomic_inc(&RQ_CIC(rq)->ioc->refcount);
ioc              1198 block/cfq-iosched.c 	hlist_for_each_entry_rcu(cic, n, &ioc->cic_list, cic_list)
ioc              1199 block/cfq-iosched.c 		func(ioc, cic);
ioc              1210 block/cfq-iosched.c 	__call_for_each_cic(ioc, func);
ioc              1249 block/cfq-iosched.c 	spin_lock_irqsave(&ioc->lock, flags);
ioc              1250 block/cfq-iosched.c 	radix_tree_delete(&ioc->radix_root, cic->dead_key);
ioc              1252 block/cfq-iosched.c 	spin_unlock_irqrestore(&ioc->lock, flags);
ioc              1270 block/cfq-iosched.c 	__call_for_each_cic(ioc, cic_free_func);
ioc              1286 block/cfq-iosched.c 	struct io_context *ioc = cic->ioc;
ioc              1297 block/cfq-iosched.c 	if (ioc->ioc_data == cic)
ioc              1298 block/cfq-iosched.c 		rcu_assign_pointer(ioc->ioc_data, NULL);
ioc              1332 block/cfq-iosched.c 	call_for_each_cic(ioc, cfq_exit_single_io_context);
ioc              1362 block/cfq-iosched.c 	ioprio_class = IOPRIO_PRIO_CLASS(ioc->ioprio);
ioc              1374 block/cfq-iosched.c 		cfqq->ioprio = task_ioprio(ioc);
ioc              1378 block/cfq-iosched.c 		cfqq->ioprio = task_ioprio(ioc);
ioc              1411 block/cfq-iosched.c 		new_cfqq = cfq_get_queue(cfqd, ASYNC, cic->ioc, GFP_ATOMIC);
ioc              1427 block/cfq-iosched.c 	call_for_each_cic(ioc, changed_ioprio);
ioc              1428 block/cfq-iosched.c 	ioc->ioprio_changed = 0;
ioc              1439 block/cfq-iosched.c 	cic = cfq_cic_lookup(cfqd, ioc);
ioc              1477 block/cfq-iosched.c 		cfq_init_prio_data(cfqq, ioc);
ioc              1515 block/cfq-iosched.c 	const int ioprio = task_ioprio(ioc);
ioc              1516 block/cfq-iosched.c 	const int ioprio_class = task_ioprio_class(ioc);
ioc              1526 block/cfq-iosched.c 		cfqq = cfq_find_alloc_queue(cfqd, is_sync, ioc, gfp_mask);
ioc              1554 block/cfq-iosched.c 	spin_lock_irqsave(&ioc->lock, flags);
ioc              1556 block/cfq-iosched.c 	BUG_ON(ioc->ioc_data == cic);
ioc              1558 block/cfq-iosched.c 	radix_tree_delete(&ioc->radix_root, (unsigned long) cfqd);
ioc              1560 block/cfq-iosched.c 	spin_unlock_irqrestore(&ioc->lock, flags);
ioc              1572 block/cfq-iosched.c 	if (unlikely(!ioc))
ioc              1580 block/cfq-iosched.c 	cic = rcu_dereference(ioc->ioc_data);
ioc              1587 block/cfq-iosched.c 		cic = radix_tree_lookup(&ioc->radix_root, (unsigned long) cfqd);
ioc              1594 block/cfq-iosched.c 			cfq_drop_dead_cic(cfqd, ioc, cic);
ioc              1599 block/cfq-iosched.c 		spin_lock_irqsave(&ioc->lock, flags);
ioc              1600 block/cfq-iosched.c 		rcu_assign_pointer(ioc->ioc_data, cic);
ioc              1601 block/cfq-iosched.c 		spin_unlock_irqrestore(&ioc->lock, flags);
ioc              1621 block/cfq-iosched.c 		cic->ioc = ioc;
ioc              1624 block/cfq-iosched.c 		spin_lock_irqsave(&ioc->lock, flags);
ioc              1625 block/cfq-iosched.c 		ret = radix_tree_insert(&ioc->radix_root,
ioc              1628 block/cfq-iosched.c 			hlist_add_head_rcu(&cic->cic_list, &ioc->cic_list);
ioc              1629 block/cfq-iosched.c 		spin_unlock_irqrestore(&ioc->lock, flags);
ioc              1654 block/cfq-iosched.c 	struct io_context *ioc = NULL;
ioc              1659 block/cfq-iosched.c 	ioc = get_io_context(gfp_mask, cfqd->queue->node);
ioc              1660 block/cfq-iosched.c 	if (!ioc)
ioc              1663 block/cfq-iosched.c 	cic = cfq_cic_lookup(cfqd, ioc);
ioc              1671 block/cfq-iosched.c 	if (cfq_cic_link(cfqd, ioc, cic, gfp_mask))
ioc              1676 block/cfq-iosched.c 	if (unlikely(ioc->ioprio_changed))
ioc              1677 block/cfq-iosched.c 		cfq_ioc_set_ioprio(ioc);
ioc              1683 block/cfq-iosched.c 	put_io_context(ioc);
ioc              1744 block/cfq-iosched.c 	if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle ||
ioc              1883 block/cfq-iosched.c 	cfq_init_prio_data(cfqq, RQ_CIC(rq)->ioc);
ioc              2018 block/cfq-iosched.c 		cfq_init_prio_data(cfqq, cic->ioc);
ioc              2040 block/cfq-iosched.c 		put_io_context(RQ_CIC(rq)->ioc);
ioc              2073 block/cfq-iosched.c 		cfqq = cfq_get_queue(cfqd, is_sync, cic->ioc, gfp_mask);
ioc              2093 block/cfq-iosched.c 		put_io_context(cic->ioc);
ioc                33 fs/ioprio.c    	struct io_context *ioc;
ioc                45 fs/ioprio.c    		ioc = task->io_context;
ioc                48 fs/ioprio.c    		if (ioc)
ioc                51 fs/ioprio.c    		ioc = alloc_io_context(GFP_ATOMIC, -1);
ioc                52 fs/ioprio.c    		if (!ioc) {
ioc                56 fs/ioprio.c    		task->io_context = ioc;
ioc                60 fs/ioprio.c    		ioc->ioprio = ioprio;
ioc                61 fs/ioprio.c    		ioc->ioprio_changed = 1;
ioc               230 include/asm-parisc/dma-mapping.h struct ioc;
ioc                85 include/asm-parisc/ropes.h 	struct ioc		ioc[MAX_IOC];
ioc                40 include/linux/iocontext.h 	struct io_context *ioc;
ioc                94 include/linux/iocontext.h 	if (ioc && atomic_inc_not_zero(&ioc->refcount)) {
ioc                95 include/linux/iocontext.h 		atomic_inc(&ioc->nr_tasks);
ioc                96 include/linux/iocontext.h 		return ioc;
ioc                51 include/linux/ioprio.h 	if (ioprio_valid(ioc->ioprio))
ioc                52 include/linux/ioprio.h 		return IOPRIO_PRIO_DATA(ioc->ioprio);
ioc                59 include/linux/ioprio.h 	if (ioprio_valid(ioc->ioprio))
ioc                60 include/linux/ioprio.h 		return IOPRIO_PRIO_CLASS(ioc->ioprio);
ioc               717 kernel/fork.c  	struct io_context *ioc = current->io_context;
ioc               719 kernel/fork.c  	if (!ioc)
ioc               725 kernel/fork.c  		tsk->io_context = ioc_task_link(ioc);
ioc               728 kernel/fork.c  	} else if (ioprio_valid(ioc->ioprio)) {
ioc               733 kernel/fork.c  		tsk->io_context->ioprio = ioc->ioprio;
ioc               839 net/rxrpc/ar-ack.c 	int genbit, loop, nbit, ioc, ret, mtu;
ioc              1197 net/rxrpc/ar-ack.c 	ioc = 1;
ioc              1199 net/rxrpc/ar-ack.c 		ioc = 5;
ioc              1205 net/rxrpc/ar-ack.c 		ioc = 4;
ioc              1210 net/rxrpc/ar-ack.c 		ioc = 3;
ioc              1214 net/rxrpc/ar-ack.c 		ioc = 2;
ioc              1219 net/rxrpc/ar-ack.c 			     &msg, iov, ioc, len);
ioc               533 net/rxrpc/ar-output.c 	int ret, ioc, segment, copied;
ioc               546 net/rxrpc/ar-output.c 	ioc = msg->msg_iovlen - 1;
ioc               648 net/rxrpc/ar-output.c 		while (segment == 0 && ioc > 0) {
ioc               652 net/rxrpc/ar-output.c 			ioc--;
ioc               656 net/rxrpc/ar-output.c 			ioc = 0;