sg                426 arch/x86/kernel/pci-calgary_64.c 	for_each_sg(sg, s, nelems, i) {
sg                450 arch/x86/kernel/pci-calgary_64.c 	calgary_unmap_sg(dev, sg, nelems, direction);
sg                451 arch/x86/kernel/pci-calgary_64.c 	for_each_sg(sg, s, nelems, i) {
sg                452 arch/x86/kernel/pci-calgary_64.c 		sg->dma_address = bad_dma_address;
sg                453 arch/x86/kernel/pci-calgary_64.c 		sg->dma_length = 0;
sg                305 arch/x86/kernel/pci-gart_64.c 	for_each_sg(sg, s, nents, i) {
sg                323 arch/x86/kernel/pci-gart_64.c 	for_each_sg(sg, s, nents, i) {
sg                330 arch/x86/kernel/pci-gart_64.c 					gart_unmap_sg(dev, sg, i, dir);
sg                332 arch/x86/kernel/pci-gart_64.c 				sg[0].dma_length = 0;
sg                418 arch/x86/kernel/pci-gart_64.c 	start_sg = sgmap = sg;
sg                422 arch/x86/kernel/pci-gart_64.c 	for_each_sg(sg, s, nents, i) {
sg                469 arch/x86/kernel/pci-gart_64.c 	gart_unmap_sg(dev, sg, out, dir);
sg                473 arch/x86/kernel/pci-gart_64.c 		out = dma_map_sg_nonforce(dev, sg, nents, dir);
sg                481 arch/x86/kernel/pci-gart_64.c 	for_each_sg(sg, s, nents, i)
sg                 62 arch/x86/kernel/pci-nommu.c 	WARN_ON(nents == 0 || sg[0].length == 0);
sg                 64 arch/x86/kernel/pci-nommu.c 	for_each_sg(sg, s, nents, i) {
sg                 73 block/blk-integrity.c 	struct scatterlist *sg;
sg                 77 block/blk-integrity.c 	sg = NULL;
sg                 86 block/blk-integrity.c 			sg->length += iv->bv_len;
sg                 89 block/blk-integrity.c 			if (!sg)
sg                 90 block/blk-integrity.c 				sg = sglist;
sg                 92 block/blk-integrity.c 				sg->page_link &= ~0x02;
sg                 93 block/blk-integrity.c 				sg = sg_next(sg);
sg                 96 block/blk-integrity.c 			sg_set_page(sg, iv->bv_page, iv->bv_len, iv->bv_offset);
sg                103 block/blk-integrity.c 	if (sg)
sg                104 block/blk-integrity.c 		sg_mark_end(sg);
sg                137 block/blk-merge.c 	struct scatterlist *sg;
sg                147 block/blk-merge.c 	sg = NULL;
sg                152 block/blk-merge.c 			if (sg->length + nbytes > q->max_segment_size)
sg                160 block/blk-merge.c 			sg->length += nbytes;
sg                163 block/blk-merge.c 			if (!sg)
sg                164 block/blk-merge.c 				sg = sglist;
sg                176 block/blk-merge.c 				sg->page_link &= ~0x02;
sg                177 block/blk-merge.c 				sg = sg_next(sg);
sg                180 block/blk-merge.c 			sg_set_page(sg, bvec->bv_page, nbytes, bvec->bv_offset);
sg                191 block/blk-merge.c 		sg->length += pad_len;
sg                199 block/blk-merge.c 		sg->page_link &= ~0x02;
sg                200 block/blk-merge.c 		sg = sg_next(sg);
sg                201 block/blk-merge.c 		sg_set_page(sg, virt_to_page(q->dma_drain_buffer),
sg                209 block/blk-merge.c 	if (sg)
sg                210 block/blk-merge.c 		sg_mark_end(sg);
sg                 46 crypto/ahash.c 	struct scatterlist *sg;
sg                 48 crypto/ahash.c 	sg = walk->sg;
sg                 49 crypto/ahash.c 	walk->pg = sg_page(sg);
sg                 50 crypto/ahash.c 	walk->offset = sg->offset;
sg                 51 crypto/ahash.c 	walk->entrylen = sg->length;
sg                 93 crypto/ahash.c 	walk->sg = scatterwalk_sg_next(walk->sg);
sg                108 crypto/ahash.c 	walk->sg = req->src;
sg                 95 crypto/authenc.c 		head->length += sg->length;
sg                 96 crypto/authenc.c 		sg = scatterwalk_sg_next(sg);
sg                 99 crypto/authenc.c 	if (sg)
sg                100 crypto/authenc.c 		scatterwalk_sg_chain(head, 2, sg);
sg                342 crypto/blkcipher.c 	scatterwalk_start(&walk->in, walk->in.sg);
sg                343 crypto/blkcipher.c 	scatterwalk_start(&walk->out, walk->out.sg);
sg                211 crypto/ccm.c   	scatterwalk_start(&walk, sg);
sg                216 crypto/ccm.c   			scatterwalk_start(&walk, sg_next(walk.sg));
sg                 45 crypto/digest.c 		struct page *pg = sg_page(sg);
sg                 46 crypto/digest.c 		unsigned int offset = sg->offset;
sg                 47 crypto/digest.c 		unsigned int l = sg->length;
sg                 81 crypto/digest.c 		sg = scatterwalk_sg_next(sg);
sg                 92 crypto/digest.c 	return update2(desc, sg, nbytes);
sg                136 crypto/digest.c 	update2(desc, sg, nbytes);
sg                 69 crypto/eseqiv.c 		head->length += sg->length;
sg                 70 crypto/eseqiv.c 		sg = scatterwalk_sg_next(sg);
sg                 73 crypto/eseqiv.c 	if (sg)
sg                 74 crypto/eseqiv.c 		scatterwalk_sg_chain(head, 2, sg);
sg                117 crypto/gcm.c   	scatterwalk_start(&walk, sg);
sg                123 crypto/gcm.c   			scatterwalk_start(&walk, scatterwalk_sg_next(walk.sg));
sg                194 crypto/gcm.c   		struct scatterlist sg[1];
sg                216 crypto/gcm.c   	sg_init_one(data->sg, &data->hash, sizeof(data->hash));
sg                222 crypto/gcm.c   	ablkcipher_request_set_crypt(&data->req, data->sg, data->sg,
sg                139 crypto/hmac.c  	return crypto_hash_update(&desc, sg, nbytes);
sg                185 crypto/hmac.c  	scatterwalk_sg_chain(sg1, 2, sg);
sg                 35 crypto/scatterwalk.c 	walk->sg = sg;
sg                 37 crypto/scatterwalk.c 	BUG_ON(!sg->length);
sg                 39 crypto/scatterwalk.c 	walk->offset = sg->offset;
sg                 56 crypto/scatterwalk.c 		page = sg_page(walk->sg) + ((walk->offset - 1) >> PAGE_SHIFT);
sg                 63 crypto/scatterwalk.c 		if (walk->offset >= walk->sg->offset + walk->sg->length)
sg                 64 crypto/scatterwalk.c 			scatterwalk_start(walk, scatterwalk_sg_next(walk->sg));
sg                112 crypto/scatterwalk.c 		scatterwalk_start(&walk, sg);
sg                114 crypto/scatterwalk.c 		if (start < offset + sg->length)
sg                117 crypto/scatterwalk.c 		offset += sg->length;
sg                118 crypto/scatterwalk.c 		sg = scatterwalk_sg_next(sg);
sg                 69 crypto/tcrypt.c 			ret = crypto_blkcipher_encrypt(desc, sg, sg, blen);
sg                 71 crypto/tcrypt.c 			ret = crypto_blkcipher_decrypt(desc, sg, sg, blen);
sg                 95 crypto/tcrypt.c 			ret = crypto_blkcipher_encrypt(desc, sg, sg, blen);
sg                 97 crypto/tcrypt.c 			ret = crypto_blkcipher_decrypt(desc, sg, sg, blen);
sg                109 crypto/tcrypt.c 			ret = crypto_blkcipher_encrypt(desc, sg, sg, blen);
sg                111 crypto/tcrypt.c 			ret = crypto_blkcipher_decrypt(desc, sg, sg, blen);
sg                166 crypto/tcrypt.c 			struct scatterlist sg[TVMEMSIZE];
sg                196 crypto/tcrypt.c 			sg_init_table(sg, TVMEMSIZE);
sg                197 crypto/tcrypt.c 			sg_set_buf(sg, tvmem[0] + *keysize,
sg                200 crypto/tcrypt.c 				sg_set_buf(sg + j, tvmem[j], PAGE_SIZE);
sg                211 crypto/tcrypt.c 				ret = test_cipher_jiffies(&desc, enc, sg,
sg                214 crypto/tcrypt.c 				ret = test_cipher_cycles(&desc, enc, sg,
sg                241 crypto/tcrypt.c 		ret = crypto_hash_digest(desc, sg, blen, out);
sg                260 crypto/tcrypt.c 		return test_hash_jiffies_digest(desc, sg, blen, out, sec);
sg                268 crypto/tcrypt.c 			ret = crypto_hash_update(desc, sg, plen);
sg                296 crypto/tcrypt.c 		ret = crypto_hash_digest(desc, sg, blen, out);
sg                307 crypto/tcrypt.c 		ret = crypto_hash_digest(desc, sg, blen, out);
sg                337 crypto/tcrypt.c 		return test_hash_cycles_digest(desc, sg, blen, out);
sg                348 crypto/tcrypt.c 			ret = crypto_hash_update(desc, sg, plen);
sg                367 crypto/tcrypt.c 			ret = crypto_hash_update(desc, sg, plen);
sg                396 crypto/tcrypt.c 	struct scatterlist sg[TVMEMSIZE];
sg                422 crypto/tcrypt.c 	sg_init_table(sg, TVMEMSIZE);
sg                424 crypto/tcrypt.c 		sg_set_buf(sg + i, tvmem[i], PAGE_SIZE);
sg                439 crypto/tcrypt.c 			ret = test_hash_jiffies(&desc, sg, speed[i].blen,
sg                442 crypto/tcrypt.c 			ret = test_hash_cycles(&desc, sg, speed[i].blen,
sg                121 crypto/testmgr.c 	struct scatterlist sg[8];
sg                146 crypto/testmgr.c 		sg_init_one(&sg[0], hash_buff, template[i].psize);
sg                160 crypto/testmgr.c 		ahash_request_set_crypt(req, sg, result, template[i].psize);
sg                197 crypto/testmgr.c 			sg_init_table(sg, template[i].np);
sg                199 crypto/testmgr.c 				sg_set_buf(&sg[k],
sg                222 crypto/testmgr.c 			ahash_request_set_crypt(req, sg, result,
sg                272 crypto/testmgr.c 	struct scatterlist sg[8];
sg                342 crypto/testmgr.c 			sg_init_one(&sg[0], input,
sg                347 crypto/testmgr.c 			aead_request_set_crypt(req, sg, sg,
sg                412 crypto/testmgr.c 			sg_init_table(sg, template[i].np);
sg                430 crypto/testmgr.c 				sg_set_buf(&sg[k], q, template[i].tap[k]);
sg                443 crypto/testmgr.c 				if (WARN_ON(sg[k - 1].offset +
sg                444 crypto/testmgr.c 					    sg[k - 1].length + authsize >
sg                450 crypto/testmgr.c 				sg[k - 1].length += authsize;
sg                464 crypto/testmgr.c 			aead_request_set_crypt(req, sg, sg,
sg                618 crypto/testmgr.c 	struct scatterlist sg[8];
sg                670 crypto/testmgr.c 			sg_init_one(&sg[0], data, template[i].ilen);
sg                672 crypto/testmgr.c 			ablkcipher_request_set_crypt(req, sg, sg,
sg                737 crypto/testmgr.c 			sg_init_table(sg, template[i].np);
sg                753 crypto/testmgr.c 				sg_set_buf(&sg[k], q, template[i].tap[k]);
sg                758 crypto/testmgr.c 			ablkcipher_request_set_crypt(req, sg, sg,
sg                500 crypto/twofish_common.c    ctx->s[2][i] = mds[2][q1[(a) ^ sc] ^ sg]; \
sg                508 crypto/twofish_common.c    ctx->s[2][i] = mds[2][q1[q0[(a) ^ sc] ^ sg] ^ sk]; \
sg                516 crypto/twofish_common.c    ctx->s[2][i] = mds[2][q1[q0[q0[(a) ^ sc] ^ sg] ^ sk] ^ so]; \
sg                596 crypto/twofish_common.c 	u8 sa = 0, sb = 0, sc = 0, sd = 0, se = 0, sf = 0, sg = 0, sh = 0;
sg                621 crypto/twofish_common.c 	CALC_S (se, sf, sg, sh, 8, 0x00, 0x2D, 0x01, 0x2D); /* 01 A4 02 A4 */
sg                622 crypto/twofish_common.c 	CALC_S (se, sf, sg, sh, 9, 0x2D, 0xA4, 0x44, 0x8A); /* A4 56 A1 55 */
sg                623 crypto/twofish_common.c 	CALC_S (se, sf, sg, sh, 10, 0x8A, 0xD5, 0xBF, 0xD1); /* 55 82 FC 87 */
sg                624 crypto/twofish_common.c 	CALC_S (se, sf, sg, sh, 11, 0xD1, 0x7F, 0x3D, 0x99); /* 87 F3 C1 5A */
sg                625 crypto/twofish_common.c 	CALC_S (se, sf, sg, sh, 12, 0x99, 0x46, 0x66, 0x96); /* 5A 1E 47 58 */
sg                626 crypto/twofish_common.c 	CALC_S (se, sf, sg, sh, 13, 0x96, 0x3C, 0x5B, 0xED); /* 58 C6 AE DB */
sg                627 crypto/twofish_common.c 	CALC_S (se, sf, sg, sh, 14, 0xED, 0x37, 0x4F, 0xE0); /* DB 68 3D 9E */
sg                628 crypto/twofish_common.c 	CALC_S (se, sf, sg, sh, 15, 0xE0, 0xD0, 0x8C, 0x17); /* 9E E5 19 03 */
sg                121 crypto/xcbc.c  		struct page *pg = sg_page(sg);
sg                122 crypto/xcbc.c  		unsigned int offset = sg->offset;
sg                123 crypto/xcbc.c  		unsigned int slen = sg->length;
sg                186 crypto/xcbc.c  		sg = scatterwalk_sg_next(sg);
sg                198 crypto/xcbc.c  	return crypto_xcbc_digest_update2(pdesc, sg, nbytes);
sg                258 crypto/xcbc.c  	crypto_xcbc_digest_update2(pdesc, sg, nbytes);
sg                 98 fs/ecryptfs/crypto.c 	struct scatterlist sg;
sg                106 fs/ecryptfs/crypto.c 	sg_init_one(&sg, (u8 *)src, len);
sg                126 fs/ecryptfs/crypto.c 	rc = crypto_hash_update(&desc, &sg, len);
sg                301 fs/ecryptfs/crypto.c 	sg_init_table(sg, sg_size);
sg                306 fs/ecryptfs/crypto.c 		if (sg)
sg                307 fs/ecryptfs/crypto.c 			sg_set_page(&sg[i], pg, 0, offset);
sg                310 fs/ecryptfs/crypto.c 			if (sg)
sg                311 fs/ecryptfs/crypto.c 				sg[i].length = remainder_of_page;
sg                315 fs/ecryptfs/crypto.c 			if (sg)
sg                316 fs/ecryptfs/crypto.c 				sg[i].length = size;
sg               2102 fs/ext4/mballoc.c 	} sg;
sg               2113 fs/ext4/mballoc.c 	i = (sb->s_blocksize_bits + 2) * sizeof(sg.info.bb_counters[0]) +
sg               2121 fs/ext4/mballoc.c 	memcpy(&sg, ext4_get_group_info(sb, group), i);
sg               2125 fs/ext4/mballoc.c 	seq_printf(seq, "#%-5lu: %-5u %-5u %-5u [", group, sg.info.bb_free,
sg               2126 fs/ext4/mballoc.c 			sg.info.bb_fragments, sg.info.bb_first_free);
sg               2129 fs/ext4/mballoc.c 				sg.info.bb_counters[i] : 0);
sg                 92 fs/nfsd/nfs4recover.c 	struct scatterlist sg;
sg                106 fs/nfsd/nfs4recover.c 	sg_init_one(&sg, clname->data, clname->len);
sg                108 fs/nfsd/nfs4recover.c 	if (crypto_hash_digest(&desc, &sg, sg.length, cksum.data))
sg                 17 include/asm-cris/scatterlist.h #define sg_dma_address(sg)	((sg)->address)
sg                 18 include/asm-cris/scatterlist.h #define sg_dma_len(sg)		((sg)->length)
sg                113 include/asm-frv/pci.h 		frv_cache_wback_inv(sg_dma_address(&sg[i]),
sg                114 include/asm-frv/pci.h 				    sg_dma_address(&sg[i])+sg_dma_len(&sg[i]));
sg                 41 include/asm-frv/scatterlist.h #define sg_dma_address(sg)	((sg)->dma_address)
sg                 42 include/asm-frv/scatterlist.h #define sg_dma_len(sg)		((sg)->length)
sg                 96 include/asm-generic/dma-mapping.h 	return pci_map_sg(to_pci_dev(dev), sg, nents, (int)direction);
sg                105 include/asm-generic/dma-mapping.h 	pci_unmap_sg(to_pci_dev(dev), sg, nhwentries, (int)direction);
sg                134 include/asm-generic/dma-mapping.h 	pci_dma_sync_sg_for_cpu(to_pci_dev(dev), sg, nelems, (int)direction);
sg                143 include/asm-generic/dma-mapping.h 	pci_dma_sync_sg_for_device(to_pci_dev(dev), sg, nelems, (int)direction);
sg                 63 include/asm-generic/pci-dma-compat.h 	return dma_map_sg(hwdev == NULL ? NULL : &hwdev->dev, sg, nents, (enum dma_data_direction)direction);
sg                 70 include/asm-generic/pci-dma-compat.h 	dma_unmap_sg(hwdev == NULL ? NULL : &hwdev->dev, sg, nents, (enum dma_data_direction)direction);
sg                 91 include/asm-generic/pci-dma-compat.h 	dma_sync_sg_for_cpu(hwdev == NULL ? NULL : &hwdev->dev, sg, nelems, (enum dma_data_direction)direction);
sg                 98 include/asm-generic/pci-dma-compat.h 	dma_sync_sg_for_device(hwdev == NULL ? NULL : &hwdev->dev, sg, nelems, (enum dma_data_direction)direction);
sg                 20 include/asm-m68k/scatterlist.h #define sg_dma_address(sg)	((sg)->dma_address)
sg                 21 include/asm-m68k/scatterlist.h #define sg_dma_len(sg)		((sg)->length)
sg                 78 include/asm-mn10300/dma-mapping.h 	struct scatterlist *sg;
sg                 84 include/asm-mn10300/dma-mapping.h 	for_each_sg(sglist, sg, nents, i) {
sg                 85 include/asm-mn10300/dma-mapping.h 		BUG_ON(!sg_page(sg));
sg                 87 include/asm-mn10300/dma-mapping.h 		sg->dma_address = sg_phys(sg);
sg                 52 include/asm-mn10300/scatterlist.h #define sg_dma_address(sg)	((sg)->dma_address)
sg                 53 include/asm-mn10300/scatterlist.h #define sg_dma_len(sg)		((sg)->length)
sg                 16 include/asm-parisc/dma-mapping.h 	int  (*map_sg)(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction);
sg                 17 include/asm-parisc/dma-mapping.h 	void (*unmap_sg)(struct device *dev, struct scatterlist *sg, int nhwents, enum dma_data_direction direction);
sg                 20 include/asm-parisc/dma-mapping.h 	void (*dma_sync_sg_for_cpu)(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction);
sg                 21 include/asm-parisc/dma-mapping.h 	void (*dma_sync_sg_for_device)(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction);
sg                 95 include/asm-parisc/dma-mapping.h 	return hppa_dma_ops->map_sg(dev, sg, nents, direction);
sg                102 include/asm-parisc/dma-mapping.h 	hppa_dma_ops->unmap_sg(dev, sg, nhwentries, direction);
sg                159 include/asm-parisc/dma-mapping.h 		hppa_dma_ops->dma_sync_sg_for_cpu(dev, sg, nelems, direction);
sg                167 include/asm-parisc/dma-mapping.h 		hppa_dma_ops->dma_sync_sg_for_device(dev, sg, nelems, direction);
sg                 21 include/asm-parisc/scatterlist.h #define sg_virt_addr(sg) ((unsigned long)sg_virt(sg))
sg                 22 include/asm-parisc/scatterlist.h #define sg_dma_address(sg) ((sg)->iova)
sg                 23 include/asm-parisc/scatterlist.h #define sg_dma_len(sg)     ((sg)->iova_length)
sg                 43 include/asm-x86/dma-mapping.h 				struct scatterlist *sg, int nelems,
sg                 46 include/asm-x86/dma-mapping.h 				struct scatterlist *sg, int nelems,
sg                 48 include/asm-x86/dma-mapping.h 	int             (*map_sg)(struct device *hwdev, struct scatterlist *sg,
sg                 51 include/asm-x86/dma-mapping.h 				struct scatterlist *sg, int nents,
sg                123 include/asm-x86/dma-mapping.h 	return ops->map_sg(hwdev, sg, nents, direction);
sg                134 include/asm-x86/dma-mapping.h 		ops->unmap_sg(hwdev, sg, nents, direction);
sg                196 include/asm-x86/dma-mapping.h 		ops->sync_sg_for_cpu(hwdev, sg, nelems, direction);
sg                208 include/asm-x86/dma-mapping.h 		ops->sync_sg_for_device(hwdev, sg, nelems, direction);
sg                 26 include/asm-x86/scatterlist.h #define sg_dma_address(sg)	((sg)->dma_address)
sg                 28 include/asm-x86/scatterlist.h # define sg_dma_len(sg)		((sg)->length)
sg                 30 include/asm-x86/scatterlist.h # define sg_dma_len(sg)		((sg)->dma_length)
sg                 60 include/asm-xtensa/dma-mapping.h 	for (i = 0; i < nents; i++, sg++ ) {
sg                 61 include/asm-xtensa/dma-mapping.h 		BUG_ON(!sg_page(sg));
sg                 63 include/asm-xtensa/dma-mapping.h 		sg->dma_address = sg_phys(sg);
sg                 64 include/asm-xtensa/dma-mapping.h 		consistent_sync(sg_virt(sg), sg->length, direction);
sg                129 include/asm-xtensa/dma-mapping.h 	for (i = 0; i < nelems; i++, sg++)
sg                130 include/asm-xtensa/dma-mapping.h 		consistent_sync(sg_virt(sg), sg->length, dir);
sg                138 include/asm-xtensa/dma-mapping.h 	for (i = 0; i < nelems; i++, sg++)
sg                139 include/asm-xtensa/dma-mapping.h 		consistent_sync(sg_virt(sg), sg->length, dir);
sg                 33 include/asm-xtensa/scatterlist.h #define sg_dma_address(sg)      ((sg)->dma_address)
sg                 34 include/asm-xtensa/scatterlist.h #define sg_dma_len(sg)          ((sg)->length)
sg                 66 include/crypto/algapi.h 	struct scatterlist *sg;
sg                252 include/crypto/algapi.h 	walk->in.sg = src;
sg                253 include/crypto/algapi.h 	walk->out.sg = dst;
sg                 32 include/crypto/internal/hash.h 	struct scatterlist *sg;
sg                 65 include/crypto/scatterwalk.h 	if (sg_is_last(sg))
sg                 68 include/crypto/scatterwalk.h 	return (++sg)->length ? sg : (void *)sg_page(sg);
sg                 74 include/crypto/scatterwalk.h 	return !(((sg_page(walk_in->sg) - sg_page(walk_out->sg)) << PAGE_SHIFT) +
sg                 80 include/crypto/scatterwalk.h 	unsigned int len = walk->sg->offset + walk->sg->length - walk->offset;
sg                106 include/crypto/scatterwalk.h 	return sg_page(walk->sg) + (walk->offset >> PAGE_SHIFT);
sg                759 include/drm/drmP.h 	struct drm_sg_mem *sg;	/**< Scatter gather memory */
sg                285 include/linux/crypto.h 	int (*update)(struct hash_desc *desc, struct scatterlist *sg,
sg                288 include/linux/crypto.h 	int (*digest)(struct hash_desc *desc, struct scatterlist *sg,
sg                424 include/linux/crypto.h 		      struct scatterlist *sg, unsigned int nsg);
sg                426 include/linux/crypto.h 	int (*digest)(struct hash_desc *desc, struct scatterlist *sg,
sg               1254 include/linux/crypto.h 	return crypto_hash_crt(desc->tfm)->update(desc, sg, nbytes);
sg               1266 include/linux/crypto.h 	return crypto_hash_crt(desc->tfm)->digest(desc, sg, nbytes, out);
sg                814 include/linux/i2o.h 	sg_count = dma_map_sg(&c->pdev->dev, sg, sg_count, direction);
sg                828 include/linux/i2o.h 		*mptr++ = cpu_to_le32(sg_flags | sg_dma_len(sg));
sg                829 include/linux/i2o.h 		*mptr++ = cpu_to_le32(i2o_dma_low(sg_dma_address(sg)));
sg                832 include/linux/i2o.h 			*mptr++ = cpu_to_le32(i2o_dma_high(sg_dma_address(sg)));
sg                834 include/linux/i2o.h 		sg = sg_next(sg);
sg                447 include/linux/ide.h 	struct scatterlist *sg;
sg                533 include/linux/libata.h 	struct scatterlist	*sg;
sg               1365 include/linux/libata.h 	qc->sg = NULL;
sg                247 include/linux/memstick.h 		struct scatterlist sg;
sg                117 include/linux/mmc/core.h 	struct scatterlist	*sg;		/* I/O scatter list */
sg                 40 include/linux/scatterlist.h #define sg_is_chain(sg)		((sg)->page_link & 0x01)
sg                 41 include/linux/scatterlist.h #define sg_is_last(sg)		((sg)->page_link & 0x02)
sg                 43 include/linux/scatterlist.h 	((struct scatterlist *) ((sg)->page_link & ~0x03))
sg                 57 include/linux/scatterlist.h 	unsigned long page_link = sg->page_link & 0x3;
sg                 65 include/linux/scatterlist.h 	BUG_ON(sg->sg_magic != SG_MAGIC);
sg                 66 include/linux/scatterlist.h 	BUG_ON(sg_is_chain(sg));
sg                 68 include/linux/scatterlist.h 	sg->page_link = page_link | (unsigned long) page;
sg                 88 include/linux/scatterlist.h 	sg_assign_page(sg, page);
sg                 89 include/linux/scatterlist.h 	sg->offset = offset;
sg                 90 include/linux/scatterlist.h 	sg->length = len;
sg                 96 include/linux/scatterlist.h 	BUG_ON(sg->sg_magic != SG_MAGIC);
sg                 97 include/linux/scatterlist.h 	BUG_ON(sg_is_chain(sg));
sg                 99 include/linux/scatterlist.h 	return (struct page *)((sg)->page_link & ~0x3);
sg                112 include/linux/scatterlist.h 	sg_set_page(sg, virt_to_page(buf), buflen, offset_in_page(buf));
sg                119 include/linux/scatterlist.h 	for (__i = 0, sg = (sglist); __i < (nr); __i++, sg = sg_next(sg))
sg                163 include/linux/scatterlist.h 	BUG_ON(sg->sg_magic != SG_MAGIC);
sg                168 include/linux/scatterlist.h 	sg->page_link |= 0x02;
sg                169 include/linux/scatterlist.h 	sg->page_link &= ~0x01;
sg                184 include/linux/scatterlist.h 	return page_to_phys(sg_page(sg)) + sg->offset;
sg                199 include/linux/scatterlist.h 	return page_address(sg_page(sg)) + sg->offset;
sg               1581 include/linux/usb.h 	struct scatterlist	*sg;
sg                 60 include/linux/virtio.h 		       struct scatterlist sg[],
sg                951 include/rdma/ib_verbs.h 				  struct scatterlist *sg, int nents,
sg                954 include/rdma/ib_verbs.h 				    struct scatterlist *sg, int nents,
sg                957 include/rdma/ib_verbs.h 				       struct scatterlist *sg);
sg                959 include/rdma/ib_verbs.h 				   struct scatterlist *sg);
sg               1695 include/rdma/ib_verbs.h 		return dev->dma_ops->map_sg(dev, sg, nents, direction);
sg               1696 include/rdma/ib_verbs.h 	return dma_map_sg(dev->dma_device, sg, nents, direction);
sg               1711 include/rdma/ib_verbs.h 		dev->dma_ops->unmap_sg(dev, sg, nents, direction);
sg               1713 include/rdma/ib_verbs.h 		dma_unmap_sg(dev->dma_device, sg, nents, direction);
sg               1721 include/rdma/ib_verbs.h 	return dma_map_sg_attrs(dev->dma_device, sg, nents, direction, attrs);
sg               1729 include/rdma/ib_verbs.h 	dma_unmap_sg_attrs(dev->dma_device, sg, nents, direction, attrs);
sg               1740 include/rdma/ib_verbs.h 		return dev->dma_ops->dma_address(dev, sg);
sg               1741 include/rdma/ib_verbs.h 	return sg_dma_address(sg);
sg               1753 include/rdma/ib_verbs.h 		return dev->dma_ops->dma_len(dev, sg);
sg               1754 include/rdma/ib_verbs.h 	return sg_dma_len(sg);
sg                180 include/scsi/scsi_cmnd.h 	for_each_sg(scsi_sglist(cmd), sg, nseg, __i)
sg                292 include/scsi/scsi_cmnd.h 	for_each_sg(scsi_prot_sglist(cmd), sg, nseg, __i)
sg                986 include/sound/pcm.h 	struct snd_sg_buf *sg = snd_pcm_substream_sgbuf(substream);
sg                987 include/sound/pcm.h 	return snd_sgbuf_get_addr(sg, ofs);
sg                993 include/sound/pcm.h 	struct snd_sg_buf *sg = snd_pcm_substream_sgbuf(substream);
sg                994 include/sound/pcm.h 	return snd_sgbuf_get_ptr(sg, ofs);
sg                126 kernel/sched.c 	return reciprocal_divide(load, sg->reciprocal_cpu_power);
sg                135 kernel/sched.c 	sg->__cpu_power += val;
sg                136 kernel/sched.c 	sg->reciprocal_cpu_power = reciprocal_value(sg->__cpu_power);
sg               6941 kernel/sched.c 		struct sched_group *sg;
sg               6942 kernel/sched.c 		int group = group_fn(i, cpu_map, &sg, tmpmask);
sg               6948 kernel/sched.c 		cpus_clear(sg->cpumask);
sg               6949 kernel/sched.c 		sg->__cpu_power = 0;
sg               6956 kernel/sched.c 			cpu_set(j, sg->cpumask);
sg               6959 kernel/sched.c 			first = sg;
sg               6961 kernel/sched.c 			last->next = sg;
sg               6962 kernel/sched.c 		last = sg;
sg               7054 kernel/sched.c 	if (sg)
sg               7055 kernel/sched.c 		*sg = &per_cpu(sched_group_cpus, cpu);
sg               7078 kernel/sched.c 	if (sg)
sg               7079 kernel/sched.c 		*sg = &per_cpu(sched_group_core, group);
sg               7087 kernel/sched.c 	if (sg)
sg               7088 kernel/sched.c 		*sg = &per_cpu(sched_group_core, cpu);
sg               7112 kernel/sched.c 	if (sg)
sg               7113 kernel/sched.c 		*sg = &per_cpu(sched_group_phys, group);
sg               7138 kernel/sched.c 	if (sg)
sg               7139 kernel/sched.c 		*sg = &per_cpu(sched_group_allnodes, group);
sg               7145 kernel/sched.c 	struct sched_group *sg = group_head;
sg               7148 kernel/sched.c 	if (!sg)
sg               7151 kernel/sched.c 		for_each_cpu_mask_nr(j, sg->cpumask) {
sg               7163 kernel/sched.c 			sg_inc_cpu_power(sg, sd->groups->__cpu_power);
sg               7165 kernel/sched.c 		sg = sg->next;
sg               7166 kernel/sched.c 	} while (sg != group_head);
sg               7184 kernel/sched.c 			struct sched_group *oldsg, *sg = sched_group_nodes[i];
sg               7191 kernel/sched.c 			if (sg == NULL)
sg               7193 kernel/sched.c 			sg = sg->next;
sg               7195 kernel/sched.c 			oldsg = sg;
sg               7196 kernel/sched.c 			sg = sg->next;
sg               7549 kernel/sched.c 		struct sched_group *sg, *prev;
sg               7567 kernel/sched.c 		sg = kmalloc_node(sizeof(struct sched_group), GFP_KERNEL, i);
sg               7568 kernel/sched.c 		if (!sg) {
sg               7573 kernel/sched.c 		sched_group_nodes[i] = sg;
sg               7578 kernel/sched.c 			sd->groups = sg;
sg               7580 kernel/sched.c 		sg->__cpu_power = 0;
sg               7581 kernel/sched.c 		sg->cpumask = *nodemask;
sg               7582 kernel/sched.c 		sg->next = sg;
sg               7584 kernel/sched.c 		prev = sg;
sg               7601 kernel/sched.c 			sg = kmalloc_node(sizeof(struct sched_group),
sg               7603 kernel/sched.c 			if (!sg) {
sg               7608 kernel/sched.c 			sg->__cpu_power = 0;
sg               7609 kernel/sched.c 			sg->cpumask = *tmpmask;
sg               7610 kernel/sched.c 			sg->next = prev->next;
sg               7612 kernel/sched.c 			prev->next = sg;
sg               7613 kernel/sched.c 			prev = sg;
sg               7645 kernel/sched.c 		struct sched_group *sg;
sg               7647 kernel/sched.c 		cpu_to_allnodes_group(first_cpu(*cpu_map), cpu_map, &sg,
sg               7649 kernel/sched.c 		init_numa_sched_groups_power(sg);
sg                 26 lib/scatterlist.c 	BUG_ON(sg->sg_magic != SG_MAGIC);
sg                 28 lib/scatterlist.c 	if (sg_is_last(sg))
sg                 31 lib/scatterlist.c 	sg++;
sg                 32 lib/scatterlist.c 	if (unlikely(sg_is_chain(sg)))
sg                 33 lib/scatterlist.c 		sg = sg_chain_ptr(sg);
sg                 35 lib/scatterlist.c 	return sg;
sg                 58 lib/scatterlist.c 	struct scatterlist *sg, *ret = NULL;
sg                 61 lib/scatterlist.c 	for_each_sg(sgl, sg, nents, i)
sg                 62 lib/scatterlist.c 		ret = sg;
sg                106 lib/scatterlist.c 	sg_init_table(sg, 1);
sg                107 lib/scatterlist.c 	sg_set_buf(sg, buf, buflen);
sg                126 lib/scatterlist.c 		free_page((unsigned long) sg);
sg                128 lib/scatterlist.c 		kfree(sg);
sg                214 lib/scatterlist.c 	struct scatterlist *sg, *prv;
sg                236 lib/scatterlist.c 		sg = alloc_fn(alloc_size, gfp_mask);
sg                237 lib/scatterlist.c 		if (unlikely(!sg))
sg                240 lib/scatterlist.c 		sg_init_table(sg, alloc_size);
sg                248 lib/scatterlist.c 			sg_chain(prv, max_ents, sg);
sg                250 lib/scatterlist.c 			table->sgl = sg;
sg                256 lib/scatterlist.c 			sg_mark_end(&sg[sg_size - 1]);
sg                266 lib/scatterlist.c 		prv = sg;
sg                 39 lib/swiotlb.c  #define SG_ENT_VIRT_ADDRESS(sg)	(sg_virt((sg)))
sg                 40 lib/swiotlb.c  #define SG_ENT_PHYS_ADDRESS(sg)	virt_to_bus(SG_ENT_VIRT_ADDRESS(sg))
sg                710 lib/swiotlb.c  	struct scatterlist *sg;
sg                717 lib/swiotlb.c  	for_each_sg(sgl, sg, nelems, i) {
sg                718 lib/swiotlb.c  		addr = SG_ENT_VIRT_ADDRESS(sg);
sg                721 lib/swiotlb.c  		    address_needs_mapping(hwdev, dev_addr, sg->length)) {
sg                722 lib/swiotlb.c  			void *map = map_single(hwdev, addr, sg->length, dir);
sg                726 lib/swiotlb.c  				swiotlb_full(hwdev, sg->length, dir, 0);
sg                732 lib/swiotlb.c  			sg->dma_address = virt_to_bus(map);
sg                734 lib/swiotlb.c  			sg->dma_address = dev_addr;
sg                735 lib/swiotlb.c  		sg->dma_length = sg->length;
sg                756 lib/swiotlb.c  	struct scatterlist *sg;
sg                761 lib/swiotlb.c  	for_each_sg(sgl, sg, nelems, i) {
sg                762 lib/swiotlb.c  		if (sg->dma_address != SG_ENT_PHYS_ADDRESS(sg))
sg                763 lib/swiotlb.c  			unmap_single(hwdev, bus_to_virt(sg->dma_address),
sg                764 lib/swiotlb.c  				     sg->dma_length, dir);
sg                766 lib/swiotlb.c  			dma_mark_clean(SG_ENT_VIRT_ADDRESS(sg), sg->dma_length);
sg                789 lib/swiotlb.c  	struct scatterlist *sg;
sg                794 lib/swiotlb.c  	for_each_sg(sgl, sg, nelems, i) {
sg                795 lib/swiotlb.c  		if (sg->dma_address != SG_ENT_PHYS_ADDRESS(sg))
sg                796 lib/swiotlb.c  			sync_single(hwdev, bus_to_virt(sg->dma_address),
sg                797 lib/swiotlb.c  				    sg->dma_length, dir, target);
sg                799 lib/swiotlb.c  			dma_mark_clean(SG_ENT_VIRT_ADDRESS(sg), sg->dma_length);
sg                807 lib/swiotlb.c  	swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_CPU);
sg                814 lib/swiotlb.c  	swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_DEVICE);
sg                132 net/9p/trans_virtio.c 	struct scatterlist sg[VIRTQUEUE_NUM];
sg                277 net/9p/trans_virtio.c 		sg_set_buf(&sg[index++], data, s);
sg                327 net/9p/trans_virtio.c 	out = pack_sg_list(chan->sg, 0, VIRTQUEUE_NUM, tc->sdata, tc->size);
sg                328 net/9p/trans_virtio.c 	in = pack_sg_list(chan->sg, out, VIRTQUEUE_NUM-out, rdata, t->msize);
sg                332 net/9p/trans_virtio.c 	if (chan->vq->vq_ops->add_buf(chan->vq, chan->sg, out, in, tc)) {
sg                407 net/9p/trans_virtio.c 	sg_init_table(chan->sg, VIRTQUEUE_NUM);
sg               2281 net/core/skbuff.c 	int sg = features & NETIF_F_SG;
sg               2305 net/core/skbuff.c 		if (hsize > len || !sg)
sg               2328 net/core/skbuff.c 		if (!sg) {
sg               2420 net/core/skbuff.c 		sg_set_buf(sg, skb->data + offset, copy);
sg               2438 net/core/skbuff.c 			sg_set_page(&sg[elt], frag->page, copy,
sg               2460 net/core/skbuff.c 				elt += __skb_to_sgvec(list, sg+elt, offset - start,
sg               2475 net/core/skbuff.c 	int nsg = __skb_to_sgvec(skb, sg, offset, len);
sg               2477 net/core/skbuff.c 	sg_mark_end(&sg[nsg - 1]);
sg                361 net/ieee80211/ieee80211_crypt_tkip.c 	struct scatterlist sg;
sg                393 net/ieee80211/ieee80211_crypt_tkip.c 	sg_init_one(&sg, pos, len + 4);
sg                394 net/ieee80211/ieee80211_crypt_tkip.c 	return crypto_blkcipher_encrypt(&desc, &sg, &sg, len + 4);
sg                421 net/ieee80211/ieee80211_crypt_tkip.c 	struct scatterlist sg;
sg                486 net/ieee80211/ieee80211_crypt_tkip.c 	sg_init_one(&sg, pos, plen + 4);
sg                487 net/ieee80211/ieee80211_crypt_tkip.c 	if (crypto_blkcipher_decrypt(&desc, &sg, &sg, plen + 4)) {
sg                532 net/ieee80211/ieee80211_crypt_tkip.c 	struct scatterlist sg[2];
sg                538 net/ieee80211/ieee80211_crypt_tkip.c 	sg_init_table(sg, 2);
sg                539 net/ieee80211/ieee80211_crypt_tkip.c 	sg_set_buf(&sg[0], hdr, 16);
sg                540 net/ieee80211/ieee80211_crypt_tkip.c 	sg_set_buf(&sg[1], data, data_len);
sg                547 net/ieee80211/ieee80211_crypt_tkip.c 	return crypto_hash_digest(&desc, sg, data_len + 16, mic);
sg                143 net/ieee80211/ieee80211_crypt_wep.c 	struct scatterlist sg;
sg                173 net/ieee80211/ieee80211_crypt_wep.c 	sg_init_one(&sg, pos, len + 4);
sg                174 net/ieee80211/ieee80211_crypt_wep.c 	return crypto_blkcipher_encrypt(&desc, &sg, &sg, len + 4);
sg                191 net/ieee80211/ieee80211_crypt_wep.c 	struct scatterlist sg;
sg                213 net/ieee80211/ieee80211_crypt_wep.c 	sg_init_one(&sg, pos, plen + 4);
sg                214 net/ieee80211/ieee80211_crypt_wep.c 	if (crypto_blkcipher_decrypt(&desc, &sg, &sg, plen + 4))
sg                110 net/ipv4/esp4.c 	struct scatterlist *sg;
sg                149 net/ipv4/esp4.c 	sg = asg + 1;
sg                204 net/ipv4/esp4.c 	sg_init_table(sg, nfrags);
sg                205 net/ipv4/esp4.c 	skb_to_sgvec(skb, sg,
sg                211 net/ipv4/esp4.c 	aead_givcrypt_set_crypt(req, sg, sg, clen, iv);
sg                335 net/ipv4/esp4.c 	struct scatterlist *sg;
sg                358 net/ipv4/esp4.c 	sg = asg + 1;
sg                367 net/ipv4/esp4.c 	sg_init_table(sg, nfrags);
sg                368 net/ipv4/esp4.c 	skb_to_sgvec(skb, sg, sizeof(*esph) + crypto_aead_ivsize(aead), elen);
sg                372 net/ipv4/esp4.c 	aead_request_set_crypt(req, sg, sg, elen, iv);
sg               2597 net/ipv4/tcp.c 	struct scatterlist sg;
sg               2603 net/ipv4/tcp.c 	sg_init_one(&sg, th, sizeof(struct tcphdr));
sg               2604 net/ipv4/tcp.c 	err = crypto_hash_update(&hp->md5_desc, &sg, sizeof(struct tcphdr));
sg               2614 net/ipv4/tcp.c 	struct scatterlist sg;
sg               2622 net/ipv4/tcp.c 	sg_init_table(&sg, 1);
sg               2624 net/ipv4/tcp.c 	sg_set_buf(&sg, ((u8 *) tp) + header_len, head_data_len);
sg               2625 net/ipv4/tcp.c 	if (crypto_hash_update(desc, &sg, head_data_len))
sg               2630 net/ipv4/tcp.c 		sg_set_page(&sg, f->page, f->size, f->page_offset);
sg               2631 net/ipv4/tcp.c 		if (crypto_hash_update(desc, &sg, f->size))
sg               2642 net/ipv4/tcp.c 	struct scatterlist sg;
sg               2644 net/ipv4/tcp.c 	sg_init_one(&sg, key->key, key->keylen);
sg               2645 net/ipv4/tcp.c 	return crypto_hash_update(&hp->md5_desc, &sg, key->keylen);
sg                996 net/ipv4/tcp_ipv4.c 	struct scatterlist sg;
sg               1011 net/ipv4/tcp_ipv4.c 	sg_init_one(&sg, bp, sizeof(*bp));
sg               1012 net/ipv4/tcp_ipv4.c 	return crypto_hash_update(&hp->md5_desc, &sg, sizeof(*bp));
sg                136 net/ipv6/esp6.c 	struct scatterlist *sg;
sg                173 net/ipv6/esp6.c 	sg = asg + 1;
sg                193 net/ipv6/esp6.c 	sg_init_table(sg, nfrags);
sg                194 net/ipv6/esp6.c 	skb_to_sgvec(skb, sg,
sg                200 net/ipv6/esp6.c 	aead_givcrypt_set_crypt(req, sg, sg, clen, iv);
sg                282 net/ipv6/esp6.c 	struct scatterlist *sg;
sg                309 net/ipv6/esp6.c 	sg = asg + 1;
sg                318 net/ipv6/esp6.c 	sg_init_table(sg, nfrags);
sg                319 net/ipv6/esp6.c 	skb_to_sgvec(skb, sg, sizeof(*esph) + crypto_aead_ivsize(aead), elen);
sg                323 net/ipv6/esp6.c 	aead_request_set_crypt(req, sg, sg, elen, iv);
sg                746 net/ipv6/tcp_ipv6.c 	struct scatterlist sg;
sg                755 net/ipv6/tcp_ipv6.c 	sg_init_one(&sg, bp, sizeof(*bp));
sg                756 net/ipv6/tcp_ipv6.c 	return crypto_hash_update(&hp->md5_desc, &sg, sizeof(*bp));
sg                263 net/ipx/ipx_route.c 	struct sockaddr_ipx *sg, *st;
sg                269 net/ipx/ipx_route.c 	sg = (struct sockaddr_ipx *)&rt.rt_gateway;
sg                274 net/ipx/ipx_route.c 	    sg->sipx_family != AF_IPX ||
sg                285 net/ipx/ipx_route.c 		f.ipx_router_network	= sg->sipx_network;
sg                286 net/ipx/ipx_route.c 		memcpy(f.ipx_router_node, sg->sipx_node, IPX_NODE_LEN);
sg                125 net/mac80211/wep.c 	struct scatterlist sg;
sg                132 net/mac80211/wep.c 	sg_init_one(&sg, data, data_len + WEP_ICV_LEN);
sg                133 net/mac80211/wep.c 	crypto_blkcipher_encrypt(&desc, &sg, &sg, sg.length);
sg                192 net/mac80211/wep.c 	struct scatterlist sg;
sg                196 net/mac80211/wep.c 	sg_init_one(&sg, data, data_len + WEP_ICV_LEN);
sg                197 net/mac80211/wep.c 	crypto_blkcipher_decrypt(&desc, &sg, &sg, sg.length);
sg                115 net/rxrpc/rxkad.c 	struct scatterlist sg[2];
sg                138 net/rxrpc/rxkad.c 	sg_init_one(&sg[0], &tmpbuf, sizeof(tmpbuf));
sg                139 net/rxrpc/rxkad.c 	sg_init_one(&sg[1], &tmpbuf, sizeof(tmpbuf));
sg                140 net/rxrpc/rxkad.c 	crypto_blkcipher_encrypt_iv(&desc, &sg[0], &sg[1], sizeof(tmpbuf));
sg                159 net/rxrpc/rxkad.c 	struct scatterlist sg[2];
sg                182 net/rxrpc/rxkad.c 	sg_init_one(&sg[0], &tmpbuf, sizeof(tmpbuf));
sg                183 net/rxrpc/rxkad.c 	sg_init_one(&sg[1], &tmpbuf, sizeof(tmpbuf));
sg                184 net/rxrpc/rxkad.c 	crypto_blkcipher_encrypt_iv(&desc, &sg[0], &sg[1], sizeof(tmpbuf));
sg                206 net/rxrpc/rxkad.c 	struct scatterlist sg[16];
sg                228 net/rxrpc/rxkad.c 	sg_init_one(&sg[0], sechdr, sizeof(rxkhdr));
sg                229 net/rxrpc/rxkad.c 	sg_init_one(&sg[1], &rxkhdr, sizeof(rxkhdr));
sg                230 net/rxrpc/rxkad.c 	crypto_blkcipher_encrypt_iv(&desc, &sg[0], &sg[1], sizeof(rxkhdr));
sg                240 net/rxrpc/rxkad.c 	sg_init_table(sg, nsg);
sg                241 net/rxrpc/rxkad.c 	skb_to_sgvec(skb, sg, 0, len);
sg                242 net/rxrpc/rxkad.c 	crypto_blkcipher_encrypt_iv(&desc, sg, sg, len);
sg                259 net/rxrpc/rxkad.c 	struct scatterlist sg[2];
sg                292 net/rxrpc/rxkad.c 	sg_init_one(&sg[0], &tmpbuf, sizeof(tmpbuf));
sg                293 net/rxrpc/rxkad.c 	sg_init_one(&sg[1], &tmpbuf, sizeof(tmpbuf));
sg                294 net/rxrpc/rxkad.c 	crypto_blkcipher_encrypt_iv(&desc, &sg[0], &sg[1], sizeof(tmpbuf));
sg                333 net/rxrpc/rxkad.c 	struct scatterlist sg[16];
sg                348 net/rxrpc/rxkad.c 	sg_init_table(sg, nsg);
sg                349 net/rxrpc/rxkad.c 	skb_to_sgvec(skb, sg, 0, 8);
sg                357 net/rxrpc/rxkad.c 	crypto_blkcipher_decrypt_iv(&desc, sg, sg, 8);
sg                408 net/rxrpc/rxkad.c 	struct scatterlist _sg[4], *sg;
sg                423 net/rxrpc/rxkad.c 	sg = _sg;
sg                425 net/rxrpc/rxkad.c 		sg = kmalloc(sizeof(*sg) * nsg, GFP_NOIO);
sg                426 net/rxrpc/rxkad.c 		if (!sg)
sg                430 net/rxrpc/rxkad.c 	sg_init_table(sg, nsg);
sg                431 net/rxrpc/rxkad.c 	skb_to_sgvec(skb, sg, 0, skb->len);
sg                440 net/rxrpc/rxkad.c 	crypto_blkcipher_decrypt_iv(&desc, sg, sg, skb->len);
sg                441 net/rxrpc/rxkad.c 	if (sg != _sg)
sg                442 net/rxrpc/rxkad.c 		kfree(sg);
sg                491 net/rxrpc/rxkad.c 	struct scatterlist sg[2];
sg                527 net/rxrpc/rxkad.c 	sg_init_one(&sg[0], &tmpbuf, sizeof(tmpbuf));
sg                528 net/rxrpc/rxkad.c 	sg_init_one(&sg[1], &tmpbuf, sizeof(tmpbuf));
sg                529 net/rxrpc/rxkad.c 	crypto_blkcipher_encrypt_iv(&desc, &sg[0], &sg[1], sizeof(tmpbuf));
sg                697 net/rxrpc/rxkad.c 	sg_init_table(sg, 2);
sg                699 net/rxrpc/rxkad.c 	sg_set_buf(&sg[0], buf, buflen);
sg                700 net/rxrpc/rxkad.c 	if (sg[0].offset + buflen > PAGE_SIZE) {
sg                702 net/rxrpc/rxkad.c 		sg[0].length = PAGE_SIZE - sg[0].offset;
sg                703 net/rxrpc/rxkad.c 		sg_set_buf(&sg[1], buf + sg[0].length, buflen - sg[0].length);
sg                707 net/rxrpc/rxkad.c 	sg_mark_end(&sg[nsg - 1]);
sg                709 net/rxrpc/rxkad.c 	ASSERTCMP(sg[0].length + sg[1].length, ==, buflen);
sg                721 net/rxrpc/rxkad.c 	struct scatterlist sg[2];
sg                729 net/rxrpc/rxkad.c 	rxkad_sg_set_buf2(sg, &resp->encrypted, sizeof(resp->encrypted));
sg                730 net/rxrpc/rxkad.c 	crypto_blkcipher_encrypt_iv(&desc, sg, sg, sizeof(resp->encrypted));
sg                825 net/rxrpc/rxkad.c 	struct scatterlist sg[1];
sg                858 net/rxrpc/rxkad.c 	sg_init_one(&sg[0], ticket, ticket_len);
sg                859 net/rxrpc/rxkad.c 	crypto_blkcipher_decrypt_iv(&desc, sg, sg, ticket_len);
sg                968 net/rxrpc/rxkad.c 	struct scatterlist sg[2];
sg                986 net/rxrpc/rxkad.c 	rxkad_sg_set_buf2(sg, &resp->encrypted, sizeof(resp->encrypted));
sg                987 net/rxrpc/rxkad.c 	crypto_blkcipher_decrypt_iv(&desc, sg, sg, sizeof(resp->encrypted));
sg                702 net/sctp/auth.c 	struct scatterlist sg;
sg                735 net/sctp/auth.c 	sg_init_one(&sg, auth, end - (unsigned char *)auth);
sg                744 net/sctp/auth.c 	crypto_hash_digest(&desc, &sg, sg.length, digest);
sg               1478 net/sctp/sm_make_chunk.c 	struct scatterlist sg;
sg               1540 net/sctp/sm_make_chunk.c 		sg_init_one(&sg, &cookie->c, bodysize);
sg               1547 net/sctp/sm_make_chunk.c 		    crypto_hash_digest(&desc, &sg, bodysize, cookie->signature))
sg               1572 net/sctp/sm_make_chunk.c 	struct scatterlist sg;
sg               1610 net/sctp/sm_make_chunk.c 	sg_init_one(&sg, bear_cookie, bodysize);
sg               1617 net/sctp/sm_make_chunk.c 	    crypto_hash_digest(&desc, &sg, bodysize, digest)) {
sg               1627 net/sctp/sm_make_chunk.c 		    crypto_hash_digest(&desc, &sg, bodysize, digest)) {
sg                 61 net/sunrpc/auth_gss/gss_krb5_crypto.c 	struct scatterlist sg[1];
sg                 78 net/sunrpc/auth_gss/gss_krb5_crypto.c 	sg_init_one(sg, out, length);
sg                 80 net/sunrpc/auth_gss/gss_krb5_crypto.c 	ret = crypto_blkcipher_encrypt_iv(&desc, sg, sg, length);
sg                 95 net/sunrpc/auth_gss/gss_krb5_crypto.c 	struct scatterlist sg[1];
sg                111 net/sunrpc/auth_gss/gss_krb5_crypto.c 	sg_init_one(sg, out, length);
sg                113 net/sunrpc/auth_gss/gss_krb5_crypto.c 	ret = crypto_blkcipher_decrypt_iv(&desc, sg, sg, length);
sg                124 net/sunrpc/auth_gss/gss_krb5_crypto.c 	return crypto_hash_update(desc, sg, sg->length);
sg                133 net/sunrpc/auth_gss/gss_krb5_crypto.c 	struct scatterlist              sg[1];
sg                145 net/sunrpc/auth_gss/gss_krb5_crypto.c 	sg_init_one(sg, header, hdrlen);
sg                146 net/sunrpc/auth_gss/gss_krb5_crypto.c 	err = crypto_hash_update(&desc, sg, hdrlen);
sg                178 net/sunrpc/auth_gss/gss_krb5_crypto.c 	int thislen = desc->fraglen + sg->length;
sg                192 net/sunrpc/auth_gss/gss_krb5_crypto.c 		in_page = sg_page(sg);
sg                194 net/sunrpc/auth_gss/gss_krb5_crypto.c 	sg_set_page(&desc->infrags[desc->fragno], in_page, sg->length,
sg                195 net/sunrpc/auth_gss/gss_krb5_crypto.c 		    sg->offset);
sg                196 net/sunrpc/auth_gss/gss_krb5_crypto.c 	sg_set_page(&desc->outfrags[desc->fragno], sg_page(sg), sg->length,
sg                197 net/sunrpc/auth_gss/gss_krb5_crypto.c 		    sg->offset);
sg                199 net/sunrpc/auth_gss/gss_krb5_crypto.c 	desc->fraglen += sg->length;
sg                200 net/sunrpc/auth_gss/gss_krb5_crypto.c 	desc->pos += sg->length;
sg                220 net/sunrpc/auth_gss/gss_krb5_crypto.c 		sg_set_page(&desc->outfrags[0], sg_page(sg), fraglen,
sg                221 net/sunrpc/auth_gss/gss_krb5_crypto.c 				sg->offset + sg->length - fraglen);
sg                271 net/sunrpc/auth_gss/gss_krb5_crypto.c 	int thislen = desc->fraglen + sg->length;
sg                277 net/sunrpc/auth_gss/gss_krb5_crypto.c 	sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length,
sg                278 net/sunrpc/auth_gss/gss_krb5_crypto.c 		    sg->offset);
sg                280 net/sunrpc/auth_gss/gss_krb5_crypto.c 	desc->fraglen += sg->length;
sg                298 net/sunrpc/auth_gss/gss_krb5_crypto.c 		sg_set_page(&desc->frags[0], sg_page(sg), fraglen,
sg                299 net/sunrpc/auth_gss/gss_krb5_crypto.c 				sg->offset + sg->length - fraglen);
sg                136 net/sunrpc/auth_gss/gss_spkm3_seal.c 	return crypto_hash_update(desc, sg, sg->length);
sg                147 net/sunrpc/auth_gss/gss_spkm3_seal.c 	struct scatterlist		sg[1];
sg                176 net/sunrpc/auth_gss/gss_spkm3_seal.c 	sg_init_one(sg, header, hdrlen);
sg                177 net/sunrpc/auth_gss/gss_spkm3_seal.c 	crypto_hash_update(&desc, sg, sg->length);
sg               1050 net/sunrpc/xdr.c 	struct scatterlist      sg[1];
sg               1052 net/sunrpc/xdr.c 	sg_init_table(sg, 1);
sg               1060 net/sunrpc/xdr.c 		sg_set_buf(sg, buf->head[0].iov_base + offset, thislen);
sg               1061 net/sunrpc/xdr.c 		ret = actor(sg, data);
sg               1083 net/sunrpc/xdr.c 			sg_set_page(sg, buf->pages[i], thislen, page_offset);
sg               1084 net/sunrpc/xdr.c 			ret = actor(sg, data);
sg               1100 net/sunrpc/xdr.c 		sg_set_buf(sg, buf->tail[0].iov_base + offset, thislen);
sg               1101 net/sunrpc/xdr.c 		ret = actor(sg, data);
sg                700 net/xfrm/xfrm_algo.c 	struct scatterlist sg;
sg                707 net/xfrm/xfrm_algo.c 		sg_init_one(&sg, skb->data + offset, copy);
sg                709 net/xfrm/xfrm_algo.c 		err = icv_update(desc, &sg, copy);
sg                730 net/xfrm/xfrm_algo.c 			sg_init_table(&sg, 1);
sg                731 net/xfrm/xfrm_algo.c 			sg_set_page(&sg, frag->page, copy,
sg                734 net/xfrm/xfrm_algo.c 			err = icv_update(desc, &sg, copy);
sg                333 sound/core/pcm_memory.c 	struct snd_sg_buf *sg = snd_pcm_substream_sgbuf(substream);
sg                339 sound/core/pcm_memory.c 	pg = sg->table[start].addr >> PAGE_SHIFT;
sg                345 sound/core/pcm_memory.c 		if ((sg->table[start].addr >> PAGE_SHIFT) != pg)