b1                 78 arch/x86/kernel/kprobes.c 	(((b0##UL << 0x0)|(b1##UL << 0x1)|(b2##UL << 0x2)|(b3##UL << 0x3) |   \
b1                123 block/blk-integrity.c 	struct blk_integrity *b1 = gd1->integrity;
b1                126 block/blk-integrity.c 	if (!b1 && !b2)
b1                129 block/blk-integrity.c 	if (!b1 || !b2)
b1                132 block/blk-integrity.c 	if (b1->sector_size != b2->sector_size) {
b1                135 block/blk-integrity.c 		       b1->sector_size, b2->sector_size);
b1                139 block/blk-integrity.c 	if (b1->tuple_size != b2->tuple_size) {
b1                142 block/blk-integrity.c 		       b1->tuple_size, b2->tuple_size);
b1                146 block/blk-integrity.c 	if (b1->tag_size && b2->tag_size && (b1->tag_size != b2->tag_size)) {
b1                149 block/blk-integrity.c 		       b1->tag_size, b2->tag_size);
b1                153 block/blk-integrity.c 	if (strcmp(b1->name, b2->name)) {
b1                156 block/blk-integrity.c 		       b1->name, b2->name);
b1                363 crypto/aes_generic.c 	u32 b0[4], b1[4];
b1                373 crypto/aes_generic.c 		f_nround(b1, b0, kp);
b1                374 crypto/aes_generic.c 		f_nround(b0, b1, kp);
b1                378 crypto/aes_generic.c 		f_nround(b1, b0, kp);
b1                379 crypto/aes_generic.c 		f_nround(b0, b1, kp);
b1                382 crypto/aes_generic.c 	f_nround(b1, b0, kp);
b1                383 crypto/aes_generic.c 	f_nround(b0, b1, kp);
b1                384 crypto/aes_generic.c 	f_nround(b1, b0, kp);
b1                385 crypto/aes_generic.c 	f_nround(b0, b1, kp);
b1                386 crypto/aes_generic.c 	f_nround(b1, b0, kp);
b1                387 crypto/aes_generic.c 	f_nround(b0, b1, kp);
b1                388 crypto/aes_generic.c 	f_nround(b1, b0, kp);
b1                389 crypto/aes_generic.c 	f_nround(b0, b1, kp);
b1                390 crypto/aes_generic.c 	f_nround(b1, b0, kp);
b1                391 crypto/aes_generic.c 	f_lround(b0, b1, kp);
b1                435 crypto/aes_generic.c 	u32 b0[4], b1[4];
b1                445 crypto/aes_generic.c 		i_nround(b1, b0, kp);
b1                446 crypto/aes_generic.c 		i_nround(b0, b1, kp);
b1                450 crypto/aes_generic.c 		i_nround(b1, b0, kp);
b1                451 crypto/aes_generic.c 		i_nround(b0, b1, kp);
b1                454 crypto/aes_generic.c 	i_nround(b1, b0, kp);
b1                455 crypto/aes_generic.c 	i_nround(b0, b1, kp);
b1                456 crypto/aes_generic.c 	i_nround(b1, b0, kp);
b1                457 crypto/aes_generic.c 	i_nround(b0, b1, kp);
b1                458 crypto/aes_generic.c 	i_nround(b1, b0, kp);
b1                459 crypto/aes_generic.c 	i_nround(b0, b1, kp);
b1                460 crypto/aes_generic.c 	i_nround(b1, b0, kp);
b1                461 crypto/aes_generic.c 	i_nround(b0, b1, kp);
b1                462 crypto/aes_generic.c 	i_nround(b1, b0, kp);
b1                463 crypto/aes_generic.c 	i_lround(b0, b1, kp);
b1                 82 crypto/xor.c   			tmpl->do_2(BENCH_SIZE, b1, b2);
b1                101 crypto/xor.c   	void *b1, *b2;
b1                104 crypto/xor.c   	b1 = (void *) __get_free_pages(GFP_KERNEL, 2);
b1                105 crypto/xor.c   	if (!b1) {
b1                109 crypto/xor.c   	b2 = b1 + 2*PAGE_SIZE + BENCH_SIZE;
b1                122 crypto/xor.c   #define xor_speed(templ)	do_xor_speed((templ), b1, b2)
b1                143 crypto/xor.c   	free_pages((unsigned long)b1, 2);
b1                 22 fs/ext3/hash.c 	__u32	b0 = buf[0], b1 = buf[1];
b1                 28 fs/ext3/hash.c 		b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b);
b1                 29 fs/ext3/hash.c 		b1 += ((b0 << 4)+c) ^ (b0+sum) ^ ((b0 >> 5)+d);
b1                 33 fs/ext3/hash.c 	buf[1] += b1;
b1               1433 fs/ext4/extents.c 	ext4_lblk_t b1, b2;
b1               1437 fs/ext4/extents.c 	b1 = le32_to_cpu(newext->ee_block);
b1               1448 fs/ext4/extents.c 	if (b2 < b1) {
b1               1455 fs/ext4/extents.c 	if (b1 + len1 < b1) {
b1               1456 fs/ext4/extents.c 		len1 = EXT_MAX_BLOCK - b1;
b1               1462 fs/ext4/extents.c 	if (b1 + len1 > b2) {
b1               1463 fs/ext4/extents.c 		newext->ee_len = cpu_to_le16(b2 - b1);
b1                 22 fs/ext4/hash.c 	__u32	b0 = buf[0], b1 = buf[1];
b1                 28 fs/ext4/hash.c 		b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b);
b1                 29 fs/ext4/hash.c 		b1 += ((b0 << 4)+c) ^ (b0+sum) ^ ((b0 >> 5)+d);
b1                 33 fs/ext4/hash.c 	buf[1] += b1;
b1                474 fs/ext4/mballoc.c 		unsigned char *b1, *b2;
b1                476 fs/ext4/mballoc.c 		b1 = (unsigned char *) e4b->bd_info->bb_bitmap;
b1                479 fs/ext4/mballoc.c 			if (b1[i] != b2[i]) {
b1                483 fs/ext4/mballoc.c 				       e4b->bd_group, i, i * 8, b1[i], b2[i]);
b1                318 fs/hpfs/ea.c   					void *b1, *b2;
b1                319 fs/hpfs/ea.c   					if (!(b1 = hpfs_map_sector(s, fnode->ea_secno + i, &bh1, len - i - 1))) {
b1                328 fs/hpfs/ea.c   					memcpy(b2, b1, 512);
b1                317 fs/nfsd/nfsctl.c 	int b1, b2, b3, b4;
b1                333 fs/nfsd/nfsctl.c 	if (sscanf(fo_path, NIPQUAD_FMT "%c", &b1, &b2, &b3, &b4, &c) != 4)
b1                335 fs/nfsd/nfsctl.c 	if (b1 > 255 || b2 > 255 || b3 > 255 || b4 > 255)
b1                337 fs/nfsd/nfsctl.c 	sin.sin_addr.s_addr = htonl((b1 << 24) | (b2 << 16) | (b3 << 8) | b4);
b1                 34 fs/reiserfs/hashes.c 		u32 b0, b1;						\
b1                 37 fs/reiserfs/hashes.c 		b1 = h1;						\
b1                 42 fs/reiserfs/hashes.c 			b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b);	\
b1                 43 fs/reiserfs/hashes.c 			b1 += ((b0 << 4)+c) ^ (b0+sum) ^ ((b0 >> 5)+d);	\
b1                 47 fs/reiserfs/hashes.c 		h1 += b1;						\
b1                 99 fs/xfs/xfs_mount.h 	(*(mp)->m_dm_ops->xfs_send_namesp)(ev,NULL,b1,r1,b2,r2,n1,n2,mode,rval,fl)
b1                101 fs/xfs/xfs_mount.h 	(*(mp)->m_dm_ops->xfs_send_namesp)(DM_EVENT_PREUNMOUNT,mp,b1,r1,b2,r2,n1,n2,mode,rval,fl)
b1                235 include/linux/bio.h 	__BIO_SEG_BOUNDARY(bvec_to_phys((b1)), bvec_to_phys((b2)) + (b2)->bv_len, (q)->seg_boundary_mask)
b1                237 include/linux/bio.h 	BIOVEC_SEG_BOUNDARY((q), __BVEC_END((b1)), __BVEC_START((b2)))
b1                125 include/linux/ide.h #define IDE_LARGE_SEEK(b1,b2,t)	(((b1) > (b2) + (t)) || ((b2) > (b1) + (t)))
b1                 98 include/linux/signal.h 	unsigned long a0, a1, a2, a3, b0, b1, b2, b3;			\
b1                107 include/linux/signal.h 		a1 = a->sig[1]; b1 = b->sig[1];				\
b1                108 include/linux/signal.h 		r->sig[1] = op(a1, b1);					\
b1                 27 include/video/newport.h 	struct { volatile unsigned char b0, b1, b2, b3; } bybytes;
b1                719 net/atm/br2684.c #define bs(var)		b1(var, 0), b1(var, 1), b1(var, 2), b1(var, 3)
b1                534 net/core/pktgen.c 		char b1[128], b2[128], b3[128];
b1                535 net/core/pktgen.c 		fmt_ip6(b1, pkt_dev->in6_saddr.s6_addr);
b1                539 net/core/pktgen.c 			   "     saddr: %s  min_saddr: %s  max_saddr: %s\n", b1,
b1                542 net/core/pktgen.c 		fmt_ip6(b1, pkt_dev->in6_daddr.s6_addr);
b1                546 net/core/pktgen.c 			   "     daddr: %s  min_daddr: %s  max_daddr: %s\n", b1,
b1                672 net/core/pktgen.c 		char b1[128], b2[128];
b1                673 net/core/pktgen.c 		fmt_ip6(b1, pkt_dev->cur_in6_daddr.s6_addr);
b1                675 net/core/pktgen.c 		seq_printf(seq, "     cur_saddr: %s  cur_daddr: %s\n", b2, b1);
b1                189 net/sunrpc/svcauth_unix.c 	int b1, b2, b3, b4, b5, b6, b7, b8;
b1                211 net/sunrpc/svcauth_unix.c 	if (sscanf(buf, NIPQUAD_FMT "%c", &b1, &b2, &b3, &b4, &c) == 4) {
b1                216 net/sunrpc/svcauth_unix.c 			htonl((((((b1<<8)|b2)<<8)|b3)<<8)|b4);
b1                218 net/sunrpc/svcauth_unix.c 			&b1, &b2, &b3, &b4, &b5, &b6, &b7, &b8, &c) == 8) {
b1                219 net/sunrpc/svcauth_unix.c 		addr.s6_addr16[0] = htons(b1);