ALIGN 499 arch/x86/kernel/amd_iommu.c boundary_size = ALIGN(dma_get_seg_boundary(dev) + 1, ALIGN 994 arch/x86/kernel/cpu/mtrr/main.c range_basek = ALIGN(state->range_startk, gran_sizek); ALIGN 998 arch/x86/kernel/cpu/mtrr/main.c range_sizek = ALIGN(state->range_sizek, gran_sizek); ALIGN 1009 arch/x86/kernel/cpu/mtrr/main.c range0_sizek = ALIGN(state->range_sizek, chunk_sizek); ALIGN 428 arch/x86/kernel/ds.c adj = ALIGN(buffer, alignment) - buffer; ALIGN 277 arch/x86/kernel/pci-calgary_64.c boundary_size = ALIGN(dma_get_seg_boundary(dev) + 1, ALIGN 92 arch/x86/kernel/pci-gart_64.c base_index = ALIGN(iommu_bus_base & dma_get_seg_boundary(dev), ALIGN 94 arch/x86/kernel/pci-gart_64.c boundary_size = ALIGN((unsigned long long)dma_get_seg_boundary(dev) + 1, ALIGN 1618 arch/x86/kvm/x86.c n = ALIGN(memslot->npages, BITS_PER_LONG) / 8; ALIGN 123 arch/x86/mm/hugetlbpage.c *addr = ALIGN(*addr, HPAGE_SIZE * PTRS_PER_PTE) - HPAGE_SIZE; ALIGN 276 arch/x86/mm/hugetlbpage.c addr = ALIGN(start_addr, huge_page_size(h)); ALIGN 298 arch/x86/mm/hugetlbpage.c addr = ALIGN(vma->vm_end, huge_page_size(h)); ALIGN 412 arch/x86/mm/hugetlbpage.c addr = ALIGN(addr, huge_page_size(h)); ALIGN 50 arch/x86/mm/memtest.c start_phys_aligned = ALIGN(start_phys, incr); ALIGN 194 arch/x86/mm/numa_32.c size = ALIGN(size, L1_CACHE_BYTES); ALIGN 308 arch/x86/mm/numa_32.c ALIGN(sizeof(pg_data_t), PAGE_SIZE); ALIGN 123 block/blk-tag.c nr_ulongs = ALIGN(depth, BITS_PER_LONG) / BITS_PER_LONG; ALIGN 259 block/blk-tag.c nr_ulongs = ALIGN(max_depth, BITS_PER_LONG) / BITS_PER_LONG; ALIGN 42 crypto/ablkcipher.c alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); ALIGN 40 crypto/aead.c alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); ALIGN 69 crypto/ahash.c walk->offset = ALIGN(walk->offset, alignmask + 1); ALIGN 129 crypto/ahash.c alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); ALIGN 119 crypto/authenc.c hash = (u8 *)ALIGN((unsigned long)hash + crypto_hash_alignmask(auth), ALIGN 78 crypto/blkcipher.c addr = (u8 *)ALIGN((unsigned long)walk->buffer, alignmask + 1); ALIGN 153 crypto/blkcipher.c unsigned aligned_bsize = ALIGN(bsize, alignmask + 1); ALIGN 169 crypto/blkcipher.c walk->dst.virt.addr = (u8 *)ALIGN((unsigned long)walk->buffer, ALIGN 284 crypto/blkcipher.c unsigned aligned_bs = ALIGN(bs, alignmask + 1); ALIGN 294 crypto/blkcipher.c iv = (u8 *)ALIGN((unsigned long)walk->buffer, alignmask + 1); ALIGN 373 crypto/blkcipher.c alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); ALIGN 438 crypto/blkcipher.c len = ALIGN(len, (unsigned long)alg->cra_alignmask + 1); ALIGN 475 crypto/blkcipher.c addr = ALIGN(addr, align); ALIGN 476 crypto/blkcipher.c addr += ALIGN(tfm->__crt_alg->cra_ctxsize, align); ALIGN 737 crypto/ccm.c ALIGN(crypto_aead_reqsize(aead), ALIGN 37 crypto/cipher.c alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); ALIGN 71 crypto/cipher.c u8 *tmp = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); ALIGN 104 crypto/digest.c u8 *dst = (u8 *)ALIGN(addr, align) + ALIGN 105 crypto/digest.c ALIGN(tfm->__crt_alg->cra_ctxsize, align); ALIGN 207 crypto/eseqiv.c reqsize = ALIGN(reqsize, crypto_tfm_ctx_alignment()); ALIGN 670 crypto/gcm.c ALIGN(crypto_aead_reqsize(aead), ALIGN 42 crypto/hash.c alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); ALIGN 35 crypto/hmac.c return (void *)ALIGN((unsigned long)p, align); ALIGN 260 crypto/hmac.c ALIGN(inst->alg.cra_blocksize * 2 + ds, ALIGN 73 crypto/internal.h len = ALIGN(len, (unsigned long)alg->cra_alignmask + 1); ALIGN 331 crypto/xcbc.c ALIGN(inst->alg.cra_blocksize * 3, sizeof(void *)); ALIGN 861 fs/compat.c int reclen = ALIGN(NAME_OFFSET(dirent) + namlen + 2, sizeof(compat_long_t)); ALIGN 952 fs/compat.c int reclen = ALIGN(jj + namlen + 1, sizeof(u64)); ALIGN 867 fs/direct-io.c i_size_aligned = ALIGN(i_size_read(dio->inode), ALIGN 149 fs/ext4/ext4.h #define EXT4_BLOCK_ALIGN(size, blkbits) ALIGN((size), (1 << (blkbits))) ALIGN 20 fs/gfs2/eattr.h ALIGN(sizeof(struct gfs2_ea_header) + (ea)->ea_name_len + \ ALIGN 28 fs/gfs2/eattr.h ALIGN(sizeof(struct gfs2_ea_header) + (er)->er_name_len + (er)->er_data_len, 8) ALIGN 31 fs/gfs2/eattr.h ALIGN(sizeof(struct gfs2_ea_header) + (er)->er_name_len + \ ALIGN 38 fs/gfs2/eattr.h ((__be64 *)(GFS2_EA2NAME(ea) + ALIGN((ea)->ea_name_len, 8))) ALIGN 163 fs/gfs2/rgrp.c g2 = ALIGN(g1, sizeof(unsigned long)); ALIGN 150 fs/hugetlbfs/inode.c addr = ALIGN(addr, huge_page_size(h)); ALIGN 163 fs/hugetlbfs/inode.c addr = ALIGN(start_addr, huge_page_size(h)); ALIGN 181 fs/hugetlbfs/inode.c addr = ALIGN(vma->vm_end, huge_page_size(h)); ALIGN 154 fs/readdir.c int reclen = ALIGN(NAME_OFFSET(dirent) + namlen + 2, sizeof(long)); ALIGN 241 fs/readdir.c int reclen = ALIGN(NAME_OFFSET(dirent) + namlen + 1, sizeof(u64)); ALIGN 203 fs/smbfs/request.c const int oparam = ALIGN(header + 3, sizeof(u32)); ALIGN 204 fs/smbfs/request.c const int odata = ALIGN(oparam + req->rq_lparm, sizeof(u32)); ALIGN 700 fs/ubifs/budget.c req.dd_growth = c->inode_budget + ALIGN(ui->data_len, 8); ALIGN 127 fs/ubifs/compress.c if (ALIGN(*out_len, 8) >= ALIGN(in_len, 8)) ALIGN 1432 fs/ubifs/debug.c add = ALIGN(add, 8); ALIGN 519 fs/ubifs/dir.c .dirtied_ino_d = ALIGN(ui->data_len, 8) }; ALIGN 781 fs/ubifs/dir.c .new_ino_d = ALIGN(devlen, 8), ALIGN 856 fs/ubifs/dir.c .new_ino_d = ALIGN(len, 8), ALIGN 1006 fs/ubifs/dir.c .dirtied_ino_d = ALIGN(old_inode_ui->data_len, 8) }; ALIGN 1189 fs/ubifs/dir.c size = ALIGN(size, UBIFS_BLOCK_SIZE); ALIGN 905 fs/ubifs/file.c .dirtied_ino_d = ALIGN(ui->data_len, 8) }; ALIGN 1067 fs/ubifs/file.c .dirtied_ino_d = ALIGN(ui->data_len, 8) }; ALIGN 242 fs/ubifs/io.c len = ALIGN(len, 8); ALIGN 243 fs/ubifs/io.c pad = ALIGN(len, c->min_io_size) - len; ALIGN 495 fs/ubifs/io.c int err, written, n, aligned_len = ALIGN(len, 8), offs; ALIGN 641 fs/ubifs/io.c int err, buf_len = ALIGN(len, c->min_io_size); ALIGN 567 fs/ubifs/journal.c aligned_dlen = ALIGN(dlen, 8); ALIGN 568 fs/ubifs/journal.c aligned_ilen = ALIGN(ilen, 8); ALIGN 933 fs/ubifs/journal.c aligned_dlen1 = ALIGN(dlen1, 8); ALIGN 934 fs/ubifs/journal.c aligned_dlen2 = ALIGN(dlen2, 8); ALIGN 935 fs/ubifs/journal.c len = aligned_dlen1 + aligned_dlen2 + ALIGN(ilen, 8) + ALIGN(plen, 8); ALIGN 974 fs/ubifs/journal.c p += ALIGN(ilen, 8); ALIGN 981 fs/ubifs/journal.c p += ALIGN(plen, 8); ALIGN 1028 fs/ubifs/journal.c offs += ALIGN(ilen, 8); ALIGN 1037 fs/ubifs/journal.c offs += ALIGN(plen, 8); ALIGN 1281 fs/ubifs/journal.c aligned_xlen = ALIGN(xlen, 8); ALIGN 1283 fs/ubifs/journal.c len = aligned_xlen + UBIFS_INO_NODE_SZ + ALIGN(hlen, 8); ALIGN 1389 fs/ubifs/journal.c aligned_len1 = ALIGN(len1, 8); ALIGN 1390 fs/ubifs/journal.c aligned_len = aligned_len1 + ALIGN(len2, 8); ALIGN 409 fs/ubifs/log.c max_len = ALIGN(max_len, c->min_io_size); ALIGN 444 fs/ubifs/log.c ubifs_pad(c, buf + len, ALIGN(len, c->min_io_size) - len); ALIGN 459 fs/ubifs/log.c len = ALIGN(len, c->min_io_size); ALIGN 659 fs/ubifs/log.c int sz = ALIGN(*offs, c->min_io_size), err; ALIGN 669 fs/ubifs/log.c *offs += ALIGN(len, 8); ALIGN 738 fs/ubifs/log.c int sz = ALIGN(offs, c->min_io_size); ALIGN 744 fs/ubifs/log.c offs = ALIGN(offs, c->min_io_size); ALIGN 596 fs/ubifs/lprops.c free = ALIGN(free, 8); ALIGN 609 fs/ubifs/lprops.c dirty = ALIGN(dirty, 8); ALIGN 1177 fs/ubifs/lprops.c used += ALIGN(snod->len, 8); ALIGN 124 fs/ubifs/lpt.c tot_wastage += ALIGN(sz, c->min_io_size) - sz; ALIGN 632 fs/ubifs/lpt.c node_sz = ALIGN(ubifs_idx_node_sz(c, 1), 8); ALIGN 633 fs/ubifs/lpt.c iopos = ALIGN(node_sz, c->min_io_size); ALIGN 639 fs/ubifs/lpt.c iopos = ALIGN(node_sz, c->min_io_size); ALIGN 672 fs/ubifs/lpt.c alen = ALIGN(len, c->min_io_size); ALIGN 702 fs/ubifs/lpt.c alen = ALIGN(len, c->min_io_size); ALIGN 751 fs/ubifs/lpt.c alen = ALIGN(len, c->min_io_size); ALIGN 777 fs/ubifs/lpt.c alen = ALIGN(len, c->min_io_size); ALIGN 792 fs/ubifs/lpt.c alen = ALIGN(len, c->min_io_size); ALIGN 805 fs/ubifs/lpt.c c->nhead_offs = ALIGN(len, c->min_io_size); ALIGN 227 fs/ubifs/lpt_commit.c alen = ALIGN(offs, c->min_io_size); ALIGN 266 fs/ubifs/lpt_commit.c alen = ALIGN(offs, c->min_io_size); ALIGN 284 fs/ubifs/lpt_commit.c alen = ALIGN(offs, c->min_io_size); ALIGN 299 fs/ubifs/lpt_commit.c alen = ALIGN(offs, c->min_io_size); ALIGN 389 fs/ubifs/lpt_commit.c alen = ALIGN(wlen, c->min_io_size); ALIGN 445 fs/ubifs/lpt_commit.c alen = ALIGN(wlen, c->min_io_size); ALIGN 470 fs/ubifs/lpt_commit.c alen = ALIGN(wlen, c->min_io_size); ALIGN 493 fs/ubifs/lpt_commit.c alen = ALIGN(wlen, c->min_io_size); ALIGN 499 fs/ubifs/lpt_commit.c c->nhead_offs = ALIGN(offs, c->min_io_size); ALIGN 1013 fs/ubifs/lpt_commit.c pad_len = ALIGN(offs, c->min_io_size) - offs; ALIGN 248 fs/ubifs/orphan.c len = ALIGN(len, c->min_io_size); ALIGN 321 fs/ubifs/orphan.c c->ohead_offs += ALIGN(len, c->min_io_size); ALIGN 322 fs/ubifs/orphan.c c->ohead_offs = ALIGN(c->ohead_offs, 8); ALIGN 375 fs/ubifs/recovery.c empty_offs = ALIGN(offs + 1, c->min_io_size); ALIGN 411 fs/ubifs/recovery.c empty_offs = ALIGN(*offs, c->min_io_size); ALIGN 443 fs/ubifs/recovery.c next_offs = ALIGN(offs + dlen, 8); ALIGN 449 fs/ubifs/recovery.c skip = ALIGN(offs + 1, c->min_io_size) - offs; ALIGN 526 fs/ubifs/recovery.c int len = ALIGN(endpt, c->min_io_size); ALIGN 536 fs/ubifs/recovery.c int pad_len = len - ALIGN(endpt, 8); ALIGN 635 fs/ubifs/recovery.c node_len = ALIGN(le32_to_cpu(ch->len), 8); ALIGN 983 fs/ubifs/recovery.c node_len = ALIGN(le32_to_cpu(ch->len), 8); ALIGN 1015 fs/ubifs/recovery.c len = ALIGN(ucleb->endpt, c->min_io_size); ALIGN 1017 fs/ubifs/recovery.c int pad_len = len - ALIGN(ucleb->endpt, 8); ALIGN 1433 fs/ubifs/recovery.c len = ALIGN(len + 1, c->min_io_size); ALIGN 362 fs/ubifs/replay.c *used += ALIGN(len, 8); ALIGN 434 fs/ubifs/replay.c *used += ALIGN(len, 8); ALIGN 157 fs/ubifs/sb.c tmp = ALIGN(UBIFS_SB_NODE_SZ, c->min_io_size); ALIGN 215 fs/ubifs/sb.c mst->ihead_offs = cpu_to_le32(ALIGN(tmp, c->min_io_size)); ALIGN 216 fs/ubifs/sb.c mst->index_size = cpu_to_le64(ALIGN(tmp, 8)); ALIGN 232 fs/ubifs/sb.c tmp64 -= ALIGN(ubifs_idx_node_sz(c, 1), c->min_io_size); ALIGN 233 fs/ubifs/sb.c tmp64 -= ALIGN(UBIFS_INO_NODE_SZ, c->min_io_size); ALIGN 236 fs/ubifs/sb.c tmp64 = ALIGN(ubifs_idx_node_sz(c, 1), c->min_io_size); ALIGN 237 fs/ubifs/sb.c ino_waste = ALIGN(UBIFS_INO_NODE_SZ, c->min_io_size) - ALIGN 240 fs/ubifs/sb.c tmp64 -= ALIGN(ubifs_idx_node_sz(c, 1), 8); ALIGN 265 fs/ubifs/sb.c idx = kzalloc(ALIGN(tmp, c->min_io_size), GFP_KERNEL); ALIGN 289 fs/ubifs/sb.c tmp = ALIGN(UBIFS_INO_NODE_SZ, c->min_io_size); ALIGN 326 fs/ubifs/sb.c tmp = ALIGN(UBIFS_CS_NODE_SZ, c->min_io_size); ALIGN 479 fs/ubifs/sb.c sup = kmalloc(ALIGN(UBIFS_SB_NODE_SZ, c->min_io_size), GFP_NOFS); ALIGN 502 fs/ubifs/sb.c int len = ALIGN(UBIFS_SB_NODE_SZ, c->min_io_size); ALIGN 119 fs/ubifs/scan.c pad_len, ALIGN(offs + node_len + pad_len, 8)); ALIGN 182 fs/ubifs/scan.c sleb->endpt = ALIGN(offs, c->min_io_size); ALIGN 314 fs/ubifs/scan.c node_len = ALIGN(le32_to_cpu(ch->len), 8); ALIGN 485 fs/ubifs/super.c c->ref_node_alsz = ALIGN(UBIFS_REF_NODE_SZ, c->min_io_size); ALIGN 486 fs/ubifs/super.c c->mst_node_alsz = ALIGN(UBIFS_MST_NODE_SZ, c->min_io_size); ALIGN 532 fs/ubifs/super.c c->dead_wm = ALIGN(MIN_WRITE_SZ, c->min_io_size); ALIGN 533 fs/ubifs/super.c c->dark_wm = ALIGN(UBIFS_MAX_NODE_SZ, c->min_io_size); ALIGN 584 fs/ubifs/super.c c->min_idx_node_sz = ALIGN(tmp, 8); ALIGN 588 fs/ubifs/super.c c->max_idx_node_sz = ALIGN(tmp, 8); ALIGN 592 fs/ubifs/super.c tmp = ALIGN(tmp, c->min_io_size); ALIGN 1019 fs/ubifs/super.c sz = ALIGN(c->max_idx_node_sz, c->min_io_size); ALIGN 1020 fs/ubifs/super.c sz = ALIGN(sz + c->max_idx_node_sz, c->min_io_size); ALIGN 255 fs/ubifs/tnc.c c->calc_idx_sz -= ALIGN(dirt, 8); ALIGN 85 fs/ubifs/tnc_commit.c c->calc_idx_sz += ALIGN(len, 8); ALIGN 126 fs/ubifs/tnc_commit.c const int alen = ALIGN(len, 8); ALIGN 144 fs/ubifs/tnc_commit.c c->ileb_len = ALIGN(gap_pos, c->min_io_size); ALIGN 270 fs/ubifs/tnc_commit.c dirt += ALIGN(snod->len, 8); ALIGN 285 fs/ubifs/tnc_commit.c gap_start = ALIGN(snod->offs + snod->len, 8); ALIGN 428 fs/ubifs/tnc_commit.c buf_len = ALIGN(buf_len, c->min_io_size); ALIGN 478 fs/ubifs/tnc_commit.c c->calc_idx_sz += ALIGN(len, 8); ALIGN 497 fs/ubifs/tnc_commit.c buf_offs += ALIGN(len, 8); ALIGN 517 fs/ubifs/tnc_commit.c used += ALIGN(len, 8); ALIGN 518 fs/ubifs/tnc_commit.c avail -= ALIGN(len, 8); ALIGN 529 fs/ubifs/tnc_commit.c blen = ALIGN(wlen, c->min_io_size); ALIGN 847 fs/ubifs/tnc_commit.c buf_len = ALIGN(c->max_idx_node_sz, c->min_io_size); ALIGN 929 fs/ubifs/tnc_commit.c used += ALIGN(len, 8); ALIGN 930 fs/ubifs/tnc_commit.c avail -= ALIGN(len, 8); ALIGN 950 fs/ubifs/tnc_commit.c buf_offs += ALIGN(wlen, 8); ALIGN 972 fs/ubifs/tnc_commit.c wlen = ALIGN(wlen, 8); ALIGN 973 fs/ubifs/tnc_commit.c blen = ALIGN(wlen, c->min_io_size); ALIGN 121 fs/ubifs/ubifs.h #define CALC_DENT_SIZE(name_len) ALIGN(UBIFS_DENT_NODE_SZ + (name_len) + 1, 8) ALIGN 124 fs/ubifs/ubifs.h #define CALC_XATTR_BYTES(data_len) ALIGN(UBIFS_INO_NODE_SZ + (data_len) + 1, 8) ALIGN 106 fs/ubifs/xattr.c .new_ino_d = ALIGN(size, 8), .dirtied_ino = 1, ALIGN 107 fs/ubifs/xattr.c .dirtied_ino_d = ALIGN(host_ui->data_len, 8) }; ALIGN 198 fs/ubifs/xattr.c .dirtied_ino_d = ALIGN(size, 8) + ALIGN(host_ui->data_len, 8) }; ALIGN 491 fs/ubifs/xattr.c .dirtied_ino_d = ALIGN(host_ui->data_len, 8) }; ALIGN 276 fs/xfs/linux-2.6/xfs_file.c reclen = ALIGN(sizeof(struct hack_dirent) + namlen, sizeof(u64)); ALIGN 347 fs/xfs/linux-2.6/xfs_file.c reclen = ALIGN(sizeof(struct hack_dirent) + de->namlen, ALIGN 10 include/asm-generic/vmlinux.lds.h #define ALIGN_FUNCTION() . = ALIGN(8) ALIGN 52 include/asm-generic/vmlinux.lds.h . = ALIGN(8); \ ALIGN 58 include/asm-generic/vmlinux.lds.h . = ALIGN((align)); \ ALIGN 204 include/asm-generic/vmlinux.lds.h . = ALIGN((align)); \ ALIGN 207 include/asm-generic/vmlinux.lds.h . = ALIGN((align)); ALIGN 334 include/asm-generic/vmlinux.lds.h . = ALIGN(8); \ ALIGN 346 include/asm-generic/vmlinux.lds.h . = ALIGN(4); \ ALIGN 385 include/asm-generic/vmlinux.lds.h . = ALIGN(align); \ ALIGN 34 include/asm-m32r/assembler.h ALIGN ALIGN 187 include/asm-m32r/assembler.h ALIGN ALIGN 16 include/asm-m68k/dvma.h #define DVMA_PAGE_ALIGN(addr) ALIGN(addr, DVMA_PAGE_SIZE) ALIGN 17 include/asm-parisc/linkage.h ALIGN !\ ALIGN 155 include/crypto/algapi.h return (void *)ALIGN(addr, align); ALIGN 151 include/linux/a.out.h #define _N_SEGMENT_ROUND(x) ALIGN(x, SEGMENT_SIZE) ALIGN 8 include/linux/cache.h #define L1_CACHE_ALIGN(x) ALIGN(x, L1_CACHE_BYTES) ALIGN 42 include/linux/kernel.h #define PTR_ALIGN(p, a) ((typeof(p))ALIGN((unsigned long)(p), (a))) ALIGN 36 include/linux/kexec.h #define KEXEC_NOTE_HEAD_BYTES ALIGN(sizeof(struct elf_note), 4) ALIGN 38 include/linux/kexec.h #define KEXEC_CORE_NOTE_NAME_BYTES ALIGN(sizeof(KEXEC_CORE_NOTE_NAME), 4) ALIGN 39 include/linux/kexec.h #define KEXEC_CORE_NOTE_DESC_BYTES ALIGN(sizeof(struct elf_prstatus), 4) ALIGN 190 include/linux/kexec.h #define VMCOREINFO_NOTE_NAME_BYTES ALIGN(sizeof(VMCOREINFO_NOTE_NAME), 4) ALIGN 59 include/linux/linkage.h ALIGN; \ ALIGN 46 include/linux/mm.h #define PAGE_ALIGN(addr) ALIGN(addr, PAGE_SIZE) ALIGN 1433 include/linux/skbuff.h return pskb_expand_head(skb, ALIGN(delta, NET_SKB_PAD), 0, ALIGN 114 include/net/neighbour.h unsigned char ha[ALIGN(MAX_ADDR_LEN, sizeof(unsigned long))]; ALIGN 95 include/scsi/scsi_transport.h t->target_private_offset = ALIGN(t->target_size, sizeof(void *)); ALIGN 102 include/scsi/scsi_transport.h t->device_private_offset = ALIGN(t->device_size, sizeof(void *)); ALIGN 399 init/main.c size = ALIGN(PERCPU_ENOUGH_ROOM, PAGE_SIZE); ALIGN 689 kernel/compat.c bitmap_size = ALIGN(bitmap_size, BITS_PER_COMPAT_LONG); ALIGN 730 kernel/compat.c bitmap_size = ALIGN(bitmap_size, BITS_PER_COMPAT_LONG); ALIGN 998 kernel/compat.c size = ALIGN(nr_bits, BITS_PER_LONG) / 8; ALIGN 394 kernel/module.c extra = ALIGN((unsigned long)ptr, align) - (unsigned long)ptr; ALIGN 1095 kernel/module.c size[0] = ALIGN(sizeof(*sect_attrs) ALIGN 1565 kernel/module.c ret = ALIGN(*size, sechdr->sh_addralign ?: 1); ALIGN 468 kernel/params.c size[0] = ALIGN(sizeof(*mp) + ALIGN 320 kernel/resource.c new->start = ALIGN(new->start, align); ALIGN 303 lib/bitmap.c i = ALIGN(nmaskbits, CHUNKSZ) - CHUNKSZ; ALIGN 325 lib/bitmap.c unsigned int nr_nibbles = ALIGN(nr_bits, 4) / 4; ALIGN 326 lib/bitmap.c return nr_nibbles + ALIGN(nr_nibbles, CHUNKSZ / 4) / (CHUNKSZ / 4) - 1; ALIGN 118 lib/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); ALIGN 140 lib/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); ALIGN 194 lib/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); ALIGN 305 lib/swiotlb.c offset_slots = ALIGN(start_dma_addr, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT; ALIGN 307 lib/swiotlb.c ? ALIGN(mask + 1, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT ALIGN 314 lib/swiotlb.c nslots = ALIGN(size, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT; ALIGN 327 lib/swiotlb.c index = ALIGN(io_tlb_index, stride); ALIGN 396 lib/swiotlb.c int i, count, nslots = ALIGN(size, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT; ALIGN 418 lib/swiotlb.c count = ((index + nslots) < ALIGN(index + 1, IO_TLB_SEGSIZE) ? ALIGN 58 mm/bootmem.c return ALIGN(bytes, sizeof(long)); ALIGN 418 mm/bootmem.c return ALIGN(base + idx, step) - base; ALIGN 428 mm/bootmem.c return ALIGN(base + off, align) - base; ALIGN 463 mm/bootmem.c start = ALIGN(goal, step); ALIGN 465 mm/bootmem.c start = ALIGN(min, step); ALIGN 149 mm/dmapool.c size = ALIGN(size, align); ALIGN 1049 mm/mempolicy.c unsigned long copy = ALIGN(maxnode-1, 64) / 8; ALIGN 1213 mm/mempolicy.c alloc_size = ALIGN(nr_bits, BITS_PER_LONG) / 8; ALIGN 1223 mm/mempolicy.c err |= clear_user(nmask, ALIGN(maxnode-1, 8) / 8); ALIGN 1239 mm/mempolicy.c alloc_size = ALIGN(nr_bits, BITS_PER_LONG) / 8; ALIGN 1263 mm/mempolicy.c alloc_size = ALIGN(nr_bits, BITS_PER_LONG) / 8; ALIGN 3506 mm/page_alloc.c end = ALIGN(end, MAX_ORDER_NR_PAGES); ALIGN 800 mm/slab.c return ALIGN(sizeof(struct slab)+nr_objs*sizeof(kmem_bufctl_t), align); ALIGN 1493 mm/slab.c cache_cache.buffer_size = ALIGN(cache_cache.buffer_size, ALIGN 1507 mm/slab.c cache_cache.slab_size = ALIGN(cache_cache.num * sizeof(kmem_bufctl_t) + ALIGN 2317 mm/slab.c size = ALIGN(size, align); ALIGN 2328 mm/slab.c slab_size = ALIGN(cachep->num * sizeof(kmem_bufctl_t) ALIGN 262 mm/slob.c aligned = (slob_t *)ALIGN((unsigned long)cur, align); ALIGN 1900 mm/slub.c return ALIGN(align, sizeof(void *)); ALIGN 1975 mm/slub.c ALIGN(sizeof(struct kmem_cache_cpu), cache_line_size()), ALIGN 2192 mm/slub.c size = ALIGN(size, sizeof(void *)); ALIGN 2267 mm/slub.c size = ALIGN(size, align); ALIGN 3067 mm/slub.c size = ALIGN(size, sizeof(void *)); ALIGN 3069 mm/slub.c size = ALIGN(size, align); ALIGN 3120 mm/slub.c s->inuse = max_t(int, s->inuse, ALIGN(size, sizeof(void *))); ALIGN 235 mm/vmalloc.c addr = ALIGN(start, align); ALIGN 254 mm/vmalloc.c addr = ALIGN(tmp->size + ALIGN 262 mm/vmalloc.c addr = ALIGN(tmp->size + (unsigned long)tmp->addr, align); ALIGN 104 net/compat.c #define CMSG_COMPAT_ALIGN(len) ALIGN((len), sizeof(s32)) ALIGN 42 net/ipv4/esp4.c len = ALIGN(len, crypto_tfm_ctx_alignment()); ALIGN 46 net/ipv4/esp4.c len = ALIGN(len, __alignof__(struct scatterlist)); ALIGN 83 net/ipv4/esp4.c return (void *)ALIGN((unsigned long)(req + 1) + ALIGN 91 net/ipv4/esp4.c return (void *)ALIGN((unsigned long)(req + 1) + ALIGN 133 net/ipv4/esp4.c blksize = ALIGN(crypto_aead_blocksize(aead), 4); ALIGN 134 net/ipv4/esp4.c clen = ALIGN(clen + 2, blksize); ALIGN 136 net/ipv4/esp4.c clen = ALIGN(clen, esp->padlen); ALIGN 388 net/ipv4/esp4.c u32 blksize = ALIGN(crypto_aead_blocksize(esp->aead), 4); ALIGN 595 net/ipv4/esp4.c align = ALIGN(crypto_aead_blocksize(aead), 4); ALIGN 638 net/ipv4/tcp.c size = ALIGN(size, 4); ALIGN 256 net/ipv4/tcp_output.c new_win = ALIGN(cur_win, 1 << tp->rx_opt.rcv_wscale); ALIGN 68 net/ipv6/esp6.c len = ALIGN(len, crypto_tfm_ctx_alignment()); ALIGN 72 net/ipv6/esp6.c len = ALIGN(len, __alignof__(struct scatterlist)); ALIGN 109 net/ipv6/esp6.c return (void *)ALIGN((unsigned long)(req + 1) + ALIGN 117 net/ipv6/esp6.c return (void *)ALIGN((unsigned long)(req + 1) + ALIGN 157 net/ipv6/esp6.c blksize = ALIGN(crypto_aead_blocksize(aead), 4); ALIGN 158 net/ipv6/esp6.c clen = ALIGN(clen + 2, blksize); ALIGN 160 net/ipv6/esp6.c clen = ALIGN(clen, esp->padlen); ALIGN 339 net/ipv6/esp6.c u32 blksize = ALIGN(crypto_aead_blocksize(esp->aead), 4); ALIGN 537 net/ipv6/esp6.c align = ALIGN(crypto_aead_blocksize(aead), 4); ALIGN 54 net/netfilter/nf_conntrack_extend.c off = ALIGN(sizeof(struct nf_ct_ext), t->align); ALIGN 94 net/netfilter/nf_conntrack_extend.c newoff = ALIGN(ct->ext->len, t->align); ALIGN 145 net/netfilter/nf_conntrack_extend.c + ALIGN(sizeof(struct nf_ct_ext), t1->align) ALIGN 153 net/netfilter/nf_conntrack_extend.c t1->alloc_size = ALIGN(t1->alloc_size, t2->align) ALIGN 172 net/netfilter/nf_conntrack_extend.c type->alloc_size = ALIGN(sizeof(struct nf_ct_ext), type->align) ALIGN 64 net/netlink/af_netlink.c #define NLGRPSZ(x) (ALIGN(x, sizeof(unsigned long) * 8) / 8) ALIGN 371 net/sctp/tsnmap.c inc = ALIGN((gap - map->len),BITS_PER_LONG) + SCTP_TSN_MAP_INCREMENT; ALIGN 30 sound/core/sgbuf.c #define sgbuf_align_table(tbl) ALIGN((tbl), SGBUF_TBL_ALIGN) ALIGN 166 sound/isa/gus/gus_mem.c ptr1 = ALIGN(pblock->ptr + pblock->size, align); ALIGN 1070 sound/isa/wavefront/wavefront_synth.c blocksize = ALIGN(length - written, 8); ALIGN 718 sound/pci/bt87x.c ALIGN(255 * 4092, 1024)); ALIGN 1350 sound/pci/es1968.c size = ALIGN(size, ESM_MEM_ALIGN); ALIGN 2224 sound/pci/maestro3.c data_bytes = ALIGN(data_bytes, 256); ALIGN 3610 sound/pci/rme9652/hdsp.c cb_bus = ALIGN(hdsp->capture_dma_buf.addr, 0x10000ul); ALIGN 3611 sound/pci/rme9652/hdsp.c pb_bus = ALIGN(hdsp->playback_dma_buf.addr, 0x10000ul); ALIGN 1806 sound/pci/rme9652/rme9652.c cb_bus = ALIGN(rme9652->capture_dma_buf.addr, 0x10000ul); ALIGN 1807 sound/pci/rme9652/rme9652.c pb_bus = ALIGN(rme9652->playback_dma_buf.addr, 0x10000ul); ALIGN 3367 sound/pci/trident/trident_main.c trident->tlb.entries = (unsigned int*)ALIGN((unsigned long)trident->tlb.buffer.area, SNDRV_TRIDENT_MAX_PAGES * 4); ALIGN 3368 sound/pci/trident/trident_main.c trident->tlb.entries_dmaaddr = ALIGN(trident->tlb.buffer.addr, SNDRV_TRIDENT_MAX_PAGES * 4); ALIGN 2086 sound/pci/ymfpci/ymfpci_main.c size = ALIGN(playback_ctrl_size, 0x100) + ALIGN 2087 sound/pci/ymfpci/ymfpci_main.c ALIGN(chip->bank_size_playback * 2 * YDSXG_PLAYBACK_VOICES, 0x100) + ALIGN 2088 sound/pci/ymfpci/ymfpci_main.c ALIGN(chip->bank_size_capture * 2 * YDSXG_CAPTURE_VOICES, 0x100) + ALIGN 2089 sound/pci/ymfpci/ymfpci_main.c ALIGN(chip->bank_size_effect * 2 * YDSXG_EFFECT_VOICES, 0x100) + ALIGN 2104 sound/pci/ymfpci/ymfpci_main.c ptr += ALIGN(playback_ctrl_size, 0x100); ALIGN 2105 sound/pci/ymfpci/ymfpci_main.c ptr_addr += ALIGN(playback_ctrl_size, 0x100); ALIGN 2116 sound/pci/ymfpci/ymfpci_main.c ptr = (char *)ALIGN((unsigned long)ptr, 0x100); ALIGN 2117 sound/pci/ymfpci/ymfpci_main.c ptr_addr = ALIGN(ptr_addr, 0x100); ALIGN 2126 sound/pci/ymfpci/ymfpci_main.c ptr = (char *)ALIGN((unsigned long)ptr, 0x100); ALIGN 2127 sound/pci/ymfpci/ymfpci_main.c ptr_addr = ALIGN(ptr_addr, 0x100); ALIGN 2136 sound/pci/ymfpci/ymfpci_main.c ptr = (char *)ALIGN((unsigned long)ptr, 0x100); ALIGN 2137 sound/pci/ymfpci/ymfpci_main.c ptr_addr = ALIGN(ptr_addr, 0x100); ALIGN 659 sound/soc/codecs/cs4270.c codec = kzalloc(ALIGN(sizeof(struct snd_soc_codec), 4) + ALIGN 675 sound/soc/codecs/cs4270.c ALIGN(sizeof(struct snd_soc_codec), 4); ALIGN 545 virt/kvm/kvm_main.c unsigned dirty_bytes = ALIGN(npages, BITS_PER_LONG) / 8; ALIGN 623 virt/kvm/kvm_main.c n = ALIGN(memslot->npages, BITS_PER_LONG) / 8;