s64 278 arch/x86/kernel/kprobes.c s64 disp; s64 336 arch/x86/kernel/kprobes.c BUG_ON((s64) (s32) disp != disp); /* Sanity check. */ s64 113 arch/x86/kernel/module_64.c if ((s64)val != *(s32 *)loc) s64 120 arch/x86/kernel/module_64.c if ((s64)val != *(s32 *)loc) s64 105 arch/x86/kvm/i8254.c s64 d, t; s64 135 arch/x86/kvm/i8254.c s64 d, t; s64 260 arch/x86/kvm/i8254.c s64 interval; s64 9 arch/x86/kvm/i8254.h s64 period; /* unit: ns */ s64 10 arch/x86/kvm/i8254.h s64 scheduled; s64 13 arch/x86/kvm/lapic.h s64 period; /* unit: ns */ s64 1926 arch/x86/kvm/x86_emulate.c rel = insn_fetch(s64, 8, c->eip); s64 118 arch/x86/xen/time.c s64 blocked, runnable, offline, stolen; s64 169 arch/x86/xen/time.c s64 offset; s64 291 arch/x86/xen/time.c static s64 get_abs_timeout(unsigned long delta) s64 1593 fs/compat.c s64 timeout = -1; s64 1650 fs/compat.c s64 timeout = MAX_SCHEDULE_TIMEOUT; s64 1762 fs/compat.c s64 timeout = -1; s64 587 fs/ext4/balloc.c s64 free_blocks, dirty_blocks; s64 588 fs/ext4/balloc.c s64 root_blocks = 0; s64 634 fs/ext4/balloc.c s64 free_blocks, dirty_blocks; s64 635 fs/ext4/balloc.c s64 root_blocks = 0; s64 2473 fs/ext4/inode.c s64 free_blocks, dirty_blocks; s64 301 fs/gfs2/incore.h s64 qd_change; s64 302 fs/gfs2/incore.h s64 qd_change_sync; s64 371 fs/gfs2/incore.h s64 sc_total; s64 372 fs/gfs2/incore.h s64 sc_free; s64 373 fs/gfs2/incore.h s64 sc_dinodes; s64 71 fs/gfs2/quota.c s64 qu_value; s64 541 fs/gfs2/quota.c s64 x; s64 616 fs/gfs2/quota.c s64 value; s64 884 fs/gfs2/quota.c s64 value; s64 902 fs/gfs2/quota.c else if ((s64)be64_to_cpu(qd->qd_qb.qb_value) >= s64 903 fs/gfs2/quota.c (s64)be64_to_cpu(qd->qd_qb.qb_limit)) s64 908 fs/gfs2/quota.c value += (s64)be64_to_cpu(qd->qd_qb.qb_value); s64 909 fs/gfs2/quota.c if (value < (s64)be64_to_cpu(qd->qd_qb.qb_limit)) s64 968 fs/gfs2/quota.c s64 value; s64 985 fs/gfs2/quota.c value = (s64)be64_to_cpu(qd->qd_qb.qb_value); s64 990 fs/gfs2/quota.c if (be64_to_cpu(qd->qd_qb.qb_limit) && (s64)be64_to_cpu(qd->qd_qb.qb_limit) < value) { s64 995 fs/gfs2/quota.c (s64)be64_to_cpu(qd->qd_qb.qb_warn) < value && s64 1414 fs/gfs2/rgrp.c gfs2_statfs_change(sdp, 0, -(s64)*n, 0); s64 1492 fs/gfs2/rgrp.c gfs2_quota_change(ip, -(s64)blen, ip->i_inode.i_uid, ip->i_inode.i_gid); s64 1520 fs/gfs2/rgrp.c gfs2_quota_change(ip, -(s64)blen, ip->i_inode.i_uid, ip->i_inode.i_gid); s64 185 fs/jfs/inode.c s64 lblock64 = lblock; s64 188 fs/jfs/inode.c s64 xaddr; s64 112 fs/jfs/jfs_btree.h s64 bn; /* 8: */ s64 106 fs/jfs/jfs_dmap.c s64 dbMapFileSizeToMapSize(struct inode *ipbmap); s64 336 fs/jfs/jfs_dmap.c s64 lblkno, rem; s64 423 fs/jfs/jfs_dmap.c s64 lblkno, rem, lastlblkno; s64 473 fs/jfs/jfs_dmap.c nblks = min(rem, (s64)BPERDMAP - dbitno); s64 599 fs/jfs/jfs_dmap.c s64 avgfree; s64 601 fs/jfs/jfs_dmap.c s64 hwm = 0; s64 700 fs/jfs/jfs_dmap.c s64 lblkno, blkno; s64 703 fs/jfs/jfs_dmap.c s64 mapSize; s64 893 fs/jfs/jfs_dmap.c s64 lblkno; s64 910 fs/jfs/jfs_dmap.c if (nblocks > ((s64) 1 << bmp->db_maxfreebud)) { s64 1023 fs/jfs/jfs_dmap.c s64 lblkno, lastblkno, extblkno; s64 1369 fs/jfs/jfs_dmap.c s64 blkno, lblkno; s64 1385 fs/jfs/jfs_dmap.c blkno = (s64) agno << bmp->db_agl2size; s64 1488 fs/jfs/jfs_dmap.c ((s64) (ti - le32_to_cpu(dcp->leafidx))) << budmin; s64 1568 fs/jfs/jfs_dmap.c s64 blkno = 0; s64 1622 fs/jfs/jfs_dmap.c s64 b, lblkno; s64 1676 fs/jfs/jfs_dmap.c b += (((s64) leafidx) << budmin); s64 1741 fs/jfs/jfs_dmap.c s64 b, lblkno, n; s64 1797 fs/jfs/jfs_dmap.c nb = min(n, (s64)BPERDMAP); s64 1893 fs/jfs/jfs_dmap.c s64 blkno; s64 2406 fs/jfs/jfs_dmap.c s64 lblkno; s64 3058 fs/jfs/jfs_dmap.c s64 mask; /* meant to be signed */ s64 3060 fs/jfs/jfs_dmap.c mask = (s64) 1 << (64 - 1); s64 3108 fs/jfs/jfs_dmap.c s64 lblkno, rem; s64 3287 fs/jfs/jfs_dmap.c s64 newsize; s64 3288 fs/jfs/jfs_dmap.c s64 p; s64 3295 fs/jfs/jfs_dmap.c s64 ag_rem; s64 3457 fs/jfs/jfs_dmap.c n = min(nblocks, (s64)BPERDMAP - n); s64 3465 fs/jfs/jfs_dmap.c n = min(nblocks, (s64)BPERDMAP); s64 3559 fs/jfs/jfs_dmap.c s64 ag_rem, actfree, inactfree, avgfree; s64 3905 fs/jfs/jfs_dmap.c s64 sz; s64 3906 fs/jfs/jfs_dmap.c s64 m; s64 3919 fs/jfs/jfs_dmap.c sz = (s64) 1 << l2sz; s64 3951 fs/jfs/jfs_dmap.c s64 dbMapFileSizeToMapSize(struct inode * ipbmap) s64 3954 fs/jfs/jfs_dmap.c s64 nblocks; s64 3955 fs/jfs/jfs_dmap.c s64 npages, ndmaps; s64 54 fs/jfs/jfs_dmap.h #define MAXL0SIZE ((s64)1 << L2MAXL0SIZE) s64 55 fs/jfs/jfs_dmap.h #define MAXL1SIZE ((s64)1 << L2MAXL1SIZE) s64 56 fs/jfs/jfs_dmap.h #define MAXL2SIZE ((s64)1 << L2MAXL2SIZE) s64 140 fs/jfs/jfs_dmap.h ((s64)(a) << (JFS_SBI((ip)->i_sb)->bmap->db_agl2size)) s64 224 fs/jfs/jfs_dmap.h s64 dn_mapsize; /* number of blocks in aggregate */ s64 225 fs/jfs/jfs_dmap.h s64 dn_nfree; /* num free blks in aggregate map */ s64 236 fs/jfs/jfs_dmap.h s64 dn_agfree[MAXAG]; /* per AG free count */ s64 237 fs/jfs/jfs_dmap.h s64 dn_agsize; /* num of blks per alloc group */ s64 285 fs/jfs/jfs_dmap.h (((b) & (((s64)1 << ((m) + L2LPERCTL)) - 1)) >> (m)) s64 313 fs/jfs/jfs_dmap.h extern s64 dbMapFileSizeToMapSize(struct inode *ipbmap); s64 210 fs/jfs/jfs_dtree.c s64 xaddr; s64 229 fs/jfs/jfs_dtree.c s64 xaddr; s64 252 fs/jfs/jfs_dtree.c s64 blkno; s64 253 fs/jfs/jfs_dtree.c s64 offset; s64 344 fs/jfs/jfs_dtree.c s64 offset; s64 347 fs/jfs/jfs_dtree.c s64 xaddr; s64 498 fs/jfs/jfs_dtree.c s64 lblock; s64 551 fs/jfs/jfs_dtree.c s64 lblock; s64 584 fs/jfs/jfs_dtree.c s64 bn; s64 826 fs/jfs/jfs_dtree.c s64 bn; s64 943 fs/jfs/jfs_dtree.c s64 xaddr, nxaddr; s64 985 fs/jfs/jfs_dtree.c if ((rc = dbAlloc(ip, 0, (s64) xlen, &xaddr))) { s64 1036 fs/jfs/jfs_dtree.c if ((rc = dbReAlloc(sbi->ipbmap, xaddr, (s64) xlen, s64 1037 fs/jfs/jfs_dtree.c (s64) n, &nxaddr))) s64 1051 fs/jfs/jfs_dtree.c dbFree(ip, nxaddr, (s64) xlen); s64 1056 fs/jfs/jfs_dtree.c dbFree(ip, xaddr, (s64) n); s64 1082 fs/jfs/jfs_dtree.c if ((rc = dbAlloc(ip, 0, (s64) xlen, &xaddr)) == 0) { s64 1304 fs/jfs/jfs_dtree.c dbFree(ip, addressPXD(pxd), (s64) lengthPXD(pxd)); s64 1338 fs/jfs/jfs_dtree.c s64 rbn; /* new right page block number */ s64 1341 fs/jfs/jfs_dtree.c s64 nextbn; s64 1581 fs/jfs/jfs_dtree.c s64 lblock; s64 1658 fs/jfs/jfs_dtree.c s64 xaddr, txaddr; s64 1705 fs/jfs/jfs_dtree.c s64 lblock; s64 1882 fs/jfs/jfs_dtree.c s64 rbn; s64 1993 fs/jfs/jfs_dtree.c s64 lblock; s64 2083 fs/jfs/jfs_dtree.c s64 bn; s64 2202 fs/jfs/jfs_dtree.c s64 lblock; s64 2448 fs/jfs/jfs_dtree.c s64 bn; s64 2452 fs/jfs/jfs_dtree.c s64 oxaddr, nextbn, prevbn; s64 2643 fs/jfs/jfs_dtree.c s64 bn; s64 2748 fs/jfs/jfs_dtree.c s64 nextbn, prevbn; s64 3013 fs/jfs/jfs_dtree.c s64 bn; s64 3335 fs/jfs/jfs_dtree.c s64 bn; s64 3413 fs/jfs/jfs_dtree.c s64 bn; s64 3909 fs/jfs/jfs_dtree.c s64 bn = 0; s64 4020 fs/jfs/jfs_dtree.c s64 lblock; s64 4513 fs/jfs/jfs_dtree.c s64 bn; s64 133 fs/jfs/jfs_dtree.h ( ((s64)((dts)->addr1)) << 32 | __le32_to_cpu((dts)->addr2) ) s64 35 fs/jfs/jfs_extent.c static s64 extRoundDown(s64 nb); s64 89 fs/jfs/jfs_extent.c s64 nxlen, nxaddr, xoff, hint, xaddr = 0; s64 145 fs/jfs/jfs_extent.c dbFree(ip, nxaddr, (s64) nxlen); s64 217 fs/jfs/jfs_extent.c s64 xaddr, xlen, nxaddr, delta, xoff; s64 218 fs/jfs/jfs_extent.c s64 ntail, nextend, ninsert; s64 260 fs/jfs/jfs_extent.c dbFree(ip, nxaddr, (s64) nxlen); s64 326 fs/jfs/jfs_extent.c dbFree(ip, xaddr, (s64) ninsert); s64 368 fs/jfs/jfs_extent.c s64 prev; s64 471 fs/jfs/jfs_extent.c s64 blkno = offsetXAD(xp) >> ip->i_blkbits; s64 525 fs/jfs/jfs_extent.c s64 nb, nblks, daddr, max; s64 536 fs/jfs/jfs_extent.c max = (s64) 1 << bmp->db_maxfreebud; s64 648 fs/jfs/jfs_extent.c static s64 extRoundDown(s64 nb) s64 136 fs/jfs/jfs_filsys.h #define MAXFILESIZE ((s64)1 << 52) s64 152 fs/jfs/jfs_filsys.h #define SIZE2PN(size) ( ((s64)((size) - 1)) >> (L2PSIZE) ) s64 154 fs/jfs/jfs_filsys.h #define SIZE2BN(size, l2bsize) ( ((s64)((size) - 1)) >> (l2bsize) ) s64 317 fs/jfs/jfs_imap.c s64 blkno, agstart; s64 605 fs/jfs/jfs_imap.c s64 blkno; s64 2182 fs/jfs/jfs_imap.c s64 blkno, hint; s64 2274 fs/jfs/jfs_imap.c hint = ((s64) agno << sbi->bmap->db_agl2size) - 1; s64 2279 fs/jfs/jfs_imap.c if ((rc = dbAlloc(ipimap, hint, (s64) imap->im_nbperiext, &blkno))) s64 2467 fs/jfs/jfs_imap.c s64 xaddr = 0; s64 2468 fs/jfs/jfs_imap.c s64 blkno; s64 2528 fs/jfs/jfs_imap.c if ((rc = dbAlloc(ipimap, 0, (s64) xlen, &xaddr))) { s64 2550 fs/jfs/jfs_imap.c dbFree(ipimap, xaddr, (s64) xlen); s64 2683 fs/jfs/jfs_imap.c s64 blkno; s64 2876 fs/jfs/jfs_imap.c s64 agstart; s64 375 fs/jfs/jfs_logmgr.h s64 base; /* @8: log extent address (inline log ) */ s64 462 fs/jfs/jfs_logmgr.h s64 l_blkno; /* 8: log page block number */ s64 258 fs/jfs/jfs_metapage.c s64 xaddr; s64 268 fs/jfs/jfs_metapage.c rc = xtLookup(inode, (s64)lblock, *len, &xflag, &xaddr, len, 0); s64 309 fs/jfs/jfs_mount.c s64 AIM_byte_addr, AIT_byte_addr, fsckwsp_addr; s64 310 fs/jfs/jfs_mount.c s64 byte_addr_diff0, byte_addr_diff1; s64 2447 fs/jfs/jfs_txnmgr.c s64 xaddr; s64 2465 fs/jfs/jfs_txnmgr.c (s64) xlen, tblk); s64 2475 fs/jfs/jfs_txnmgr.c dbUpdatePMap(ipbmap, false, xaddr, (s64) xlen, tblk); s64 2484 fs/jfs/jfs_txnmgr.c dbUpdatePMap(ipbmap, false, xaddr, (s64) xlen, s64 2505 fs/jfs/jfs_txnmgr.c s64 xaddr; s64 2527 fs/jfs/jfs_txnmgr.c (s64) xlen, tblk); s64 2537 fs/jfs/jfs_txnmgr.c dbUpdatePMap(ipbmap, true, xaddr, (s64) xlen, s64 2549 fs/jfs/jfs_txnmgr.c (s64) xlen, tblk); s64 2566 fs/jfs/jfs_txnmgr.c dbFree(ip, xaddr, (s64) xlen); s64 2575 fs/jfs/jfs_txnmgr.c dbFree(ip, xaddr, (s64) xlen); s64 2585 fs/jfs/jfs_txnmgr.c dbFree(ip, xaddr, (s64) xlen); s64 262 fs/jfs/jfs_txnmgr.h s64 pad; /* 8: Force 64-bit xdlist size */ s64 73 fs/jfs/jfs_types.h (lxd)->off1 = ((s64)offset64) >> 32;\ s64 80 fs/jfs/jfs_types.h ( ((s64)((lxd)->off1)) << 32 | (lxd)->off2 ) s64 103 fs/jfs/jfs_types.h (pxd)->addr1 = ((s64)address64) >> 32;\ s64 110 fs/jfs/jfs_types.h ( ((s64)((pxd)->addr1)) << 32 | __le32_to_cpu((pxd)->addr2)) s64 94 fs/jfs/jfs_xtree.c s64 off; s64 95 fs/jfs/jfs_xtree.c s64 addr; s64 151 fs/jfs/jfs_xtree.c s64 bn; s64 156 fs/jfs/jfs_xtree.c s64 next, size, xoff, xend; s64 158 fs/jfs/jfs_xtree.c s64 xaddr; s64 253 fs/jfs/jfs_xtree.c s64 bn; s64 259 fs/jfs/jfs_xtree.c s64 size, lstart, lend, xstart, xend, pstart; s64 260 fs/jfs/jfs_xtree.c s64 llen, xlen, plen; s64 261 fs/jfs/jfs_xtree.c s64 xaddr, paddr; s64 505 fs/jfs/jfs_xtree.c s64 bn; /* block number */ s64 512 fs/jfs/jfs_xtree.c s64 t64; s64 513 fs/jfs/jfs_xtree.c s64 next = 0; s64 802 fs/jfs/jfs_xtree.c s64 xaddr, hint; s64 805 fs/jfs/jfs_xtree.c s64 bn; s64 811 fs/jfs/jfs_xtree.c s64 next; s64 851 fs/jfs/jfs_xtree.c if ((rc = dbAlloc(ip, hint, (s64) xlen, &xaddr))) { s64 880 fs/jfs/jfs_xtree.c dbFree(ip, xaddr, (s64) xlen); s64 955 fs/jfs/jfs_xtree.c s64 rbn; /* new right page block number */ s64 958 fs/jfs/jfs_xtree.c s64 rcbn; /* right child page block number */ s64 963 fs/jfs/jfs_xtree.c s64 xaddr; s64 1029 fs/jfs/jfs_xtree.c if ((rc = dbAlloc(ip, (s64) 0, (s64) xlen, &xaddr)) s64 1226 fs/jfs/jfs_xtree.c s64 rbn; /* new right page block number */ s64 1229 fs/jfs/jfs_xtree.c s64 nextbn; s64 1492 fs/jfs/jfs_xtree.c s64 rbn; s64 1628 fs/jfs/jfs_xtree.c s64 bn; s64 1633 fs/jfs/jfs_xtree.c s64 xaddr; s64 1791 fs/jfs/jfs_xtree.c s64 bn; s64 1929 fs/jfs/jfs_xtree.c dbFree(ip, addressXAD(xad) + llen, (s64) rlen); s64 1971 fs/jfs/jfs_xtree.c s64 bn; s64 1977 fs/jfs/jfs_xtree.c s64 nxoff, xoff; s64 1979 fs/jfs/jfs_xtree.c s64 nxaddr, xaddr; s64 2439 fs/jfs/jfs_xtree.c s64 bn, xaddr; s64 2450 fs/jfs/jfs_xtree.c s64 next; s64 2503 fs/jfs/jfs_xtree.c if ((rc = dbAllocBottomUp(ip, xaddr, (s64) nblocks)) == 0) { s64 2522 fs/jfs/jfs_xtree.c if ((rc = dbAllocBottomUp(ip, xaddr, (s64) xlen))) s64 2533 fs/jfs/jfs_xtree.c dbFree(ip, *xaddrp, (s64) * xlenp); s64 2549 fs/jfs/jfs_xtree.c if ((rc = dbAllocBottomUp(ip, xaddr, (s64) xlen))) s64 2605 fs/jfs/jfs_xtree.c s64 bn; s64 2678 fs/jfs/jfs_xtree.c s64 xaddr; s64 2709 fs/jfs/jfs_xtree.c dbFree(ip, xaddr, (s64) xlen); s64 2757 fs/jfs/jfs_xtree.c (s64) JFS_SBI(ip->i_sb)->nbperpage); s64 2832 fs/jfs/jfs_xtree.c s64 xoff, xsize; s64 2834 fs/jfs/jfs_xtree.c s64 oxaddr, sxaddr, dxaddr, nextbn, prevbn; s64 2836 fs/jfs/jfs_xtree.c s64 offset, nbytes, nbrd, pno; s64 2838 fs/jfs/jfs_xtree.c s64 bn; s64 3156 fs/jfs/jfs_xtree.c s64 xoff, xaddr; s64 3159 fs/jfs/jfs_xtree.c s64 bn; /* block number */ s64 3164 fs/jfs/jfs_xtree.c s64 t64; s64 3276 fs/jfs/jfs_xtree.c s64 nextbn, prevbn; s64 3423 fs/jfs/jfs_xtree.c s64 xtTruncate(tid_t tid, struct inode *ip, s64 newsize, int flag) s64 3426 fs/jfs/jfs_xtree.c s64 teof; s64 3429 fs/jfs/jfs_xtree.c s64 bn; s64 3432 fs/jfs/jfs_xtree.c s64 xoff, xaddr; s64 3441 fs/jfs/jfs_xtree.c s64 nfreed; s64 3979 fs/jfs/jfs_xtree.c s64 xtTruncate_pmap(tid_t tid, struct inode *ip, s64 committed_size) s64 3981 fs/jfs/jfs_xtree.c s64 bn; s64 3994 fs/jfs/jfs_xtree.c s64 xoff; s64 61 fs/jfs/jfs_xtree.h ( ((s64)((xad)->off1)) << 32 | __le32_to_cpu((xad)->off2)) s64 63 fs/jfs/jfs_xtree.h ( ((s64)((xad)->addr1)) << 32 | __le32_to_cpu((xad)->addr2)) s64 127 fs/jfs/jfs_xtree.h extern s64 xtTruncate(tid_t tid, struct inode *ip, s64 newsize, int type); s64 128 fs/jfs/jfs_xtree.h extern s64 xtTruncate_pmap(tid_t tid, struct inode *ip, s64 committed_size); s64 40 fs/jfs/namei.c static s64 commitZeroLink(tid_t, struct inode *); s64 476 fs/jfs/namei.c s64 new_size = 0; s64 621 fs/jfs/namei.c static s64 commitZeroLink(tid_t tid, struct inode *ip) s64 714 fs/jfs/namei.c s64 xaddr = addressDXD(&JFS_IP(ip)->ea); s64 735 fs/jfs/namei.c s64 xaddr = addressDXD(&JFS_IP(ip)->acl); s64 884 fs/jfs/namei.c s64 xlen = 0; s64 886 fs/jfs/namei.c s64 extent = 0, xaddr; s64 1077 fs/jfs/namei.c s64 new_size = 0; s64 71 fs/jfs/resize.c s64 newLogAddress, newFSCKAddress; s64 73 fs/jfs/resize.c s64 newMapSize = 0, mapSize; s64 74 fs/jfs/resize.c s64 XAddress, XSize, nblocks, xoff, xaddr, t64; s64 75 fs/jfs/resize.c s64 oldLVSize; s64 76 fs/jfs/resize.c s64 newFSSize; s64 77 fs/jfs/resize.c s64 VolumeSize; s64 147 fs/jfs/super.c s64 maxinodes; s64 165 fs/jfs/super.c maxinodes = min((s64) atomic_read(&imap->im_numinos) + s64 167 fs/jfs/super.c << L2INOSPEREXT), (s64) 0xffffffffLL); s64 366 fs/jfs/super.c s64 newLVSize = 0; s64 417 fs/jfs/super.c s64 newLVSize = 0; s64 243 fs/jfs/xattr.c s64 blkno; s64 392 fs/jfs/xattr.c s64 blkno; s64 468 fs/jfs/xattr.c s64 blkno; s64 546 fs/jfs/xattr.c rc = dbAlloc(inode, INOHINT(inode), (s64) blocks_needed, s64 562 fs/jfs/xattr.c dbFree(inode, blkno, (s64) blocks_needed); s64 573 fs/jfs/xattr.c dbFree(inode, blkno, (s64) blocks_needed); s64 31 fs/lockd/xdr4.c static inline s64 s64 34 fs/lockd/xdr4.c s64 res; s64 192 fs/lockd/xdr4.c s64 start, len; s64 73 fs/ntfs/aops.c s64 file_ofs, init_size; s64 77 fs/ntfs/aops.c file_ofs = ((s64)page->index << PAGE_CACHE_SHIFT) + s64 189 fs/ntfs/aops.c s64 init_size; s64 232 fs/ntfs/aops.c iblock = (s64)page->index << (PAGE_CACHE_SHIFT - blocksize_bits); s64 554 fs/ntfs/aops.c s64 initialized_size; s64 602 fs/ntfs/aops.c block = (s64)page->index << (PAGE_CACHE_SHIFT - blocksize_bits); s64 1136 fs/ntfs/aops.c mft_no = (((s64)page->index << PAGE_CACHE_SHIFT) + ofs) s64 1252 fs/ntfs/aops.c mft_no = (((s64)page->index << PAGE_CACHE_SHIFT) + ofs) s64 1536 fs/ntfs/attrib.c s64 new_size; s64 1904 fs/ntfs/attrib.c s64 ntfs_attr_extend_allocation(ntfs_inode *ni, s64 new_alloc_size, s64 1908 fs/ntfs/attrib.c s64 ll, allocated_size, start = data_start; s64 1942 fs/ntfs/attrib.c start &= ~(s64)vol->cluster_size_mask; s64 1944 fs/ntfs/attrib.c ~(s64)vol->cluster_size_mask; s64 81 fs/ntfs/attrib.h static inline s64 ntfs_attr_size(const ATTR_RECORD *a) s64 84 fs/ntfs/attrib.h return (s64)le32_to_cpu(a->data.resident.value_length); s64 108 fs/ntfs/attrib.h extern s64 ntfs_attr_extend_allocation(ntfs_inode *ni, s64 new_alloc_size, s64 50 fs/ntfs/bitmap.c s64 cnt = count; s64 111 fs/ntfs/bitmap.c len = min_t(s64, cnt >> 3, PAGE_CACHE_SIZE - pos); s64 135 fs/ntfs/bitmap.c len = min_t(s64, cnt >> 3, PAGE_CACHE_SIZE); s64 106 fs/ntfs/compress.c if (((s64)page->index << PAGE_CACHE_SHIFT) >= initialized_size) { s64 484 fs/ntfs/compress.c s64 initialized_size; s64 500 fs/ntfs/compress.c VCN start_vcn = (((s64)index << PAGE_CACHE_SHIFT) & ~cb_size_mask) >> s64 506 fs/ntfs/compress.c VCN end_vcn = ((((s64)(index + 1UL) << PAGE_CACHE_SHIFT) + cb_size - 1) s64 1101 fs/ntfs/dir.c s64 ia_pos, ia_start, prev_ia_pos, bmp_pos; s64 1172 fs/ntfs/dir.c ir_pos = (s64)fpos; s64 1249 fs/ntfs/dir.c ia_pos = (s64)fpos - vol->mft_record_size; s64 1306 fs/ntfs/dir.c if ((prev_ia_pos & (s64)PAGE_CACHE_MASK) != s64 1307 fs/ntfs/dir.c (ia_pos & (s64)PAGE_CACHE_MASK)) { s64 1329 fs/ntfs/dir.c ~(s64)(ndir->itype.index.block_size - 1))); s64 1345 fs/ntfs/dir.c ~(s64)(ndir->itype.index.block_size - 1)) >> s64 1377 fs/ntfs/dir.c ia_start = ia_pos & ~(s64)(ndir->itype.index.block_size - 1); s64 43 fs/ntfs/endian.h static inline s64 sle64_to_cpu(sle64 x) s64 58 fs/ntfs/endian.h static inline s64 sle64_to_cpup(sle64 *x) s64 115 fs/ntfs/file.c s64 old_init_size; s64 249 fs/ntfs/file.c ni->initialized_size = (s64)(index + 1) << PAGE_CACHE_SHIFT; s64 496 fs/ntfs/file.c s64 bh_pos, vcn_len, end, initialized_size; s64 563 fs/ntfs/file.c bh_pos = (s64)page->index << PAGE_CACHE_SHIFT; s64 567 fs/ntfs/file.c s64 bh_end; s64 1118 fs/ntfs/file.c bh_pos = ((s64)page->index << PAGE_CACHE_SHIFT) + s64 1260 fs/ntfs/file.c ((s64)page->index << PAGE_CACHE_SHIFT) + s64 1485 fs/ntfs/file.c s64 end, initialized_size; s64 1502 fs/ntfs/file.c s64 bh_pos; s64 1507 fs/ntfs/file.c bh_pos = (s64)page->index << PAGE_CACHE_SHIFT; s64 1511 fs/ntfs/file.c s64 bh_end; s64 1639 fs/ntfs/file.c s64 end, initialized_size; s64 1794 fs/ntfs/file.c s64 end, ll; s64 1989 fs/ntfs/file.c start_idx = (pos & ~(s64) s64 1773 fs/ntfs/inode.c s64 block; s64 2362 fs/ntfs/inode.c s64 new_size, old_size, nr_freed, new_alloc_size, old_alloc_size; s64 2431 fs/ntfs/inode.c ~(s64)vol->cluster_size_mask; s64 49 fs/ntfs/inode.h s64 initialized_size; /* Copy from the attribute record. */ s64 50 fs/ntfs/inode.h s64 allocated_size; /* Copy from the attribute record. */ s64 113 fs/ntfs/inode.h s64 size; /* Copy of compressed_size from s64 327 fs/ntfs/layout.h #define ERR_MREF(x) ((u64)((s64)(x))) s64 328 fs/ntfs/layout.h #define MREF_ERR(x) ((int)((s64)(x))) s64 153 fs/ntfs/lcnalloc.c s64 clusters; s64 849 fs/ntfs/lcnalloc.c s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, s64 count, s64 852 fs/ntfs/lcnalloc.c s64 delta, to_free, total_freed, real_freed; s64 48 fs/ntfs/lcnalloc.h extern s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, s64 107 fs/ntfs/lcnalloc.h static inline s64 ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, s64 485 fs/ntfs/logfile.c s64 size, pos; s64 522 fs/ntfs/logfile.c size &= ~(s64)(log_page_size - 1); s64 776 fs/ntfs/logfile.c s64 len; s64 1134 fs/ntfs/mft.c s64 pass_end, ll, data_pos, pass_start, ofs, bit; s64 1281 fs/ntfs/mft.c s64 ll; s64 1579 fs/ntfs/mft.c s64 old_data_size, old_initialized_size; s64 1718 fs/ntfs/mft.c s64 min_nr, nr, ll; s64 2246 fs/ntfs/mft.c s64 ll, bit, old_data_initialized, old_data_size; s64 2438 fs/ntfs/mft.c s64 new_initialized_size, mft_no; s64 302 fs/ntfs/runlist.c s64 merged_length; s64 709 fs/ntfs/runlist.c drl[ds].length = (s64)0; s64 754 fs/ntfs/runlist.c s64 deltaxcn; /* Change in [vl]cn. */ s64 835 fs/ntfs/runlist.c deltaxcn = (s64)-1; s64 959 fs/ntfs/runlist.c rl[rlpos].length = (s64)0; s64 1086 fs/ntfs/runlist.c s64 l = n; s64 1158 fs/ntfs/runlist.c s64 delta, length = rl->length; s64 1168 fs/ntfs/runlist.c s64 s1 = last_vcn + 1; s64 1195 fs/ntfs/runlist.c s64 length = rl->length; s64 1204 fs/ntfs/runlist.c s64 s1 = last_vcn + 1; s64 1255 fs/ntfs/runlist.c s64 l = n; s64 1360 fs/ntfs/runlist.c s64 delta, length = rl->length; s64 1370 fs/ntfs/runlist.c s64 s1 = last_vcn + 1; s64 1413 fs/ntfs/runlist.c s64 length = rl->length; s64 1422 fs/ntfs/runlist.c s64 s1 = last_vcn + 1; s64 1648 fs/ntfs/runlist.c s64 delta; s64 45 fs/ntfs/runlist.h s64 length; /* Run length in clusters. */ s64 741 fs/ntfs/super.c s64 ll; s64 2469 fs/ntfs/super.c static s64 get_nr_free_clusters(ntfs_volume *vol) s64 2471 fs/ntfs/super.c s64 nr_free = vol->nr_clusters; s64 2513 fs/ntfs/super.c nr_free -= (s64)hweight32(kaddr[i]); s64 2584 fs/ntfs/super.c nr_free -= (s64)hweight32(kaddr[i]); s64 2618 fs/ntfs/super.c s64 size; s64 30 fs/ntfs/time.h #define NTFS_TIME_OFFSET ((s64)(369 * 365 + 89) * 24 * 3600 * 10000000) s64 54 fs/ntfs/time.h return cpu_to_sle64((s64)ts.tv_sec * 10000000 + ts.tv_nsec / 100 + s64 43 fs/ntfs/types.h typedef s64 VCN; s64 45 fs/ntfs/types.h typedef s64 LCN; s64 53 fs/ntfs/types.h typedef s64 LSN; s64 61 fs/ntfs/types.h typedef s64 USN; s64 84 fs/ntfs/volume.h s64 mft_data_pos; /* Mft record number at which to s64 1493 fs/ocfs2/file.c s64 llen; s64 1661 fs/ocfs2/file.c sr.l_start = (s64)offset; s64 1662 fs/ocfs2/file.c sr.l_len = (s64)len; s64 272 fs/select.c } else if (unlikely(*timeout >= (s64)MAX_SCHEDULE_TIMEOUT - 1)) { s64 380 fs/select.c s64 timeout = -1; s64 434 fs/select.c s64 timeout = MAX_SCHEDULE_TIMEOUT; s64 626 fs/select.c } else if (unlikely(*timeout >= (s64)MAX_SCHEDULE_TIMEOUT-1)) { s64 717 fs/select.c s64 timeout = ((s64)restart_block->arg3<<32) | (s64)restart_block->arg2; s64 733 fs/select.c s64 timeout_jiffies; s64 739 fs/select.c if (timeout_msecs / 1000 > (s64)0x7fffffffffffffffULL / (s64)HZ) s64 770 fs/select.c s64 timeout = -1; s64 193 fs/sync.c if ((s64)offset < 0) s64 195 fs/sync.c if ((s64)endbyte < 0) s64 86 fs/xfs/linux-2.6/xfs_ioctl32.c copy_in_user(&p->l_start, &p32->l_start, sizeof(s64)) || s64 87 fs/xfs/linux-2.6/xfs_ioctl32.c copy_in_user(&p->l_len, &p32->l_len, sizeof(s64)) || s64 150 include/acpi/actypes.h typedef s64 acpi_native_int; s64 263 include/asm-parisc/atomic.h typedef struct { volatile s64 counter; } atomic64_t; s64 291 include/asm-parisc/atomic.h static __inline__ s64 s64 297 include/asm-parisc/atomic.h #define atomic64_add(i,v) ((void)(__atomic64_add_return( ((s64)i),(v)))) s64 298 include/asm-parisc/atomic.h #define atomic64_sub(i,v) ((void)(__atomic64_add_return(-((s64)i),(v)))) s64 302 include/asm-parisc/atomic.h #define atomic64_add_return(i,v) (__atomic64_add_return( ((s64)i),(v))) s64 303 include/asm-parisc/atomic.h #define atomic64_sub_return(i,v) (__atomic64_add_return(-((s64)i),(v))) s64 25 include/asm-parisc/compat.h typedef s64 compat_loff_t; s64 34 include/asm-parisc/compat.h typedef s64 compat_s64; s64 26 include/asm-x86/compat.h typedef s64 compat_loff_t; s64 37 include/asm-x86/compat.h typedef s64 __attribute__((aligned(4))) compat_s64; s64 57 include/asm-x86/uv/bios.h s64 status; s64 113 include/linux/backing-dev.h static inline s64 bdi_stat(struct backing_dev_info *bdi, s64 119 include/linux/backing-dev.h static inline s64 __bdi_stat_sum(struct backing_dev_info *bdi, s64 125 include/linux/backing-dev.h static inline s64 bdi_stat_sum(struct backing_dev_info *bdi, s64 128 include/linux/backing-dev.h s64 sum; s64 229 include/linux/cgroup.h s64 (*read_s64)(struct cgroup *cgrp, struct cftype *cft); s64 258 include/linux/cgroup.h int (*write_s64)(struct cgroup *cgrp, struct cftype *cft, s64 val); s64 87 include/linux/clocksource.h s64 error; s64 181 include/linux/clocksource.h static inline s64 cyc2ns(struct clocksource *cs, cycle_t cycles) s64 6 include/linux/dirent.h s64 d_off; s64 47 include/linux/ktime.h s64 tv64; s64 61 include/linux/ktime.h #define KTIME_MAX ((s64)~((u64)1 << 63)) s64 87 include/linux/ktime.h return (ktime_t) { .tv64 = (s64)secs * NSEC_PER_SEC + (s64)nsecs }; s64 273 include/linux/ktime.h static inline s64 ktime_to_ns(const ktime_t kt) s64 275 include/linux/ktime.h return (s64) kt.tv.sec * NSEC_PER_SEC + kt.tv.nsec; s64 292 include/linux/ktime.h static inline s64 ktime_to_us(const ktime_t kt) s64 295 include/linux/ktime.h return (s64) tv.tv_sec * USEC_PER_SEC + tv.tv_usec; s64 298 include/linux/ktime.h static inline s64 ktime_us_delta(const ktime_t later, const ktime_t earlier) s64 15 include/linux/lockd/nlm.h # define NLM4_OFFSET_MAX ((s64) ((~(u64)0) >> 1)) s64 128 include/linux/lockdep.h s64 min; s64 129 include/linux/lockdep.h s64 max; s64 130 include/linux/lockdep.h s64 total; s64 24 include/linux/math64.h static inline s64 div_s64_rem(s64 dividend, s32 divisor, s32 *remainder) s64 49 include/linux/math64.h extern s64 div_s64_rem(s64 dividend, s32 divisor, s32 *remainder); s64 77 include/linux/math64.h static inline s64 div_s64(s64 dividend, s32 divisor) s64 20 include/linux/percpu_counter.h s64 count; s64 38 include/linux/percpu_counter.h s64 __percpu_counter_sum(struct percpu_counter *fbc); s64 45 include/linux/percpu_counter.h static inline s64 percpu_counter_sum_positive(struct percpu_counter *fbc) s64 47 include/linux/percpu_counter.h s64 ret = __percpu_counter_sum(fbc); s64 51 include/linux/percpu_counter.h static inline s64 percpu_counter_sum(struct percpu_counter *fbc) s64 56 include/linux/percpu_counter.h static inline s64 percpu_counter_read(struct percpu_counter *fbc) s64 66 include/linux/percpu_counter.h static inline s64 percpu_counter_read_positive(struct percpu_counter *fbc) s64 68 include/linux/percpu_counter.h s64 ret = fbc->count; s64 79 include/linux/percpu_counter.h s64 count; s64 110 include/linux/percpu_counter.h static inline s64 percpu_counter_read(struct percpu_counter *fbc) s64 115 include/linux/percpu_counter.h static inline s64 percpu_counter_read_positive(struct percpu_counter *fbc) s64 120 include/linux/percpu_counter.h static inline s64 percpu_counter_sum_positive(struct percpu_counter *fbc) s64 125 include/linux/percpu_counter.h static inline s64 percpu_counter_sum(struct percpu_counter *fbc) s64 115 include/linux/poll.h #define MAX_INT64_SECONDS (((s64)(~((u64)0)>>1)/HZ)-1) s64 979 include/linux/sched.h s64 sum_sleep_runtime; s64 135 include/linux/time.h static inline s64 timespec_to_ns(const struct timespec *ts) s64 137 include/linux/time.h return ((s64) ts->tv_sec * NSEC_PER_SEC) + ts->tv_nsec; s64 147 include/linux/time.h static inline s64 timeval_to_ns(const struct timeval *tv) s64 149 include/linux/time.h return ((s64) tv->tv_sec * NSEC_PER_SEC) + s64 91 include/linux/timex.h #define MAXFREQ_SCALED ((s64)MAXFREQ << NTP_SCALE_SHIFT) s64 45 include/net/pkt_sched.h #define PSCHED_US2NS(x) ((s64)(x) << 10) s64 197 ipc/compat.c return __get_compat_ipc64_perm(&s64->sem_perm, &up64->sem_perm); s64 215 ipc/compat.c err = __put_compat_ipc64_perm(&s64->sem_perm, &up64->sem_perm); s64 216 ipc/compat.c err |= __put_user(s64->sem_otime, &up64->sem_otime); s64 217 ipc/compat.c err |= __put_user(s64->sem_ctime, &up64->sem_ctime); s64 218 ipc/compat.c err |= __put_user(s64->sem_nsems, &up64->sem_nsems); s64 241 ipc/compat.c struct semid64_ds s64; s64 269 ipc/compat.c up64 = compat_alloc_user_space(sizeof(s64)); s64 274 ipc/compat.c if (copy_from_user(&s64, up64, sizeof(s64))) s64 277 ipc/compat.c err2 = put_compat_semid64_ds(&s64, compat_ptr(pad)); s64 279 ipc/compat.c err2 = put_compat_semid_ds(&s64, compat_ptr(pad)); s64 286 ipc/compat.c err = get_compat_semid64_ds(&s64, compat_ptr(pad)); s64 288 ipc/compat.c err = get_compat_semid_ds(&s64, compat_ptr(pad)); s64 290 ipc/compat.c up64 = compat_alloc_user_space(sizeof(s64)); s64 291 ipc/compat.c if (copy_to_user(up64, &s64, sizeof(s64))) s64 491 ipc/compat.c return __get_compat_ipc64_perm(&s64->shm_perm, &up64->shm_perm); s64 509 ipc/compat.c err = __put_compat_ipc64_perm(&s64->shm_perm, &up64->shm_perm); s64 510 ipc/compat.c err |= __put_user(s64->shm_atime, &up64->shm_atime); s64 511 ipc/compat.c err |= __put_user(s64->shm_dtime, &up64->shm_dtime); s64 512 ipc/compat.c err |= __put_user(s64->shm_ctime, &up64->shm_ctime); s64 513 ipc/compat.c err |= __put_user(s64->shm_segsz, &up64->shm_segsz); s64 514 ipc/compat.c err |= __put_user(s64->shm_nattch, &up64->shm_nattch); s64 515 ipc/compat.c err |= __put_user(s64->shm_cpid, &up64->shm_cpid); s64 516 ipc/compat.c err |= __put_user(s64->shm_lpid, &up64->shm_lpid); s64 593 ipc/compat.c struct shmid64_ds s64; s64 623 ipc/compat.c err = get_compat_shmid64_ds(&s64, uptr); s64 625 ipc/compat.c err = get_compat_shmid_ds(&s64, uptr); s64 629 ipc/compat.c p = compat_alloc_user_space(sizeof(s64)); s64 630 ipc/compat.c if (copy_to_user(p, &s64, sizeof(s64))) s64 638 ipc/compat.c p = compat_alloc_user_space(sizeof(s64)); s64 642 ipc/compat.c if (copy_from_user(&s64, p, sizeof(s64))) s64 645 ipc/compat.c err2 = put_compat_shmid64_ds(&s64, uptr); s64 647 ipc/compat.c err2 = put_compat_shmid_ds(&s64, uptr); s64 460 kernel/acct.c while ((s64)value > 0){ s64 1391 kernel/cgroup.c s64 val = simple_strtoll(buffer, &end, 0); s64 1476 kernel/cgroup.c s64 val = cft->read_s64(cgrp, cft); s64 1562 kernel/cpuset.c static s64 cpuset_read_s64(struct cgroup *cont, struct cftype *cft) s64 64 kernel/delayacct.c s64 ns; s64 101 kernel/delayacct.c s64 tmp; s64 113 kernel/delayacct.c tmp = (s64)d->cpu_run_real_total; s64 116 kernel/delayacct.c d->cpu_run_real_total = (tmp < (s64)d->cpu_run_real_total) ? 0 : tmp; s64 118 kernel/delayacct.c tmp = (s64)d->cpu_scaled_run_real_total; s64 122 kernel/delayacct.c (tmp < (s64)d->cpu_scaled_run_real_total) ? 0 : tmp; s64 134 kernel/delayacct.c tmp = (s64)d->cpu_delay_total + t2; s64 135 kernel/delayacct.c d->cpu_delay_total = (tmp < (s64)d->cpu_delay_total) ? 0 : tmp; s64 137 kernel/delayacct.c tmp = (s64)d->cpu_run_virtual_total + t3; s64 139 kernel/delayacct.c (tmp < (s64)d->cpu_run_virtual_total) ? 0 : tmp; s64 807 kernel/hrtimer.c s64 incr = ktime_to_ns(interval); s64 227 kernel/lockdep.c s64 holdtime; s64 3046 kernel/lockdep.c s64 waittime = 0; s64 179 kernel/power/swsusp.c s64 elapsed_centisecs64; s64 1567 kernel/sched.c s64 elapsed = now - sd->last_update; s64 1569 kernel/sched.c if (elapsed >= (s64)(u64)sysctl_sched_shares_ratelimit) { s64 1655 kernel/sched.c s64 diff = sample - *avg; s64 1798 kernel/sched.c s64 delta; s64 1816 kernel/sched.c return delta < (s64)sysctl_sched_migration_cost; s64 4069 kernel/sched.c if ((s64)delta_exec > 0) s64 9166 kernel/sched.c static s64 cpu_rt_runtime_read(struct cgroup *cgrp, struct cftype *cft) s64 95 kernel/sched_clock.c return (s64)(x - y) < 0 ? x : y; s64 100 kernel/sched_clock.c return (s64)(x - y) > 0 ? x : y; s64 111 kernel/sched_clock.c s64 delta = now - scd->tick_raw; s64 174 kernel/sched_clock.c if (likely((s64)(remote_clock - this_clock) < 0)) { s64 116 kernel/sched_debug.c s64 MIN_vruntime = -1, min_vruntime, max_vruntime = -1, s64 203 kernel/sched_fair.c s64 delta = (s64)(vruntime - min_vruntime); s64 212 kernel/sched_fair.c s64 delta = (s64)(vruntime - min_vruntime); s64 219 kernel/sched_fair.c static inline s64 entity_key(struct cfs_rq *cfs_rq, struct sched_entity *se) s64 232 kernel/sched_fair.c s64 key = entity_key(cfs_rq, se); s64 560 kernel/sched_fair.c if ((s64)delta < 0) s64 575 kernel/sched_fair.c if ((s64)delta < 0) s64 602 kernel/sched_fair.c s64 d = se->vruntime - cfs_rq->min_vruntime; s64 833 kernel/sched_fair.c s64 delta = slice - ran; s64 846 kernel/sched_fair.c delta = max_t(s64, 10000LL, delta); s64 1242 kernel/sched_fair.c s64 delta_exec; s64 250 kernel/sched_rt.c s64 diff; s64 302 kernel/sched_rt.c s64 want; s64 329 kernel/sched_rt.c s64 diff; s64 339 kernel/sched_rt.c diff = min_t(s64, iter->rt_runtime, want); s64 523 kernel/sched_rt.c if (unlikely((s64)delta_exec < 0)) s64 43 kernel/time/ntp.c static s64 time_offset; /* time adjustment (ns) */ s64 47 kernel/time/ntp.c static s64 time_freq; /* frequency offset (scaled ns/s)*/ s64 56 kernel/time/ntp.c second_length += (s64)ntp_tick_adj << NTP_SCALE_SHIFT; s64 68 kernel/time/ntp.c s64 freq_adj; s64 92 kernel/time/ntp.c freq_adj = (s64)offset * mtemp; s64 96 kernel/time/ntp.c freq_adj += div_s64((s64)offset << (NTP_SCALE_SHIFT - SHIFT_FLL), s64 104 kernel/time/ntp.c time_offset = div_s64((s64)offset << NTP_SCALE_SHIFT, NTP_INTERVAL_FREQ); s64 183 kernel/time/ntp.c s64 time_adj; s64 209 kernel/time/ntp.c tick_length += (s64)(time_adjust * NSEC_PER_USEC / s64 357 kernel/time/ntp.c time_freq = (s64)txc->freq * PPM_SCALE; s64 406 kernel/time/ntp.c (s64)PPM_SCALE_INV, s64 71 kernel/time/tick-sched.c s64 incr = ktime_to_ns(tick_period); s64 67 kernel/time/timekeeping.c static inline s64 __get_nsec_offset(void) s64 70 kernel/time/timekeeping.c s64 ns_offset; s64 93 kernel/time/timekeeping.c s64 nsecs; s64 204 kernel/time/timekeeping.c static inline s64 __get_nsec_offset(void) { return 0; } s64 269 kernel/time/timekeeping.c static s64 timekeeping_suspend_nsecs; s64 361 kernel/time/timekeeping.c s64 tick_error, i; s64 411 kernel/time/timekeeping.c s64 error, interval = clock->cycle_interval; s64 457 kernel/time/timekeeping.c clock->xtime_nsec += (s64)xtime.tv_nsec << clock->shift; s64 483 kernel/time/timekeeping.c xtime.tv_nsec = (s64)clock->xtime_nsec >> clock->shift; s64 484 kernel/time/timekeeping.c clock->xtime_nsec -= (s64)xtime.tv_nsec << clock->shift; s64 61 lib/div64.c s64 div_s64_rem(s64 dividend, s32 divisor, s32 *remainder) s64 33 lib/percpu_counter.c s64 count; s64 55 lib/percpu_counter.c s64 __percpu_counter_sum(struct percpu_counter *fbc) s64 57 lib/percpu_counter.c s64 ret; s64 237 lib/proportions.c s64 val = percpu_counter_read(&pl->events); s64 58 mm/memcontrol.c s64 count[MEM_CGROUP_STAT_NSTATS]; s64 75 mm/memcontrol.c static s64 mem_cgroup_read_stat(struct mem_cgroup_stat *stat, s64 79 mm/memcontrol.c s64 ret = 0; s64 1007 mm/memcontrol.c s64 val; s64 160 net/dccp/ccids/ccid3.c - (s64)hctx->ccid3hctx_rtt >= 0) { s64 298 net/dccp/ccids/ccid3.c s64 delay; s64 358 net/dccp/ccids/ccid3.c if (delay - (s64)hctx->ccid3hctx_delta >= 1000) s64 679 net/dccp/ccids/ccid3.c s64 delta = 0; s64 95 net/dccp/ccids/lib/loss_interval.c s64 len; s64 102 net/dccp/ccids/lib/loss_interval.c if (len - (s64)cur->li_length <= 0) /* duplicate or reordered */ s64 129 net/dccp/dccp.h static inline s64 dccp_delta_seqno(const u64 seqno1, const u64 seqno2) s64 139 net/dccp/dccp.h return (s64)((seq2 << 16) - (seq1 << 16)) > 0; s64 165 net/dccp/dccp.h s64 delta = dccp_delta_seqno(s1, s2); s64 291 net/dccp/timer.c s64 delta = ktime_us_delta(ktime_get_real(), dccp_timestamp_seed); s64 137 net/ipv4/tcp_lp.c s64 rhz = lp->remote_hz << 6; /* remote HZ << 6 */ s64 138 net/ipv4/tcp_lp.c s64 m = 0; s64 190 net/ipv4/tcp_lp.c s64 owd = 0; s64 223 net/ipv4/tcp_lp.c s64 mowd = tcp_lp_owd_calculator(sk); s64 161 net/netfilter/xt_time.c s64 stamp; s64 59 samples/kprobes/kretprobe_example.c s64 delta; s64 164 sound/core/control_compat.c s64 integer64[64]; s64 205 sound/core/control_compat.c return sizeof(s64) * count; s64 417 sound/mips/sgio2audio.c s64 l, r;