u64 111 arch/x86/boot/pm.c static const u64 boot_gdt[] __attribute__((aligned(16))) = { u64 806 arch/x86/ia32/sys_ia32.c (((u64)offset_high)<<32) | offset_low, u64 807 arch/x86/ia32/sys_ia32.c (((u64)len_high)<<32) | len_low, u64 828 arch/x86/ia32/sys_ia32.c return sys_lookup_dcookie(((u64)addr_high << 32) | addr_low, buf, len); u64 834 arch/x86/ia32/sys_ia32.c return sys_readahead(fd, ((u64)off_hi << 32) | off_lo, count); u64 841 arch/x86/ia32/sys_ia32.c ((u64)off_hi << 32) | off_low, u64 842 arch/x86/ia32/sys_ia32.c ((u64)n_hi << 32) | n_low, flags); u64 848 arch/x86/ia32/sys_ia32.c return sys_fadvise64_64(fd, ((u64)offset_hi << 32) | offset_lo, u64 856 arch/x86/ia32/sys_ia32.c return sys_fallocate(fd, mode, ((u64)offset_hi << 32) | offset_lo, u64 857 arch/x86/ia32/sys_ia32.c ((u64)len_hi << 32) | len_lo); u64 92 arch/x86/kernel/acpi/boot.c static u64 acpi_lapic_addr __initdata = APIC_DEFAULT_PHYS_BASE; u64 235 arch/x86/kernel/acpi/boot.c acpi_lapic_addr = (u64) madt->address; u64 23 arch/x86/kernel/acpi/realmode/wakeup.h u64 pmode_gdt; u64 31 arch/x86/kernel/acpi/realmode/wakeup.h u64 wakeup_gdt[3]; u64 66 arch/x86/kernel/acpi/sleep.c (u64)(sizeof(header->wakeup_gdt) - 1) + u64 67 arch/x86/kernel/acpi/sleep.c ((u64)(acpi_wakeup_address + u64 69 arch/x86/kernel/amd_iommu.c u64 address = (u64)(((u64)event[3]) << 32) | event[2]; u64 319 arch/x86/kernel/amd_iommu.c u64 address = CMD_INV_IOMMU_ALL_PAGES_ADDRESS; u64 343 arch/x86/kernel/amd_iommu.c u64 __pte, *pte, *page; u64 355 arch/x86/kernel/amd_iommu.c page = (u64 *)get_zeroed_page(GFP_KERNEL); u64 365 arch/x86/kernel/amd_iommu.c page = (u64 *)get_zeroed_page(GFP_KERNEL); u64 435 arch/x86/kernel/amd_iommu.c u64 addr; u64 587 arch/x86/kernel/amd_iommu.c u64 *p1, *p2, *p3; u64 640 arch/x86/kernel/amd_iommu.c u64 *l2_pde; u64 641 arch/x86/kernel/amd_iommu.c u64 address; u64 698 arch/x86/kernel/amd_iommu.c l2_pde = (u64 *)get_zeroed_page(GFP_KERNEL); u64 705 arch/x86/kernel/amd_iommu.c dma_dom->pte_pages[i] = (u64 *)get_zeroed_page(GFP_KERNEL); u64 746 arch/x86/kernel/amd_iommu.c u64 pte_root = virt_to_phys(domain->pt_root); u64 870 arch/x86/kernel/amd_iommu.c u64 *pte, __pte; u64 902 arch/x86/kernel/amd_iommu.c u64 *pte; u64 1009 arch/x86/kernel/amd_iommu.c u64 dma_mask; u64 1096 arch/x86/kernel/amd_iommu.c u64 dma_mask; u64 1188 arch/x86/kernel/amd_iommu.c u64 dma_mask = dev->coherent_dma_mask; u64 85 arch/x86/kernel/amd_iommu_init.c u64 mmio_phys; u64 112 arch/x86/kernel/amd_iommu_init.c u64 resv; u64 113 arch/x86/kernel/amd_iommu_init.c u64 range_start; u64 114 arch/x86/kernel/amd_iommu_init.c u64 range_length; u64 196 arch/x86/kernel/amd_iommu_init.c u64 start = iommu->exclusion_start & PAGE_MASK; u64 197 arch/x86/kernel/amd_iommu_init.c u64 limit = (start + iommu->exclusion_length) & PAGE_MASK; u64 198 arch/x86/kernel/amd_iommu_init.c u64 entry; u64 418 arch/x86/kernel/amd_iommu_init.c u64 entry; u64 425 arch/x86/kernel/amd_iommu_init.c entry = (u64)virt_to_phys(cmd_buf); u64 444 arch/x86/kernel/amd_iommu_init.c u64 entry; u64 451 arch/x86/kernel/amd_iommu_init.c entry = (u64)virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK; u64 147 arch/x86/kernel/aperture_64.c u64 aper; u64 171 arch/x86/kernel/aperture_64.c aper = (aper_low & ~((1<<22)-1)) | ((u64)aper_hi << 32); u64 278 arch/x86/kernel/aperture_64.c u64 aper_base = 0, last_aper_base = 0; u64 367 arch/x86/kernel/aperture_64.c u64 aper_base, last_aper_base = 0; u64 183 arch/x86/kernel/apic_32.c u64 xapic_icr_read(void) u64 190 arch/x86/kernel/apic_32.c return icr1 | ((u64)icr2 << 32); u64 499 arch/x86/kernel/apic_32.c u64 res; u64 507 arch/x86/kernel/apic_32.c res = (((u64) deltapm) * mult) >> 22; u64 513 arch/x86/kernel/apic_32.c res = (((u64) delta) * pm_100ms); u64 169 arch/x86/kernel/apic_64.c u64 xapic_icr_read(void) u64 176 arch/x86/kernel/apic_64.c return icr1 | ((u64)icr2 << 32); u64 208 arch/x86/kernel/apic_64.c u64 x2apic_icr_read(void) u64 59 arch/x86/kernel/cpu/cmpxchg.c u64 prev; u64 64 arch/x86/kernel/cpu/cmpxchg.c prev = *(u64 *)ptr; u64 66 arch/x86/kernel/cpu/cmpxchg.c *(u64 *)ptr = new; u64 775 arch/x86/kernel/cpu/common.c u64 val; u64 914 arch/x86/kernel/cpu/common.c wrmsrl(MSR_STAR, ((u64)__USER32_CS)<<48 | ((u64)__KERNEL_CS)<<32); u64 267 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c u64 whole; u64 155 arch/x86/kernel/cpu/cpufreq/e_powersaver.c u64 val; u64 50 arch/x86/kernel/cpu/cpufreq/powernow-k6.c u64 invalue = 0; u64 180 arch/x86/kernel/cpu/mcheck/mce_64.c u64 mcestart = 0; u64 443 arch/x86/kernel/cpu/mcheck/mce_64.c u64 cap; u64 40 arch/x86/kernel/cpu/mtrr/generic.c u64 mtrr_tom2; u64 57 arch/x86/kernel/cpu/mtrr/generic.c u64 base, mask; u64 104 arch/x86/kernel/cpu/mtrr/generic.c base = (((u64)mtrr_state.var_ranges[i].base_hi) << 32) + u64 106 arch/x86/kernel/cpu/mtrr/generic.c mask = (((u64)mtrr_state.var_ranges[i].mask_hi) << 32) + u64 55 arch/x86/kernel/cpu/mtrr/main.c u64 size_or_mask, size_and_mask; u64 860 arch/x86/kernel/cpu/mtrr/main.c u64 base, mask; u64 868 arch/x86/kernel/cpu/mtrr/main.c mask &= ~((((u64)sizek) << 10) - 1); u64 870 arch/x86/kernel/cpu/mtrr/main.c base = ((u64)basek) << 10; u64 1122 arch/x86/kernel/cpu/mtrr/main.c static u64 mtrr_chunk_size __initdata = (256ULL<<20); u64 1134 arch/x86/kernel/cpu/mtrr/main.c static u64 mtrr_gran_size __initdata; u64 1218 arch/x86/kernel/cpu/mtrr/main.c u64 chunk_size, gran_size; u64 1537 arch/x86/kernel/cpu/mtrr/main.c static u64 __init real_trim_memory(unsigned long start_pfn, u64 1540 arch/x86/kernel/cpu/mtrr/main.c u64 trim_start, trim_size; u64 1566 arch/x86/kernel/cpu/mtrr/main.c u64 total_trim_size; u64 90 arch/x86/kernel/cpu/mtrr/mtrr.h extern u64 size_or_mask, size_and_mask; u64 97 arch/x86/kernel/cpu/mtrr/mtrr.h extern u64 mtrr_tom2; u64 38 arch/x86/kernel/cpu/perfctr-watchdog.c u64 checkbit; u64 226 arch/x86/kernel/cpu/perfctr-watchdog.c u64 counter_val; u64 236 arch/x86/kernel/cpu/perfctr-watchdog.c counter_val = (u64)cpu_khz * 1000; u64 239 arch/x86/kernel/cpu/perfctr-watchdog.c u64 count = (u64)cpu_khz * 1000; u64 249 arch/x86/kernel/cpu/perfctr-watchdog.c u64 count = (u64)cpu_khz * 1000; u64 260 arch/x86/kernel/cpu/perfctr-watchdog.c u64 count = (u64)cpu_khz * 1000; u64 790 arch/x86/kernel/cpu/perfctr-watchdog.c u64 ctr; u64 90 arch/x86/kernel/cpuid.c u64 pos = *ppos; u64 774 arch/x86/kernel/ds.c *value = *(u64 *)(context->ds + (ds_cfg.sizeof_field * 8)); u64 792 arch/x86/kernel/ds.c *(u64 *)(context->ds + (ds_cfg.sizeof_field * 8)) = value; u64 385 arch/x86/kernel/e820.c u64 start = biosmap->addr; u64 386 arch/x86/kernel/e820.c u64 size = biosmap->size; u64 387 arch/x86/kernel/e820.c u64 end = start + size; u64 420 arch/x86/kernel/e820.c static u64 __init e820_update_range_map(struct e820map *e820x, u64 start, u64 425 arch/x86/kernel/e820.c u64 real_updated_size = 0; u64 434 arch/x86/kernel/e820.c u64 final_start, final_end; u64 461 arch/x86/kernel/e820.c u64 __init e820_update_range(u64 start, u64 size, unsigned old_type, u64 467 arch/x86/kernel/e820.c static u64 __init e820_update_range_saved(u64 start, u64 size, u64 475 arch/x86/kernel/e820.c u64 __init e820_remove_range(u64 start, u64 size, unsigned old_type, u64 479 arch/x86/kernel/e820.c u64 real_removed_size = 0; u64 486 arch/x86/kernel/e820.c u64 final_start, final_end; u64 674 arch/x86/kernel/e820.c u64 start, end; u64 745 arch/x86/kernel/e820.c u64 lower_start, lower_end; u64 746 arch/x86/kernel/e820.c u64 upper_start, upper_end; u64 877 arch/x86/kernel/e820.c u64 final_start, final_end; u64 906 arch/x86/kernel/e820.c u64 addr = *addrp; u64 924 arch/x86/kernel/e820.c u64 addr = *addrp, last; u64 925 arch/x86/kernel/e820.c u64 size = *sizep; u64 957 arch/x86/kernel/e820.c u64 __init find_e820_area(u64 start, u64 end, u64 size, u64 align) u64 963 arch/x86/kernel/e820.c u64 addr, last; u64 964 arch/x86/kernel/e820.c u64 ei_last; u64 989 arch/x86/kernel/e820.c u64 __init find_e820_area_size(u64 start, u64 *sizep, u64 align) u64 995 arch/x86/kernel/e820.c u64 addr, last; u64 996 arch/x86/kernel/e820.c u64 ei_last; u64 1022 arch/x86/kernel/e820.c u64 __init early_reserve_e820(u64 startt, u64 sizet, u64 align) u64 1024 arch/x86/kernel/e820.c u64 size = 0; u64 1025 arch/x86/kernel/e820.c u64 addr; u64 1026 arch/x86/kernel/e820.c u64 start; u64 1111 arch/x86/kernel/e820.c u64 align = PAGE_SIZE; u64 1154 arch/x86/kernel/e820.c u64 __init e820_hole_size(u64 start, u64 end) u64 1167 arch/x86/kernel/e820.c return end - start - ((u64)ram << PAGE_SHIFT); u64 1181 arch/x86/kernel/e820.c u64 mem_size; u64 1204 arch/x86/kernel/e820.c u64 start_at, mem_size; u64 1279 arch/x86/kernel/e820.c u64 end; u64 1345 arch/x86/kernel/e820.c u64 mem_size; u64 437 arch/x86/kernel/efi.c u64 addr, npages; u64 466 arch/x86/kernel/efi.c u64 end, systab, addr, npages; u64 483 arch/x86/kernel/efi.c md->virt_addr = (u64) (unsigned long) va; u64 498 arch/x86/kernel/efi.c systab = (u64) (unsigned long) efi_phys.systab; u64 559 arch/x86/kernel/efi.c u64 efi_mem_attributes(unsigned long phys_addr) u64 23 arch/x86/kernel/head32.c u64 ramdisk_image = boot_params.hdr.ramdisk_image; u64 24 arch/x86/kernel/head32.c u64 ramdisk_size = boot_params.hdr.ramdisk_size; u64 25 arch/x86/kernel/head32.c u64 ramdisk_end = ramdisk_image + ramdisk_size; u64 325 arch/x86/kernel/hpet.c u64 start, now; u64 569 arch/x86/kernel/i387.c u64 mask; u64 1487 arch/x86/kernel/io_apic_32.c u64 icr; u64 360 arch/x86/kernel/irq_32.c u64 arch_irq_stat_cpu(unsigned int cpu) u64 362 arch/x86/kernel/irq_32.c u64 sum = nmi_count(cpu); u64 381 arch/x86/kernel/irq_32.c u64 arch_irq_stat(void) u64 383 arch/x86/kernel/irq_32.c u64 sum = atomic_read(&irq_err_count); u64 53 arch/x86/kernel/irq_64.c u64 curbase = (u64)task_stack_page(current); u64 160 arch/x86/kernel/irq_64.c u64 arch_irq_stat_cpu(unsigned int cpu) u64 162 arch/x86/kernel/irq_64.c u64 sum = cpu_pda(cpu)->__nmi_count; u64 178 arch/x86/kernel/irq_64.c u64 arch_irq_stat(void) u64 24 arch/x86/kernel/kdebugfs.c u64 paddr; u64 38 arch/x86/kernel/kdebugfs.c u64 pa; u64 122 arch/x86/kernel/kdebugfs.c u64 pa_data; u64 185 arch/x86/kernel/kvm.c .pt_phys = (u64)pfn << PAGE_SHIFT, u64 55 arch/x86/kernel/kvmclock.c high = ((u64)__pa(&wall_clock) >> 32); u64 96 arch/x86/kernel/kvmclock.c high = ((u64)__pa(&per_cpu(hv_clock, cpu)) >> 32); u64 22 arch/x86/kernel/machine_kexec_64.c static u64 kexec_pgd[512] PAGE_ALIGNED; u64 23 arch/x86/kernel/machine_kexec_64.c static u64 kexec_pud0[512] PAGE_ALIGNED; u64 24 arch/x86/kernel/machine_kexec_64.c static u64 kexec_pmd0[512] PAGE_ALIGNED; u64 25 arch/x86/kernel/machine_kexec_64.c static u64 kexec_pte0[512] PAGE_ALIGNED; u64 26 arch/x86/kernel/machine_kexec_64.c static u64 kexec_pud1[512] PAGE_ALIGNED; u64 27 arch/x86/kernel/machine_kexec_64.c static u64 kexec_pmd1[512] PAGE_ALIGNED; u64 28 arch/x86/kernel/machine_kexec_64.c static u64 kexec_pte1[512] PAGE_ALIGNED; u64 26 arch/x86/kernel/mmconf-fam10h_64.c static u64 __cpuinitdata fam10h_pci_mmconf_base; u64 35 arch/x86/kernel/mmconf-fam10h_64.c u64 start; u64 36 arch/x86/kernel/mmconf-fam10h_64.c u64 end; u64 62 arch/x86/kernel/mmconf-fam10h_64.c u64 val; u64 64 arch/x86/kernel/mmconf-fam10h_64.c u64 tom2; u64 65 arch/x86/kernel/mmconf-fam10h_64.c u64 base = FAM10H_PCI_MMCONF_BASE; u64 124 arch/x86/kernel/mmconf-fam10h_64.c u64 start; u64 125 arch/x86/kernel/mmconf-fam10h_64.c u64 end; u64 130 arch/x86/kernel/mmconf-fam10h_64.c start = (((u64)reg) << 8) & (0xffULL << 32); /* 39:16 on 31:8*/ u64 132 arch/x86/kernel/mmconf-fam10h_64.c end = (((u64)reg) << 8) & (0xffULL << 32); /* 39:16 on 31:8*/ u64 180 arch/x86/kernel/mmconf-fam10h_64.c u64 val; u64 197 arch/x86/kernel/mmconf-fam10h_64.c u64 base; u64 80 arch/x86/kernel/module_64.c u64 val; u64 96 arch/x86/kernel/module_64.c sym->st_value, rel[i].r_addend, (u64)loc); u64 104 arch/x86/kernel/module_64.c *(u64 *)loc = val; u64 117 arch/x86/kernel/module_64.c val -= (u64)loc; u64 1004 arch/x86/kernel/mpparse.c u64 startt = 0; u64 591 arch/x86/kernel/pci-calgary_64.c u64 val; u64 630 arch/x86/kernel/pci-calgary_64.c u64 val64; u64 720 arch/x86/kernel/pci-calgary_64.c u64 low, high, sizelow; u64 721 arch/x86/kernel/pci-calgary_64.c u64 start, limit; u64 744 arch/x86/kernel/pci-calgary_64.c u64 low, high, sizelow, sizehigh; u64 745 arch/x86/kernel/pci-calgary_64.c u64 start, limit; u64 781 arch/x86/kernel/pci-calgary_64.c u64 start; u64 805 arch/x86/kernel/pci-calgary_64.c u64 val64; u64 806 arch/x86/kernel/pci-calgary_64.c u64 table_phys; u64 839 arch/x86/kernel/pci-calgary_64.c table_phys = (u64)__pa(tbl->it_base); u64 844 arch/x86/kernel/pci-calgary_64.c val64 |= (u64) specified_table_size; u64 856 arch/x86/kernel/pci-calgary_64.c u64 val64; u64 977 arch/x86/kernel/pci-calgary_64.c u64 val64; u64 980 arch/x86/kernel/pci-calgary_64.c u64 mask; u64 1354 arch/x86/kernel/pci-calgary_64.c u64 *tp; u64 1356 arch/x86/kernel/pci-calgary_64.c tp = ((u64 *)tbl->it_base); u64 1396 arch/x86/kernel/pci-calgary_64.c info->tce_space = (u64 *)__va(tce_space); u64 68 arch/x86/kernel/pci-gart_64.c #define GPTE_DECODE(x) (((x) & 0xfffff000) | (((u64)(x) & 0xff0) << 28)) u64 554 arch/x86/kernel/pci-gart_64.c u64 aper_base; u64 736 arch/x86/kernel/process_64.c u64 fp, ip; u64 744 arch/x86/kernel/process_64.c fp = *(u64 *)(p->thread.sp); u64 749 arch/x86/kernel/process_64.c ip = *(u64 *)(fp+8); u64 752 arch/x86/kernel/process_64.c fp = *(u64 *)fp; u64 29 arch/x86/kernel/pvclock.c u64 tsc_timestamp; /* TSC at last update of time vals. */ u64 30 arch/x86/kernel/pvclock.c u64 system_timestamp; /* Time, in nanosecs, since boot. */ u64 40 arch/x86/kernel/pvclock.c static inline u64 scale_delta(u64 delta, u32 mul_frac, int shift) u64 42 arch/x86/kernel/pvclock.c u64 product; u64 66 arch/x86/kernel/pvclock.c : "=a" (product) : "0" (delta), "d" ((u64)mul_frac) ); u64 74 arch/x86/kernel/pvclock.c static u64 pvclock_get_nsec_offset(struct pvclock_shadow_time *shadow) u64 76 arch/x86/kernel/pvclock.c u64 delta = native_read_tsc() - shadow->tsc_timestamp; u64 122 arch/x86/kernel/pvclock.c u64 delta; u64 135 arch/x86/kernel/pvclock.c delta += now.tv_sec * (u64)NSEC_PER_SEC + now.tv_nsec; u64 262 arch/x86/kernel/setup.c u64 ramdisk_image = boot_params.hdr.ramdisk_image; u64 263 arch/x86/kernel/setup.c u64 ramdisk_size = boot_params.hdr.ramdisk_size; u64 264 arch/x86/kernel/setup.c u64 end_of_lowmem = max_low_pfn << PAGE_SHIFT; u64 265 arch/x86/kernel/setup.c u64 ramdisk_here; u64 324 arch/x86/kernel/setup.c u64 ramdisk_image = boot_params.hdr.ramdisk_image; u64 325 arch/x86/kernel/setup.c u64 ramdisk_size = boot_params.hdr.ramdisk_size; u64 326 arch/x86/kernel/setup.c u64 ramdisk_end = ramdisk_image + ramdisk_size; u64 327 arch/x86/kernel/setup.c u64 end_of_lowmem = max_low_pfn << PAGE_SHIFT; u64 376 arch/x86/kernel/setup.c u64 pa_data; u64 398 arch/x86/kernel/setup.c u64 pa_data; u64 424 arch/x86/kernel/setup.c u64 pa_data; u64 639 arch/x86/kernel/setup.c u64 addr = PAGE_SIZE; /* assume first page is reserved anyway */ u64 660 arch/x86/kernel/setup.c u64 size; u64 51 arch/x86/kernel/tce_64.c u64* tp; u64 52 arch/x86/kernel/tce_64.c u64 t; u64 53 arch/x86/kernel/tce_64.c u64 rpn; u64 59 arch/x86/kernel/tce_64.c tp = ((u64*)tbl->it_base) + index; u64 76 arch/x86/kernel/tce_64.c u64* tp; u64 78 arch/x86/kernel/tce_64.c tp = ((u64*)tbl->it_base) + index; u64 37 arch/x86/kernel/tsc.c u64 native_sched_clock(void) u64 39 arch/x86/kernel/tsc.c u64 this_offset; u64 107 arch/x86/kernel/tsc.c static u64 tsc_read_refs(u64 *p, int hpet) u64 109 arch/x86/kernel/tsc.c u64 t1, t2; u64 130 arch/x86/kernel/tsc.c u64 tmp; u64 135 arch/x86/kernel/tsc.c tmp = ((u64)hpet2 * hpet_readl(HPET_PERIOD)); u64 147 arch/x86/kernel/tsc.c u64 tmp; u64 153 arch/x86/kernel/tsc.c pm2 += (u64)ACPI_PM_OVRRUN; u64 180 arch/x86/kernel/tsc.c u64 tsc, t1, t2, delta; u64 309 arch/x86/kernel/tsc.c u64 t1, t2, delta; u64 353 arch/x86/kernel/tsc.c u64 tsc1, tsc2, delta, ref1, ref2; u64 429 arch/x86/kernel/tsc.c delta = ((u64) tsc_pit_min) * 100; u64 796 arch/x86/kernel/tsc.c u64 lpj; u64 816 arch/x86/kernel/tsc.c lpj = ((u64)tsc_khz * 1000); u64 44 arch/x86/kernel/vmi_32.c typedef u64 __attribute__((regparm(2))) (VROMLONGFUNC)(int); u64 103 arch/x86/kernel/vmi_32.c u64 reloc; u64 695 arch/x86/kernel/vmi_32.c u64 reloc; u64 748 arch/x86/kernel/vmi_32.c u64 reloc; u64 17 arch/x86/kernel/xsave.c u64 pcntxt_mask; u64 120 arch/x86/kernel/xsave.c u64 xstate_bv; u64 165 arch/x86/kernel/xsave.c u64 mask; u64 318 arch/x86/kernel/xsave.c pcntxt_mask = eax + ((u64)edx << 32); u64 49 arch/x86/kvm/i8254.c static u64 muldiv64(u64 a, u32 b, u32 c) u64 52 arch/x86/kvm/i8254.c u64 ll; u64 57 arch/x86/kvm/i8254.c u64 rl, rh; u64 60 arch/x86/kvm/i8254.c rl = (u64)u.l.low * (u64)b; u64 61 arch/x86/kvm/i8254.c rh = (u64)u.l.high * (u64)b; u64 34 arch/x86/kvm/kvm_svm.h u64 next_rip; u64 36 arch/x86/kvm/kvm_svm.h u64 host_user_msrs[NR_HOST_SAVE_USER_MSRS]; u64 37 arch/x86/kvm/kvm_svm.h u64 host_gs_base; u64 505 arch/x86/kvm/lapic.c u64 counter_passed; u64 833 arch/x86/kvm/lapic.c u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu) u64 836 arch/x86/kvm/lapic.c u64 tpr; u64 840 arch/x86/kvm/lapic.c tpr = (u64) apic_get_reg(apic, APIC_TASKPRI); u64 868 arch/x86/kvm/lapic.c u64 kvm_lapic_get_base(struct kvm_vcpu *vcpu) u64 31 arch/x86/kvm/lapic.h u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu); u64 34 arch/x86/kvm/lapic.h u64 kvm_lapic_get_base(struct kvm_vcpu *vcpu); u64 40 arch/x86/kvm/lapic.h u64 kvm_get_apic_base(struct kvm_vcpu *vcpu); u64 112 arch/x86/kvm/mmu.c #define PT64_BASE_ADDR_MASK (((1ULL << 52) - 1) & ~(u64)(PAGE_SIZE-1)) u64 146 arch/x86/kvm/mmu.c u64 *shadow_ptes[RMAP_EXT]; u64 154 arch/x86/kvm/mmu.c static u64 __read_mostly shadow_trap_nonpresent_pte; u64 155 arch/x86/kvm/mmu.c static u64 __read_mostly shadow_notrap_nonpresent_pte; u64 156 arch/x86/kvm/mmu.c static u64 __read_mostly shadow_base_present_pte; u64 157 arch/x86/kvm/mmu.c static u64 __read_mostly shadow_nx_mask; u64 158 arch/x86/kvm/mmu.c static u64 __read_mostly shadow_x_mask; /* mutual exclusive with nx_mask */ u64 159 arch/x86/kvm/mmu.c static u64 __read_mostly shadow_user_mask; u64 160 arch/x86/kvm/mmu.c static u64 __read_mostly shadow_accessed_mask; u64 161 arch/x86/kvm/mmu.c static u64 __read_mostly shadow_dirty_mask; u64 485 arch/x86/kvm/mmu.c desc->shadow_ptes[0] = (u64 *)*rmapp; u64 551 arch/x86/kvm/mmu.c if ((u64 *)*rmapp != spte) { u64 576 arch/x86/kvm/mmu.c static u64 *rmap_next(struct kvm *kvm, unsigned long *rmapp, u64 *spte) u64 580 arch/x86/kvm/mmu.c u64 *prev_spte; u64 587 arch/x86/kvm/mmu.c return (u64 *)*rmapp; u64 607 arch/x86/kvm/mmu.c u64 *spte; u64 658 arch/x86/kvm/mmu.c u64 *spte; u64 711 arch/x86/kvm/mmu.c u64 *spte; u64 721 arch/x86/kvm/mmu.c u64 _spte = *spte; u64 741 arch/x86/kvm/mmu.c u64 *pos; u64 742 arch/x86/kvm/mmu.c u64 *end; u64 744 arch/x86/kvm/mmu.c for (pos = spt, end = pos + PAGE_SIZE / sizeof(u64); pos != end; pos++) u64 797 arch/x86/kvm/mmu.c u64 *old = sp->parent_pte; u64 947 arch/x86/kvm/mmu.c u64 *pt; u64 948 arch/x86/kvm/mmu.c u64 ent; u64 996 arch/x86/kvm/mmu.c u64 *parent_pte; u64 1126 arch/x86/kvm/mmu.c u64 spte; u64 1142 arch/x86/kvm/mmu.c u64 pte = *shadow_pte; u64 1177 arch/x86/kvm/mmu.c spte |= (u64)pfn << PAGE_SHIFT; u64 1242 arch/x86/kvm/mmu.c u64 *table; u64 1667 arch/x86/kvm/mmu.c u64 pte; u64 1728 arch/x86/kvm/mmu.c u64 *spte = vcpu->arch.last_pte_updated; u64 1738 arch/x86/kvm/mmu.c u64 gpte = 0; u64 1755 arch/x86/kvm/mmu.c r = kvm_read_guest(vcpu->kvm, gpa & ~(u64)7, &gpte, 8); u64 1790 arch/x86/kvm/mmu.c u64 *spte = vcpu->arch.last_pte_updated; u64 1808 arch/x86/kvm/mmu.c u64 entry, gentry; u64 1809 arch/x86/kvm/mmu.c u64 *spte; u64 1886 arch/x86/kvm/mmu.c gpa & ~(u64)(pte_size - 1), u64 2060 arch/x86/kvm/mmu.c u64 *pt; u64 2333 arch/x86/kvm/mmu.c u64 *pt = __va(page_pte & PT64_BASE_ADDR_MASK); u64 2338 arch/x86/kvm/mmu.c u64 ent = pt[i]; u64 2427 arch/x86/kvm/mmu.c u64 *pt = sp->spt; u64 2433 arch/x86/kvm/mmu.c u64 ent = pt[i]; u64 26 arch/x86/kvm/paging_tmpl.h #define pt_element_t u64 u64 225 arch/x86/kvm/paging_tmpl.h __func__, (u64)pte, pt_access, pte_access); u64 246 arch/x86/kvm/paging_tmpl.h u64 *spte, const void *pte) u64 259 arch/x86/kvm/paging_tmpl.h pgprintk("%s: gpte %llx spte %p\n", __func__, (u64)gpte, spte); u64 277 arch/x86/kvm/paging_tmpl.h static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr, u64 284 arch/x86/kvm/paging_tmpl.h u64 *shadow_ent; u64 301 arch/x86/kvm/paging_tmpl.h u64 shadow_pte; u64 305 arch/x86/kvm/paging_tmpl.h shadow_ent = ((u64 *)__va(shadow_addr)) + index; u64 380 arch/x86/kvm/paging_tmpl.h u64 *shadow_pte; u64 86 arch/x86/kvm/svm.c u64 asid_generation; u64 695 arch/x86/kvm/svm.c u64 tsc_this, delta; u64 769 arch/x86/kvm/svm.c static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg) u64 1014 arch/x86/kvm/svm.c u64 fault_address; u64 1209 arch/x86/kvm/svm.c u64 tsc; u64 1268 arch/x86/kvm/svm.c u64 data; u64 1290 arch/x86/kvm/svm.c u64 tsc; u64 1362 arch/x86/kvm/svm.c u64 data = (svm->vmcb->save.rax & -1u) u64 1363 arch/x86/kvm/svm.c | ((u64)(svm->vcpu.arch.regs[VCPU_REGS_RDX] & -1u) << 32); u64 1464 arch/x86/kvm/svm.c (u32)((u64)svm->vmcb->save.rip >> 32), entryexit); u64 1709 arch/x86/kvm/svm.c u64 cr8; u64 1904 arch/x86/kvm/svm.c u64 vm_cr; u64 59 arch/x86/kvm/svm.h u64 intercept; u64 61 arch/x86/kvm/svm.h u64 iopm_base_pa; u64 62 arch/x86/kvm/svm.h u64 msrpm_base_pa; u64 63 arch/x86/kvm/svm.h u64 tsc_offset; u64 73 arch/x86/kvm/svm.h u64 exit_info_1; u64 74 arch/x86/kvm/svm.h u64 exit_info_2; u64 77 arch/x86/kvm/svm.h u64 nested_ctl; u64 81 arch/x86/kvm/svm.h u64 nested_cr3; u64 82 arch/x86/kvm/svm.h u64 lbr_ctl; u64 121 arch/x86/kvm/svm.h u64 base; u64 138 arch/x86/kvm/svm.h u64 efer; u64 140 arch/x86/kvm/svm.h u64 cr4; u64 141 arch/x86/kvm/svm.h u64 cr3; u64 142 arch/x86/kvm/svm.h u64 cr0; u64 143 arch/x86/kvm/svm.h u64 dr7; u64 144 arch/x86/kvm/svm.h u64 dr6; u64 145 arch/x86/kvm/svm.h u64 rflags; u64 146 arch/x86/kvm/svm.h u64 rip; u64 148 arch/x86/kvm/svm.h u64 rsp; u64 150 arch/x86/kvm/svm.h u64 rax; u64 151 arch/x86/kvm/svm.h u64 star; u64 152 arch/x86/kvm/svm.h u64 lstar; u64 153 arch/x86/kvm/svm.h u64 cstar; u64 154 arch/x86/kvm/svm.h u64 sfmask; u64 155 arch/x86/kvm/svm.h u64 kernel_gs_base; u64 156 arch/x86/kvm/svm.h u64 sysenter_cs; u64 157 arch/x86/kvm/svm.h u64 sysenter_esp; u64 158 arch/x86/kvm/svm.h u64 sysenter_eip; u64 159 arch/x86/kvm/svm.h u64 cr2; u64 161 arch/x86/kvm/svm.h u64 g_pat; u64 162 arch/x86/kvm/svm.h u64 dbgctl; u64 163 arch/x86/kvm/svm.h u64 br_from; u64 164 arch/x86/kvm/svm.h u64 br_to; u64 165 arch/x86/kvm/svm.h u64 last_excp_from; u64 166 arch/x86/kvm/svm.h u64 last_excp_to; u64 94 arch/x86/kvm/vmx.c static u64 construct_eptp(unsigned long root_hpa); u64 286 arch/x86/kvm/vmx.c u64 vpid : 16; u64 287 arch/x86/kvm/vmx.c u64 rsvd : 48; u64 288 arch/x86/kvm/vmx.c u64 gva; u64 300 arch/x86/kvm/vmx.c u64 eptp, gpa; u64 321 arch/x86/kvm/vmx.c u64 phys_addr = __pa(vmcs); u64 408 arch/x86/kvm/vmx.c static u64 vmcs_read64(unsigned long field) u64 413 arch/x86/kvm/vmx.c return vmcs_readl(field) | ((u64)vmcs_readl(field+1) << 32); u64 496 arch/x86/kvm/vmx.c u64 host_efer = vmx->host_msrs[efer_offset].data; u64 497 arch/x86/kvm/vmx.c u64 guest_efer = vmx->guest_msrs[efer_offset].data; u64 498 arch/x86/kvm/vmx.c u64 ignore_bits; u64 511 arch/x86/kvm/vmx.c ignore_bits &= ~(u64)EFER_SCE; u64 623 arch/x86/kvm/vmx.c u64 phys_addr = __pa(vmx->vmcs); u64 624 arch/x86/kvm/vmx.c u64 tsc_this, delta, new_offset; u64 817 arch/x86/kvm/vmx.c static u64 guest_read_tsc(void) u64 819 arch/x86/kvm/vmx.c u64 host_tsc, tsc_offset; u64 832 arch/x86/kvm/vmx.c u64 host_tsc; u64 845 arch/x86/kvm/vmx.c u64 data; u64 1031 arch/x86/kvm/vmx.c u64 msr; u64 1043 arch/x86/kvm/vmx.c u64 phys_addr = __pa(per_cpu(vmxarea, cpu)); u64 1044 arch/x86/kvm/vmx.c u64 old; u64 1522 arch/x86/kvm/vmx.c static u64 construct_eptp(unsigned long root_hpa) u64 1524 arch/x86/kvm/vmx.c u64 eptp; u64 1537 arch/x86/kvm/vmx.c u64 eptp; u64 1592 arch/x86/kvm/vmx.c static u64 vmx_get_segment_base(struct kvm_vcpu *vcpu, int seg) u64 1980 arch/x86/kvm/vmx.c u64 data; u64 1987 arch/x86/kvm/vmx.c data = data_low | ((u64)data_high << 32); u64 2019 arch/x86/kvm/vmx.c u64 msr; u64 2300 arch/x86/kvm/vmx.c (u32)((u64)cr2 >> 32), handler); u64 2390 arch/x86/kvm/vmx.c (u32)((u64)vcpu->arch.regs[reg] >> 32), handler); u64 2434 arch/x86/kvm/vmx.c (u32)((u64)vcpu->arch.regs[reg] >> 32), u64 2507 arch/x86/kvm/vmx.c u64 data; u64 2527 arch/x86/kvm/vmx.c u64 data = (vcpu->arch.regs[VCPU_REGS_RAX] & -1u) u64 2528 arch/x86/kvm/vmx.c | ((u64)(vcpu->arch.regs[VCPU_REGS_RDX] & -1u) << 32); u64 2595 arch/x86/kvm/vmx.c u64 exit_qualification; u64 2629 arch/x86/kvm/vmx.c u64 exit_qualification; u64 2739 arch/x86/kvm/vmx.c (u32)((u64)vmcs_readl(GUEST_RIP) >> 32), entryexit); u64 52 arch/x86/kvm/x86.c static u64 __read_mostly efer_reserved_bits = 0xfffffffffffffafeULL; u64 54 arch/x86/kvm/x86.c static u64 __read_mostly efer_reserved_bits = 0xfffffffffffffffeULL; u64 129 arch/x86/kvm/x86.c u64 kvm_get_apic_base(struct kvm_vcpu *vcpu) u64 209 arch/x86/kvm/x86.c u64 pdpte[ARRAY_SIZE(vcpu->arch.pdptrs)]; u64 212 arch/x86/kvm/x86.c offset * sizeof(u64), sizeof(pdpte)); u64 234 arch/x86/kvm/x86.c u64 pdpte[ARRAY_SIZE(vcpu->arch.pdptrs)]; u64 596 arch/x86/kvm/x86.c (NSEC_PER_SEC * (u64)ts.tv_sec); u64 738 arch/x86/kvm/x86.c u64 data; u64 985 arch/x86/kvm/x86.c u64 efer; u64 2008 arch/x86/kvm/x86.c u64 val; u64 2019 arch/x86/kvm/x86.c val = *(u64 *)new; u64 2026 arch/x86/kvm/x86.c set_64bit((u64 *)(kaddr + offset_in_page(gpa)), val); u64 2580 arch/x86/kvm/x86.c static u64 mk_cr_64(u64 curr_cr, u32 new_val) u64 2632 arch/x86/kvm/x86.c (u32)((u64)value >> 32), handler); u64 2641 arch/x86/kvm/x86.c (u32)((u64)val >> 32), handler); u64 3665 arch/x86/kvm/x86.c u64 rip; u64 3666 arch/x86/kvm/x86.c u64 rdp; u64 688 arch/x86/kvm/x86_emulate.c op->val = *(u64 *) op->ptr; u64 824 arch/x86/kvm/x86_emulate.c c->modrm_ea = insn_fetch(u64, 8, c->eip); u64 1169 arch/x86/kvm/x86_emulate.c u64 old, new; u64 1184 arch/x86/kvm/x86_emulate.c new = ((u64)c->regs[VCPU_REGS_RCX] << 32) | u64 1251 arch/x86/kvm/x86_emulate.c u64 msr_data; u64 1888 arch/x86/kvm/x86_emulate.c | ((u64)c->regs[VCPU_REGS_RDX] << 32); u64 2006 arch/x86/kvm/x86_emulate.c (u64) c->src.val; u64 796 arch/x86/lguest/boot.c static u64 lguest_apic_icr_read(void) u64 228 arch/x86/mm/fault.c printk("*pte = %0*Lx ", sizeof(page)*2, (u64)page); u64 332 arch/x86/mm/fault.c if ((address >= (u64)_stext && address <= (u64)_etext) || u64 243 arch/x86/mm/ioremap.c retval = reserve_memtype(phys_addr, (u64)phys_addr + size, u64 94 arch/x86/mm/memtest.c u64 t_start, t_size; u64 231 arch/x86/mm/numa_32.c u64 node_kva_target; u64 232 arch/x86/mm/numa_32.c u64 node_kva_final; u64 260 arch/x86/mm/numa_32.c ((u64)node_end_pfn[nid])<<PAGE_SHIFT, u64 261 arch/x86/mm/numa_32.c ((u64)size)<<PAGE_SHIFT, u64 289 arch/x86/mm/numa_32.c node_kva_final+(((u64)size)<<PAGE_SHIFT), u64 331 arch/x86/mm/numa_64.c u64 size; u64 356 arch/x86/mm/numa_64.c u64 end = *addr + size; u64 404 arch/x86/mm/numa_64.c u64 size, addr = start_pfn << PAGE_SHIFT; u64 405 arch/x86/mm/numa_64.c u64 max_addr = last_pfn << PAGE_SHIFT; u64 440 arch/x86/mm/numa_64.c size = ((u64)num << 20) & FAKE_NODE_MIN_HASH_MASK; u64 74 arch/x86/mm/pageattr-test.c addr, level, (u64)pte_val(*pte)); u64 186 arch/x86/mm/pageattr-test.c pte ? (u64)pte_val(*pte) : 0ULL); u64 218 arch/x86/mm/pageattr-test.c addr[i], pte ? (u64)pte_val(*pte) : 0ULL); u64 61 arch/x86/mm/pat.c static u64 __read_mostly boot_pat_state; u64 72 arch/x86/mm/pat.c #define PAT(x, y) ((u64)PAT_ ## y << ((x)*8)) u64 76 arch/x86/mm/pat.c u64 pat; u64 149 arch/x86/mm/pat.c u64 start; u64 150 arch/x86/mm/pat.c u64 end; u64 212 arch/x86/mm/pat.c static u64 cached_start; u64 230 arch/x86/mm/pat.c u64 pfn, end_pfn; u64 254 arch/x86/mm/pat.c u64 pfn, end_pfn; u64 486 arch/x86/mm/pat.c u64 from = ((u64)pfn) << PAGE_SHIFT; u64 487 arch/x86/mm/pat.c u64 to = from + size; u64 488 arch/x86/mm/pat.c u64 cursor = from; u64 507 arch/x86/mm/pat.c u64 offset = ((u64) pfn) << PAGE_SHIFT; u64 574 arch/x86/mm/pat.c u64 addr = (u64)pfn << PAGE_SHIFT; u64 590 arch/x86/mm/pat.c u64 addr = (u64)pfn << PAGE_SHIFT; u64 480 arch/x86/mm/srat_64.c u64 total_mb; u64 242 arch/x86/pci/amd_bus.c static u64 __initdata fam10h_mmconf_start; u64 243 arch/x86/pci/amd_bus.c static u64 __initdata fam10h_mmconf_end; u64 247 arch/x86/pci/amd_bus.c u64 base, msr; u64 293 arch/x86/pci/amd_bus.c u64 val; u64 385 arch/x86/pci/amd_bus.c node, link, (u64)start, (u64)end); u64 458 arch/x86/pci/amd_bus.c node, link, (u64)start, (u64)end); u64 465 arch/x86/pci/amd_bus.c u64 endx = 0; u64 484 arch/x86/pci/amd_bus.c printk(KERN_CONT " ==> [%llx, %llx]", (u64)start, endx); u64 490 arch/x86/pci/amd_bus.c printk(KERN_CONT " %s [%llx, %llx]", endx?"and":"==>", (u64)start, (u64)end); u64 571 arch/x86/pci/amd_bus.c u64 reg; u64 263 arch/x86/pci/i386.c u64 addr = (u64)vma->vm_pgoff << PAGE_SHIFT; u64 269 arch/x86/pci/i386.c u64 addr = (u64)vma->vm_pgoff << PAGE_SHIFT; u64 286 arch/x86/pci/i386.c u64 addr = vma->vm_pgoff << PAGE_SHIFT; u64 106 arch/x86/pci/mmconfig-shared.c u64 base, msr; u64 319 arch/x86/pci/mmconfig-shared.c u64 old_size = size; u64 362 arch/x86/pci/mmconfig-shared.c u64 addr, size; u64 208 arch/x86/vdso/vdso32-setup.c checking_wrmsrl(MSR_IA32_SYSENTER_CS, (u64)__KERNEL_CS); u64 210 arch/x86/vdso/vdso32-setup.c checking_wrmsrl(MSR_IA32_SYSENTER_EIP, (u64)ia32_sysenter_target); u64 404 arch/x86/xen/enlighten.c u64 entry = *(u64 *)ptr; u64 526 arch/x86/xen/enlighten.c if (HYPERVISOR_update_descriptor(maddr.maddr, *(u64 *)desc)) u64 568 arch/x86/xen/enlighten.c static u64 xen_apic_icr_read(void) u64 799 arch/x86/xen/enlighten.c u64 base; u64 806 arch/x86/xen/enlighten.c base = ((u64)high << 32) | low; u64 527 arch/x86/xen/mmu.c set_64bit((u64 *)ptep, native_pte_val(pte)); u64 187 arch/x86/xen/multicalls.c unsigned argidx = roundup(b->argidx, sizeof(u64)); u64 196 arch/x86/xen/multicalls.c argidx = roundup(b->argidx, sizeof(u64)); u64 45 arch/x86/xen/setup.c e820_add_region(0, PFN_PHYS((u64)max_pfn), E820_RAM); u64 21 arch/x86/xen/spinlock.c u64 taken; u64 28 arch/x86/xen/spinlock.c u64 released; u64 37 arch/x86/xen/spinlock.c u64 time_total; u64 38 arch/x86/xen/spinlock.c u64 time_spinning; u64 39 arch/x86/xen/spinlock.c u64 time_blocked; u64 58 arch/x86/xen/spinlock.c static inline u64 spin_time_start(void) u64 102 arch/x86/xen/spinlock.c static inline u64 spin_time_start(void) u64 191 arch/x86/xen/spinlock.c u64 start; u64 259 arch/x86/xen/spinlock.c u64 start_spin; u64 266 arch/x86/xen/spinlock.c u64 start_spin_fast = spin_time_start(); u64 44 arch/x86/xen/time.c static u64 get64(const u64 *p) u64 46 arch/x86/xen/time.c u64 ret; u64 65 arch/x86/xen/time.c ret = (((u64)h) << 32) | l; u64 77 arch/x86/xen/time.c u64 state_time; u64 168 arch/x86/xen/time.c u64 ret; u64 201 arch/x86/xen/time.c u64 xen_khz = 1000000ULL << 32; u64 105 block/as-iosched.c u64 new_seek_total; /* mean seek on new proc */ u64 519 block/as-iosched.c u64 total; u64 522 block/as-iosched.c ad->new_seek_total = (7*ad->new_seek_total + 256*(u64)sdist)/8; u64 536 block/as-iosched.c aic->seek_total = (7*aic->seek_total + (u64)256*sdist) / 8; u64 177 block/blk-settings.c if (b_pfn < (min_t(u64, 0x100000000UL, BLK_BOUNCE_HIGH) >> PAGE_SHIFT)) u64 1703 block/cfq-iosched.c u64 total; u64 1720 block/cfq-iosched.c cic->seek_total = (7*cic->seek_total + (u64)256*sdist) / 8; u64 435 block/compat_ioctl.c err |= __put_user((u64)f->name, (compat_caddr_t __user *)&uf->name); u64 485 block/compat_ioctl.c err |= __put_user((u64)f->dmabuf, &uf->dmabuf); u64 486 block/compat_ioctl.c err |= __put_user((u64)f->bufblocks, &uf->bufblocks); u64 201 block/ioctl.c return put_user(val, (u64 __user *)arg); u64 145 crypto/eseqiv.c if (ivsize > sizeof(u64)) { u64 146 crypto/eseqiv.c memset(req->giv, 0, ivsize - sizeof(u64)); u64 147 crypto/eseqiv.c len = sizeof(u64); u64 311 crypto/fcrypt.c u64 k; /* k holds all 56 non-parity bits */ u64 492 crypto/gcm.c inst->alg.cra_alignmask = ctr->cra_alignmask | (__alignof__(u64) - 1); u64 127 crypto/gf128mul.c u64 a = be64_to_cpu(x->a); u64 128 crypto/gf128mul.c u64 b = be64_to_cpu(x->b); u64 129 crypto/gf128mul.c u64 _tt = gf128mul_table_lle[(b << 7) & 0xff]; u64 137 crypto/gf128mul.c u64 a = be64_to_cpu(x->a); u64 138 crypto/gf128mul.c u64 b = be64_to_cpu(x->b); u64 139 crypto/gf128mul.c u64 _tt = gf128mul_table_bbe[a >> 63]; u64 147 crypto/gf128mul.c u64 a = le64_to_cpu(x->a); u64 148 crypto/gf128mul.c u64 b = le64_to_cpu(x->b); u64 149 crypto/gf128mul.c u64 _tt = gf128mul_table_bbe[b >> 63]; u64 158 crypto/gf128mul.c u64 a = be64_to_cpu(x->a); u64 159 crypto/gf128mul.c u64 b = be64_to_cpu(x->b); u64 160 crypto/gf128mul.c u64 _tt = gf128mul_table_lle[b & 0xff]; u64 168 crypto/gf128mul.c u64 a = be64_to_cpu(x->a); u64 169 crypto/gf128mul.c u64 b = be64_to_cpu(x->b); u64 170 crypto/gf128mul.c u64 _tt = gf128mul_table_bbe[a >> 56]; u64 34 crypto/khazad.c u64 E[KHAZAD_ROUNDS + 1]; u64 35 crypto/khazad.c u64 D[KHAZAD_ROUNDS + 1]; u64 38 crypto/khazad.c static const u64 T0[256] = { u64 127 crypto/khazad.c static const u64 T1[256] = { u64 216 crypto/khazad.c static const u64 T2[256] = { u64 305 crypto/khazad.c static const u64 T3[256] = { u64 394 crypto/khazad.c static const u64 T4[256] = { u64 483 crypto/khazad.c static const u64 T5[256] = { u64 572 crypto/khazad.c static const u64 T6[256] = { u64 661 crypto/khazad.c static const u64 T7[256] = { u64 750 crypto/khazad.c static const u64 c[KHAZAD_ROUNDS + 1] = { u64 762 crypto/khazad.c const u64 *S = T7; u64 763 crypto/khazad.c u64 K2, K1; u64 766 crypto/khazad.c K2 = ((u64)be32_to_cpu(key[0]) << 32) | be32_to_cpu(key[1]); u64 767 crypto/khazad.c K1 = ((u64)be32_to_cpu(key[2]) << 32) | be32_to_cpu(key[3]); u64 808 crypto/khazad.c u64 state; u64 38 crypto/md4.c u64 byte_count; u64 205 crypto/md4.c memset(p, 0x00, padding + sizeof (u64)); u64 215 crypto/md4.c sizeof(u64)) / sizeof(u32)); u64 41 crypto/md5.c u64 byte_count; u64 200 crypto/md5.c memset(p, 0x00, padding + sizeof (u64)); u64 210 crypto/md5.c sizeof(u64)) / sizeof(u32)); u64 27 crypto/rmd128.c u64 byte_count; u64 27 crypto/rmd160.c u64 byte_count; u64 27 crypto/rmd256.c u64 byte_count; u64 27 crypto/rmd320.c u64 byte_count; u64 88 crypto/seqiv.c if (ivsize > sizeof(u64)) { u64 89 crypto/seqiv.c memset(info, 0, ivsize - sizeof(u64)); u64 90 crypto/seqiv.c len = sizeof(u64); u64 29 crypto/sha1_generic.c u64 count; u64 25 crypto/sha512_generic.c u64 state[8]; u64 28 crypto/sha512_generic.c u64 W[80]; u64 31 crypto/sha512_generic.c static inline u64 Ch(u64 x, u64 y, u64 z) u64 36 crypto/sha512_generic.c static inline u64 Maj(u64 x, u64 y, u64 z) u64 41 crypto/sha512_generic.c static inline u64 RORu64(u64 x, u64 y) u64 46 crypto/sha512_generic.c static const u64 sha512_K[80] = { u64 94 crypto/sha512_generic.c u64 a, b, c, d, e, f, g, h, t1, t2; u64 38 crypto/tgr192.c u64 a, b, c; u64 44 crypto/tgr192.c static const u64 sbox1[256] = { u64 133 crypto/tgr192.c static const u64 sbox2[256] = { u64 222 crypto/tgr192.c static const u64 sbox3[256] = { u64 311 crypto/tgr192.c static const u64 sbox4[256] = { u64 403 crypto/tgr192.c u64 a = *ra; u64 404 crypto/tgr192.c u64 b = *rb; u64 405 crypto/tgr192.c u64 c = *rc; u64 422 crypto/tgr192.c u64 a = *ra; u64 423 crypto/tgr192.c u64 b = *rb; u64 424 crypto/tgr192.c u64 c = *rc; u64 468 crypto/tgr192.c u64 a, b, c, aa, bb, cc; u64 469 crypto/tgr192.c u64 x[8]; u64 43 crypto/wp512.c u64 hash[WP512_DIGEST_SIZE/8]; u64 53 crypto/wp512.c static const u64 C0[256] = { u64 142 crypto/wp512.c static const u64 C1[256] = { u64 231 crypto/wp512.c static const u64 C2[256] = { u64 320 crypto/wp512.c static const u64 C3[256] = { u64 409 crypto/wp512.c static const u64 C4[256] = { u64 498 crypto/wp512.c static const u64 C5[256] = { u64 587 crypto/wp512.c static const u64 C6[256] = { u64 676 crypto/wp512.c static const u64 C7[256] = { u64 765 crypto/wp512.c static const u64 rc[WHIRLPOOL_ROUNDS + 1] = { u64 778 crypto/wp512.c u64 K[8]; /* the round key */ u64 779 crypto/wp512.c u64 block[8]; /* mu(buffer) */ u64 780 crypto/wp512.c u64 state[8]; /* the cipher state */ u64 781 crypto/wp512.c u64 L[8]; u64 1010 crypto/wp512.c u64 value = bits_len; u64 882 fs/9p/vfs_inode.c u64 path = qid->path + 2; u64 119 fs/afs/afs.h u64 size; /* file size */ u64 42 fs/afs/fsclient.c u64 data_version, size; u64 69 fs/afs/fsclient.c data_version |= (u64) ntohl(*bp++) << 32; u64 71 fs/afs/fsclient.c size |= (u64) ntohl(*bp++) << 32; u64 52 fs/afs/main.c u64 uuidtime; u64 63 fs/afs/main.c uuidtime = (u64) ts.tv_sec * 1000 * 1000 * 10; u64 968 fs/aio.c event->obj = (u64)(unsigned long)iocb->ki_obj.user; u64 1760 fs/aio.c tmp.obj = (u64)(unsigned long)kiocb->ki_obj.user; u64 83 fs/autofs4/autofs_i.h u64 ino; u64 212 fs/autofs4/autofs_i.h static inline u64 autofs4_get_ino(struct autofs_sb_info *sbi) u64 17 fs/befs/befs.h typedef u64 befs_blocknr_t; u64 82 fs/befs/befs_fs_types.h typedef u64 __bitwise fs64; u64 86 fs/befs/befs_fs_types.h typedef u64 befs_off_t; u64 14 fs/befs/endian.h static inline u64 u64 227 fs/bfs/inode.c u64 id = huge_encode_dev(s->s_bdev->bd_dev); u64 90 fs/cifs/cifsproto.h extern u64 cifs_UnixTimeToNT(struct timespec); u64 2212 fs/cifs/connect.c sb->s_maxbytes = (u64) 1 << 63; u64 2214 fs/cifs/connect.c sb->s_maxbytes = (u64) 1 << 31; /* 2 GB */ u64 852 fs/cifs/netmisc.c #define NTFS_TIME_OFFSET ((u64)(369*365 + 89) * 24 * 3600 * 10000000) u64 865 fs/cifs/netmisc.c u64 t; u64 874 fs/cifs/netmisc.c u64 u64 878 fs/cifs/netmisc.c return (u64) t.tv_sec * 10000000 + t.tv_nsec/100 + NTFS_TIME_OFFSET; u64 952 fs/compat.c int reclen = ALIGN(jj + namlen + 1, sizeof(u64)); u64 953 fs/compat.c u64 off; u64 1605 fs/compat.c if ((u64)tv.tv_sec >= (u64)MAX_INT64_SECONDS) u64 1620 fs/compat.c rtv.tv_usec = jiffies_to_usecs(do_div((*(u64*)&timeout), HZ)); u64 1811 fs/compat.c rts.tv_nsec = jiffies_to_usecs(do_div((*(u64*)&timeout), HZ)) * u64 190 fs/debugfs/file.c *(u64 *)data = val; u64 196 fs/debugfs/file.c *val = *(u64 *)data; u64 951 fs/dquot.c 2 * nla_total_size(sizeof(u64)); u64 203 fs/ecryptfs/ecryptfs_kernel.h #define ECRYPTFS_FILE_SIZE_BYTES (sizeof(u64)) u64 268 fs/ecryptfs/inode.c u64 file_size; u64 408 fs/ecryptfs/inode.c u64 file_size_save; u64 379 fs/ecryptfs/mmap.c file_size_virt = kmalloc(sizeof(u64), GFP_KERNEL); u64 386 fs/ecryptfs/mmap.c sizeof(u64)); u64 1229 fs/ext2/super.c u64 fsid; u64 1282 fs/ext2/super.c le64_to_cpup((void *)es->s_uuid + sizeof(u64)); u64 209 fs/ext3/dir.c u64 version = filp->f_version; u64 2559 fs/ext3/super.c u64 fsid; u64 2614 fs/ext3/super.c le64_to_cpup((void *)es->s_uuid + sizeof(u64)); u64 215 fs/ext4/dir.c u64 version = filp->f_version; u64 4014 fs/ext4/inode.c i_blocks = ((u64)le16_to_cpu(raw_inode->i_blocks_high)) << 32 | u64 4176 fs/ext4/inode.c u64 i_blocks = inode->i_blocks; u64 1787 fs/ext4/super.c if (sizeof(blkcnt_t) < sizeof(u64)) { u64 1830 fs/ext4/super.c if (sizeof(blkcnt_t) < sizeof(u64)) { u64 2089 fs/ext4/super.c if (sizeof(root->i_blocks) < sizeof(u64) && u64 3183 fs/ext4/super.c u64 fsid; u64 3239 fs/ext4/super.c le64_to_cpup((void *)es->s_uuid + sizeof(u64)); u64 688 fs/fat/dir.c loff_t offset, u64 ino, unsigned int d_type) \ u64 223 fs/fuse/dev.c static u64 fuse_get_unique(struct fuse_conn *fc) u64 23 fs/fuse/dir.c static inline u64 fuse_dentry_time(struct dentry *entry) u64 37 fs/fuse/dir.c static u64 fuse_dentry_time(struct dentry *entry) u64 39 fs/fuse/dir.c return (u64) entry->d_time + u64 40 fs/fuse/dir.c ((u64) (unsigned long) entry->d_fsdata << 32); u64 53 fs/fuse/dir.c static u64 time_to_jiffies(unsigned long sec, unsigned long nsec) u64 73 fs/fuse/dir.c static u64 attr_timeout(struct fuse_attr_out *o) u64 78 fs/fuse/dir.c static u64 entry_attr_timeout(struct fuse_entry_out *o) u64 133 fs/fuse/dir.c u64 fuse_get_attr_version(struct fuse_conn *fc) u64 135 fs/fuse/dir.c u64 curr_version; u64 170 fs/fuse/dir.c u64 attr_version; u64 263 fs/fuse/dir.c u64 attr_version; u64 786 fs/fuse/dir.c u64 attr_version; u64 195 fs/fuse/file.c u64 fuse_lock_owner_id(struct fuse_conn *fc, fl_owner_t id) u64 198 fs/fuse/file.c u64 v = (unsigned long) id; u64 210 fs/fuse/file.c return (u64) v0 + ((u64) v1 << 32); u64 423 fs/fuse/file.c u64 attr_ver; u64 60 fs/fuse/fuse_i.h u64 nodeid; u64 63 fs/fuse/fuse_i.h u64 nlookup; u64 69 fs/fuse/fuse_i.h u64 i_time; u64 76 fs/fuse/fuse_i.h u64 attr_version; u64 101 fs/fuse/fuse_i.h u64 fh; u64 191 fs/fuse/fuse_i.h u64 intr_unique; u64 244 fs/fuse/fuse_i.h u64 attr_ver; u64 346 fs/fuse/fuse_i.h u64 reqctr; u64 444 fs/fuse/fuse_i.h u64 attr_version; u64 462 fs/fuse/fuse_i.h static inline u64 get_node_id(struct inode *inode) u64 647 fs/fuse/fuse_i.h u64 fuse_lock_owner_id(struct fuse_conn *fc, fl_owner_t id); u64 657 fs/fuse/fuse_i.h u64 fuse_get_attr_version(struct fuse_conn *fc); u64 211 fs/fuse/inode.c u64 nodeid = *(u64 *) _nodeidp; u64 220 fs/fuse/inode.c u64 nodeid = *(u64 *) _nodeidp; u64 558 fs/fuse/inode.c u64 nodeid; u64 623 fs/fuse/inode.c u64 nodeid; u64 662 fs/fuse/inode.c handle.nodeid = (u64) fid->raw[0] << 32; u64 663 fs/fuse/inode.c handle.nodeid |= (u64) fid->raw[1]; u64 676 fs/fuse/inode.c parent.nodeid = (u64) fid->raw[3] << 32; u64 677 fs/fuse/inode.c parent.nodeid |= (u64) fid->raw[4]; u64 125 fs/gfs2/bmap.c u64 block = 0; u64 296 fs/gfs2/bmap.c u64 dblock; u64 343 fs/gfs2/bmap.c u64 d = be64_to_cpu(*ptr); u64 434 fs/gfs2/bmap.c u64 bn, dblock = 0; u64 572 fs/gfs2/bmap.c const u64 *arr = sdp->sd_heightsize; u64 574 fs/gfs2/bmap.c u64 size; u64 688 fs/gfs2/bmap.c u64 bn; u64 753 fs/gfs2/bmap.c u64 bn, bstart; u64 1044 fs/gfs2/bmap.c if (size & (u64)(sdp->sd_sb.sb_bsize - 1)) u64 1067 fs/gfs2/bmap.c u64 lblock; u64 1278 fs/gfs2/bmap.c u64 lblock, lblock_stop, size; u64 1301 fs/gfs2/bmap.c u64 end_of_file = (ip->i_di.di_size + sdp->sd_sb.sb_bsize - 1) >> shift; u64 80 fs/gfs2/dir.c #define gfs2_disk_hash2offset(h) (((u64)(h)) >> 1) u64 81 fs/gfs2/dir.c #define gfs2_dir_offset2hash(p) ((u32)(((u64)(p)) << 1)) u64 157 fs/gfs2/dir.c u64 lblock, dblock; u64 274 fs/gfs2/dir.c u64 lblock, dblock; u64 727 fs/gfs2/dir.c if (error != sizeof(u64)) u64 738 fs/gfs2/dir.c u64 leaf_no; u64 762 fs/gfs2/dir.c u64 ln; u64 763 fs/gfs2/dir.c if (hsize * sizeof(u64) != ip->i_di.di_size) { u64 807 fs/gfs2/dir.c u64 bn = gfs2_alloc_block(ip, &n); u64 847 fs/gfs2/dir.c u64 bn; u64 938 fs/gfs2/dir.c u64 bn, leaf_no; u64 989 fs/gfs2/dir.c error = gfs2_dir_write_data(dip, (char *)lp, start * sizeof(u64), u64 990 fs/gfs2/dir.c half_len * sizeof(u64)); u64 991 fs/gfs2/dir.c if (error != half_len * sizeof(u64)) { u64 1078 fs/gfs2/dir.c u64 *buf; u64 1079 fs/gfs2/dir.c u64 *from, *to; u64 1080 fs/gfs2/dir.c u64 block; u64 1085 fs/gfs2/dir.c if (hsize * sizeof(u64) != dip->i_di.di_size) { u64 1105 fs/gfs2/dir.c to = (u64 *)((char *)buf + sdp->sd_hash_bsize); u64 1205 fs/gfs2/dir.c u64 off, off_next; u64 1275 fs/gfs2/dir.c u64 lfn = leaf_no; u64 1373 fs/gfs2/dir.c if (hsize * sizeof(u64) != dip->i_di.di_size) { u64 1549 fs/gfs2/dir.c u64 bn; u64 1783 fs/gfs2/dir.c u64 leaf_no; u64 1787 fs/gfs2/dir.c if (hsize * sizeof(u64) != dip->i_di.di_size) { u64 1859 fs/gfs2/dir.c u64 blk, nblk; u64 1862 fs/gfs2/dir.c unsigned int x, size = len * sizeof(u64); u64 1928 fs/gfs2/dir.c error = gfs2_dir_write_data(dip, ht, index * sizeof(u64), size); u64 135 fs/gfs2/eattr.c u64 bn; u64 227 fs/gfs2/eattr.c u64 bn = 0; u64 228 fs/gfs2/eattr.c u64 bstart = 0; u64 585 fs/gfs2/eattr.c u64 block; u64 642 fs/gfs2/eattr.c u64 block; u64 965 fs/gfs2/eattr.c u64 blk; u64 987 fs/gfs2/eattr.c *eablk = cpu_to_be64((u64)newbh->b_blocknr); u64 1304 fs/gfs2/eattr.c u64 bstart = 0; u64 1325 fs/gfs2/eattr.c u64 bn; u64 1370 fs/gfs2/eattr.c u64 bn; u64 135 fs/gfs2/glock.c h = jhash(&name->ln_number, sizeof(u64), 0); u64 33 fs/gfs2/incore.h u64 lh_sequence; /* Sequence number of this transaction */ u64 74 fs/gfs2/incore.h u64 rg_igeneration; u64 81 fs/gfs2/incore.h u64 rd_addr; /* grp block disk address */ u64 82 fs/gfs2/incore.h u64 rd_data0; /* first data location */ u64 116 fs/gfs2/incore.h u64 blkno; u64 237 fs/gfs2/incore.h u64 di_size; /* number of bytes in file */ u64 238 fs/gfs2/incore.h u64 di_generation; /* generation number for NFS */ u64 242 fs/gfs2/incore.h u64 di_eattr; /* extended attribute block number */ u64 247 fs/gfs2/incore.h u64 i_no_addr; u64 248 fs/gfs2/incore.h u64 i_no_formal_ino; u64 257 fs/gfs2/incore.h u64 i_goal; /* goal block for allocations */ u64 284 fs/gfs2/incore.h u64 rr_blkno; u64 314 fs/gfs2/incore.h u64 qd_sync_gen; u64 348 fs/gfs2/incore.h u64 ai_sync_gen; u64 355 fs/gfs2/incore.h u64 dblock; /* First disk block */ u64 356 fs/gfs2/incore.h u64 blocks; u64 441 fs/gfs2/incore.h u64 no_formal_ino; u64 442 fs/gfs2/incore.h u64 no_addr; u64 481 fs/gfs2/incore.h u64 sd_heightsize[GFS2_MAX_META_HEIGHT + 1]; u64 483 fs/gfs2/incore.h u64 sd_jheightsize[GFS2_MAX_META_HEIGHT + 1]; u64 569 fs/gfs2/incore.h u64 sd_quota_sync_gen; u64 595 fs/gfs2/incore.h u64 sd_log_sequence; u64 606 fs/gfs2/incore.h u64 sd_log_flush_wrapped; u64 610 fs/gfs2/incore.h u64 sd_ail_sync_gen; u64 42 fs/gfs2/inode.c u64 ir_start; u64 43 fs/gfs2/inode.c u64 ir_length; u64 49 fs/gfs2/inode.c u64 *no_addr = opaque; u64 60 fs/gfs2/inode.c u64 *no_addr = opaque; u64 81 fs/gfs2/inode.c u64 no_addr; u64 610 fs/gfs2/inode.c u64 x, y; u64 984 fs/gfs2/inode.c u64 generation; u64 48 fs/gfs2/inode.h static inline u64 gfs2_get_inode_blocks(const struct inode *inode) u64 174 fs/gfs2/log.c u64 sync_gen; u64 343 fs/gfs2/log.c static u64 log_bmap(struct gfs2_sbd *sdp, unsigned int lbn) u64 420 fs/gfs2/log.c sizeof(u64)); u64 487 fs/gfs2/log.c u64 blkno = log_bmap(sdp, sdp->sd_log_flush_head); u64 535 fs/gfs2/log.c u64 blkno = log_bmap(sdp, sdp->sd_log_flush_head); u64 577 fs/gfs2/log.c u64 blkno = log_bmap(sdp, sdp->sd_log_flush_head); u64 251 fs/gfs2/lops.c u64 blkno; u64 335 fs/gfs2/lops.c sizeof(u64))); u64 344 fs/gfs2/lops.c if (offset + sizeof(u64) > sdp->sd_sb.sb_bsize) { u64 358 fs/gfs2/lops.c offset += sizeof(u64); u64 386 fs/gfs2/lops.c u64 blkno; u64 403 fs/gfs2/lops.c while (offset + sizeof(u64) <= sdp->sd_sb.sb_bsize) { u64 416 fs/gfs2/lops.c offset += sizeof(u64); u64 647 fs/gfs2/lops.c u64 blkno; u64 648 fs/gfs2/lops.c u64 esc; u64 730 fs/gfs2/ops_address.c u64 fs_total, new_free; u64 766 fs/gfs2/ops_address.c u64 to = pos + copied; u64 99 fs/gfs2/ops_export.c u64 offset = 0; u64 266 fs/gfs2/ops_export.c this.no_formal_ino = ((u64)be32_to_cpu(fh[0])) << 32; u64 268 fs/gfs2/ops_export.c this.no_addr = ((u64)be32_to_cpu(fh[2])) << 32; u64 285 fs/gfs2/ops_export.c parent.no_formal_ino = ((u64)be32_to_cpu(fh[4])) << 32; u64 287 fs/gfs2/ops_export.c parent.no_addr = ((u64)be32_to_cpu(fh[6])) << 32; u64 89 fs/gfs2/ops_file.c u64 offset = file->f_pos; u64 318 fs/gfs2/ops_file.c u64 lblock = page->index << (PAGE_CACHE_SHIFT - inode->i_blkbits); u64 347 fs/gfs2/ops_file.c u64 pos = page->index << (PAGE_CACHE_SIZE - inode->i_blkbits); u64 337 fs/gfs2/ops_fstype.c sizeof(struct gfs2_dinode)) / sizeof(u64); u64 339 fs/gfs2/ops_fstype.c sizeof(struct gfs2_meta_header)) / sizeof(u64); u64 343 fs/gfs2/ops_fstype.c sdp->sd_hash_ptrs = sdp->sd_hash_bsize / sizeof(u64); u64 350 fs/gfs2/ops_fstype.c hash_blocks = DIV_ROUND_UP(sizeof(u64) * (1 << GFS2_DIR_MAX_DEPTH), u64 367 fs/gfs2/ops_fstype.c u64 space, d; u64 386 fs/gfs2/ops_fstype.c u64 space, d; u64 542 fs/gfs2/ops_fstype.c u64 no_addr; u64 615 fs/gfs2/ops_fstype.c u64 db, prev_db; /* logical block, disk block, prev disk block */ u64 1057 fs/gfs2/ops_inode.c u64 blocks = gfs2_get_inode_blocks(&ip->i_inode); u64 69 fs/gfs2/quota.c u64 qu_limit; u64 70 fs/gfs2/quota.c u64 qu_warn; u64 76 fs/gfs2/quota.c u64 qc_change; u64 81 fs/gfs2/quota.c static u64 qd2offset(struct gfs2_quota_data *qd) u64 83 fs/gfs2/quota.c u64 offset; u64 85 fs/gfs2/quota.c offset = 2 * (u64)qd->qd_id + !test_bit(QDF_USER, &qd->qd_flags); u64 107 fs/gfs2/quota.c error = gfs2_glock_get(sdp, 2 * (u64)id + !user, u64 1106 fs/gfs2/quota.c u64 dblock; u64 36 fs/gfs2/recovery.c u64 dblock; u64 35 fs/gfs2/rgrp.c #define NO_BLOCK ((u64)~0) u64 307 fs/gfs2/rgrp.c u64 first = rgd->rd_data0; u64 308 fs/gfs2/rgrp.c u64 last = first + rgd->rd_data; u64 490 fs/gfs2/rgrp.c u64 gfs2_ri_total(struct gfs2_sbd *sdp) u64 492 fs/gfs2/rgrp.c u64 total_data = 0; u64 593 fs/gfs2/rgrp.c u64 rgrp_count = ip->i_di.di_size; u64 911 fs/gfs2/rgrp.c u64 no_addr; u64 1141 fs/gfs2/rgrp.c u64 last_unlinked = NO_BLOCK; u64 1386 fs/gfs2/rgrp.c u64 gfs2_alloc_block(struct gfs2_inode *ip, unsigned int *n) u64 1392 fs/gfs2/rgrp.c u64 block; u64 1431 fs/gfs2/rgrp.c u64 gfs2_alloc_di(struct gfs2_inode *dip, u64 *generation) u64 1437 fs/gfs2/rgrp.c u64 block; u64 1529 fs/gfs2/rgrp.c u64 blkno = ip->i_no_addr; u64 49 fs/gfs2/rgrp.h u64 gfs2_alloc_block(struct gfs2_inode *ip, unsigned int *n); u64 50 fs/gfs2/rgrp.h u64 gfs2_alloc_di(struct gfs2_inode *ip, u64 *generation); u64 68 fs/gfs2/rgrp.h u64 gfs2_ri_total(struct gfs2_sbd *sdp); u64 49 fs/gfs2/trans.c sizeof(u64)); u64 190 fs/ioctl.c u64 len; u64 254 fs/ioctl.c u64 logical = 0, phys = 0, size = 0; u64 266 fs/ioctl.c length = (long long)min_t(u64, len, i_size_read(inode)); u64 319 fs/jbd2/recovery.c block |= (u64)be32_to_cpu(tag->t_blocknr_high) << 32; u64 3913 fs/jfs/jfs_dmap.c m = ((u64) 1 << (64 - 1)); u64 338 fs/jfs/jfs_dtree.c u64 blkno; u64 128 fs/jfs/jfs_dtree.h (dir_table_slot)->addr1 = ((u64)address64) >> 32;\ u64 651 fs/jfs/jfs_extent.c u64 m, k; u64 653 fs/jfs/jfs_extent.c for (i = 0, m = (u64) 1 << 63; i < 64; i++, m >>= 1) { u64 659 fs/jfs/jfs_extent.c k = (u64) 1 << i; u64 174 fs/jfs/jfs_types.h (((u64)((dasdp)->limit_hi) << 32) + __le32_to_cpu((dasdp)->limit_lo)) u64 177 fs/jfs/jfs_types.h (dasdp)->limit_hi = ((u64)limit) >> 32;\ u64 181 fs/jfs/jfs_types.h (((u64)((dasdp)->used_hi) << 32) + __le32_to_cpu((dasdp)->used_lo)) u64 184 fs/jfs/jfs_types.h (dasdp)->used_hi = ((u64)used) >> 32;\ u64 165 fs/jfs/jfs_xtree.c size = ((u64) ip->i_size + (JFS_SBI(ip->i_sb)->bsize - 1)) >> u64 49 fs/jfs/jfs_xtree.h (xad)->off1 = ((u64)offset64) >> 32;\ u64 54 fs/jfs/jfs_xtree.h (xad)->addr1 = ((u64)address64) >> 32;\ u64 513 fs/jfs/super.c sb->s_maxbytes = ((u64) sb->s_blocksize) << 40; u64 519 fs/jfs/super.c sb->s_maxbytes = min(((u64) PAGE_CACHE_SIZE << 32) - 1, sb->s_maxbytes); u64 631 fs/libfs.c int (*get)(void *, u64 *); u64 632 fs/libfs.c int (*set)(void *, u64); u64 689 fs/libfs.c u64 val; u64 709 fs/libfs.c u64 val; u64 170 fs/ncpfs/ioctl.c #define NCP_IOC_GETMOUNTUID64 _IOW('n', 2, u64) u64 374 fs/ncpfs/ioctl.c (u64 __user *)argp)) u64 153 fs/nfs/dir.c u64 *dir_cookie; u64 405 fs/nfs/dir.c u64 fileid; u64 75 fs/nfs/inode.c u64 nfs_compat_user_ino64(u64 fileid) u64 2178 fs/nfs/nfs4xdr.c (x) = (u64)ntohl(*p++) << 32; \ u64 506 fs/nfs/proc.c stat->tbytes = (u64)fsinfo.blocks * fsinfo.bsize; u64 507 fs/nfs/proc.c stat->fbytes = (u64)fsinfo.bfree * fsinfo.bsize; u64 508 fs/nfs/proc.c stat->abytes = (u64)fsinfo.bavail * fsinfo.bsize; u64 127 fs/nfsd/nfs3xdr.c u64 newsize; u64 155 fs/nfsd/nfs3xdr.c u64 f; u64 159 fs/nfsd/nfs3xdr.c p = xdr_encode_hyper(p, (u64)huge_encode_dev u64 163 fs/nfsd/nfs3xdr.c p = xdr_encode_hyper(p, (u64) fhp->fh_export->ex_fsid); u64 166 fs/nfsd/nfs3xdr.c f = ((u64*)fhp->fh_export->ex_uuid)[0]; u64 167 fs/nfsd/nfs3xdr.c f ^= ((u64*)fhp->fh_export->ex_uuid)[1]; u64 184 fs/nfsd/nfs3xdr.c p = xdr_encode_hyper(p, (u64) NFS3_MAXPATHLEN); u64 186 fs/nfsd/nfs3xdr.c p = xdr_encode_hyper(p, (u64) stat->size); u64 188 fs/nfsd/nfs3xdr.c p = xdr_encode_hyper(p, ((u64)stat->blocks) << 9); u64 250 fs/nfsd/nfs3xdr.c p = xdr_encode_hyper(p, (u64) fhp->fh_pre_size); u64 887 fs/nfsd/nfs3xdr.c u64 offset64 = offset; u64 1040 fs/nfsd/nfs3xdr.c u64 bs = s->f_bsize; u64 125 fs/nfsd/nfs4callback.c (x) = (u64)ntohl(*p++) << 32; \ u64 543 fs/nfsd/nfs4proc.c u64 cookie = readdir->rd_cookie; u64 2402 fs/nfsd/nfs4state.c #define LOFF_OVERFLOW(start, len) ((u64)(len) > ~(u64)(start)) u64 2509 fs/nfsd/nfs4state.c deny->ld_length = ~(u64)0; u64 2510 fs/nfsd/nfs4state.c if (fl->fl_end != ~(u64)0) u64 2607 fs/nfsd/nfs4state.c return ((length == 0) || ((length != ~(u64)0) && u64 2727 fs/nfsd/nfs4state.c if ((lock->lk_length == ~(u64)0) || u64 2729 fs/nfsd/nfs4state.c file_lock.fl_end = ~(u64)0; u64 2830 fs/nfsd/nfs4state.c if ((lockt->lt_length == ~(u64)0) || LOFF_OVERFLOW(lockt->lt_offset, lockt->lt_length)) u64 2831 fs/nfsd/nfs4state.c file_lock.fl_end = ~(u64)0; u64 2897 fs/nfsd/nfs4state.c if ((locku->lu_length == ~(u64)0) || LOFF_OVERFLOW(locku->lu_offset, locku->lu_length)) u64 2898 fs/nfsd/nfs4state.c file_lock.fl_end = ~(u64)0; u64 112 fs/nfsd/nfs4xdr.c (x) = (u64)ntohl(*p++) << 32; \ u64 1423 fs/nfsd/nfs4xdr.c u64 dummy64; u64 1554 fs/nfsd/nfs4xdr.c WRITE64((u64)exp->ex_fsid); u64 1555 fs/nfsd/nfs4xdr.c WRITE64((u64)0); u64 1654 fs/nfsd/nfs4xdr.c WRITE64((u64) statfs.f_ffree); u64 1659 fs/nfsd/nfs4xdr.c WRITE64((u64) statfs.f_ffree); u64 1664 fs/nfsd/nfs4xdr.c WRITE64((u64) statfs.f_files); u64 1681 fs/nfsd/nfs4xdr.c WRITE64(~(u64)0); u64 1696 fs/nfsd/nfs4xdr.c WRITE64((u64) svc_max_payload(rqstp)); u64 1701 fs/nfsd/nfs4xdr.c WRITE64((u64) svc_max_payload(rqstp)); u64 1741 fs/nfsd/nfs4xdr.c dummy64 = (u64)statfs.f_bavail * (u64)statfs.f_bsize; u64 1747 fs/nfsd/nfs4xdr.c dummy64 = (u64)statfs.f_bfree * (u64)statfs.f_bsize; u64 1753 fs/nfsd/nfs4xdr.c dummy64 = (u64)statfs.f_blocks * (u64)statfs.f_bsize; u64 1759 fs/nfsd/nfs4xdr.c dummy64 = (u64)stat.blocks << 9; u64 1919 fs/nfsd/nfs4xdr.c xdr_encode_hyper(cd->offset, (u64) offset); u64 2086 fs/nfsd/nfs4xdr.c WRITE64((u64)0); /* clientid */ u64 1136 fs/nfsd/vfs.c if ((u64)count > ~(u64)offset) u64 496 fs/ntfs/compress.c u64 cb_size_mask = cb_size - 1UL; u64 89 fs/ntfs/dir.c u64 mref; u64 632 fs/ntfs/dir.c u64 ntfs_lookup_inode_by_name(ntfs_inode *dir_ni, const ntfschar *uname, u64 642 fs/ntfs/dir.c u64 mref; u64 1268 fs/ntfs/dir.c bmp_pos &= ~(u64)((PAGE_CACHE_SIZE * 8) - 1); u64 314 fs/ntfs/layout.h typedef u64 MFT_REF; u64 327 fs/ntfs/layout.h #define ERR_MREF(x) ((u64)((s64)(x))) u64 158 fs/ntfs/logfile.c u64 file_size; u64 227 fs/ntfs/logfile.c file_size = (u64)sle64_to_cpu(ra->file_size); u64 62 fs/ntfs/mft.c index = (u64)ni->mft_no << vol->mft_record_size_bits >> u64 852 fs/ntfs/super.c if ((u64)ll >= 1ULL << 32) { u64 90 fs/ntfs/time.h u64 t = (u64)(sle64_to_cpu(time) - NTFS_TIME_OFFSET); u64 72 fs/ntfs/volume.h u64 serial_no; /* The volume serial number. */ u64 70 fs/ocfs2/alloc.c u64 blkno); u64 71 fs/ocfs2/alloc.c u64 (*eo_get_last_eb_blk)(struct ocfs2_extent_tree *et); u64 118 fs/ocfs2/alloc.c static u64 ocfs2_dinode_get_last_eb_blk(struct ocfs2_extent_tree *et); u64 148 fs/ocfs2/alloc.c static u64 ocfs2_dinode_get_last_eb_blk(struct ocfs2_extent_tree *et) u64 230 fs/ocfs2/alloc.c static u64 ocfs2_xattr_value_get_last_eb_blk(struct ocfs2_extent_tree *et) u64 279 fs/ocfs2/alloc.c static u64 ocfs2_xattr_tree_get_last_eb_blk(struct ocfs2_extent_tree *et) u64 353 fs/ocfs2/alloc.c static inline u64 ocfs2_et_get_last_eb_blk(struct ocfs2_extent_tree *et) u64 611 fs/ocfs2/alloc.c u64 blk_end = le64_to_cpu(ext->e_blkno); u64 635 fs/ocfs2/alloc.c u64 blkno = le64_to_cpu(insert_rec->e_blkno); u64 700 fs/ocfs2/alloc.c u64 last_eb_blk = 0; u64 743 fs/ocfs2/alloc.c u64 first_blkno; u64 858 fs/ocfs2/alloc.c u64 next_blkno, new_last_eb_blk; u64 1145 fs/ocfs2/alloc.c u64 blkno; u64 1496 fs/ocfs2/alloc.c u64 blkno; u64 1945 fs/ocfs2/alloc.c u64 blkno; u64 2531 fs/ocfs2/alloc.c u64 blkno; u64 3532 fs/ocfs2/alloc.c u64 len_blocks; u64 4490 fs/ocfs2/alloc.c u64 block; u64 4830 fs/ocfs2/alloc.c u64 start_blkno = ocfs2_clusters_to_blocks(inode->i_sb, phys); u64 5382 fs/ocfs2/alloc.c u64 start_blk; u64 5681 fs/ocfs2/alloc.c u64 start_blk; u64 5807 fs/ocfs2/alloc.c u64 free_blk; u64 5824 fs/ocfs2/alloc.c u64 bg_blkno; u64 6108 fs/ocfs2/alloc.c u64 deleted_eb = 0; u64 6307 fs/ocfs2/alloc.c u64 delete_blk = 0; u64 6544 fs/ocfs2/alloc.c u64 phys; u64 6658 fs/ocfs2/alloc.c u64 uninitialized_var(block); u64 6702 fs/ocfs2/alloc.c u64 phys; u64 85 fs/ocfs2/aops.c if ((u64)iblock >= ocfs2_clusters_to_blocks(inode->i_sb, u64 95 fs/ocfs2/aops.c u64 blkno = le64_to_cpu(fe->id2.i_list.l_recs[0].e_blkno) + u64 140 fs/ocfs2/aops.c u64 max_blocks = bh_result->b_size >> inode->i_blkbits; u64 141 fs/ocfs2/aops.c u64 p_blkno, count, past_eof; u64 505 fs/ocfs2/aops.c u64 p_blkno = 0; u64 565 fs/ocfs2/aops.c u64 p_blkno, inode_blocks, contig_blocks; u64 779 fs/ocfs2/aops.c u64 offset = page_offset(page) + block_start; u64 1245 fs/ocfs2/aops.c u64 v_blkno, p_blkno; u64 1942 fs/ocfs2/aops.c di->i_size = cpu_to_le64((u64)i_size_read(inode)); u64 363 fs/ocfs2/buffer_head_io.c u64 backup_blkno; u64 100 fs/ocfs2/cluster/heartbeat.c u64 ds_last_time; u64 101 fs/ocfs2/cluster/heartbeat.c u64 ds_last_generation; u64 142 fs/ocfs2/cluster/heartbeat.c u64 hr_generation; u64 412 fs/ocfs2/cluster/heartbeat.c u64 cputime; u64 557 fs/ocfs2/cluster/heartbeat.c u64 cputime; u64 69 fs/ocfs2/cluster/masklog.c u64 mask; u64 150 fs/ocfs2/cluster/masklog.h ((u64)(mask) >> 32) & bits.words[1] ) u64 153 fs/ocfs2/cluster/masklog.h bits.words[1] |= (u64)(mask) >> 32; \ u64 157 fs/ocfs2/cluster/masklog.h bits.words[1] &= ~((u64)(mask) >> 32); \ u64 162 fs/ocfs2/cluster/masklog.h [1] = (u64)(mask) >> 32, \ u64 201 fs/ocfs2/cluster/masklog.h u64 __m = MLOG_MASK_PREFIX | (mask); \ u64 33 fs/ocfs2/dcache.h u64 dl_parent_blkno; u64 90 fs/ocfs2/dir.c u64 p_blkno; u64 96 fs/ocfs2/dir.c if (((u64)block << inode->i_sb->s_blocksize_bits) >= u64 689 fs/ocfs2/dir.c u64 version = *f_version; u64 858 fs/ocfs2/dir.c u64 version = inode->i_version; u64 1222 fs/ocfs2/dir.c u64 blkno, bytes = blocks_wanted << sb->s_blocksize_bits; u64 1414 fs/ocfs2/dir.c u64 p_blkno, v_blkno; u64 276 fs/ocfs2/dlm/dlmast.c u64 cookie; u64 59 fs/ocfs2/dlm/dlmlock.c static u64 dlm_next_cookie = 1; u64 550 fs/ocfs2/dlm/dlmlock.c u64 tmpnode = node_num; u64 659 fs/ocfs2/dlm/dlmlock.c u64 tmpcookie; u64 102 fs/ocfs2/dlm/dlmrecovery.c static u64 dlm_get_next_mig_cookie(void); u64 106 fs/ocfs2/dlm/dlmrecovery.c static u64 dlm_mig_cookie = 1; u64 108 fs/ocfs2/dlm/dlmrecovery.c static u64 dlm_get_next_mig_cookie(void) u64 110 fs/ocfs2/dlm/dlmrecovery.c u64 c; u64 1100 fs/ocfs2/dlm/dlmrecovery.c u64 mig_cookie = be64_to_cpu(mres->mig_cookie); u64 1243 fs/ocfs2/dlm/dlmrecovery.c u64 mig_cookie = 0; u64 538 fs/ocfs2/dlmglue.c u64 inode_blkno = OCFS2_I(inode)->ip_blkno; u64 1810 fs/ocfs2/dlmglue.c static u64 ocfs2_pack_timespec(struct timespec *spec) u64 1812 fs/ocfs2/dlmglue.c u64 res; u64 1813 fs/ocfs2/dlmglue.c u64 sec = spec->tv_sec; u64 44 fs/ocfs2/export.c u64 ih_blkno; u64 87 fs/ocfs2/export.c u64 blkno; u64 143 fs/ocfs2/export.c u64 blkno; u64 205 fs/ocfs2/export.c handle.ih_blkno = (u64)le32_to_cpu(fid->raw[0]) << 32; u64 206 fs/ocfs2/export.c handle.ih_blkno |= (u64)le32_to_cpu(fid->raw[1]); u64 219 fs/ocfs2/export.c parent.ih_blkno = (u64)le32_to_cpu(fid->raw[3]) << 32; u64 220 fs/ocfs2/export.c parent.ih_blkno |= (u64)le32_to_cpu(fid->raw[4]); u64 291 fs/ocfs2/extent_map.c u64 last_eb_blk = le64_to_cpu(di->i_last_eb_blk); u64 682 fs/ocfs2/extent_map.c u64 boff = 0; u64 698 fs/ocfs2/extent_map.c boff += (v_blkno & (u64)(bpc - 1)); u64 705 fs/ocfs2/extent_map.c *ret_count -= v_blkno & (u64)(bpc - 1); u64 719 fs/ocfs2/extent_map.c u64 phys; u64 748 fs/ocfs2/extent_map.c u64 len_bytes, phys_bytes, virt_bytes; u64 797 fs/ocfs2/extent_map.c len_bytes = (u64)le16_to_cpu(rec.e_leaf_clusters) << osb->s_clustersize_bits; u64 799 fs/ocfs2/extent_map.c virt_bytes = (u64)le32_to_cpu(rec.e_cpos) << osb->s_clustersize_bits; u64 339 fs/ocfs2/file.c u64 cluster_bytes; u64 747 fs/ocfs2/file.c u64 start_off; u64 1156 fs/ocfs2/file.c u64 end = start + len; u64 1236 fs/ocfs2/file.c u64 phys_blkno = ocfs2_clusters_to_blocks(inode->i_sb, phys_cpos); u64 1332 fs/ocfs2/file.c u64 tmpend, end = start + len; u64 1365 fs/ocfs2/file.c tmpend = (u64)osb->s_clustersize + (start & ~(osb->s_clustersize - 1)); u64 58 fs/ocfs2/inode.c u64 fi_blkno; u64 34 fs/ocfs2/inode.h u64 ip_blkno; u64 830 fs/ocfs2/journal.c u64 v_blkno, p_blkno, p_blocks, num_blocks; u64 577 fs/ocfs2/localalloc.c u64 block_off; u64 914 fs/ocfs2/localalloc.c u64 la_start_blk; u64 915 fs/ocfs2/localalloc.c u64 blkno; u64 97 fs/ocfs2/namei.c #define OCFS2_ORPHAN_NAMELEN ((int)(2 * sizeof(u64))) u64 103 fs/ocfs2/namei.c u64 blkno; u64 362 fs/ocfs2/namei.c u64 fe_blkno = 0; u64 693 fs/ocfs2/namei.c u64 blkno; u64 948 fs/ocfs2/namei.c u64 newfe_blkno, old_de_ino; u64 1038 fs/ocfs2/namei.c u64 old_inode_parent; u64 1388 fs/ocfs2/namei.c u64 p_blkno, p_blocks; u64 1485 fs/ocfs2/namei.c u64 newsize; u64 222 fs/ocfs2/ocfs2.h u64 root_blkno; u64 223 fs/ocfs2/ocfs2.h u64 system_dir_blkno; u64 224 fs/ocfs2/ocfs2.h u64 bitmap_blkno; u64 230 fs/ocfs2/ocfs2.h u64 first_cluster_group_blkno; u64 282 fs/ocfs2/ocfs2.h u64 la_last_gd; u64 479 fs/ocfs2/ocfs2.h return (unsigned long)(blkno & (u64)ULONG_MAX); u64 482 fs/ocfs2/ocfs2.h static inline u64 ocfs2_clusters_to_blocks(struct super_block *sb, u64 488 fs/ocfs2/ocfs2.h return (u64)clusters << c_to_b_bits; u64 513 fs/ocfs2/ocfs2.h static inline u64 ocfs2_blocks_for_bytes(struct super_block *sb, u64 520 fs/ocfs2/ocfs2.h static inline u64 ocfs2_clusters_to_bytes(struct super_block *sb, u64 523 fs/ocfs2/ocfs2.h return (u64)clusters << OCFS2_SB(sb)->s_clustersize_bits; u64 526 fs/ocfs2/ocfs2.h static inline u64 ocfs2_align_bytes_to_clusters(struct super_block *sb, u64 533 fs/ocfs2/ocfs2.h return (u64)clusters << cl_bits; u64 536 fs/ocfs2/ocfs2.h static inline u64 ocfs2_align_bytes_to_blocks(struct super_block *sb, u64 539 fs/ocfs2/ocfs2.h u64 blocks; u64 969 fs/ocfs2/ocfs2_fs.h static inline u64 ocfs2_backup_super_blkno(struct super_block *sb, int index) u64 971 fs/ocfs2/ocfs2_fs.h u64 offset = OCFS2_BACKUP_SB_START; u64 64 fs/ocfs2/resize.c u64 blkno, gd_blkno, lgd_blkno = le64_to_cpu(gd->bg_blkno); u64 191 fs/ocfs2/resize.c u64 blkno; u64 286 fs/ocfs2/resize.c u64 lgd_blkno; u64 359 fs/ocfs2/slot_map.c u64 blkno; u64 199 fs/ocfs2/stackglue.h u64 ino, u64 293 fs/ocfs2/suballoc.c u64 bg_blkno; u64 929 fs/ocfs2/suballoc.c u64 fe_ptr, bg_ptr, prev_bg_ptr; u64 1032 fs/ocfs2/suballoc.c u64 blkoff; u64 1105 fs/ocfs2/suballoc.c u64 blkoff; u64 1231 fs/ocfs2/suballoc.c u64 next_group; u64 1384 fs/ocfs2/suballoc.c u64 hint_blkno = ac->ac_last_group; u64 1499 fs/ocfs2/suballoc.c u64 bg_blkno; u64 1519 fs/ocfs2/suballoc.c *blkno_start = bg_blkno + (u64) *suballoc_bit_start; u64 1535 fs/ocfs2/suballoc.c u64 bg_blkno; u64 1560 fs/ocfs2/suballoc.c *fe_blkno = bg_blkno + (u64) (*suballoc_bit); u64 1587 fs/ocfs2/suballoc.c u64 ocfs2_which_cluster_group(struct inode *inode, u32 cluster) u64 1639 fs/ocfs2/suballoc.c u64 bg_blkno = 0; u64 1848 fs/ocfs2/suballoc.c u64 blk = le64_to_cpu(di->i_blkno); u64 1850 fs/ocfs2/suballoc.c u64 bg_blkno = ocfs2_which_suballoc_group(blk, bit); u64 1864 fs/ocfs2/suballoc.c u64 bg_blkno; u64 54 fs/ocfs2/suballoc.h u64 ac_last_group; u64 55 fs/ocfs2/suballoc.h u64 ac_max_block; /* Highest block number to allocate. 0 is u64 128 fs/ocfs2/suballoc.h static inline u64 ocfs2_which_suballoc_group(u64 block, unsigned int bit) u64 130 fs/ocfs2/suballoc.h u64 group = block - (u64) bit; u64 165 fs/ocfs2/suballoc.h u64 ocfs2_which_cluster_group(struct inode *inode, u32 cluster); u64 102 fs/ocfs2/sysfile.c u64 blkno; u64 310 fs/ocfs2/xattr.c u64 phys_blkno = ocfs2_clusters_to_blocks(inode->i_sb, phys_cpos); u64 387 fs/ocfs2/xattr.c u64 block; u64 650 fs/ocfs2/xattr.c u64 blkno; u64 895 fs/ocfs2/xattr.c u64 blkno; u64 1508 fs/ocfs2/xattr.c u64 blk, bg_blkno; u64 1872 fs/ocfs2/xattr.c u64 first_blkno; u64 2113 fs/ocfs2/xattr.c u64 e_blkno = 0; u64 2260 fs/ocfs2/xattr.c u64 blkno; u64 2381 fs/ocfs2/xattr.c u64 p_blkno = 0; u64 2536 fs/ocfs2/xattr.c u64 p_blkno = 0; u64 2716 fs/ocfs2/xattr.c u64 blkno; u64 2888 fs/ocfs2/xattr.c u64 blkno = bucket->bhs[0]->b_blocknr; u64 3429 fs/ocfs2/xattr.c u64 to_blk_start = to_blk; u64 3574 fs/ocfs2/xattr.c u64 last_blk = prev_blk + bpc * (prev_clusters - 1); u64 3622 fs/ocfs2/xattr.c u64 block; u64 3756 fs/ocfs2/xattr.c u64 start_blk = start_bh->b_blocknr, end_blk; u64 3837 fs/ocfs2/xattr.c u64 p_blkno; u64 4187 fs/ocfs2/xattr.c u64 value_blk; u64 4689 fs/ocfs2/xattr.c u64 p_blkno; u64 97 fs/omfs/bitmap.c u64 tmp; u64 178 fs/omfs/bitmap.c u64 tmp; u64 74 fs/omfs/dir.c u64 block, dummy; u64 121 fs/omfs/dir.c u64 block; u64 168 fs/omfs/dir.c u64 block, prev; u64 224 fs/omfs/dir.c u64 *ptr; u64 233 fs/omfs/dir.c ptr = (u64 *) &bh->b_data[OMFS_DIR_START]; u64 334 fs/omfs/dir.c u64 ino = be64_to_cpu(header->h_self); u64 350 fs/omfs/dir.c u64 self; u64 443 fs/omfs/dir.c u64 fsblock; u64 53 fs/omfs/file.c u64 next, last; u64 91 fs/omfs/file.c u64 start, count; u64 139 fs/omfs/file.c u64 new_block = 0; u64 190 fs/omfs/file.c entry->e_blocks = cpu_to_be64((u64) new_count); u64 193 fs/omfs/file.c be64_to_cpu(~terminator->e_blocks) + (u64) new_count)); u64 240 fs/omfs/file.c u64 new_block; u64 24 fs/omfs/inode.c u64 new_block; u64 99 fs/omfs/inode.c u64 ctime; u64 194 fs/omfs/inode.c u64 ctime; u64 11 fs/omfs/omfs.h u64 s_num_blocks; u64 12 fs/omfs/omfs.h u64 s_bitmap_ino; u64 13 fs/omfs/omfs.h u64 s_root_ino; u64 139 fs/partitions/efi.c static u64 u64 292 fs/partitions/efi.c u64 lastlba; u64 511 fs/partitions/efi.c u64 lastlba; u64 86 fs/partitions/efi.h u64 required_to_function:1; u64 87 fs/partitions/efi.h u64 reserved:47; u64 88 fs/partitions/efi.h u64 type_guid_specific:16; u64 716 fs/partitions/ldm.c static u64 ldm_get_vnum (const u8 *block) u64 718 fs/partitions/ldm.c u64 tmp = 0; u64 120 fs/partitions/ldm.h u64 logical_disk_start; u64 121 fs/partitions/ldm.h u64 logical_disk_size; u64 122 fs/partitions/ldm.h u64 config_start; u64 123 fs/partitions/ldm.h u64 config_size; u64 129 fs/partitions/ldm.h u64 bitmap1_start; u64 130 fs/partitions/ldm.h u64 bitmap1_size; u64 132 fs/partitions/ldm.h u64 bitmap2_start; u64 133 fs/partitions/ldm.h u64 bitmap2_size; u64 146 fs/partitions/ldm.h u64 parent_id; u64 162 fs/partitions/ldm.h u64 start; u64 163 fs/partitions/ldm.h u64 size; /* start, size and vol_off in sectors */ u64 164 fs/partitions/ldm.h u64 volume_offset; u64 165 fs/partitions/ldm.h u64 parent_id; u64 166 fs/partitions/ldm.h u64 disk_id; u64 175 fs/partitions/ldm.h u64 size; u64 187 fs/partitions/ldm.h u64 obj_id; u64 3073 fs/proc/base.c filp->f_version = (u64)tid; u64 508 fs/proc/proc_misc.c u64 sum = 0; u64 730 fs/proc/proc_misc.c #define KPMSIZE sizeof(u64) u64 740 fs/proc/proc_misc.c u64 __user *out = (u64 __user *)buf; u64 745 fs/proc/proc_misc.c u64 pcount; u64 806 fs/proc/proc_misc.c u64 __user *out = (u64 __user *)buf; u64 811 fs/proc/proc_misc.c u64 kflags, uflags; u64 311 fs/proc/task_mmu.c u64 pss; u64 502 fs/proc/task_mmu.c u64 __user *out, *end; u64 505 fs/proc/task_mmu.c #define PM_ENTRY_BYTES sizeof(u64) u64 513 fs/proc/task_mmu.c #define PM_PSHIFT(x) (((u64) (x) << PM_PSHIFT_OFFSET) & PM_PSHIFT_MASK) u64 547 fs/proc/task_mmu.c static u64 swap_pte_to_pagemap_entry(pte_t pte) u64 576 fs/proc/task_mmu.c u64 pfn = PM_NOT_PRESENT; u64 685 fs/proc/task_mmu.c pm.out = (u64 *)buf; u64 686 fs/proc/task_mmu.c pm.end = (u64 *)(buf + count); u64 33 fs/proc/vmcore.c static u64 vmcore_size; u64 77 fs/proc/vmcore.c static u64 map_offset_to_paddr(loff_t offset, struct list_head *vc_list, u64 81 fs/proc/vmcore.c u64 paddr; u64 84 fs/proc/vmcore.c u64 start, end; u64 105 fs/proc/vmcore.c u64 start, nr_bytes; u64 182 fs/proc/vmcore.c static u64 __init get_vmcore_size_elf64(char *elfptr) u64 185 fs/proc/vmcore.c u64 size; u64 199 fs/proc/vmcore.c static u64 __init get_vmcore_size_elf32(char *elfptr) u64 202 fs/proc/vmcore.c u64 size; u64 225 fs/proc/vmcore.c u64 phdr_sz = 0, note_off; u64 233 fs/proc/vmcore.c u64 offset, max_sz, sz, real_sz = 0; u64 306 fs/proc/vmcore.c u64 phdr_sz = 0, note_off; u64 314 fs/proc/vmcore.c u64 offset, max_sz, sz, real_sz = 0; u64 498 fs/proc/vmcore.c u64 addr; u64 554 fs/proc/vmcore.c u64 addr; u64 609 fs/proc/vmcore.c u64 addr; u64 241 fs/readdir.c int reclen = ALIGN(NAME_OFFSET(dirent) + namlen + 1, sizeof(u64)); u64 392 fs/select.c if ((u64)tv.tv_sec >= (u64)MAX_INT64_SECONDS) u64 407 fs/select.c rtv.tv_usec = jiffies_to_usecs(do_div((*(u64*)&timeout), HZ)); u64 447 fs/select.c if ((u64)ts.tv_sec >= (u64)MAX_INT64_SECONDS) u64 473 fs/select.c rts.tv_nsec = jiffies_to_usecs(do_div((*(u64*)&timeout), HZ)) * u64 724 fs/select.c restart_block->arg3 = (u64)timeout >> 32; u64 757 fs/select.c restart_block->arg3 = (u64)timeout_jiffies >> 32; u64 778 fs/select.c if ((u64)ts.tv_sec >= (u64)MAX_INT64_SECONDS) u64 821 fs/select.c rts.tv_nsec = jiffies_to_usecs(do_div((*(u64*)&timeout), HZ)) * u64 521 fs/smbfs/proc.c #define NTFS_TIME_OFFSET ((u64)(369*365 + 89) * 24 * 3600 * 10000000) u64 533 fs/smbfs/proc.c u64 t = ntutc - NTFS_TIME_OFFSET; u64 540 fs/smbfs/proc.c static u64 u64 545 fs/smbfs/proc.c return ((u64)ts.tv_sec) * 10000000 + ts.tv_nsec/100 + NTFS_TIME_OFFSET; u64 2079 fs/smbfs/proc.c u64 size, disk_bytes; u64 3078 fs/smbfs/proc.c u64 nttime; u64 3398 fs/smbfs/proc.c u64 caps; u64 29 fs/timerfd.c u64 ticks; u64 113 fs/timerfd.c u64 ticks = 0; u64 157 fs/timerfd.c res = put_user(ticks, (u64 __user *) buf) ? -EFAULT: sizeof(ticks); u64 477 fs/ubifs/key.h to->u64[0] = from->u64[0]; u64 261 fs/ubifs/ubifs.h uint64_t u64[CUR_MAX_KEY_LEN/8]; u64 26 fs/ufs/balloc.c #define INVBLOCK ((u64)-1L) u64 28 fs/ufs/balloc.c static u64 ufs_add_fragments(struct inode *, u64, unsigned, unsigned, int *); u64 29 fs/ufs/balloc.c static u64 ufs_alloc_fragments(struct inode *, unsigned, u64, unsigned, int *); u64 30 fs/ufs/balloc.c static u64 ufs_alloccg_block(struct inode *, struct ufs_cg_private_info *, u64, int *); u64 31 fs/ufs/balloc.c static u64 ufs_bitmap_search (struct super_block *, struct ufs_cg_private_info *, u64, unsigned); u64 46 fs/ufs/balloc.c u64 blkno; u64 148 fs/ufs/balloc.c u64 blkno; u64 354 fs/ufs/balloc.c u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment, u64 362 fs/ufs/balloc.c u64 tmp, request, result; u64 505 fs/ufs/balloc.c static u64 ufs_add_fragments(struct inode *inode, u64 fragment, u64 589 fs/ufs/balloc.c static u64 ufs_alloc_fragments(struct inode *inode, unsigned cgno, u64 598 fs/ufs/balloc.c u64 result; u64 708 fs/ufs/balloc.c static u64 ufs_alloccg_block(struct inode *inode, u64 716 fs/ufs/balloc.c u64 result, blkno; u64 807 fs/ufs/balloc.c static u64 ufs_bitmap_search(struct super_block *sb, u64 826 fs/ufs/balloc.c u64 result; u64 45 fs/ufs/inode.c static u64 ufs_frag_map(struct inode *inode, sector_t frag); u64 86 fs/ufs/inode.c static u64 ufs_frag_map(struct inode *inode, sector_t frag) u64 91 fs/ufs/inode.c u64 mask = (u64) uspi->s_apbmask>>uspi->s_fpbshift; u64 95 fs/ufs/inode.c u64 ret = 0L; u64 99 fs/ufs/inode.c u64 temp = 0L; u64 130 fs/ufs/inode.c ret = (u64) (uspi->s_sbbase + fs32_to_cpu(sb, block) + (frag & uspi->s_fpbmask)); u64 143 fs/ufs/inode.c temp = (u64)(uspi->s_sbbase) + fs64_to_cpu(sb, u2_block); u64 144 fs/ufs/inode.c bh = sb_bread(sb, temp +(u64) (n>>shift)); u64 152 fs/ufs/inode.c temp = (u64)uspi->s_sbbase + fs64_to_cpu(sb, u2_block); u64 153 fs/ufs/inode.c ret = temp + (u64) (frag & uspi->s_fpbmask); u64 183 fs/ufs/inode.c u64 tmp, goal, lastfrag, block, lastblock; u64 339 fs/ufs/inode.c u64 tmp, goal, block; u64 423 fs/ufs/inode.c u64 phys64 = 0; u64 25 fs/ufs/swab.h static inline u64 u64 73 fs/ufs/truncate.c u64 frag1, frag2, frag3, frag4, block1, block2; u64 178 fs/ufs/truncate.c u64 tmp, indirect_block, i, frag_to_free; u64 262 fs/ufs/truncate.c u64 i, tmp, dindirect_block; u64 327 fs/ufs/truncate.c u64 tindirect_block, tmp, i; u64 390 fs/ufs/truncate.c u64 phys64; u64 79 fs/ufs/ufs.h extern u64 ufs_new_fragments(struct inode *, void *, u64, u64, u64 145 fs/ufs/ufs.h static inline u64 ufs_dtog(struct ufs_sb_private_info * uspi, u64 b) u64 222 fs/ufs/ufs_fs.h #define ufs_inotofsba(x) (((u64)ufs_cgimin(ufs_inotocg(x))) + ufs_inotocgoff(x) / uspi->s_inopf) u64 23 fs/ufs/util.c u64 count = 0; u64 52 fs/ufs/util.c u64 count = 0; u64 88 fs/ufs/util.h static inline u64 u64 112 fs/ufs/util.h static inline u64 u64 360 fs/ufs/util.h static inline u64 u64 564 fs/ufs/util.h static inline u64 ufs_data_ptr_to_cpu(struct super_block *sb, void *p) u64 63 fs/xfs/linux-2.6/xfs_export.h u64 ino; u64 65 fs/xfs/linux-2.6/xfs_export.h u64 parent_ino; u64 250 fs/xfs/linux-2.6/xfs_file.c u64 ino; u64 276 fs/xfs/linux-2.6/xfs_file.c reclen = ALIGN(sizeof(struct hack_dirent) + namlen, sizeof(u64)); u64 348 fs/xfs/linux-2.6/xfs_file.c sizeof(u64)); u64 937 include/acpi/aclocal.h ACPI_COMMON_DEBUG_MEM_HEADER u64 user_space; u64 72 include/acpi/acmacros.h #define ACPI_GET64(ptr) *ACPI_CAST_PTR (u64, ptr) u64 76 include/acpi/acmacros.h #define ACPI_SET64(ptr) *ACPI_CAST_PTR (u64, ptr) u64 152 include/acpi/acmacros.h #define ACPI_MOVE_16_TO_64(d,s) {(*(u64 *)(void *)(d))=0;\ u64 165 include/acpi/acmacros.h #define ACPI_MOVE_32_TO_64(d,s) {(*(u64 *)(void *)(d))=0;\ u64 202 include/acpi/acmacros.h #define ACPI_MOVE_16_TO_64(d,s) *(u64 *)(void *)(d) = *(u16 *)(void *)(s) u64 208 include/acpi/acmacros.h #define ACPI_MOVE_32_TO_64(d,s) *(u64 *)(void *)(d) = *(u32 *)(void *)(s) u64 214 include/acpi/acmacros.h #define ACPI_MOVE_64_TO_64(d,s) *(u64 *)(void *)(d) = *(u64 *)(void *)(s) u64 229 include/acpi/acmacros.h #define ACPI_MOVE_16_TO_64(d,s) {(*(u64 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d,s);} u64 240 include/acpi/acmacros.h #define ACPI_MOVE_32_TO_64(d,s) {(*(u64 *)(void *)(d)) = 0; ACPI_MOVE_32_TO_32(d,s);} u64 248 include/acpi/acpiosxf.h u64 acpi_os_get_timer(void); u64 110 include/acpi/actbl.h u64 address; /* 64-bit address of struct or register */ u64 126 include/acpi/actbl.h u64 xsdt_physical_address; /* 64-bit physical address of the XSDT (ACPI 2.0+) */ u64 146 include/acpi/actbl.h u64 table_offset_entry[1]; /* Array of pointers to ACPI tables */ u64 162 include/acpi/actbl.h u64 xfirmware_waking_vector; /* 64-bit version of the Firmware Waking Vector (ACPI 2.0+) */ u64 227 include/acpi/actbl.h u64 Xfacs; /* 64-bit physical address of FACS */ u64 228 include/acpi/actbl.h u64 Xdsdt; /* 64-bit physical address of DSDT */ u64 112 include/acpi/actbl1.h u64 value; /* Value used with Read/Write register */ u64 113 include/acpi/actbl1.h u64 mask; /* Bitmask required for this register instruction */ u64 234 include/acpi/actbl1.h u64 address; /* Physical addresss of the error region */ u64 274 include/acpi/actbl1.h u64 reserved; u64 369 include/acpi/actbl1.h u64 address; /* Register Base Address */ u64 382 include/acpi/actbl1.h u64 base_address; /* 4_k aligned base address */ u64 383 include/acpi/actbl1.h u64 end_address; /* 4_k aligned limit address */ u64 584 include/acpi/actbl1.h u64 control_init_data; u64 676 include/acpi/actbl1.h u64 global_capability_data; u64 677 include/acpi/actbl1.h u64 global_control_data; u64 969 include/acpi/actbl1.h u64 address; /* APIC physical address */ u64 979 include/acpi/actbl1.h u64 address; /* SAPIC physical address */ u64 1051 include/acpi/actbl1.h u64 address; /* Base address, processor-relative */ u64 1079 include/acpi/actbl1.h u64 locality_count; u64 1145 include/acpi/actbl1.h u64 reserved; /* Reserved, must be zero */ u64 1176 include/acpi/actbl1.h u64 base_address; u64 1177 include/acpi/actbl1.h u64 length; u64 1180 include/acpi/actbl1.h u64 reserved2; /* Reserved, must be zero */ u64 1199 include/acpi/actbl1.h u64 log_address; /* Address of the event log area */ u64 152 include/acpi/actypes.h typedef u64 acpi_size; u64 153 include/acpi/actypes.h typedef u64 acpi_io_address; u64 154 include/acpi/actypes.h typedef u64 acpi_physical_address; u64 307 include/acpi/actypes.h u64 full; u64 1137 include/acpi/actypes.h ACPI_RESOURCE_ADDRESS_COMMON u64 granularity; u64 1138 include/acpi/actypes.h u64 minimum; u64 1139 include/acpi/actypes.h u64 maximum; u64 1140 include/acpi/actypes.h u64 translation_offset; u64 1141 include/acpi/actypes.h u64 address_length; u64 1147 include/acpi/actypes.h u64 granularity; u64 1148 include/acpi/actypes.h u64 minimum; u64 1149 include/acpi/actypes.h u64 maximum; u64 1150 include/acpi/actypes.h u64 translation_offset; u64 1151 include/acpi/actypes.h u64 address_length; u64 1152 include/acpi/actypes.h u64 type_specific; u64 1170 include/acpi/actypes.h u64 address; u64 214 include/acpi/amlresrc.h u64 granularity; u64 215 include/acpi/amlresrc.h u64 minimum; u64 216 include/acpi/amlresrc.h u64 maximum; u64 217 include/acpi/amlresrc.h u64 translation_offset; u64 218 include/acpi/amlresrc.h u64 address_length; u64 219 include/acpi/amlresrc.h u64 type_specific; u64 226 include/acpi/amlresrc.h AML_RESOURCE_ADDRESS_COMMON u64 granularity; u64 227 include/acpi/amlresrc.h u64 minimum; u64 228 include/acpi/amlresrc.h u64 maximum; u64 229 include/acpi/amlresrc.h u64 translation_offset; u64 230 include/acpi/amlresrc.h u64 address_length; u64 263 include/acpi/amlresrc.h u64 address; u64 52 include/acpi/processor.h u64 address; u64 76 include/acpi/processor.h u64 time; u64 110 include/acpi/processor.h u64 address; u64 151 include/acpi/processor.h u64 address; u64 282 include/asm-frv/bitops.h u64 ll; u64 364 include/asm-frv/bitops.h u64 ll; u64 37 include/asm-generic/cmpxchg-local.h case 8: prev = *(u64 *)ptr; u64 39 include/asm-generic/cmpxchg-local.h *(u64 *)ptr = (u64)new; u64 51 include/asm-generic/cmpxchg-local.h static inline u64 __cmpxchg64_local_generic(volatile void *ptr, u64 54 include/asm-generic/cmpxchg-local.h u64 prev; u64 58 include/asm-generic/cmpxchg-local.h prev = *(u64 *)ptr; u64 60 include/asm-generic/cmpxchg-local.h *(u64 *)ptr = new; u64 24 include/asm-generic/cputime.h typedef u64 cputime64_t; u64 31 include/asm-generic/cputime.h #define cputime_to_cputime64(__ct) ((u64) __ct) u64 24 include/asm-m32r/types.h typedef u64 dma64_addr_t; u64 37 include/asm-parisc/compat.h typedef u64 compat_u64; u64 113 include/asm-parisc/compat.h u64 sc_fr[32]; u64 10 include/asm-parisc/dma-mapping.h int (*dma_supported)(struct device *dev, u64 mask); u64 86 include/asm-parisc/io.h ret = ((u64) gsc_readl(addr)) << 32; u64 27 include/asm-parisc/parisc-device.h u64 dma_mask; /* DMA mask for I/O */ u64 31 include/asm-parisc/ropes.h u64 *pdir_base; /* physical base address */ u64 237 include/asm-parisc/ropes.h u64 bus_mode; u64 30 include/asm-parisc/types.h typedef u64 dma64_addr_t; u64 184 include/asm-parisc/uaccess.h u64 __val64 = (u64)(__val); \ u64 198 include/asm-parisc/uaccess.h u64 __val64 = (u64)(__val); \ u64 201 include/asm-x86/amd_iommu_types.h u64 *pt_root; /* page table root pointer */ u64 229 include/asm-x86/amd_iommu_types.h u64 **pte_pages; u64 262 include/asm-x86/amd_iommu_types.h u64 mmio_phys; u64 278 include/asm-x86/amd_iommu_types.h u64 exclusion_start; u64 280 include/asm-x86/amd_iommu_types.h u64 exclusion_length; u64 329 include/asm-x86/amd_iommu_types.h u64 address_start; u64 331 include/asm-x86/amd_iommu_types.h u64 address_end; u64 59 include/asm-x86/apic.h extern u64 xapic_icr_read(void); u64 108 include/asm-x86/apic.h u64 (*icr_read)(void); u64 39 include/asm-x86/calgary.h u64 tar_val; /* Table Address Register */ u64 40 include/asm-x86/compat.h typedef u64 __attribute__((aligned(4))) compat_u64; u64 36 include/asm-x86/div64.h static inline u64 div_u64_rem(u64 dividend, u32 divisor, u32 *remainder) u64 39 include/asm-x86/div64.h u64 v64; u64 53 include/asm-x86/dma-mapping.h int (*dma_supported)(struct device *hwdev, u64 mask); u64 75 include/asm-x86/e820.h extern u64 e820_update_range(u64 start, u64 size, unsigned old_type, u64 77 include/asm-x86/e820.h extern u64 e820_remove_range(u64 start, u64 size, unsigned old_type, u64 105 include/asm-x86/e820.h extern u64 find_e820_area(u64 start, u64 end, u64 size, u64 align); u64 106 include/asm-x86/e820.h extern u64 find_e820_area_size(u64 start, u64 *sizep, u64 align); u64 111 include/asm-x86/e820.h extern u64 early_reserve_e820(u64 startt, u64 sizet, u64 align); u64 122 include/asm-x86/e820.h extern u64 e820_hole_size(u64 start, u64 end); u64 42 include/asm-x86/efi.h extern u64 efi_call0(void *fp); u64 43 include/asm-x86/efi.h extern u64 efi_call1(void *fp, u64 arg1); u64 44 include/asm-x86/efi.h extern u64 efi_call2(void *fp, u64 arg1, u64 arg2); u64 45 include/asm-x86/efi.h extern u64 efi_call3(void *fp, u64 arg1, u64 arg2, u64 arg3); u64 46 include/asm-x86/efi.h extern u64 efi_call4(void *fp, u64 arg1, u64 arg2, u64 arg3, u64 arg4); u64 47 include/asm-x86/efi.h extern u64 efi_call5(void *fp, u64 arg1, u64 arg2, u64 arg3, u64 49 include/asm-x86/efi.h extern u64 efi_call6(void *fp, u64 arg1, u64 arg2, u64 arg3, u64 55 include/asm-x86/efi.h efi_call1((void *)(f), (u64)(a1)) u64 57 include/asm-x86/efi.h efi_call2((void *)(f), (u64)(a1), (u64)(a2)) u64 59 include/asm-x86/efi.h efi_call3((void *)(f), (u64)(a1), (u64)(a2), (u64)(a3)) u64 61 include/asm-x86/efi.h efi_call4((void *)(f), (u64)(a1), (u64)(a2), (u64)(a3), \ u64 62 include/asm-x86/efi.h (u64)(a4)) u64 64 include/asm-x86/efi.h efi_call5((void *)(f), (u64)(a1), (u64)(a2), (u64)(a3), \ u64 65 include/asm-x86/efi.h (u64)(a4), (u64)(a5)) u64 67 include/asm-x86/efi.h efi_call6((void *)(f), (u64)(a1), (u64)(a2), (u64)(a3), \ u64 68 include/asm-x86/efi.h (u64)(a4), (u64)(a5), (u64)(a6)) u64 73 include/asm-x86/efi.h efi_call1((void *)(efi.systab->runtime->f), (u64)(a1)) u64 75 include/asm-x86/efi.h efi_call2((void *)(efi.systab->runtime->f), (u64)(a1), (u64)(a2)) u64 77 include/asm-x86/efi.h efi_call3((void *)(efi.systab->runtime->f), (u64)(a1), (u64)(a2), \ u64 78 include/asm-x86/efi.h (u64)(a3)) u64 80 include/asm-x86/efi.h efi_call4((void *)(efi.systab->runtime->f), (u64)(a1), (u64)(a2), \ u64 81 include/asm-x86/efi.h (u64)(a3), (u64)(a4)) u64 83 include/asm-x86/efi.h efi_call5((void *)(efi.systab->runtime->f), (u64)(a1), (u64)(a2), \ u64 84 include/asm-x86/efi.h (u64)(a3), (u64)(a4), (u64)(a5)) u64 86 include/asm-x86/efi.h efi_call6((void *)(efi.systab->runtime->f), (u64)(a1), (u64)(a2), \ u64 87 include/asm-x86/efi.h (u64)(a3), (u64)(a4), (u64)(a5), (u64)(a6)) u64 7 include/asm-x86/hardirq.h extern u64 arch_irq_stat_cpu(unsigned int cpu); u64 10 include/asm-x86/hardirq.h extern u64 arch_irq_stat(void); u64 147 include/asm-x86/kvm_host.h u64 *parent_ptes[NR_PTE_CHAIN_ENTRIES]; u64 184 include/asm-x86/kvm_host.h u64 *spt; u64 193 include/asm-x86/kvm_host.h u64 *parent_pte; /* !multimapped */ u64 214 include/asm-x86/kvm_host.h u64 *pae_root; u64 218 include/asm-x86/kvm_host.h u64 host_tsc; u64 230 include/asm-x86/kvm_host.h u64 pdptrs[4]; /* pae */ u64 231 include/asm-x86/kvm_host.h u64 shadow_efer; u64 232 include/asm-x86/kvm_host.h u64 apic_base; u64 236 include/asm-x86/kvm_host.h u64 ia32_misc_enable_msr; u64 248 include/asm-x86/kvm_host.h u64 *last_pte_updated; u64 298 include/asm-x86/kvm_host.h u64 mtrr[0x100]; u64 396 include/asm-x86/kvm_host.h int (*get_msr)(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata); u64 397 include/asm-x86/kvm_host.h int (*set_msr)(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); u64 398 include/asm-x86/kvm_host.h u64 (*get_segment_base)(struct kvm_vcpu *vcpu, int seg); u64 409 include/asm-x86/kvm_host.h void (*set_efer)(struct kvm_vcpu *vcpu, u64 efer); u64 629 include/asm-x86/kvm_host.h u64 value; u64 21 include/asm-x86/msr.h return low | ((u64)high << 32); u64 32 include/asm-x86/msr.h #define EAX_EDX_VAL(val, low, high) ((low) | ((u64)(high) << 32)) u64 138 include/asm-x86/msr.h u64 __val = native_read_msr((msr)); \ u64 152 include/asm-x86/msr.h native_write_msr((msr), (u32)((u64)(val)), (u32)((u64)(val) >> 32)) u64 164 include/asm-x86/msr.h u64 __val = native_read_msr_safe((msr), &__err); \ u64 193 include/asm-x86/msr.h u64 _l = native_read_pmc((counter)); \ u64 8 include/asm-x86/numa_64.h u64 start; u64 9 include/asm-x86/numa_64.h u64 end; u64 37 include/asm-x86/page_32.h typedef u64 pteval_t; u64 38 include/asm-x86/page_32.h typedef u64 pmdval_t; u64 39 include/asm-x86/page_32.h typedef u64 pudval_t; u64 40 include/asm-x86/page_32.h typedef u64 pgdval_t; u64 41 include/asm-x86/page_32.h typedef u64 pgprotval_t; u64 42 include/asm-x86/page_32.h typedef u64 phys_addr_t; u64 143 include/asm-x86/paravirt.h u64 (*read_msr_amd)(unsigned int msr, int *err); u64 144 include/asm-x86/paravirt.h u64 (*read_msr)(unsigned int msr, int *err); u64 147 include/asm-x86/paravirt.h u64 (*read_tsc)(void); u64 148 include/asm-x86/paravirt.h u64 (*read_pmc)(int counter); u64 524 include/asm-x86/paravirt.h __ret = (rettype)((((u64)__edx) << 32) | __eax); \ u64 724 include/asm-x86/paravirt.h static inline u64 paravirt_read_msr(unsigned msr, int *err) u64 726 include/asm-x86/paravirt.h return PVOP_CALL2(u64, pv_cpu_ops.read_msr, msr, err); u64 728 include/asm-x86/paravirt.h static inline u64 paravirt_read_msr_amd(unsigned msr, int *err) u64 730 include/asm-x86/paravirt.h return PVOP_CALL2(u64, pv_cpu_ops.read_msr_amd, msr, err); u64 741 include/asm-x86/paravirt.h u64 _l = paravirt_read_msr(msr, &_err); \ u64 757 include/asm-x86/paravirt.h #define wrmsrl(msr, val) wrmsr(msr, (u32)((u64)(val)), ((u64)(val))>>32) u64 764 include/asm-x86/paravirt.h u64 _l = paravirt_read_msr(msr, &_err); \ u64 785 include/asm-x86/paravirt.h static inline u64 paravirt_read_tsc(void) u64 787 include/asm-x86/paravirt.h return PVOP_CALL0(u64, pv_cpu_ops.read_tsc); u64 792 include/asm-x86/paravirt.h u64 _l = paravirt_read_tsc(); \ u64 806 include/asm-x86/paravirt.h return PVOP_CALL1(u64, pv_cpu_ops.read_pmc, counter); u64 811 include/asm-x86/paravirt.h u64 _l = paravirt_read_pmc(counter); \ u64 818 include/asm-x86/paravirt.h return PVOP_CALL1(u64, pv_cpu_ops.read_tscp, aux); u64 1064 include/asm-x86/paravirt.h val, (u64)val >> 32); u64 1079 include/asm-x86/paravirt.h pte.pte, (u64)pte.pte >> 32); u64 1093 include/asm-x86/paravirt.h pte.pte, (u64)pte.pte >> 32); u64 1110 include/asm-x86/paravirt.h val, (u64)val >> 32); u64 1124 include/asm-x86/paravirt.h pgd.pgd, (u64)pgd.pgd >> 32); u64 1159 include/asm-x86/paravirt.h pte.pte, (u64)pte.pte >> 32); u64 1180 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.set_pmd, pmdp, val, (u64)val >> 32); u64 1192 include/asm-x86/paravirt.h val, (u64)val >> 32); u64 1206 include/asm-x86/paravirt.h pmd.pmd, (u64)pmd.pmd >> 32); u64 1220 include/asm-x86/paravirt.h val, (u64)val >> 32); u64 1232 include/asm-x86/paravirt.h val, (u64)val >> 32); u64 1246 include/asm-x86/paravirt.h pud.pud, (u64)pud.pud >> 32); u64 1260 include/asm-x86/paravirt.h val, (u64)val >> 32); u64 225 include/asm-x86/processor.h u64 sp0; u64 226 include/asm-x86/processor.h u64 sp1; u64 227 include/asm-x86/processor.h u64 sp2; u64 228 include/asm-x86/processor.h u64 reserved2; u64 229 include/asm-x86/processor.h u64 ist[7]; u64 308 include/asm-x86/processor.h u64 rip; /* Instruction Pointer */ u64 309 include/asm-x86/processor.h u64 rdp; /* Data Pointer */ u64 357 include/asm-x86/processor.h u64 xstate_bv; u64 358 include/asm-x86/processor.h u64 reserved1[2]; u64 359 include/asm-x86/processor.h u64 reserved2[5]; u64 28 include/asm-x86/pvclock-abi.h u64 tsc_timestamp; u64 29 include/asm-x86/pvclock-abi.h u64 system_time; u64 25 include/asm-x86/types.h typedef u64 dma64_addr_t; u64 28 include/asm-x86/types.h typedef u64 dma_addr_t; u64 44 include/asm-x86/uaccess_64.h case 8:__get_user_asm(*(u64 *)dst, (u64 __user *)src, u64 48 include/asm-x86/uaccess_64.h __get_user_asm(*(u64 *)dst, (u64 __user *)src, u64 57 include/asm-x86/uaccess_64.h __get_user_asm(*(u64 *)dst, (u64 __user *)src, u64 61 include/asm-x86/uaccess_64.h __get_user_asm(*(u64 *)(8 + (char *)dst), u64 62 include/asm-x86/uaccess_64.h (u64 __user *)(8 + (char __user *)src), u64 86 include/asm-x86/uaccess_64.h case 8:__put_user_asm(*(u64 *)src, (u64 __user *)dst, u64 90 include/asm-x86/uaccess_64.h __put_user_asm(*(u64 *)src, (u64 __user *)dst, u64 99 include/asm-x86/uaccess_64.h __put_user_asm(*(u64 *)src, (u64 __user *)dst, u64 104 include/asm-x86/uaccess_64.h __put_user_asm(1[(u64 *)src], 1 + (u64 __user *)dst, u64 149 include/asm-x86/uaccess_64.h u64 tmp; u64 150 include/asm-x86/uaccess_64.h __get_user_asm(tmp, (u64 __user *)src, u64 153 include/asm-x86/uaccess_64.h __put_user_asm(tmp, (u64 __user *)dst, u64 58 include/asm-x86/uv/bios.h u64 v0; u64 59 include/asm-x86/uv/bios.h u64 v1; u64 60 include/asm-x86/uv/bios.h u64 v2; u64 237 include/asm-x86/vmi.h u64 efer; u64 40 include/asm-x86/vmi_time.h u64 (*get_cycle_frequency)(void); u64 41 include/asm-x86/vmi_time.h u64 (*get_cycle_counter)(int); u64 42 include/asm-x86/vmi_time.h u64 (*get_wallclock)(void); u64 44 include/asm-x86/vmi_time.h void (*set_alarm)(u32 flags, u64 expiry, u64 period); u64 27 include/asm-x86/xcr.h static inline u64 xgetbv(u32 index) u64 34 include/asm-x86/xcr.h return eax + ((u64)edx << 32); u64 63 include/asm-x86/xen/hypervisor.h u64 jiffies_to_st(unsigned long jiffies); u64 31 include/asm-x86/xen/page.h ((unsigned long)((u64)CONFIG_XEN_MAX_DOMAIN_MEMORY * 1024 * 1024 * 1024 / PAGE_SIZE)) u64 79 include/asm-x86/xen/page.h return XMADDR(PFN_PHYS((u64)pfn_to_mfn(PFN_DOWN(phys.paddr))) | offset); u64 85 include/asm-x86/xen/page.h return XPADDR(PFN_PHYS((u64)mfn_to_pfn(PFN_DOWN(machine.maddr))) | offset); u64 27 include/asm-x86/xsave.h extern u64 pcntxt_mask; u64 27 include/crypto/aead.h u64 seq; u64 53 include/crypto/b128ops.h u64 a, b; u64 27 include/crypto/skcipher.h u64 seq; u64 185 include/linux/acct.h u64 tmp = (u64)x * TICK_NSEC; u64 191 include/linux/acct.h static inline u64 nsec_to_AHZ(u64 x) u64 18 include/linux/amba/bus.h u64 dma_mask; u64 520 include/linux/ata.h ( ((u64) (id)[(n) + 3] << 48) | \ u64 521 include/linux/ata.h ((u64) (id)[(n) + 2] << 32) | \ u64 522 include/linux/ata.h ((u64) (id)[(n) + 1] << 16) | \ u64 523 include/linux/ata.h ((u64) (id)[(n) + 0]) ) u64 850 include/linux/ata.h *(u64 *)&id[ATA_ID_LBA_CAPACITY_2] = u64 873 include/linux/ata.h return ((block + n_block) < ((u64)1 << 28)) && (n_block <= 256); u64 879 include/linux/ata.h return ((block + n_block - 1) < ((u64)1 << 48)) && (n_block <= 65536); u64 645 include/linux/blkdev.h #define BLK_BOUNCE_HIGH ((u64)blk_max_low_pfn << PAGE_SHIFT) u64 148 include/linux/blktrace_api.h u64 start_lba; u64 149 include/linux/blktrace_api.h u64 end_lba; u64 187 include/linux/cgroup.h int (*fill)(struct cgroup_map_cb *cb, const char *key, u64 value); u64 225 include/linux/cgroup.h u64 (*read_u64)(struct cgroup *cgrp, struct cftype *cft); u64 254 include/linux/cgroup.h int (*write_u64)(struct cgroup *cgrp, struct cftype *cft, u64 val); u64 21 include/linux/clocksource.h typedef u64 cycle_t; u64 79 include/linux/clocksource.h u64 xtime_interval; u64 86 include/linux/clocksource.h u64 xtime_nsec; u64 127 include/linux/clocksource.h u64 tmp = ((u64)1000000) << shift_constant; u64 153 include/linux/clocksource.h u64 tmp = ((u64)1000000000) << shift_constant; u64 183 include/linux/clocksource.h u64 ret = (u64)cycles; u64 202 include/linux/clocksource.h u64 tmp; u64 214 include/linux/clocksource.h c->xtime_interval = (u64)c->cycle_interval * c->mult; u64 29 include/linux/cnt32_to_63.h u64 val; u64 144 include/linux/cpufreq.h u64 result = ((u64) old) * ((u64) mult); u64 150 include/linux/cpufreq.h unsigned long result = old * ((u64) mult); u64 329 include/linux/dccp.h static inline u64 dccp_hdr_ack_seq(const struct sk_buff *skb) u64 332 include/linux/dccp.h return ((u64)ntohs(dhack->dccph_ack_nr_high) << 32) + ntohl(dhack->dccph_ack_nr_low); u64 438 include/linux/dccp.h u64 dccpor_ndp:48; u64 392 include/linux/device.h u64 *dma_mask; /* dma mask (if dma'able device) */ u64 393 include/linux/device.h u64 coherent_dma_mask;/* Like dma_mask, but for u64 5 include/linux/dirent.h u64 d_ino; u64 66 include/linux/dma-mapping.h static inline u64 dma_get_mask(struct device *dev) u64 73 include/linux/dma-mapping.h extern u64 dma_get_required_mask(struct device *dev); u64 34 include/linux/dmar.h u64 reg_base_addr; /* register base address*/ u64 138 include/linux/dmar.h u64 base_address; /* reserved base address*/ u64 139 include/linux/dmar.h u64 end_address; /* reserved end address */ u64 54 include/linux/efi.h u64 signature; u64 83 include/linux/efi.h #define EFI_MEMORY_UC ((u64)0x0000000000000001ULL) /* uncached */ u64 84 include/linux/efi.h #define EFI_MEMORY_WC ((u64)0x0000000000000002ULL) /* write-coalescing */ u64 85 include/linux/efi.h #define EFI_MEMORY_WT ((u64)0x0000000000000004ULL) /* write-through */ u64 86 include/linux/efi.h #define EFI_MEMORY_WB ((u64)0x0000000000000008ULL) /* write-back */ u64 87 include/linux/efi.h #define EFI_MEMORY_WP ((u64)0x0000000000001000ULL) /* write-protect */ u64 88 include/linux/efi.h #define EFI_MEMORY_RP ((u64)0x0000000000002000ULL) /* read-protect */ u64 89 include/linux/efi.h #define EFI_MEMORY_XP ((u64)0x0000000000004000ULL) /* execute-protect */ u64 90 include/linux/efi.h #define EFI_MEMORY_RUNTIME ((u64)0x8000000000000000ULL) /* range requires runtime mapping */ u64 98 include/linux/efi.h u64 phys_addr; u64 99 include/linux/efi.h u64 virt_addr; u64 100 include/linux/efi.h u64 num_pages; u64 101 include/linux/efi.h u64 attribute; u64 143 include/linux/efi.h #define EFI_RUNTIME_SERVICES_SIGNATURE ((u64)0x5652453544e5552ULL) u64 216 include/linux/efi.h #define EFI_SYSTEM_TABLE_SIGNATURE ((u64)0x5453595320494249ULL) u64 293 include/linux/efi.h extern u64 efi_get_iobase (void); u64 295 include/linux/efi.h extern u64 efi_mem_attributes (unsigned long phys_addr); u64 296 include/linux/efi.h extern u64 efi_mem_attribute (unsigned long phys_addr, unsigned long size); u64 40 include/linux/eisa.h u64 dma_mask; u64 98 include/linux/eisa.h u64 dma_mask; /* from bridge device */ u64 403 include/linux/ethtool.h void (*self_test)(struct net_device *, struct ethtool_test *, u64 *); u64 406 include/linux/ethtool.h void (*get_ethtool_stats)(struct net_device *, struct ethtool_stats *, u64 *); u64 907 include/linux/fb.h #define fb_readq(addr) (*(volatile u64 *) (addr)) u64 911 include/linux/fb.h #define fb_writeq(b,addr) (*(volatile u64 *) (addr) = (b)) u64 635 include/linux/fs.h u64 i_version; u64 834 include/linux/fs.h u64 f_version; u64 906 include/linux/fs.h #define MAX_LFS_FILESIZE (((u64)PAGE_CACHE_SIZE << (BITS_PER_LONG-1))-1) u64 1350 include/linux/fs.h int (*fiemap)(struct inode *, struct fiemap_extent_info *, u64 start, u64 1351 include/linux/fs.h u64 len); u64 34 include/linux/hash.h static inline u64 hash_64(u64 val, unsigned int bits) u64 36 include/linux/hash.h u64 hash = val; u64 39 include/linux/hash.h u64 n = hash; u64 13 include/linux/hpet.h u64 hpet_cap; /* capabilities */ u64 14 include/linux/hpet.h u64 res0; /* reserved */ u64 15 include/linux/hpet.h u64 hpet_config; /* configuration */ u64 16 include/linux/hpet.h u64 res1; /* reserved */ u64 17 include/linux/hpet.h u64 hpet_isr; /* interrupt status reg */ u64 18 include/linux/hpet.h u64 res2[25]; /* reserved */ u64 20 include/linux/hpet.h u64 _hpet_mc64; u64 24 include/linux/hpet.h u64 res3; /* reserved */ u64 26 include/linux/hpet.h u64 hpet_config; /* configuration/cap */ u64 28 include/linux/hpet.h u64 _hpet_hc64; u64 32 include/linux/hpet.h u64 hpet_fsb[2]; /* FSB route */ u64 339 include/linux/hrtimer.h extern u64 u64 343 include/linux/hrtimer.h static inline u64 hrtimer_forward_now(struct hrtimer *timer, u64 367 include/linux/hrtimer.h extern u64 ktime_divns(const ktime_t kt, s64 div); u64 369 include/linux/hrtimer.h # define ktime_divns(kt, div) (u64)((kt).tv64 / (div)) u64 635 include/linux/i2o.h return (u32) (u64) ptr; u64 640 include/linux/i2o.h return (u32) ((u64) ptr >> 32); u64 645 include/linux/i2o.h return (u32) (u64) dma_addr; u64 650 include/linux/i2o.h return (u32) ((u64) dma_addr >> 32); u64 639 include/linux/ide.h u64 probed_capacity;/* initial reported media capacity (ide-cd only currently) */ u64 641 include/linux/ide.h u64 capacity64; /* total number of sectors */ u64 1534 include/linux/ide.h u64 ide_get_lba_addr(struct ide_taskfile *, int); u64 109 include/linux/inet_lro.h void **tcpudp_hdr, u64 *hdr_flags, void *priv); u64 122 include/linux/inet_lro.h void **ip_hdr, void **tcpudp_hdr, u64 *hdr_flags, u64 29 include/linux/iocontext.h u64 seek_total; u64 50 include/linux/iocontext.h u64 seek_total; u64 81 include/linux/jiffies.h extern u64 __jiffy_data jiffies_64; u64 85 include/linux/jiffies.h u64 get_jiffies_64(void); u64 87 include/linux/jiffies.h static inline u64 get_jiffies_64(void) u64 89 include/linux/jiffies.h return (u64)jiffies; u64 253 include/linux/jiffies.h #define SEC_CONVERSION ((unsigned long)((((u64)NSEC_PER_SEC << SEC_JIFFIE_SC) +\ u64 254 include/linux/jiffies.h TICK_NSEC -1) / (u64)TICK_NSEC)) u64 256 include/linux/jiffies.h #define NSEC_CONVERSION ((unsigned long)((((u64)1 << NSEC_JIFFIE_SC) +\ u64 257 include/linux/jiffies.h TICK_NSEC -1) / (u64)TICK_NSEC)) u64 259 include/linux/jiffies.h ((unsigned long)((((u64)NSEC_PER_USEC << USEC_JIFFIE_SC) +\ u64 260 include/linux/jiffies.h TICK_NSEC -1) / (u64)TICK_NSEC)) u64 268 include/linux/jiffies.h #define USEC_ROUND (u64)(((u64)1 << USEC_JIFFIE_SC) - 1) u64 276 include/linux/jiffies.h (long)((u64)((u64)MAX_JIFFY_OFFSET * TICK_NSEC) / NSEC_PER_SEC) u64 298 include/linux/jiffies.h extern u64 jiffies_64_to_clock_t(u64 x); u64 299 include/linux/jiffies.h extern u64 nsec_to_clock_t(u64 x); u64 60 include/linux/kdev_t.h static inline u64 huge_encode_dev(dev_t dev) u64 37 include/linux/kobject.h extern u64 uevent_seqnum; u64 61 include/linux/ktime.h #define KTIME_MAX ((s64)~((u64)1 << 63)) u64 34 include/linux/kvm_types.h typedef u64 gpa_t; u64 38 include/linux/kvm_types.h typedef u64 hpa_t; u64 552 include/linux/libata.h u64 timestamp; u64 571 include/linux/libata.h u64 n_sectors; /* size of device, if ATA */ u64 158 include/linux/lm_interface.h u64 ln_number; u64 22 include/linux/lmb.h u64 base; u64 23 include/linux/lmb.h u64 size; u64 28 include/linux/lmb.h u64 size; u64 34 include/linux/lmb.h u64 rmo_size; u64 46 include/linux/lmb.h extern u64 __init lmb_alloc_nid(u64 size, u64 align, int nid, u64 48 include/linux/lmb.h extern u64 __init lmb_alloc(u64 size, u64 align); u64 49 include/linux/lmb.h extern u64 __init lmb_alloc_base(u64 size, u64 51 include/linux/lmb.h extern u64 __init __lmb_alloc_base(u64 size, u64 53 include/linux/lmb.h extern u64 __init lmb_phys_mem_size(void); u64 54 include/linux/lmb.h extern u64 __init lmb_end_of_DRAM(void); u64 61 include/linux/lmb.h static inline u64 u64 66 include/linux/lmb.h static inline u64 u64 71 include/linux/lmb.h static inline u64 u64 76 include/linux/lmb.h static inline u64 u64 15 include/linux/lockd/nlm.h # define NLM4_OFFSET_MAX ((s64) ((~(u64)0) >> 1)) u64 190 include/linux/lockdep.h u64 chain_key; u64 216 include/linux/lockdep.h u64 prev_chain_key; u64 221 include/linux/lockdep.h u64 waittime_stamp; u64 222 include/linux/lockdep.h u64 holdtime_stamp; u64 15 include/linux/math64.h static inline u64 div_u64_rem(u64 dividend, u32 divisor, u32 *remainder) u64 33 include/linux/math64.h static inline u64 div64_u64(u64 dividend, u64 divisor) u64 41 include/linux/math64.h static inline u64 div_u64_rem(u64 dividend, u32 divisor, u32 *remainder) u64 53 include/linux/math64.h extern u64 div64_u64(u64 dividend, u64 divisor); u64 66 include/linux/math64.h static inline u64 div_u64(u64 dividend, u32 divisor) u64 45 include/linux/mca.h u64 dma_mask; u64 86 include/linux/mca.h u64 default_dma_mask; u64 148 include/linux/mlx4/device.h static inline u64 mlx4_fw_ver(u64 major, u64 minor, u64 subminor) u64 154 include/linux/mlx4/device.h u64 fw_ver; u64 259 include/linux/mlx4/device.h u64 iova; u64 260 include/linux/mlx4/device.h u64 size; u64 353 include/linux/mlx4/device.h u64 guid0; u64 354 include/linux/mlx4/device.h u64 node_guid; u64 355 include/linux/mlx4/device.h u64 si_guid; u64 379 include/linux/mlx4/device.h u64 mlx4_mtt_addr(struct mlx4_dev *dev, struct mlx4_mtt *mtt); u64 58 include/linux/mlx4/doorbell.h __raw_writeq(*(u64 *) val, dest); u64 68 include/linux/nbd.h u64 bytesize; u64 96 include/linux/netfilter/x_tables.h u_int64_t u64; u64 347 include/linux/nfs_fs.h extern u64 nfs_compat_user_ino64(u64 fileid); u64 557 include/linux/nfs_fs.h if (sizeof(ino_t) < sizeof(u64)) u64 558 include/linux/nfs_fs.h ino ^= fileid >> (sizeof(u64)-sizeof(ino_t)) * 8; u64 38 include/linux/nfs_fs_sb.h u64 cl_clientid; /* constant */ u64 110 include/linux/nfs_xdr.h u64 before; u64 111 include/linux/nfs_xdr.h u64 after; u64 544 include/linux/nfs_xdr.h typedef u64 clientid4; u64 639 include/linux/nfs_xdr.h u64 cookie; u64 815 include/linux/nfs_xdr.h u64, struct page *, unsigned int, int); u64 225 include/linux/nfsd/nfsfh.h *(u64*)fsidv = (u64)ino; u64 74 include/linux/nfsd/xdr4.h u64 co_offset; /* request */ u64 121 include/linux/nfsd/xdr4.h u64 ld_start; u64 122 include/linux/nfsd/xdr4.h u64 ld_length; u64 130 include/linux/nfsd/xdr4.h u64 lk_offset; u64 131 include/linux/nfsd/xdr4.h u64 lk_length; u64 175 include/linux/nfsd/xdr4.h u64 lt_offset; u64 176 include/linux/nfsd/xdr4.h u64 lt_length; u64 186 include/linux/nfsd/xdr4.h u64 lu_offset; u64 187 include/linux/nfsd/xdr4.h u64 lu_length; u64 248 include/linux/nfsd/xdr4.h u64 rd_offset; /* request */ u64 258 include/linux/nfsd/xdr4.h u64 rd_cookie; /* request */ u64 337 include/linux/nfsd/xdr4.h u64 wr_offset; /* request */ u64 28 include/linux/of.h #define OF_BAD_ADDR ((u64)-1) u64 171 include/linux/pci.h u64 dma_mask; /* Mask of the bits of bus address this u64 243 include/linux/pnp.h u64 dma_mask; u64 115 include/linux/poll.h #define MAX_INT64_SECONDS (((s64)(~((u64)0)>>1)/HZ)-1) u64 62 include/linux/random.h extern u64 secure_dccp_sequence_number(__be32 saddr, __be32 daddr, u64 30 include/linux/reciprocal_div.h return (u32)(((u64)A * R) >> 32); u64 61 include/linux/res_counter.h u64 res_counter_read_u64(struct res_counter *counter, int member); u64 110 include/linux/rio.h u64 dma_mask; u64 657 include/linux/rtnetlink.h ({ u64 _tmp = (value); \ u64 658 include/linux/rtnetlink.h RTA_PUT(skb, attrtype, sizeof(u64), &_tmp); }) u64 714 include/linux/rtnetlink.h ({ u64 _tmp; \ u64 715 include/linux/rtnetlink.h if (!rta || RTA_PAYLOAD(rta) < sizeof(u64)) \ u64 666 include/linux/sched.h u64 blkio_delay; /* wait for sync block io completion */ u64 667 include/linux/sched.h u64 swapin_delay; /* wait for swapin block io completion */ u64 674 include/linux/sched.h u64 freepages_delay; /* wait for memory reclaim */ u64 794 include/linux/sched.h u64 last_update; u64 963 include/linux/sched.h u64 exec_start; u64 964 include/linux/sched.h u64 sum_exec_runtime; u64 965 include/linux/sched.h u64 vruntime; u64 966 include/linux/sched.h u64 prev_sum_exec_runtime; u64 968 include/linux/sched.h u64 last_wakeup; u64 969 include/linux/sched.h u64 avg_overlap; u64 972 include/linux/sched.h u64 wait_start; u64 973 include/linux/sched.h u64 wait_max; u64 974 include/linux/sched.h u64 wait_count; u64 975 include/linux/sched.h u64 wait_sum; u64 977 include/linux/sched.h u64 sleep_start; u64 978 include/linux/sched.h u64 sleep_max; u64 981 include/linux/sched.h u64 block_start; u64 982 include/linux/sched.h u64 block_max; u64 983 include/linux/sched.h u64 exec_max; u64 984 include/linux/sched.h u64 slice_max; u64 986 include/linux/sched.h u64 nr_migrations; u64 987 include/linux/sched.h u64 nr_migrations_cold; u64 988 include/linux/sched.h u64 nr_failed_migrations_affine; u64 989 include/linux/sched.h u64 nr_failed_migrations_running; u64 990 include/linux/sched.h u64 nr_failed_migrations_hot; u64 991 include/linux/sched.h u64 nr_forced_migrations; u64 992 include/linux/sched.h u64 nr_forced2_migrations; u64 994 include/linux/sched.h u64 nr_wakeups; u64 995 include/linux/sched.h u64 nr_wakeups_sync; u64 996 include/linux/sched.h u64 nr_wakeups_migrate; u64 997 include/linux/sched.h u64 nr_wakeups_local; u64 998 include/linux/sched.h u64 nr_wakeups_remote; u64 999 include/linux/sched.h u64 nr_wakeups_affine; u64 1000 include/linux/sched.h u64 nr_wakeups_affine_attempts; u64 1001 include/linux/sched.h u64 nr_wakeups_passive; u64 1002 include/linux/sched.h u64 nr_wakeups_idle; u64 1239 include/linux/sched.h u64 curr_chain_key; u64 1262 include/linux/sched.h u64 acct_rss_mem1; /* accumulated rss usage */ u64 1263 include/linux/sched.h u64 acct_vm_mem1; /* accumulated virtual memory usage */ u64 1563 include/linux/sched.h extern u64 sched_clock_cpu(int cpu); u64 22 include/linux/seq_file.h u64 version; u64 63 include/linux/stat.h u64 ino; u64 11 include/linux/statfs.h u64 f_blocks; u64 12 include/linux/statfs.h u64 f_bfree; u64 13 include/linux/statfs.h u64 f_bavail; u64 14 include/linux/statfs.h u64 f_files; u64 15 include/linux/statfs.h u64 f_ffree; u64 14 include/linux/task_io_accounting.h u64 rchar; u64 16 include/linux/task_io_accounting.h u64 wchar; u64 18 include/linux/task_io_accounting.h u64 syscr; u64 20 include/linux/task_io_accounting.h u64 syscw; u64 28 include/linux/task_io_accounting.h u64 read_bytes; u64 34 include/linux/task_io_accounting.h u64 write_bytes; u64 43 include/linux/task_io_accounting.h u64 cancelled_write_bytes; u64 30 include/linux/thread_info.h u64 time; u64 39 include/linux/thread_info.h u64 expires; u64 117 include/linux/tick.h extern u64 get_cpu_idle_time_us(int cpu, u64 *last_update_time); u64 129 include/linux/tick.h static inline u64 get_cpu_idle_time_us(int cpu, u64 *unused) { return -1; } u64 244 include/linux/timex.h extern u64 tick_length; u64 140 include/linux/types.h typedef u64 sector_t; u64 149 include/linux/types.h typedef u64 blkcnt_t; u64 195 include/linux/types.h typedef u64 resource_size_t; u64 17 include/linux/unaligned/access_ok.h static inline u64 get_unaligned_le64(const void *p) u64 32 include/linux/unaligned/access_ok.h static inline u64 get_unaligned_be64(const void *p) u64 16 include/linux/unaligned/be_byteshift.h static inline u64 __get_unaligned_be64(const u8 *p) u64 18 include/linux/unaligned/be_byteshift.h return (u64)__get_unaligned_be32(p) << 32 | u64 50 include/linux/unaligned/be_byteshift.h static inline u64 get_unaligned_be64(const void *p) u64 16 include/linux/unaligned/be_memmove.h static inline u64 get_unaligned_be64(const void *p) u64 16 include/linux/unaligned/be_struct.h static inline u64 get_unaligned_be64(const void *p) u64 39 include/linux/unaligned/generic.h put_unaligned_le64((__force u64)(val), __gu_p); \ u64 60 include/linux/unaligned/generic.h put_unaligned_be64((__force u64)(val), __gu_p); \ u64 16 include/linux/unaligned/le_byteshift.h static inline u64 __get_unaligned_le64(const u8 *p) u64 18 include/linux/unaligned/le_byteshift.h return (u64)__get_unaligned_le32(p + 4) << 32 | u64 50 include/linux/unaligned/le_byteshift.h static inline u64 get_unaligned_le64(const void *p) u64 16 include/linux/unaligned/le_memmove.h static inline u64 get_unaligned_le64(const void *p) u64 16 include/linux/unaligned/le_struct.h static inline u64 get_unaligned_le64(const void *p) u64 23 include/linux/unaligned/memmove.h static inline u64 __get_unaligned_memmove64(const void *p) u64 25 include/linux/unaligned/memmove.h u64 tmp; u64 8 include/linux/unaligned/packed_struct.h struct __una_u64 { u64 x __attribute__((packed)); }; u64 22 include/linux/unaligned/packed_struct.h static inline u64 __get_unaligned_cpu64(const void *p) u64 303 include/net/9p/9p.h u64 path; u64 336 include/net/9p/9p.h u64 length; u64 363 include/net/9p/9p.h u64 length; u64 457 include/net/9p/9p.h u64 offset; u64 468 include/net/9p/9p.h u64 offset; u64 404 include/net/ieee80211.h u64 tsf; u64 243 include/net/ip_vs.h u64 last_inbytes; u64 244 include/net/ip_vs.h u64 last_outbytes; u64 207 include/net/mac80211.h u64 timestamp; u64 208 include/net/mac80211.h u64 basic_rates; u64 357 include/net/mac80211.h u64 ampdu_ack_map; u64 429 include/net/mac80211.h u64 mactime; u64 692 include/net/mac80211.h u64 supp_rates[IEEE80211_NUM_BANDS]; u64 1256 include/net/mac80211.h u64 (*get_tsf)(struct ieee80211_hw *hw); u64 156 include/net/netlabel.h #define NETLBL_CATMAP_MAPTYPE u64 u64 791 include/net/netlink.h return nla_put(skb, attrtype, sizeof(u64), &value); u64 825 include/net/netlink.h u64 tmp = jiffies_to_msecs(jiffies); u64 826 include/net/netlink.h return nla_put(skb, attrtype, sizeof(u64), &tmp); u64 860 include/net/netlink.h NLA_PUT_TYPE(skb, u64, attrtype, value) u64 932 include/net/netlink.h static inline u64 nla_get_u64(struct nlattr *nla) u64 934 include/net/netlink.h u64 tmp; u64 958 include/net/netlink.h u64 msecs = nla_get_u64(nla); u64 41 include/net/pkt_sched.h typedef u64 psched_time_t; u64 215 include/net/red.h us_idle = (p->qavg * (u64)us_idle) >> p->Scell_log; u64 562 include/net/xfrm.h u64 output; u64 125 include/pcmcia/ds.h u64 dma_mask; u64 74 include/rdma/ib_fmr_pool.h u64 io_virtual_address; u64 76 include/rdma/ib_fmr_pool.h u64 page_list[0]; u64 152 include/rdma/ib_mad.h typedef u64 __bitwise ib_sa_comp_mask; u64 241 include/rdma/ib_sa.h u64 id; u64 251 include/rdma/ib_sa.h u64 data64[2]; u64 117 include/rdma/ib_verbs.h u64 fw_ver; u64 119 include/rdma/ib_verbs.h u64 max_mr_size; u64 120 include/rdma/ib_verbs.h u64 page_size_cap; u64 235 include/rdma/ib_verbs.h u64 ipInReceives; u64 236 include/rdma/ib_verbs.h u64 ipInHdrErrors; u64 237 include/rdma/ib_verbs.h u64 ipInTooBigErrors; u64 238 include/rdma/ib_verbs.h u64 ipInNoRoutes; u64 239 include/rdma/ib_verbs.h u64 ipInAddrErrors; u64 240 include/rdma/ib_verbs.h u64 ipInUnknownProtos; u64 241 include/rdma/ib_verbs.h u64 ipInTruncatedPkts; u64 242 include/rdma/ib_verbs.h u64 ipInDiscards; u64 243 include/rdma/ib_verbs.h u64 ipInDelivers; u64 244 include/rdma/ib_verbs.h u64 ipOutForwDatagrams; u64 245 include/rdma/ib_verbs.h u64 ipOutRequests; u64 246 include/rdma/ib_verbs.h u64 ipOutDiscards; u64 247 include/rdma/ib_verbs.h u64 ipOutNoRoutes; u64 248 include/rdma/ib_verbs.h u64 ipReasmTimeout; u64 249 include/rdma/ib_verbs.h u64 ipReasmReqds; u64 250 include/rdma/ib_verbs.h u64 ipReasmOKs; u64 251 include/rdma/ib_verbs.h u64 ipReasmFails; u64 252 include/rdma/ib_verbs.h u64 ipFragOKs; u64 253 include/rdma/ib_verbs.h u64 ipFragFails; u64 254 include/rdma/ib_verbs.h u64 ipFragCreates; u64 255 include/rdma/ib_verbs.h u64 ipInMcastPkts; u64 256 include/rdma/ib_verbs.h u64 ipOutMcastPkts; u64 257 include/rdma/ib_verbs.h u64 ipInBcastPkts; u64 258 include/rdma/ib_verbs.h u64 ipOutBcastPkts; u64 260 include/rdma/ib_verbs.h u64 tcpRtoAlgorithm; u64 261 include/rdma/ib_verbs.h u64 tcpRtoMin; u64 262 include/rdma/ib_verbs.h u64 tcpRtoMax; u64 263 include/rdma/ib_verbs.h u64 tcpMaxConn; u64 264 include/rdma/ib_verbs.h u64 tcpActiveOpens; u64 265 include/rdma/ib_verbs.h u64 tcpPassiveOpens; u64 266 include/rdma/ib_verbs.h u64 tcpAttemptFails; u64 267 include/rdma/ib_verbs.h u64 tcpEstabResets; u64 268 include/rdma/ib_verbs.h u64 tcpCurrEstab; u64 269 include/rdma/ib_verbs.h u64 tcpInSegs; u64 270 include/rdma/ib_verbs.h u64 tcpOutSegs; u64 271 include/rdma/ib_verbs.h u64 tcpRetransSegs; u64 272 include/rdma/ib_verbs.h u64 tcpInErrs; u64 273 include/rdma/ib_verbs.h u64 tcpOutRsts; u64 309 include/rdma/ib_verbs.h u64 sys_image_guid; u64 485 include/rdma/ib_verbs.h u64 wr_id; u64 703 include/rdma/ib_verbs.h u64 addr; u64 710 include/rdma/ib_verbs.h u64 *page_list; u64 716 include/rdma/ib_verbs.h u64 wr_id; u64 727 include/rdma/ib_verbs.h u64 remote_addr; u64 731 include/rdma/ib_verbs.h u64 remote_addr; u64 732 include/rdma/ib_verbs.h u64 compare_add; u64 733 include/rdma/ib_verbs.h u64 swap; u64 747 include/rdma/ib_verbs.h u64 iova_start; u64 760 include/rdma/ib_verbs.h u64 wr_id; u64 774 include/rdma/ib_verbs.h u64 addr; u64 775 include/rdma/ib_verbs.h u64 size; u64 780 include/rdma/ib_verbs.h u64 device_virt_addr; u64 781 include/rdma/ib_verbs.h u64 size; u64 795 include/rdma/ib_verbs.h u64 wr_id; u64 796 include/rdma/ib_verbs.h u64 addr; u64 821 include/rdma/ib_verbs.h u64 user_handle; /* handle given to us by userspace */ u64 936 include/rdma/ib_verbs.h u64 dma_addr); u64 937 include/rdma/ib_verbs.h u64 (*map_single)(struct ib_device *dev, u64 941 include/rdma/ib_verbs.h u64 addr, size_t size, u64 943 include/rdma/ib_verbs.h u64 (*map_page)(struct ib_device *dev, u64 948 include/rdma/ib_verbs.h u64 addr, size_t size, u64 956 include/rdma/ib_verbs.h u64 (*dma_address)(struct ib_device *dev, u64 961 include/rdma/ib_verbs.h u64 dma_handle, u64 965 include/rdma/ib_verbs.h u64 dma_handle, u64 970 include/rdma/ib_verbs.h u64 *dma_handle, u64 974 include/rdma/ib_verbs.h u64 dma_handle); u64 1086 include/rdma/ib_verbs.h u64 *iova_start); u64 1088 include/rdma/ib_verbs.h u64 start, u64 length, u64 1089 include/rdma/ib_verbs.h u64 virt_addr, u64 1106 include/rdma/ib_verbs.h u64 *iova_start); u64 1116 include/rdma/ib_verbs.h u64 *page_list, int list_len, u64 1117 include/rdma/ib_verbs.h u64 iova); u64 1147 include/rdma/ib_verbs.h u64 uverbs_cmd_mask; u64 1603 include/rdma/ib_verbs.h static inline u64 ib_dma_map_single(struct ib_device *dev, u64 1629 include/rdma/ib_verbs.h static inline u64 ib_dma_map_single_attrs(struct ib_device *dev, u64 1655 include/rdma/ib_verbs.h static inline u64 ib_dma_map_page(struct ib_device *dev, u64 1736 include/rdma/ib_verbs.h static inline u64 ib_sg_dma_address(struct ib_device *dev, u64 723 include/scsi/scsi_host.h extern u64 scsi_calculate_bounce_limit(struct Scsi_Host *); u64 108 include/scsi/scsi_netlink.h #define SCSI_NL_VID_TYPE_MASK ((u64)0xFF << SCSI_NL_VID_TYPE_SHIFT) u64 109 include/scsi/scsi_netlink.h #define SCSI_NL_VID_TYPE_PCI ((u64)0x01 << SCSI_NL_VID_TYPE_SHIFT) u64 74 include/scsi/scsi_transport.h int (* it_nexus_response)(struct Scsi_Host *, u64, int); u64 80 include/scsi/scsi_transport.h int (* tsk_mgmt_response)(struct Scsi_Host *, u64, u64, int); u64 183 include/scsi/scsi_transport_fc.h u64 node_name; u64 184 include/scsi/scsi_transport_fc.h u64 port_name; u64 228 include/scsi/scsi_transport_fc.h u64 node_name; u64 229 include/scsi/scsi_transport_fc.h u64 port_name; u64 284 include/scsi/scsi_transport_fc.h u64 node_name; u64 285 include/scsi/scsi_transport_fc.h u64 port_name; u64 333 include/scsi/scsi_transport_fc.h u64 node_name; u64 334 include/scsi/scsi_transport_fc.h u64 port_name; u64 382 include/scsi/scsi_transport_fc.h u64 node_name; u64 383 include/scsi/scsi_transport_fc.h u64 port_name; u64 405 include/scsi/scsi_transport_fc.h u64 seconds_since_last_reset; u64 406 include/scsi/scsi_transport_fc.h u64 tx_frames; u64 407 include/scsi/scsi_transport_fc.h u64 tx_words; u64 408 include/scsi/scsi_transport_fc.h u64 rx_frames; u64 409 include/scsi/scsi_transport_fc.h u64 rx_words; u64 410 include/scsi/scsi_transport_fc.h u64 lip_count; u64 411 include/scsi/scsi_transport_fc.h u64 nos_count; u64 412 include/scsi/scsi_transport_fc.h u64 error_frames; u64 413 include/scsi/scsi_transport_fc.h u64 dumped_frames; u64 414 include/scsi/scsi_transport_fc.h u64 link_failure_count; u64 415 include/scsi/scsi_transport_fc.h u64 loss_of_sync_count; u64 416 include/scsi/scsi_transport_fc.h u64 loss_of_signal_count; u64 417 include/scsi/scsi_transport_fc.h u64 prim_seq_protocol_err_count; u64 418 include/scsi/scsi_transport_fc.h u64 invalid_tx_word_count; u64 419 include/scsi/scsi_transport_fc.h u64 invalid_crc_count; u64 422 include/scsi/scsi_transport_fc.h u64 fcp_input_requests; u64 423 include/scsi/scsi_transport_fc.h u64 fcp_output_requests; u64 424 include/scsi/scsi_transport_fc.h u64 fcp_control_requests; u64 425 include/scsi/scsi_transport_fc.h u64 fcp_input_megabytes; u64 426 include/scsi/scsi_transport_fc.h u64 fcp_output_megabytes; u64 478 include/scsi/scsi_transport_fc.h u64 node_name; u64 479 include/scsi/scsi_transport_fc.h u64 port_name; u64 480 include/scsi/scsi_transport_fc.h u64 permanent_port_name; u64 494 include/scsi/scsi_transport_fc.h u64 fabric_name; u64 612 include/scsi/scsi_transport_fc.h int (* tsk_mgmt_response)(struct Scsi_Host *, u64, u64, int); u64 613 include/scsi/scsi_transport_fc.h int (* it_nexus_response)(struct Scsi_Host *, u64, int); u64 695 include/scsi/scsi_transport_fc.h static inline u64 wwn_to_u64(u8 *wwn) u64 697 include/scsi/scsi_transport_fc.h return (u64)wwn[0] << 56 | (u64)wwn[1] << 48 | u64 698 include/scsi/scsi_transport_fc.h (u64)wwn[2] << 40 | (u64)wwn[3] << 32 | u64 699 include/scsi/scsi_transport_fc.h (u64)wwn[4] << 24 | (u64)wwn[5] << 16 | u64 700 include/scsi/scsi_transport_fc.h (u64)wwn[6] << 8 | (u64)wwn[7]; u64 50 include/scsi/scsi_transport_sas.h u64 sas_address; u64 166 include/scsi/scsi_transport_sas.h int (*get_enclosure_identifier)(struct sas_rphy *, u64 *); u64 25 include/scsi/scsi_transport_srp.h int (* tsk_mgmt_response)(struct Scsi_Host *, u64, u64, int); u64 26 include/scsi/scsi_transport_srp.h int (* it_nexus_response)(struct Scsi_Host *, u64, int); u64 121 include/scsi/srp.h u64 tag; u64 140 include/scsi/srp.h u64 tag; u64 152 include/scsi/srp.h u64 tag; u64 161 include/scsi/srp.h u64 tag; u64 169 include/scsi/srp.h u64 tag; u64 180 include/scsi/srp.h u64 tag; u64 186 include/scsi/srp.h u64 task_tag; u64 201 include/scsi/srp.h u64 tag; u64 231 include/scsi/srp.h u64 tag; u64 46 include/sound/pcm.h u64 formats; /* SNDRV_PCM_FMTBIT_* */ u64 324 include/sound/soc.h u64 formats; /* SNDRV_PCM_FMTBIT_* */ u64 54 include/sound/vx_core.h typedef u64 pcx_time_t; u64 85 include/sound/vx_core.h u64 cur_count; /* current sample position (for playback) */ u64 488 kernel/acct.c u64 elapsed; u64 489 kernel/acct.c u64 run_time; u64 511 kernel/acct.c run_time = (u64)uptime.tv_sec*NSEC_PER_SEC + uptime.tv_nsec; u64 512 kernel/acct.c run_time -= (u64)current->group_leader->start_time.tv_sec * NSEC_PER_SEC u64 1386 kernel/cgroup.c u64 val = simple_strtoull(buffer, &end, 0); u64 1464 kernel/cgroup.c u64 val = cft->read_u64(cgrp, cft); u64 2188 kernel/cgroup.c static u64 cgroup_read_notify_on_release(struct cgroup *cgrp, u64 34 kernel/cgroup_debug.c static u64 cgroup_refcount_read(struct cgroup *cont, struct cftype *cft) u64 39 kernel/cgroup_debug.c static u64 taskcount_read(struct cgroup *cont, struct cftype *cft) u64 41 kernel/cgroup_debug.c u64 count; u64 49 kernel/cgroup_debug.c static u64 current_css_set_read(struct cgroup *cont, struct cftype *cft) u64 51 kernel/cgroup_debug.c return (u64)(long)current->cgroups; u64 54 kernel/cgroup_debug.c static u64 current_css_set_refcount_read(struct cgroup *cont, u64 57 kernel/cgroup_debug.c u64 count; u64 65 kernel/cgroup_debug.c static u64 releasable_read(struct cgroup *cgrp, struct cftype *cft) u64 1531 kernel/cpuset.c static u64 cpuset_read_u64(struct cgroup *cont, struct cftype *cft) u64 301 kernel/hrtimer.c u64 ktime_divns(const ktime_t kt, s64 div) u64 303 kernel/hrtimer.c u64 dclc; u64 793 kernel/hrtimer.c u64 hrtimer_forward(struct hrtimer *timer, ktime_t now, ktime_t interval) u64 795 kernel/hrtimer.c u64 orun = 1; u64 1756 kernel/lockdep.c u64 chain_key = 0; u64 2529 kernel/lockdep.c u64 chain_key; u64 3045 kernel/lockdep.c u64 now; u64 277 kernel/panic.c static u64 oops_id; u64 152 kernel/power/process.c u64 elapsed_csecs64; u64 101 kernel/res_counter.c u64 res_counter_read_u64(struct res_counter *counter, int member) u64 270 kernel/resource.c u64 orig_end; u64 272 kernel/resource.c res.start = (u64) start_pfn << PAGE_SHIFT; u64 273 kernel/resource.c res.end = ((u64)(start_pfn + nr_pages) << PAGE_SHIFT) - 1; u64 117 kernel/sched.c #define RUNTIME_INF ((u64)~0ULL) u64 164 kernel/sched.c u64 rt_runtime; u64 386 kernel/sched.c u64 exec_clock; u64 387 kernel/sched.c u64 min_vruntime; u64 388 kernel/sched.c u64 pair_start; u64 457 kernel/sched.c u64 rt_time; u64 458 kernel/sched.c u64 rt_runtime; u64 532 kernel/sched.c u64 nr_switches; u64 557 kernel/sched.c u64 clock; u64 834 kernel/sched.c static inline u64 global_rt_period(void) u64 836 kernel/sched.c return (u64)sysctl_sched_rt_period * NSEC_PER_USEC; u64 839 kernel/sched.c static inline u64 global_rt_runtime(void) u64 844 kernel/sched.c return (u64)sysctl_sched_rt_runtime * NSEC_PER_USEC; u64 1258 kernel/sched.c u64 tmp; u64 1268 kernel/sched.c tmp = (u64)delta_exec * weight; u64 1278 kernel/sched.c return (unsigned long)min(tmp, (u64)(unsigned long)LONG_MAX); u64 1566 kernel/sched.c u64 now = cpu_clock(raw_smp_processor_id()); u64 1569 kernel/sched.c if (elapsed >= (s64)(u64)sysctl_sched_shares_ratelimit) { u64 1826 kernel/sched.c u64 clock_offset; u64 4061 kernel/sched.c u64 ns, delta_exec; u64 4214 kernel/sched.c u64 temp; u64 4219 kernel/sched.c temp = (u64)nsec_to_clock_t(p->se.sum_exec_runtime); u64 8023 kernel/sched.c cfs_rq->min_vruntime = (u64)(-(1LL << 20)); u64 8860 kernel/sched.c u64 rt_period; u64 8861 kernel/sched.c u64 rt_runtime; u64 8869 kernel/sched.c u64 period, runtime; u64 8963 kernel/sched.c u64 rt_runtime, rt_period; u64 8966 kernel/sched.c rt_runtime = (u64)rt_runtime_us * NSEC_PER_USEC; u64 8975 kernel/sched.c u64 rt_runtime_us; u64 8987 kernel/sched.c u64 rt_runtime, rt_period; u64 8989 kernel/sched.c rt_period = (u64)rt_period_us * NSEC_PER_USEC; u64 9000 kernel/sched.c u64 rt_period_us; u64 9009 kernel/sched.c u64 runtime, period; u64 9151 kernel/sched.c static u64 cpu_shares_read_u64(struct cgroup *cgrp, struct cftype *cft) u64 9155 kernel/sched.c return (u64) tg->shares; u64 9177 kernel/sched.c static u64 cpu_rt_period_read_uint(struct cgroup *cgrp, struct cftype *cft) u64 9236 kernel/sched.c u64 *cpuusage; u64 9264 kernel/sched.c ca->cpuusage = alloc_percpu(u64); u64 9284 kernel/sched.c static u64 cpuusage_read(struct cgroup *cgrp, struct cftype *cft) u64 9287 kernel/sched.c u64 totalcpuusage = 0; u64 9291 kernel/sched.c u64 *cpuusage = percpu_ptr(ca->cpuusage, i); u64 9318 kernel/sched.c u64 *cpuusage = percpu_ptr(ca->cpuusage, i); u64 9355 kernel/sched.c u64 *cpuusage = percpu_ptr(ca->cpuusage, task_cpu(tsk)); u64 55 kernel/sched_clock.c u64 tick_raw; u64 56 kernel/sched_clock.c u64 tick_gtod; u64 57 kernel/sched_clock.c u64 clock; u64 74 kernel/sched_clock.c u64 ktime_now = ktime_to_ns(ktime_get()); u64 93 kernel/sched_clock.c static inline u64 wrap_min(u64 x, u64 y) u64 98 kernel/sched_clock.c static inline u64 wrap_max(u64 x, u64 y) u64 109 kernel/sched_clock.c static u64 __update_sched_clock(struct sched_clock_data *scd, u64 now) u64 112 kernel/sched_clock.c u64 clock, min_clock, max_clock; u64 149 kernel/sched_clock.c u64 sched_clock_cpu(int cpu) u64 152 kernel/sched_clock.c u64 now, clock, this_clock, remote_clock; u64 199 kernel/sched_clock.c u64 now, now_gtod; u64 242 kernel/sched_clock.c u64 sched_clock_cpu(int cpu) u64 271 kernel/sched_debug.c u64 now = ktime_to_ns(ktime_get()); u64 390 kernel/sched_debug.c u64 avg_atom, avg_per_cpu; u64 425 kernel/sched_debug.c u64 t0, t1; u64 201 kernel/sched_fair.c static inline u64 max_vruntime(u64 min_vruntime, u64 vruntime) u64 210 kernel/sched_fair.c static inline u64 min_vruntime(u64 min_vruntime, u64 vruntime) u64 372 kernel/sched_fair.c static u64 __sched_period(unsigned long nr_running) u64 374 kernel/sched_fair.c u64 period = sysctl_sched_latency; u64 391 kernel/sched_fair.c static u64 sched_slice(struct cfs_rq *cfs_rq, struct sched_entity *se) u64 401 kernel/sched_fair.c static u64 sched_vslice_add(struct cfs_rq *cfs_rq, struct sched_entity *se) u64 421 kernel/sched_fair.c schedstat_set(curr->exec_max, max((u64)delta_exec, curr->exec_max)); u64 432 kernel/sched_fair.c u64 now = rq_of(cfs_rq)->clock; u64 557 kernel/sched_fair.c u64 delta = rq_of(cfs_rq)->clock - se->sleep_start; u64 572 kernel/sched_fair.c u64 delta = rq_of(cfs_rq)->clock - se->block_start; u64 615 kernel/sched_fair.c u64 vruntime; u64 748 kernel/sched_fair.c u64 pair_slice = rq->clock - cfs_rq->pair_start; u64 831 kernel/sched_fair.c u64 slice = sched_slice(cfs_rq, se); u64 832 kernel/sched_fair.c u64 ran = se->sum_exec_runtime - se->prev_sum_exec_runtime; u64 1396 kernel/sched_fair.c u64 rem_load, moved_load; u64 1404 kernel/sched_fair.c rem_load = (u64)rem_load_move * busiest_weight; u64 66 kernel/sched_rt.c static inline u64 sched_rt_runtime(struct rt_rq *rt_rq) u64 74 kernel/sched_rt.c static inline u64 sched_rt_period(struct rt_rq *rt_rq) u64 166 kernel/sched_rt.c static inline u64 sched_rt_runtime(struct rt_rq *rt_rq) u64 171 kernel/sched_rt.c static inline u64 sched_rt_period(struct rt_rq *rt_rq) u64 242 kernel/sched_rt.c u64 rt_period; u64 270 kernel/sched_rt.c diff = div_u64((u64)diff, weight); u64 445 kernel/sched_rt.c u64 runtime; u64 484 kernel/sched_rt.c u64 runtime = sched_rt_runtime(rt_rq); u64 517 kernel/sched_rt.c u64 delta_exec; u64 521 kernel/time.c return (((u64)sec * SEC_CONVERSION) + u64 522 kernel/time.c (((u64)nsec * NSEC_CONVERSION) >> u64 536 kernel/time.c value->tv_sec = div_u64_rem((u64)jiffies * TICK_NSEC, u64 564 kernel/time.c return (((u64)sec * SEC_CONVERSION) + u64 565 kernel/time.c (((u64)usec * USEC_CONVERSION + USEC_ROUND) >> u64 578 kernel/time.c value->tv_sec = div_u64_rem((u64)jiffies * TICK_NSEC, u64 596 kernel/time.c return div_u64((u64)x * TICK_NSEC, NSEC_PER_SEC / USER_HZ); u64 613 kernel/time.c return div_u64((u64)x * HZ, USER_HZ); u64 618 kernel/time.c u64 jiffies_64_to_clock_t(u64 x) u64 640 kernel/time.c u64 nsec_to_clock_t(u64 x) u64 657 kernel/time.c u64 get_jiffies_64(void) u64 660 kernel/time.c u64 ret; u64 42 kernel/time/clockevents.c u64 clc = ((u64) latch << evt->shift); u64 37 kernel/time/jiffies.c #define NSEC_PER_JIFFY ((u32)((((u64)NSEC_PER_SEC)<<8)/ACTHZ)) u64 27 kernel/time/ntp.c u64 tick_length; u64 28 kernel/time/ntp.c static u64 tick_length_base; u64 33 kernel/time/ntp.c #define MAX_TICKADJ_SCALED (((u64)(MAX_TICKADJ * NSEC_PER_USEC) << \ u64 54 kernel/time/ntp.c u64 second_length = (u64)(tick_usec * NSEC_PER_USEC * USER_HZ) u64 190 kernel/time/tick-sched.c u64 get_cpu_idle_time_us(int cpu, u64 *last_update_time) u64 630 kernel/time/tick-sched.c u64 offset; u64 174 kernel/time/timekeeping.c u64 nsec; u64 468 kernel/time/timekeeping.c if (clock->xtime_nsec >= (u64)NSEC_PER_SEC << clock->shift) { u64 469 kernel/time/timekeeping.c clock->xtime_nsec -= (u64)NSEC_PER_SEC << clock->shift; u64 244 kernel/time/timer_list.c u64 now = ktime_to_ns(ktime_get()); u64 47 kernel/timer.c u64 jiffies_64 __cacheline_aligned_in_smp = INITIAL_JIFFIES; u64 31 kernel/tsacct.c u64 ac_etime; u64 123 kernel/tsacct.c u64 delta; u64 63 lib/div64.c u64 quotient; u64 82 lib/div64.c u64 div64_u64(u64 dividend, u64 divisor) u64 168 lib/find_next_bit.c return (unsigned long) __swab64p((u64 *) x); u64 180 lib/find_next_bit.c return (unsigned long) __swab64((u64) y); u64 63 lib/hexdump.c const u64 *ptr8 = buf; u64 59 lib/iomap_copy.c u64 __iomem *dst = to; u64 60 lib/iomap_copy.c const u64 *src = from; u64 61 lib/iomap_copy.c const u64 *end = src + count; u64 28 lib/kobject_uevent.c u64 uevent_seqnum; u64 99 lib/kobject_uevent.c u64 seq; u64 80 lib/lmb.c u64 base1 = rgn->region[r1].base; u64 81 lib/lmb.c u64 size1 = rgn->region[r1].size; u64 82 lib/lmb.c u64 base2 = rgn->region[r2].base; u64 83 lib/lmb.c u64 size2 = rgn->region[r2].size; u64 145 lib/lmb.c u64 rgnbase = rgn->region[i].base; u64 146 lib/lmb.c u64 rgnsize = rgn->region[i].size; u64 211 lib/lmb.c u64 rgnbegin, rgnend; u64 212 lib/lmb.c u64 end = base + size; u64 271 lib/lmb.c u64 rgnbase = rgn->region[i].base; u64 272 lib/lmb.c u64 rgnsize = rgn->region[i].size; u64 280 lib/lmb.c static u64 lmb_align_down(u64 addr, u64 size) u64 285 lib/lmb.c static u64 lmb_align_up(u64 addr, u64 size) u64 290 lib/lmb.c static u64 __init lmb_alloc_nid_unreserved(u64 start, u64 end, u64 293 lib/lmb.c u64 base, res_base; u64 302 lib/lmb.c base = ~(u64)0; u64 311 lib/lmb.c return ~(u64)0; u64 314 lib/lmb.c static u64 __init lmb_alloc_nid_region(struct lmb_property *mp, u64 318 lib/lmb.c u64 start, end; u64 325 lib/lmb.c u64 this_end; u64 330 lib/lmb.c u64 ret = lmb_alloc_nid_unreserved(start, this_end, u64 332 lib/lmb.c if (ret != ~(u64)0) u64 338 lib/lmb.c return ~(u64)0; u64 341 lib/lmb.c u64 __init lmb_alloc_nid(u64 size, u64 align, int nid, u64 352 lib/lmb.c u64 ret = lmb_alloc_nid_region(&mem->region[i], u64 355 lib/lmb.c if (ret != ~(u64)0) u64 362 lib/lmb.c u64 __init lmb_alloc(u64 size, u64 align) u64 367 lib/lmb.c u64 __init lmb_alloc_base(u64 size, u64 align, u64 max_addr) u64 369 lib/lmb.c u64 alloc; u64 380 lib/lmb.c u64 __init __lmb_alloc_base(u64 size, u64 align, u64 max_addr) u64 383 lib/lmb.c u64 base = 0; u64 384 lib/lmb.c u64 res_base; u64 396 lib/lmb.c u64 lmbbase = lmb.memory.region[i].base; u64 397 lib/lmb.c u64 lmbsize = lmb.memory.region[i].size; u64 427 lib/lmb.c u64 __init lmb_phys_mem_size(void) u64 432 lib/lmb.c u64 __init lmb_end_of_DRAM(void) u64 443 lib/lmb.c u64 limit; u64 488 lib/lmb.c u64 upper = lmb.reserved.region[i].base + u64 503 lib/lmb.c u64 rstart, rend; u64 509 lib/lmb.c u64 start = lmb.memory.region[i].base; u64 510 lib/lmb.c u64 end = start + lmb.memory.region[i].size - 1; u64 109 lib/proportions.c u64 events; u64 6 lib/reciprocal_div.c u64 val = (1LL << 32) + (k - 1); u64 103 lib/spinlock_debug.c u64 i; u64 104 lib/spinlock_debug.c u64 loops = loops_per_jiffy * HZ; u64 173 lib/spinlock_debug.c u64 i; u64 174 lib/spinlock_debug.c u64 loops = loops_per_jiffy * HZ; u64 248 lib/spinlock_debug.c u64 i; u64 249 lib/spinlock_debug.c u64 loops = loops_per_jiffy * HZ; u64 39 lib/string_helpers.c u64 remainder = 0, sf_cap; u64 234 mm/memcontrol.c u64 total = 0; u64 938 mm/memcontrol.c static u64 mem_cgroup_read(struct cgroup *cont, struct cftype *cft) u64 991 mm/memcontrol.c u64 unit; u64 242 mm/page-writeback.c u64 inv = dirty >> 3; u64 398 mm/page-writeback.c u64 bdi_dirty; u64 4211 mm/page_alloc.c u64 tmp; u64 4214 mm/page_alloc.c tmp = (u64)pages_min * zone->present_pages; u64 2049 mm/shmem.c u64 inum = fid->raw[2]; u64 400 net/802/garp.c delay = (u64)msecs_to_jiffies(garp_join_time) * net_random() >> 32; u64 167 net/9p/conv.c static u64 buf_get_int64(struct cbuf *buf) u64 169 net/9p/conv.c u64 ret = 0; u64 164 net/bluetooth/bnep/bnep.h u64 mc_filter; u64 79 net/can/bcm.c static inline u64 GET_U64(const struct can_frame *cp) u64 81 net/can/bcm.c return *(u64 *)cp->data; u64 1746 net/core/dev.c return (u16) (((u64) hash * dev->real_num_tx_queues) >> 32); u64 606 net/core/ethtool.c u64 *data; u64 627 net/core/ethtool.c data = kmalloc(test_len * sizeof(u64), GFP_USER); u64 637 net/core/ethtool.c if (copy_to_user(useraddr, data, test.len * sizeof(u64))) u64 720 net/core/ethtool.c u64 *data; u64 741 net/core/ethtool.c data = kmalloc(n_stats * sizeof(u64), GFP_USER); u64 751 net/core/ethtool.c if (copy_to_user(useraddr, data, stats.n_stats * sizeof(u64))) u64 139 net/core/flow.c typedef u64 flow_compare_t; u64 87 net/core/gen_estimator.c u64 last_bytes; u64 112 net/core/gen_estimator.c u64 nbytes; u64 3349 net/core/pktgen.c u64 now; u64 267 net/dccp/ackvec.c const u64 delta = dccp_delta_seqno(av->av_buf_ackno, ackno); u64 288 net/dccp/ackvec.c u64 delta = dccp_delta_seqno(ackno, av->av_buf_ackno); u64 392 net/dccp/ackvec.c u64 ackno_end_rl; u64 53 net/dccp/ackvec.h u64 av_buf_ackno; u64 78 net/dccp/ackvec.h u64 avr_ack_seqno; u64 79 net/dccp/ackvec.h u64 avr_ack_ackno; u64 509 net/dccp/ccids/ccid2.c u64 ackno, seqno; u64 587 net/dccp/ccids/ccid2.c u64 ackno_end_rl = SUB48(ackno, rl); u64 33 net/dccp/ccids/ccid2.h u64 ccid2s_seq; u64 65 net/dccp/ccids/ccid2.h u64 ccid2hctx_rpseq; u64 68 net/dccp/ccids/ccid2.h u64 ccid2hctx_high_ack; u64 88 net/dccp/ccids/ccid3.c static inline u64 rfc3390_initial_rate(struct sock *sk) u64 103 net/dccp/ccids/ccid3.c hctx->ccid3hctx_t_ipi = scaled_div32(((u64)hctx->ccid3hctx_s) << 6, u64 766 net/dccp/ccids/ccid3.c u64 fval; u64 797 net/dccp/ccids/ccid3.c const u64 ndp = dccp_sk(sk)->dccps_options_received.dccpor_ndp; u64 63 net/dccp/ccids/ccid3.h u64 ccid3or_seqno:48, u64 35 net/dccp/ccids/lib/loss_interval.h u64 li_seqno:48, u64 51 net/dccp/ccids/lib/packet_history.c u64 seqno; u64 180 net/dccp/ccids/lib/packet_history.c const u64 seq = DCCP_SKB_CB(skb)->dccpd_seq; u64 215 net/dccp/ccids/lib/packet_history.c u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, u64 226 net/dccp/ccids/lib/packet_history.c u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, u64 239 net/dccp/ccids/lib/packet_history.c u64 n1 = tfrc_rx_hist_entry(h, 1)->tfrchrx_ndp; u64 263 net/dccp/ccids/lib/packet_history.c u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, u64 289 net/dccp/ccids/lib/packet_history.c u64 n1 = tfrc_rx_hist_entry(h, 1)->tfrchrx_ndp; u64 293 net/dccp/ccids/lib/packet_history.c u64 n2 = tfrc_rx_hist_entry(h, 2)->tfrchrx_ndp; u64 332 net/dccp/ccids/lib/packet_history.c u64 s1 = tfrc_rx_hist_entry(h, 1)->tfrchrx_seqno, u64 335 net/dccp/ccids/lib/packet_history.c u64 n2 = tfrc_rx_hist_entry(h, 2)->tfrchrx_ndp, u64 64 net/dccp/ccids/lib/packet_history.h u64 tfrchrx_seqno:48, u64 67 net/dccp/ccids/lib/packet_history.h u64 tfrchrx_ndp:48; u64 32 net/dccp/ccids/lib/tfrc.h static inline u64 scaled_div(u64 a, u64 b) u64 40 net/dccp/ccids/lib/tfrc.h u64 result = scaled_div(a, b); u64 623 net/dccp/ccids/lib/tfrc_equation.c u64 result; u64 131 net/dccp/dccp.h u64 delta = SUB48(seqno2, seqno1); u64 151 net/dccp/dccp.h static inline u64 max48(const u64 seq1, const u64 seq2) u64 168 net/dccp/dccp.h return (u64)delta <= ndp + 1; u64 190 net/dccp/input.c u64 lswl, lawl, seqno = DCCP_SKB_CB(skb)->dccpd_seq, u64 360 net/dccp/ipv4.c static inline u64 dccp_v4_init_sequence(const struct sk_buff *skb) u64 60 net/dccp/options.c u64 ackno = DCCP_SKB_CB(skb)->dccpd_ack_seq; u64 346 net/dccp/options.c u64 ndp = dp->dccps_ndp_count; u64 55 net/dccp/output.c u64 ackno = dp->dccps_gsr; u64 518 net/dsa/mv88e6xxx.c data[i] = (((u64)high) << 32) | low; u64 339 net/ipv4/inet_lro.c u64 flags; u64 443 net/ipv4/inet_lro.c u64 flags; u64 270 net/ipv4/netfilter/ipt_CLUSTERIP.c return (((u64)hashval * config->num_total_nodes) >> 32) + 1; u64 83 net/ipv4/netfilter/nf_nat_core.c return ((u64)hash * nf_nat_htable_size) >> 32; u64 215 net/ipv4/netfilter/nf_nat_core.c j = ((u64)j * (maxip - minip + 1)) >> 32; u64 761 net/ipv4/route.c u64 mult; u64 763 net/ipv4/route.c mult = ((u64)ip_rt_gc_interval) << rt_hash_log; u64 34 net/ipv4/tcp_cubic.c static u64 cube_factor __read_mostly; u64 131 net/ipv4/tcp_cubic.c x = (2 * x + (u32)div64_u64(a, (u64)x * (u64)(x - 1))); u64 141 net/ipv4/tcp_cubic.c u64 offs; u64 45 net/ipv4/tcp_illinois.c u64 sum_rtt; /* sum of rtt's measured within last rtt */ u64 121 net/ipv4/tcp_illinois.c u64 t = ca->sum_rtt; u64 316 net/ipv4/tcp_illinois.c u64 t = ca->sum_rtt; u64 235 net/ipv4/tcp_vegas.c u64 target_cwnd; u64 258 net/ipv4/tcp_vegas.c target_cwnd = ((u64)old_wnd * vegas->baseRTT); u64 138 net/ipv4/tcp_veno.c u64 target_cwnd; u64 142 net/ipv4/tcp_yeah.c u64 bw; u64 3716 net/ipv6/addrconf.c int pad = bytes - sizeof(u64) * items; u64 3791 net/ipv6/addrconf.c nla = nla_reserve(skb, IFLA_INET6_STATS, IPSTATS_MIB_MAX * sizeof(u64)); u64 3796 net/ipv6/addrconf.c nla = nla_reserve(skb, IFLA_INET6_ICMP6STATS, ICMP6_MIB_MAX * sizeof(u64)); u64 98 net/mac80211/ieee80211_i.h u64 timestamp; u64 916 net/mac80211/ieee80211_i.h u64 ieee80211_sta_get_rates(struct ieee80211_local *local, u64 1016 net/mac80211/ieee80211_i.h u64 ieee80211_mandatory_rates(struct ieee80211_local *local, u64 479 net/mac80211/mesh.c u64 supp_rates = 0; u64 215 net/mac80211/mesh_hwmp.c u64 result; u64 496 net/mac80211/mesh_plink.c u64 rates; u64 97 net/mac80211/mlme.c u64 ieee80211_sta_get_rates(struct ieee80211_local *local, u64 104 net/mac80211/mlme.c u64 supp_rates; u64 245 net/mac80211/mlme.c u64 rates = 0; u64 1192 net/mac80211/mlme.c u64 rates, basic_rates; u64 1509 net/mac80211/mlme.c u64 beacon_timestamp, rx_timestamp; u64 1510 net/mac80211/mlme.c u64 supp_rates = 0; u64 1533 net/mac80211/mlme.c u64 prev_rates; u64 37 net/mac80211/rc80211_minstrel.h u64 succ_hist; u64 38 net/mac80211/rc80211_minstrel.h u64 att_hist; u64 659 net/mac80211/util.c u64 ieee80211_mandatory_rates(struct ieee80211_local *local, u64 664 net/mac80211/util.c u64 mandatory_rates; u64 60 net/netfilter/ipvs/ip_vs_est.c u64 n_inbytes, n_outbytes; u64 76 net/netfilter/nf_conntrack_core.c return ((u64)h * size) >> 32; u64 82 net/netfilter/nf_conntrack_expect.c return ((u64)hash * nf_ct_expect_hsize) >> 32; u64 454 net/netfilter/nf_conntrack_proto_dccp.c static u64 dccp_ack_seq(const struct dccp_hdr *dh) u64 459 net/netfilter/nf_conntrack_proto_dccp.c return ((u64)ntohs(dhack->dccph_ack_nr_high) << 32) + u64 123 net/netfilter/xt_hashlimit.c return ((u64)hash * ht->cfg.size) >> 32; u64 22 net/netlink/attr.c [NLA_U64] = sizeof(u64), u64 610 net/sched/cls_u32.c n->pf = kzalloc(sizeof(struct tc_u32_pcnt) + s->nkeys*sizeof(u64), GFP_KERNEL); u64 735 net/sched/cls_u32.c sizeof(struct tc_u32_pcnt) + n->sel.nkeys*sizeof(u64), u64 86 net/sched/sch_hfsc.c u64 sm1; /* scaled slope of the 1st segment */ u64 87 net/sched/sch_hfsc.c u64 ism1; /* scaled inverse-slope of the 1st segment */ u64 88 net/sched/sch_hfsc.c u64 dx; /* the x-projection of the 1st segment */ u64 89 net/sched/sch_hfsc.c u64 dy; /* the y-projection of the 1st segment */ u64 90 net/sched/sch_hfsc.c u64 sm2; /* scaled slope of the 2nd segment */ u64 91 net/sched/sch_hfsc.c u64 ism2; /* scaled inverse-slope of the 2nd segment */ u64 97 net/sched/sch_hfsc.c u64 x; /* current starting position on x-axis */ u64 98 net/sched/sch_hfsc.c u64 y; /* current starting position on y-axis */ u64 99 net/sched/sch_hfsc.c u64 sm1; /* scaled slope of the 1st segment */ u64 100 net/sched/sch_hfsc.c u64 ism1; /* scaled inverse-slope of the 1st segment */ u64 101 net/sched/sch_hfsc.c u64 dx; /* the x-projection of the 1st segment */ u64 102 net/sched/sch_hfsc.c u64 dy; /* the y-projection of the 1st segment */ u64 103 net/sched/sch_hfsc.c u64 sm2; /* scaled slope of the 2nd segment */ u64 104 net/sched/sch_hfsc.c u64 ism2; /* scaled inverse-slope of the 2nd segment */ u64 139 net/sched/sch_hfsc.c u64 cl_total; /* total work in bytes */ u64 140 net/sched/sch_hfsc.c u64 cl_cumul; /* cumulative work in bytes done by u64 143 net/sched/sch_hfsc.c u64 cl_d; /* deadline*/ u64 144 net/sched/sch_hfsc.c u64 cl_e; /* eligible time */ u64 145 net/sched/sch_hfsc.c u64 cl_vt; /* virtual time */ u64 146 net/sched/sch_hfsc.c u64 cl_f; /* time when this class will fit for u64 148 net/sched/sch_hfsc.c u64 cl_myf; /* my fit-time (calculated from this u64 150 net/sched/sch_hfsc.c u64 cl_myfadj; /* my fit-time adjustment (to cancel u64 152 net/sched/sch_hfsc.c u64 cl_cfmin; /* earliest children's fit-time (used u64 154 net/sched/sch_hfsc.c u64 cl_cvtmin; /* minimal virtual time among the u64 157 net/sched/sch_hfsc.c u64 cl_vtadj; /* intra-period cumulative vt u64 159 net/sched/sch_hfsc.c u64 cl_vtoff; /* inter-period cumulative vt offset */ u64 160 net/sched/sch_hfsc.c u64 cl_cvtmax; /* max child's vt in the last period */ u64 161 net/sched/sch_hfsc.c u64 cl_cvtoff; /* cumulative cvtmax of all periods */ u64 162 net/sched/sch_hfsc.c u64 cl_pcvtoff; /* parent's cvtoff at initialization u64 394 net/sched/sch_hfsc.c static inline u64 u64 397 net/sched/sch_hfsc.c u64 y; u64 408 net/sched/sch_hfsc.c static inline u64 u64 411 net/sched/sch_hfsc.c u64 x; u64 425 net/sched/sch_hfsc.c static u64 u64 428 net/sched/sch_hfsc.c u64 sm; u64 430 net/sched/sch_hfsc.c sm = ((u64)m << SM_SHIFT); u64 437 net/sched/sch_hfsc.c static u64 u64 440 net/sched/sch_hfsc.c u64 ism; u64 445 net/sched/sch_hfsc.c ism = ((u64)PSCHED_TICKS_PER_SEC << ISM_SHIFT); u64 453 net/sched/sch_hfsc.c static u64 u64 456 net/sched/sch_hfsc.c u64 dx; u64 458 net/sched/sch_hfsc.c dx = ((u64)d * PSCHED_TICKS_PER_SEC); u64 468 net/sched/sch_hfsc.c u64 m; u64 478 net/sched/sch_hfsc.c u64 d; u64 517 net/sched/sch_hfsc.c static u64 u64 520 net/sched/sch_hfsc.c u64 x; u64 538 net/sched/sch_hfsc.c static u64 u64 541 net/sched/sch_hfsc.c u64 y; u64 562 net/sched/sch_hfsc.c u64 y1, y2, dx, dy; u64 625 net/sched/sch_hfsc.c u64 cur_time = psched_get_time(); u64 682 net/sched/sch_hfsc.c u64 vt, f, cur_time; u64 772 net/sched/sch_hfsc.c u64 f; /* , myf_bound, delta; */ u64 992 net/sched/sch_hfsc.c u64 cur_time; u64 1414 net/sched/sch_hfsc.c u64 next_time = 0; u64 1614 net/sched/sch_hfsc.c u64 cur_time; u64 107 net/sched/sch_netem.c u64 value, rho; u64 114 net/sched/sch_netem.c rho = (u64)state->rho + 1; u64 93 net/sunrpc/auth_gss/gss_krb5_wrap.c static u64 i = 0; u64 94 net/sunrpc/auth_gss/gss_krb5_wrap.c u64 *q = (u64 *)p; u64 568 net/sunrpc/xprtrdma/rpc_rdma.c u64 off; u64 63 net/sunrpc/xprtrdma/svc_rdma_marshal.c u64 ch_offset; u64 77 net/sunrpc/xprtrdma/svc_rdma_marshal.c put_unaligned(ch_offset, (u64 *)va); u64 134 net/sunrpc/xprtrdma/svc_rdma_marshal.c u64 ch_offset; u64 142 net/sunrpc/xprtrdma/svc_rdma_marshal.c put_unaligned(ch_offset, (u64 *)va); u64 177 net/sunrpc/xprtrdma/svc_rdma_marshal.c u64 ch_offset; u64 185 net/sunrpc/xprtrdma/svc_rdma_marshal.c put_unaligned(ch_offset, (u64 *)va); u64 389 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c u64 sgl_offset; u64 368 net/sunrpc/xprtrdma/svc_rdma_sendto.c u64 rs_offset; u64 441 net/sunrpc/xprtrdma/svc_rdma_sendto.c u64 rs_offset; u64 518 net/sunrpc/xprtrdma/svc_rdma_transport.c recv_wr.wr_id = (u64)(unsigned long)ctxt; u64 1358 net/sunrpc/xprtrdma/verbs.c u64 physaddrs[RPCRDMA_MAX_DATA_SEGS]; u64 1512 net/sunrpc/xprtrdma/verbs.c param.wr_id = (u64) (unsigned long) r; u64 1614 net/sunrpc/xprtrdma/verbs.c recv_wr.wr_id = (u64) (unsigned long) rep; u64 163 net/sunrpc/xprtrdma/xprt_rdma.h u64 mr_base; /* registration result */ u64 2107 net/xfrm/xfrm_user.c l += nla_total_size(sizeof(u64)); u64 171 security/selinux/ss/ebitmap.c u64 map = c_iter->bitmap[c_idx]; u64 352 security/selinux/ss/ebitmap.c u64 map; u64 366 security/selinux/ss/ebitmap.c if (mapunit != sizeof(u64) * 8) { u64 369 security/selinux/ss/ebitmap.c mapunit, sizeof(u64) * 8, e->highbit); u64 426 security/selinux/ss/ebitmap.c rc = next_entry(&map, fp, sizeof(u64)); u64 1880 security/selinux/ss/services.c if (addrlen != sizeof(u64) * 2) { u64 78 sound/aoa/soundbus/i2sbus/i2sbus-pcm.c u64 formats = 0; u64 37 sound/aoa/soundbus/soundbus.h u64 formats; /* SNDRV_PCM_FMTBIT_* */ u64 82 sound/arm/pxa2xx-pcm.c static u64 pxa2xx_pcm_dmamask = 0xffffffff; u64 77 sound/core/control_compat.c u64 min; u64 78 sound/core/control_compat.c u64 max; u64 79 sound/core/control_compat.c u64 step; u64 619 sound/core/oss/pcm_oss.c u64 bsize = (u64)runtime->oss.buffer_bytes * (u64)bytes; u64 270 sound/core/oss/pcm_plugin.c u64 linfmts = (SNDRV_PCM_FMTBIT_U8 | SNDRV_PCM_FMTBIT_S8 | u64 133 sound/core/seq/seq_queue.h #define u64_div(x,y,q) ((q) = (u32)((u64)(x) / (u64)(y))) u64 134 sound/core/seq/seq_queue.h #define u64_mod(x,y,r) ((r) = (u32)((u64)(x) % (u64)(y))) u64 47 sound/drivers/pcsp/pcsp.h u64 __val = 1000000000ULL * (div); \ u64 69 sound/drivers/pcsp/pcsp.h u64 ns_rem; u64 26 sound/drivers/pcsp/pcsp_lib.c u64 ns; u64 721 sound/drivers/vx/vx_pcm.c u64 count; u64 729 sound/drivers/vx/vx_pcm.c count = ((u64)(rmh.Stat[0] & 0xfffff) << 24) | (u64)rmh.Stat[1]; u64 372 sound/mips/sgio2audio.c u64 *src; u64 374 sound/mips/sgio2audio.c u64 x; u64 392 sound/mips/sgio2audio.c src = (u64 *)(src_base + src_pos); u64 399 sound/mips/sgio2audio.c src_pos = (src_pos + sizeof(u64)) & CHANNEL_RING_MASK; u64 401 sound/mips/sgio2audio.c count -= sizeof(u64); u64 421 sound/mips/sgio2audio.c u64 *dst; u64 441 sound/mips/sgio2audio.c dst = (u64 *)(dst_base + dst_pos); u64 449 sound/mips/sgio2audio.c dst_pos = (dst_pos + sizeof(u64)) & CHANNEL_RING_MASK; u64 451 sound/mips/sgio2audio.c count -= sizeof(u64); u64 207 sound/oss/swarm_cs4297a.c u64 descr_a; u64 208 sound/oss/swarm_cs4297a.c u64 descr_b; u64 222 sound/oss/swarm_cs4297a.c u64 *dma_buf; // buffer for DMA contents (frames) u64 310 sound/oss/swarm_cs4297a.c volatile u64 reg_request; u64 724 sound/oss/swarm_cs4297a.c u64 *data_p; u64 777 sound/oss/swarm_cs4297a.c if (serdma_reg_access(s, (0xCLL << 60) | (1LL << 47) | ((u64)(offset & 0x7F) << 40))) u64 797 sound/oss/swarm_cs4297a.c return (serdma_reg_access(s, (0xELL << 60) | ((u64)(offset & 0x7F) << 40) | ((value & 0xffff) << 12))); u64 922 sound/oss/swarm_cs4297a.c u64 *data_p, data; u64 948 sound/oss/swarm_cs4297a.c u64 data = be64_to_cpu(*(u64 *)s_ptr); u64 949 sound/oss/swarm_cs4297a.c u64 descr_a; u64 1854 sound/oss/swarm_cs4297a.c *(u64 *)(d->dma_buf + (swptr * 4)) |= u64 2584 sound/oss/swarm_cs4297a.c u64 cfg; u64 143 sound/pci/echoaudio/echoaudio.c u64 fmask; u64 146 sound/pci/echoaudio/echoaudio.c fmask = fmt.bits[0] + ((u64)fmt.bits[1] << 32); u64 178 sound/pci/echoaudio/echoaudio.c u64 fmask; u64 182 sound/pci/echoaudio/echoaudio.c fmask = f->bits[0] + ((u64)f->bits[1] << 32); u64 2070 sound/pci/hda/hda_codec.c u64 formats = 0; u64 664 sound/pci/hda/hda_codec.h u64 formats; /* supported formats (SNDRV_PCM_FMTBIT_) */ u64 235 sound/pci/hda/hda_local.h u64 analog_formats; u64 238 sound/pci/hda/hda_local.h u64 spdif_formats; u64 147 sound/pci/mixart/mixart.h u64 abs_period_elapsed; /* last absolute stream position where period_elapsed was called (multiple of runtime->period_size) */ u64 488 sound/pci/mixart/mixart_core.c u64 sample_count = ((u64)notify->streams[i].sample_pos_high_part) << 32; u64 492 sound/pci/mixart/mixart_core.c u64 new_elapse_pos = stream->abs_period_elapsed + runtime->period_size; u64 230 sound/pci/mixart/mixart_core.h u64 scheduler; u64 243 sound/pci/mixart/mixart_core.h u64 scheduler; u64 252 sound/pci/mixart/mixart_core.h u64 scheduler; u64 393 sound/pci/mixart/mixart_core.h u64 scheduler; u64 444 sound/pci/mixart/mixart_core.h u64 scheduler; u64 504 sound/pci/mixart/mixart_core.h u64 scheduler; u64 549 sound/pci/mixart/mixart_core.h u64 scheduler; u64 1028 sound/pci/rme9652/hdsp.c u64 n; u64 3078 sound/pci/rme9652/hdsp.c u64 n; u64 4433 sound/pci/rme9652/hdsp.c u64 rms; u64 4436 sound/pci/rme9652/hdsp.c rms = ((u64)rms_high << 32) | rms_low; u64 4443 sound/pci/rme9652/hdsp.c u64 rms; u64 4446 sound/pci/rme9652/hdsp.c rms = ((u64)rms_high << 32) | rms_low; u64 833 sound/pci/rme9652/hdspm.c u64 n; u64 772 sound/ppc/snd_ps3.c u64 lpar_addr, lpar_size; u64 773 sound/ppc/snd_ps3.c u64 __iomem *mapped; u64 846 sound/ppc/snd_ps3.c u64 lpar_addr, lpar_size; u64 67 sound/ppc/snd_ps3.h u64 audio_lpar_addr; u64 68 sound/ppc/snd_ps3.h u64 audio_lpar_size; u64 74 sound/ppc/snd_ps3.h u64 audio_irq_outlet; u64 366 sound/soc/at32/at32-pcm.c static u64 at32_pcm_dmamask = 0xffffffff; u64 318 sound/soc/at91/at91-pcm.c static u64 at91_pcm_dmamask = 0xffffffff; u64 417 sound/soc/blackfin/bf5xx-ac97-pcm.c static u64 bf5xx_pcm_dmamask = DMA_32BIT_MASK; u64 248 sound/soc/blackfin/bf5xx-i2s-pcm.c static u64 bf5xx_pcm_dmamask = DMA_32BIT_MASK; u64 421 sound/soc/codecs/wm8580.c u64 Kpart; u64 708 sound/soc/codecs/wm8753.c u64 Kpart; u64 776 sound/soc/codecs/wm8900.c u64 Kpart; u64 1001 sound/soc/codecs/wm8990.c u64 Kpart; u64 674 sound/soc/codecs/wm9713.c u64 Kpart; u64 350 sound/soc/davinci/davinci-pcm.c static u64 davinci_pcm_dmamask = 0xffffffff; u64 293 sound/soc/fsl/fsl_dma.c static u64 fsl_dma_dmamask = DMA_BIT_MASK(32); u64 394 sound/soc/fsl/fsl_dma.c u64 temp_link; /* Pointer to next link descriptor */ u64 579 sound/soc/fsl/mpc5200_psc_i2s.c static u64 psc_i2s_pcm_dmamask = 0xffffffff; u64 279 sound/soc/omap/omap-pcm.c static u64 omap_pcm_dmamask = DMA_BIT_MASK(32); u64 84 sound/soc/pxa/pxa2xx-pcm.c static u64 pxa2xx_pcm_dmamask = DMA_32BIT_MASK; u64 429 sound/soc/s3c24xx/s3c24xx-pcm.c static u64 s3c24xx_pcm_dmamask = DMA_32BIT_MASK; u64 188 sound/usb/usbaudio.c u64 formats; /* format bitmasks (all or'ed) */ u64 1725 sound/usb/usbaudio.c u64 fbits; u64 70 virt/kvm/ioapic.c u64 redir_content; u64 124 virt/kvm/ioapic.c ioapic->redirtbl[index].bits |= (u64) val << 32; u64 344 virt/kvm/ioapic.c *(u64 *) val = result; u64 38 virt/kvm/ioapic.h u64 base_address; u64 44 virt/kvm/ioapic.h u64 bits;