__pa 96 arch/x86/kernel/aperture_64.c if (!p || __pa(p)+aper_size > 0xffffffff) {
__pa 101 arch/x86/kernel/aperture_64.c free_bootmem(__pa(p), aper_size);
__pa 105 arch/x86/kernel/aperture_64.c aper_size >> 10, __pa(p));
__pa 106 arch/x86/kernel/aperture_64.c insert_aperture_resource((u32)__pa(p), aper_size);
__pa 107 arch/x86/kernel/aperture_64.c register_nosave_region((u32)__pa(p) >> PAGE_SHIFT,
__pa 108 arch/x86/kernel/aperture_64.c (u32)__pa(p+aper_size) >> PAGE_SHIFT);
__pa 110 arch/x86/kernel/aperture_64.c return (u32)__pa(p);
__pa 1271 arch/x86/kernel/apic_32.c apic_phys = __pa(apic_phys);
__pa 1217 arch/x86/kernel/apic_64.c apic_phys = __pa(apic_phys);
__pa 69 arch/x86/kernel/cpu/intel.c __set_fixmap(FIX_F00F_IDT, __pa(&idt_table), PAGE_KERNEL_RO);
__pa 80 arch/x86/kernel/efi_32.c gdt_descr.address = __pa(get_cpu_gdt_table(0));
__pa 335 arch/x86/kernel/es7000_32.c vect = ((unsigned long)__pa(eip)/0x1000) << 16;
__pa 2900 arch/x86/kernel/io_apic_32.c ioapic_phys = __pa(ioapic_phys);
__pa 2936 arch/x86/kernel/io_apic_64.c ioapic_phys = __pa(ioapic_phys);
__pa 59 arch/x86/kernel/kvm.c a1 = __pa(buffer);
__pa 103 arch/x86/kernel/kvm.c pte_phys = (unsigned long)__pa(dest);
__pa 54 arch/x86/kernel/kvmclock.c low = (int)__pa(&wall_clock);
__pa 55 arch/x86/kernel/kvmclock.c high = ((u64)__pa(&wall_clock) >> 32);
__pa 95 arch/x86/kernel/kvmclock.c low = (int)__pa(&per_cpu(hv_clock, cpu)) | 1;
__pa 96 arch/x86/kernel/kvmclock.c high = ((u64)__pa(&per_cpu(hv_clock, cpu)) >> 32);
__pa 151 arch/x86/kernel/machine_kexec_32.c page_list[PA_CONTROL_PAGE] = __pa(control_page);
__pa 153 arch/x86/kernel/machine_kexec_32.c page_list[PA_PGD] = __pa(kexec_pgd);
__pa 156 arch/x86/kernel/machine_kexec_32.c page_list[PA_PMD_0] = __pa(kexec_pmd0);
__pa 158 arch/x86/kernel/machine_kexec_32.c page_list[PA_PMD_1] = __pa(kexec_pmd1);
__pa 161 arch/x86/kernel/machine_kexec_32.c page_list[PA_PTE_0] = __pa(kexec_pte0);
__pa 163 arch/x86/kernel/machine_kexec_32.c page_list[PA_PTE_1] = __pa(kexec_pte1);
__pa 62 arch/x86/kernel/machine_kexec_64.c set_pud(level3p++, __pud(__pa(level2p) | _KERNPG_TABLE));
__pa 98 arch/x86/kernel/machine_kexec_64.c set_pgd(level4p++, __pgd(__pa(level3p) | _KERNPG_TABLE));
__pa 215 arch/x86/kernel/machine_kexec_64.c (unsigned long)__pa(page_address(image->control_code_page));
__pa 839 arch/x86/kernel/pci-calgary_64.c table_phys = (u64)__pa(tbl->it_base);
__pa 103 arch/x86/kernel/pci-dma.c free_bootmem(__pa(dma32_bootmem_ptr), dma32_bootmem_size);
__pa 578 arch/x86/kernel/pci-gart_64.c enable_gart_translation(dev, __pa(agp_gatt_table));
__pa 854 arch/x86/kernel/pci-gart_64.c gart_unmapped_entry = GPTE_ENCODE(__pa(scratch));
__pa 706 arch/x86/kernel/setup.c addr, __pa(addr), *addr);
__pa 167 arch/x86/kernel/setup_percpu.c cpu, __pa(ptr));
__pa 173 arch/x86/kernel/setup_percpu.c cpu, node, __pa(ptr));
__pa 188 arch/x86/kernel/tce_64.c free_bootmem(__pa(tbl), size);
__pa 659 arch/x86/kernel/tlb_uv.c pa = __pa((unsigned long)adp);
__pa 453 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(ptep) >> PAGE_SHIFT, VMI_PAGE_PTE);
__pa 459 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(ptep) >> PAGE_SHIFT, VMI_PAGE_PTE);
__pa 466 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(ptep) >> PAGE_SHIFT, VMI_PAGE_PTE | VMI_PAGE_PD);
__pa 472 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(ptep) >> PAGE_SHIFT, VMI_PAGE_PTE);
__pa 480 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(pmdp) >> PAGE_SHIFT, VMI_PAGE_PMD);
__pa 483 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(pmdp) >> PAGE_SHIFT, VMI_PAGE_PGD);
__pa 505 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(ptep) >> PAGE_SHIFT, VMI_PAGE_PTE);
__pa 513 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(pudp) >> PAGE_SHIFT, VMI_PAGE_PGD);
__pa 520 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(ptep) >> PAGE_SHIFT, VMI_PAGE_PTE);
__pa 527 arch/x86/kernel/vmi_32.c vmi_check_page_type(__pa(pmd) >> PAGE_SHIFT, VMI_PAGE_PMD);
__pa 571 arch/x86/kernel/vmi_32.c ap.cr3 = __pa(swapper_pg_dir);
__pa 476 arch/x86/kvm/mmu.c sp = page_header(__pa(spte));
__pa 537 arch/x86/kvm/mmu.c sp = page_header(__pa(spte));
__pa 1099 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp = page_header(__pa(pte));
__pa 1275 arch/x86/kvm/mmu.c __pa(new_table->spt)
__pa 1378 arch/x86/kvm/mmu.c root = __pa(sp->spt);
__pa 1401 arch/x86/kvm/mmu.c root = __pa(sp->spt);
__pa 1405 arch/x86/kvm/mmu.c vcpu->arch.mmu.root_hpa = __pa(vcpu->arch.mmu.pae_root);
__pa 345 arch/x86/kvm/paging_tmpl.h shadow_addr = __pa(shadow_page->spt);
__pa 550 arch/x86/kvm/svm.c control->msrpm_base_pa = __pa(svm->msrpm);
__pa 321 arch/x86/kvm/vmx.c u64 phys_addr = __pa(vmcs);
__pa 623 arch/x86/kvm/vmx.c u64 phys_addr = __pa(vmx->vmcs);
__pa 1043 arch/x86/kvm/vmx.c u64 phys_addr = __pa(per_cpu(vmxarea, cpu));
__pa 260 arch/x86/lguest/boot.c hcall(LHCALL_LOAD_GDT, __pa(desc->address), GDT_ENTRIES, 0);
__pa 270 arch/x86/lguest/boot.c hcall(LHCALL_LOAD_GDT, __pa(dt), GDT_ENTRIES, 0);
__pa 282 arch/x86/lguest/boot.c lazy_hcall(LHCALL_LOAD_TLS, __pa(&t->tls_array), cpu, 0);
__pa 482 arch/x86/lguest/boot.c lazy_hcall(LHCALL_SET_PTE, __pa(mm->pgd), addr, pteval.pte_low);
__pa 491 arch/x86/lguest/boot.c lazy_hcall(LHCALL_SET_PMD, __pa(pmdp)&PAGE_MASK,
__pa 492 arch/x86/lguest/boot.c (__pa(pmdp)&(PAGE_SIZE-1))/4, 0);
__pa 840 arch/x86/lguest/boot.c hcall(LHCALL_SHUTDOWN, __pa("Power down"), LGUEST_SHUTDOWN_POWEROFF, 0);
__pa 850 arch/x86/lguest/boot.c hcall(LHCALL_SHUTDOWN, __pa(p), LGUEST_SHUTDOWN_POWEROFF, 0);
__pa 891 arch/x86/lguest/boot.c hcall(LHCALL_NOTIFY, __pa(scratch), 0, 0);
__pa 901 arch/x86/lguest/boot.c hcall(LHCALL_SHUTDOWN, __pa(reason), LGUEST_SHUTDOWN_RESTART, 0);
__pa 1044 arch/x86/lguest/boot.c init_pg_tables_start = __pa(pg0);
__pa 1045 arch/x86/lguest/boot.c init_pg_tables_end = __pa(pg0);
__pa 101 arch/x86/mm/init_32.c paravirt_alloc_pmd(&init_mm, __pa(pmd_table) >> PAGE_SHIFT);
__pa 102 arch/x86/mm/init_32.c set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT));
__pa 134 arch/x86/mm/init_32.c paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT);
__pa 135 arch/x86/mm/init_32.c set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
__pa 473 arch/x86/mm/init_32.c paravirt_alloc_pmd(&init_mm, __pa(base) >> PAGE_SHIFT);
__pa 551 arch/x86/mm/init_32.c set_pgd(swapper_pg_dir+i, __pgd(1 + __pa(empty_zero_page)));
__pa 951 arch/x86/mm/init_32.c __set_fixmap(FIX_WP_TEST, __pa(&swapper_pg_dir), PAGE_READONLY);
__pa 241 arch/x86/mm/init_64.c set_pgd(pgd, __pgd(__pa(pud) | _KERNPG_TABLE |
__pa 247 arch/x86/mm/init_64.c set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE |
__pa 305 arch/x86/mm/init_64.c *phys = __pa(adr);
__pa 613 arch/x86/mm/init_64.c last_map_addr = phys_pud_update(pgd, __pa(start),
__pa 614 arch/x86/mm/init_64.c __pa(end), page_size_mask);
__pa 619 arch/x86/mm/init_64.c last_map_addr = phys_pud_init(pud, __pa(start), __pa(next),
__pa 1188 arch/x86/mm/init_64.c entry = pfn_pte(__pa(p) >> PAGE_SHIFT,
__pa 562 arch/x86/mm/ioremap.c paravirt_release_pte(__pa(bm_pte) >> PAGE_SHIFT);
__pa 163 arch/x86/mm/numa_64.c ptr = __alloc_bootmem_nopanic(size, align, __pa(MAX_DMA_ADDRESS));
__pa 194 arch/x86/mm/numa_64.c nodedata_phys = __pa(node_data[nodeid]);
__pa 228 arch/x86/mm/numa_64.c bootmap_start = __pa(bootmap);
__pa 94 arch/x86/mm/pageattr.c return __pa(_text) >> PAGE_SHIFT;
__pa 99 arch/x86/mm/pageattr.c return __pa(roundup((unsigned long)_end, PMD_SIZE)) >> PAGE_SHIFT;
__pa 270 arch/x86/mm/pageattr.c if (within(pfn, __pa((unsigned long)__start_rodata) >> PAGE_SHIFT,
__pa 271 arch/x86/mm/pageattr.c __pa((unsigned long)__end_rodata) >> PAGE_SHIFT))
__pa 869 arch/x86/mm/pageattr.c if (reserve_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE,
__pa 886 arch/x86/mm/pageattr.c start = __pa(addr[i]);
__pa 888 arch/x86/mm/pageattr.c if (end != __pa(addr[i + 1]))
__pa 900 arch/x86/mm/pageattr.c unsigned long tmp = __pa(addr[i]);
__pa 905 arch/x86/mm/pageattr.c if (end != __pa(addr[i + 1]))
__pa 926 arch/x86/mm/pageattr.c if (reserve_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE,
__pa 942 arch/x86/mm/pageattr.c free_memtype(__pa(addr), __pa(addr) + numpages * PAGE_SIZE);
__pa 953 arch/x86/mm/pageattr.c unsigned long start = __pa(addr[i]);
__pa 957 arch/x86/mm/pageattr.c if (end != __pa(addr[i + 1]))
__pa 532 arch/x86/mm/pat.c (pfn << PAGE_SHIFT) >= __pa(high_memory)) {
__pa 36 arch/x86/mm/pgtable.c paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT);
__pa 43 arch/x86/mm/pgtable.c paravirt_release_pud(__pa(pud) >> PAGE_SHIFT);
__pa 77 arch/x86/mm/pgtable.c paravirt_alloc_pmd_clone(__pa(pgd) >> PAGE_SHIFT,
__pa 78 arch/x86/mm/pgtable.c __pa(swapper_pg_dir) >> PAGE_SHIFT,
__pa 127 arch/x86/mm/pgtable.c paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
__pa 131 arch/x86/mm/pgtable.c set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT));
__pa 44 arch/x86/power/hibernate_32.c set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT));
__pa 67 arch/x86/power/hibernate_32.c set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE));
__pa 141 arch/x86/power/hibernate_32.c __pgd(__pa(empty_zero_page) | _PAGE_PRESENT));
__pa 57 arch/x86/power/hibernate_64.c set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE));
__pa 95 arch/x86/power/hibernate_64.c if ((error = res_phys_pud_init(pud, __pa(start), __pa(next))))
__pa 98 arch/x86/power/hibernate_64.c mk_kernel_pgd(__pa(pud)));
__pa 781 arch/x86/xen/enlighten.c __xen_write_cr3(false, __pa(user_pgd));
__pa 908 arch/x86/xen/enlighten.c __pgd(__pa(level3_user_vsyscall) | _PAGE_TABLE);
__pa 1439 arch/x86/xen/enlighten.c unsigned long pfn = __pa(addr) >> PAGE_SHIFT;
__pa 1468 arch/x86/xen/enlighten.c pmd[pmdidx] = __pmd(__pa(pte_page) | _PAGE_TABLE);
__pa 1554 arch/x86/xen/enlighten.c pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
__pa 1565 arch/x86/xen/enlighten.c __xen_write_cr3(true, __pa(pgd));
__pa 1568 arch/x86/xen/enlighten.c reserve_early(__pa(xen_start_info->pt_base),
__pa 1569 arch/x86/xen/enlighten.c __pa(xen_start_info->pt_base +
__pa 1582 arch/x86/xen/enlighten.c init_pg_tables_start = __pa(pgd);
__pa 1583 arch/x86/xen/enlighten.c init_pg_tables_end = __pa(pgd) + xen_start_info->nr_pt_frames*PAGE_SIZE;
__pa 1593 arch/x86/xen/enlighten.c __pgd(__pa(level2_kernel_pgt) | _PAGE_PRESENT));
__pa 1599 arch/x86/xen/enlighten.c pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
__pa 1601 arch/x86/xen/enlighten.c xen_write_cr3(__pa(swapper_pg_dir));
__pa 1603 arch/x86/xen/enlighten.c pin_pagetable_pfn(MMUEXT_PIN_L3_TABLE, PFN_DOWN(__pa(swapper_pg_dir)));
__pa 1694 arch/x86/xen/enlighten.c ? __pa(xen_start_info->mod_start) : 0;
__pa 1696 arch/x86/xen/enlighten.c boot_params.hdr.cmd_line_ptr = __pa(xen_start_info->cmd_line);
__pa 856 arch/x86/xen/mmu.c xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd)));
__pa 860 arch/x86/xen/mmu.c xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(user_pgd)));
__pa 869 arch/x86/xen/mmu.c xen_do_pin(MMUEXT_PIN_L3_TABLE, PFN_DOWN(__pa(pgd)));
__pa 968 arch/x86/xen/mmu.c xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
__pa 975 arch/x86/xen/mmu.c xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(user_pgd)));
__pa 1053 arch/x86/xen/mmu.c if (x86_read_percpu(xen_current_cr3) == __pa(mm->pgd)) {
__pa 1081 arch/x86/xen/mmu.c if (per_cpu(xen_current_cr3, cpu) == __pa(mm->pgd))
__pa 61 arch/x86/xen/setup.c e820_add_region(__pa(xen_start_info->mfn_list),
__pa 268 arch/x86/xen/smp.c per_cpu(xen_cr3, cpu) = __pa(swapper_pg_dir);
__pa 28 include/asm-cris/io.h return __pa(address);
__pa 61 include/asm-cris/page.h #define page_to_phys(page) __pa((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET)
__pa 183 include/asm-cris/pgtable.h pte_val(pte) = __pa(page) | pgprot_val(pgprot);
__pa 62 include/asm-frv/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa 63 include/asm-frv/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa 23 include/asm-frv/pgalloc.h #define pmd_populate_kernel(mm, pmd, pte) __set_pmd(pmd, __pa(pte) | _PAGE_TABLE)
__pa 27 include/asm-m32r/io.h return __pa(address);
__pa 76 include/asm-m32r/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa 77 include/asm-m32r/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa 9 include/asm-m32r/pgalloc.h set_pmd(pmd, __pmd(_PAGE_TABLE + __pa(pte)))
__pa 120 include/asm-m68k/cacheflush.h : : "a" (__pa(vaddr)));
__pa 121 include/asm-m68k/motorola_pgtable.h pgd_val(*pgdp) = _PAGE_TABLE | _PAGE_ACCESSED | __pa(pmdp);
__pa 173 include/asm-m68k/page.h #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
__pa 69 include/asm-m68k/sun3_pgalloc.h pmd_val(*pmd) = __pa((unsigned long)pte);
__pa 74 include/asm-m68k/sun3_pgalloc.h pmd_val(*pmd) = __pa((unsigned long)page_address(page));
__pa 17 include/asm-m68k/sun3_pgtable.h #define VTOP(addr) __pa(addr)
__pa 20 include/asm-m68k/virtconvert.h return __pa(address);
__pa 31 include/asm-m68k/virtconvert.h __pa(PAGE_OFFSET + (((page) - pg_data_map[0].node_mem_map) << PAGE_SHIFT))
__pa 230 include/asm-mn10300/io.h return __pa(address);
__pa 117 include/asm-mn10300/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa 118 include/asm-mn10300/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa 24 include/asm-mn10300/pgalloc.h set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE))
__pa 13 include/asm-parisc/io.h #define virt_to_phys(a) ((unsigned long)__pa(a))
__pa 51 include/asm-parisc/mmu_context.h mtctl(__pa(next->pgd), 25);
__pa 162 include/asm-parisc/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa 165 include/asm-parisc/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa 37 include/asm-parisc/pgalloc.h + (__u32)(__pa((unsigned long)pgd) >> PxD_VALUE_SHIFT));
__pa 61 include/asm-parisc/pgalloc.h (__u32)(__pa((unsigned long)pmd) >> PxD_VALUE_SHIFT));
__pa 109 include/asm-parisc/pgalloc.h + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT));
__pa 113 include/asm-parisc/pgalloc.h + (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT));
__pa 17 include/asm-um/io.h return __pa((void *) address);
__pa 116 include/asm-um/page.h #define virt_addr_valid(v) pfn_valid(phys_to_pfn(__pa(v)))
__pa 15 include/asm-um/pgalloc.h set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte)))
__pa 60 include/asm-um/pgtable-3level.h set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd)))
__pa 274 include/asm-um/pgtable.h #define __virt_to_page(virt) phys_to_page(__pa(virt))
__pa 71 include/asm-x86/io_32.h return __pa(address);
__pa 152 include/asm-x86/io_64.h return __pa(address);
__pa 99 include/asm-x86/mmzone_32.h __alloc_bootmem_node(NODE_DATA(0), (x), SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS))
__pa 102 include/asm-x86/mmzone_32.h __pa(MAX_DMA_ADDRESS))
__pa 106 include/asm-x86/mmzone_32.h __alloc_bootmem_node(NODE_DATA(0), (x), PAGE_SIZE, __pa(MAX_DMA_ADDRESS))
__pa 109 include/asm-x86/mmzone_32.h __pa(MAX_DMA_ADDRESS))
__pa 117 include/asm-x86/mmzone_32.h __pa(MAX_DMA_ADDRESS)); \
__pa 124 include/asm-x86/mmzone_32.h __pa(MAX_DMA_ADDRESS)); \
__pa 185 include/asm-x86/page.h #define __pa_symbol(x) __pa(__phys_reloc_hide((unsigned long)(x)))
__pa 190 include/asm-x86/page.h #define __boot_pa(x) __pa(x)
__pa 196 include/asm-x86/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa 53 include/asm-x86/pgalloc.h paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
__pa 54 include/asm-x86/pgalloc.h set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
__pa 87 include/asm-x86/pgalloc.h paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
__pa 88 include/asm-x86/pgalloc.h set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd)));
__pa 95 include/asm-x86/pgalloc.h paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT);
__pa 96 include/asm-x86/pgalloc.h set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(pud)));
__pa 118 include/asm-x86/pgtable-3level.h if (__pa(pudp) >= pgd && __pa(pudp) <
__pa 186 include/asm-x86/processor.h write_cr3(__pa(pgdir));
__pa 47 include/asm-x86/uv/uv_bau.h #define uv_physnodeaddr(x) ((__pa((unsigned long)(x)) & uv_mmask))
__pa 184 include/asm-x86/uv/uv_hub.h return __pa(v) | uv_hub_info->gnode_upper;
__pa 121 include/asm-x86/xen/page.h #define virt_to_machine(v) (phys_to_machine(XPADDR(__pa(v))))
__pa 122 include/asm-x86/xen/page.h #define virt_to_mfn(v) (pfn_to_mfn(PFN_DOWN(__pa(v))))
__pa 48 include/asm-xtensa/io.h return __pa(address);
__pa 160 include/asm-xtensa/page.h #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
__pa 162 include/asm-xtensa/page.h #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
__pa 99 include/linux/bootmem.h __alloc_bootmem(x, SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS))
__pa 101 include/linux/bootmem.h __alloc_bootmem_nopanic(x, SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS))
__pa 105 include/linux/bootmem.h __alloc_bootmem(x, PAGE_SIZE, __pa(MAX_DMA_ADDRESS))
__pa 107 include/linux/bootmem.h __alloc_bootmem_nopanic(x, PAGE_SIZE, __pa(MAX_DMA_ADDRESS))
__pa 111 include/linux/bootmem.h __alloc_bootmem_node(pgdat, x, SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS))
__pa 113 include/linux/bootmem.h __alloc_bootmem_node(pgdat, x, PAGE_SIZE, __pa(MAX_DMA_ADDRESS))
__pa 317 include/linux/efi.h unsigned long paddr = __pa(start + i);
__pa 1362 kernel/kexec.c return __pa((unsigned long)(char *)&vmcoreinfo_note);
__pa 4084 mm/page_alloc.c __pa(PAGE_OFFSET) >> PAGE_SHIFT, NULL);
__pa 58 mm/sparse-vmemmap.c __pa(MAX_DMA_ADDRESS));
__pa 80 mm/sparse-vmemmap.c entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
__pa 287 mm/sparse.c section_nr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT);
__pa 299 mm/sparse.c usemap_snr = pfn_to_section_nr(__pa(usemap) >> PAGE_SHIFT);
__pa 300 mm/sparse.c pgdat_snr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT);
__pa 450 mm/sparse.c free_bootmem(__pa(usemap_map), size);
__pa 588 sound/parisc/harmony.c ss->runtime->dma_addr = __pa(ss->runtime->dma_area);