pv_mmu_ops 72 arch/x86/kernel/asm-offsets_64.c OFFSET(PV_MMU_read_cr2, pv_mmu_ops, read_cr2); pv_mmu_ops 217 arch/x86/kernel/kvm.c pv_mmu_ops.set_pte = kvm_set_pte; pv_mmu_ops 218 arch/x86/kernel/kvm.c pv_mmu_ops.set_pte_at = kvm_set_pte_at; pv_mmu_ops 219 arch/x86/kernel/kvm.c pv_mmu_ops.set_pmd = kvm_set_pmd; pv_mmu_ops 222 arch/x86/kernel/kvm.c pv_mmu_ops.set_pte_atomic = kvm_set_pte_atomic; pv_mmu_ops 223 arch/x86/kernel/kvm.c pv_mmu_ops.set_pte_present = kvm_set_pte_present; pv_mmu_ops 224 arch/x86/kernel/kvm.c pv_mmu_ops.pte_clear = kvm_pte_clear; pv_mmu_ops 225 arch/x86/kernel/kvm.c pv_mmu_ops.pmd_clear = kvm_pmd_clear; pv_mmu_ops 227 arch/x86/kernel/kvm.c pv_mmu_ops.set_pud = kvm_set_pud; pv_mmu_ops 229 arch/x86/kernel/kvm.c pv_mmu_ops.set_pgd = kvm_set_pgd; pv_mmu_ops 232 arch/x86/kernel/kvm.c pv_mmu_ops.flush_tlb_user = kvm_flush_tlb; pv_mmu_ops 233 arch/x86/kernel/kvm.c pv_mmu_ops.release_pte = kvm_release_pt; pv_mmu_ops 234 arch/x86/kernel/kvm.c pv_mmu_ops.release_pmd = kvm_release_pt; pv_mmu_ops 235 arch/x86/kernel/kvm.c pv_mmu_ops.release_pud = kvm_release_pt; pv_mmu_ops 237 arch/x86/kernel/kvm.c pv_mmu_ops.lazy_mode.enter = kvm_enter_lazy_mmu; pv_mmu_ops 238 arch/x86/kernel/kvm.c pv_mmu_ops.lazy_mode.leave = kvm_leave_lazy_mmu; pv_mmu_ops 126 arch/x86/kernel/paravirt.c .pv_mmu_ops = pv_mmu_ops, pv_mmu_ops 376 arch/x86/kernel/paravirt.c struct pv_mmu_ops pv_mmu_ops = { pv_mmu_ops 33 arch/x86/kernel/paravirt_patch_32.c PATCH_SITE(pv_mmu_ops, read_cr2); pv_mmu_ops 34 arch/x86/kernel/paravirt_patch_32.c PATCH_SITE(pv_mmu_ops, read_cr3); pv_mmu_ops 35 arch/x86/kernel/paravirt_patch_32.c PATCH_SITE(pv_mmu_ops, write_cr3); pv_mmu_ops 43 arch/x86/kernel/paravirt_patch_64.c PATCH_SITE(pv_mmu_ops, read_cr2); pv_mmu_ops 44 arch/x86/kernel/paravirt_patch_64.c PATCH_SITE(pv_mmu_ops, read_cr3); pv_mmu_ops 45 arch/x86/kernel/paravirt_patch_64.c PATCH_SITE(pv_mmu_ops, write_cr3); pv_mmu_ops 47 arch/x86/kernel/paravirt_patch_64.c PATCH_SITE(pv_mmu_ops, flush_tlb_single); pv_mmu_ops 785 arch/x86/kernel/vmi_32.c para_fill(pv_mmu_ops.read_cr2, GetCR2); pv_mmu_ops 786 arch/x86/kernel/vmi_32.c para_fill(pv_mmu_ops.read_cr3, GetCR3); pv_mmu_ops 789 arch/x86/kernel/vmi_32.c para_fill(pv_mmu_ops.write_cr2, SetCR2); pv_mmu_ops 790 arch/x86/kernel/vmi_32.c para_fill(pv_mmu_ops.write_cr3, SetCR3); pv_mmu_ops 832 arch/x86/kernel/vmi_32.c para_wrap(pv_mmu_ops.lazy_mode.enter, vmi_enter_lazy_mmu, pv_mmu_ops 834 arch/x86/kernel/vmi_32.c para_wrap(pv_mmu_ops.lazy_mode.leave, vmi_leave_lazy, pv_mmu_ops 838 arch/x86/kernel/vmi_32.c para_wrap(pv_mmu_ops.flush_tlb_user, vmi_flush_tlb_user, _flush_tlb, FlushTLB); pv_mmu_ops 839 arch/x86/kernel/vmi_32.c para_wrap(pv_mmu_ops.flush_tlb_kernel, vmi_flush_tlb_kernel, _flush_tlb, FlushTLB); pv_mmu_ops 840 arch/x86/kernel/vmi_32.c para_fill(pv_mmu_ops.flush_tlb_single, InvalPage); pv_mmu_ops 856 arch/x86/kernel/vmi_32.c pv_mmu_ops.set_pte = vmi_set_pte; pv_mmu_ops 857 arch/x86/kernel/vmi_32.c pv_mmu_ops.set_pte_at = vmi_set_pte_at; pv_mmu_ops 858 arch/x86/kernel/vmi_32.c pv_mmu_ops.set_pmd = vmi_set_pmd; pv_mmu_ops 860 arch/x86/kernel/vmi_32.c pv_mmu_ops.set_pte_atomic = vmi_set_pte_atomic; pv_mmu_ops 861 arch/x86/kernel/vmi_32.c pv_mmu_ops.set_pte_present = vmi_set_pte_present; pv_mmu_ops 862 arch/x86/kernel/vmi_32.c pv_mmu_ops.set_pud = vmi_set_pud; pv_mmu_ops 863 arch/x86/kernel/vmi_32.c pv_mmu_ops.pte_clear = vmi_pte_clear; pv_mmu_ops 864 arch/x86/kernel/vmi_32.c pv_mmu_ops.pmd_clear = vmi_pmd_clear; pv_mmu_ops 869 arch/x86/kernel/vmi_32.c pv_mmu_ops.pte_update = vmi_update_pte; pv_mmu_ops 870 arch/x86/kernel/vmi_32.c pv_mmu_ops.pte_update_defer = vmi_update_pte_defer; pv_mmu_ops 875 arch/x86/kernel/vmi_32.c pv_mmu_ops.alloc_pte = vmi_allocate_pte; pv_mmu_ops 876 arch/x86/kernel/vmi_32.c pv_mmu_ops.alloc_pmd = vmi_allocate_pmd; pv_mmu_ops 877 arch/x86/kernel/vmi_32.c pv_mmu_ops.alloc_pmd_clone = vmi_allocate_pmd_clone; pv_mmu_ops 882 arch/x86/kernel/vmi_32.c pv_mmu_ops.release_pte = vmi_release_pte; pv_mmu_ops 883 arch/x86/kernel/vmi_32.c pv_mmu_ops.release_pmd = vmi_release_pmd; pv_mmu_ops 890 arch/x86/kernel/vmi_32.c pv_mmu_ops.kmap_atomic_pte = vmi_kmap_atomic_pte; pv_mmu_ops 1010 arch/x86/lguest/boot.c pv_mmu_ops.write_cr3 = lguest_write_cr3; pv_mmu_ops 1011 arch/x86/lguest/boot.c pv_mmu_ops.flush_tlb_user = lguest_flush_tlb_user; pv_mmu_ops 1012 arch/x86/lguest/boot.c pv_mmu_ops.flush_tlb_single = lguest_flush_tlb_single; pv_mmu_ops 1013 arch/x86/lguest/boot.c pv_mmu_ops.flush_tlb_kernel = lguest_flush_tlb_kernel; pv_mmu_ops 1014 arch/x86/lguest/boot.c pv_mmu_ops.set_pte = lguest_set_pte; pv_mmu_ops 1015 arch/x86/lguest/boot.c pv_mmu_ops.set_pte_at = lguest_set_pte_at; pv_mmu_ops 1016 arch/x86/lguest/boot.c pv_mmu_ops.set_pmd = lguest_set_pmd; pv_mmu_ops 1017 arch/x86/lguest/boot.c pv_mmu_ops.read_cr2 = lguest_read_cr2; pv_mmu_ops 1018 arch/x86/lguest/boot.c pv_mmu_ops.read_cr3 = lguest_read_cr3; pv_mmu_ops 1019 arch/x86/lguest/boot.c pv_mmu_ops.lazy_mode.enter = paravirt_enter_lazy_mmu; pv_mmu_ops 1020 arch/x86/lguest/boot.c pv_mmu_ops.lazy_mode.leave = lguest_leave_lazy_mode; pv_mmu_ops 1035 arch/x86/xen/enlighten.c pv_mmu_ops.set_pte = xen_set_pte; pv_mmu_ops 1036 arch/x86/xen/enlighten.c pv_mmu_ops.set_pmd = xen_set_pmd; pv_mmu_ops 1037 arch/x86/xen/enlighten.c pv_mmu_ops.set_pud = xen_set_pud; pv_mmu_ops 1039 arch/x86/xen/enlighten.c pv_mmu_ops.set_pgd = xen_set_pgd; pv_mmu_ops 1044 arch/x86/xen/enlighten.c pv_mmu_ops.alloc_pte = xen_alloc_pte; pv_mmu_ops 1045 arch/x86/xen/enlighten.c pv_mmu_ops.alloc_pmd = xen_alloc_pmd; pv_mmu_ops 1046 arch/x86/xen/enlighten.c pv_mmu_ops.release_pte = xen_release_pte; pv_mmu_ops 1047 arch/x86/xen/enlighten.c pv_mmu_ops.release_pmd = xen_release_pmd; pv_mmu_ops 1049 arch/x86/xen/enlighten.c pv_mmu_ops.alloc_pud = xen_alloc_pud; pv_mmu_ops 1050 arch/x86/xen/enlighten.c pv_mmu_ops.release_pud = xen_release_pud; pv_mmu_ops 1076 arch/x86/xen/enlighten.c pv_mmu_ops.read_cr2 = xen_read_cr2_direct; pv_mmu_ops 1271 arch/x86/xen/enlighten.c static const struct pv_mmu_ops xen_mmu_ops __initdata = { pv_mmu_ops 1629 arch/x86/xen/enlighten.c pv_mmu_ops = xen_mmu_ops; pv_mmu_ops 1641 arch/x86/xen/enlighten.c pv_mmu_ops.ptep_modify_prot_start = xen_ptep_modify_prot_start; pv_mmu_ops 1642 arch/x86/xen/enlighten.c pv_mmu_ops.ptep_modify_prot_commit = xen_ptep_modify_prot_commit; pv_mmu_ops 345 include/asm-x86/paravirt.h struct pv_mmu_ops pv_mmu_ops; pv_mmu_ops 355 include/asm-x86/paravirt.h extern struct pv_mmu_ops pv_mmu_ops; pv_mmu_ops 663 include/asm-x86/paravirt.h return PVOP_CALL0(unsigned long, pv_mmu_ops.read_cr2); pv_mmu_ops 668 include/asm-x86/paravirt.h PVOP_VCALL1(pv_mmu_ops.write_cr2, x); pv_mmu_ops 673 include/asm-x86/paravirt.h return PVOP_CALL0(unsigned long, pv_mmu_ops.read_cr3); pv_mmu_ops 678 include/asm-x86/paravirt.h PVOP_VCALL1(pv_mmu_ops.write_cr3, x); pv_mmu_ops 940 include/asm-x86/paravirt.h (*pv_mmu_ops.pagetable_setup_start)(base); pv_mmu_ops 945 include/asm-x86/paravirt.h (*pv_mmu_ops.pagetable_setup_done)(base); pv_mmu_ops 960 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.activate_mm, prev, next); pv_mmu_ops 966 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.dup_mmap, oldmm, mm); pv_mmu_ops 971 include/asm-x86/paravirt.h PVOP_VCALL1(pv_mmu_ops.exit_mmap, mm); pv_mmu_ops 976 include/asm-x86/paravirt.h PVOP_VCALL0(pv_mmu_ops.flush_tlb_user); pv_mmu_ops 980 include/asm-x86/paravirt.h PVOP_VCALL0(pv_mmu_ops.flush_tlb_kernel); pv_mmu_ops 984 include/asm-x86/paravirt.h PVOP_VCALL1(pv_mmu_ops.flush_tlb_single, addr); pv_mmu_ops 990 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.flush_tlb_others, &cpumask, mm, va); pv_mmu_ops 995 include/asm-x86/paravirt.h return PVOP_CALL1(int, pv_mmu_ops.pgd_alloc, mm); pv_mmu_ops 1000 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.pgd_free, mm, pgd); pv_mmu_ops 1005 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.alloc_pte, mm, pfn); pv_mmu_ops 1009 include/asm-x86/paravirt.h PVOP_VCALL1(pv_mmu_ops.release_pte, pfn); pv_mmu_ops 1014 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.alloc_pmd, mm, pfn); pv_mmu_ops 1020 include/asm-x86/paravirt.h PVOP_VCALL4(pv_mmu_ops.alloc_pmd_clone, pfn, clonepfn, start, count); pv_mmu_ops 1024 include/asm-x86/paravirt.h PVOP_VCALL1(pv_mmu_ops.release_pmd, pfn); pv_mmu_ops 1029 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.alloc_pud, mm, pfn); pv_mmu_ops 1033 include/asm-x86/paravirt.h PVOP_VCALL1(pv_mmu_ops.release_pud, pfn); pv_mmu_ops 1040 include/asm-x86/paravirt.h ret = PVOP_CALL2(unsigned long, pv_mmu_ops.kmap_atomic_pte, page, type); pv_mmu_ops 1048 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.pte_update, mm, addr, ptep); pv_mmu_ops 1054 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.pte_update_defer, mm, addr, ptep); pv_mmu_ops 1063 include/asm-x86/paravirt.h pv_mmu_ops.make_pte, pv_mmu_ops 1067 include/asm-x86/paravirt.h pv_mmu_ops.make_pte, pv_mmu_ops 1078 include/asm-x86/paravirt.h ret = PVOP_CALL2(pteval_t, pv_mmu_ops.pte_val, pv_mmu_ops 1081 include/asm-x86/paravirt.h ret = PVOP_CALL1(pteval_t, pv_mmu_ops.pte_val, pv_mmu_ops 1092 include/asm-x86/paravirt.h ret = PVOP_CALL2(pteval_t, pv_mmu_ops.pte_flags, pv_mmu_ops 1095 include/asm-x86/paravirt.h ret = PVOP_CALL1(pteval_t, pv_mmu_ops.pte_flags, pv_mmu_ops 1109 include/asm-x86/paravirt.h ret = PVOP_CALL2(pgdval_t, pv_mmu_ops.make_pgd, pv_mmu_ops 1112 include/asm-x86/paravirt.h ret = PVOP_CALL1(pgdval_t, pv_mmu_ops.make_pgd, pv_mmu_ops 1123 include/asm-x86/paravirt.h ret = PVOP_CALL2(pgdval_t, pv_mmu_ops.pgd_val, pv_mmu_ops 1126 include/asm-x86/paravirt.h ret = PVOP_CALL1(pgdval_t, pv_mmu_ops.pgd_val, pv_mmu_ops 1138 include/asm-x86/paravirt.h ret = PVOP_CALL3(pteval_t, pv_mmu_ops.ptep_modify_prot_start, pv_mmu_ops 1149 include/asm-x86/paravirt.h pv_mmu_ops.ptep_modify_prot_commit(mm, addr, ptep, pte); pv_mmu_ops 1151 include/asm-x86/paravirt.h PVOP_VCALL4(pv_mmu_ops.ptep_modify_prot_commit, pv_mmu_ops 1158 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.set_pte, ptep, pv_mmu_ops 1161 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.set_pte, ptep, pv_mmu_ops 1170 include/asm-x86/paravirt.h pv_mmu_ops.set_pte_at(mm, addr, ptep, pte); pv_mmu_ops 1172 include/asm-x86/paravirt.h PVOP_VCALL4(pv_mmu_ops.set_pte_at, mm, addr, ptep, pte.pte); pv_mmu_ops 1180 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.set_pmd, pmdp, val, (u64)val >> 32); pv_mmu_ops 1182 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.set_pmd, pmdp, val); pv_mmu_ops 1191 include/asm-x86/paravirt.h ret = PVOP_CALL2(pmdval_t, pv_mmu_ops.make_pmd, pv_mmu_ops 1194 include/asm-x86/paravirt.h ret = PVOP_CALL1(pmdval_t, pv_mmu_ops.make_pmd, pv_mmu_ops 1205 include/asm-x86/paravirt.h ret = PVOP_CALL2(pmdval_t, pv_mmu_ops.pmd_val, pv_mmu_ops 1208 include/asm-x86/paravirt.h ret = PVOP_CALL1(pmdval_t, pv_mmu_ops.pmd_val, pv_mmu_ops 1219 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.set_pud, pudp, pv_mmu_ops 1222 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.set_pud, pudp, pv_mmu_ops 1231 include/asm-x86/paravirt.h ret = PVOP_CALL2(pudval_t, pv_mmu_ops.make_pud, pv_mmu_ops 1234 include/asm-x86/paravirt.h ret = PVOP_CALL1(pudval_t, pv_mmu_ops.make_pud, pv_mmu_ops 1245 include/asm-x86/paravirt.h ret = PVOP_CALL2(pudval_t, pv_mmu_ops.pud_val, pv_mmu_ops 1248 include/asm-x86/paravirt.h ret = PVOP_CALL1(pudval_t, pv_mmu_ops.pud_val, pv_mmu_ops 1259 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.set_pgd, pgdp, pv_mmu_ops 1262 include/asm-x86/paravirt.h PVOP_VCALL2(pv_mmu_ops.set_pgd, pgdp, pv_mmu_ops 1285 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.set_pte_atomic, ptep, pv_mmu_ops 1293 include/asm-x86/paravirt.h pv_mmu_ops.set_pte_present(mm, addr, ptep, pte); pv_mmu_ops 1299 include/asm-x86/paravirt.h PVOP_VCALL3(pv_mmu_ops.pte_clear, mm, addr, ptep); pv_mmu_ops 1304 include/asm-x86/paravirt.h PVOP_VCALL1(pv_mmu_ops.pmd_clear, pmdp); pv_mmu_ops 1367 include/asm-x86/paravirt.h PVOP_VCALL0(pv_mmu_ops.lazy_mode.enter); pv_mmu_ops 1372 include/asm-x86/paravirt.h PVOP_VCALL0(pv_mmu_ops.lazy_mode.leave); pv_mmu_ops 1386 include/asm-x86/paravirt.h pv_mmu_ops.set_fixmap(idx, phys, flags); pv_mmu_ops 1628 include/asm-x86/paravirt.h call PARA_INDIRECT(pv_mmu_ops+PV_MMU_read_cr2); \