page 36 arch/x86/boot/video.c u16 ax, page; page 42 arch/x86/boot/video.c : "+a" (ax), "=b" (page) page 47 arch/x86/boot/video.c boot_params.screen_info.orig_video_page = page >> 8; page 504 arch/x86/kernel/alternative.c struct page *pages[2]; page 343 arch/x86/kernel/amd_iommu.c u64 __pte, *pte, *page; page 355 arch/x86/kernel/amd_iommu.c page = (u64 *)get_zeroed_page(GFP_KERNEL); page 356 arch/x86/kernel/amd_iommu.c if (!page) page 358 arch/x86/kernel/amd_iommu.c *pte = IOMMU_L2_PDE(virt_to_phys(page)); page 365 arch/x86/kernel/amd_iommu.c page = (u64 *)get_zeroed_page(GFP_KERNEL); page 366 arch/x86/kernel/amd_iommu.c if (!page) page 368 arch/x86/kernel/amd_iommu.c *pte = IOMMU_L1_PDE(virt_to_phys(page)); page 48 arch/x86/kernel/cpu/mtrr/if.c if (!page) { page 67 arch/x86/kernel/cpu/mtrr/if.c if (!page) { page 36 arch/x86/kernel/kdebugfs.c struct page *pg; page 121 arch/x86/kernel/kdebugfs.c struct page *pg; page 95 arch/x86/kernel/kvm.c struct page *page; page 98 arch/x86/kernel/kvm.c page = kmap_atomic_to_page(dest); page 99 arch/x86/kernel/kvm.c pte_phys = page_to_pfn(page); page 52 arch/x86/kernel/machine_kexec_64.c struct page *page; page 55 arch/x86/kernel/machine_kexec_64.c page = kimage_alloc_control_pages(image, 0); page 56 arch/x86/kernel/machine_kexec_64.c if (!page) { page 60 arch/x86/kernel/machine_kexec_64.c level2p = (pmd_t *)page_address(page); page 85 arch/x86/kernel/machine_kexec_64.c struct page *page; page 88 arch/x86/kernel/machine_kexec_64.c page = kimage_alloc_control_pages(image, 0); page 89 arch/x86/kernel/machine_kexec_64.c if (!page) { page 93 arch/x86/kernel/machine_kexec_64.c level3p = (pud_t *)page_address(page); page 141 arch/x86/kernel/pci-dma.c struct page *page; page 148 arch/x86/kernel/pci-dma.c page = alloc_pages_node(dev_to_node(dev), flag, get_order(size)); page 149 arch/x86/kernel/pci-dma.c if (!page) page 152 arch/x86/kernel/pci-dma.c addr = page_to_phys(page); page 154 arch/x86/kernel/pci-dma.c __free_pages(page, get_order(size)); page 165 arch/x86/kernel/pci-dma.c return page_address(page); page 493 arch/x86/kernel/pci-gart_64.c struct page *page; page 497 arch/x86/kernel/pci-gart_64.c page = alloc_pages(flag | __GFP_ZERO, get_order(size)); page 498 arch/x86/kernel/pci-gart_64.c if (!page) page 502 arch/x86/kernel/pci-gart_64.c paddr = dma_map_area(dev, page_to_phys(page), size, page 508 arch/x86/kernel/pci-gart_64.c return page_address(page); page 510 arch/x86/kernel/pci-gart_64.c __free_pages(page, get_order(size)); page 293 arch/x86/kernel/vmi_32.c struct page *page = pfn_to_page(boot_page_allocations[i].pfn); page 294 arch/x86/kernel/vmi_32.c page->type = boot_page_allocations[i].type; page 295 arch/x86/kernel/vmi_32.c page->type = boot_page_allocations[i].type & page 315 arch/x86/kernel/vmi_32.c if (page_address(page)) page 316 arch/x86/kernel/vmi_32.c ptr = (u32 *)page_address(page); page 340 arch/x86/kernel/vmi_32.c struct page *page = pfn_to_page(pfn); page 342 arch/x86/kernel/vmi_32.c BUG_ON(page->type); page 344 arch/x86/kernel/vmi_32.c BUG_ON(page->type == VMI_PAGE_NORMAL); page 345 arch/x86/kernel/vmi_32.c page->type = type & ~(VMI_PAGE_ZEROED | VMI_PAGE_CLONE); page 347 arch/x86/kernel/vmi_32.c check_zeroed_page(pfn, type, page); page 361 arch/x86/kernel/vmi_32.c struct page *page = pfn_to_page(pfn); page 362 arch/x86/kernel/vmi_32.c BUG_ON((page->type ^ type) & VMI_PAGE_PAE); page 363 arch/x86/kernel/vmi_32.c BUG_ON(type == VMI_PAGE_NORMAL && page->type); page 364 arch/x86/kernel/vmi_32.c BUG_ON((type & page->type) == 0); page 375 arch/x86/kernel/vmi_32.c void *va = kmap_atomic(page, type); page 390 arch/x86/kernel/vmi_32.c vmi_ops.set_linear_mapping((type - KM_PTE0)+1, va, 1, page_to_pfn(page)); page 19 arch/x86/kvm/lapic.h struct page *regs_page; page 22 arch/x86/kvm/lapic.h struct page *vapic_page; page 279 arch/x86/kvm/mmu.c struct page *page; page 284 arch/x86/kvm/mmu.c page = alloc_page(GFP_KERNEL); page 285 arch/x86/kvm/mmu.c if (!page) page 287 arch/x86/kvm/mmu.c set_page_private(page, 0); page 288 arch/x86/kvm/mmu.c cache->objects[cache->nobjs++] = page_address(page); page 1048 arch/x86/kvm/mmu.c struct kvm_mmu_page *page; page 1050 arch/x86/kvm/mmu.c page = container_of(kvm->arch.active_mmu_pages.prev, page 1052 arch/x86/kvm/mmu.c kvm_mmu_zap_page(kvm, page); page 1104 arch/x86/kvm/mmu.c struct page *gva_to_page(struct kvm_vcpu *vcpu, gva_t gva) page 1106 arch/x86/kvm/mmu.c struct page *page; page 1114 arch/x86/kvm/mmu.c page = gfn_to_page(vcpu->kvm, gpa >> PAGE_SHIFT); page 1117 arch/x86/kvm/mmu.c return page; page 1999 arch/x86/kvm/mmu.c struct page *page; page 2015 arch/x86/kvm/mmu.c page = alloc_page(GFP_KERNEL | __GFP_DMA32); page 2016 arch/x86/kvm/mmu.c if (!page) page 2018 arch/x86/kvm/mmu.c vcpu->arch.mmu.pae_root = page_address(page); page 2087 arch/x86/kvm/mmu.c struct kvm_mmu_page *page; page 2089 arch/x86/kvm/mmu.c page = container_of(kvm->arch.active_mmu_pages.prev, page 2091 arch/x86/kvm/mmu.c kvm_mmu_zap_page(kvm, page); page 92 arch/x86/kvm/paging_tmpl.h struct page *page; page 95 arch/x86/kvm/paging_tmpl.h page = gfn_to_page(kvm, table_gfn); page 98 arch/x86/kvm/paging_tmpl.h table = kmap_atomic(page, KM_USER0); page 104 arch/x86/kvm/paging_tmpl.h kvm_release_page_dirty(page); page 245 arch/x86/kvm/paging_tmpl.h static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *page, page 260 arch/x86/kvm/paging_tmpl.h pte_access = page->role.access & FNAME(gpte_access)(vcpu, gpte); page 269 arch/x86/kvm/paging_tmpl.h mmu_set_spte(vcpu, spte, page->role.access, pte_access, 0, 0, page 91 arch/x86/kvm/svm.c struct page *save_area; page 421 arch/x86/kvm/svm.c struct page *iopm_pages; page 629 arch/x86/kvm/svm.c struct page *page; page 630 arch/x86/kvm/svm.c struct page *msrpm_pages; page 643 arch/x86/kvm/svm.c page = alloc_page(GFP_KERNEL); page 644 arch/x86/kvm/svm.c if (!page) { page 656 arch/x86/kvm/svm.c svm->vmcb = page_address(page); page 658 arch/x86/kvm/svm.c svm->vmcb_pa = page_to_pfn(page) << PAGE_SHIFT; page 100 arch/x86/kvm/vmx.c static struct page *vmx_io_bitmap_a; page 101 arch/x86/kvm/vmx.c static struct page *vmx_io_bitmap_b; page 102 arch/x86/kvm/vmx.c static struct page *vmx_msr_bitmap; page 1208 arch/x86/kvm/vmx.c struct page *pages; page 2006 arch/x86/kvm/x86.c struct page *page; page 2022 arch/x86/kvm/x86.c page = gfn_to_page(vcpu->kvm, gpa >> PAGE_SHIFT); page 2025 arch/x86/kvm/x86.c kaddr = kmap_atomic(page, KM_USER0); page 2028 arch/x86/kvm/x86.c kvm_release_page_dirty(page); page 2358 arch/x86/kvm/x86.c struct page *page; page 2414 arch/x86/kvm/x86.c page = gva_to_page(vcpu, address + i * PAGE_SIZE); page 2415 arch/x86/kvm/x86.c vcpu->arch.pio.guest_pages[i] = page; page 2416 arch/x86/kvm/x86.c if (!page) { page 2774 arch/x86/kvm/x86.c struct page *page; page 2780 arch/x86/kvm/x86.c page = gfn_to_page(vcpu->kvm, apic->vapic_addr >> PAGE_SHIFT); page 2783 arch/x86/kvm/x86.c vcpu->arch.apic->vapic_page = page; page 3864 arch/x86/kvm/x86.c struct page *page; page 3877 arch/x86/kvm/x86.c page = alloc_page(GFP_KERNEL | __GFP_ZERO); page 3878 arch/x86/kvm/x86.c if (!page) { page 3882 arch/x86/kvm/x86.c vcpu->arch.pio_data = page_address(page); page 145 arch/x86/lib/mmx_32.c : : "r" (page) : "memory"); page 146 arch/x86/lib/mmx_32.c page += 64; page 273 arch/x86/lib/mmx_32.c : : "r" (page) : "memory"); page 274 arch/x86/lib/mmx_32.c page += 128; page 345 arch/x86/lib/mmx_32.c :"a" (0), "1" (page), "0" (1024) page 352 arch/x86/lib/mmx_32.c slow_zero_page(page); page 354 arch/x86/lib/mmx_32.c fast_clear_page(page); page 741 arch/x86/lib/usercopy_32.c struct page *pg; page 198 arch/x86/mm/fault.c __typeof__(pte_val(__pte(0))) page; page 200 arch/x86/mm/fault.c page = read_cr3(); page 201 arch/x86/mm/fault.c page = ((__typeof__(page) *) __va(page))[address >> PGDIR_SHIFT]; page 203 arch/x86/mm/fault.c printk("*pdpt = %016Lx ", page); page 204 arch/x86/mm/fault.c if ((page >> PAGE_SHIFT) < max_low_pfn page 205 arch/x86/mm/fault.c && page & _PAGE_PRESENT) { page 206 arch/x86/mm/fault.c page &= PAGE_MASK; page 207 arch/x86/mm/fault.c page = ((__typeof__(page) *) __va(page))[(address >> PMD_SHIFT) page 209 arch/x86/mm/fault.c printk(KERN_CONT "*pde = %016Lx ", page); page 210 arch/x86/mm/fault.c page &= ~_PAGE_NX; page 213 arch/x86/mm/fault.c printk("*pde = %08lx ", page); page 222 arch/x86/mm/fault.c if ((page >> PAGE_SHIFT) < max_low_pfn page 223 arch/x86/mm/fault.c && (page & _PAGE_PRESENT) page 224 arch/x86/mm/fault.c && !(page & _PAGE_PSE)) { page 225 arch/x86/mm/fault.c page &= PAGE_MASK; page 226 arch/x86/mm/fault.c page = ((__typeof__(page) *) __va(page))[(address >> PAGE_SHIFT) page 228 arch/x86/mm/fault.c printk("*pte = %0*Lx ", sizeof(page)*2, (u64)page); page 922 arch/x86/mm/fault.c struct page *page; page 925 arch/x86/mm/fault.c list_for_each_entry(page, &pgd_list, lru) { page 926 arch/x86/mm/fault.c if (!vmalloc_sync_one(page_address(page), page 937 arch/x86/mm/fault.c struct page *page; page 942 arch/x86/mm/fault.c list_for_each_entry(page, &pgd_list, lru) { page 944 arch/x86/mm/fault.c pgd = (pgd_t *)page_address(page) + pgd_index(address); page 83 arch/x86/mm/gup.c struct page *page; page 90 arch/x86/mm/gup.c page = pte_page(pte); page 91 arch/x86/mm/gup.c get_page(page); page 92 arch/x86/mm/gup.c pages[*nr] = page; page 103 arch/x86/mm/gup.c VM_BUG_ON(page != compound_head(page)); page 104 arch/x86/mm/gup.c VM_BUG_ON(page_count(page) == 0); page 105 arch/x86/mm/gup.c atomic_add(nr, &page->_count); page 113 arch/x86/mm/gup.c struct page *head, *page; page 127 arch/x86/mm/gup.c page = head + ((addr & ~PMD_MASK) >> PAGE_SHIFT); page 129 arch/x86/mm/gup.c VM_BUG_ON(compound_head(page) != head); page 130 arch/x86/mm/gup.c pages[*nr] = page; page 132 arch/x86/mm/gup.c page++; page 170 arch/x86/mm/gup.c struct page *head, *page; page 184 arch/x86/mm/gup.c page = head + ((addr & ~PUD_MASK) >> PAGE_SHIFT); page 186 arch/x86/mm/gup.c VM_BUG_ON(compound_head(page) != head); page 187 arch/x86/mm/gup.c pages[*nr] = page; page 189 arch/x86/mm/gup.c page++; page 7 arch/x86/mm/highmem_32.c if (!PageHighMem(page)) page 8 arch/x86/mm/highmem_32.c return page_address(page); page 9 arch/x86/mm/highmem_32.c return kmap_high(page); page 16 arch/x86/mm/highmem_32.c if (!PageHighMem(page)) page 18 arch/x86/mm/highmem_32.c kunmap_high(page); page 80 arch/x86/mm/highmem_32.c if (!PageHighMem(page)) page 81 arch/x86/mm/highmem_32.c return page_address(page); page 88 arch/x86/mm/highmem_32.c set_pte(kmap_pte-idx, mk_pte(page, prot)); page 96 arch/x86/mm/highmem_32.c return kmap_atomic_prot(page, type, kmap_prot); page 141 arch/x86/mm/highmem_32.c struct page *kmap_atomic_to_page(void *ptr) page 170 arch/x86/mm/hugetlbpage.c struct page * page 176 arch/x86/mm/hugetlbpage.c struct page *page; page 188 arch/x86/mm/hugetlbpage.c page = &pte_page(*pte)[vpfn % (HPAGE_SIZE/PAGE_SIZE)]; page 190 arch/x86/mm/hugetlbpage.c WARN_ON(!PageHead(page)); page 192 arch/x86/mm/hugetlbpage.c return page; page 205 arch/x86/mm/hugetlbpage.c struct page * page 214 arch/x86/mm/hugetlbpage.c struct page * page 230 arch/x86/mm/hugetlbpage.c struct page * page 234 arch/x86/mm/hugetlbpage.c struct page *page; page 236 arch/x86/mm/hugetlbpage.c page = pte_page(*(pte_t *)pmd); page 237 arch/x86/mm/hugetlbpage.c if (page) page 238 arch/x86/mm/hugetlbpage.c page += ((address & ~PMD_MASK) >> PAGE_SHIFT); page 239 arch/x86/mm/hugetlbpage.c return page; page 242 arch/x86/mm/hugetlbpage.c struct page * page 246 arch/x86/mm/hugetlbpage.c struct page *page; page 248 arch/x86/mm/hugetlbpage.c page = pte_page(*(pte_t *)pud); page 249 arch/x86/mm/hugetlbpage.c if (page) page 250 arch/x86/mm/hugetlbpage.c page += ((address & ~PUD_MASK) >> PAGE_SHIFT); page 251 arch/x86/mm/hugetlbpage.c return page; page 380 arch/x86/mm/init_32.c ClearPageReserved(page); page 381 arch/x86/mm/init_32.c init_page_count(page); page 382 arch/x86/mm/init_32.c __free_page(page); page 395 arch/x86/mm/init_32.c struct page *page; page 410 arch/x86/mm/init_32.c page = pfn_to_page(node_pfn); page 411 arch/x86/mm/init_32.c add_one_highpage_init(page, node_pfn); page 34 arch/x86/mm/kmmio.c unsigned long page; /* location of the fault page */ page 68 arch/x86/mm/kmmio.c return &kmmio_page_table[hash_long(page, KMMIO_PAGE_HASH_BITS)]; page 99 arch/x86/mm/kmmio.c page &= PAGE_MASK; page 100 arch/x86/mm/kmmio.c head = kmmio_page_list(page); page 102 arch/x86/mm/kmmio.c if (p->page == page) page 152 arch/x86/mm/kmmio.c set_page_present(page & PAGE_MASK, false, pglevel); page 158 arch/x86/mm/kmmio.c set_page_present(page & PAGE_MASK, true, pglevel); page 205 arch/x86/mm/kmmio.c disarm_kmmio_fault_page(faultpage->page, NULL); page 247 arch/x86/mm/kmmio.c disarm_kmmio_fault_page(ctx->fpage->page, NULL); page 286 arch/x86/mm/kmmio.c arm_kmmio_fault_page(ctx->fpage->page, NULL); page 314 arch/x86/mm/kmmio.c page &= PAGE_MASK; page 315 arch/x86/mm/kmmio.c f = get_kmmio_fault_page(page); page 318 arch/x86/mm/kmmio.c arm_kmmio_fault_page(f->page, NULL); page 328 arch/x86/mm/kmmio.c f->page = page; page 329 arch/x86/mm/kmmio.c list_add_rcu(&f->list, kmmio_page_list(f->page)); page 331 arch/x86/mm/kmmio.c arm_kmmio_fault_page(f->page, NULL); page 342 arch/x86/mm/kmmio.c page &= PAGE_MASK; page 343 arch/x86/mm/kmmio.c f = get_kmmio_fault_page(page); page 350 arch/x86/mm/kmmio.c disarm_kmmio_fault_page(f->page, NULL); page 97 arch/x86/mm/numa_32.c return (nr_pages + 1) * sizeof(struct page); page 70 arch/x86/mm/pageattr.c int n = sprintf(page, "DirectMap4k: %8lu kB\n", page 73 arch/x86/mm/pageattr.c n += sprintf(page + n, "DirectMap2M: %8lu kB\n", page 76 arch/x86/mm/pageattr.c n += sprintf(page + n, "DirectMap4M: %8lu kB\n", page 81 arch/x86/mm/pageattr.c n += sprintf(page + n, "DirectMap1G: %8lu kB\n", page 329 arch/x86/mm/pageattr.c struct page *page; page 331 arch/x86/mm/pageattr.c list_for_each_entry(page, &pgd_list, lru) { page 336 arch/x86/mm/pageattr.c pgd = (pgd_t *)page_address(page) + pgd_index(address); page 464 arch/x86/mm/pageattr.c struct page *base; page 1005 arch/x86/mm/pageattr.c unsigned long addr = (unsigned long)page_address(page); page 1013 arch/x86/mm/pageattr.c unsigned long addr = (unsigned long)page_address(page); page 1021 arch/x86/mm/pageattr.c unsigned long addr = (unsigned long)page_address(page); page 1029 arch/x86/mm/pageattr.c unsigned long addr = (unsigned long)page_address(page); page 1037 arch/x86/mm/pageattr.c unsigned long addr = (unsigned long)page_address(page); page 1044 arch/x86/mm/pageattr.c unsigned long addr = (unsigned long)page_address(page); page 1053 arch/x86/mm/pageattr.c unsigned long tempaddr = (unsigned long) page_address(page); page 1071 arch/x86/mm/pageattr.c unsigned long tempaddr = (unsigned long) page_address(page); page 1089 arch/x86/mm/pageattr.c if (PageHighMem(page)) page 1092 arch/x86/mm/pageattr.c debug_check_no_locks_freed(page_address(page), page 1108 arch/x86/mm/pageattr.c __set_pages_p(page, numpages); page 1110 arch/x86/mm/pageattr.c __set_pages_np(page, numpages); page 1126 arch/x86/mm/pageattr.c if (PageHighMem(page)) page 1129 arch/x86/mm/pageattr.c pte = lookup_address((unsigned long)page_address(page), &level); page 229 arch/x86/mm/pat.c struct page *page; page 233 arch/x86/mm/pat.c page = pfn_to_page(pfn); page 234 arch/x86/mm/pat.c if (page_mapped(page) || PageNonWB(page)) page 237 arch/x86/mm/pat.c SetPageNonWB(page); page 244 arch/x86/mm/pat.c page = pfn_to_page(pfn); page 245 arch/x86/mm/pat.c ClearPageNonWB(page); page 253 arch/x86/mm/pat.c struct page *page; page 257 arch/x86/mm/pat.c page = pfn_to_page(pfn); page 258 arch/x86/mm/pat.c if (page_mapped(page) || !PageNonWB(page)) page 261 arch/x86/mm/pat.c ClearPageNonWB(page); page 268 arch/x86/mm/pat.c page = pfn_to_page(pfn); page 269 arch/x86/mm/pat.c SetPageNonWB(page); page 14 arch/x86/mm/pgtable.c struct page *pte; page 51 arch/x86/mm/pgtable.c struct page *page = virt_to_page(pgd); page 53 arch/x86/mm/pgtable.c list_add(&page->lru, &pgd_list); page 58 arch/x86/mm/pgtable.c struct page *page = virt_to_page(pgd); page 60 arch/x86/mm/pgtable.c list_del(&page->lru); page 487 arch/x86/mm/srat_64.c total_mb *= sizeof(struct page); page 359 arch/x86/pci/pcbios.c unsigned long page; page 363 arch/x86/pci/pcbios.c page = __get_free_page(GFP_KERNEL); page 364 arch/x86/pci/pcbios.c if (!page) page 366 arch/x86/pci/pcbios.c opt.table = (struct irq_info *) page; page 397 arch/x86/pci/pcbios.c memcpy(rt->slots, (void *) page, opt.size); page 401 arch/x86/pci/pcbios.c free_page(page); page 196 arch/x86/vdso/vdso32-setup.c static struct page *vdso32_pages[1]; page 24 arch/x86/vdso/vma.c static struct page **vdso_pages; page 43 arch/x86/vdso/vma.c vdso_pages = kmalloc(sizeof(struct page *) * npages, GFP_KERNEL); page 47 arch/x86/vdso/vma.c struct page *p; page 262 arch/x86/xen/enlighten.c struct page *page; page 268 arch/x86/xen/enlighten.c page = pfn_to_page(pfn); page 275 arch/x86/xen/enlighten.c if (!PageHighMem(page)) { page 861 arch/x86/xen/enlighten.c struct page *page = pfn_to_page(pfn); page 864 arch/x86/xen/enlighten.c SetPagePinned(page); page 866 arch/x86/xen/enlighten.c if (!PageHighMem(page)) { page 896 arch/x86/xen/enlighten.c struct page *page = virt_to_page(pgd); page 899 arch/x86/xen/enlighten.c BUG_ON(page->private != 0); page 904 arch/x86/xen/enlighten.c page->private = (unsigned long)user_pgd; page 932 arch/x86/xen/enlighten.c struct page *page = pfn_to_page(pfn); page 934 arch/x86/xen/enlighten.c if (PagePinned(page)) { page 935 arch/x86/xen/enlighten.c if (!PageHighMem(page)) { page 940 arch/x86/xen/enlighten.c ClearPagePinned(page); page 971 arch/x86/xen/enlighten.c if (PagePinned(page)) page 974 arch/x86/xen/enlighten.c if (0 && PageHighMem(page)) page 976 arch/x86/xen/enlighten.c page_to_pfn(page), type, page 979 arch/x86/xen/enlighten.c return kmap_atomic_prot(page, type, prot); page 290 arch/x86/xen/mmu.c struct page *page = virt_to_page(ptr); page 292 arch/x86/xen/mmu.c return PagePinned(page); page 569 arch/x86/xen/mmu.c struct page *page = virt_to_page(pgd_page); page 570 arch/x86/xen/mmu.c user_ptr = (pgd_t *)page->private; page 722 arch/x86/xen/mmu.c struct page *pte; page 753 arch/x86/xen/mmu.c ptl = __pte_lockptr(page); page 781 arch/x86/xen/mmu.c unsigned pgfl = TestSetPagePinned(page); page 786 arch/x86/xen/mmu.c else if (PageHighMem(page)) page 791 arch/x86/xen/mmu.c void *pt = lowmem_page_address(page); page 792 arch/x86/xen/mmu.c unsigned long pfn = page_to_pfn(page); page 820 arch/x86/xen/mmu.c ptl = xen_pte_lock(page, mm); page 892 arch/x86/xen/mmu.c struct page *page; page 896 arch/x86/xen/mmu.c list_for_each_entry(page, &pgd_list, lru) { page 897 arch/x86/xen/mmu.c if (!PagePinned(page)) { page 898 arch/x86/xen/mmu.c __xen_pgd_pin(&init_mm, (pgd_t *)page_address(page)); page 899 arch/x86/xen/mmu.c SetPageSavePinned(page); page 914 arch/x86/xen/mmu.c SetPagePinned(page); page 926 arch/x86/xen/mmu.c unsigned pgfl = TestClearPagePinned(page); page 928 arch/x86/xen/mmu.c if (pgfl && !PageHighMem(page)) { page 929 arch/x86/xen/mmu.c void *pt = lowmem_page_address(page); page 930 arch/x86/xen/mmu.c unsigned long pfn = page_to_pfn(page); page 942 arch/x86/xen/mmu.c ptl = xen_pte_lock(page, mm); page 1004 arch/x86/xen/mmu.c struct page *page; page 1008 arch/x86/xen/mmu.c list_for_each_entry(page, &pgd_list, lru) { page 1009 arch/x86/xen/mmu.c if (PageSavePinned(page)) { page 1010 arch/x86/xen/mmu.c BUG_ON(!PagePinned(page)); page 1011 arch/x86/xen/mmu.c __xen_pgd_unpin(&init_mm, (pgd_t *)page_address(page)); page 1012 arch/x86/xen/mmu.c ClearPageSavePinned(page); page 1400 block/as-iosched.c return sprintf(page, "%d\n", var); page 1406 block/as-iosched.c char *p = (char *) page; page 1417 block/as-iosched.c pos += sprintf(page+pos, "%lu %% exit probability\n", page 1419 block/as-iosched.c pos += sprintf(page+pos, "%lu %% probability of exiting without a " page 1422 block/as-iosched.c pos += sprintf(page+pos, "%lu ms new thinktime\n", ad->new_ttime_mean); page 1423 block/as-iosched.c pos += sprintf(page+pos, "%llu sectors new seek distance\n", page 1430 block/as-iosched.c static ssize_t __FUNC(elevator_t *e, char *page) \ page 1433 block/as-iosched.c return as_var_show(jiffies_to_msecs((__VAR)), (page)); \ page 1443 block/as-iosched.c static ssize_t __FUNC(elevator_t *e, const char *page, size_t count) \ page 1446 block/as-iosched.c int ret = as_var_store(__PTR, (page), count); \ page 178 block/blk-integrity.c return entry->show(bi, page); page 192 block/blk-integrity.c ret = entry->store(bi, page, count); page 200 block/blk-integrity.c return sprintf(page, "%s\n", bi->name); page 202 block/blk-integrity.c return sprintf(page, "none\n"); page 208 block/blk-integrity.c return sprintf(page, "%u\n", bi->tag_size); page 210 block/blk-integrity.c return sprintf(page, "0\n"); page 216 block/blk-integrity.c char *p = (char *) page; page 229 block/blk-integrity.c return sprintf(page, "%d\n", (bi->flags & INTEGRITY_FLAG_READ) != 0); page 235 block/blk-integrity.c char *p = (char *) page; page 248 block/blk-integrity.c return sprintf(page, "%d\n", (bi->flags & INTEGRITY_FLAG_WRITE) != 0); page 21 block/blk-sysfs.c return sprintf(page, "%d\n", var); page 27 block/blk-sysfs.c char *p = (char *) page; page 35 block/blk-sysfs.c return queue_var_show(q->nr_requests, (page)); page 43 block/blk-sysfs.c int ret = queue_var_store(&nr, page, count); page 82 block/blk-sysfs.c return queue_var_show(ra_kb, (page)); page 89 block/blk-sysfs.c ssize_t ret = queue_var_store(&ra_kb, page, count); page 102 block/blk-sysfs.c return queue_var_show(max_sectors_kb, (page)); page 107 block/blk-sysfs.c return queue_var_show(q->hardsect_size, page); page 116 block/blk-sysfs.c ssize_t ret = queue_var_store(&max_sectors_kb, page, count); page 135 block/blk-sysfs.c return queue_var_show(max_hw_sectors_kb, (page)); page 140 block/blk-sysfs.c return queue_var_show(blk_queue_nomerges(q), page); page 147 block/blk-sysfs.c ssize_t ret = queue_var_store(&nm, page, count); page 163 block/blk-sysfs.c return queue_var_show(set != 0, page); page 173 block/blk-sysfs.c ret = queue_var_store(&val, page, count); page 259 block/blk-sysfs.c res = entry->show(q, page); page 281 block/blk-sysfs.c res = entry->store(q, page, length); page 2278 block/cfq-iosched.c return sprintf(page, "%d\n", var); page 2284 block/cfq-iosched.c char *p = (char *) page; page 2291 block/cfq-iosched.c static ssize_t __FUNC(elevator_t *e, char *page) \ page 2297 block/cfq-iosched.c return cfq_var_show(__data, (page)); \ page 2311 block/cfq-iosched.c static ssize_t __FUNC(elevator_t *e, const char *page, size_t count) \ page 2315 block/cfq-iosched.c int ret = cfq_var_store(&__data, (page), count); \ page 57 block/cmd-filter.c char *npage = page; page 74 block/cmd-filter.c if (npage != page) page 77 block/cmd-filter.c return npage - page; page 82 block/cmd-filter.c return rcf_cmds_show(filter, page, READ); page 88 block/cmd-filter.c return rcf_cmds_show(filter, page, WRITE); page 106 block/cmd-filter.c while ((p = strsep((char **)&page, " ")) != NULL) { page 135 block/cmd-filter.c return rcf_cmds_store(filter, page, count, READ); page 141 block/cmd-filter.c return rcf_cmds_store(filter, page, count, WRITE); page 178 block/cmd-filter.c return entry->show(filter, page); page 197 block/cmd-filter.c return entry->store(filter, page, length); page 377 block/deadline-iosched.c return sprintf(page, "%d\n", var); page 383 block/deadline-iosched.c char *p = (char *) page; page 390 block/deadline-iosched.c static ssize_t __FUNC(elevator_t *e, char *page) \ page 396 block/deadline-iosched.c return deadline_var_show(__data, (page)); \ page 406 block/deadline-iosched.c static ssize_t __FUNC(elevator_t *e, const char *page, size_t count) \ page 410 block/deadline-iosched.c int ret = deadline_var_store(&__data, (page), count); \ page 971 block/elevator.c error = e->ops ? entry->show(e, page) : -ENOENT; page 988 block/elevator.c error = e->ops ? entry->store(e, page, length) : -ENOENT; page 306 crypto/async_tx/async_xor.c BUILD_BUG_ON(sizeof(dma_addr_t) > sizeof(struct page *)); page 151 crypto/authenc.c struct page *dstp; page 280 crypto/authenc.c struct page *srcp; page 89 crypto/blkcipher.c memcpy(walk->dst.virt.addr, walk->page, n); page 138 crypto/blkcipher.c if (walk->buffer != walk->page) page 140 crypto/blkcipher.c if (walk->page) page 141 crypto/blkcipher.c free_page((unsigned long)walk->page); page 158 crypto/blkcipher.c walk->buffer = walk->page; page 185 crypto/blkcipher.c u8 *tmp = walk->page; page 202 crypto/blkcipher.c walk->src.phys.page = scatterwalk_page(&walk->in); page 204 crypto/blkcipher.c walk->dst.phys.page = scatterwalk_page(&walk->out); page 211 crypto/blkcipher.c diff |= walk->src.virt.page - walk->dst.virt.page; page 244 crypto/blkcipher.c if (!walk->page) { page 245 crypto/blkcipher.c walk->page = (void *)__get_free_page(GFP_ATOMIC); page 246 crypto/blkcipher.c if (!walk->page) page 270 crypto/blkcipher.c walk->src.phys.page = virt_to_page(walk->src.virt.addr); page 271 crypto/blkcipher.c walk->dst.phys.page = virt_to_page(walk->dst.virt.addr); page 344 crypto/blkcipher.c walk->page = NULL; page 45 crypto/digest.c struct page *pg = sg_page(sg); page 89 crypto/eseqiv.c struct page *srcp; page 90 crypto/eseqiv.c struct page *dstp; page 54 crypto/scatterwalk.c struct page *page; page 56 crypto/scatterwalk.c page = sg_page(walk->sg) + ((walk->offset - 1) >> PAGE_SHIFT); page 57 crypto/scatterwalk.c flush_dcache_page(page); page 121 crypto/xcbc.c struct page *pg = sg_page(sg); page 60 fs/9p/vfs_addr.c buffer = kmap(page); page 61 fs/9p/vfs_addr.c offset = page_offset(page); page 68 fs/9p/vfs_addr.c flush_dcache_page(page); page 69 fs/9p/vfs_addr.c SetPageUptodate(page); page 73 fs/9p/vfs_addr.c kunmap(page); page 74 fs/9p/vfs_addr.c unlock_page(page); page 56 fs/adfs/inode.c return block_write_full_page(page, adfs_get_block, wbc); page 61 fs/adfs/inode.c return block_read_full_page(page, adfs_get_block); page 397 fs/affs/file.c return block_write_full_page(page, affs_get_block, wbc); page 402 fs/affs/file.c return block_read_full_page(page, affs_get_block); page 489 fs/affs/file.c struct inode *inode = page->mapping->host; page 496 fs/affs/file.c pr_debug("AFFS: read_page(%u, %ld, %d, %d)\n", (u32)inode->i_ino, page->index, from, to); page 498 fs/affs/file.c kmap(page); page 499 fs/affs/file.c data = page_address(page); page 501 fs/affs/file.c tmp = (page->index << PAGE_CACHE_SHIFT) + from; page 517 fs/affs/file.c flush_dcache_page(page); page 518 fs/affs/file.c kunmap(page); page 593 fs/affs/file.c struct inode *inode = page->mapping->host; page 597 fs/affs/file.c pr_debug("AFFS: read_page(%u, %ld)\n", (u32)inode->i_ino, page->index); page 599 fs/affs/file.c if (((page->index + 1) << PAGE_CACHE_SHIFT) > inode->i_size) { page 601 fs/affs/file.c memset(page_address(page) + to, 0, PAGE_CACHE_SIZE - to); page 604 fs/affs/file.c err = affs_do_readpage_ofs(file, page, 0, to); page 606 fs/affs/file.c SetPageUptodate(page); page 607 fs/affs/file.c unlock_page(page); page 616 fs/affs/file.c struct page *page; page 631 fs/affs/file.c page = __grab_cache_page(mapping, index); page 632 fs/affs/file.c if (!page) page 634 fs/affs/file.c *pagep = page; page 636 fs/affs/file.c if (PageUptodate(page)) page 640 fs/affs/file.c err = affs_do_readpage_ofs(file, page, 0, PAGE_CACHE_SIZE); page 642 fs/affs/file.c unlock_page(page); page 643 fs/affs/file.c page_cache_release(page); page 671 fs/affs/file.c data = page_address(page); page 675 fs/affs/file.c tmp = (page->index << PAGE_CACHE_SHIFT) + from; page 757 fs/affs/file.c SetPageUptodate(page); page 761 fs/affs/file.c tmp = (page->index << PAGE_CACHE_SHIFT) + from; page 765 fs/affs/file.c unlock_page(page); page 766 fs/affs/file.c page_cache_release(page); page 824 fs/affs/file.c struct page *page; page 829 fs/affs/file.c res = mapping->a_ops->write_begin(NULL, mapping, size, 0, 0, &page, &fsdata); page 831 fs/affs/file.c res = mapping->a_ops->write_end(NULL, mapping, size, 0, 0, page, fsdata); page 16 fs/affs/symlink.c struct inode *inode = page->mapping->host; page 17 fs/affs/symlink.c char *link = kmap(page); page 58 fs/affs/symlink.c SetPageUptodate(page); page 59 fs/affs/symlink.c kunmap(page); page 60 fs/affs/symlink.c unlock_page(page); page 63 fs/affs/symlink.c SetPageError(page); page 64 fs/affs/symlink.c kunmap(page); page 65 fs/affs/symlink.c unlock_page(page); page 141 fs/afs/dir.c if (page->index == 0 && qty != ntohs(dbuf->blocks[0].pagehdr.npages)) { page 150 fs/afs/dir.c latter = dir->i_size - page_offset(page); page 158 fs/afs/dir.c dbuf = page_address(page); page 168 fs/afs/dir.c SetPageChecked(page); page 172 fs/afs/dir.c SetPageChecked(page); page 173 fs/afs/dir.c SetPageError(page); page 181 fs/afs/dir.c kunmap(page); page 182 fs/afs/dir.c page_cache_release(page); page 188 fs/afs/dir.c static struct page *afs_dir_get_page(struct inode *dir, unsigned long index, page 191 fs/afs/dir.c struct page *page; page 198 fs/afs/dir.c page = read_mapping_page(dir->i_mapping, index, &file); page 199 fs/afs/dir.c if (!IS_ERR(page)) { page 200 fs/afs/dir.c kmap(page); page 201 fs/afs/dir.c if (!PageChecked(page)) page 202 fs/afs/dir.c afs_dir_check_page(dir, page); page 203 fs/afs/dir.c if (PageError(page)) page 206 fs/afs/dir.c return page; page 209 fs/afs/dir.c afs_dir_put_page(page); page 334 fs/afs/dir.c struct page *page; page 355 fs/afs/dir.c page = afs_dir_get_page(dir, blkoff / PAGE_SIZE, key); page 356 fs/afs/dir.c if (IS_ERR(page)) { page 357 fs/afs/dir.c ret = PTR_ERR(page); page 363 fs/afs/dir.c dbuf = page_address(page); page 372 fs/afs/dir.c afs_dir_put_page(page); page 380 fs/afs/dir.c afs_dir_put_page(page); page 110 fs/afs/file.c _enter("%p,%p,%p,%d", cookie_data, page, data, error); page 113 fs/afs/file.c SetPageError(page); page 115 fs/afs/file.c SetPageUptodate(page); page 116 fs/afs/file.c unlock_page(page); page 130 fs/afs/file.c _enter("%p,%p,%p,%d", cookie_data, page, data, error); page 132 fs/afs/file.c unlock_page(page); page 148 fs/afs/file.c inode = page->mapping->host; page 154 fs/afs/file.c _enter("{%x},{%lu},{%lu}", key_serial(key), inode->i_ino, page->index); page 158 fs/afs/file.c BUG_ON(!PageLocked(page)); page 167 fs/afs/file.c page, page 188 fs/afs/file.c offset = page->index << PAGE_CACHE_SHIFT; page 193 fs/afs/file.c ret = afs_vnode_fetch_data(vnode, key, offset, len, page); page 202 fs/afs/file.c cachefs_uncache_page(vnode->cache, page); page 207 fs/afs/file.c SetPageUptodate(page); page 211 fs/afs/file.c page, page 216 fs/afs/file.c cachefs_uncache_page(vnode->cache, page); page 217 fs/afs/file.c unlock_page(page); page 220 fs/afs/file.c unlock_page(page); page 228 fs/afs/file.c SetPageError(page); page 229 fs/afs/file.c unlock_page(page); page 241 fs/afs/file.c _enter("{%lu},%lu", page->index, offset); page 243 fs/afs/file.c BUG_ON(!PageLocked(page)); page 245 fs/afs/file.c if (PagePrivate(page)) { page 252 fs/afs/file.c BUG_ON(!PageLocked(page)); page 255 fs/afs/file.c if (!PageWriteback(page)) page 256 fs/afs/file.c ret = page->mapping->a_ops->releasepage(page, page 270 fs/afs/file.c _enter("{%lu}", page->index); page 280 fs/afs/file.c struct afs_vnode *vnode = AFS_FS_I(page->mapping->host); page 284 fs/afs/file.c vnode->fid.vid, vnode->fid.vnode, page->index, page->flags, page 287 fs/afs/file.c if (PagePrivate(page)) { page 288 fs/afs/file.c wb = (struct afs_writeback *) page_private(page); page 290 fs/afs/file.c set_page_private(page, 0); page 291 fs/afs/file.c ClearPagePrivate(page); page 309 fs/afs/fsclient.c struct page *page; page 365 fs/afs/fsclient.c page = call->reply3; page 366 fs/afs/fsclient.c buffer = kmap_atomic(page, KM_USER0); page 411 fs/afs/fsclient.c page = call->reply3; page 412 fs/afs/fsclient.c buffer = kmap_atomic(page, KM_USER0); page 56 fs/afs/mntpt.c struct page *page; page 65 fs/afs/mntpt.c page = read_mapping_page(AFS_VNODE_TO_I(vnode)->i_mapping, 0, &file); page 66 fs/afs/mntpt.c if (IS_ERR(page)) { page 67 fs/afs/mntpt.c ret = PTR_ERR(page); page 72 fs/afs/mntpt.c if (PageError(page)) page 75 fs/afs/mntpt.c buf = kmap(page); page 93 fs/afs/mntpt.c kunmap(page); page 95 fs/afs/mntpt.c page_cache_release(page); page 142 fs/afs/mntpt.c struct page *page = NULL; page 166 fs/afs/mntpt.c page = read_mapping_page(mntpt->d_inode->i_mapping, 0, NULL); page 167 fs/afs/mntpt.c if (IS_ERR(page)) { page 168 fs/afs/mntpt.c ret = PTR_ERR(page); page 173 fs/afs/mntpt.c if (PageError(page)) page 176 fs/afs/mntpt.c buf = kmap(page); page 178 fs/afs/mntpt.c kunmap(page); page 179 fs/afs/mntpt.c page_cache_release(page); page 180 fs/afs/mntpt.c page = NULL; page 200 fs/afs/mntpt.c if (page) page 201 fs/afs/mntpt.c page_cache_release(page); page 245 fs/afs/rxrpc.c struct page *pages[8]; page 425 fs/afs/vnode.c page, &afs_sync_call); page 28 fs/afs/write.c return __set_page_dirty_nobuffers(page); page 95 fs/afs/write.c ret = afs_vnode_fetch_data(vnode, key, start, len, page); page 125 fs/afs/write.c p = kmap_atomic(page, KM_USER0); page 128 fs/afs/write.c pos = (loff_t) page->index << PAGE_SHIFT; page 166 fs/afs/write.c ret = afs_fill_page(vnode, key, start, len, page); page 188 fs/afs/write.c vnode->fid.vid, vnode->fid.vnode, page->index, offset, to); page 194 fs/afs/write.c candidate->first = candidate->last = page->index; page 201 fs/afs/write.c if (!PageUptodate(page)) { page 203 fs/afs/write.c ret = afs_prepare_page(vnode, page, key, offset, to); page 212 fs/afs/write.c index = page->index; page 217 fs/afs/write.c wb = (struct afs_writeback *) page_private(page); page 237 fs/afs/write.c SetPagePrivate(page); page 238 fs/afs/write.c set_page_private(page, (unsigned long) candidate); page 260 fs/afs/write.c SetPagePrivate(page); page 261 fs/afs/write.c set_page_private(page, (unsigned long) wb); page 273 fs/afs/write.c if (PageDirty(page)) { page 274 fs/afs/write.c ret = afs_write_back_from_locked_page(wb, page); page 284 fs/afs/write.c set_page_private(page, 0); page 285 fs/afs/write.c ClearPagePrivate(page); page 299 fs/afs/write.c vnode->fid.vid, vnode->fid.vnode, page->index, offset, to); page 301 fs/afs/write.c maybe_i_size = (loff_t) page->index << PAGE_SHIFT; page 313 fs/afs/write.c SetPageUptodate(page); page 314 fs/afs/write.c set_page_dirty(page); page 315 fs/afs/write.c if (PageDirty(page)) page 365 fs/afs/write.c struct page *pages[8], *page; page 404 fs/afs/write.c page = pages[loop]; page 405 fs/afs/write.c if (page->index > wb->last) page 407 fs/afs/write.c if (!trylock_page(page)) page 409 fs/afs/write.c if (!PageDirty(page) || page 410 fs/afs/write.c page_private(page) != (unsigned long) wb) { page 411 fs/afs/write.c unlock_page(page); page 414 fs/afs/write.c if (!clear_page_dirty_for_io(page)) page 416 fs/afs/write.c if (test_set_page_writeback(page)) page 418 fs/afs/write.c unlock_page(page); page 419 fs/afs/write.c put_page(page); page 486 fs/afs/write.c struct backing_dev_info *bdi = page->mapping->backing_dev_info; page 490 fs/afs/write.c _enter("{%lx},", page->index); page 492 fs/afs/write.c wb = (struct afs_writeback *) page_private(page); page 495 fs/afs/write.c ret = afs_write_back_from_locked_page(wb, page); page 496 fs/afs/write.c unlock_page(page); page 519 fs/afs/write.c struct page *page; page 526 fs/afs/write.c 1, &page); page 530 fs/afs/write.c _debug("wback %lx", page->index); page 532 fs/afs/write.c if (page->index > end) { page 534 fs/afs/write.c page_cache_release(page); page 544 fs/afs/write.c lock_page(page); page 546 fs/afs/write.c if (page->mapping != mapping) { page 547 fs/afs/write.c unlock_page(page); page 548 fs/afs/write.c page_cache_release(page); page 553 fs/afs/write.c wait_on_page_writeback(page); page 555 fs/afs/write.c if (PageWriteback(page) || !PageDirty(page)) { page 556 fs/afs/write.c unlock_page(page); page 560 fs/afs/write.c wb = (struct afs_writeback *) page_private(page); page 567 fs/afs/write.c ret = afs_write_back_from_locked_page(wb, page); page 568 fs/afs/write.c unlock_page(page); page 569 fs/afs/write.c page_cache_release(page); page 683 fs/afs/write.c struct page *page = pv.pages[loop]; page 684 fs/afs/write.c end_page_writeback(page); page 685 fs/afs/write.c if (page_private(page) == (unsigned long) wb) { page 686 fs/afs/write.c set_page_private(page, 0); page 687 fs/afs/write.c ClearPagePrivate(page); page 77 fs/aio.c pr_debug("aio_setup: sizeof(struct page) = %d\n", (int)sizeof(struct page)); page 125 fs/aio.c info->ring_pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); page 98 fs/befs/linuxvfs.c return block_read_full_page(page, befs_get_block); page 160 fs/bfs/file.c return block_write_full_page(page, bfs_get_block, wbc); page 165 fs/bfs/file.c return block_read_full_page(page, bfs_get_block); page 2011 fs/binfmt_elf.c struct page *page; page 2015 fs/binfmt_elf.c &page, &tmp_vma) <= 0) { page 2018 fs/binfmt_elf.c if (page == ZERO_PAGE(0)) { page 2020 fs/binfmt_elf.c page_cache_release(page); page 2026 fs/binfmt_elf.c page_to_pfn(page)); page 2027 fs/binfmt_elf.c kaddr = kmap(page); page 2031 fs/binfmt_elf.c kunmap(page); page 2032 fs/binfmt_elf.c page_cache_release(page); page 2035 fs/binfmt_elf.c kunmap(page); page 2037 fs/binfmt_elf.c page_cache_release(page); page 669 fs/binfmt_elf_fdpic.c src = kmap(bprm->page[index]); page 673 fs/binfmt_elf_fdpic.c kunmap(bprm->page[index]); page 1489 fs/binfmt_elf_fdpic.c struct page *page; page 1492 fs/binfmt_elf_fdpic.c &page, &vma) <= 0) { page 1495 fs/binfmt_elf_fdpic.c else if (page == ZERO_PAGE(0)) { page 1496 fs/binfmt_elf_fdpic.c page_cache_release(page); page 1502 fs/binfmt_elf_fdpic.c flush_cache_page(vma, addr, page_to_pfn(page)); page 1503 fs/binfmt_elf_fdpic.c kaddr = kmap(page); page 1507 fs/binfmt_elf_fdpic.c kunmap(page); page 1508 fs/binfmt_elf_fdpic.c page_cache_release(page); page 1511 fs/binfmt_elf_fdpic.c kunmap(page); page 1512 fs/binfmt_elf_fdpic.c page_cache_release(page); page 894 fs/binfmt_flat.c ((char *) page_address(bprm->page[i/PAGE_SIZE]))[i % PAGE_SIZE]; page 447 fs/binfmt_misc.c sprintf(page, "%s\n", status); page 451 fs/binfmt_misc.c sprintf(page, "%s\ninterpreter %s\n", status, e->interpreter); page 452 fs/binfmt_misc.c dp = page + strlen(page); page 475 fs/binfmt_misc.c dp = page + strlen(page); page 540 fs/binfmt_misc.c char *page; page 542 fs/binfmt_misc.c if (!(page = (char*) __get_free_page(GFP_KERNEL))) page 545 fs/binfmt_misc.c entry_status(e, page); page 547 fs/binfmt_misc.c res = simple_read_from_buffer(buf, nbytes, ppos, page, strlen(page)); page 549 fs/binfmt_misc.c free_page((unsigned long) page); page 145 fs/bio-integrity.c iv->bv_page = page; page 368 fs/bio.c if (page == prev->bv_page && page 413 fs/bio.c bvec->bv_page = page; page 470 fs/bio.c return __bio_add_page(q, bio, page, len, offset, q->max_hw_sectors); page 490 fs/bio.c return __bio_add_page(q, bio, page, len, offset, q->max_sectors); page 633 fs/bio.c struct page *page; page 681 fs/bio.c page = map_data->pages[i++]; page 683 fs/bio.c page = alloc_page(q->bounce_gfp | gfp_mask); page 684 fs/bio.c if (!page) { page 689 fs/bio.c if (bio_add_pc_page(q, bio, page, bytes, 0) < bytes) page 752 fs/bio.c struct page **pages; page 779 fs/bio.c pages = kcalloc(nr_pages, sizeof(struct page *), gfp_mask); page 1110 fs/bio.c struct page *page = bvec[i].bv_page; page 1112 fs/bio.c if (page && !PageCompound(page)) page 1113 fs/bio.c set_page_dirty_lock(page); page 1123 fs/bio.c struct page *page = bvec[i].bv_page; page 1125 fs/bio.c if (page) page 1126 fs/bio.c put_page(page); page 1177 fs/bio.c struct page *page = bvec[i].bv_page; page 1179 fs/bio.c if (PageDirty(page) || PageCompound(page)) { page 1180 fs/bio.c page_cache_release(page); page 179 fs/block_dev.c return block_write_full_page(page, blkdev_get_block, wbc); page 184 fs/block_dev.c return block_read_full_page(page, blkdev_get_block); page 201 fs/block_dev.c ret = block_write_end(file, mapping, pos, len, copied, page, fsdata); page 203 fs/block_dev.c unlock_page(page); page 204 fs/block_dev.c page_cache_release(page); page 98 fs/buffer.c ClearPagePrivate(page); page 99 fs/buffer.c set_page_private(page, 0); page 100 fs/buffer.c page_cache_release(page); page 270 fs/buffer.c struct page *page; page 274 fs/buffer.c page = find_get_page(bd_mapping, index); page 275 fs/buffer.c if (!page) page 279 fs/buffer.c if (!page_has_buffers(page)) page 281 fs/buffer.c head = page_buffers(page); page 310 fs/buffer.c page_cache_release(page); page 388 fs/buffer.c struct page *page; page 393 fs/buffer.c page = bh->b_page; page 400 fs/buffer.c SetPageError(page); page 408 fs/buffer.c first = page_buffers(page); page 430 fs/buffer.c if (page_uptodate && !PageError(page)) page 431 fs/buffer.c SetPageUptodate(page); page 432 fs/buffer.c unlock_page(page); page 451 fs/buffer.c struct page *page; page 455 fs/buffer.c page = bh->b_page; page 465 fs/buffer.c set_bit(AS_EIO, &page->mapping->flags); page 468 fs/buffer.c SetPageError(page); page 471 fs/buffer.c first = page_buffers(page); page 487 fs/buffer.c end_page_writeback(page); page 704 fs/buffer.c return !TestSetPageDirty(page); page 706 fs/buffer.c if (TestSetPageDirty(page)) page 710 fs/buffer.c if (page->mapping) { /* Race with truncate? */ page 711 fs/buffer.c WARN_ON_ONCE(warn && !PageUptodate(page)); page 714 fs/buffer.c __inc_zone_page_state(page, NR_FILE_DIRTY); page 720 fs/buffer.c page_index(page), PAGECACHE_TAG_DIRTY); page 755 fs/buffer.c struct address_space *mapping = page_mapping(page); page 758 fs/buffer.c return !TestSetPageDirty(page); page 761 fs/buffer.c if (page_has_buffers(page)) { page 762 fs/buffer.c struct buffer_head *head = page_buffers(page); page 772 fs/buffer.c return __set_page_dirty(page, mapping, 1); page 946 fs/buffer.c set_bh_page(bh, page, offset); page 994 fs/buffer.c attach_page_buffers(page, head); page 1004 fs/buffer.c struct buffer_head *head = page_buffers(page); page 1006 fs/buffer.c int uptodate = PageUptodate(page); page 1027 fs/buffer.c static struct page * page 1032 fs/buffer.c struct page *page; page 1035 fs/buffer.c page = find_or_create_page(inode->i_mapping, index, page 1037 fs/buffer.c if (!page) page 1040 fs/buffer.c BUG_ON(!PageLocked(page)); page 1042 fs/buffer.c if (page_has_buffers(page)) { page 1043 fs/buffer.c bh = page_buffers(page); page 1045 fs/buffer.c init_page_buffers(page, bdev, block, size); page 1046 fs/buffer.c return page; page 1048 fs/buffer.c if (!try_to_free_buffers(page)) page 1055 fs/buffer.c bh = alloc_page_buffers(page, size, 0); page 1065 fs/buffer.c link_dev_buffers(page, bh); page 1066 fs/buffer.c init_page_buffers(page, bdev, block, size); page 1068 fs/buffer.c return page; page 1072 fs/buffer.c unlock_page(page); page 1073 fs/buffer.c page_cache_release(page); page 1084 fs/buffer.c struct page *page; page 1110 fs/buffer.c page = grow_dev_page(bdev, block, index, size); page 1111 fs/buffer.c if (!page) page 1113 fs/buffer.c unlock_page(page); page 1114 fs/buffer.c page_cache_release(page); page 1473 fs/buffer.c bh->b_page = page; page 1475 fs/buffer.c if (PageHighMem(page)) page 1481 fs/buffer.c bh->b_data = page_address(page) + offset; page 1521 fs/buffer.c BUG_ON(!PageLocked(page)); page 1522 fs/buffer.c if (!page_has_buffers(page)) page 1525 fs/buffer.c head = page_buffers(page); page 1546 fs/buffer.c try_to_release_page(page, 0); page 1562 fs/buffer.c head = alloc_page_buffers(page, blocksize, 1); page 1571 fs/buffer.c spin_lock(&page->mapping->private_lock); page 1572 fs/buffer.c if (PageUptodate(page) || PageDirty(page)) { page 1575 fs/buffer.c if (PageDirty(page)) page 1577 fs/buffer.c if (PageUptodate(page)) page 1582 fs/buffer.c attach_page_buffers(page, head); page 1583 fs/buffer.c spin_unlock(&page->mapping->private_lock); page 1654 fs/buffer.c BUG_ON(!PageLocked(page)); page 1658 fs/buffer.c if (!page_has_buffers(page)) { page 1659 fs/buffer.c create_empty_buffers(page, blocksize, page 1673 fs/buffer.c block = (sector_t)page->index << (PAGE_CACHE_SHIFT - inode->i_blkbits); page 1674 fs/buffer.c head = page_buffers(page); page 1724 fs/buffer.c redirty_page_for_writepage(wbc, page); page 1738 fs/buffer.c BUG_ON(PageWriteback(page)); page 1739 fs/buffer.c set_page_writeback(page); page 1749 fs/buffer.c unlock_page(page); page 1759 fs/buffer.c end_page_writeback(page); page 1790 fs/buffer.c SetPageError(page); page 1791 fs/buffer.c BUG_ON(PageWriteback(page)); page 1792 fs/buffer.c mapping_set_error(page->mapping, err); page 1793 fs/buffer.c set_page_writeback(page); page 1803 fs/buffer.c unlock_page(page); page 1817 fs/buffer.c BUG_ON(!PageLocked(page)); page 1818 fs/buffer.c if (!page_has_buffers(page)) page 1821 fs/buffer.c bh = head = page_buffers(page); page 1828 fs/buffer.c if (!PageUptodate(page)) { page 1834 fs/buffer.c zero_user(page, start, size); page 1858 fs/buffer.c BUG_ON(!PageLocked(page)); page 1864 fs/buffer.c if (!page_has_buffers(page)) page 1865 fs/buffer.c create_empty_buffers(page, blocksize, 0); page 1866 fs/buffer.c head = page_buffers(page); page 1869 fs/buffer.c block = (sector_t)page->index << (PAGE_CACHE_SHIFT - bbits); page 1875 fs/buffer.c if (PageUptodate(page)) { page 1891 fs/buffer.c if (PageUptodate(page)) { page 1898 fs/buffer.c zero_user_segments(page, page 1904 fs/buffer.c if (PageUptodate(page)) { page 1925 fs/buffer.c page_zero_new_buffers(page, from, to); page 1939 fs/buffer.c for(bh = head = page_buffers(page), block_start = 0; page 1960 fs/buffer.c SetPageUptodate(page); page 1979 fs/buffer.c struct page *page; page 1988 fs/buffer.c page = *pagep; page 1989 fs/buffer.c if (page == NULL) { page 1991 fs/buffer.c page = __grab_cache_page(mapping, index); page 1992 fs/buffer.c if (!page) { page 1996 fs/buffer.c *pagep = page; page 1998 fs/buffer.c BUG_ON(!PageLocked(page)); page 2000 fs/buffer.c status = __block_prepare_write(inode, page, start, end, get_block); page 2002 fs/buffer.c ClearPageUptodate(page); page 2005 fs/buffer.c unlock_page(page); page 2006 fs/buffer.c page_cache_release(page); page 2047 fs/buffer.c if (!PageUptodate(page)) page 2050 fs/buffer.c page_zero_new_buffers(page, start+copied, start+len); page 2052 fs/buffer.c flush_dcache_page(page); page 2055 fs/buffer.c __block_commit_write(inode, page, start, start+copied); page 2068 fs/buffer.c copied = block_write_end(file, mapping, pos, len, copied, page, fsdata); page 2082 fs/buffer.c unlock_page(page); page 2083 fs/buffer.c page_cache_release(page); page 2108 fs/buffer.c struct inode *inode = page->mapping->host; page 2114 fs/buffer.c if (!page_has_buffers(page)) page 2123 fs/buffer.c head = page_buffers(page); page 2153 fs/buffer.c struct inode *inode = page->mapping->host; page 2160 fs/buffer.c BUG_ON(!PageLocked(page)); page 2162 fs/buffer.c if (!page_has_buffers(page)) page 2163 fs/buffer.c create_empty_buffers(page, blocksize, 0); page 2164 fs/buffer.c head = page_buffers(page); page 2166 fs/buffer.c iblock = (sector_t)page->index << (PAGE_CACHE_SHIFT - inode->i_blkbits); page 2184 fs/buffer.c SetPageError(page); page 2187 fs/buffer.c zero_user(page, i * blocksize, blocksize); page 2203 fs/buffer.c SetPageMappedToDisk(page); page 2210 fs/buffer.c if (!PageError(page)) page 2211 fs/buffer.c SetPageUptodate(page); page 2212 fs/buffer.c unlock_page(page); page 2245 fs/buffer.c struct page *page; page 2261 fs/buffer.c &page, &fsdata); page 2265 fs/buffer.c err = pagecache_write_end(NULL, mapping, size, 0, 0, page, fsdata); page 2277 fs/buffer.c struct page *page; page 2297 fs/buffer.c &page, &fsdata); page 2300 fs/buffer.c zero_user(page, zerofrom, len); page 2302 fs/buffer.c page, fsdata); page 2326 fs/buffer.c &page, &fsdata); page 2329 fs/buffer.c zero_user(page, zerofrom, len); page 2331 fs/buffer.c page, fsdata); page 2375 fs/buffer.c struct inode *inode = page->mapping->host; page 2376 fs/buffer.c int err = __block_prepare_write(inode, page, from, to, get_block); page 2378 fs/buffer.c ClearPageUptodate(page); page 2384 fs/buffer.c struct inode *inode = page->mapping->host; page 2385 fs/buffer.c __block_commit_write(inode,page,from,to); page 2413 fs/buffer.c lock_page(page); page 2415 fs/buffer.c if ((page->mapping != inode->i_mapping) || page 2416 fs/buffer.c (page_offset(page) > size)) { page 2422 fs/buffer.c if (((page->index + 1) << PAGE_CACHE_SHIFT) > size) page 2427 fs/buffer.c ret = block_prepare_write(page, 0, end, get_block); page 2429 fs/buffer.c ret = block_commit_write(page, 0, end); page 2432 fs/buffer.c unlock_page(page); page 2455 fs/buffer.c BUG_ON(!PageLocked(page)); page 2457 fs/buffer.c spin_lock(&page->mapping->private_lock); page 2460 fs/buffer.c if (PageDirty(page)) page 2466 fs/buffer.c attach_page_buffers(page, head); page 2467 fs/buffer.c spin_unlock(&page->mapping->private_lock); page 2483 fs/buffer.c struct page *page; page 2497 fs/buffer.c page = __grab_cache_page(mapping, index); page 2498 fs/buffer.c if (!page) page 2500 fs/buffer.c *pagep = page; page 2503 fs/buffer.c if (page_has_buffers(page)) { page 2504 fs/buffer.c unlock_page(page); page 2505 fs/buffer.c page_cache_release(page); page 2511 fs/buffer.c if (PageMappedToDisk(page)) page 2523 fs/buffer.c head = alloc_page_buffers(page, blocksize, 0); page 2529 fs/buffer.c block_in_file = (sector_t)page->index << (PAGE_CACHE_SHIFT - blkbits); page 2554 fs/buffer.c if (PageUptodate(page)) { page 2559 fs/buffer.c zero_user_segments(page, block_start, from, page 2589 fs/buffer.c SetPageMappedToDisk(page); page 2604 fs/buffer.c attach_nobh_buffers(page, head); page 2605 fs/buffer.c page_zero_new_buffers(page, from, to); page 2608 fs/buffer.c unlock_page(page); page 2609 fs/buffer.c page_cache_release(page); page 2623 fs/buffer.c struct inode *inode = page->mapping->host; page 2626 fs/buffer.c BUG_ON(fsdata != NULL && page_has_buffers(page)); page 2628 fs/buffer.c if (unlikely(copied < len) && !page_has_buffers(page)) page 2629 fs/buffer.c attach_nobh_buffers(page, head); page 2630 fs/buffer.c if (page_has_buffers(page)) page 2632 fs/buffer.c copied, page, fsdata); page 2634 fs/buffer.c SetPageUptodate(page); page 2635 fs/buffer.c set_page_dirty(page); page 2641 fs/buffer.c unlock_page(page); page 2642 fs/buffer.c page_cache_release(page); page 2662 fs/buffer.c struct inode * const inode = page->mapping->host; page 2669 fs/buffer.c if (page->index < end_index) page 2674 fs/buffer.c if (page->index >= end_index+1 || !offset) { page 2682 fs/buffer.c if (page->mapping->a_ops->invalidatepage) page 2683 fs/buffer.c page->mapping->a_ops->invalidatepage(page, offset); page 2685 fs/buffer.c unlock_page(page); page 2696 fs/buffer.c zero_user_segment(page, offset, PAGE_CACHE_SIZE); page 2698 fs/buffer.c ret = mpage_writepage(page, get_block, wbc); page 2700 fs/buffer.c ret = __block_write_full_page(inode, page, get_block, wbc); page 2714 fs/buffer.c struct page *page; page 2728 fs/buffer.c page = grab_cache_page(mapping, index); page 2730 fs/buffer.c if (!page) page 2733 fs/buffer.c if (page_has_buffers(page)) { page 2735 fs/buffer.c unlock_page(page); page 2736 fs/buffer.c page_cache_release(page); page 2755 fs/buffer.c if (!PageUptodate(page)) { page 2756 fs/buffer.c err = mapping->a_ops->readpage(NULL, page); page 2758 fs/buffer.c page_cache_release(page); page 2761 fs/buffer.c lock_page(page); page 2762 fs/buffer.c if (!PageUptodate(page)) { page 2766 fs/buffer.c if (page_has_buffers(page)) page 2769 fs/buffer.c zero_user(page, offset, length); page 2770 fs/buffer.c set_page_dirty(page); page 2774 fs/buffer.c unlock_page(page); page 2775 fs/buffer.c page_cache_release(page); page 2790 fs/buffer.c struct page *page; page 2804 fs/buffer.c page = grab_cache_page(mapping, index); page 2806 fs/buffer.c if (!page) page 2809 fs/buffer.c if (!page_has_buffers(page)) page 2810 fs/buffer.c create_empty_buffers(page, blocksize, 0); page 2813 fs/buffer.c bh = page_buffers(page); page 2833 fs/buffer.c if (PageUptodate(page)) page 2845 fs/buffer.c zero_user(page, offset, length); page 2850 fs/buffer.c unlock_page(page); page 2851 fs/buffer.c page_cache_release(page); page 2862 fs/buffer.c struct inode * const inode = page->mapping->host; page 2868 fs/buffer.c if (page->index < end_index) page 2869 fs/buffer.c return __block_write_full_page(inode, page, get_block, wbc); page 2873 fs/buffer.c if (page->index >= end_index+1 || !offset) { page 2879 fs/buffer.c do_invalidatepage(page, 0); page 2880 fs/buffer.c unlock_page(page); page 2891 fs/buffer.c zero_user_segment(page, offset, PAGE_CACHE_SIZE); page 2892 fs/buffer.c return __block_write_full_page(inode, page, get_block, wbc); page 3088 fs/buffer.c struct buffer_head *head = page_buffers(page); page 3093 fs/buffer.c if (buffer_write_io_error(bh) && page->mapping) page 3094 fs/buffer.c set_bit(AS_EIO, &page->mapping->flags); page 3108 fs/buffer.c __clear_page_buffers(page); page 3116 fs/buffer.c struct address_space * const mapping = page->mapping; page 3120 fs/buffer.c BUG_ON(!PageLocked(page)); page 3121 fs/buffer.c if (PageWriteback(page)) page 3125 fs/buffer.c ret = drop_buffers(page, &buffers_to_free); page 3130 fs/buffer.c ret = drop_buffers(page, &buffers_to_free); page 3147 fs/buffer.c cancel_dirty_page(page, PAGE_CACHE_SIZE); page 3168 fs/buffer.c mapping = page_mapping(page); page 3170 fs/buffer.c blk_run_backing_dev(mapping->backing_dev_info, page); page 1148 fs/cifs/file.c struct address_space *mapping = page->mapping; page 1149 fs/cifs/file.c loff_t offset = (loff_t)page->index << PAGE_CACHE_SHIFT; page 1161 fs/cifs/file.c inode = page->mapping->host; page 1166 fs/cifs/file.c write_data = kmap(page); page 1170 fs/cifs/file.c kunmap(page); page 1176 fs/cifs/file.c kunmap(page); page 1200 fs/cifs/file.c kunmap(page); page 1222 fs/cifs/file.c struct page *page; page 1285 fs/cifs/file.c page = pvec.pages[i]; page 1295 fs/cifs/file.c lock_page(page); page 1296 fs/cifs/file.c else if (!trylock_page(page)) page 1299 fs/cifs/file.c if (unlikely(page->mapping != mapping)) { page 1300 fs/cifs/file.c unlock_page(page); page 1304 fs/cifs/file.c if (!wbc->range_cyclic && page->index > end) { page 1306 fs/cifs/file.c unlock_page(page); page 1310 fs/cifs/file.c if (next && (page->index != next)) { page 1312 fs/cifs/file.c unlock_page(page); page 1317 fs/cifs/file.c wait_on_page_writeback(page); page 1319 fs/cifs/file.c if (PageWriteback(page) || page 1320 fs/cifs/file.c !clear_page_dirty_for_io(page)) { page 1321 fs/cifs/file.c unlock_page(page); page 1329 fs/cifs/file.c set_page_writeback(page); page 1331 fs/cifs/file.c if (page_offset(page) >= mapping->host->i_size) { page 1333 fs/cifs/file.c unlock_page(page); page 1334 fs/cifs/file.c end_page_writeback(page); page 1341 fs/cifs/file.c page_cache_get(page); page 1343 fs/cifs/file.c len = min(mapping->host->i_size - page_offset(page), page 1348 fs/cifs/file.c iov[n_iov].iov_base = kmap(page); page 1354 fs/cifs/file.c offset = page_offset(page); page 1356 fs/cifs/file.c next = page->index + 1; page 1391 fs/cifs/file.c page = pvec.pages[first + i]; page 1398 fs/cifs/file.c SetPageError(page); page 1399 fs/cifs/file.c kunmap(page); page 1400 fs/cifs/file.c unlock_page(page); page 1401 fs/cifs/file.c end_page_writeback(page); page 1402 fs/cifs/file.c page_cache_release(page); page 1434 fs/cifs/file.c page_cache_get(page); page 1435 fs/cifs/file.c if (!PageUptodate(page)) page 1448 fs/cifs/file.c set_page_writeback(page); page 1449 fs/cifs/file.c rc = cifs_partialpagewrite(page, 0, PAGE_CACHE_SIZE); page 1450 fs/cifs/file.c SetPageUptodate(page); /* BB add check for error and Clearuptodate? */ page 1451 fs/cifs/file.c unlock_page(page); page 1452 fs/cifs/file.c end_page_writeback(page); page 1453 fs/cifs/file.c page_cache_release(page); page 1466 fs/cifs/file.c page, pos, copied)); page 1468 fs/cifs/file.c if (!PageUptodate(page) && copied == PAGE_CACHE_SIZE) page 1469 fs/cifs/file.c SetPageUptodate(page); page 1471 fs/cifs/file.c if (!PageUptodate(page)) { page 1482 fs/cifs/file.c page_data = kmap(page); page 1485 fs/cifs/file.c kunmap(page); page 1491 fs/cifs/file.c set_page_dirty(page); page 1501 fs/cifs/file.c unlock_page(page); page 1502 fs/cifs/file.c page_cache_release(page); page 1757 fs/cifs/file.c struct page *page; page 1764 fs/cifs/file.c page = list_entry(pages->prev, struct page, lru); page 1765 fs/cifs/file.c list_del(&page->lru); page 1767 fs/cifs/file.c if (add_to_page_cache(page, mapping, page->index, page 1769 fs/cifs/file.c page_cache_release(page); page 1776 fs/cifs/file.c target = kmap_atomic(page, KM_USER0); page 1790 fs/cifs/file.c flush_dcache_page(page); page 1791 fs/cifs/file.c SetPageUptodate(page); page 1792 fs/cifs/file.c unlock_page(page); page 1793 fs/cifs/file.c if (!pagevec_add(plru_pvec, page)) page 1806 fs/cifs/file.c struct page *page; page 1830 fs/cifs/file.c struct page *tmp_page; page 1836 fs/cifs/file.c page = list_entry(page_list->prev, struct page, lru); page 1837 fs/cifs/file.c offset = (loff_t)page->index << PAGE_CACHE_SHIFT; page 1842 fs/cifs/file.c list_entry(page_list->prev, struct page, lru)->index; page 1949 fs/cifs/file.c page_cache_get(page); page 1950 fs/cifs/file.c read_data = kmap(page); page 1966 fs/cifs/file.c flush_dcache_page(page); page 1967 fs/cifs/file.c SetPageUptodate(page); page 1971 fs/cifs/file.c kunmap(page); page 1972 fs/cifs/file.c page_cache_release(page); page 1978 fs/cifs/file.c loff_t offset = (loff_t)page->index << PAGE_CACHE_SHIFT; page 1990 fs/cifs/file.c page, (int)offset, (int)offset)); page 1992 fs/cifs/file.c rc = cifs_readpage_worker(file, page, &offset); page 1994 fs/cifs/file.c unlock_page(page); page 1494 fs/cifs/inode.c struct page *page; page 1497 fs/cifs/inode.c page = grab_cache_page(mapping, index); page 1498 fs/cifs/inode.c if (!page) page 1501 fs/cifs/inode.c zero_user_segment(page, offset, PAGE_CACHE_SIZE); page 1502 fs/cifs/inode.c unlock_page(page); page 1503 fs/cifs/inode.c page_cache_release(page); page 26 fs/coda/symlink.c struct inode *inode = page->mapping->host; page 30 fs/coda/symlink.c char *p = kmap(page); page 39 fs/coda/symlink.c SetPageUptodate(page); page 40 fs/coda/symlink.c kunmap(page); page 41 fs/coda/symlink.c unlock_page(page); page 45 fs/coda/symlink.c SetPageError(page); page 46 fs/coda/symlink.c kunmap(page); page 47 fs/coda/symlink.c unlock_page(page); page 1256 fs/compat.c struct page *kmapped_page = NULL; page 1299 fs/compat.c struct page *page; page 1310 fs/compat.c 1, 1, 1, &page, NULL); page 1322 fs/compat.c kmapped_page = page; page 47 fs/configfs/file.c char * page; page 72 fs/configfs/file.c if (!buffer->page) page 73 fs/configfs/file.c buffer->page = (char *) get_zeroed_page(GFP_KERNEL); page 74 fs/configfs/file.c if (!buffer->page) page 77 fs/configfs/file.c count = ops->show_attribute(item,attr,buffer->page); page 118 fs/configfs/file.c __func__, count, *ppos, buffer->page); page 119 fs/configfs/file.c retval = simple_read_from_buffer(buf, count, ppos, buffer->page, page 142 fs/configfs/file.c if (!buffer->page) page 143 fs/configfs/file.c buffer->page = (char *)__get_free_pages(GFP_KERNEL, 0); page 144 fs/configfs/file.c if (!buffer->page) page 149 fs/configfs/file.c error = copy_from_user(buffer->page,buf,count); page 153 fs/configfs/file.c buffer->page[count] = 0; page 176 fs/configfs/file.c return ops->store_attribute(item,attr,buffer->page,count); page 300 fs/configfs/file.c if (buffer->page) page 301 fs/configfs/file.c free_page((unsigned long)buffer->page); page 288 fs/configfs/symlink.c unsigned long page = get_zeroed_page(GFP_KERNEL); page 290 fs/configfs/symlink.c if (page) { page 291 fs/configfs/symlink.c error = configfs_getlink(dentry, (char *)page); page 293 fs/configfs/symlink.c nd_set_link(nd, (char *)page); page 294 fs/configfs/symlink.c return (void *)page; page 306 fs/configfs/symlink.c unsigned long page = (unsigned long)cookie; page 307 fs/configfs/symlink.c free_page(page); page 140 fs/cramfs/inode.c struct page *pages[BLKS_PER_BUF]; page 169 fs/cramfs/inode.c struct page *page = NULL; page 172 fs/cramfs/inode.c page = read_mapping_page_async(mapping, blocknr + i, page 175 fs/cramfs/inode.c if (IS_ERR(page)) page 176 fs/cramfs/inode.c page = NULL; page 178 fs/cramfs/inode.c pages[i] = page; page 182 fs/cramfs/inode.c struct page *page = pages[i]; page 183 fs/cramfs/inode.c if (page) { page 184 fs/cramfs/inode.c wait_on_page_locked(page); page 185 fs/cramfs/inode.c if (!PageUptodate(page)) { page 187 fs/cramfs/inode.c page_cache_release(page); page 200 fs/cramfs/inode.c struct page *page = pages[i]; page 201 fs/cramfs/inode.c if (page) { page 202 fs/cramfs/inode.c memcpy(data, kmap(page), PAGE_CACHE_SIZE); page 203 fs/cramfs/inode.c kunmap(page); page 204 fs/cramfs/inode.c page_cache_release(page); page 463 fs/cramfs/inode.c struct inode *inode = page->mapping->host; page 469 fs/cramfs/inode.c if (page->index < maxblock) { page 471 fs/cramfs/inode.c u32 blkptr_offset = OFFSET(inode) + page->index*4; page 476 fs/cramfs/inode.c if (page->index) page 480 fs/cramfs/inode.c pgdata = kmap(page); page 494 fs/cramfs/inode.c pgdata = kmap(page); page 496 fs/cramfs/inode.c kunmap(page); page 497 fs/cramfs/inode.c flush_dcache_page(page); page 498 fs/cramfs/inode.c SetPageUptodate(page); page 499 fs/cramfs/inode.c unlock_page(page); page 2092 fs/dcache.c char *page = (char *) __get_free_page(GFP_USER); page 2094 fs/dcache.c if (!page) page 2112 fs/dcache.c cwd = __d_path(&pwd, &tmp, page, PAGE_SIZE); page 2120 fs/dcache.c len = PAGE_SIZE + page - cwd; page 2132 fs/dcache.c free_page((unsigned long) page); page 102 fs/direct-io.c struct page *cur_page; /* The page */ page 118 fs/direct-io.c struct page *pages[DIO_PAGES]; /* page buffer */ page 160 fs/direct-io.c struct page *page = ZERO_PAGE(0); page 168 fs/direct-io.c page_cache_get(page); page 169 fs/direct-io.c dio->pages[0] = page; page 193 fs/direct-io.c static struct page *dio_get_page(struct dio *dio) page 412 fs/direct-io.c struct page *page = bvec[page_no].bv_page; page 414 fs/direct-io.c if (dio->rw == READ && !PageCompound(page)) page 415 fs/direct-io.c set_page_dirty_lock(page); page 416 fs/direct-io.c page_cache_release(page); page 669 fs/direct-io.c if ( (dio->cur_page == page) && page 698 fs/direct-io.c page_cache_get(page); /* It is in dio */ page 699 fs/direct-io.c dio->cur_page = page; page 739 fs/direct-io.c struct page *page; page 760 fs/direct-io.c page = ZERO_PAGE(0); page 761 fs/direct-io.c if (submit_page_section(dio, page, 0, this_chunk_bytes, page 788 fs/direct-io.c struct page *page; page 797 fs/direct-io.c page = dio_get_page(dio); page 798 fs/direct-io.c if (IS_ERR(page)) { page 799 fs/direct-io.c ret = PTR_ERR(page); page 818 fs/direct-io.c page_cache_release(page); page 859 fs/direct-io.c page_cache_release(page); page 872 fs/direct-io.c page_cache_release(page); page 875 fs/direct-io.c zero_user(page, block_in_page << blkbits, page 905 fs/direct-io.c ret = submit_page_section(dio, page, offset_in_page, page 908 fs/direct-io.c page_cache_release(page); page 923 fs/direct-io.c page_cache_release(page); page 112 fs/dlm/lowcomms.c struct page *rx_page; page 126 fs/dlm/lowcomms.c struct page *page; page 771 fs/dlm/lowcomms.c __free_page(e->page); page 826 fs/dlm/lowcomms.c kmap(e->page); page 829 fs/dlm/lowcomms.c iov[0].iov_base = page_address(e->page)+offset; page 857 fs/dlm/lowcomms.c kunmap(e->page); page 1166 fs/dlm/lowcomms.c entry->page = alloc_page(allocation); page 1167 fs/dlm/lowcomms.c if (!entry->page) { page 1207 fs/dlm/lowcomms.c kmap(e->page); page 1208 fs/dlm/lowcomms.c *ppc = page_address(e->page) + offset; page 1236 fs/dlm/lowcomms.c kunmap(e->page); page 1253 fs/dlm/lowcomms.c ssize_t(*sendpage) (struct socket *, struct page *, int, size_t, int); page 1275 fs/dlm/lowcomms.c kmap(e->page); page 1279 fs/dlm/lowcomms.c ret = sendpage(con->sock, e->page, offset, len, page 1297 fs/dlm/lowcomms.c kunmap(e->page); page 297 fs/ecryptfs/crypto.c struct page *pg; page 412 fs/ecryptfs/crypto.c extent_base = (((loff_t)page->index) page 430 fs/ecryptfs/crypto.c (page_address(page) page 435 fs/ecryptfs/crypto.c page, (extent_offset page 441 fs/ecryptfs/crypto.c "rc = [%d]\n", __func__, page->index, extent_offset, page 479 fs/ecryptfs/crypto.c struct page *enc_extent_page = NULL; page 483 fs/ecryptfs/crypto.c ecryptfs_inode = page->mapping->host; page 487 fs/ecryptfs/crypto.c rc = ecryptfs_write_lower_page_segment(ecryptfs_inode, page, page 492 fs/ecryptfs/crypto.c page->index); page 508 fs/ecryptfs/crypto.c rc = ecryptfs_encrypt_extent(enc_extent_page, crypt_stat, page, page 516 fs/ecryptfs/crypto.c &offset, ((((loff_t)page->index) page 546 fs/ecryptfs/crypto.c extent_base = (((loff_t)page->index) page 568 fs/ecryptfs/crypto.c rc = ecryptfs_decrypt_page_offset(crypt_stat, page, page 576 fs/ecryptfs/crypto.c "rc = [%d]\n", __func__, page->index, extent_offset, page 587 fs/ecryptfs/crypto.c ecryptfs_dump_hex((char *)(page_address(page) page 616 fs/ecryptfs/crypto.c struct page *enc_extent_page = NULL; page 620 fs/ecryptfs/crypto.c ecryptfs_inode = page->mapping->host; page 624 fs/ecryptfs/crypto.c rc = ecryptfs_read_lower_page_segment(page, page->index, 0, page 630 fs/ecryptfs/crypto.c page->index); page 647 fs/ecryptfs/crypto.c &offset, ((page->index * (PAGE_CACHE_SIZE page 659 fs/ecryptfs/crypto.c rc = ecryptfs_decrypt_extent(page, crypt_stat, enc_extent_page, page 178 fs/ecryptfs/ecryptfs_kernel.h struct page *page; page 691 fs/ecryptfs/ecryptfs_kernel.h struct page *ecryptfs_get_locked_page(struct file *file, loff_t index); page 46 fs/ecryptfs/mmap.c struct page *ecryptfs_get_locked_page(struct file *file, loff_t index) page 51 fs/ecryptfs/mmap.c struct page *page; page 56 fs/ecryptfs/mmap.c page = read_mapping_page(mapping, index, (void *)file); page 57 fs/ecryptfs/mmap.c if (!IS_ERR(page)) page 58 fs/ecryptfs/mmap.c lock_page(page); page 59 fs/ecryptfs/mmap.c return page; page 72 fs/ecryptfs/mmap.c rc = ecryptfs_encrypt_page(page); page 75 fs/ecryptfs/mmap.c "page (upper index [0x%.16x])\n", page->index); page 76 fs/ecryptfs/mmap.c ClearPageUptodate(page); page 79 fs/ecryptfs/mmap.c SetPageUptodate(page); page 80 fs/ecryptfs/mmap.c unlock_page(page); page 134 fs/ecryptfs/mmap.c loff_t view_extent_num = ((((loff_t)page->index) page 145 fs/ecryptfs/mmap.c page_virt = kmap_atomic(page, KM_USER0); page 150 fs/ecryptfs/mmap.c page_virt, page->mapping->host); page 154 fs/ecryptfs/mmap.c flush_dcache_page(page); page 167 fs/ecryptfs/mmap.c page, (lower_offset >> PAGE_CACHE_SHIFT), page 169 fs/ecryptfs/mmap.c crypt_stat->extent_size, page->mapping->host); page 204 fs/ecryptfs/mmap.c rc = ecryptfs_read_lower_page_segment(page, page->index, 0, page 206 fs/ecryptfs/mmap.c page->mapping->host); page 209 fs/ecryptfs/mmap.c rc = ecryptfs_copy_up_encrypted_with_header(page, page 222 fs/ecryptfs/mmap.c page, page->index, 0, PAGE_CACHE_SIZE, page 223 fs/ecryptfs/mmap.c page->mapping->host); page 231 fs/ecryptfs/mmap.c rc = ecryptfs_decrypt_page(page); page 240 fs/ecryptfs/mmap.c ClearPageUptodate(page); page 242 fs/ecryptfs/mmap.c SetPageUptodate(page); page 244 fs/ecryptfs/mmap.c page->index); page 245 fs/ecryptfs/mmap.c unlock_page(page); page 254 fs/ecryptfs/mmap.c struct inode *inode = page->mapping->host; page 257 fs/ecryptfs/mmap.c if ((i_size_read(inode) / PAGE_CACHE_SIZE) != page->index) page 262 fs/ecryptfs/mmap.c zero_user_segment(page, end_byte_in_page, PAGE_CACHE_SIZE); page 284 fs/ecryptfs/mmap.c if (!PageUptodate(page)) { page 292 fs/ecryptfs/mmap.c page, page->index, 0, PAGE_CACHE_SIZE, page 293 fs/ecryptfs/mmap.c page->mapping->host); page 298 fs/ecryptfs/mmap.c ClearPageUptodate(page); page 301 fs/ecryptfs/mmap.c SetPageUptodate(page); page 305 fs/ecryptfs/mmap.c page, crypt_stat); page 313 fs/ecryptfs/mmap.c ClearPageUptodate(page); page 316 fs/ecryptfs/mmap.c SetPageUptodate(page); page 319 fs/ecryptfs/mmap.c page, page->index, 0, PAGE_CACHE_SIZE, page 320 fs/ecryptfs/mmap.c page->mapping->host); page 325 fs/ecryptfs/mmap.c ClearPageUptodate(page); page 328 fs/ecryptfs/mmap.c SetPageUptodate(page); page 331 fs/ecryptfs/mmap.c rc = ecryptfs_decrypt_page(page); page 335 fs/ecryptfs/mmap.c __func__, page->index, rc); page 336 fs/ecryptfs/mmap.c ClearPageUptodate(page); page 339 fs/ecryptfs/mmap.c SetPageUptodate(page); page 342 fs/ecryptfs/mmap.c prev_page_end_size = ((loff_t)page->index << PAGE_CACHE_SHIFT); page 345 fs/ecryptfs/mmap.c if (page->index != 0) { page 346 fs/ecryptfs/mmap.c if (prev_page_end_size > i_size_read(page->mapping->host)) { page 360 fs/ecryptfs/mmap.c if ((i_size_read(page->mapping->host) == prev_page_end_size) page 362 fs/ecryptfs/mmap.c zero_user(page, 0, PAGE_CACHE_SIZE); page 462 fs/ecryptfs/mmap.c struct inode *ecryptfs_inode = page->mapping->host; page 474 fs/ecryptfs/mmap.c "(page w/ index = [0x%.16x], to = [%d])\n", page->index, page 477 fs/ecryptfs/mmap.c rc = fill_zeros_to_end_of_page(page, to); page 481 fs/ecryptfs/mmap.c page->index); page 484 fs/ecryptfs/mmap.c rc = ecryptfs_encrypt_page(page); page 487 fs/ecryptfs/mmap.c "index [0x%.16x])\n", page->index); page 490 fs/ecryptfs/mmap.c pos = (((loff_t)page->index) << PAGE_CACHE_SHIFT) + to; page 119 fs/ecryptfs/read_write.c struct page *ecryptfs_page; page 303 fs/ecryptfs/read_write.c struct page *ecryptfs_page; page 18 fs/efs/inode.c return block_read_full_page(page,efs_get_block); page 17 fs/efs/symlink.c char *link = kmap(page); page 19 fs/efs/symlink.c struct inode * inode = page->mapping->host; page 44 fs/efs/symlink.c SetPageUptodate(page); page 45 fs/efs/symlink.c kunmap(page); page 46 fs/efs/symlink.c unlock_page(page); page 51 fs/efs/symlink.c SetPageError(page); page 52 fs/efs/symlink.c kunmap(page); page 53 fs/efs/symlink.c unlock_page(page); page 171 fs/exec.c static struct page *get_arg_page(struct linux_binprm *bprm, unsigned long pos, page 174 fs/exec.c struct page *page; page 185 fs/exec.c 1, write, 1, &page, NULL); page 198 fs/exec.c return page; page 209 fs/exec.c put_page(page); page 214 fs/exec.c return page; page 219 fs/exec.c put_page(page); page 233 fs/exec.c flush_cache_page(bprm->vma, pos, page_to_pfn(page)); page 289 fs/exec.c static struct page *get_arg_page(struct linux_binprm *bprm, unsigned long pos, page 292 fs/exec.c struct page *page; page 294 fs/exec.c page = bprm->page[pos / PAGE_SIZE]; page 295 fs/exec.c if (!page && write) { page 296 fs/exec.c page = alloc_page(GFP_HIGHUSER|__GFP_ZERO); page 297 fs/exec.c if (!page) page 299 fs/exec.c bprm->page[pos / PAGE_SIZE] = page; page 302 fs/exec.c return page; page 311 fs/exec.c if (bprm->page[i]) { page 312 fs/exec.c __free_page(bprm->page[i]); page 313 fs/exec.c bprm->page[i] = NULL; page 410 fs/exec.c struct page *kmapped_page = NULL; page 453 fs/exec.c struct page *page; page 455 fs/exec.c page = get_arg_page(bprm, pos, 1); page 456 fs/exec.c if (!page) { page 466 fs/exec.c kmapped_page = page; page 1128 fs/exec.c struct page *page; page 1135 fs/exec.c page = get_arg_page(bprm, bprm->p, 0); page 1136 fs/exec.c if (!page) { page 1140 fs/exec.c kaddr = kmap_atomic(page, KM_USER0); page 1147 fs/exec.c put_arg_page(page); page 60 fs/ext2/dir.c kunmap(page); page 61 fs/ext2/dir.c page_cache_release(page); page 86 fs/ext2/dir.c struct address_space *mapping = page->mapping; page 91 fs/ext2/dir.c block_write_end(NULL, mapping, pos, len, len, page, NULL); page 99 fs/ext2/dir.c err = write_one_page(page, 1); page 101 fs/ext2/dir.c unlock_page(page); page 108 fs/ext2/dir.c struct inode *dir = page->mapping->host; page 111 fs/ext2/dir.c char *kaddr = page_address(page); page 118 fs/ext2/dir.c if ((dir->i_size >> PAGE_CACHE_SHIFT) == page->index) { page 143 fs/ext2/dir.c SetPageChecked(page); page 171 fs/ext2/dir.c dir->i_ino, error, (page->index<<PAGE_CACHE_SHIFT)+offs, page 180 fs/ext2/dir.c dir->i_ino, (page->index<<PAGE_CACHE_SHIFT)+offs, page 183 fs/ext2/dir.c SetPageChecked(page); page 184 fs/ext2/dir.c SetPageError(page); page 187 fs/ext2/dir.c static struct page * ext2_get_page(struct inode *dir, unsigned long n) page 190 fs/ext2/dir.c struct page *page = read_mapping_page(mapping, n, NULL); page 191 fs/ext2/dir.c if (!IS_ERR(page)) { page 192 fs/ext2/dir.c kmap(page); page 193 fs/ext2/dir.c if (!PageChecked(page)) page 194 fs/ext2/dir.c ext2_check_page(page); page 195 fs/ext2/dir.c if (PageError(page)) page 198 fs/ext2/dir.c return page; page 201 fs/ext2/dir.c ext2_put_page(page); page 295 fs/ext2/dir.c struct page *page = ext2_get_page(inode, n); page 297 fs/ext2/dir.c if (IS_ERR(page)) { page 302 fs/ext2/dir.c return PTR_ERR(page); page 304 fs/ext2/dir.c kaddr = page_address(page); page 319 fs/ext2/dir.c ext2_put_page(page); page 334 fs/ext2/dir.c ext2_put_page(page); page 340 fs/ext2/dir.c ext2_put_page(page); page 361 fs/ext2/dir.c struct page *page = NULL; page 377 fs/ext2/dir.c page = ext2_get_page(dir, n); page 378 fs/ext2/dir.c if (!IS_ERR(page)) { page 379 fs/ext2/dir.c kaddr = page_address(page); page 386 fs/ext2/dir.c ext2_put_page(page); page 393 fs/ext2/dir.c ext2_put_page(page); page 410 fs/ext2/dir.c *res_page = page; page 417 fs/ext2/dir.c struct page *page = ext2_get_page(dir, 0); page 420 fs/ext2/dir.c if (!IS_ERR(page)) { page 421 fs/ext2/dir.c de = ext2_next_entry((ext2_dirent *) page_address(page)); page 422 fs/ext2/dir.c *p = page; page 431 fs/ext2/dir.c struct page *page; page 433 fs/ext2/dir.c de = ext2_find_entry (dir, dentry, &page); page 436 fs/ext2/dir.c ext2_put_page(page); page 445 fs/ext2/dir.c loff_t pos = page_offset(page) + page 446 fs/ext2/dir.c (char *) de - (char *) page_address(page); page 450 fs/ext2/dir.c lock_page(page); page 451 fs/ext2/dir.c err = __ext2_write_begin(NULL, page->mapping, pos, len, page 452 fs/ext2/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 456 fs/ext2/dir.c err = ext2_commit_chunk(page, pos, len); page 457 fs/ext2/dir.c ext2_put_page(page); page 474 fs/ext2/dir.c struct page *page = NULL; page 490 fs/ext2/dir.c page = ext2_get_page(dir, n); page 491 fs/ext2/dir.c err = PTR_ERR(page); page 492 fs/ext2/dir.c if (IS_ERR(page)) page 494 fs/ext2/dir.c lock_page(page); page 495 fs/ext2/dir.c kaddr = page_address(page); page 525 fs/ext2/dir.c unlock_page(page); page 526 fs/ext2/dir.c ext2_put_page(page); page 532 fs/ext2/dir.c pos = page_offset(page) + page 533 fs/ext2/dir.c (char*)de - (char*)page_address(page); page 534 fs/ext2/dir.c err = __ext2_write_begin(NULL, page->mapping, pos, rec_len, 0, page 535 fs/ext2/dir.c &page, NULL); page 548 fs/ext2/dir.c err = ext2_commit_chunk(page, pos, rec_len); page 554 fs/ext2/dir.c ext2_put_page(page); page 558 fs/ext2/dir.c unlock_page(page); page 568 fs/ext2/dir.c struct address_space *mapping = page->mapping; page 570 fs/ext2/dir.c char *kaddr = page_address(page); page 590 fs/ext2/dir.c from = (char*)pde - (char*)page_address(page); page 591 fs/ext2/dir.c pos = page_offset(page) + from; page 592 fs/ext2/dir.c lock_page(page); page 593 fs/ext2/dir.c err = __ext2_write_begin(NULL, page->mapping, pos, to - from, 0, page 594 fs/ext2/dir.c &page, NULL); page 599 fs/ext2/dir.c err = ext2_commit_chunk(page, pos, to - from); page 604 fs/ext2/dir.c ext2_put_page(page); page 614 fs/ext2/dir.c struct page *page = grab_cache_page(mapping, 0); page 620 fs/ext2/dir.c if (!page) page 623 fs/ext2/dir.c err = __ext2_write_begin(NULL, page->mapping, 0, chunk_size, 0, page 624 fs/ext2/dir.c &page, NULL); page 626 fs/ext2/dir.c unlock_page(page); page 629 fs/ext2/dir.c kaddr = kmap_atomic(page, KM_USER0); page 645 fs/ext2/dir.c err = ext2_commit_chunk(page, 0, chunk_size); page 647 fs/ext2/dir.c page_cache_release(page); page 656 fs/ext2/dir.c struct page *page = NULL; page 662 fs/ext2/dir.c page = ext2_get_page(inode, i); page 664 fs/ext2/dir.c if (IS_ERR(page)) page 667 fs/ext2/dir.c kaddr = page_address(page); page 693 fs/ext2/dir.c ext2_put_page(page); page 698 fs/ext2/dir.c ext2_put_page(page); page 717 fs/ext2/inode.c return block_write_full_page(page, ext2_get_block, wbc); page 722 fs/ext2/inode.c return mpage_readpage(page, ext2_get_block); page 766 fs/ext2/inode.c return nobh_writepage(page, ext2_get_block, wbc); page 257 fs/ext2/namei.c struct page * page; page 260 fs/ext2/namei.c de = ext2_find_entry (dir, dentry, &page); page 264 fs/ext2/namei.c err = ext2_delete_entry (de, page); page 296 fs/ext2/namei.c struct page * dir_page = NULL; page 298 fs/ext2/namei.c struct page * old_page; page 314 fs/ext2/namei.c struct page *new_page; page 1154 fs/ext3/inode.c struct page *page; page 1163 fs/ext3/inode.c page = __grab_cache_page(mapping, index); page 1164 fs/ext3/inode.c if (!page) page 1166 fs/ext3/inode.c *pagep = page; page 1170 fs/ext3/inode.c unlock_page(page); page 1171 fs/ext3/inode.c page_cache_release(page); page 1181 fs/ext3/inode.c ret = walk_page_buffers(handle, page_buffers(page), page 1187 fs/ext3/inode.c unlock_page(page); page 1188 fs/ext3/inode.c page_cache_release(page); page 1228 fs/ext3/inode.c copied = block_write_end(file, mapping, pos, len, copied, page, fsdata); page 1258 fs/ext3/inode.c ret = walk_page_buffers(handle, page_buffers(page), page 1273 fs/ext3/inode.c page, fsdata); page 1281 fs/ext3/inode.c unlock_page(page); page 1282 fs/ext3/inode.c page_cache_release(page); page 1302 fs/ext3/inode.c page, fsdata); page 1310 fs/ext3/inode.c unlock_page(page); page 1311 fs/ext3/inode.c page_cache_release(page); page 1331 fs/ext3/inode.c if (!PageUptodate(page)) page 1333 fs/ext3/inode.c page_zero_new_buffers(page, from+copied, to); page 1336 fs/ext3/inode.c ret = walk_page_buffers(handle, page_buffers(page), from, page 1339 fs/ext3/inode.c SetPageUptodate(page); page 1353 fs/ext3/inode.c unlock_page(page); page 1354 fs/ext3/inode.c page_cache_release(page); page 1485 fs/ext3/inode.c struct inode *inode = page->mapping->host; page 1491 fs/ext3/inode.c J_ASSERT(PageLocked(page)); page 1507 fs/ext3/inode.c if (!page_has_buffers(page)) { page 1508 fs/ext3/inode.c create_empty_buffers(page, inode->i_sb->s_blocksize, page 1511 fs/ext3/inode.c page_bufs = page_buffers(page); page 1515 fs/ext3/inode.c ret = block_write_full_page(page, ext3_get_block, wbc); page 1543 fs/ext3/inode.c redirty_page_for_writepage(wbc, page); page 1544 fs/ext3/inode.c unlock_page(page); page 1551 fs/ext3/inode.c struct inode *inode = page->mapping->host; page 1566 fs/ext3/inode.c ret = nobh_writepage(page, ext3_get_block, wbc); page 1568 fs/ext3/inode.c ret = block_write_full_page(page, ext3_get_block, wbc); page 1576 fs/ext3/inode.c redirty_page_for_writepage(wbc, page); page 1577 fs/ext3/inode.c unlock_page(page); page 1584 fs/ext3/inode.c struct inode *inode = page->mapping->host; page 1598 fs/ext3/inode.c if (!page_has_buffers(page) || PageChecked(page)) { page 1603 fs/ext3/inode.c ClearPageChecked(page); page 1604 fs/ext3/inode.c ret = block_prepare_write(page, 0, PAGE_CACHE_SIZE, page 1610 fs/ext3/inode.c ret = walk_page_buffers(handle, page_buffers(page), 0, page 1613 fs/ext3/inode.c err = walk_page_buffers(handle, page_buffers(page), 0, page 1618 fs/ext3/inode.c unlock_page(page); page 1625 fs/ext3/inode.c ret = block_write_full_page(page, ext3_get_block, wbc); page 1634 fs/ext3/inode.c redirty_page_for_writepage(wbc, page); page 1636 fs/ext3/inode.c unlock_page(page); page 1642 fs/ext3/inode.c return mpage_readpage(page, ext3_get_block); page 1654 fs/ext3/inode.c journal_t *journal = EXT3_JOURNAL(page->mapping->host); page 1660 fs/ext3/inode.c ClearPageChecked(page); page 1662 fs/ext3/inode.c journal_invalidatepage(journal, page, offset); page 1667 fs/ext3/inode.c journal_t *journal = EXT3_JOURNAL(page->mapping->host); page 1669 fs/ext3/inode.c WARN_ON(PageChecked(page)); page 1670 fs/ext3/inode.c if (!page_has_buffers(page)) page 1672 fs/ext3/inode.c return journal_try_to_free_buffers(journal, page, wait); page 1773 fs/ext3/inode.c SetPageChecked(page); page 1774 fs/ext3/inode.c return __set_page_dirty_nobuffers(page); page 1855 fs/ext3/inode.c if (!page_has_buffers(page) && test_opt(inode->i_sb, NOBH) && page 1856 fs/ext3/inode.c ext3_should_writeback_data(inode) && PageUptodate(page)) { page 1857 fs/ext3/inode.c zero_user(page, offset, length); page 1858 fs/ext3/inode.c set_page_dirty(page); page 1862 fs/ext3/inode.c if (!page_has_buffers(page)) page 1863 fs/ext3/inode.c create_empty_buffers(page, blocksize, 0); page 1866 fs/ext3/inode.c bh = page_buffers(page); page 1891 fs/ext3/inode.c if (PageUptodate(page)) page 1910 fs/ext3/inode.c zero_user(page, offset, length); page 1923 fs/ext3/inode.c unlock_page(page); page 1924 fs/ext3/inode.c page_cache_release(page); page 2336 fs/ext3/inode.c struct page *page; page 2347 fs/ext3/inode.c page = NULL; page 2349 fs/ext3/inode.c page = grab_cache_page(mapping, page 2351 fs/ext3/inode.c if (!page) page 2357 fs/ext3/inode.c if (page) { page 2358 fs/ext3/inode.c clear_highpage(page); page 2359 fs/ext3/inode.c flush_dcache_page(page); page 2360 fs/ext3/inode.c unlock_page(page); page 2361 fs/ext3/inode.c page_cache_release(page); page 2369 fs/ext3/inode.c if (page) page 2370 fs/ext3/inode.c ext3_block_truncate_page(handle, page, mapping, inode->i_size); page 3102 fs/ext4/extents.c struct page *page; page 3106 fs/ext4/extents.c page = find_get_page(inode->i_mapping, offset); page 3107 fs/ext4/extents.c if (!page || !page_has_buffers(page)) page 3110 fs/ext4/extents.c bh = page_buffers(page); page 3117 fs/ext4/extents.c page_cache_release(page); page 3119 fs/ext4/extents.c page_cache_release(page); page 1333 fs/ext4/inode.c struct page *page; page 1348 fs/ext4/inode.c page = __grab_cache_page(mapping, index); page 1349 fs/ext4/inode.c if (!page) { page 1354 fs/ext4/inode.c *pagep = page; page 1360 fs/ext4/inode.c ret = walk_page_buffers(handle, page_buffers(page), page 1365 fs/ext4/inode.c unlock_page(page); page 1367 fs/ext4/inode.c page_cache_release(page); page 1424 fs/ext4/inode.c page, fsdata); page 1457 fs/ext4/inode.c page, fsdata); page 1485 fs/ext4/inode.c if (!PageUptodate(page)) page 1487 fs/ext4/inode.c page_zero_new_buffers(page, from+copied, to); page 1490 fs/ext4/inode.c ret = walk_page_buffers(handle, page_buffers(page), from, page 1493 fs/ext4/inode.c SetPageUptodate(page); page 1505 fs/ext4/inode.c unlock_page(page); page 1509 fs/ext4/inode.c page_cache_release(page); page 1602 fs/ext4/inode.c head = page_buffers(page); page 1613 fs/ext4/inode.c ext4_da_release_space(page->mapping->host, to_release); page 1663 fs/ext4/inode.c struct page *page = pvec.pages[i]; page 1665 fs/ext4/inode.c index = page->index; page 1670 fs/ext4/inode.c err = mapping->a_ops->writepage(page, mpd->wbc); page 1721 fs/ext4/inode.c struct page *page = pvec.pages[i]; page 1723 fs/ext4/inode.c index = page->index; page 1728 fs/ext4/inode.c BUG_ON(!PageLocked(page)); page 1729 fs/ext4/inode.c BUG_ON(PageWriteback(page)); page 1730 fs/ext4/inode.c BUG_ON(!page_has_buffers(page)); page 1732 fs/ext4/inode.c bh = page_buffers(page); page 1799 fs/ext4/inode.c struct page *page = pvec.pages[i]; page 1800 fs/ext4/inode.c index = page->index; page 1805 fs/ext4/inode.c BUG_ON(!PageLocked(page)); page 1806 fs/ext4/inode.c BUG_ON(PageWriteback(page)); page 1807 fs/ext4/inode.c block_invalidatepage(page, 0); page 1808 fs/ext4/inode.c ClearPageUptodate(page); page 1809 fs/ext4/inode.c unlock_page(page); page 2016 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); page 2017 fs/ext4/inode.c unlock_page(page); page 2023 fs/ext4/inode.c if (mpd->next_page != page->index) { page 2035 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); page 2036 fs/ext4/inode.c unlock_page(page); page 2043 fs/ext4/inode.c mpd->first_page = page->index; page 2053 fs/ext4/inode.c mpd->next_page = page->index + 1; page 2054 fs/ext4/inode.c logical = (sector_t) page->index << page 2057 fs/ext4/inode.c if (!page_has_buffers(page)) { page 2074 fs/ext4/inode.c head = page_buffers(page); page 2264 fs/ext4/inode.c struct inode *inode = page->mapping->host; page 2267 fs/ext4/inode.c if (page->index == size >> PAGE_CACHE_SHIFT) page 2272 fs/ext4/inode.c if (page_has_buffers(page)) { page 2273 fs/ext4/inode.c page_bufs = page_buffers(page); page 2284 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); page 2285 fs/ext4/inode.c unlock_page(page); page 2302 fs/ext4/inode.c ret = block_prepare_write(page, 0, PAGE_CACHE_SIZE, page 2305 fs/ext4/inode.c page_bufs = page_buffers(page); page 2309 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); page 2310 fs/ext4/inode.c unlock_page(page); page 2319 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); page 2320 fs/ext4/inode.c unlock_page(page); page 2326 fs/ext4/inode.c ret = nobh_writepage(page, ext4_normal_get_block_write, wbc); page 2328 fs/ext4/inode.c ret = block_write_full_page(page, page 2502 fs/ext4/inode.c struct page *page; page 2531 fs/ext4/inode.c page = __grab_cache_page(mapping, index); page 2532 fs/ext4/inode.c if (!page) { page 2537 fs/ext4/inode.c *pagep = page; page 2542 fs/ext4/inode.c unlock_page(page); page 2544 fs/ext4/inode.c page_cache_release(page); page 2568 fs/ext4/inode.c struct inode *inode = page->mapping->host; page 2572 fs/ext4/inode.c bh = page_buffers(page); page 2598 fs/ext4/inode.c len, copied, page, fsdata); page 2601 fs/ext4/inode.c len, copied, page, fsdata); page 2618 fs/ext4/inode.c if (ext4_da_should_update_i_disksize(page, end)) { page 2640 fs/ext4/inode.c page, fsdata); page 2656 fs/ext4/inode.c BUG_ON(!PageLocked(page)); page 2657 fs/ext4/inode.c if (!page_has_buffers(page)) page 2660 fs/ext4/inode.c ext4_da_page_release_reservation(page, offset); page 2663 fs/ext4/inode.c ext4_invalidatepage(page, offset); page 2799 fs/ext4/inode.c struct inode *inode = page->mapping->host; page 2802 fs/ext4/inode.c return nobh_writepage(page, page 2805 fs/ext4/inode.c return block_write_full_page(page, page 2813 fs/ext4/inode.c struct inode *inode = page->mapping->host; page 2817 fs/ext4/inode.c J_ASSERT(PageLocked(page)); page 2818 fs/ext4/inode.c if (page->index == size >> PAGE_CACHE_SHIFT) page 2823 fs/ext4/inode.c if (page_has_buffers(page)) { page 2833 fs/ext4/inode.c BUG_ON(walk_page_buffers(NULL, page_buffers(page), 0, len, NULL, page 2838 fs/ext4/inode.c return __ext4_normal_writepage(page, wbc); page 2840 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); page 2841 fs/ext4/inode.c unlock_page(page); page 2848 fs/ext4/inode.c struct address_space *mapping = page->mapping; page 2855 fs/ext4/inode.c ret = block_prepare_write(page, 0, PAGE_CACHE_SIZE, page 2860 fs/ext4/inode.c page_bufs = page_buffers(page); page 2865 fs/ext4/inode.c unlock_page(page); page 2890 fs/ext4/inode.c unlock_page(page); page 2898 fs/ext4/inode.c struct inode *inode = page->mapping->host; page 2902 fs/ext4/inode.c J_ASSERT(PageLocked(page)); page 2903 fs/ext4/inode.c if (page->index == size >> PAGE_CACHE_SHIFT) page 2908 fs/ext4/inode.c if (page_has_buffers(page)) { page 2918 fs/ext4/inode.c BUG_ON(walk_page_buffers(NULL, page_buffers(page), 0, len, NULL, page 2925 fs/ext4/inode.c if (PageChecked(page)) { page 2930 fs/ext4/inode.c ClearPageChecked(page); page 2931 fs/ext4/inode.c return __ext4_journalled_writepage(page, wbc); page 2938 fs/ext4/inode.c return block_write_full_page(page, page 2943 fs/ext4/inode.c redirty_page_for_writepage(wbc, page); page 2944 fs/ext4/inode.c unlock_page(page); page 2950 fs/ext4/inode.c return mpage_readpage(page, ext4_get_block); page 2962 fs/ext4/inode.c journal_t *journal = EXT4_JOURNAL(page->mapping->host); page 2968 fs/ext4/inode.c ClearPageChecked(page); page 2970 fs/ext4/inode.c jbd2_journal_invalidatepage(journal, page, offset); page 2975 fs/ext4/inode.c journal_t *journal = EXT4_JOURNAL(page->mapping->host); page 2977 fs/ext4/inode.c WARN_ON(PageChecked(page)); page 2978 fs/ext4/inode.c if (!page_has_buffers(page)) page 2980 fs/ext4/inode.c return jbd2_journal_try_to_free_buffers(journal, page, wait); page 3081 fs/ext4/inode.c SetPageChecked(page); page 3082 fs/ext4/inode.c return __set_page_dirty_nobuffers(page); page 3176 fs/ext4/inode.c struct page *page; page 3179 fs/ext4/inode.c page = grab_cache_page(mapping, from >> PAGE_CACHE_SHIFT); page 3180 fs/ext4/inode.c if (!page) page 3191 fs/ext4/inode.c if (!page_has_buffers(page) && test_opt(inode->i_sb, NOBH) && page 3192 fs/ext4/inode.c ext4_should_writeback_data(inode) && PageUptodate(page)) { page 3193 fs/ext4/inode.c zero_user(page, offset, length); page 3194 fs/ext4/inode.c set_page_dirty(page); page 3198 fs/ext4/inode.c if (!page_has_buffers(page)) page 3199 fs/ext4/inode.c create_empty_buffers(page, blocksize, 0); page 3202 fs/ext4/inode.c bh = page_buffers(page); page 3227 fs/ext4/inode.c if (PageUptodate(page)) page 3246 fs/ext4/inode.c zero_user(page, offset, length); page 3260 fs/ext4/inode.c unlock_page(page); page 3261 fs/ext4/inode.c page_cache_release(page); page 4952 fs/ext4/inode.c if (page->mapping != mapping || size <= page_offset(page) page 4953 fs/ext4/inode.c || !PageUptodate(page)) { page 4958 fs/ext4/inode.c if (PageMappedToDisk(page)) page 4961 fs/ext4/inode.c if (page->index == size >> PAGE_CACHE_SHIFT) page 4966 fs/ext4/inode.c if (page_has_buffers(page)) { page 4968 fs/ext4/inode.c if (!walk_page_buffers(NULL, page_buffers(page), 0, len, NULL, page 4979 fs/ext4/inode.c ret = mapping->a_ops->write_begin(file, mapping, page_offset(page), page 4980 fs/ext4/inode.c len, AOP_FLAG_UNINTERRUPTIBLE, &page, &fsdata); page 4983 fs/ext4/inode.c ret = mapping->a_ops->write_end(file, mapping, page_offset(page), page 4984 fs/ext4/inode.c len, len, page, fsdata); page 745 fs/ext4/mballoc.c mb_debug("init page %lu\n", page->index); page 747 fs/ext4/mballoc.c inode = page->mapping->host; page 766 fs/ext4/mballoc.c first_group = page->index * blocks_per_page / 2; page 816 fs/ext4/mballoc.c first_block = page->index * blocks_per_page; page 831 fs/ext4/mballoc.c data = page_address(page) + (i * blocksize); page 842 fs/ext4/mballoc.c group, page->index, i * blocksize); page 857 fs/ext4/mballoc.c group, page->index, i * blocksize); page 873 fs/ext4/mballoc.c SetPageUptodate(page); page 895 fs/ext4/mballoc.c struct page *page; page 920 fs/ext4/mballoc.c page = find_get_page(inode->i_mapping, pnum); page 921 fs/ext4/mballoc.c if (page == NULL || !PageUptodate(page)) { page 922 fs/ext4/mballoc.c if (page) page 923 fs/ext4/mballoc.c page_cache_release(page); page 924 fs/ext4/mballoc.c page = find_or_create_page(inode->i_mapping, pnum, GFP_NOFS); page 925 fs/ext4/mballoc.c if (page) { page 926 fs/ext4/mballoc.c BUG_ON(page->mapping != inode->i_mapping); page 927 fs/ext4/mballoc.c if (!PageUptodate(page)) { page 928 fs/ext4/mballoc.c ret = ext4_mb_init_cache(page, NULL); page 930 fs/ext4/mballoc.c unlock_page(page); page 933 fs/ext4/mballoc.c mb_cmp_bitmaps(e4b, page_address(page) + page 936 fs/ext4/mballoc.c unlock_page(page); page 939 fs/ext4/mballoc.c if (page == NULL || !PageUptodate(page)) { page 943 fs/ext4/mballoc.c e4b->bd_bitmap_page = page; page 944 fs/ext4/mballoc.c e4b->bd_bitmap = page_address(page) + (poff * sb->s_blocksize); page 945 fs/ext4/mballoc.c mark_page_accessed(page); page 951 fs/ext4/mballoc.c page = find_get_page(inode->i_mapping, pnum); page 952 fs/ext4/mballoc.c if (page == NULL || !PageUptodate(page)) { page 953 fs/ext4/mballoc.c if (page) page 954 fs/ext4/mballoc.c page_cache_release(page); page 955 fs/ext4/mballoc.c page = find_or_create_page(inode->i_mapping, pnum, GFP_NOFS); page 956 fs/ext4/mballoc.c if (page) { page 957 fs/ext4/mballoc.c BUG_ON(page->mapping != inode->i_mapping); page 958 fs/ext4/mballoc.c if (!PageUptodate(page)) { page 959 fs/ext4/mballoc.c ret = ext4_mb_init_cache(page, e4b->bd_bitmap); page 961 fs/ext4/mballoc.c unlock_page(page); page 965 fs/ext4/mballoc.c unlock_page(page); page 968 fs/ext4/mballoc.c if (page == NULL || !PageUptodate(page)) { page 972 fs/ext4/mballoc.c e4b->bd_buddy_page = page; page 973 fs/ext4/mballoc.c e4b->bd_buddy = page_address(page) + (poff * sb->s_blocksize); page 974 fs/ext4/mballoc.c mark_page_accessed(page); page 2340 fs/ext4/mballoc.c struct page *page; page 2358 fs/ext4/mballoc.c page = find_get_page(inode->i_mapping, pnum); page 2359 fs/ext4/mballoc.c if (page != NULL) { page 2360 fs/ext4/mballoc.c ClearPageUptodate(page); page 2361 fs/ext4/mballoc.c page_cache_release(page); page 2367 fs/ext4/mballoc.c page = find_get_page(inode->i_mapping, pnum); page 2368 fs/ext4/mballoc.c if (page != NULL) { page 2369 fs/ext4/mballoc.c ClearPageUptodate(page); page 2370 fs/ext4/mballoc.c page_cache_release(page); page 210 fs/ext4/mballoc.h struct page *ac_bitmap_page; page 211 fs/ext4/mballoc.h struct page *ac_buddy_page; page 237 fs/ext4/mballoc.h struct page *bd_buddy_page; page 239 fs/ext4/mballoc.h struct page *bd_bitmap_page; page 1102 fs/ext4/resize.c struct page *page; page 1108 fs/ext4/resize.c page = find_get_page(inode->i_mapping, pnum); page 1109 fs/ext4/resize.c if (page != NULL) { page 1110 fs/ext4/resize.c ClearPageUptodate(page); page 1111 fs/ext4/resize.c page_cache_release(page); page 1117 fs/ext4/resize.c page = find_get_page(inode->i_mapping, pnum); page 1118 fs/ext4/resize.c if (page != NULL) { page 1119 fs/ext4/resize.c ClearPageUptodate(page); page 1120 fs/ext4/resize.c page_cache_release(page); page 124 fs/fat/inode.c return block_write_full_page(page, fat_get_block, wbc); page 135 fs/fat/inode.c return mpage_readpage(page, fat_get_block); page 77 fs/freevxfs/vxfs_extern.h extern struct page * vxfs_get_page(struct address_space *, u_long); page 132 fs/freevxfs/vxfs_inode.c struct page *pp; page 115 fs/freevxfs/vxfs_lookup.c u_long npages, page, nblocks, pblocks, block; page 124 fs/freevxfs/vxfs_lookup.c for (page = 0; page < npages; page++) { page 126 fs/freevxfs/vxfs_lookup.c struct page *pp; page 128 fs/freevxfs/vxfs_lookup.c pp = vxfs_get_page(ip->i_mapping, page); page 177 fs/freevxfs/vxfs_lookup.c struct page *pp; page 246 fs/freevxfs/vxfs_lookup.c u_long page, npages, block, pblocks, nblocks, offset; page 275 fs/freevxfs/vxfs_lookup.c page = pos >> PAGE_CACHE_SHIFT; page 279 fs/freevxfs/vxfs_lookup.c for (; page < npages; page++, block = 0) { page 281 fs/freevxfs/vxfs_lookup.c struct page *pp; page 283 fs/freevxfs/vxfs_lookup.c pp = vxfs_get_page(ip->i_mapping, page); page 312 fs/freevxfs/vxfs_lookup.c ((page << PAGE_CACHE_SHIFT) | offset) + 2, page 326 fs/freevxfs/vxfs_lookup.c fp->f_pos = ((page << PAGE_CACHE_SHIFT) | offset) + 2; page 69 fs/freevxfs/vxfs_subr.c struct page * page 72 fs/freevxfs/vxfs_subr.c struct page * pp; page 163 fs/freevxfs/vxfs_subr.c return block_read_full_page(page, vxfs_getblk); page 495 fs/fuse/dev.c struct page *pg; page 586 fs/fuse/dev.c if (page && zeroing && count < PAGE_SIZE) { page 587 fs/fuse/dev.c void *mapaddr = kmap_atomic(page, KM_USER1); page 595 fs/fuse/dev.c if (page) { page 596 fs/fuse/dev.c void *mapaddr = kmap_atomic(page, KM_USER1); page 603 fs/fuse/dev.c if (page && !cs->write) page 604 fs/fuse/dev.c flush_dcache_page(page); page 618 fs/fuse/dev.c struct page *page = req->pages[i]; page 619 fs/fuse/dev.c int err = fuse_copy_page(cs, page, offset, count, zeroing); page 1009 fs/fuse/dir.c struct page *page; page 1021 fs/fuse/dir.c page = alloc_page(GFP_KERNEL); page 1022 fs/fuse/dir.c if (!page) { page 1027 fs/fuse/dir.c req->pages[0] = page; page 1034 fs/fuse/dir.c err = parse_dirfile(page_address(page), nbytes, file, dstbuf, page 1037 fs/fuse/dir.c __free_page(page); page 417 fs/fuse/file.c struct inode *inode = page->mapping->host; page 421 fs/fuse/file.c loff_t pos = page_offset(page); page 435 fs/fuse/file.c fuse_wait_on_page_writeback(inode, page->index); page 446 fs/fuse/file.c req->pages[0] = page; page 458 fs/fuse/file.c SetPageUptodate(page); page 463 fs/fuse/file.c unlock_page(page); page 485 fs/fuse/file.c struct page *page = req->pages[i]; page 487 fs/fuse/file.c SetPageUptodate(page); page 489 fs/fuse/file.c SetPageError(page); page 490 fs/fuse/file.c unlock_page(page); page 530 fs/fuse/file.c fuse_wait_on_page_writeback(inode, page->index); page 535 fs/fuse/file.c req->pages[req->num_pages - 1]->index + 1 != page->index)) { page 539 fs/fuse/file.c unlock_page(page); page 543 fs/fuse/file.c req->pages[req->num_pages] = page; page 681 fs/fuse/file.c fuse_wait_on_page_writeback(inode, page->index); page 688 fs/fuse/file.c req->pages[0] = page; page 699 fs/fuse/file.c SetPageUptodate(page); page 713 fs/fuse/file.c res = fuse_buffered_write(file, inode, pos, copied, page); page 715 fs/fuse/file.c unlock_page(page); page 716 fs/fuse/file.c page_cache_release(page); page 736 fs/fuse/file.c struct page *page = req->pages[i]; page 739 fs/fuse/file.c SetPageUptodate(page); page 747 fs/fuse/file.c unlock_page(page); page 748 fs/fuse/file.c page_cache_release(page); page 767 fs/fuse/file.c struct page *page; page 780 fs/fuse/file.c page = __grab_cache_page(mapping, index); page 781 fs/fuse/file.c if (!page) page 785 fs/fuse/file.c tmp = iov_iter_copy_from_user_atomic(page, ii, offset, bytes); page 787 fs/fuse/file.c flush_dcache_page(page); page 790 fs/fuse/file.c unlock_page(page); page 791 fs/fuse/file.c page_cache_release(page); page 797 fs/fuse/file.c req->pages[req->num_pages] = page; page 919 fs/fuse/file.c struct page *page = req->pages[i]; page 921 fs/fuse/file.c set_page_dirty_lock(page); page 922 fs/fuse/file.c put_page(page); page 1124 fs/fuse/file.c struct address_space *mapping = page->mapping; page 1130 fs/fuse/file.c struct page *tmp_page; page 1132 fs/fuse/file.c set_page_writeback(page); page 1148 fs/fuse/file.c fuse_write_fill(req, NULL, ff, inode, page_offset(page), 0, 1); page 1150 fs/fuse/file.c copy_highpage(tmp_page, page); page 1159 fs/fuse/file.c end_page_writeback(page); page 1172 fs/fuse/file.c end_page_writeback(page); page 1180 fs/fuse/file.c err = fuse_writepage_locked(page); page 1181 fs/fuse/file.c unlock_page(page); page 1189 fs/fuse/file.c if (clear_page_dirty_for_io(page)) { page 1190 fs/fuse/file.c struct inode *inode = page->mapping->host; page 1191 fs/fuse/file.c err = fuse_writepage_locked(page); page 1193 fs/fuse/file.c fuse_wait_on_page_writeback(inode, page->index); page 1230 fs/fuse/file.c fuse_wait_on_page_writeback(inode, page->index); page 254 fs/fuse/fuse_i.h struct page *pages[FUSE_MAX_PAGES_PER_REQ]; page 67 fs/gfs2/bmap.c if (!page || page->index) { page 68 fs/gfs2/bmap.c page = grab_cache_page(inode->i_mapping, 0); page 69 fs/gfs2/bmap.c if (!page) page 74 fs/gfs2/bmap.c if (!PageUptodate(page)) { page 75 fs/gfs2/bmap.c void *kaddr = kmap(page); page 81 fs/gfs2/bmap.c kunmap(page); page 83 fs/gfs2/bmap.c SetPageUptodate(page); page 86 fs/gfs2/bmap.c if (!page_has_buffers(page)) page 87 fs/gfs2/bmap.c create_empty_buffers(page, 1 << inode->i_blkbits, page 90 fs/gfs2/bmap.c bh = page_buffers(page); page 102 fs/gfs2/bmap.c unlock_page(page); page 103 fs/gfs2/bmap.c page_cache_release(page); page 150 fs/gfs2/bmap.c error = gfs2_unstuffer_page(ip, dibh, block, page); page 962 fs/gfs2/bmap.c struct page *page; page 965 fs/gfs2/bmap.c page = grab_cache_page(mapping, index); page 966 fs/gfs2/bmap.c if (!page) page 973 fs/gfs2/bmap.c if (!page_has_buffers(page)) page 974 fs/gfs2/bmap.c create_empty_buffers(page, blocksize, 0); page 977 fs/gfs2/bmap.c bh = page_buffers(page); page 995 fs/gfs2/bmap.c if (PageUptodate(page)) page 1011 fs/gfs2/bmap.c zero_user(page, offset, length); page 1014 fs/gfs2/bmap.c unlock_page(page); page 1015 fs/gfs2/bmap.c page_cache_release(page); page 15 fs/gfs2/bmap.h struct page; page 47 fs/gfs2/meta_io.c return block_write_full_page(page, aspace_get_block, wbc); page 144 fs/gfs2/meta_io.c struct page *page; page 156 fs/gfs2/meta_io.c page = grab_cache_page(mapping, index); page 157 fs/gfs2/meta_io.c if (page) page 162 fs/gfs2/meta_io.c page = find_lock_page(mapping, index); page 163 fs/gfs2/meta_io.c if (!page) page 167 fs/gfs2/meta_io.c if (!page_has_buffers(page)) page 168 fs/gfs2/meta_io.c create_empty_buffers(page, sdp->sd_sb.sb_bsize, 0); page 171 fs/gfs2/meta_io.c for (bh = page_buffers(page); bufnum--; bh = bh->b_this_page) page 178 fs/gfs2/meta_io.c unlock_page(page); page 179 fs/gfs2/meta_io.c mark_page_accessed(page); page 180 fs/gfs2/meta_io.c page_cache_release(page); page 44 fs/gfs2/ops_address.c struct buffer_head *head = page_buffers(page); page 100 fs/gfs2/ops_address.c struct inode *inode = page->mapping->host; page 113 fs/gfs2/ops_address.c if (page->index > end_index || (page->index == end_index && !offset)) { page 114 fs/gfs2/ops_address.c page->mapping->a_ops->invalidatepage(page, 0); page 119 fs/gfs2/ops_address.c redirty_page_for_writepage(wbc, page); page 121 fs/gfs2/ops_address.c unlock_page(page); page 137 fs/gfs2/ops_address.c ret = gfs2_writepage_common(page, wbc); page 141 fs/gfs2/ops_address.c ret = mpage_writepage(page, gfs2_get_block_noalloc, wbc); page 143 fs/gfs2/ops_address.c ret = block_write_full_page(page, gfs2_get_block_noalloc, wbc); page 157 fs/gfs2/ops_address.c struct inode *inode = page->mapping->host; page 161 fs/gfs2/ops_address.c ret = gfs2_writepage_common(page, wbc); page 165 fs/gfs2/ops_address.c if (!page_has_buffers(page)) { page 166 fs/gfs2/ops_address.c create_empty_buffers(page, inode->i_sb->s_blocksize, page 169 fs/gfs2/ops_address.c gfs2_page_add_databufs(ip, page, 0, inode->i_sb->s_blocksize-1); page 170 fs/gfs2/ops_address.c return block_write_full_page(page, gfs2_get_block_noalloc, wbc); page 186 fs/gfs2/ops_address.c struct inode *inode = page->mapping->host; page 190 fs/gfs2/ops_address.c if (PageChecked(page)) { page 191 fs/gfs2/ops_address.c ClearPageChecked(page); page 192 fs/gfs2/ops_address.c if (!page_has_buffers(page)) { page 193 fs/gfs2/ops_address.c create_empty_buffers(page, inode->i_sb->s_blocksize, page 196 fs/gfs2/ops_address.c gfs2_page_add_databufs(ip, page, 0, sdp->sd_vfs->s_blocksize-1); page 198 fs/gfs2/ops_address.c return block_write_full_page(page, gfs2_get_block_noalloc, wbc); page 211 fs/gfs2/ops_address.c struct inode *inode = page->mapping->host; page 216 fs/gfs2/ops_address.c error = gfs2_writepage_common(page, wbc); page 220 fs/gfs2/ops_address.c if (PageChecked(page)) { page 228 fs/gfs2/ops_address.c error = __gfs2_jdata_writepage(page, wbc); page 234 fs/gfs2/ops_address.c redirty_page_for_writepage(wbc, page); page 235 fs/gfs2/ops_address.c unlock_page(page); page 285 fs/gfs2/ops_address.c struct page *page = pvec->pages[i]; page 287 fs/gfs2/ops_address.c lock_page(page); page 289 fs/gfs2/ops_address.c if (unlikely(page->mapping != mapping)) { page 290 fs/gfs2/ops_address.c unlock_page(page); page 294 fs/gfs2/ops_address.c if (!wbc->range_cyclic && page->index > end) { page 296 fs/gfs2/ops_address.c unlock_page(page); page 301 fs/gfs2/ops_address.c wait_on_page_writeback(page); page 303 fs/gfs2/ops_address.c if (PageWriteback(page) || page 304 fs/gfs2/ops_address.c !clear_page_dirty_for_io(page)) { page 305 fs/gfs2/ops_address.c unlock_page(page); page 310 fs/gfs2/ops_address.c if (page->index > end_index || (page->index == end_index && !offset)) { page 311 fs/gfs2/ops_address.c page->mapping->a_ops->invalidatepage(page, 0); page 312 fs/gfs2/ops_address.c unlock_page(page); page 316 fs/gfs2/ops_address.c ret = __gfs2_jdata_writepage(page, wbc); page 445 fs/gfs2/ops_address.c if (unlikely(page->index)) { page 446 fs/gfs2/ops_address.c zero_user(page, 0, PAGE_CACHE_SIZE); page 454 fs/gfs2/ops_address.c kaddr = kmap_atomic(page, KM_USER0); page 459 fs/gfs2/ops_address.c flush_dcache_page(page); page 461 fs/gfs2/ops_address.c SetPageUptodate(page); page 480 fs/gfs2/ops_address.c struct gfs2_inode *ip = GFS2_I(page->mapping->host); page 481 fs/gfs2/ops_address.c struct gfs2_sbd *sdp = GFS2_SB(page->mapping->host); page 485 fs/gfs2/ops_address.c error = stuffed_readpage(ip, page); page 486 fs/gfs2/ops_address.c unlock_page(page); page 488 fs/gfs2/ops_address.c error = mpage_readpage(page, gfs2_block_map); page 509 fs/gfs2/ops_address.c struct address_space *mapping = page->mapping; page 514 fs/gfs2/ops_address.c unlock_page(page); page 520 fs/gfs2/ops_address.c lock_page(page); page 521 fs/gfs2/ops_address.c if (page->mapping == mapping && !PageUptodate(page)) page 522 fs/gfs2/ops_address.c error = __gfs2_readpage(file, page); page 524 fs/gfs2/ops_address.c unlock_page(page); page 529 fs/gfs2/ops_address.c lock_page(page); page 551 fs/gfs2/ops_address.c struct page *page; page 558 fs/gfs2/ops_address.c page = read_cache_page(mapping, index, __gfs2_readpage, NULL); page 559 fs/gfs2/ops_address.c if (IS_ERR(page)) page 560 fs/gfs2/ops_address.c return PTR_ERR(page); page 561 fs/gfs2/ops_address.c p = kmap_atomic(page, KM_USER0); page 564 fs/gfs2/ops_address.c mark_page_accessed(page); page 565 fs/gfs2/ops_address.c page_cache_release(page); page 637 fs/gfs2/ops_address.c struct page *page; page 678 fs/gfs2/ops_address.c page = __grab_cache_page(mapping, index); page 679 fs/gfs2/ops_address.c *pagep = page; page 680 fs/gfs2/ops_address.c if (unlikely(!page)) page 686 fs/gfs2/ops_address.c error = gfs2_unstuff_dinode(ip, page); page 689 fs/gfs2/ops_address.c } else if (!PageUptodate(page)) { page 690 fs/gfs2/ops_address.c error = stuffed_readpage(ip, page); page 696 fs/gfs2/ops_address.c error = block_prepare_write(page, from, to, gfs2_block_map); page 701 fs/gfs2/ops_address.c page_cache_release(page); page 772 fs/gfs2/ops_address.c kaddr = kmap_atomic(page, KM_USER0); page 775 fs/gfs2/ops_address.c flush_dcache_page(page); page 778 fs/gfs2/ops_address.c if (!PageUptodate(page)) page 779 fs/gfs2/ops_address.c SetPageUptodate(page); page 780 fs/gfs2/ops_address.c unlock_page(page); page 781 fs/gfs2/ops_address.c page_cache_release(page); page 821 fs/gfs2/ops_address.c struct inode *inode = page->mapping->host; page 835 fs/gfs2/ops_address.c unlock_page(page); page 836 fs/gfs2/ops_address.c page_cache_release(page); page 843 fs/gfs2/ops_address.c return gfs2_stuffed_write_end(inode, dibh, pos, len, copied, page); page 846 fs/gfs2/ops_address.c gfs2_page_add_databufs(ip, page, from, to); page 848 fs/gfs2/ops_address.c ret = generic_write_end(file, mapping, pos, len, copied, page, fsdata); page 882 fs/gfs2/ops_address.c SetPageChecked(page); page 883 fs/gfs2/ops_address.c return __set_page_dirty_buffers(page); page 937 fs/gfs2/ops_address.c struct gfs2_sbd *sdp = GFS2_SB(page->mapping->host); page 941 fs/gfs2/ops_address.c BUG_ON(!PageLocked(page)); page 943 fs/gfs2/ops_address.c ClearPageChecked(page); page 944 fs/gfs2/ops_address.c if (!page_has_buffers(page)) page 947 fs/gfs2/ops_address.c bh = head = page_buffers(page); page 956 fs/gfs2/ops_address.c try_to_release_page(page, 0); page 1033 fs/gfs2/ops_address.c struct inode *aspace = page->mapping->host; page 1038 fs/gfs2/ops_address.c if (!page_has_buffers(page)) page 1042 fs/gfs2/ops_address.c head = bh = page_buffers(page); page 1055 fs/gfs2/ops_address.c head = bh = page_buffers(page); page 1079 fs/gfs2/ops_address.c return try_to_free_buffers(page); page 315 fs/gfs2/ops_file.c struct inode *inode = page->mapping->host; page 318 fs/gfs2/ops_file.c u64 lblock = page->index << (PAGE_CACHE_SHIFT - inode->i_blkbits); page 347 fs/gfs2/ops_file.c u64 pos = page->index << (PAGE_CACHE_SIZE - inode->i_blkbits); page 386 fs/gfs2/ops_file.c lock_page(page); page 389 fs/gfs2/ops_file.c if (page->index > last_index) page 392 fs/gfs2/ops_file.c if (!PageUptodate(page) || page->mapping != ip->i_inode.i_mapping) page 395 fs/gfs2/ops_file.c ret = gfs2_unstuff_dinode(ip, page); page 399 fs/gfs2/ops_file.c ret = gfs2_allocate_page_backing(page); page 402 fs/gfs2/ops_file.c unlock_page(page); page 219 fs/gfs2/ops_fstype.c struct page *page = bio->bi_private; page 222 fs/gfs2/ops_fstype.c SetPageUptodate(page); page 225 fs/gfs2/ops_fstype.c unlock_page(page); page 271 fs/gfs2/ops_fstype.c struct page *page; page 274 fs/gfs2/ops_fstype.c page = alloc_page(GFP_NOFS); page 275 fs/gfs2/ops_fstype.c if (unlikely(!page)) page 278 fs/gfs2/ops_fstype.c ClearPageUptodate(page); page 279 fs/gfs2/ops_fstype.c ClearPageDirty(page); page 280 fs/gfs2/ops_fstype.c lock_page(page); page 284 fs/gfs2/ops_fstype.c __free_page(page); page 290 fs/gfs2/ops_fstype.c bio_add_page(bio, page, PAGE_SIZE, 0); page 293 fs/gfs2/ops_fstype.c bio->bi_private = page; page 295 fs/gfs2/ops_fstype.c wait_on_page_locked(page); page 297 fs/gfs2/ops_fstype.c if (!PageUptodate(page)) { page 298 fs/gfs2/ops_fstype.c __free_page(page); page 301 fs/gfs2/ops_fstype.c p = kmap(page); page 303 fs/gfs2/ops_fstype.c kunmap(page); page 304 fs/gfs2/ops_fstype.c __free_page(page); page 612 fs/gfs2/quota.c struct page *page; page 622 fs/gfs2/quota.c page = grab_cache_page(mapping, index); page 623 fs/gfs2/quota.c if (!page) page 629 fs/gfs2/quota.c if (!page_has_buffers(page)) page 630 fs/gfs2/quota.c create_empty_buffers(page, blocksize, 0); page 632 fs/gfs2/quota.c bh = page_buffers(page); page 646 fs/gfs2/quota.c if (PageUptodate(page)) page 658 fs/gfs2/quota.c kaddr = kmap_atomic(page, KM_USER0); page 664 fs/gfs2/quota.c flush_dcache_page(page); page 672 fs/gfs2/quota.c unlock_page(page); page 673 fs/gfs2/quota.c page_cache_release(page); page 19 fs/hfs/bnode.c struct page *page; page 22 fs/hfs/bnode.c page = node->page[0]; page 24 fs/hfs/bnode.c memcpy(buf, kmap(page) + off, len); page 25 fs/hfs/bnode.c kunmap(page); page 61 fs/hfs/bnode.c struct page *page; page 64 fs/hfs/bnode.c page = node->page[0]; page 66 fs/hfs/bnode.c memcpy(kmap(page) + off, buf, len); page 67 fs/hfs/bnode.c kunmap(page); page 68 fs/hfs/bnode.c set_page_dirty(page); page 86 fs/hfs/bnode.c struct page *page; page 89 fs/hfs/bnode.c page = node->page[0]; page 91 fs/hfs/bnode.c memset(kmap(page) + off, 0, len); page 92 fs/hfs/bnode.c kunmap(page); page 93 fs/hfs/bnode.c set_page_dirty(page); page 100 fs/hfs/bnode.c struct page *src_page, *dst_page; page 108 fs/hfs/bnode.c src_page = src_node->page[0]; page 109 fs/hfs/bnode.c dst_page = dst_node->page[0]; page 119 fs/hfs/bnode.c struct page *page; page 127 fs/hfs/bnode.c page = node->page[0]; page 128 fs/hfs/bnode.c ptr = kmap(page); page 130 fs/hfs/bnode.c kunmap(page); page 131 fs/hfs/bnode.c set_page_dirty(page); page 240 fs/hfs/bnode.c struct page *page; page 251 fs/hfs/bnode.c sizeof(struct page *); page 282 fs/hfs/bnode.c page = read_mapping_page(mapping, block++, NULL); page 283 fs/hfs/bnode.c if (IS_ERR(page)) page 285 fs/hfs/bnode.c if (PageError(page)) { page 286 fs/hfs/bnode.c page_cache_release(page); page 289 fs/hfs/bnode.c page_cache_release(page); page 290 fs/hfs/bnode.c node->page[i] = page; page 340 fs/hfs/bnode.c desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset); page 346 fs/hfs/bnode.c kunmap(node->page[0]); page 410 fs/hfs/bnode.c struct page **pagep; page 425 fs/hfs/bnode.c pagep = node->page; page 463 fs/hfs/bnode.c if (!node->page[i]) page 465 fs/hfs/bnode.c mark_page_accessed(node->page[i]); page 22 fs/hfs/btree.c struct page *page; page 62 fs/hfs/btree.c page = read_mapping_page(mapping, 0, NULL); page 63 fs/hfs/btree.c if (IS_ERR(page)) page 67 fs/hfs/btree.c head = (struct hfs_btree_header_rec *)(kmap(page) + sizeof(struct hfs_bnode_desc)); page 106 fs/hfs/btree.c kunmap(page); page 107 fs/hfs/btree.c page_cache_release(page); page 111 fs/hfs/btree.c page_cache_release(page); page 147 fs/hfs/btree.c struct page *page; page 154 fs/hfs/btree.c page = node->page[0]; page 155 fs/hfs/btree.c head = (struct hfs_btree_header_rec *)(kmap(page) + sizeof(struct hfs_bnode_desc)); page 166 fs/hfs/btree.c kunmap(page); page 167 fs/hfs/btree.c set_page_dirty(page); page 209 fs/hfs/btree.c struct page **pagep; page 244 fs/hfs/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); page 289 fs/hfs/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); page 298 fs/hfs/btree.c struct page *page; page 333 fs/hfs/btree.c page = node->page[off >> PAGE_CACHE_SHIFT]; page 334 fs/hfs/btree.c data = kmap(page); page 340 fs/hfs/btree.c kunmap(page); page 345 fs/hfs/btree.c set_page_dirty(page); page 346 fs/hfs/btree.c kunmap(page); page 62 fs/hfs/btree.h struct page *page[0]; page 468 fs/hfs/extent.c struct page *page; page 474 fs/hfs/extent.c AOP_FLAG_UNINTERRUPTIBLE, &page, &fsdata); page 477 fs/hfs/extent.c page, fsdata); page 30 fs/hfs/inode.c return block_write_full_page(page, hfs_get_block, wbc); page 35 fs/hfs/inode.c return block_read_full_page(page, hfs_get_block); page 55 fs/hfs/inode.c struct inode *inode = page->mapping->host; page 74 fs/hfs/inode.c nidx = page->index >> (tree->node_size_shift - PAGE_CACHE_SHIFT); page 87 fs/hfs/inode.c nidx = page->index << (PAGE_CACHE_SHIFT - tree->node_size_shift); page 103 fs/hfs/inode.c return res ? try_to_free_buffers(page) : 0; page 20 fs/hfsplus/bitmap.c struct page *page; page 34 fs/hfsplus/bitmap.c page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, NULL); page 35 fs/hfsplus/bitmap.c pptr = kmap(page); page 70 fs/hfsplus/bitmap.c kunmap(page); page 74 fs/hfsplus/bitmap.c page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, page 76 fs/hfsplus/bitmap.c curr = pptr = kmap(page); page 118 fs/hfsplus/bitmap.c set_page_dirty(page); page 119 fs/hfsplus/bitmap.c kunmap(page); page 121 fs/hfsplus/bitmap.c page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, page 123 fs/hfsplus/bitmap.c pptr = kmap(page); page 138 fs/hfsplus/bitmap.c set_page_dirty(page); page 139 fs/hfsplus/bitmap.c kunmap(page); page 151 fs/hfsplus/bitmap.c struct page *page; page 169 fs/hfsplus/bitmap.c page = read_mapping_page(mapping, pnr, NULL); page 170 fs/hfsplus/bitmap.c pptr = kmap(page); page 199 fs/hfsplus/bitmap.c set_page_dirty(page); page 200 fs/hfsplus/bitmap.c kunmap(page); page 201 fs/hfsplus/bitmap.c page = read_mapping_page(mapping, ++pnr, NULL); page 202 fs/hfsplus/bitmap.c pptr = kmap(page); page 213 fs/hfsplus/bitmap.c set_page_dirty(page); page 214 fs/hfsplus/bitmap.c kunmap(page); page 23 fs/hfsplus/bnode.c struct page **pagep; page 27 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); page 75 fs/hfsplus/bnode.c struct page **pagep; page 79 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); page 105 fs/hfsplus/bnode.c struct page **pagep; page 109 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); page 129 fs/hfsplus/bnode.c struct page **src_page, **dst_page; page 138 fs/hfsplus/bnode.c src_page = src_node->page + (src >> PAGE_CACHE_SHIFT); page 140 fs/hfsplus/bnode.c dst_page = dst_node->page + (dst >> PAGE_CACHE_SHIFT); page 187 fs/hfsplus/bnode.c struct page **src_page, **dst_page; page 197 fs/hfsplus/bnode.c src_page = node->page + (src >> PAGE_CACHE_SHIFT); page 200 fs/hfsplus/bnode.c dst_page = node->page + (dst >> PAGE_CACHE_SHIFT); page 246 fs/hfsplus/bnode.c src_page = node->page + (src >> PAGE_CACHE_SHIFT); page 248 fs/hfsplus/bnode.c dst_page = node->page + (dst >> PAGE_CACHE_SHIFT); page 400 fs/hfsplus/bnode.c struct page *page; page 411 fs/hfsplus/bnode.c sizeof(struct page *); page 442 fs/hfsplus/bnode.c page = read_mapping_page(mapping, block, NULL); page 443 fs/hfsplus/bnode.c if (IS_ERR(page)) page 445 fs/hfsplus/bnode.c if (PageError(page)) { page 446 fs/hfsplus/bnode.c page_cache_release(page); page 449 fs/hfsplus/bnode.c page_cache_release(page); page 450 fs/hfsplus/bnode.c node->page[i] = page; page 500 fs/hfsplus/bnode.c desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset); page 506 fs/hfsplus/bnode.c kunmap(node->page[0]); page 570 fs/hfsplus/bnode.c struct page **pagep; page 589 fs/hfsplus/bnode.c pagep = node->page; page 627 fs/hfsplus/bnode.c if (!node->page[i]) page 629 fs/hfsplus/bnode.c mark_page_accessed(node->page[i]); page 26 fs/hfsplus/btree.c struct page *page; page 43 fs/hfsplus/btree.c page = read_mapping_page(mapping, 0, NULL); page 44 fs/hfsplus/btree.c if (IS_ERR(page)) page 48 fs/hfsplus/btree.c head = (struct hfs_btree_header_rec *)(kmap(page) + sizeof(struct hfs_bnode_desc)); page 85 fs/hfsplus/btree.c kunmap(page); page 86 fs/hfsplus/btree.c page_cache_release(page); page 91 fs/hfsplus/btree.c page_cache_release(page); page 125 fs/hfsplus/btree.c struct page *page; page 132 fs/hfsplus/btree.c page = node->page[0]; page 133 fs/hfsplus/btree.c head = (struct hfs_btree_header_rec *)(kmap(page) + sizeof(struct hfs_bnode_desc)); page 144 fs/hfsplus/btree.c kunmap(page); page 145 fs/hfsplus/btree.c set_page_dirty(page); page 185 fs/hfsplus/btree.c struct page **pagep; page 220 fs/hfsplus/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); page 265 fs/hfsplus/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); page 274 fs/hfsplus/btree.c struct page *page; page 310 fs/hfsplus/btree.c page = node->page[off >> PAGE_CACHE_SHIFT]; page 311 fs/hfsplus/btree.c data = kmap(page); page 317 fs/hfsplus/btree.c kunmap(page); page 322 fs/hfsplus/btree.c set_page_dirty(page); page 323 fs/hfsplus/btree.c kunmap(page); page 445 fs/hfsplus/extents.c struct page *page; page 452 fs/hfsplus/extents.c &page, &fsdata); page 455 fs/hfsplus/extents.c res = pagecache_write_end(NULL, mapping, size, 0, 0, page, fsdata); page 73 fs/hfsplus/hfsplus_fs.h struct page; page 93 fs/hfsplus/hfsplus_fs.h struct page *page[0]; page 389 fs/hfsplus/hfsplus_fs.h #define hfsplus_kmap(p) ({ struct page *__p = (p); kmap(__p); }) page 390 fs/hfsplus/hfsplus_fs.h #define hfsplus_kunmap(p) ({ struct page *__p = (p); kunmap(__p); __p; }) page 22 fs/hfsplus/inode.c return block_read_full_page(page, hfsplus_get_block); page 27 fs/hfsplus/inode.c return block_write_full_page(page, hfsplus_get_block, wbc); page 47 fs/hfsplus/inode.c struct inode *inode = page->mapping->host; page 71 fs/hfsplus/inode.c nidx = page->index >> (tree->node_size_shift - PAGE_CACHE_SHIFT); page 84 fs/hfsplus/inode.c nidx = page->index << (PAGE_CACHE_SHIFT - tree->node_size_shift); page 100 fs/hfsplus/inode.c return res ? try_to_free_buffers(page) : 0; page 438 fs/hostfs/hostfs_kern.c struct address_space *mapping = page->mapping; page 446 fs/hostfs/hostfs_kern.c if (page->index >= end_index) page 449 fs/hostfs/hostfs_kern.c buffer = kmap(page); page 450 fs/hostfs/hostfs_kern.c base = ((unsigned long long) page->index) << PAGE_CACHE_SHIFT; page 454 fs/hostfs/hostfs_kern.c ClearPageUptodate(page); page 461 fs/hostfs/hostfs_kern.c if (PageError(page)) page 462 fs/hostfs/hostfs_kern.c ClearPageError(page); page 466 fs/hostfs/hostfs_kern.c kunmap(page); page 468 fs/hostfs/hostfs_kern.c unlock_page(page); page 478 fs/hostfs/hostfs_kern.c start = (long long) page->index << PAGE_CACHE_SHIFT; page 479 fs/hostfs/hostfs_kern.c buffer = kmap(page); page 487 fs/hostfs/hostfs_kern.c flush_dcache_page(page); page 488 fs/hostfs/hostfs_kern.c SetPageUptodate(page); page 489 fs/hostfs/hostfs_kern.c if (PageError(page)) ClearPageError(page); page 492 fs/hostfs/hostfs_kern.c kunmap(page); page 493 fs/hostfs/hostfs_kern.c unlock_page(page); page 518 fs/hostfs/hostfs_kern.c buffer = kmap(page); page 520 fs/hostfs/hostfs_kern.c kunmap(page); page 522 fs/hostfs/hostfs_kern.c if (!PageUptodate(page) && err == PAGE_CACHE_SIZE) page 523 fs/hostfs/hostfs_kern.c SetPageUptodate(page); page 531 fs/hostfs/hostfs_kern.c unlock_page(page); page 532 fs/hostfs/hostfs_kern.c page_cache_release(page); page 941 fs/hostfs/hostfs_kern.c buffer = kmap(page); page 942 fs/hostfs/hostfs_kern.c name = inode_name(page->mapping->host, 0); page 950 fs/hostfs/hostfs_kern.c flush_dcache_page(page); page 951 fs/hostfs/hostfs_kern.c SetPageUptodate(page); page 952 fs/hostfs/hostfs_kern.c if (PageError(page)) ClearPageError(page); page 955 fs/hostfs/hostfs_kern.c kunmap(page); page 956 fs/hostfs/hostfs_kern.c unlock_page(page); page 87 fs/hpfs/file.c return block_write_full_page(page,hpfs_get_block, wbc); page 92 fs/hpfs/file.c return block_read_full_page(page,hpfs_get_block); page 513 fs/hpfs/namei.c char *link = kmap(page); page 514 fs/hpfs/namei.c struct inode *i = page->mapping->host; page 528 fs/hpfs/namei.c SetPageUptodate(page); page 529 fs/hpfs/namei.c kunmap(page); page 530 fs/hpfs/namei.c unlock_page(page); page 535 fs/hpfs/namei.c SetPageError(page); page 536 fs/hpfs/namei.c kunmap(page); page 537 fs/hpfs/namei.c unlock_page(page); page 208 fs/hugetlbfs/inode.c kaddr = kmap(&page[i]); page 210 fs/hugetlbfs/inode.c kunmap(&page[i]); page 253 fs/hugetlbfs/inode.c struct page *page; page 269 fs/hugetlbfs/inode.c page = find_get_page(mapping, index); page 270 fs/hugetlbfs/inode.c if (unlikely(page == NULL)) { page 282 fs/hugetlbfs/inode.c ret = hugetlbfs_read_actor(page, offset, buf, len, nr); page 287 fs/hugetlbfs/inode.c if (page) page 288 fs/hugetlbfs/inode.c page_cache_release(page); page 298 fs/hugetlbfs/inode.c if (page) page 299 fs/hugetlbfs/inode.c page_cache_release(page); page 317 fs/hugetlbfs/inode.c unlock_page(page); page 339 fs/hugetlbfs/inode.c cancel_dirty_page(page, /* No IO accounting for huge pages? */0); page 340 fs/hugetlbfs/inode.c ClearPageUptodate(page); page 341 fs/hugetlbfs/inode.c remove_from_page_cache(page); page 342 fs/hugetlbfs/inode.c put_page(page); page 365 fs/hugetlbfs/inode.c struct page *page = pvec.pages[i]; page 367 fs/hugetlbfs/inode.c lock_page(page); page 368 fs/hugetlbfs/inode.c if (page->index > next) page 369 fs/hugetlbfs/inode.c next = page->index; page 371 fs/hugetlbfs/inode.c truncate_huge_page(page); page 372 fs/hugetlbfs/inode.c unlock_page(page); page 612 fs/hugetlbfs/inode.c struct page *head = compound_head(page); page 62 fs/isofs/compress.c struct page *pages[zisofs_block_pages]; page 63 fs/isofs/compress.c unsigned long index = page->index; page 69 fs/isofs/compress.c pages[xpage] = page; page 80 fs/isofs/compress.c if (page->index >= maxpage) { page 81 fs/isofs/compress.c SetPageUptodate(page); page 82 fs/isofs/compress.c unlock_page(page); page 92 fs/isofs/compress.c page = pages[i]; page 93 fs/isofs/compress.c if ( page ) { page 94 fs/isofs/compress.c ClearPageError(page); page 95 fs/isofs/compress.c kmap(page); page 159 fs/isofs/compress.c if ( (page = pages[fpage]) != NULL ) { page 160 fs/isofs/compress.c memset(page_address(page), 0, PAGE_CACHE_SIZE); page 162 fs/isofs/compress.c flush_dcache_page(page); page 163 fs/isofs/compress.c SetPageUptodate(page); page 164 fs/isofs/compress.c kunmap(page); page 165 fs/isofs/compress.c unlock_page(page); page 169 fs/isofs/compress.c page_cache_release(page); page 220 fs/isofs/compress.c page = pages[fpage]; page 221 fs/isofs/compress.c if ( page ) page 222 fs/isofs/compress.c stream.next_out = page_address(page); page 279 fs/isofs/compress.c if ( page ) { page 280 fs/isofs/compress.c flush_dcache_page(page); page 281 fs/isofs/compress.c SetPageUptodate(page); page 282 fs/isofs/compress.c kunmap(page); page 283 fs/isofs/compress.c unlock_page(page); page 287 fs/isofs/compress.c page_cache_release(page); page 308 fs/isofs/compress.c page = pages[fpage]; page 309 fs/isofs/compress.c if ( page ) { page 310 fs/isofs/compress.c flush_dcache_page(page); page 312 fs/isofs/compress.c SetPageError(page); page 313 fs/isofs/compress.c kunmap(page); page 314 fs/isofs/compress.c unlock_page(page); page 316 fs/isofs/compress.c page_cache_release(page); page 1082 fs/isofs/inode.c return block_read_full_page(page,isofs_get_block); page 171 fs/isofs/namei.c struct page *page; page 175 fs/isofs/namei.c page = alloc_page(GFP_USER); page 176 fs/isofs/namei.c if (!page) page 182 fs/isofs/namei.c page_address(page), page 183 fs/isofs/namei.c 1024 + page_address(page)); page 184 fs/isofs/namei.c __free_page(page); page 663 fs/isofs/rock.c struct inode *inode = page->mapping->host; page 665 fs/isofs/rock.c char *link = kmap(page); page 753 fs/isofs/rock.c SetPageUptodate(page); page 754 fs/isofs/rock.c kunmap(page); page 755 fs/isofs/rock.c unlock_page(page); page 771 fs/isofs/rock.c SetPageError(page); page 772 fs/isofs/rock.c kunmap(page); page 773 fs/isofs/rock.c unlock_page(page); page 53 fs/jbd/commit.c struct page *page; page 59 fs/jbd/commit.c page = bh->b_page; page 60 fs/jbd/commit.c if (!page) page 62 fs/jbd/commit.c if (page->mapping) page 66 fs/jbd/commit.c if (!trylock_page(page)) page 69 fs/jbd/commit.c page_cache_get(page); page 71 fs/jbd/commit.c try_to_free_buffers(page); page 72 fs/jbd/commit.c unlock_page(page); page 73 fs/jbd/commit.c page_cache_release(page); page 287 fs/jbd/journal.c struct page *new_page; page 723 fs/jbd/transaction.c struct page *page; page 729 fs/jbd/transaction.c page = jh2bh(jh)->b_page; page 731 fs/jbd/transaction.c source = kmap_atomic(page, KM_USER0); page 1725 fs/jbd/transaction.c J_ASSERT(PageLocked(page)); page 1727 fs/jbd/transaction.c head = page_buffers(page); page 1749 fs/jbd/transaction.c ret = try_to_free_buffers(page); page 1767 fs/jbd/transaction.c ret = try_to_free_buffers(page); page 2003 fs/jbd/transaction.c if (!PageLocked(page)) page 2005 fs/jbd/transaction.c if (!page_has_buffers(page)) page 2012 fs/jbd/transaction.c head = bh = page_buffers(page); page 2029 fs/jbd/transaction.c if (may_free && try_to_free_buffers(page)) page 2030 fs/jbd/transaction.c J_ASSERT(!page_has_buffers(page)); page 58 fs/jbd2/commit.c struct page *page; page 64 fs/jbd2/commit.c page = bh->b_page; page 65 fs/jbd2/commit.c if (!page) page 67 fs/jbd2/commit.c if (page->mapping) page 71 fs/jbd2/commit.c if (!trylock_page(page)) page 74 fs/jbd2/commit.c page_cache_get(page); page 76 fs/jbd2/commit.c try_to_free_buffers(page); page 77 fs/jbd2/commit.c unlock_page(page); page 78 fs/jbd2/commit.c page_cache_release(page); page 299 fs/jbd2/commit.c struct page *page = bh->b_page; page 303 fs/jbd2/commit.c addr = kmap_atomic(page, KM_USER0); page 290 fs/jbd2/journal.c struct page *new_page; page 732 fs/jbd2/transaction.c struct page *page; page 738 fs/jbd2/transaction.c page = jh2bh(jh)->b_page; page 740 fs/jbd2/transaction.c source = kmap_atomic(page, KM_USER0); page 1543 fs/jbd2/transaction.c J_ASSERT(PageLocked(page)); page 1545 fs/jbd2/transaction.c head = page_buffers(page); page 1568 fs/jbd2/transaction.c ret = try_to_free_buffers(page); page 1586 fs/jbd2/transaction.c ret = try_to_free_buffers(page); page 1815 fs/jbd2/transaction.c if (!PageLocked(page)) page 1817 fs/jbd2/transaction.c if (!page_has_buffers(page)) page 1824 fs/jbd2/transaction.c head = bh = page_buffers(page); page 1841 fs/jbd2/transaction.c if (may_free && try_to_free_buffers(page)) page 1842 fs/jbd2/transaction.c J_ASSERT(!page_has_buffers(page)); page 128 fs/jffs2/file.c struct page *pg; page 657 fs/jffs2/fs.c struct page *pg; page 672 fs/jffs2/fs.c struct page *pg = (void *)*priv; page 270 fs/jfs/inode.c return block_write_full_page(page, jfs_get_block, wbc); page 281 fs/jfs/inode.c return mpage_readpage(page, jfs_get_block); page 508 fs/jfs/jfs_logmgr.c lsn = (log->page << L2LOGPSIZE) + dstoffset; page 530 fs/jfs/jfs_logmgr.c tblk->pn = log->page; page 540 fs/jfs/jfs_logmgr.c le16_to_cpu(lrd->type), log->bp, log->page, dstoffset); page 581 fs/jfs/jfs_logmgr.c pn = log->page; page 584 fs/jfs/jfs_logmgr.c lspn = le32_to_cpu(lp->h.page); page 645 fs/jfs/jfs_logmgr.c log->page = (pn == log->size - 1) ? 2 : pn + 1; page 649 fs/jfs/jfs_logmgr.c nextbp = lbmAllocate(log, log->page); page 655 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(lspn + 1); page 1356 fs/jfs/jfs_logmgr.c log->page = le32_to_cpu(logsuper->end) / LOGPSIZE; page 1357 fs/jfs/jfs_logmgr.c log->eor = le32_to_cpu(logsuper->end) - (LOGPSIZE * log->page); page 1363 fs/jfs/jfs_logmgr.c if ((rc = lbmRead(log, log->page, &bp))) page 1369 fs/jfs/jfs_logmgr.c le32_to_cpu(logsuper->end), log->page, log->eor, page 1373 fs/jfs/jfs_logmgr.c bp->l_pn = log->page; page 1633 fs/jfs/jfs_logmgr.c sizeof(long), mp->page, page 1634 fs/jfs/jfs_logmgr.c sizeof(struct page), 0); page 1708 fs/jfs/jfs_logmgr.c lsn, log->page, log->eor); page 1845 fs/jfs/jfs_logmgr.c struct page *page; page 1850 fs/jfs/jfs_logmgr.c page = virt_to_page(buffer); page 1859 fs/jfs/jfs_logmgr.c get_page(page); page 1862 fs/jfs/jfs_logmgr.c lbuf->l_page = page; page 2465 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(npages - 3); page 2485 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(lspn); page 124 fs/jfs/jfs_logmgr.h __le32 page; /* 4: log sequence page number */ page 132 fs/jfs/jfs_logmgr.h __le32 page; /* 4: normally the same as h.page */ page 386 fs/jfs/jfs_logmgr.h int page; /* 4: page number of eol page */ page 464 fs/jfs/jfs_logmgr.h struct page *l_page; /* The page itself */ page 60 fs/jfs/jfs_metapage.c unlock_page(mp->page); page 62 fs/jfs/jfs_metapage.c lock_page(mp->page); page 91 fs/jfs/jfs_metapage.c #define mp_anchor(page) ((struct meta_anchor *)page_private(page)) page 95 fs/jfs/jfs_metapage.c if (!PagePrivate(page)) page 97 fs/jfs/jfs_metapage.c return mp_anchor(page)->mp[offset >> L2PSIZE]; page 106 fs/jfs/jfs_metapage.c if (PagePrivate(page)) page 107 fs/jfs/jfs_metapage.c a = mp_anchor(page); page 112 fs/jfs/jfs_metapage.c set_page_private(page, (unsigned long)a); page 113 fs/jfs/jfs_metapage.c SetPagePrivate(page); page 114 fs/jfs/jfs_metapage.c kmap(page); page 118 fs/jfs/jfs_metapage.c l2mp_blocks = L2PSIZE - page->mapping->host->i_blkbits; page 129 fs/jfs/jfs_metapage.c struct meta_anchor *a = mp_anchor(page); page 130 fs/jfs/jfs_metapage.c int l2mp_blocks = L2PSIZE - page->mapping->host->i_blkbits; page 140 fs/jfs/jfs_metapage.c set_page_private(page, 0); page 141 fs/jfs/jfs_metapage.c ClearPagePrivate(page); page 142 fs/jfs/jfs_metapage.c kunmap(page); page 148 fs/jfs/jfs_metapage.c atomic_inc(&mp_anchor(page)->io_count); page 153 fs/jfs/jfs_metapage.c if (atomic_dec_and_test(&mp_anchor(page)->io_count)) page 154 fs/jfs/jfs_metapage.c handler(page); page 160 fs/jfs/jfs_metapage.c return PagePrivate(page) ? (struct metapage *)page_private(page) : NULL; page 166 fs/jfs/jfs_metapage.c set_page_private(page, (unsigned long)mp); page 167 fs/jfs/jfs_metapage.c SetPagePrivate(page); page 168 fs/jfs/jfs_metapage.c kmap(page); page 175 fs/jfs/jfs_metapage.c set_page_private(page, 0); page 176 fs/jfs/jfs_metapage.c ClearPagePrivate(page); page 177 fs/jfs/jfs_metapage.c kunmap(page); page 181 fs/jfs/jfs_metapage.c #define dec_io(page, handler) handler(page) page 244 fs/jfs/jfs_metapage.c remove_metapage(page, mp); page 280 fs/jfs/jfs_metapage.c if (!PageError(page)) page 281 fs/jfs/jfs_metapage.c SetPageUptodate(page); page 282 fs/jfs/jfs_metapage.c unlock_page(page); page 287 fs/jfs/jfs_metapage.c struct page *page = bio->bi_private; page 291 fs/jfs/jfs_metapage.c SetPageError(page); page 294 fs/jfs/jfs_metapage.c dec_io(page, last_read_complete); page 326 fs/jfs/jfs_metapage.c mp = page_to_mp(page, offset); page 337 fs/jfs/jfs_metapage.c end_page_writeback(page); page 342 fs/jfs/jfs_metapage.c struct page *page = bio->bi_private; page 344 fs/jfs/jfs_metapage.c BUG_ON(!PagePrivate(page)); page 348 fs/jfs/jfs_metapage.c SetPageError(page); page 350 fs/jfs/jfs_metapage.c dec_io(page, last_write_complete); page 358 fs/jfs/jfs_metapage.c struct inode *inode = page->mapping->host; page 373 fs/jfs/jfs_metapage.c page_start = (sector_t)page->index << page 375 fs/jfs/jfs_metapage.c BUG_ON(!PageLocked(page)); page 376 fs/jfs/jfs_metapage.c BUG_ON(PageWriteback(page)); page 377 fs/jfs/jfs_metapage.c set_page_writeback(page); page 380 fs/jfs/jfs_metapage.c mp = page_to_mp(page, offset); page 409 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, bio_bytes, bio_offset) < page 416 fs/jfs/jfs_metapage.c inc_io(page); page 423 fs/jfs/jfs_metapage.c inc_io(page); page 429 fs/jfs/jfs_metapage.c dec_io(page, last_write_complete); page 439 fs/jfs/jfs_metapage.c bio->bi_private = page; page 449 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, bio_bytes, bio_offset) < bio_bytes) page 458 fs/jfs/jfs_metapage.c redirty_page_for_writepage(wbc, page); page 460 fs/jfs/jfs_metapage.c unlock_page(page); page 463 fs/jfs/jfs_metapage.c end_page_writeback(page); page 475 fs/jfs/jfs_metapage.c unlock_page(page); page 476 fs/jfs/jfs_metapage.c dec_io(page, last_write_complete); page 483 fs/jfs/jfs_metapage.c struct inode *inode = page->mapping->host; page 493 fs/jfs/jfs_metapage.c BUG_ON(!PageLocked(page)); page 494 fs/jfs/jfs_metapage.c page_start = (sector_t)page->index << page 503 fs/jfs/jfs_metapage.c if (!PagePrivate(page)) page 504 fs/jfs/jfs_metapage.c insert_metapage(page, NULL); page 505 fs/jfs/jfs_metapage.c inc_io(page); page 513 fs/jfs/jfs_metapage.c bio->bi_private = page; page 516 fs/jfs/jfs_metapage.c if (bio_add_page(bio, page, len, offset) < len) page 525 fs/jfs/jfs_metapage.c unlock_page(page); page 532 fs/jfs/jfs_metapage.c dec_io(page, last_read_complete); page 543 fs/jfs/jfs_metapage.c mp = page_to_mp(page, offset); page 558 fs/jfs/jfs_metapage.c remove_metapage(page, mp); page 569 fs/jfs/jfs_metapage.c BUG_ON(PageWriteback(page)); page 571 fs/jfs/jfs_metapage.c metapage_releasepage(page, 0); page 591 fs/jfs/jfs_metapage.c struct page *page; page 622 fs/jfs/jfs_metapage.c page = grab_cache_page(mapping, page_index); page 623 fs/jfs/jfs_metapage.c if (!page) { page 627 fs/jfs/jfs_metapage.c SetPageUptodate(page); page 629 fs/jfs/jfs_metapage.c page = read_mapping_page(mapping, page_index, NULL); page 630 fs/jfs/jfs_metapage.c if (IS_ERR(page) || !PageUptodate(page)) { page 634 fs/jfs/jfs_metapage.c lock_page(page); page 637 fs/jfs/jfs_metapage.c mp = page_to_mp(page, page_offset); page 662 fs/jfs/jfs_metapage.c mp->page = page; page 668 fs/jfs/jfs_metapage.c mp->data = page_address(page) + page_offset; page 670 fs/jfs/jfs_metapage.c if (unlikely(insert_metapage(page, mp))) { page 682 fs/jfs/jfs_metapage.c unlock_page(page); page 687 fs/jfs/jfs_metapage.c unlock_page(page); page 694 fs/jfs/jfs_metapage.c page_cache_get(mp->page); page 695 fs/jfs/jfs_metapage.c lock_page(mp->page); page 698 fs/jfs/jfs_metapage.c unlock_page(mp->page); page 703 fs/jfs/jfs_metapage.c struct page *page = mp->page; page 707 fs/jfs/jfs_metapage.c page_cache_get(page); page 708 fs/jfs/jfs_metapage.c lock_page(page); page 709 fs/jfs/jfs_metapage.c set_page_dirty(page); page 710 fs/jfs/jfs_metapage.c write_one_page(page, 1); page 712 fs/jfs/jfs_metapage.c page_cache_release(page); page 717 fs/jfs/jfs_metapage.c lock_page(mp->page); page 724 fs/jfs/jfs_metapage.c unlock_page(mp->page); page 727 fs/jfs/jfs_metapage.c page_cache_get(mp->page); page 730 fs/jfs/jfs_metapage.c unlock_page(mp->page); page 736 fs/jfs/jfs_metapage.c struct page *page = mp->page; page 739 fs/jfs/jfs_metapage.c BUG_ON(!page); page 741 fs/jfs/jfs_metapage.c lock_page(page); page 746 fs/jfs/jfs_metapage.c unlock_page(page); page 747 fs/jfs/jfs_metapage.c page_cache_release(page); page 752 fs/jfs/jfs_metapage.c set_page_dirty(page); page 755 fs/jfs/jfs_metapage.c write_one_page(page, 1); page 756 fs/jfs/jfs_metapage.c lock_page(page); /* write_one_page unlocks the page */ page 762 fs/jfs/jfs_metapage.c drop_metapage(page, mp); page 764 fs/jfs/jfs_metapage.c unlock_page(page); page 765 fs/jfs/jfs_metapage.c page_cache_release(page); page 777 fs/jfs/jfs_metapage.c struct page *page; page 786 fs/jfs/jfs_metapage.c page = find_lock_page(mapping, lblock >> l2BlocksPerPage); page 787 fs/jfs/jfs_metapage.c if (!page) page 790 fs/jfs/jfs_metapage.c mp = page_to_mp(page, offset); page 803 fs/jfs/jfs_metapage.c unlock_page(page); page 804 fs/jfs/jfs_metapage.c page_cache_release(page); page 40 fs/jfs/jfs_metapage.h struct page *page; page 106 fs/jfs/jfs_metapage.h struct page *page = mp->page; page 107 fs/jfs/jfs_metapage.h lock_page(page); page 110 fs/jfs/jfs_metapage.h page_cache_get(page); page 111 fs/jfs/jfs_metapage.h wait_on_page_writeback(page); page 113 fs/jfs/jfs_metapage.h unlock_page(page); page 123 fs/jfs/jfs_metapage.h wait_on_page_writeback(mp->page); page 132 fs/jfs/jfs_metapage.h page_cache_release(mp->page); page 333 fs/libfs.c clear_highpage(page); page 334 fs/libfs.c flush_dcache_page(page); page 335 fs/libfs.c SetPageUptodate(page); page 336 fs/libfs.c unlock_page(page); page 343 fs/libfs.c if (!PageUptodate(page)) { page 345 fs/libfs.c zero_user_segments(page, page 356 fs/libfs.c struct page *page; page 363 fs/libfs.c page = __grab_cache_page(mapping, index); page 364 fs/libfs.c if (!page) page 367 fs/libfs.c *pagep = page; page 369 fs/libfs.c return simple_prepare_write(file, page, from, from+len); page 375 fs/libfs.c struct inode *inode = page->mapping->host; page 376 fs/libfs.c loff_t pos = ((loff_t)page->index << PAGE_CACHE_SHIFT) + to; page 378 fs/libfs.c if (!PageUptodate(page)) page 379 fs/libfs.c SetPageUptodate(page); page 386 fs/libfs.c set_page_dirty(page); page 398 fs/libfs.c void *kaddr = kmap_atomic(page, KM_USER0); page 400 fs/libfs.c flush_dcache_page(page); page 404 fs/libfs.c simple_commit_write(file, page, from, from+copied); page 406 fs/libfs.c unlock_page(page); page 407 fs/libfs.c page_cache_release(page); page 30 fs/minix/dir.c kunmap(page); page 31 fs/minix/dir.c page_cache_release(page); page 55 fs/minix/dir.c struct address_space *mapping = page->mapping; page 58 fs/minix/dir.c block_write_end(NULL, mapping, pos, len, len, page, NULL); page 65 fs/minix/dir.c err = write_one_page(page, 1); page 67 fs/minix/dir.c unlock_page(page); page 71 fs/minix/dir.c static struct page * dir_get_page(struct inode *dir, unsigned long n) page 74 fs/minix/dir.c struct page *page = read_mapping_page(mapping, n, NULL); page 75 fs/minix/dir.c if (!IS_ERR(page)) { page 76 fs/minix/dir.c kmap(page); page 77 fs/minix/dir.c if (!PageUptodate(page)) page 80 fs/minix/dir.c return page; page 83 fs/minix/dir.c dir_put_page(page); page 113 fs/minix/dir.c struct page *page = dir_get_page(inode, n); page 115 fs/minix/dir.c if (IS_ERR(page)) page 117 fs/minix/dir.c kaddr = (char *)page_address(page); page 139 fs/minix/dir.c dir_put_page(page); page 144 fs/minix/dir.c dir_put_page(page); page 178 fs/minix/dir.c struct page *page = NULL; page 188 fs/minix/dir.c page = dir_get_page(dir, n); page 189 fs/minix/dir.c if (IS_ERR(page)) page 192 fs/minix/dir.c kaddr = (char*)page_address(page); page 209 fs/minix/dir.c dir_put_page(page); page 214 fs/minix/dir.c *res_page = page; page 225 fs/minix/dir.c struct page *page = NULL; page 244 fs/minix/dir.c page = dir_get_page(dir, n); page 245 fs/minix/dir.c err = PTR_ERR(page); page 246 fs/minix/dir.c if (IS_ERR(page)) page 248 fs/minix/dir.c lock_page(page); page 249 fs/minix/dir.c kaddr = (char*)page_address(page); page 276 fs/minix/dir.c unlock_page(page); page 277 fs/minix/dir.c dir_put_page(page); page 283 fs/minix/dir.c pos = (page->index >> PAGE_CACHE_SHIFT) + p - (char*)page_address(page); page 284 fs/minix/dir.c err = __minix_write_begin(NULL, page->mapping, pos, sbi->s_dirsize, page 285 fs/minix/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 296 fs/minix/dir.c err = dir_commit_chunk(page, pos, sbi->s_dirsize); page 300 fs/minix/dir.c dir_put_page(page); page 304 fs/minix/dir.c unlock_page(page); page 310 fs/minix/dir.c struct address_space *mapping = page->mapping; page 312 fs/minix/dir.c char *kaddr = page_address(page); page 313 fs/minix/dir.c loff_t pos = page_offset(page) + (char*)de - kaddr; page 317 fs/minix/dir.c lock_page(page); page 319 fs/minix/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 322 fs/minix/dir.c err = dir_commit_chunk(page, pos, len); page 324 fs/minix/dir.c unlock_page(page); page 326 fs/minix/dir.c dir_put_page(page); page 335 fs/minix/dir.c struct page *page = grab_cache_page(mapping, 0); page 340 fs/minix/dir.c if (!page) page 343 fs/minix/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 345 fs/minix/dir.c unlock_page(page); page 349 fs/minix/dir.c kaddr = kmap_atomic(page, KM_USER0); page 371 fs/minix/dir.c err = dir_commit_chunk(page, 0, 2 * sbi->s_dirsize); page 373 fs/minix/dir.c page_cache_release(page); page 382 fs/minix/dir.c struct page *page = NULL; page 391 fs/minix/dir.c page = dir_get_page(inode, i); page 392 fs/minix/dir.c if (IS_ERR(page)) page 395 fs/minix/dir.c kaddr = (char *)page_address(page); page 421 fs/minix/dir.c dir_put_page(page); page 426 fs/minix/dir.c dir_put_page(page); page 434 fs/minix/dir.c struct address_space *mapping = page->mapping; page 437 fs/minix/dir.c loff_t pos = page_offset(page) + page 438 fs/minix/dir.c (char *)de-(char*)page_address(page); page 441 fs/minix/dir.c lock_page(page); page 444 fs/minix/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 447 fs/minix/dir.c err = dir_commit_chunk(page, pos, sbi->s_dirsize); page 449 fs/minix/dir.c unlock_page(page); page 451 fs/minix/dir.c dir_put_page(page); page 458 fs/minix/dir.c struct page *page = dir_get_page(dir, 0); page 462 fs/minix/dir.c if (!IS_ERR(page)) { page 463 fs/minix/dir.c de = minix_next_entry(page_address(page), sbi); page 464 fs/minix/dir.c *p = page; page 471 fs/minix/dir.c struct page *page; page 472 fs/minix/dir.c struct minix_dir_entry *de = minix_find_entry(dentry, &page); page 477 fs/minix/dir.c dir_put_page(page); page 347 fs/minix/inode.c return block_write_full_page(page, minix_get_block, wbc); page 352 fs/minix/inode.c return block_read_full_page(page,minix_get_block); page 156 fs/minix/namei.c struct page * page; page 159 fs/minix/namei.c de = minix_find_entry(dentry, &page); page 163 fs/minix/namei.c err = minix_delete_entry(de, page); page 194 fs/minix/namei.c struct page * dir_page = NULL; page 196 fs/minix/namei.c struct page * old_page; page 212 fs/minix/namei.c struct page * new_page; page 48 fs/mpage.c struct page *page = bvec->bv_page; page 54 fs/mpage.c SetPageUptodate(page); page 56 fs/mpage.c ClearPageUptodate(page); page 57 fs/mpage.c SetPageError(page); page 59 fs/mpage.c unlock_page(page); page 70 fs/mpage.c struct page *page = bvec->bv_page; page 76 fs/mpage.c SetPageError(page); page 77 fs/mpage.c if (page->mapping) page 78 fs/mpage.c set_bit(AS_EIO, &page->mapping->flags); page 80 fs/mpage.c end_page_writeback(page); page 129 fs/mpage.c struct inode *inode = page->mapping->host; page 133 fs/mpage.c if (!page_has_buffers(page)) { page 140 fs/mpage.c SetPageUptodate(page); page 143 fs/mpage.c create_empty_buffers(page, 1 << inode->i_blkbits, 0); page 145 fs/mpage.c head = page_buffers(page); page 173 fs/mpage.c struct inode *inode = page->mapping->host; page 189 fs/mpage.c if (page_has_buffers(page)) page 192 fs/mpage.c block_in_file = (sector_t)page->index << (PAGE_CACHE_SHIFT - blkbits); page 226 fs/mpage.c map_bh->b_page = page; page 255 fs/mpage.c map_buffer_to_page(page, map_bh, page_block); page 280 fs/mpage.c zero_user_segment(page, first_hole << blkbits, PAGE_CACHE_SIZE); page 282 fs/mpage.c SetPageUptodate(page); page 283 fs/mpage.c unlock_page(page); page 287 fs/mpage.c SetPageMappedToDisk(page); page 306 fs/mpage.c if (bio_add_page(bio, page, length, 0) < length) { page 321 fs/mpage.c if (!PageUptodate(page)) page 322 fs/mpage.c block_read_full_page(page, get_block); page 324 fs/mpage.c unlock_page(page); page 383 fs/mpage.c struct page *page = list_entry(pages->prev, struct page, lru); page 385 fs/mpage.c prefetchw(&page->flags); page 386 fs/mpage.c list_del(&page->lru); page 387 fs/mpage.c if (!add_to_page_cache_lru(page, mapping, page 388 fs/mpage.c page->index, GFP_KERNEL)) { page 389 fs/mpage.c bio = do_mpage_readpage(bio, page, page 395 fs/mpage.c page_cache_release(page); page 415 fs/mpage.c bio = do_mpage_readpage(bio, page, 1, &last_block_in_bio, page 445 fs/mpage.c struct address_space *mapping = page->mapping; page 446 fs/mpage.c struct inode *inode = page->mapping->host; page 464 fs/mpage.c if (page_has_buffers(page)) { page 465 fs/mpage.c struct buffer_head *head = page_buffers(page); page 517 fs/mpage.c BUG_ON(!PageUptodate(page)); page 518 fs/mpage.c block_in_file = (sector_t)page->index << (PAGE_CACHE_SHIFT - blkbits); page 520 fs/mpage.c map_bh.b_page = page; page 551 fs/mpage.c if (page->index >= end_index) { page 562 fs/mpage.c if (page->index > end_index || !offset) page 564 fs/mpage.c zero_user_segment(page, offset, PAGE_CACHE_SIZE); page 587 fs/mpage.c if (bio_add_page(bio, page, length, 0) < length) { page 596 fs/mpage.c if (page_has_buffers(page)) { page 597 fs/mpage.c struct buffer_head *head = page_buffers(page); page 613 fs/mpage.c if (buffer_heads_over_limit && PageUptodate(page)) page 614 fs/mpage.c try_to_free_buffers(page); page 617 fs/mpage.c BUG_ON(PageWriteback(page)); page 618 fs/mpage.c set_page_writeback(page); page 619 fs/mpage.c unlock_page(page); page 636 fs/mpage.c ret = mapping->a_ops->writepage(page, wbc); page 703 fs/mpage.c int ret = __mpage_writepage(page, wbc, &mpd); page 130 fs/namei.c retval = strncpy_from_user(page, filename, len); page 2762 fs/namei.c struct page * page; page 2764 fs/namei.c page = read_mapping_page(mapping, 0, NULL); page 2765 fs/namei.c if (IS_ERR(page)) page 2766 fs/namei.c return (char*)page; page 2767 fs/namei.c *ppage = page; page 2768 fs/namei.c return kmap(page); page 2773 fs/namei.c struct page *page = NULL; page 2774 fs/namei.c char *s = page_getlink(dentry, &page); page 2776 fs/namei.c if (page) { page 2777 fs/namei.c kunmap(page); page 2778 fs/namei.c page_cache_release(page); page 2785 fs/namei.c struct page *page = NULL; page 2786 fs/namei.c nd_set_link(nd, page_getlink(dentry, &page)); page 2787 fs/namei.c return page; page 2792 fs/namei.c struct page *page = cookie; page 2794 fs/namei.c if (page) { page 2795 fs/namei.c kunmap(page); page 2796 fs/namei.c page_cache_release(page); page 2804 fs/namei.c struct page *page; page 2811 fs/namei.c AOP_FLAG_UNINTERRUPTIBLE, &page, &fsdata); page 2815 fs/namei.c kaddr = kmap_atomic(page, KM_USER0); page 2820 fs/namei.c page, fsdata); page 1858 fs/namespace.c unsigned long page; page 1865 fs/namespace.c if (!(page = __get_free_page(GFP_KERNEL))) page 1877 fs/namespace.c i = size - exact_copy_from_user((void *)page, data, size); page 1879 fs/namespace.c free_page(page); page 1883 fs/namespace.c memset((char *)page + i, 0, PAGE_SIZE - i); page 1884 fs/namespace.c *where = page; page 407 fs/ncpfs/dir.c struct page *page = NULL; page 416 fs/ncpfs/dir.c ctl.page = NULL; page 439 fs/ncpfs/dir.c page = grab_cache_page(&inode->i_data, 0); page 440 fs/ncpfs/dir.c if (!page) page 443 fs/ncpfs/dir.c ctl.cache = cache = kmap(page); page 446 fs/ncpfs/dir.c if (!PageUptodate(page) || !ctl.head.eof) page 468 fs/ncpfs/dir.c ctl.page = find_lock_page(&inode->i_data, ctl.ofs); page 469 fs/ncpfs/dir.c if (!ctl.page) page 471 fs/ncpfs/dir.c ctl.cache = kmap(ctl.page); page 472 fs/ncpfs/dir.c if (!PageUptodate(ctl.page)) page 494 fs/ncpfs/dir.c if (ctl.page) { page 495 fs/ncpfs/dir.c kunmap(ctl.page); page 496 fs/ncpfs/dir.c SetPageUptodate(ctl.page); page 497 fs/ncpfs/dir.c unlock_page(ctl.page); page 498 fs/ncpfs/dir.c page_cache_release(ctl.page); page 499 fs/ncpfs/dir.c ctl.page = NULL; page 505 fs/ncpfs/dir.c if (ctl.page) { page 506 fs/ncpfs/dir.c kunmap(ctl.page); page 507 fs/ncpfs/dir.c unlock_page(ctl.page); page 508 fs/ncpfs/dir.c page_cache_release(ctl.page); page 509 fs/ncpfs/dir.c ctl.page = NULL; page 535 fs/ncpfs/dir.c if (page) { page 537 fs/ncpfs/dir.c kunmap(page); page 538 fs/ncpfs/dir.c SetPageUptodate(page); page 539 fs/ncpfs/dir.c unlock_page(page); page 540 fs/ncpfs/dir.c page_cache_release(page); page 542 fs/ncpfs/dir.c if (ctl.page) { page 543 fs/ncpfs/dir.c kunmap(ctl.page); page 544 fs/ncpfs/dir.c SetPageUptodate(ctl.page); page 545 fs/ncpfs/dir.c unlock_page(ctl.page); page 546 fs/ncpfs/dir.c page_cache_release(ctl.page); page 611 fs/ncpfs/dir.c if (ctl.page) { page 612 fs/ncpfs/dir.c kunmap(ctl.page); page 613 fs/ncpfs/dir.c SetPageUptodate(ctl.page); page 614 fs/ncpfs/dir.c unlock_page(ctl.page); page 615 fs/ncpfs/dir.c page_cache_release(ctl.page); page 620 fs/ncpfs/dir.c ctl.page = grab_cache_page(&inode->i_data, ctl.ofs); page 621 fs/ncpfs/dir.c if (ctl.page) page 622 fs/ncpfs/dir.c ctl.cache = kmap(ctl.page); page 46 fs/ncpfs/mmap.c vmf->page = alloc_page(GFP_HIGHUSER); page 47 fs/ncpfs/mmap.c if (!vmf->page) page 49 fs/ncpfs/mmap.c pg_addr = kmap(vmf->page); page 86 fs/ncpfs/mmap.c flush_dcache_page(vmf->page); page 87 fs/ncpfs/mmap.c kunmap(vmf->page); page 249 fs/ncpfs/ncplib_kernel.h struct page *page; page 45 fs/ncpfs/symlink.c struct inode *inode = page->mapping->host; page 48 fs/ncpfs/symlink.c char *buf = kmap(page); page 83 fs/ncpfs/symlink.c SetPageUptodate(page); page 84 fs/ncpfs/symlink.c kunmap(page); page 85 fs/ncpfs/symlink.c unlock_page(page); page 92 fs/ncpfs/symlink.c SetPageError(page); page 93 fs/ncpfs/symlink.c kunmap(page); page 94 fs/ncpfs/symlink.c unlock_page(page); page 150 fs/nfs/dir.c struct page *page; page 185 fs/nfs/dir.c page->index); page 189 fs/nfs/dir.c error = NFS_PROTO(inode)->readdir(file->f_path.dentry, cred, desc->entry->cookie, page, page 203 fs/nfs/dir.c SetPageUptodate(page); page 208 fs/nfs/dir.c if (invalidate_inode_pages2_range(inode->i_mapping, page->index + 1, -1) < 0) { page 212 fs/nfs/dir.c unlock_page(page); page 215 fs/nfs/dir.c unlock_page(page); page 237 fs/nfs/dir.c kunmap(desc->page); page 238 fs/nfs/dir.c page_cache_release(desc->page); page 239 fs/nfs/dir.c desc->page = NULL; page 315 fs/nfs/dir.c struct page *page; page 326 fs/nfs/dir.c page = read_cache_page(inode->i_mapping, desc->page_index, page 328 fs/nfs/dir.c if (IS_ERR(page)) { page 329 fs/nfs/dir.c status = PTR_ERR(page); page 334 fs/nfs/dir.c desc->page = page; page 335 fs/nfs/dir.c desc->ptr = kmap(page); /* matching kunmap in nfs_do_filldir */ page 472 fs/nfs/dir.c struct page *page = NULL; page 479 fs/nfs/dir.c page = alloc_page(GFP_HIGHUSER); page 480 fs/nfs/dir.c if (!page) { page 486 fs/nfs/dir.c *desc->dir_cookie, page, page 489 fs/nfs/dir.c desc->page = page; page 490 fs/nfs/dir.c desc->ptr = kmap(page); /* matching kunmap in nfs_do_filldir */ page 1467 fs/nfs/dir.c struct page *page; page 1482 fs/nfs/dir.c page = alloc_page(GFP_HIGHUSER); page 1483 fs/nfs/dir.c if (!page) page 1486 fs/nfs/dir.c kaddr = kmap_atomic(page, KM_USER0); page 1492 fs/nfs/dir.c error = NFS_PROTO(dir)->symlink(dir, dentry, page, pathlen, &attr); page 1498 fs/nfs/dir.c __free_page(page); page 1507 fs/nfs/dir.c if (!add_to_page_cache(page, dentry->d_inode->i_mapping, 0, page 1509 fs/nfs/dir.c pagevec_add(&lru_pvec, page); page 1511 fs/nfs/dir.c SetPageUptodate(page); page 1512 fs/nfs/dir.c unlock_page(page); page 1514 fs/nfs/dir.c __free_page(page); page 135 fs/nfs/direct.c struct page *page = pages[i]; page 136 fs/nfs/direct.c if (!PageCompound(page)) page 137 fs/nfs/direct.c set_page_dirty(page); page 346 fs/nfs/file.c struct page *page; page 354 fs/nfs/file.c page = __grab_cache_page(mapping, index); page 355 fs/nfs/file.c if (!page) page 357 fs/nfs/file.c *pagep = page; page 359 fs/nfs/file.c ret = nfs_flush_incompatible(file, page); page 361 fs/nfs/file.c unlock_page(page); page 362 fs/nfs/file.c page_cache_release(page); page 383 fs/nfs/file.c if (!PageUptodate(page)) { page 384 fs/nfs/file.c unsigned pglen = nfs_page_length(page); page 388 fs/nfs/file.c zero_user_segments(page, 0, offset, page 390 fs/nfs/file.c SetPageUptodate(page); page 392 fs/nfs/file.c zero_user_segment(page, end, PAGE_CACHE_SIZE); page 394 fs/nfs/file.c SetPageUptodate(page); page 396 fs/nfs/file.c zero_user_segment(page, pglen, PAGE_CACHE_SIZE); page 399 fs/nfs/file.c status = nfs_updatepage(file, page, offset, copied); page 401 fs/nfs/file.c unlock_page(page); page 402 fs/nfs/file.c page_cache_release(page); page 411 fs/nfs/file.c dfprintk(PAGECACHE, "NFS: invalidate_page(%p, %lu)\n", page, offset); page 416 fs/nfs/file.c nfs_wb_page_cancel(page->mapping->host, page); page 421 fs/nfs/file.c dfprintk(PAGECACHE, "NFS: release_page(%p)\n", page); page 429 fs/nfs/file.c struct inode *inode = page->mapping->host; page 432 fs/nfs/file.c inode->i_ino, (long long)page_offset(page)); page 434 fs/nfs/file.c return nfs_wb_page(inode, page); page 462 fs/nfs/file.c (long long)page_offset(page)); page 464 fs/nfs/file.c lock_page(page); page 465 fs/nfs/file.c mapping = page->mapping; page 470 fs/nfs/file.c pagelen = nfs_page_length(page); page 474 fs/nfs/file.c ret = nfs_flush_incompatible(filp, page); page 478 fs/nfs/file.c ret = nfs_updatepage(filp, page, 0, pagelen); page 482 fs/nfs/file.c unlock_page(page); page 256 fs/nfs/internal.h loff_t i_size = i_size_read(page->mapping->host); page 260 fs/nfs/internal.h if (page->index < end_index) page 262 fs/nfs/internal.h if (page->index == end_index) page 227 fs/nfs/namespace.c char *page = (char *) __get_free_page(GFP_USER); page 235 fs/nfs/namespace.c if (page == NULL) page 237 fs/nfs/namespace.c devname = nfs_devname(mnt_parent, dentry, page, PAGE_SIZE); page 243 fs/nfs/namespace.c free_page((unsigned long)page); page 427 fs/nfs/nfs2xdr.c struct page **page; page 451 fs/nfs/nfs2xdr.c page = rcvbuf->pages; page 452 fs/nfs/nfs2xdr.c kaddr = p = kmap_atomic(*page, KM_USER0); page 188 fs/nfs/nfs3acl.c struct page *pages[NFSACL_MAXPAGES] = { }; page 294 fs/nfs/nfs3acl.c struct page *pages[NFSACL_MAXPAGES] = { }; page 236 fs/nfs/nfs3proc.c .pages = &page page 510 fs/nfs/nfs3proc.c data->arg.symlink.pages = &page; page 600 fs/nfs/nfs3proc.c .pages = &page page 508 fs/nfs/nfs3xdr.c struct page **page; page 541 fs/nfs/nfs3xdr.c page = rcvbuf->pages; page 542 fs/nfs/nfs3xdr.c kaddr = p = kmap_atomic(*page, KM_USER0); page 79 fs/nfs/nfs4namespace.c path = nfs4_path(mnt_parent, dentry, page, PAGE_SIZE); page 130 fs/nfs/nfs4namespace.c char *page = NULL, *page2 = NULL; page 140 fs/nfs/nfs4namespace.c page = (char *) __get_free_page(GFP_USER); page 141 fs/nfs/nfs4namespace.c if (!page) page 149 fs/nfs/nfs4namespace.c error = nfs4_validate_fspath(mnt_parent, dentry, locations, page, page2); page 191 fs/nfs/nfs4namespace.c snprintf(page, PAGE_SIZE, "%s:%s", page 195 fs/nfs/nfs4namespace.c mnt = vfs_kern_mount(&nfs4_referral_fs_type, 0, page, &mountdata); page 205 fs/nfs/nfs4namespace.c free_page((unsigned long) page); page 222 fs/nfs/nfs4namespace.c struct page *page; page 228 fs/nfs/nfs4namespace.c page = alloc_page(GFP_KERNEL); page 229 fs/nfs/nfs4namespace.c if (page == NULL) page 243 fs/nfs/nfs4namespace.c err = nfs4_proc_fs_locations(parent->d_inode, &dentry->d_name, fs_locations, page); page 252 fs/nfs/nfs4namespace.c __free_page(page); page 1564 fs/nfs/nfs4proc.c struct page *page = NULL; page 1567 fs/nfs/nfs4proc.c page = alloc_page(GFP_KERNEL); page 1568 fs/nfs/nfs4proc.c if (page == NULL) page 1574 fs/nfs/nfs4proc.c status = nfs4_proc_fs_locations(dir, name, locations, page); page 1590 fs/nfs/nfs4proc.c if (page) page 1591 fs/nfs/nfs4proc.c __free_page(page); page 1839 fs/nfs/nfs4proc.c .pages = &page, page 1857 fs/nfs/nfs4proc.c _nfs4_proc_readlink(inode, page, pgbase, pglen), page 2147 fs/nfs/nfs4proc.c data->arg.u.symlink.pages = &page; page 2164 fs/nfs/nfs4proc.c _nfs4_proc_symlink(dir, dentry, page, page 2207 fs/nfs/nfs4proc.c .pages = &page, page 2245 fs/nfs/nfs4proc.c page, count, plus), page 2622 fs/nfs/nfs4proc.c struct page *pages[NFS4ACL_MAXPAGES]; page 2635 fs/nfs/nfs4proc.c struct page *localpage = NULL; page 2707 fs/nfs/nfs4proc.c struct page *pages[NFS4ACL_MAXPAGES]; page 3645 fs/nfs/nfs4proc.c .page = page, page 2160 fs/nfs/nfs4xdr.c xdr_inline_pages(&req->rq_rcv_buf, replen, &args->page, page 3495 fs/nfs/nfs4xdr.c struct page *page = *rcvbuf->pages; page 3521 fs/nfs/nfs4xdr.c kaddr = p = kmap_atomic(page, KM_USER0); page 77 fs/nfs/pagelist.c req->wb_page = page; page 79 fs/nfs/pagelist.c req->wb_index = page->index; page 80 fs/nfs/pagelist.c page_cache_get(page); page 81 fs/nfs/pagelist.c BUG_ON(PagePrivate(page)); page 82 fs/nfs/pagelist.c BUG_ON(!PageLocked(page)); page 83 fs/nfs/pagelist.c BUG_ON(page->mapping->host != inode); page 150 fs/nfs/pagelist.c struct page *page = req->wb_page; page 151 fs/nfs/pagelist.c if (page != NULL) { page 152 fs/nfs/pagelist.c page_cache_release(page); page 176 fs/nfs/proc.c .pages = &page page 368 fs/nfs/proc.c .pages = &page, page 470 fs/nfs/proc.c .pages = &page, page 51 fs/nfs/read.c p->pagevec = kcalloc(pagecount, sizeof(struct page *), GFP_NOFS); page 79 fs/nfs/read.c zero_user(page, 0, PAGE_CACHE_SIZE); page 80 fs/nfs/read.c SetPageUptodate(page); page 81 fs/nfs/read.c unlock_page(page); page 90 fs/nfs/read.c struct page **pages; page 121 fs/nfs/read.c len = nfs_page_length(page); page 123 fs/nfs/read.c return nfs_return_empty_page(page); page 124 fs/nfs/read.c new = nfs_create_request(ctx, inode, page, 0, len); page 126 fs/nfs/read.c unlock_page(page); page 130 fs/nfs/read.c zero_user_segment(page, len, PAGE_CACHE_SIZE); page 240 fs/nfs/read.c struct page *page = req->wb_page; page 263 fs/nfs/read.c ClearPageError(page); page 272 fs/nfs/read.c data->pagevec[0] = page; page 292 fs/nfs/read.c SetPageError(page); page 300 fs/nfs/read.c struct page **pages; page 389 fs/nfs/read.c struct page *page = req->wb_page; page 393 fs/nfs/read.c SetPageError(page); page 396 fs/nfs/read.c if (!PageError(page)) page 397 fs/nfs/read.c SetPageUptodate(page); page 412 fs/nfs/read.c struct page **pages; page 480 fs/nfs/read.c struct inode *inode = page->mapping->host; page 484 fs/nfs/read.c page, PAGE_CACHE_SIZE, page->index); page 495 fs/nfs/read.c error = nfs_wb_page(inode, page); page 498 fs/nfs/read.c if (PageUptodate(page)) page 513 fs/nfs/read.c error = nfs_readpage_async(ctx, inode, page); page 518 fs/nfs/read.c unlock_page(page); page 531 fs/nfs/read.c struct inode *inode = page->mapping->host; page 536 fs/nfs/read.c error = nfs_wb_page(inode, page); page 539 fs/nfs/read.c if (PageUptodate(page)) page 542 fs/nfs/read.c len = nfs_page_length(page); page 544 fs/nfs/read.c return nfs_return_empty_page(page); page 546 fs/nfs/read.c new = nfs_create_request(desc->ctx, inode, page, 0, len); page 551 fs/nfs/read.c zero_user_segment(page, len, PAGE_CACHE_SIZE); page 559 fs/nfs/read.c SetPageError(page); page 561 fs/nfs/read.c unlock_page(page); page 34 fs/nfs/symlink.c error = NFS_PROTO(inode)->readlink(inode, page, 0, PAGE_SIZE); page 37 fs/nfs/symlink.c SetPageUptodate(page); page 38 fs/nfs/symlink.c unlock_page(page); page 42 fs/nfs/symlink.c SetPageError(page); page 43 fs/nfs/symlink.c unlock_page(page); page 50 fs/nfs/symlink.c struct page *page; page 56 fs/nfs/symlink.c page = read_cache_page(&inode->i_data, 0, page 58 fs/nfs/symlink.c if (IS_ERR(page)) { page 59 fs/nfs/symlink.c err = page; page 62 fs/nfs/symlink.c nd_set_link(nd, kmap(page)); page 63 fs/nfs/symlink.c return page; page 77 fs/nfs/write.c p->pagevec = kcalloc(pagecount, sizeof(struct page *), GFP_NOFS); page 113 fs/nfs/write.c if (PagePrivate(page)) { page 114 fs/nfs/write.c req = (struct nfs_page *)page_private(page); page 123 fs/nfs/write.c struct inode *inode = page->mapping->host; page 127 fs/nfs/write.c req = nfs_page_find_request_locked(page); page 135 fs/nfs/write.c struct inode *inode = page->mapping->host; page 142 fs/nfs/write.c if (i_size > 0 && page->index < end_index) page 144 fs/nfs/write.c end = ((loff_t)page->index << PAGE_CACHE_SHIFT) + ((loff_t)offset+count); page 156 fs/nfs/write.c SetPageError(page); page 157 fs/nfs/write.c nfs_zap_mapping(page->mapping->host, page->mapping); page 165 fs/nfs/write.c if (PageUptodate(page)) page 169 fs/nfs/write.c if (count != nfs_page_length(page)) page 171 fs/nfs/write.c SetPageUptodate(page); page 195 fs/nfs/write.c int ret = test_set_page_writeback(page); page 198 fs/nfs/write.c struct inode *inode = page->mapping->host; page 210 fs/nfs/write.c struct inode *inode = page->mapping->host; page 213 fs/nfs/write.c end_page_writeback(page); page 225 fs/nfs/write.c struct inode *inode = page->mapping->host; page 231 fs/nfs/write.c req = nfs_page_find_request_locked(page); page 254 fs/nfs/write.c if (nfs_set_page_writeback(page) != 0) { page 268 fs/nfs/write.c struct inode *inode = page->mapping->host; page 273 fs/nfs/write.c nfs_pageio_cond_complete(pgio, page->index); page 274 fs/nfs/write.c return nfs_page_async_flush(pgio, page); page 285 fs/nfs/write.c nfs_pageio_init_write(&pgio, page->mapping->host, wb_priority(wbc)); page 286 fs/nfs/write.c err = nfs_do_writepage(page, wbc, &pgio); page 299 fs/nfs/write.c ret = nfs_writepage_locked(page, wbc); page 300 fs/nfs/write.c unlock_page(page); page 308 fs/nfs/write.c ret = nfs_do_writepage(page, wbc, data); page 309 fs/nfs/write.c unlock_page(page); page 421 fs/nfs/write.c struct page *page = req->wb_page; page 424 fs/nfs/write.c dec_zone_page_state(page, NR_UNSTABLE_NFS); page 425 fs/nfs/write.c dec_bdi_stat(page->mapping->backing_dev_info, BDI_RECLAIMABLE); page 574 fs/nfs/write.c if (!PagePrivate(page)) page 581 fs/nfs/write.c req = nfs_page_find_request_locked(page); page 627 fs/nfs/write.c error = nfs_wb_page(inode, page); page 642 fs/nfs/write.c struct inode *inode = page->mapping->host; page 646 fs/nfs/write.c req = nfs_try_to_update_request(inode, page, offset, bytes); page 649 fs/nfs/write.c req = nfs_create_request(ctx, inode, page, offset, bytes); page 666 fs/nfs/write.c req = nfs_setup_write_request(ctx, page, offset, count); page 670 fs/nfs/write.c nfs_grow_file(page, offset, count); page 671 fs/nfs/write.c nfs_mark_uptodate(page, req->wb_pgbase, req->wb_bytes); page 690 fs/nfs/write.c req = nfs_page_find_request(page); page 693 fs/nfs/write.c do_flush = req->wb_page != page || req->wb_context != ctx; page 697 fs/nfs/write.c status = nfs_wb_page(page->mapping->host, page); page 709 fs/nfs/write.c return PageUptodate(page) && page 723 fs/nfs/write.c struct inode *inode = page->mapping->host; page 731 fs/nfs/write.c (long long)(page_offset(page) + offset)); page 738 fs/nfs/write.c if (nfs_write_pageuptodate(page, inode) && page 741 fs/nfs/write.c count = max(count + offset, nfs_page_length(page)); page 745 fs/nfs/write.c status = nfs_writepage_setup(ctx, page, offset, count); page 747 fs/nfs/write.c nfs_set_pageerror(page); page 749 fs/nfs/write.c __set_page_dirty_nobuffers(page); page 868 fs/nfs/write.c struct page *page = req->wb_page; page 891 fs/nfs/write.c ClearPageError(page); page 900 fs/nfs/write.c data->pagevec[0] = page; page 935 fs/nfs/write.c struct page **pages; page 995 fs/nfs/write.c struct page *page = req->wb_page; page 999 fs/nfs/write.c nfs_set_pageerror(page); page 1006 fs/nfs/write.c struct inode *inode = page->mapping->host; page 1056 fs/nfs/write.c struct page *page = req->wb_page; page 1068 fs/nfs/write.c nfs_set_pageerror(page); page 1077 fs/nfs/write.c nfs_end_page_writeback(page); page 1083 fs/nfs/write.c nfs_end_page_writeback(page); page 1458 fs/nfs/write.c loff_t range_start = page_offset(page); page 1461 fs/nfs/write.c .bdi = page->mapping->backing_dev_info, page 1469 fs/nfs/write.c BUG_ON(!PageLocked(page)); page 1471 fs/nfs/write.c req = nfs_page_find_request(page); page 1484 fs/nfs/write.c cancel_dirty_page(page, PAGE_CACHE_SIZE); page 1492 fs/nfs/write.c if (!PagePrivate(page)) page 1494 fs/nfs/write.c ret = nfs_sync_mapping_wait(page->mapping, &wbc, FLUSH_INVALIDATE); page 1502 fs/nfs/write.c loff_t range_start = page_offset(page); page 1505 fs/nfs/write.c .bdi = page->mapping->backing_dev_info, page 1514 fs/nfs/write.c if (clear_page_dirty_for_io(page)) { page 1515 fs/nfs/write.c ret = nfs_writepage_locked(page, &wbc); page 1518 fs/nfs/write.c } else if (!PagePrivate(page)) page 1520 fs/nfs/write.c ret = nfs_sync_mapping_wait(page->mapping, &wbc, how); page 1523 fs/nfs/write.c } while (PagePrivate(page)); page 1535 fs/nfs/write.c return nfs_wb_page_priority(inode, page, FLUSH_STABLE); page 2300 fs/nfsd/nfs4xdr.c char *page; page 2308 fs/nfsd/nfs4xdr.c page = page_address(resp->rqstp->rq_respages[resp->rqstp->rq_resused++]); page 2319 fs/nfsd/nfs4xdr.c nfserr = nfsd_readlink(readlink->rl_rqstp, readlink->rl_fhp, page, &maxcount); page 2349 fs/nfsd/nfs4xdr.c __be32 *page, *savep, *tailbase; page 2382 fs/nfsd/nfs4xdr.c page = page_address(resp->rqstp->rq_respages[resp->rqstp->rq_resused++]); page 2385 fs/nfsd/nfs4xdr.c readdir->buffer = page; page 2394 fs/nfsd/nfs4xdr.c readdir->buffer == page) page 840 fs/nfsd/vfs.c struct page **pp = rqstp->rq_respages + rqstp->rq_resused; page 841 fs/nfsd/vfs.c struct page *page = buf->page; page 852 fs/nfsd/vfs.c get_page(page); page 854 fs/nfsd/vfs.c *pp = page; page 858 fs/nfsd/vfs.c } else if (page != pp[-1]) { page 859 fs/nfsd/vfs.c get_page(page); page 862 fs/nfsd/vfs.c *pp = page; page 62 fs/ntfs/aops.c struct page *page; page 67 fs/ntfs/aops.c page = bh->b_page; page 68 fs/ntfs/aops.c vi = page->mapping->host; page 77 fs/ntfs/aops.c file_ofs = ((s64)page->index << PAGE_CACHE_SHIFT) + page 96 fs/ntfs/aops.c kaddr = kmap_atomic(page, KM_BIO_SRC_IRQ); page 99 fs/ntfs/aops.c flush_dcache_page(page); page 105 fs/ntfs/aops.c SetPageError(page); page 109 fs/ntfs/aops.c first = page_buffers(page); page 137 fs/ntfs/aops.c if (likely(page_uptodate && !PageError(page))) page 138 fs/ntfs/aops.c SetPageUptodate(page); page 149 fs/ntfs/aops.c kaddr = kmap_atomic(page, KM_BIO_SRC_IRQ); page 155 fs/ntfs/aops.c flush_dcache_page(page); page 156 fs/ntfs/aops.c if (likely(page_uptodate && !PageError(page))) page 157 fs/ntfs/aops.c SetPageUptodate(page); page 159 fs/ntfs/aops.c unlock_page(page); page 201 fs/ntfs/aops.c vi = page->mapping->host; page 211 fs/ntfs/aops.c if (!page_has_buffers(page)) { page 212 fs/ntfs/aops.c create_empty_buffers(page, blocksize, 0); page 213 fs/ntfs/aops.c if (unlikely(!page_has_buffers(page))) { page 214 fs/ntfs/aops.c unlock_page(page); page 218 fs/ntfs/aops.c bh = head = page_buffers(page); page 232 fs/ntfs/aops.c iblock = (s64)page->index << (PAGE_CACHE_SHIFT - blocksize_bits); page 322 fs/ntfs/aops.c SetPageError(page); page 341 fs/ntfs/aops.c zero_user(page, i * blocksize, blocksize); page 372 fs/ntfs/aops.c if (likely(!PageError(page))) page 373 fs/ntfs/aops.c SetPageUptodate(page); page 376 fs/ntfs/aops.c unlock_page(page); page 411 fs/ntfs/aops.c BUG_ON(!PageLocked(page)); page 412 fs/ntfs/aops.c vi = page->mapping->host; page 415 fs/ntfs/aops.c if (unlikely(page->index >= (i_size + PAGE_CACHE_SIZE - 1) >> page 417 fs/ntfs/aops.c zero_user(page, 0, PAGE_CACHE_SIZE); page 425 fs/ntfs/aops.c if (PageUptodate(page)) { page 426 fs/ntfs/aops.c unlock_page(page); page 449 fs/ntfs/aops.c return ntfs_read_compressed_block(page); page 455 fs/ntfs/aops.c return ntfs_read_block(page); page 465 fs/ntfs/aops.c if (unlikely(page->index > 0)) { page 466 fs/ntfs/aops.c zero_user(page, 0, PAGE_CACHE_SIZE); page 506 fs/ntfs/aops.c addr = kmap_atomic(page, KM_USER0); page 513 fs/ntfs/aops.c flush_dcache_page(page); page 520 fs/ntfs/aops.c SetPageUptodate(page); page 522 fs/ntfs/aops.c unlock_page(page); page 568 fs/ntfs/aops.c vi = page->mapping->host; page 573 fs/ntfs/aops.c "0x%lx.", ni->mft_no, ni->type, page->index); page 579 fs/ntfs/aops.c if (!page_has_buffers(page)) { page 580 fs/ntfs/aops.c BUG_ON(!PageUptodate(page)); page 581 fs/ntfs/aops.c create_empty_buffers(page, blocksize, page 583 fs/ntfs/aops.c if (unlikely(!page_has_buffers(page))) { page 591 fs/ntfs/aops.c redirty_page_for_writepage(wbc, page); page 592 fs/ntfs/aops.c unlock_page(page); page 596 fs/ntfs/aops.c bh = head = page_buffers(page); page 602 fs/ntfs/aops.c block = (s64)page->index << (PAGE_CACHE_SHIFT - blocksize_bits); page 690 fs/ntfs/aops.c if (!PageUptodate(page)) { page 748 fs/ntfs/aops.c kaddr = kmap_atomic(page, KM_USER0); page 795 fs/ntfs/aops.c zero_user(page, bh_offset(bh), blocksize); page 822 fs/ntfs/aops.c if (unlikely(!PageUptodate(page))) { page 832 fs/ntfs/aops.c SetPageUptodate(page); page 866 fs/ntfs/aops.c redirty_page_for_writepage(wbc, page); page 869 fs/ntfs/aops.c SetPageError(page); page 872 fs/ntfs/aops.c BUG_ON(PageWriteback(page)); page 873 fs/ntfs/aops.c set_page_writeback(page); /* Keeps try_to_free_buffers() away. */ page 885 fs/ntfs/aops.c unlock_page(page); page 889 fs/ntfs/aops.c end_page_writeback(page); page 923 fs/ntfs/aops.c struct inode *vi = page->mapping->host; page 938 fs/ntfs/aops.c "0x%lx.", vi->i_ino, ni->type, page->index); page 960 fs/ntfs/aops.c bh = head = page_buffers(page); page 969 fs/ntfs/aops.c rec_block = block = (sector_t)page->index << page 1119 fs/ntfs/aops.c kaddr = kmap(page); page 1121 fs/ntfs/aops.c BUG_ON(!PageUptodate(page)); page 1122 fs/ntfs/aops.c ClearPageUptodate(page); page 1136 fs/ntfs/aops.c mft_no = (((s64)page->index << PAGE_CACHE_SHIFT) + ofs) page 1175 fs/ntfs/aops.c ni->type, page->index, ofs); page 1191 fs/ntfs/aops.c flush_dcache_page(page); page 1223 fs/ntfs/aops.c page->index, bh_offset(tbh)); page 1252 fs/ntfs/aops.c mft_no = (((s64)page->index << PAGE_CACHE_SHIFT) + ofs) page 1272 fs/ntfs/aops.c flush_dcache_page(page); page 1295 fs/ntfs/aops.c SetPageUptodate(page); page 1296 fs/ntfs/aops.c kunmap(page); page 1304 fs/ntfs/aops.c SetPageError(page); page 1310 fs/ntfs/aops.c "record 0x%lx.", page->index << page 1312 fs/ntfs/aops.c redirty_page_for_writepage(wbc, page); page 1313 fs/ntfs/aops.c unlock_page(page); page 1320 fs/ntfs/aops.c BUG_ON(PageWriteback(page)); page 1321 fs/ntfs/aops.c set_page_writeback(page); page 1322 fs/ntfs/aops.c unlock_page(page); page 1323 fs/ntfs/aops.c end_page_writeback(page); page 1356 fs/ntfs/aops.c struct inode *vi = page->mapping->host; page 1365 fs/ntfs/aops.c BUG_ON(!PageLocked(page)); page 1368 fs/ntfs/aops.c if (unlikely(page->index >= (i_size + PAGE_CACHE_SIZE - 1) >> page 1374 fs/ntfs/aops.c block_invalidatepage(page, 0); page 1375 fs/ntfs/aops.c unlock_page(page); page 1390 fs/ntfs/aops.c unlock_page(page); page 1401 fs/ntfs/aops.c unlock_page(page); page 1408 fs/ntfs/aops.c unlock_page(page); page 1417 fs/ntfs/aops.c if (page->index >= (i_size >> PAGE_CACHE_SHIFT)) { page 1420 fs/ntfs/aops.c zero_user_segment(page, ofs, PAGE_CACHE_SIZE); page 1424 fs/ntfs/aops.c return ntfs_write_mst_block(page, wbc); page 1426 fs/ntfs/aops.c return ntfs_write_block(page, wbc); page 1436 fs/ntfs/aops.c BUG_ON(page_has_buffers(page)); page 1437 fs/ntfs/aops.c BUG_ON(!PageUptodate(page)); page 1438 fs/ntfs/aops.c if (unlikely(page->index > 0)) { page 1440 fs/ntfs/aops.c "Aborting write.", page->index); page 1441 fs/ntfs/aops.c BUG_ON(PageWriteback(page)); page 1442 fs/ntfs/aops.c set_page_writeback(page); page 1443 fs/ntfs/aops.c unlock_page(page); page 1444 fs/ntfs/aops.c end_page_writeback(page); page 1480 fs/ntfs/aops.c BUG_ON(PageWriteback(page)); page 1481 fs/ntfs/aops.c set_page_writeback(page); page 1482 fs/ntfs/aops.c unlock_page(page); page 1497 fs/ntfs/aops.c addr = kmap_atomic(page, KM_USER0); page 1505 fs/ntfs/aops.c flush_dcache_page(page); page 1508 fs/ntfs/aops.c end_page_writeback(page); page 1522 fs/ntfs/aops.c redirty_page_for_writepage(wbc, page); page 1527 fs/ntfs/aops.c SetPageError(page); page 1530 fs/ntfs/aops.c unlock_page(page); page 1592 fs/ntfs/aops.c struct address_space *mapping = page->mapping; page 1597 fs/ntfs/aops.c BUG_ON(!PageUptodate(page)); page 1601 fs/ntfs/aops.c if (unlikely(!page_has_buffers(page))) { page 1603 fs/ntfs/aops.c bh = head = alloc_page_buffers(page, bh_size, 1); page 1605 fs/ntfs/aops.c if (likely(!page_has_buffers(page))) { page 1614 fs/ntfs/aops.c attach_page_buffers(page, head); page 1618 fs/ntfs/aops.c bh = head = page_buffers(page); page 1629 fs/ntfs/aops.c __set_page_dirty_nobuffers(page); page 42 fs/ntfs/aops.h kunmap(page); page 43 fs/ntfs/aops.h page_cache_release(page); page 86 fs/ntfs/aops.h static inline struct page *ntfs_map_page(struct address_space *mapping, page 89 fs/ntfs/aops.h struct page *page = read_mapping_page(mapping, index, NULL); page 91 fs/ntfs/aops.h if (!IS_ERR(page)) { page 92 fs/ntfs/aops.h kmap(page); page 93 fs/ntfs/aops.h if (!PageError(page)) page 94 fs/ntfs/aops.h return page; page 95 fs/ntfs/aops.h ntfs_unmap_page(page); page 98 fs/ntfs/aops.h return page; page 91 fs/ntfs/attrib.c struct page *put_this_page = NULL; page 153 fs/ntfs/attrib.c put_this_page = old_ctx.ntfs_ino->page; page 1543 fs/ntfs/attrib.c struct page *page; page 1579 fs/ntfs/attrib.c page = find_or_create_page(vi->i_mapping, 0, page 1581 fs/ntfs/attrib.c if (unlikely(!page)) page 1596 fs/ntfs/attrib.c page = NULL; page 1657 fs/ntfs/attrib.c if (page && !PageUptodate(page)) { page 1658 fs/ntfs/attrib.c kaddr = kmap_atomic(page, KM_USER0); page 1664 fs/ntfs/attrib.c flush_dcache_page(page); page 1665 fs/ntfs/attrib.c SetPageUptodate(page); page 1747 fs/ntfs/attrib.c if (page) { page 1748 fs/ntfs/attrib.c set_page_dirty(page); page 1749 fs/ntfs/attrib.c unlock_page(page); page 1750 fs/ntfs/attrib.c mark_page_accessed(page); page 1751 fs/ntfs/attrib.c page_cache_release(page); page 1807 fs/ntfs/attrib.c if (page) { page 1808 fs/ntfs/attrib.c kaddr = kmap_atomic(page, KM_USER0); page 1837 fs/ntfs/attrib.c unlock_page(page); page 1838 fs/ntfs/attrib.c page_cache_release(page); page 2497 fs/ntfs/attrib.c struct page *page; page 2529 fs/ntfs/attrib.c page = read_mapping_page(mapping, idx, NULL); page 2530 fs/ntfs/attrib.c if (IS_ERR(page)) { page 2533 fs/ntfs/attrib.c return PTR_ERR(page); page 2542 fs/ntfs/attrib.c kaddr = kmap_atomic(page, KM_USER0); page 2544 fs/ntfs/attrib.c flush_dcache_page(page); page 2546 fs/ntfs/attrib.c set_page_dirty(page); page 2547 fs/ntfs/attrib.c page_cache_release(page); page 2557 fs/ntfs/attrib.c page = grab_cache_page(mapping, idx); page 2558 fs/ntfs/attrib.c if (unlikely(!page)) { page 2563 fs/ntfs/attrib.c kaddr = kmap_atomic(page, KM_USER0); page 2565 fs/ntfs/attrib.c flush_dcache_page(page); page 2571 fs/ntfs/attrib.c if (page_has_buffers(page)) { page 2574 fs/ntfs/attrib.c bh = head = page_buffers(page); page 2580 fs/ntfs/attrib.c SetPageUptodate(page); page 2585 fs/ntfs/attrib.c set_page_dirty(page); page 2587 fs/ntfs/attrib.c unlock_page(page); page 2588 fs/ntfs/attrib.c page_cache_release(page); page 2594 fs/ntfs/attrib.c page = read_mapping_page(mapping, idx, NULL); page 2595 fs/ntfs/attrib.c if (IS_ERR(page)) { page 2598 fs/ntfs/attrib.c return PTR_ERR(page); page 2600 fs/ntfs/attrib.c kaddr = kmap_atomic(page, KM_USER0); page 2602 fs/ntfs/attrib.c flush_dcache_page(page); page 2604 fs/ntfs/attrib.c set_page_dirty(page); page 2605 fs/ntfs/attrib.c page_cache_release(page); page 53 fs/ntfs/bitmap.c struct page *page; page 75 fs/ntfs/bitmap.c page = ntfs_map_page(mapping, index); page 76 fs/ntfs/bitmap.c if (IS_ERR(page)) { page 79 fs/ntfs/bitmap.c "%li), aborting.", PTR_ERR(page)); page 80 fs/ntfs/bitmap.c return PTR_ERR(page); page 82 fs/ntfs/bitmap.c kaddr = page_address(page); page 124 fs/ntfs/bitmap.c flush_dcache_page(page); page 125 fs/ntfs/bitmap.c set_page_dirty(page); page 126 fs/ntfs/bitmap.c ntfs_unmap_page(page); page 127 fs/ntfs/bitmap.c page = ntfs_map_page(mapping, ++index); page 128 fs/ntfs/bitmap.c if (IS_ERR(page)) page 130 fs/ntfs/bitmap.c kaddr = page_address(page); page 160 fs/ntfs/bitmap.c flush_dcache_page(page); page 161 fs/ntfs/bitmap.c set_page_dirty(page); page 162 fs/ntfs/bitmap.c ntfs_unmap_page(page); page 172 fs/ntfs/bitmap.c return PTR_ERR(page); page 181 fs/ntfs/bitmap.c "%li), aborting.", PTR_ERR(page)); page 187 fs/ntfs/bitmap.c "Unmount and run chkdsk.", PTR_ERR(page), pos); page 190 fs/ntfs/bitmap.c return PTR_ERR(page); page 102 fs/ntfs/compress.c u8 *kp = page_address(page); page 106 fs/ntfs/compress.c if (((s64)page->index << PAGE_CACHE_SHIFT) >= initialized_size) { page 125 fs/ntfs/compress.c if ((page->index >= (initialized_size >> PAGE_CACHE_SHIFT)) && page 127 fs/ntfs/compress.c zero_partial_compressed_page(page, initialized_size); page 184 fs/ntfs/compress.c struct page *dp; /* Current destination page being worked on. */ page 485 fs/ntfs/compress.c struct address_space *mapping = page->mapping; page 494 fs/ntfs/compress.c unsigned long offset, index = page->index; page 521 fs/ntfs/compress.c struct page **pages; page 533 fs/ntfs/compress.c pages = kmalloc(nr_pages * sizeof(struct page *), GFP_NOFS); page 542 fs/ntfs/compress.c unlock_page(page); page 553 fs/ntfs/compress.c pages[xpage] = page; page 568 fs/ntfs/compress.c zero_user(page, 0, PAGE_CACHE_SIZE); page 570 fs/ntfs/compress.c SetPageUptodate(page); page 571 fs/ntfs/compress.c unlock_page(page); page 579 fs/ntfs/compress.c page = pages[i]; page 580 fs/ntfs/compress.c if (page) { page 586 fs/ntfs/compress.c if (!PageDirty(page) && (!PageUptodate(page) || page 587 fs/ntfs/compress.c PageError(page))) { page 588 fs/ntfs/compress.c ClearPageError(page); page 589 fs/ntfs/compress.c kmap(page); page 592 fs/ntfs/compress.c unlock_page(page); page 593 fs/ntfs/compress.c page_cache_release(page); page 754 fs/ntfs/compress.c page = pages[cur_page]; page 755 fs/ntfs/compress.c if (page) { page 762 fs/ntfs/compress.c clear_page(page_address(page)); page 764 fs/ntfs/compress.c memset(page_address(page) + cur_ofs, 0, page 767 fs/ntfs/compress.c flush_dcache_page(page); page 768 fs/ntfs/compress.c kunmap(page); page 769 fs/ntfs/compress.c SetPageUptodate(page); page 770 fs/ntfs/compress.c unlock_page(page); page 774 fs/ntfs/compress.c page_cache_release(page); page 784 fs/ntfs/compress.c page = pages[cur_page]; page 785 fs/ntfs/compress.c if (page) page 786 fs/ntfs/compress.c memset(page_address(page) + cur_ofs, 0, page 816 fs/ntfs/compress.c page = pages[cur_page]; page 817 fs/ntfs/compress.c if (page) page 818 fs/ntfs/compress.c memcpy(page_address(page) + cur_ofs, cb_pos, page 827 fs/ntfs/compress.c page = pages[cur_page]; page 828 fs/ntfs/compress.c if (page) page 829 fs/ntfs/compress.c memcpy(page_address(page) + cur_ofs, cb_pos, page 838 fs/ntfs/compress.c page = pages[cur2_page]; page 839 fs/ntfs/compress.c if (page) { page 844 fs/ntfs/compress.c handle_bounds_compressed_page(page, i_size, page 846 fs/ntfs/compress.c flush_dcache_page(page); page 847 fs/ntfs/compress.c kunmap(page); page 848 fs/ntfs/compress.c SetPageUptodate(page); page 849 fs/ntfs/compress.c unlock_page(page); page 853 fs/ntfs/compress.c page_cache_release(page); page 881 fs/ntfs/compress.c page = pages[prev_cur_page]; page 882 fs/ntfs/compress.c if (page) { page 883 fs/ntfs/compress.c flush_dcache_page(page); page 884 fs/ntfs/compress.c kunmap(page); page 885 fs/ntfs/compress.c unlock_page(page); page 887 fs/ntfs/compress.c page_cache_release(page); page 907 fs/ntfs/compress.c page = pages[cur_page]; page 908 fs/ntfs/compress.c if (page) { page 912 fs/ntfs/compress.c "0x%lx.", ni->mft_no, page->index); page 913 fs/ntfs/compress.c flush_dcache_page(page); page 914 fs/ntfs/compress.c kunmap(page); page 915 fs/ntfs/compress.c unlock_page(page); page 917 fs/ntfs/compress.c page_cache_release(page); page 958 fs/ntfs/compress.c page = pages[i]; page 959 fs/ntfs/compress.c if (page) { page 960 fs/ntfs/compress.c flush_dcache_page(page); page 961 fs/ntfs/compress.c kunmap(page); page 962 fs/ntfs/compress.c unlock_page(page); page 964 fs/ntfs/compress.c page_cache_release(page); page 94 fs/ntfs/dir.c struct page *page; page 320 fs/ntfs/dir.c page = ntfs_map_page(ia_mapping, vcn << page 322 fs/ntfs/dir.c if (IS_ERR(page)) { page 324 fs/ntfs/dir.c -PTR_ERR(page)); page 325 fs/ntfs/dir.c err = PTR_ERR(page); page 328 fs/ntfs/dir.c lock_page(page); page 329 fs/ntfs/dir.c kaddr = (u8*)page_address(page); page 449 fs/ntfs/dir.c unlock_page(page); page 450 fs/ntfs/dir.c ntfs_unmap_page(page); page 482 fs/ntfs/dir.c unlock_page(page); page 483 fs/ntfs/dir.c ntfs_unmap_page(page); page 565 fs/ntfs/dir.c unlock_page(page); page 566 fs/ntfs/dir.c ntfs_unmap_page(page); page 579 fs/ntfs/dir.c unlock_page(page); page 580 fs/ntfs/dir.c ntfs_unmap_page(page); page 586 fs/ntfs/dir.c unlock_page(page); page 587 fs/ntfs/dir.c ntfs_unmap_page(page); page 648 fs/ntfs/dir.c struct page *page; page 798 fs/ntfs/dir.c page = ntfs_map_page(ia_mapping, vcn << page 800 fs/ntfs/dir.c if (IS_ERR(page)) { page 802 fs/ntfs/dir.c -PTR_ERR(page)); page 803 fs/ntfs/dir.c err = PTR_ERR(page); page 806 fs/ntfs/dir.c lock_page(page); page 807 fs/ntfs/dir.c kaddr = (u8*)page_address(page); page 910 fs/ntfs/dir.c unlock_page(page); page 911 fs/ntfs/dir.c ntfs_unmap_page(page); page 974 fs/ntfs/dir.c unlock_page(page); page 975 fs/ntfs/dir.c ntfs_unmap_page(page); page 986 fs/ntfs/dir.c unlock_page(page); page 987 fs/ntfs/dir.c ntfs_unmap_page(page); page 1114 fs/ntfs/dir.c struct page *bmp_page = NULL, *ia_page = NULL; page 125 fs/ntfs/file.c struct page *page = NULL; page 234 fs/ntfs/file.c page = read_mapping_page(mapping, index, NULL); page 235 fs/ntfs/file.c if (IS_ERR(page)) { page 236 fs/ntfs/file.c err = PTR_ERR(page); page 239 fs/ntfs/file.c if (unlikely(PageError(page))) { page 240 fs/ntfs/file.c page_cache_release(page); page 254 fs/ntfs/file.c set_page_dirty(page); page 255 fs/ntfs/file.c page_cache_release(page); page 498 fs/ntfs/file.c struct page *page; page 534 fs/ntfs/file.c page = pages[u]; page 535 fs/ntfs/file.c BUG_ON(!page); page 540 fs/ntfs/file.c if (!page_has_buffers(page)) { page 541 fs/ntfs/file.c create_empty_buffers(page, blocksize, 0); page 542 fs/ntfs/file.c if (unlikely(!page_has_buffers(page))) page 562 fs/ntfs/file.c page = pages[u]; page 563 fs/ntfs/file.c bh_pos = (s64)page->index << PAGE_CACHE_SHIFT; page 564 fs/ntfs/file.c bh = head = page_buffers(page); page 587 fs/ntfs/file.c if (PageUptodate(page)) { page 610 fs/ntfs/file.c zero_user(page, bh_offset(bh), page 645 fs/ntfs/file.c if (PageUptodate(page)) { page 686 fs/ntfs/file.c zero_user(page, bh_offset(bh), page 705 fs/ntfs/file.c zero_user(page, bh_offset(bh), page 718 fs/ntfs/file.c kaddr = kmap_atomic(page, KM_USER0); page 728 fs/ntfs/file.c flush_dcache_page(page); page 741 fs/ntfs/file.c if (PageUptodate(page)) { page 745 fs/ntfs/file.c zero_user(page, bh_offset(bh), blocksize); page 865 fs/ntfs/file.c if (PageUptodate(page)) { page 869 fs/ntfs/file.c zero_user(page, bh_offset(bh), page 1117 fs/ntfs/file.c page = bh->b_page; page 1118 fs/ntfs/file.c bh_pos = ((s64)page->index << PAGE_CACHE_SHIFT) + page 1129 fs/ntfs/file.c zero_user_segment(page, bh_offset(bh) + ofs, page 1256 fs/ntfs/file.c page = pages[u]; page 1257 fs/ntfs/file.c bh = head = page_buffers(page); page 1260 fs/ntfs/file.c ((s64)page->index << PAGE_CACHE_SHIFT) + page 1267 fs/ntfs/file.c if (PageUptodate(page)) page 1270 fs/ntfs/file.c zero_user(page, bh_offset(bh), page 1291 fs/ntfs/file.c struct page **last_page = pages + nr_pages; page 1408 fs/ntfs/file.c struct page **last_page = pages + nr_pages; page 1503 fs/ntfs/file.c struct page *page; page 1506 fs/ntfs/file.c page = pages[u]; page 1507 fs/ntfs/file.c bh_pos = (s64)page->index << PAGE_CACHE_SHIFT; page 1508 fs/ntfs/file.c bh = head = page_buffers(page); page 1526 fs/ntfs/file.c if (!partial && !PageUptodate(page)) page 1527 fs/ntfs/file.c SetPageUptodate(page); page 1643 fs/ntfs/file.c struct page *page; page 1654 fs/ntfs/file.c page = pages[0]; page 1655 fs/ntfs/file.c BUG_ON(!page); page 1656 fs/ntfs/file.c vi = page->mapping->host; page 1660 fs/ntfs/file.c vi->i_ino, ni->type, page->index, nr_pages, page 1706 fs/ntfs/file.c kaddr = kmap_atomic(page, KM_USER0); page 1718 fs/ntfs/file.c if (!PageUptodate(page)) { page 1725 fs/ntfs/file.c flush_dcache_page(page); page 1726 fs/ntfs/file.c SetPageUptodate(page); page 1752 fs/ntfs/file.c if (PageUptodate(page)) { page 1760 fs/ntfs/file.c __set_page_dirty_nobuffers(page); page 1791 fs/ntfs/file.c struct page *pages[NTFS_MAX_PAGES_PER_CLUSTER]; page 1792 fs/ntfs/file.c struct page *cached_page = NULL; page 64 fs/ntfs/index.c struct page *page = ictx->page; page 65 fs/ntfs/index.c if (page) { page 66 fs/ntfs/index.c BUG_ON(!PageLocked(page)); page 67 fs/ntfs/index.c unlock_page(page); page 68 fs/ntfs/index.c ntfs_unmap_page(page); page 132 fs/ntfs/index.c struct page *page; page 212 fs/ntfs/index.c ictx->page = NULL; page 276 fs/ntfs/index.c page = ntfs_map_page(ia_mapping, vcn << page 278 fs/ntfs/index.c if (IS_ERR(page)) { page 280 fs/ntfs/index.c -PTR_ERR(page)); page 281 fs/ntfs/index.c err = PTR_ERR(page); page 284 fs/ntfs/index.c lock_page(page); page 285 fs/ntfs/index.c kaddr = (u8*)page_address(page); page 381 fs/ntfs/index.c ictx->page = page; page 432 fs/ntfs/index.c unlock_page(page); page 433 fs/ntfs/index.c ntfs_unmap_page(page); page 439 fs/ntfs/index.c unlock_page(page); page 440 fs/ntfs/index.c ntfs_unmap_page(page); page 85 fs/ntfs/index.h struct page *page; page 116 fs/ntfs/index.h flush_dcache_page(ictx->page); page 142 fs/ntfs/index.h mark_ntfs_record_dirty(ictx->page, page 143 fs/ntfs/index.h (u8*)ictx->ia - (u8*)page_address(ictx->page)); page 339 fs/ntfs/inode.c BUG_ON(ni->page); page 362 fs/ntfs/inode.c BUG_ON(ni->page); page 396 fs/ntfs/inode.c ni->page = NULL; page 3052 fs/ntfs/inode.c mark_ntfs_record_dirty(ctx->ntfs_ino->page, page 87 fs/ntfs/inode.h struct page *page; /* The page containing the mft record of the page 158 fs/ntfs/lcnalloc.c struct page *page = NULL; page 276 fs/ntfs/lcnalloc.c if (likely(page)) { page 279 fs/ntfs/lcnalloc.c flush_dcache_page(page); page 280 fs/ntfs/lcnalloc.c set_page_dirty(page); page 283 fs/ntfs/lcnalloc.c ntfs_unmap_page(page); page 285 fs/ntfs/lcnalloc.c page = ntfs_map_page(mapping, last_read_pos >> page 287 fs/ntfs/lcnalloc.c if (IS_ERR(page)) { page 288 fs/ntfs/lcnalloc.c err = PTR_ERR(page); page 293 fs/ntfs/lcnalloc.c buf = page_address(page) + buf_size; page 744 fs/ntfs/lcnalloc.c if (likely(page && !IS_ERR(page))) { page 747 fs/ntfs/lcnalloc.c flush_dcache_page(page); page 748 fs/ntfs/lcnalloc.c set_page_dirty(page); page 751 fs/ntfs/lcnalloc.c ntfs_unmap_page(page); page 388 fs/ntfs/logfile.c struct page *page; page 399 fs/ntfs/logfile.c page = ntfs_map_page(vi->i_mapping, idx); page 400 fs/ntfs/logfile.c if (IS_ERR(page)) { page 403 fs/ntfs/logfile.c err = PTR_ERR(page); page 409 fs/ntfs/logfile.c memcpy((u8*)trp + have_read, page_address(page), size); page 410 fs/ntfs/logfile.c ntfs_unmap_page(page); page 489 fs/ntfs/logfile.c struct page *page = NULL; page 542 fs/ntfs/logfile.c if (!page || page->index != idx) { page 543 fs/ntfs/logfile.c if (page) page 544 fs/ntfs/logfile.c ntfs_unmap_page(page); page 545 fs/ntfs/logfile.c page = ntfs_map_page(mapping, idx); page 546 fs/ntfs/logfile.c if (IS_ERR(page)) { page 552 fs/ntfs/logfile.c kaddr = (u8*)page_address(page) + (pos & ~PAGE_CACHE_MASK); page 605 fs/ntfs/logfile.c ntfs_unmap_page(page); page 612 fs/ntfs/logfile.c if (page) page 613 fs/ntfs/logfile.c ntfs_unmap_page(page); page 51 fs/ntfs/mft.c struct page *page; page 55 fs/ntfs/mft.c BUG_ON(ni->page); page 74 fs/ntfs/mft.c page = ERR_PTR(-ENOENT); page 83 fs/ntfs/mft.c page = ntfs_map_page(mft_vi->i_mapping, index); page 84 fs/ntfs/mft.c if (likely(!IS_ERR(page))) { page 86 fs/ntfs/mft.c if (likely(ntfs_is_mft_recordp((le32*)(page_address(page) + page 88 fs/ntfs/mft.c ni->page = page; page 90 fs/ntfs/mft.c return page_address(page) + ofs; page 94 fs/ntfs/mft.c ntfs_unmap_page(page); page 95 fs/ntfs/mft.c page = ERR_PTR(-EIO); page 99 fs/ntfs/mft.c ni->page = NULL; page 101 fs/ntfs/mft.c return (void*)page; page 192 fs/ntfs/mft.c BUG_ON(!ni->page); page 195 fs/ntfs/mft.c ntfs_unmap_page(ni->page); page 196 fs/ntfs/mft.c ni->page = NULL; page 214 fs/ntfs/mft.c struct page *page = ni->page; page 216 fs/ntfs/mft.c BUG_ON(!page); page 401 fs/ntfs/mft.c mark_ntfs_record_dirty(ni->page, ni->page_ofs); page 468 fs/ntfs/mft.c struct page *page; page 489 fs/ntfs/mft.c page = ntfs_map_page(vol->mftmirr_ino->i_mapping, mft_no >> page 491 fs/ntfs/mft.c if (IS_ERR(page)) { page 493 fs/ntfs/mft.c err = PTR_ERR(page); page 496 fs/ntfs/mft.c lock_page(page); page 497 fs/ntfs/mft.c BUG_ON(!PageUptodate(page)); page 498 fs/ntfs/mft.c ClearPageUptodate(page); page 502 fs/ntfs/mft.c kmirr = page_address(page) + page_ofs; page 506 fs/ntfs/mft.c if (unlikely(!page_has_buffers(page))) { page 509 fs/ntfs/mft.c bh = head = alloc_page_buffers(page, blocksize, 1); page 516 fs/ntfs/mft.c attach_page_buffers(page, head); page 518 fs/ntfs/mft.c bh = head = page_buffers(page); page 523 fs/ntfs/mft.c m_start = kmirr - (u8*)page_address(page); page 619 fs/ntfs/mft.c flush_dcache_page(page); page 620 fs/ntfs/mft.c SetPageUptodate(page); page 621 fs/ntfs/mft.c unlock_page(page); page 622 fs/ntfs/mft.c ntfs_unmap_page(page); page 672 fs/ntfs/mft.c struct page *page = ni->page; page 685 fs/ntfs/mft.c BUG_ON(!PageLocked(page)); page 694 fs/ntfs/mft.c bh = head = page_buffers(page); page 804 fs/ntfs/mft.c if (PageUptodate(page)) page 1138 fs/ntfs/mft.c struct page *page; page 1192 fs/ntfs/mft.c page = ntfs_map_page(mftbmp_mapping, page 1194 fs/ntfs/mft.c if (IS_ERR(page)) { page 1197 fs/ntfs/mft.c return PTR_ERR(page); page 1199 fs/ntfs/mft.c buf = (u8*)page_address(page) + page_ofs; page 1214 fs/ntfs/mft.c ntfs_unmap_page(page); page 1218 fs/ntfs/mft.c flush_dcache_page(page); page 1219 fs/ntfs/mft.c set_page_dirty(page); page 1220 fs/ntfs/mft.c ntfs_unmap_page(page); page 1232 fs/ntfs/mft.c ntfs_unmap_page(page); page 1283 fs/ntfs/mft.c struct page *page; page 1330 fs/ntfs/mft.c page = ntfs_map_page(vol->lcnbmp_ino->i_mapping, page 1332 fs/ntfs/mft.c if (IS_ERR(page)) { page 1335 fs/ntfs/mft.c return PTR_ERR(page); page 1337 fs/ntfs/mft.c b = (u8*)page_address(page) + (ll & ~PAGE_CACHE_MASK); page 1343 fs/ntfs/mft.c flush_dcache_page(page); page 1344 fs/ntfs/mft.c set_page_dirty(page); page 1346 fs/ntfs/mft.c ntfs_unmap_page(page); page 1354 fs/ntfs/mft.c ntfs_unmap_page(page); page 2095 fs/ntfs/mft.c struct page *page; page 2120 fs/ntfs/mft.c page = ntfs_map_page(mft_vi->i_mapping, index); page 2121 fs/ntfs/mft.c if (IS_ERR(page)) { page 2124 fs/ntfs/mft.c return PTR_ERR(page); page 2126 fs/ntfs/mft.c lock_page(page); page 2127 fs/ntfs/mft.c BUG_ON(!PageUptodate(page)); page 2128 fs/ntfs/mft.c ClearPageUptodate(page); page 2129 fs/ntfs/mft.c m = (MFT_RECORD*)((u8*)page_address(page) + ofs); page 2134 fs/ntfs/mft.c SetPageUptodate(page); page 2135 fs/ntfs/mft.c unlock_page(page); page 2136 fs/ntfs/mft.c ntfs_unmap_page(page); page 2139 fs/ntfs/mft.c flush_dcache_page(page); page 2140 fs/ntfs/mft.c SetPageUptodate(page); page 2141 fs/ntfs/mft.c unlock_page(page); page 2147 fs/ntfs/mft.c mark_ntfs_record_dirty(page, ofs); page 2148 fs/ntfs/mft.c ntfs_unmap_page(page); page 2249 fs/ntfs/mft.c struct page *page; page 2521 fs/ntfs/mft.c page = ntfs_map_page(vol->mft_ino->i_mapping, index); page 2522 fs/ntfs/mft.c if (IS_ERR(page)) { page 2525 fs/ntfs/mft.c err = PTR_ERR(page); page 2528 fs/ntfs/mft.c lock_page(page); page 2529 fs/ntfs/mft.c BUG_ON(!PageUptodate(page)); page 2530 fs/ntfs/mft.c ClearPageUptodate(page); page 2531 fs/ntfs/mft.c m = (MFT_RECORD*)((u8*)page_address(page) + ofs); page 2543 fs/ntfs/mft.c SetPageUptodate(page); page 2544 fs/ntfs/mft.c unlock_page(page); page 2545 fs/ntfs/mft.c ntfs_unmap_page(page); page 2562 fs/ntfs/mft.c SetPageUptodate(page); page 2563 fs/ntfs/mft.c unlock_page(page); page 2564 fs/ntfs/mft.c ntfs_unmap_page(page); page 2576 fs/ntfs/mft.c flush_dcache_page(page); page 2577 fs/ntfs/mft.c SetPageUptodate(page); page 2599 fs/ntfs/mft.c flush_dcache_page(page); page 2601 fs/ntfs/mft.c mark_ntfs_record_dirty(page, ofs); page 2602 fs/ntfs/mft.c unlock_page(page); page 2603 fs/ntfs/mft.c ntfs_unmap_page(page); page 2613 fs/ntfs/mft.c mark_ntfs_record_dirty(page, ofs); page 2614 fs/ntfs/mft.c unlock_page(page); page 2619 fs/ntfs/mft.c ntfs_unmap_page(page); page 2632 fs/ntfs/mft.c flush_dcache_page(page); page 2634 fs/ntfs/mft.c mark_ntfs_record_dirty(page, ofs); page 2635 fs/ntfs/mft.c unlock_page(page); page 2636 fs/ntfs/mft.c ntfs_unmap_page(page); page 2708 fs/ntfs/mft.c ni->page = page; page 2720 fs/ntfs/mft.c mark_ntfs_record_dirty(page, ofs); page 2721 fs/ntfs/mft.c unlock_page(page); page 57 fs/ntfs/mft.h flush_dcache_page(ni->page); page 104 fs/ntfs/mft.h struct page *page = ni->page; page 107 fs/ntfs/mft.h BUG_ON(!page); page 108 fs/ntfs/mft.h lock_page(page); page 110 fs/ntfs/mft.h unlock_page(page); page 1061 fs/ntfs/super.c struct page *mft_page, *mirr_page; page 1241 fs/ntfs/super.c struct page *page; page 1292 fs/ntfs/super.c page = ntfs_map_page(vi->i_mapping, 0); page 1293 fs/ntfs/super.c if (IS_ERR(page)) { page 1295 fs/ntfs/super.c ret = PTR_ERR(page); page 1298 fs/ntfs/super.c kaddr = (u32*)page_address(page); page 1322 fs/ntfs/super.c ntfs_unmap_page(page); page 1416 fs/ntfs/super.c struct page *page; page 1512 fs/ntfs/super.c page = ntfs_map_page(vol->usnjrnl_max_ino->i_mapping, 0); page 1513 fs/ntfs/super.c if (IS_ERR(page)) { page 1518 fs/ntfs/super.c uh = (USN_HEADER*)page_address(page); page 1526 fs/ntfs/super.c ntfs_unmap_page(page); page 1537 fs/ntfs/super.c ntfs_unmap_page(page); page 1550 fs/ntfs/super.c ntfs_unmap_page(page); page 1553 fs/ntfs/super.c ntfs_unmap_page(page); page 1569 fs/ntfs/super.c struct page *page; page 1595 fs/ntfs/super.c page = ntfs_map_page(ino->i_mapping, index); page 1596 fs/ntfs/super.c if (IS_ERR(page)) page 1599 fs/ntfs/super.c page_address(page), size); page 1600 fs/ntfs/super.c ntfs_unmap_page(page); page 1634 fs/ntfs/super.c struct page *page; page 1664 fs/ntfs/super.c page = ntfs_map_page(ino->i_mapping, index); page 1665 fs/ntfs/super.c if (IS_ERR(page)) page 1668 fs/ntfs/super.c page_address(page), size); page 1669 fs/ntfs/super.c ntfs_unmap_page(page); page 2474 fs/ntfs/super.c struct page *page; page 2496 fs/ntfs/super.c page = read_mapping_page(mapping, index, NULL); page 2498 fs/ntfs/super.c if (IS_ERR(page)) { page 2504 fs/ntfs/super.c kaddr = (u32*)kmap_atomic(page, KM_USER0); page 2515 fs/ntfs/super.c page_cache_release(page); page 2554 fs/ntfs/super.c struct page *page; page 2567 fs/ntfs/super.c page = read_mapping_page(mapping, index, NULL); page 2569 fs/ntfs/super.c if (IS_ERR(page)) { page 2575 fs/ntfs/super.c kaddr = (u32*)kmap_atomic(page, KM_USER0); page 2586 fs/ntfs/super.c page_cache_release(page); page 52 fs/ntfs/usnjrnl.c struct page *page; page 55 fs/ntfs/usnjrnl.c page = ntfs_map_page(vol->usnjrnl_max_ino->i_mapping, 0); page 56 fs/ntfs/usnjrnl.c if (IS_ERR(page)) { page 61 fs/ntfs/usnjrnl.c uh = (USN_HEADER*)page_address(page); page 74 fs/ntfs/usnjrnl.c flush_dcache_page(page); page 75 fs/ntfs/usnjrnl.c set_page_dirty(page); page 76 fs/ntfs/usnjrnl.c ntfs_unmap_page(page); page 6420 fs/ocfs2/alloc.c ret = ocfs2_map_page_blocks(page, phys, inode, from, to, 0); page 6425 fs/ocfs2/alloc.c zero_user_segment(page, from, to); page 6432 fs/ocfs2/alloc.c ret = walk_page_buffers(handle, page_buffers(page), page 6440 fs/ocfs2/alloc.c ret = walk_page_buffers(handle, page_buffers(page), page 6449 fs/ocfs2/alloc.c SetPageUptodate(page); page 6451 fs/ocfs2/alloc.c flush_dcache_page(page); page 6459 fs/ocfs2/alloc.c struct page *page; page 6470 fs/ocfs2/alloc.c page = pages[i]; page 6473 fs/ocfs2/alloc.c if ((end >> PAGE_CACHE_SHIFT) == page->index) page 6479 fs/ocfs2/alloc.c ocfs2_map_and_dirty_page(inode, handle, from, to, page, 1, page 6482 fs/ocfs2/alloc.c start = (page->index + 1) << PAGE_CACHE_SHIFT; page 6543 fs/ocfs2/alloc.c struct page **pages = NULL; page 6556 fs/ocfs2/alloc.c sizeof(struct page *), GFP_NOFS); page 6663 fs/ocfs2/alloc.c struct page **pages = NULL; page 6671 fs/ocfs2/alloc.c sizeof(struct page *), GFP_NOFS); page 244 fs/ocfs2/aops.c kaddr = kmap_atomic(page, KM_USER0); page 249 fs/ocfs2/aops.c flush_dcache_page(page); page 252 fs/ocfs2/aops.c SetPageUptodate(page); page 262 fs/ocfs2/aops.c BUG_ON(!PageLocked(page)); page 271 fs/ocfs2/aops.c ret = ocfs2_read_inline_data(inode, page, di_bh); page 273 fs/ocfs2/aops.c unlock_page(page); page 281 fs/ocfs2/aops.c struct inode *inode = page->mapping->host; page 283 fs/ocfs2/aops.c loff_t start = (loff_t)page->index << PAGE_CACHE_SHIFT; page 286 fs/ocfs2/aops.c mlog_entry("(0x%p, %lu)\n", file, (page ? page->index : 0)); page 288 fs/ocfs2/aops.c ret = ocfs2_inode_lock_with_page(inode, NULL, 0, page); page 312 fs/ocfs2/aops.c zero_user(page, 0, PAGE_SIZE); page 313 fs/ocfs2/aops.c SetPageUptodate(page); page 319 fs/ocfs2/aops.c ret = ocfs2_readpage_inline(inode, page); page 321 fs/ocfs2/aops.c ret = block_read_full_page(page, ocfs2_get_block); page 330 fs/ocfs2/aops.c unlock_page(page); page 351 fs/ocfs2/aops.c struct page *last; page 377 fs/ocfs2/aops.c last = list_entry(pages->prev, struct page, lru); page 406 fs/ocfs2/aops.c mlog_entry("(0x%p)\n", page); page 408 fs/ocfs2/aops.c ret = block_write_full_page(page, ocfs2_get_block, wbc); page 425 fs/ocfs2/aops.c ret = block_prepare_write(page, from, to, ocfs2_get_block); page 486 fs/ocfs2/aops.c page_buffers(page), page 668 fs/ocfs2/aops.c journal_t *journal = OCFS2_SB(page->mapping->host->i_sb)->journal->j_journal; page 670 fs/ocfs2/aops.c jbd2_journal_invalidatepage(journal, page, offset); page 675 fs/ocfs2/aops.c journal_t *journal = OCFS2_SB(page->mapping->host->i_sb)->journal->j_journal; page 677 fs/ocfs2/aops.c if (!page_has_buffers(page)) page 679 fs/ocfs2/aops.c return jbd2_journal_try_to_free_buffers(journal, page, wait); page 755 fs/ocfs2/aops.c kaddr = kmap_atomic(page, KM_USER0); page 779 fs/ocfs2/aops.c u64 offset = page_offset(page) + block_start; page 806 fs/ocfs2/aops.c if (!page_has_buffers(page)) page 807 fs/ocfs2/aops.c create_empty_buffers(page, bsize, 0); page 809 fs/ocfs2/aops.c head = page_buffers(page); page 821 fs/ocfs2/aops.c if (PageUptodate(page)) page 838 fs/ocfs2/aops.c if (PageUptodate(page)) { page 843 fs/ocfs2/aops.c ocfs2_should_read_blk(inode, page, block_start) && page 877 fs/ocfs2/aops.c zero_user(page, block_start, bh->b_size); page 943 fs/ocfs2/aops.c struct page *w_pages[OCFS2_MAX_CTXT_PAGES]; page 945 fs/ocfs2/aops.c struct page *w_target_page; page 1025 fs/ocfs2/aops.c BUG_ON(!PageLocked(page)); page 1026 fs/ocfs2/aops.c if (!page_has_buffers(page)) page 1029 fs/ocfs2/aops.c bh = head = page_buffers(page); page 1036 fs/ocfs2/aops.c if (!PageUptodate(page)) { page 1042 fs/ocfs2/aops.c zero_user_segment(page, start, end); page 1067 fs/ocfs2/aops.c struct page *tmppage; page 1104 fs/ocfs2/aops.c if (page == wc->w_target_page) { page 1109 fs/ocfs2/aops.c ret = ocfs2_map_page_blocks(page, p_blkno, inode, page 1113 fs/ocfs2/aops.c ret = ocfs2_map_page_blocks(page, p_blkno, inode, page 1137 fs/ocfs2/aops.c ret = ocfs2_map_page_blocks(page, p_blkno, inode, page 1155 fs/ocfs2/aops.c if (new && !PageUptodate(page)) page 1156 fs/ocfs2/aops.c ocfs2_clear_page_regions(page, OCFS2_SB(inode->i_sb), page 1159 fs/ocfs2/aops.c flush_dcache_page(page); page 1510 fs/ocfs2/aops.c struct page *page; page 1514 fs/ocfs2/aops.c page = find_or_create_page(mapping, 0, GFP_NOFS); page 1515 fs/ocfs2/aops.c if (!page) { page 1524 fs/ocfs2/aops.c wc->w_pages[0] = wc->w_target_page = page; page 1546 fs/ocfs2/aops.c if (!PageUptodate(page)) { page 1547 fs/ocfs2/aops.c ret = ocfs2_read_inline_data(inode, page, wc->w_di_bh); page 1885 fs/ocfs2/aops.c struct page *tmppage; page 1964 fs/ocfs2/aops.c ret = ocfs2_write_end_nolock(mapping, pos, len, copied, page, fsdata); page 127 fs/ocfs2/cluster/heartbeat.c struct page **hr_slot_data; page 246 fs/ocfs2/cluster/heartbeat.c struct page *page; page 268 fs/ocfs2/cluster/heartbeat.c page = reg->hr_slot_data[current_page]; page 276 fs/ocfs2/cluster/heartbeat.c len = bio_add_page(bio, page, vec_len, vec_start); page 963 fs/ocfs2/cluster/heartbeat.c struct page *page; page 971 fs/ocfs2/cluster/heartbeat.c page = reg->hr_slot_data[i]; page 972 fs/ocfs2/cluster/heartbeat.c if (page) page 973 fs/ocfs2/cluster/heartbeat.c __free_page(page); page 998 fs/ocfs2/cluster/heartbeat.c char *p = (char *)page; page 1021 fs/ocfs2/cluster/heartbeat.c return sprintf(page, "%u\n", reg->hr_block_bytes); page 1035 fs/ocfs2/cluster/heartbeat.c status = o2hb_read_block_input(reg, page, count, page 1049 fs/ocfs2/cluster/heartbeat.c return sprintf(page, "%llu\n", reg->hr_start_block); page 1057 fs/ocfs2/cluster/heartbeat.c char *p = (char *)page; page 1074 fs/ocfs2/cluster/heartbeat.c return sprintf(page, "%d\n", reg->hr_blocks); page 1082 fs/ocfs2/cluster/heartbeat.c char *p = (char *)page; page 1105 fs/ocfs2/cluster/heartbeat.c ret = sprintf(page, "%s\n", reg->hr_dev_name); page 1128 fs/ocfs2/cluster/heartbeat.c struct page *page; page 1157 fs/ocfs2/cluster/heartbeat.c reg->hr_slot_data = kcalloc(reg->hr_num_pages, sizeof(struct page *), page 1165 fs/ocfs2/cluster/heartbeat.c page = alloc_page(GFP_KERNEL); page 1166 fs/ocfs2/cluster/heartbeat.c if (!page) { page 1171 fs/ocfs2/cluster/heartbeat.c reg->hr_slot_data[i] = page; page 1174 fs/ocfs2/cluster/heartbeat.c raw = page_address(page); page 1235 fs/ocfs2/cluster/heartbeat.c char *p = (char *)page; page 1381 fs/ocfs2/cluster/heartbeat.c return sprintf(page, "%u\n", pid); page 1448 fs/ocfs2/cluster/heartbeat.c ret = o2hb_region_attr->show(reg, page); page 1462 fs/ocfs2/cluster/heartbeat.c ret = o2hb_region_attr->store(reg, page, count); page 1553 fs/ocfs2/cluster/heartbeat.c ret = o2hb_heartbeat_group_attr->show(reg, page); page 1567 fs/ocfs2/cluster/heartbeat.c ret = o2hb_heartbeat_group_attr->store(reg, page, count); page 1574 fs/ocfs2/cluster/heartbeat.c return sprintf(page, "%u\n", o2hb_dead_threshold); page 1582 fs/ocfs2/cluster/heartbeat.c char *p = (char *)page; page 173 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%d\n", node->nd_num); page 195 fs/ocfs2/cluster/nodemanager.c char *p = (char *)page; page 228 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u\n", ntohs(node->nd_ipv4_port)); page 235 fs/ocfs2/cluster/nodemanager.c char *p = (char *)page; page 253 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u.%u.%u.%u\n", NIPQUAD(node->nd_ipv4_address)); page 266 fs/ocfs2/cluster/nodemanager.c ret = sscanf(page, "%3u.%3u.%3u.%3u", &octets[3], &octets[2], page 296 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%d\n", node->nd_local); page 304 fs/ocfs2/cluster/nodemanager.c char *p = (char *)page; page 415 fs/ocfs2/cluster/nodemanager.c ret = o2nm_node_attr->show(node, page); page 437 fs/ocfs2/cluster/nodemanager.c ret = o2nm_node_attr->store(node, page, count); page 484 fs/ocfs2/cluster/nodemanager.c char *p = (char *)page; page 503 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u\n", cluster->cl_idle_timeout_ms); page 512 fs/ocfs2/cluster/nodemanager.c ret = o2nm_cluster_attr_write(page, count, &val); page 538 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u\n", cluster->cl_keepalive_delay_ms); page 547 fs/ocfs2/cluster/nodemanager.c ret = o2nm_cluster_attr_write(page, count, &val); page 573 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u\n", cluster->cl_reconnect_delay_ms); page 579 fs/ocfs2/cluster/nodemanager.c return o2nm_cluster_attr_write(page, count, page 622 fs/ocfs2/cluster/nodemanager.c ret = o2nm_cluster_attr->show(cluster, page); page 640 fs/ocfs2/cluster/nodemanager.c ret = o2nm_cluster_attr->store(cluster, page, count); page 378 fs/ocfs2/cluster/tcp.c struct page *page = NULL; page 380 fs/ocfs2/cluster/tcp.c page = alloc_page(GFP_NOFS); page 382 fs/ocfs2/cluster/tcp.c if (sc == NULL || page == NULL) page 401 fs/ocfs2/cluster/tcp.c sc->sc_page = page; page 404 fs/ocfs2/cluster/tcp.c page = NULL; page 407 fs/ocfs2/cluster/tcp.c if (page) page 408 fs/ocfs2/cluster/tcp.c __free_page(page); page 163 fs/ocfs2/cluster/tcp_internal.h struct page *sc_page; page 2239 fs/ocfs2/dlmglue.c unlock_page(page); page 688 fs/ocfs2/file.c struct page *page; page 704 fs/ocfs2/file.c page = grab_cache_page(mapping, index); page 705 fs/ocfs2/file.c if (!page) { page 711 fs/ocfs2/file.c ret = ocfs2_prepare_write_nolock(inode, page, offset, offset); page 718 fs/ocfs2/file.c handle = ocfs2_start_walk_page_trans(inode, page, offset, page 728 fs/ocfs2/file.c ret = block_commit_write(page, offset, offset); page 737 fs/ocfs2/file.c unlock_page(page); page 738 fs/ocfs2/file.c page_cache_release(page); page 83 fs/ocfs2/mmap.c mlog_exit_ptr(vmf->page); page 92 fs/ocfs2/mmap.c loff_t pos = page_offset(page); page 95 fs/ocfs2/mmap.c struct page *locked_page = NULL; page 104 fs/ocfs2/mmap.c if (page->index > last_index) { page 115 fs/ocfs2/mmap.c if (!PageUptodate(page) || page->mapping != inode->i_mapping) { page 130 fs/ocfs2/mmap.c if (page->index == last_index) page 134 fs/ocfs2/mmap.c &fsdata, di_bh, page); page 184 fs/ocfs2/mmap.c ret = __ocfs2_page_mkwrite(inode, di_bh, page); page 66 fs/ocfs2/symlink.c struct page * page; page 68 fs/ocfs2/symlink.c page = read_mapping_page(mapping, 0, NULL); page 69 fs/ocfs2/symlink.c if (IS_ERR(page)) page 71 fs/ocfs2/symlink.c *ppage = page; page 72 fs/ocfs2/symlink.c return kmap(page); page 75 fs/ocfs2/symlink.c return (char*)page; page 137 fs/ocfs2/symlink.c struct page *page = NULL; page 143 fs/ocfs2/symlink.c link = ocfs2_page_getlink(dentry, &page); page 153 fs/ocfs2/symlink.c if (page) { page 154 fs/ocfs2/symlink.c kunmap(page); page 155 fs/ocfs2/symlink.c page_cache_release(page); page 306 fs/omfs/file.c return block_read_full_page(page, omfs_get_block); page 317 fs/omfs/file.c return block_write_full_page(page, omfs_get_block, wbc); page 564 fs/partitions/check.c struct page *page; page 566 fs/partitions/check.c page = read_mapping_page(mapping, (pgoff_t)(n >> (PAGE_CACHE_SHIFT-9)), page 568 fs/partitions/check.c if (!IS_ERR(page)) { page 569 fs/partitions/check.c if (PageError(page)) page 571 fs/partitions/check.c p->v = page; page 572 fs/partitions/check.c return (unsigned char *)page_address(page) + ((n & ((1 << (PAGE_CACHE_SHIFT - 9)) - 1)) << 9); page 574 fs/partitions/check.c page_cache_release(page); page 154 fs/pipe.c struct page *page = buf->page; page 161 fs/pipe.c if (page_count(page) == 1 && !pipe->tmp_page) page 162 fs/pipe.c pipe->tmp_page = page; page 164 fs/pipe.c page_cache_release(page); page 186 fs/pipe.c return kmap_atomic(buf->page, KM_USER0); page 189 fs/pipe.c return kmap(buf->page); page 208 fs/pipe.c kunmap(buf->page); page 226 fs/pipe.c struct page *page = buf->page; page 233 fs/pipe.c if (page_count(page) == 1) { page 234 fs/pipe.c lock_page(page); page 253 fs/pipe.c page_cache_get(buf->page); page 476 fs/pipe.c struct page *page = pipe->tmp_page; page 480 fs/pipe.c if (!page) { page 481 fs/pipe.c page = alloc_page(GFP_HIGHUSER); page 482 fs/pipe.c if (unlikely(!page)) { page 486 fs/pipe.c pipe->tmp_page = page; page 501 fs/pipe.c src = kmap_atomic(page, KM_USER0); page 503 fs/pipe.c src = kmap(page); page 510 fs/pipe.c kunmap(page); page 524 fs/pipe.c buf->page = page; page 697 fs/proc/base.c unsigned long page; page 709 fs/proc/base.c if (!(page = __get_free_page(GFP_TEMPORARY))) page 712 fs/proc/base.c length = PROC_I(inode)->op.proc_read(task, (char*)page); page 715 fs/proc/base.c length = simple_read_from_buffer(buf, count, ppos, (char *)page, length); page 716 fs/proc/base.c free_page(page); page 776 fs/proc/base.c char *page; page 788 fs/proc/base.c page = (char *)__get_free_page(GFP_TEMPORARY); page 789 fs/proc/base.c if (!page) page 809 fs/proc/base.c retval = access_process_vm(task, src, page, this_len, 0); page 816 fs/proc/base.c if (copy_to_user(buf, page, retval)) { page 831 fs/proc/base.c free_page((unsigned long) page); page 846 fs/proc/base.c char *page; page 858 fs/proc/base.c page = (char *)__get_free_page(GFP_TEMPORARY); page 859 fs/proc/base.c if (!page) page 867 fs/proc/base.c if (copy_from_user(page, buf, this_len)) { page 871 fs/proc/base.c retval = access_process_vm(task, dst, page, this_len, 1); page 883 fs/proc/base.c free_page((unsigned long) page); page 918 fs/proc/base.c char *page; page 930 fs/proc/base.c page = (char *)__get_free_page(GFP_TEMPORARY); page 931 fs/proc/base.c if (!page) page 952 fs/proc/base.c page, this_len, 0); page 959 fs/proc/base.c if (copy_to_user(buf, page, retval)) { page 973 fs/proc/base.c free_page((unsigned long) page); page 1061 fs/proc/base.c char *page, *tmp; page 1078 fs/proc/base.c page = (char*)__get_free_page(GFP_TEMPORARY); page 1079 fs/proc/base.c if (!page) page 1082 fs/proc/base.c if (copy_from_user(page, buf, count)) page 1085 fs/proc/base.c page[count] = '\0'; page 1086 fs/proc/base.c loginuid = simple_strtoul(page, &tmp, 10); page 1087 fs/proc/base.c if (tmp == page) { page 1097 fs/proc/base.c free_page((unsigned long) page); page 2084 fs/proc/base.c char *page; page 2100 fs/proc/base.c page = (char*)__get_free_page(GFP_TEMPORARY); page 2101 fs/proc/base.c if (!page) page 2105 fs/proc/base.c if (copy_from_user(page, buf, count)) page 2110 fs/proc/base.c (void*)page, count); page 2112 fs/proc/base.c free_page((unsigned long) page); page 45 fs/proc/generic.c char *page; page 65 fs/proc/generic.c if (!(page = (char*) __get_free_page(GFP_TEMPORARY))) page 120 fs/proc/generic.c n = dp->read_proc(page, &start, *ppos, page 144 fs/proc/generic.c start = page + *ppos; page 145 fs/proc/generic.c } else if (start < page) { page 160 fs/proc/generic.c unsigned long startoff = (unsigned long)(start - page); page 170 fs/proc/generic.c n -= copy_to_user(buf, start < page ? page : start, n); page 177 fs/proc/generic.c *ppos += start < page ? (unsigned long)start : n; page 182 fs/proc/generic.c free_page((unsigned long) page); page 41 fs/proc/proc_devtree.c memcpy(page, (char *)pp->value + off, n); page 42 fs/proc/proc_devtree.c *start = page; page 76 fs/proc/proc_misc.c *start = page + off; page 97 fs/proc/proc_misc.c len = sprintf(page,"%d.%02d %d.%02d %d.%02d %ld/%d %d\n", page 103 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 117 fs/proc/proc_misc.c len = sprintf(page,"%lu.%02lu %lu.%02lu\n", page 123 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 161 fs/proc/proc_misc.c len = sprintf(page, page 233 fs/proc/proc_misc.c len += hugetlb_report_meminfo(page + len); page 235 fs/proc/proc_misc.c len += arch_report_meminfo(page + len); page 237 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 283 fs/proc/proc_misc.c len = snprintf(page, PAGE_SIZE, linux_proc_banner, page 287 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 377 fs/proc/proc_misc.c int len = get_hardware_list(page); page 378 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 386 fs/proc/proc_misc.c int len = get_stram_list(page); page 387 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 673 fs/proc/proc_misc.c int len = get_filesystem_list(page); page 674 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 682 fs/proc/proc_misc.c len = sprintf(page, "%s\n", saved_command_line); page 683 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 703 fs/proc/proc_misc.c int len = get_exec_domain_list(page); page 704 fs/proc/proc_misc.c return proc_calc_metrics(page, start, off, count, eof, len); page 741 fs/proc/proc_misc.c struct page *ppage; page 807 fs/proc/proc_misc.c struct page *ppage; page 321 fs/proc/task_mmu.c struct page *page; page 338 fs/proc/task_mmu.c page = vm_normal_page(vma, addr, ptent); page 339 fs/proc/task_mmu.c if (!page) page 343 fs/proc/task_mmu.c if (pte_young(ptent) || PageReferenced(page)) page 345 fs/proc/task_mmu.c mapcount = page_mapcount(page); page 433 fs/proc/task_mmu.c struct page *page; page 441 fs/proc/task_mmu.c page = vm_normal_page(vma, addr, ptent); page 442 fs/proc/task_mmu.c if (!page) page 447 fs/proc/task_mmu.c ClearPageReferenced(page); page 630 fs/proc/task_mmu.c struct page **pages, *page; page 666 fs/proc/task_mmu.c pages = kcalloc(pagecount, sizeof(struct page *), GFP_KERNEL); page 718 fs/proc/task_mmu.c page = pages[pagecount-1]; page 719 fs/proc/task_mmu.c if (!PageReserved(page)) page 720 fs/proc/task_mmu.c SetPageDirty(page); page 721 fs/proc/task_mmu.c page_cache_release(page); page 435 fs/qnx4/inode.c return block_write_full_page(page,qnx4_get_block, wbc); page 440 fs/qnx4/inode.c return block_read_full_page(page,qnx4_get_block); page 65 fs/ramfs/file-nommu.c struct page *pages; page 108 fs/ramfs/file-nommu.c struct page *page = pages + loop; page 110 fs/ramfs/file-nommu.c ret = add_to_page_cache(page, inode->i_mapping, loop, GFP_KERNEL); page 114 fs/ramfs/file-nommu.c if (!pagevec_add(&lru_pvec, page)) page 117 fs/ramfs/file-nommu.c unlock_page(page); page 243 fs/ramfs/file-nommu.c struct page **pages = NULL, **ptr, *page; page 263 fs/ramfs/file-nommu.c pages = kzalloc(lpages * sizeof(struct page *), GFP_KERNEL); page 273 fs/ramfs/file-nommu.c page = *ptr++; page 274 fs/ramfs/file-nommu.c page++; page 276 fs/ramfs/file-nommu.c if (*ptr++ != page++) page 182 fs/reiserfs/file.c for (bh = head = page_buffers(page), block_start = 0; page 203 fs/reiserfs/file.c (new || page->index >= i_size_index)) { page 221 fs/reiserfs/file.c SetPageUptodate(page); page 162 fs/reiserfs/inode.c if (page && page_has_buffers(page)) { page 163 fs/reiserfs/inode.c head = page_buffers(page); page 520 fs/reiserfs/inode.c struct page *tail_page; page 521 fs/reiserfs/inode.c struct page *hole_page = bh_result->b_page; page 2012 fs/reiserfs/inode.c struct page *page; page 2023 fs/reiserfs/inode.c page = grab_cache_page(p_s_inode->i_mapping, index); page 2025 fs/reiserfs/inode.c if (!page) { page 2031 fs/reiserfs/inode.c error = block_prepare_write(page, start, offset, page 2036 fs/reiserfs/inode.c head = page_buffers(page); page 2060 fs/reiserfs/inode.c *page_result = page; page 2066 fs/reiserfs/inode.c unlock_page(page); page 2067 fs/reiserfs/inode.c page_cache_release(page); page 2084 fs/reiserfs/inode.c struct page *page = NULL; page 2092 fs/reiserfs/inode.c if ((error = grab_tail_page(p_s_inode, &page, &bh))) { page 2100 fs/reiserfs/inode.c page = NULL; page 2125 fs/reiserfs/inode.c err2 = reiserfs_do_truncate(&th, p_s_inode, page, update_timestamps); page 2143 fs/reiserfs/inode.c if (page) { page 2148 fs/reiserfs/inode.c zero_user(page, offset, length); page 2153 fs/reiserfs/inode.c unlock_page(page); page 2154 fs/reiserfs/inode.c page_cache_release(page); page 2160 fs/reiserfs/inode.c if (page) { page 2161 fs/reiserfs/inode.c unlock_page(page); page 2162 fs/reiserfs/inode.c page_cache_release(page); page 2328 fs/reiserfs/inode.c struct inode *inode = page->mapping->host; page 2336 fs/reiserfs/inode.c int checked = PageChecked(page); page 2344 fs/reiserfs/inode.c redirty_page_for_writepage(wbc, page); page 2345 fs/reiserfs/inode.c unlock_page(page); page 2354 fs/reiserfs/inode.c if (!page_has_buffers(page)) { page 2355 fs/reiserfs/inode.c create_empty_buffers(page, s->s_blocksize, page 2358 fs/reiserfs/inode.c head = page_buffers(page); page 2363 fs/reiserfs/inode.c if (page->index >= end_index) { page 2368 fs/reiserfs/inode.c if (page->index >= end_index + 1 || !last_offset) { page 2369 fs/reiserfs/inode.c unlock_page(page); page 2372 fs/reiserfs/inode.c zero_user_segment(page, last_offset, PAGE_CACHE_SIZE); page 2375 fs/reiserfs/inode.c block = page->index << (PAGE_CACHE_SHIFT - s->s_blocksize_bits); page 2410 fs/reiserfs/inode.c ClearPageChecked(page); page 2439 fs/reiserfs/inode.c redirty_page_for_writepage(wbc, page); page 2456 fs/reiserfs/inode.c BUG_ON(PageWriteback(page)); page 2457 fs/reiserfs/inode.c set_page_writeback(page); page 2458 fs/reiserfs/inode.c unlock_page(page); page 2493 fs/reiserfs/inode.c SetPageUptodate(page); page 2494 fs/reiserfs/inode.c end_page_writeback(page); page 2503 fs/reiserfs/inode.c ClearPageUptodate(page); page 2519 fs/reiserfs/inode.c SetPageError(page); page 2520 fs/reiserfs/inode.c BUG_ON(PageWriteback(page)); page 2521 fs/reiserfs/inode.c set_page_writeback(page); page 2522 fs/reiserfs/inode.c unlock_page(page); page 2538 fs/reiserfs/inode.c return block_read_full_page(page, reiserfs_get_block); page 2543 fs/reiserfs/inode.c struct inode *inode = page->mapping->host; page 2545 fs/reiserfs/inode.c return reiserfs_write_full_page(page, wbc); page 2554 fs/reiserfs/inode.c struct page *page; page 2568 fs/reiserfs/inode.c page = __grab_cache_page(mapping, index); page 2569 fs/reiserfs/inode.c if (!page) page 2571 fs/reiserfs/inode.c *pagep = page; page 2574 fs/reiserfs/inode.c fix_tail_page_for_writing(page); page 2612 fs/reiserfs/inode.c unlock_page(page); page 2613 fs/reiserfs/inode.c page_cache_release(page); page 2621 fs/reiserfs/inode.c struct inode *inode = page->mapping->host; page 2626 fs/reiserfs/inode.c fix_tail_page_for_writing(page); page 2637 fs/reiserfs/inode.c ret = block_prepare_write(page, from, to, reiserfs_get_block); page 2676 fs/reiserfs/inode.c struct inode *inode = page->mapping->host; page 2693 fs/reiserfs/inode.c if (!PageUptodate(page)) page 2696 fs/reiserfs/inode.c page_zero_new_buffers(page, start + copied, start + len); page 2698 fs/reiserfs/inode.c flush_dcache_page(page); page 2700 fs/reiserfs/inode.c reiserfs_commit_page(inode, page, start, start + copied); page 2750 fs/reiserfs/inode.c unlock_page(page); page 2751 fs/reiserfs/inode.c page_cache_release(page); page 2769 fs/reiserfs/inode.c struct inode *inode = page->mapping->host; page 2770 fs/reiserfs/inode.c loff_t pos = ((loff_t) page->index << PAGE_CACHE_SHIFT) + to; page 2779 fs/reiserfs/inode.c reiserfs_commit_page(inode, page, from, to); page 2948 fs/reiserfs/inode.c struct inode *inode = page->mapping->host; page 2952 fs/reiserfs/inode.c BUG_ON(!PageLocked(page)); page 2955 fs/reiserfs/inode.c ClearPageChecked(page); page 2957 fs/reiserfs/inode.c if (!page_has_buffers(page)) page 2960 fs/reiserfs/inode.c head = page_buffers(page); page 2985 fs/reiserfs/inode.c ret = try_to_release_page(page, 0); page 2994 fs/reiserfs/inode.c struct inode *inode = page->mapping->host; page 2996 fs/reiserfs/inode.c SetPageChecked(page); page 2997 fs/reiserfs/inode.c return __set_page_dirty_nobuffers(page); page 2999 fs/reiserfs/inode.c return __set_page_dirty_buffers(page); page 3013 fs/reiserfs/inode.c struct inode *inode = page->mapping->host; page 3019 fs/reiserfs/inode.c WARN_ON(PageChecked(page)); page 3021 fs/reiserfs/inode.c head = page_buffers(page); page 3035 fs/reiserfs/inode.c ret = try_to_free_buffers(page); page 164 fs/reiserfs/ioctl.c struct page *page; page 197 fs/reiserfs/ioctl.c page = grab_cache_page(mapping, index); page 199 fs/reiserfs/ioctl.c if (!page) { page 202 fs/reiserfs/ioctl.c retval = reiserfs_prepare_write(NULL, page, write_from, write_from); page 207 fs/reiserfs/ioctl.c flush_dcache_page(page); page 208 fs/reiserfs/ioctl.c retval = reiserfs_commit_write(NULL, page, write_from, write_from); page 212 fs/reiserfs/ioctl.c unlock_page(page); page 213 fs/reiserfs/ioctl.c page_cache_release(page); page 629 fs/reiserfs/journal.c struct page *page = bh->b_page; page 630 fs/reiserfs/journal.c if (!page->mapping && trylock_page(page)) { page 631 fs/reiserfs/journal.c page_cache_get(page); page 633 fs/reiserfs/journal.c if (!page->mapping) page 634 fs/reiserfs/journal.c try_to_free_buffers(page); page 635 fs/reiserfs/journal.c unlock_page(page); page 636 fs/reiserfs/journal.c page_cache_release(page); page 4117 fs/reiserfs/journal.c struct page *page; page 4125 fs/reiserfs/journal.c page = cn->bh->b_page; page 4126 fs/reiserfs/journal.c addr = kmap(page); page 4130 fs/reiserfs/journal.c kunmap(page); page 1439 fs/reiserfs/stree.c if (page) { page 1440 fs/reiserfs/stree.c if (page_has_buffers(page)) { page 1443 fs/reiserfs/stree.c head = page_buffers(page); page 1483 fs/reiserfs/stree.c !page || (REISERFS_I(p_s_inode)->i_flags & i_nopack_mask)) { page 1493 fs/reiserfs/stree.c return indirect2direct(th, p_s_inode, page, p_s_path, p_s_item_key, page 1588 fs/reiserfs/stree.c maybe_indirect_to_direct(th, p_s_inode, page, page 1729 fs/reiserfs/stree.c unmap_buffers(page, tail_pos); page 1851 fs/reiserfs/stree.c p_s_inode, page, n_new_file_size); page 211 fs/reiserfs/tail_conversion.c tail = (char *)kmap(page); /* this can schedule */ page 257 fs/reiserfs/tail_conversion.c kunmap(page); page 260 fs/reiserfs/tail_conversion.c kunmap(page); page 374 fs/reiserfs/xattr.c kunmap(page); page 375 fs/reiserfs/xattr.c page_cache_release(page); page 378 fs/reiserfs/xattr.c static struct page *reiserfs_get_page(struct inode *dir, unsigned long n) page 381 fs/reiserfs/xattr.c struct page *page; page 385 fs/reiserfs/xattr.c page = read_mapping_page(mapping, n, NULL); page 386 fs/reiserfs/xattr.c if (!IS_ERR(page)) { page 387 fs/reiserfs/xattr.c kmap(page); page 388 fs/reiserfs/xattr.c if (PageError(page)) page 391 fs/reiserfs/xattr.c return page; page 394 fs/reiserfs/xattr.c reiserfs_put_page(page); page 420 fs/reiserfs/xattr.c struct page *page; page 476 fs/reiserfs/xattr.c page = reiserfs_get_page(xinode, file_pos >> PAGE_CACHE_SHIFT); page 477 fs/reiserfs/xattr.c if (IS_ERR(page)) { page 478 fs/reiserfs/xattr.c err = PTR_ERR(page); page 482 fs/reiserfs/xattr.c lock_page(page); page 483 fs/reiserfs/xattr.c data = page_address(page); page 495 fs/reiserfs/xattr.c err = reiserfs_prepare_write(NULL, page, page_offset, page 500 fs/reiserfs/xattr.c err = reiserfs_commit_write(NULL, page, page_offset, page 504 fs/reiserfs/xattr.c unlock_page(page); page 505 fs/reiserfs/xattr.c reiserfs_put_page(page); page 542 fs/reiserfs/xattr.c struct page *page; page 584 fs/reiserfs/xattr.c page = reiserfs_get_page(xinode, file_pos >> PAGE_CACHE_SHIFT); page 585 fs/reiserfs/xattr.c if (IS_ERR(page)) { page 586 fs/reiserfs/xattr.c err = PTR_ERR(page); page 590 fs/reiserfs/xattr.c lock_page(page); page 591 fs/reiserfs/xattr.c data = page_address(page); page 599 fs/reiserfs/xattr.c unlock_page(page); page 600 fs/reiserfs/xattr.c reiserfs_put_page(page); page 611 fs/reiserfs/xattr.c unlock_page(page); page 612 fs/reiserfs/xattr.c reiserfs_put_page(page); page 420 fs/romfs/inode.c struct inode *inode = page->mapping->host; page 426 fs/romfs/inode.c page_cache_get(page); page 428 fs/romfs/inode.c buf = kmap(page); page 433 fs/romfs/inode.c offset = page_offset(page); page 446 fs/romfs/inode.c SetPageError(page); page 456 fs/romfs/inode.c SetPageUptodate(page); page 457 fs/romfs/inode.c flush_dcache_page(page); page 459 fs/romfs/inode.c unlock_page(page); page 461 fs/romfs/inode.c kunmap(page); page 463 fs/romfs/inode.c page_cache_release(page); page 34 fs/smbfs/cache.c struct page *page = NULL; page 36 fs/smbfs/cache.c page = grab_cache_page(&dir->i_data, 0); page 37 fs/smbfs/cache.c if (!page) page 40 fs/smbfs/cache.c if (!PageUptodate(page)) page 43 fs/smbfs/cache.c cache = kmap(page); page 46 fs/smbfs/cache.c kunmap(page); page 47 fs/smbfs/cache.c SetPageUptodate(page); page 49 fs/smbfs/cache.c unlock_page(page); page 50 fs/smbfs/cache.c page_cache_release(page); page 172 fs/smbfs/cache.c if (ctl.page) { page 173 fs/smbfs/cache.c kunmap(ctl.page); page 174 fs/smbfs/cache.c SetPageUptodate(ctl.page); page 175 fs/smbfs/cache.c unlock_page(ctl.page); page 176 fs/smbfs/cache.c page_cache_release(ctl.page); page 181 fs/smbfs/cache.c ctl.page = grab_cache_page(&inode->i_data, ctl.ofs); page 182 fs/smbfs/cache.c if (ctl.page) page 183 fs/smbfs/cache.c ctl.cache = kmap(ctl.page); page 87 fs/smbfs/dir.c struct page *page = NULL; page 90 fs/smbfs/dir.c ctl.page = NULL; page 120 fs/smbfs/dir.c page = grab_cache_page(&dir->i_data, 0); page 121 fs/smbfs/dir.c if (!page) page 124 fs/smbfs/dir.c ctl.cache = cache = kmap(page); page 127 fs/smbfs/dir.c if (!PageUptodate(page) || !ctl.head.eof) { page 129 fs/smbfs/dir.c DENTRY_PATH(dentry), PageUptodate(page),ctl.head.eof); page 157 fs/smbfs/dir.c ctl.page = find_lock_page(&dir->i_data, ctl.ofs); page 158 fs/smbfs/dir.c if (!ctl.page) page 160 fs/smbfs/dir.c ctl.cache = kmap(ctl.page); page 161 fs/smbfs/dir.c if (!PageUptodate(ctl.page)) page 184 fs/smbfs/dir.c if (ctl.page) { page 185 fs/smbfs/dir.c kunmap(ctl.page); page 186 fs/smbfs/dir.c SetPageUptodate(ctl.page); page 187 fs/smbfs/dir.c unlock_page(ctl.page); page 188 fs/smbfs/dir.c page_cache_release(ctl.page); page 189 fs/smbfs/dir.c ctl.page = NULL; page 195 fs/smbfs/dir.c if (ctl.page) { page 196 fs/smbfs/dir.c kunmap(ctl.page); page 197 fs/smbfs/dir.c unlock_page(ctl.page); page 198 fs/smbfs/dir.c page_cache_release(ctl.page); page 199 fs/smbfs/dir.c ctl.page = NULL; page 213 fs/smbfs/dir.c if (result == -ERESTARTSYS && page) page 214 fs/smbfs/dir.c ClearPageUptodate(page); page 220 fs/smbfs/dir.c if (page) { page 222 fs/smbfs/dir.c kunmap(page); page 224 fs/smbfs/dir.c SetPageUptodate(page); page 225 fs/smbfs/dir.c unlock_page(page); page 226 fs/smbfs/dir.c page_cache_release(page); page 228 fs/smbfs/dir.c if (ctl.page) { page 229 fs/smbfs/dir.c kunmap(ctl.page); page 230 fs/smbfs/dir.c SetPageUptodate(ctl.page); page 231 fs/smbfs/dir.c unlock_page(ctl.page); page 232 fs/smbfs/dir.c page_cache_release(ctl.page); page 57 fs/smbfs/file.c char *buffer = kmap(page); page 58 fs/smbfs/file.c loff_t offset = (loff_t)page->index << PAGE_CACHE_SHIFT; page 89 fs/smbfs/file.c flush_dcache_page(page); page 90 fs/smbfs/file.c SetPageUptodate(page); page 94 fs/smbfs/file.c kunmap(page); page 95 fs/smbfs/file.c unlock_page(page); page 108 fs/smbfs/file.c page_cache_get(page); page 109 fs/smbfs/file.c error = smb_readpage_sync(dentry, page); page 110 fs/smbfs/file.c page_cache_release(page); page 123 fs/smbfs/file.c char *buffer = kmap(page) + pageoffset; page 128 fs/smbfs/file.c offset = ((loff_t)page->index << PAGE_CACHE_SHIFT) + pageoffset; page 163 fs/smbfs/file.c kunmap(page); page 176 fs/smbfs/file.c struct address_space *mapping = page->mapping; page 189 fs/smbfs/file.c if (page->index < end_index) page 194 fs/smbfs/file.c if (page->index >= end_index+1 || !offset) page 197 fs/smbfs/file.c page_cache_get(page); page 198 fs/smbfs/file.c err = smb_writepage_sync(inode, page, 0, offset); page 199 fs/smbfs/file.c SetPageUptodate(page); page 200 fs/smbfs/file.c unlock_page(page); page 201 fs/smbfs/file.c page_cache_release(page); page 212 fs/smbfs/file.c ((unsigned long long)page->index << PAGE_CACHE_SHIFT) + offset); page 214 fs/smbfs/file.c return smb_writepage_sync(dentry->d_inode, page, offset, count); page 314 fs/smbfs/file.c status = smb_updatepage(file, page, offset, copied); page 318 fs/smbfs/file.c if (!PageUptodate(page) && copied == PAGE_CACHE_SIZE) page 319 fs/smbfs/file.c SetPageUptodate(page); page 323 fs/smbfs/file.c unlock_page(page); page 324 fs/smbfs/file.c page_cache_release(page); page 42 fs/splice.c struct page *page = buf->page; page 45 fs/splice.c lock_page(page); page 47 fs/splice.c mapping = page_mapping(page); page 49 fs/splice.c WARN_ON(!PageUptodate(page)); page 59 fs/splice.c wait_on_page_writeback(page); page 61 fs/splice.c if (PagePrivate(page) && !try_to_release_page(page, GFP_KERNEL)) page 68 fs/splice.c if (remove_mapping(mapping, page)) { page 79 fs/splice.c unlock_page(page); page 86 fs/splice.c page_cache_release(buf->page); page 97 fs/splice.c struct page *page = buf->page; page 100 fs/splice.c if (!PageUptodate(page)) { page 101 fs/splice.c lock_page(page); page 107 fs/splice.c if (!page->mapping) { page 115 fs/splice.c if (!PageUptodate(page)) { page 123 fs/splice.c unlock_page(page); page 128 fs/splice.c unlock_page(page); page 198 fs/splice.c buf->page = spd->pages[page_nr]; page 275 fs/splice.c struct page *pages[PIPE_BUFFERS]; page 277 fs/splice.c struct page *page; page 314 fs/splice.c page = find_get_page(mapping, index); page 315 fs/splice.c if (!page) { page 319 fs/splice.c page = page_cache_alloc_cold(mapping); page 320 fs/splice.c if (!page) page 323 fs/splice.c error = add_to_page_cache_lru(page, mapping, index, page 326 fs/splice.c page_cache_release(page); page 335 fs/splice.c unlock_page(page); page 338 fs/splice.c pages[spd.nr_pages++] = page; page 359 fs/splice.c page = pages[page_nr]; page 361 fs/splice.c if (PageReadahead(page)) page 363 fs/splice.c page, index, req_pages - page_nr); page 368 fs/splice.c if (!PageUptodate(page)) { page 374 fs/splice.c if (!trylock_page(page)) { page 379 fs/splice.c lock_page(page); page 387 fs/splice.c if (!page->mapping) { page 388 fs/splice.c unlock_page(page); page 389 fs/splice.c page = find_or_create_page(mapping, index, page 392 fs/splice.c if (!page) { page 397 fs/splice.c pages[page_nr] = page; page 402 fs/splice.c if (PageUptodate(page)) { page 403 fs/splice.c unlock_page(page); page 410 fs/splice.c error = mapping->a_ops->readpage(in, page); page 529 fs/splice.c ret = file->f_op->sendpage(file, buf->page, buf->offset, page 562 fs/splice.c struct page *page; page 580 fs/splice.c AOP_FLAG_UNINTERRUPTIBLE, &page, &fsdata); page 584 fs/splice.c if (buf->page != page) { page 589 fs/splice.c char *dst = kmap_atomic(page, KM_USER1); page 592 fs/splice.c flush_dcache_page(page); page 597 fs/splice.c page, fsdata); page 1399 fs/splice.c struct page *pages[PIPE_BUFFERS]; page 47 fs/sysfs/file.c char * page; page 74 fs/sysfs/file.c if (!buffer->page) page 75 fs/sysfs/file.c buffer->page = (char *) get_zeroed_page(GFP_KERNEL); page 76 fs/sysfs/file.c if (!buffer->page) page 84 fs/sysfs/file.c count = ops->show(kobj, attr_sd->s_attr.attr, buffer->page); page 139 fs/sysfs/file.c __func__, count, *ppos, buffer->page); page 140 fs/sysfs/file.c retval = simple_read_from_buffer(buf, count, ppos, buffer->page, page 162 fs/sysfs/file.c if (!buffer->page) page 163 fs/sysfs/file.c buffer->page = (char *)get_zeroed_page(GFP_KERNEL); page 164 fs/sysfs/file.c if (!buffer->page) page 169 fs/sysfs/file.c error = copy_from_user(buffer->page,buf,count); page 173 fs/sysfs/file.c buffer->page[count] = 0; page 201 fs/sysfs/file.c rc = ops->store(kobj, attr_sd->s_attr.attr, buffer->page, count); page 399 fs/sysfs/file.c if (buffer->page) page 400 fs/sysfs/file.c free_page((unsigned long)buffer->page); page 194 fs/sysfs/symlink.c unsigned long page = get_zeroed_page(GFP_KERNEL); page 195 fs/sysfs/symlink.c if (page) page 196 fs/sysfs/symlink.c error = sysfs_getlink(dentry, (char *) page); page 197 fs/sysfs/symlink.c nd_set_link(nd, error ? ERR_PTR(error) : (char *)page); page 203 fs/sysfs/symlink.c char *page = nd_get_link(nd); page 204 fs/sysfs/symlink.c if (!IS_ERR(page)) page 205 fs/sysfs/symlink.c free_page((unsigned long)page); page 32 fs/sysv/dir.c kunmap(page); page 33 fs/sysv/dir.c page_cache_release(page); page 43 fs/sysv/dir.c struct address_space *mapping = page->mapping; page 47 fs/sysv/dir.c block_write_end(NULL, mapping, pos, len, len, page, NULL); page 53 fs/sysv/dir.c err = write_one_page(page, 1); page 55 fs/sysv/dir.c unlock_page(page); page 59 fs/sysv/dir.c static struct page * dir_get_page(struct inode *dir, unsigned long n) page 62 fs/sysv/dir.c struct page *page = read_mapping_page(mapping, n, NULL); page 63 fs/sysv/dir.c if (!IS_ERR(page)) page 64 fs/sysv/dir.c kmap(page); page 65 fs/sysv/dir.c return page; page 86 fs/sysv/dir.c struct page *page = dir_get_page(inode, n); page 88 fs/sysv/dir.c if (IS_ERR(page)) page 90 fs/sysv/dir.c kaddr = (char *)page_address(page); page 107 fs/sysv/dir.c dir_put_page(page); page 111 fs/sysv/dir.c dir_put_page(page); page 146 fs/sysv/dir.c struct page *page = NULL; page 158 fs/sysv/dir.c page = dir_get_page(dir, n); page 159 fs/sysv/dir.c if (!IS_ERR(page)) { page 160 fs/sysv/dir.c kaddr = (char*)page_address(page); page 171 fs/sysv/dir.c dir_put_page(page); page 181 fs/sysv/dir.c *res_page = page; page 190 fs/sysv/dir.c struct page *page = NULL; page 200 fs/sysv/dir.c page = dir_get_page(dir, n); page 201 fs/sysv/dir.c err = PTR_ERR(page); page 202 fs/sysv/dir.c if (IS_ERR(page)) page 204 fs/sysv/dir.c kaddr = (char*)page_address(page); page 215 fs/sysv/dir.c dir_put_page(page); page 221 fs/sysv/dir.c pos = page_offset(page) + page 222 fs/sysv/dir.c (char*)de - (char*)page_address(page); page 223 fs/sysv/dir.c lock_page(page); page 224 fs/sysv/dir.c err = __sysv_write_begin(NULL, page->mapping, pos, SYSV_DIRSIZE, page 225 fs/sysv/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 231 fs/sysv/dir.c err = dir_commit_chunk(page, pos, SYSV_DIRSIZE); page 235 fs/sysv/dir.c dir_put_page(page); page 239 fs/sysv/dir.c unlock_page(page); page 245 fs/sysv/dir.c struct address_space *mapping = page->mapping; page 247 fs/sysv/dir.c char *kaddr = (char*)page_address(page); page 248 fs/sysv/dir.c loff_t pos = page_offset(page) + (char *)de - kaddr; page 251 fs/sysv/dir.c lock_page(page); page 253 fs/sysv/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 256 fs/sysv/dir.c err = dir_commit_chunk(page, pos, SYSV_DIRSIZE); page 257 fs/sysv/dir.c dir_put_page(page); page 266 fs/sysv/dir.c struct page *page = grab_cache_page(mapping, 0); page 271 fs/sysv/dir.c if (!page) page 274 fs/sysv/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 276 fs/sysv/dir.c unlock_page(page); page 279 fs/sysv/dir.c kmap(page); page 281 fs/sysv/dir.c base = (char*)page_address(page); page 291 fs/sysv/dir.c kunmap(page); page 292 fs/sysv/dir.c err = dir_commit_chunk(page, 0, 2 * SYSV_DIRSIZE); page 294 fs/sysv/dir.c page_cache_release(page); page 304 fs/sysv/dir.c struct page *page = NULL; page 310 fs/sysv/dir.c page = dir_get_page(inode, i); page 312 fs/sysv/dir.c if (IS_ERR(page)) page 315 fs/sysv/dir.c kaddr = (char *)page_address(page); page 334 fs/sysv/dir.c dir_put_page(page); page 339 fs/sysv/dir.c dir_put_page(page); page 347 fs/sysv/dir.c struct address_space *mapping = page->mapping; page 349 fs/sysv/dir.c loff_t pos = page_offset(page) + page 350 fs/sysv/dir.c (char *)de-(char*)page_address(page); page 353 fs/sysv/dir.c lock_page(page); page 355 fs/sysv/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 358 fs/sysv/dir.c err = dir_commit_chunk(page, pos, SYSV_DIRSIZE); page 359 fs/sysv/dir.c dir_put_page(page); page 366 fs/sysv/dir.c struct page *page = dir_get_page(dir, 0); page 369 fs/sysv/dir.c if (!IS_ERR(page)) { page 370 fs/sysv/dir.c de = (struct sysv_dir_entry*) page_address(page) + 1; page 371 fs/sysv/dir.c *p = page; page 378 fs/sysv/dir.c struct page *page; page 379 fs/sysv/dir.c struct sysv_dir_entry *de = sysv_find_entry (dentry, &page); page 384 fs/sysv/dir.c dir_put_page(page); page 454 fs/sysv/itree.c return block_write_full_page(page,get_block,wbc); page 459 fs/sysv/itree.c return block_read_full_page(page,get_block); page 176 fs/sysv/namei.c struct page * page; page 180 fs/sysv/namei.c de = sysv_find_entry(dentry, &page); page 184 fs/sysv/namei.c err = sysv_delete_entry (de, page); page 219 fs/sysv/namei.c struct page * dir_page = NULL; page 221 fs/sysv/namei.c struct page * old_page; page 237 fs/sysv/namei.c struct page * new_page; page 111 fs/ubifs/file.c struct inode *inode = page->mapping->host; page 115 fs/ubifs/file.c inode->i_ino, page->index, i_size, page->flags); page 116 fs/ubifs/file.c ubifs_assert(!PageChecked(page)); page 117 fs/ubifs/file.c ubifs_assert(!PagePrivate(page)); page 119 fs/ubifs/file.c addr = kmap(page); page 121 fs/ubifs/file.c block = page->index << UBIFS_BLOCKS_PER_PAGE_SHIFT; page 125 fs/ubifs/file.c SetPageChecked(page); page 160 fs/ubifs/file.c SetPageChecked(page); page 165 fs/ubifs/file.c page->index, inode->i_ino, err); page 172 fs/ubifs/file.c SetPageUptodate(page); page 173 fs/ubifs/file.c ClearPageError(page); page 174 fs/ubifs/file.c flush_dcache_page(page); page 175 fs/ubifs/file.c kunmap(page); page 180 fs/ubifs/file.c ClearPageUptodate(page); page 181 fs/ubifs/file.c SetPageError(page); page 182 fs/ubifs/file.c flush_dcache_page(page); page 183 fs/ubifs/file.c kunmap(page); page 223 fs/ubifs/file.c struct page *page; page 244 fs/ubifs/file.c page = __grab_cache_page(mapping, index); page 245 fs/ubifs/file.c if (unlikely(!page)) { page 250 fs/ubifs/file.c if (!PageUptodate(page)) { page 252 fs/ubifs/file.c SetPageChecked(page); page 254 fs/ubifs/file.c err = do_readpage(page); page 256 fs/ubifs/file.c unlock_page(page); page 257 fs/ubifs/file.c page_cache_release(page); page 262 fs/ubifs/file.c SetPageUptodate(page); page 263 fs/ubifs/file.c ClearPageError(page); page 266 fs/ubifs/file.c if (PagePrivate(page)) page 278 fs/ubifs/file.c else if (!PageChecked(page)) page 304 fs/ubifs/file.c *pagep = page; page 326 fs/ubifs/file.c if (PagePrivate(page)) { page 353 fs/ubifs/file.c if (PageChecked(page)) page 426 fs/ubifs/file.c struct page *page; page 435 fs/ubifs/file.c page = __grab_cache_page(mapping, index); page 436 fs/ubifs/file.c if (unlikely(!page)) page 439 fs/ubifs/file.c if (!PageUptodate(page)) { page 449 fs/ubifs/file.c SetPageChecked(page); page 451 fs/ubifs/file.c err = do_readpage(page); page 453 fs/ubifs/file.c unlock_page(page); page 454 fs/ubifs/file.c page_cache_release(page); page 459 fs/ubifs/file.c SetPageUptodate(page); page 460 fs/ubifs/file.c ClearPageError(page); page 463 fs/ubifs/file.c err = allocate_budget(c, page, ui, appending); page 477 fs/ubifs/file.c unlock_page(page); page 478 fs/ubifs/file.c page_cache_release(page); page 489 fs/ubifs/file.c *pagep = page; page 512 fs/ubifs/file.c if (!PagePrivate(page)) { page 513 fs/ubifs/file.c if (PageChecked(page)) page 531 fs/ubifs/file.c inode->i_ino, pos, page->index, len, copied, inode->i_size); page 545 fs/ubifs/file.c cancel_budget(c, page, ui, appending); page 551 fs/ubifs/file.c copied = do_readpage(page); page 555 fs/ubifs/file.c if (!PagePrivate(page)) { page 556 fs/ubifs/file.c SetPagePrivate(page); page 558 fs/ubifs/file.c __set_page_dirty_nobuffers(page); page 575 fs/ubifs/file.c unlock_page(page); page 576 fs/ubifs/file.c page_cache_release(page); page 582 fs/ubifs/file.c do_readpage(page); page 583 fs/ubifs/file.c unlock_page(page); page 593 fs/ubifs/file.c struct inode *inode = page->mapping->host; page 598 fs/ubifs/file.c ubifs_assert(page->index <= ui->synced_i_size << PAGE_CACHE_SIZE); page 603 fs/ubifs/file.c set_page_writeback(page); page 605 fs/ubifs/file.c addr = kmap(page); page 606 fs/ubifs/file.c block = page->index << UBIFS_BLOCKS_PER_PAGE_SHIFT; page 621 fs/ubifs/file.c SetPageError(page); page 623 fs/ubifs/file.c page->index, inode->i_ino, err); page 627 fs/ubifs/file.c ubifs_assert(PagePrivate(page)); page 628 fs/ubifs/file.c if (PageChecked(page)) page 634 fs/ubifs/file.c ClearPagePrivate(page); page 635 fs/ubifs/file.c ClearPageChecked(page); page 637 fs/ubifs/file.c kunmap(page); page 638 fs/ubifs/file.c unlock_page(page); page 639 fs/ubifs/file.c end_page_writeback(page); page 687 fs/ubifs/file.c struct inode *inode = page->mapping->host; page 695 fs/ubifs/file.c inode->i_ino, page->index, page->flags); page 696 fs/ubifs/file.c ubifs_assert(PagePrivate(page)); page 699 fs/ubifs/file.c if (page->index > end_index || (page->index == end_index && !len)) { page 709 fs/ubifs/file.c if (page->index < end_index) { page 710 fs/ubifs/file.c if (page->index >= synced_i_size >> PAGE_CACHE_SHIFT) { page 723 fs/ubifs/file.c return do_writepage(page, PAGE_CACHE_SIZE); page 733 fs/ubifs/file.c kaddr = kmap_atomic(page, KM_USER0); page 735 fs/ubifs/file.c flush_dcache_page(page); page 744 fs/ubifs/file.c return do_writepage(page, len); page 747 fs/ubifs/file.c unlock_page(page); page 830 fs/ubifs/file.c struct page *page; page 832 fs/ubifs/file.c page = find_lock_page(inode->i_mapping, index); page 833 fs/ubifs/file.c if (page) { page 834 fs/ubifs/file.c if (PageDirty(page)) { page 843 fs/ubifs/file.c ubifs_assert(PagePrivate(page)); page 845 fs/ubifs/file.c clear_page_dirty_for_io(page); page 849 fs/ubifs/file.c err = do_writepage(page, offset); page 850 fs/ubifs/file.c page_cache_release(page); page 863 fs/ubifs/file.c unlock_page(page); page 864 fs/ubifs/file.c page_cache_release(page); page 977 fs/ubifs/file.c struct inode *inode = page->mapping->host; page 980 fs/ubifs/file.c ubifs_assert(PagePrivate(page)); page 985 fs/ubifs/file.c if (PageChecked(page)) page 991 fs/ubifs/file.c ClearPagePrivate(page); page 992 fs/ubifs/file.c ClearPageChecked(page); page 1114 fs/ubifs/file.c ret = __set_page_dirty_nobuffers(page); page 1129 fs/ubifs/file.c if (PageWriteback(page)) page 1131 fs/ubifs/file.c ubifs_assert(PagePrivate(page)); page 1133 fs/ubifs/file.c ClearPagePrivate(page); page 1134 fs/ubifs/file.c ClearPageChecked(page); page 1150 fs/ubifs/file.c dbg_gen("ino %lu, pg %lu, i_size %lld", inode->i_ino, page->index, page 1191 fs/ubifs/file.c lock_page(page); page 1192 fs/ubifs/file.c if (unlikely(page->mapping != inode->i_mapping || page 1193 fs/ubifs/file.c page_offset(page) > i_size_read(inode))) { page 1199 fs/ubifs/file.c if (PagePrivate(page)) page 1202 fs/ubifs/file.c if (!PageChecked(page)) page 1204 fs/ubifs/file.c SetPagePrivate(page); page 1206 fs/ubifs/file.c __set_page_dirty_nobuffers(page); page 1222 fs/ubifs/file.c unlock_page(page); page 1226 fs/ubifs/file.c unlock_page(page); page 45 fs/udf/file.c struct inode *inode = page->mapping->host; page 49 fs/udf/file.c BUG_ON(!PageLocked(page)); page 51 fs/udf/file.c kaddr = kmap(page); page 54 fs/udf/file.c flush_dcache_page(page); page 55 fs/udf/file.c SetPageUptodate(page); page 56 fs/udf/file.c kunmap(page); page 57 fs/udf/file.c unlock_page(page); page 65 fs/udf/file.c struct inode *inode = page->mapping->host; page 69 fs/udf/file.c BUG_ON(!PageLocked(page)); page 71 fs/udf/file.c kaddr = kmap(page); page 74 fs/udf/file.c SetPageUptodate(page); page 75 fs/udf/file.c kunmap(page); page 76 fs/udf/file.c unlock_page(page); page 91 fs/udf/file.c kaddr = kmap_atomic(page, KM_USER0); page 96 fs/udf/file.c return simple_write_end(file, mapping, pos, len, copied, page, fsdata); page 117 fs/udf/inode.c return block_write_full_page(page, udf_get_block, wbc); page 122 fs/udf/inode.c return block_read_full_page(page, udf_get_block); page 150 fs/udf/inode.c struct page *page; page 170 fs/udf/inode.c page = grab_cache_page(inode->i_mapping, 0); page 171 fs/udf/inode.c BUG_ON(!PageLocked(page)); page 173 fs/udf/inode.c if (!PageUptodate(page)) { page 174 fs/udf/inode.c kaddr = kmap(page); page 179 fs/udf/inode.c flush_dcache_page(page); page 180 fs/udf/inode.c SetPageUptodate(page); page 181 fs/udf/inode.c kunmap(page); page 191 fs/udf/inode.c inode->i_data.a_ops->writepage(page, &udf_wbc); page 192 fs/udf/inode.c page_cache_release(page); page 76 fs/udf/symlink.c struct inode *inode = page->mapping->host; page 80 fs/udf/symlink.c char *p = kmap(page); page 100 fs/udf/symlink.c SetPageUptodate(page); page 101 fs/udf/symlink.c kunmap(page); page 102 fs/udf/symlink.c unlock_page(page); page 107 fs/udf/symlink.c SetPageError(page); page 108 fs/udf/symlink.c kunmap(page); page 109 fs/udf/symlink.c unlock_page(page); page 259 fs/ufs/balloc.c struct page *page; page 276 fs/ufs/balloc.c page = ufs_get_locked_page(mapping, index); page 277 fs/ufs/balloc.c if (!page)/* it was truncated */ page 279 fs/ufs/balloc.c if (IS_ERR(page)) {/* or EIO */ page 286 fs/ufs/balloc.c page = locked_page; page 288 fs/ufs/balloc.c head = page_buffers(page); page 330 fs/ufs/balloc.c ufs_put_locked_page(page); page 45 fs/ufs/dir.c struct address_space *mapping = page->mapping; page 50 fs/ufs/dir.c block_write_end(NULL, mapping, pos, len, len, page, NULL); page 56 fs/ufs/dir.c err = write_one_page(page, 1); page 58 fs/ufs/dir.c unlock_page(page); page 64 fs/ufs/dir.c kunmap(page); page 65 fs/ufs/dir.c page_cache_release(page); page 77 fs/ufs/dir.c struct page *page; page 79 fs/ufs/dir.c de = ufs_find_entry(dir, dentry, &page); page 82 fs/ufs/dir.c ufs_put_page(page); page 92 fs/ufs/dir.c loff_t pos = page_offset(page) + page 93 fs/ufs/dir.c (char *) de - (char *) page_address(page); page 97 fs/ufs/dir.c lock_page(page); page 98 fs/ufs/dir.c err = __ufs_write_begin(NULL, page->mapping, pos, len, page 99 fs/ufs/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 105 fs/ufs/dir.c err = ufs_commit_chunk(page, pos, len); page 106 fs/ufs/dir.c ufs_put_page(page); page 114 fs/ufs/dir.c struct inode *dir = page->mapping->host; page 116 fs/ufs/dir.c char *kaddr = page_address(page); page 123 fs/ufs/dir.c if ((dir->i_size >> PAGE_CACHE_SHIFT) == page->index) { page 149 fs/ufs/dir.c SetPageChecked(page); page 177 fs/ufs/dir.c dir->i_ino, error, (page->index<<PAGE_CACHE_SHIFT)+offs, page 185 fs/ufs/dir.c dir->i_ino, (page->index<<PAGE_CACHE_SHIFT)+offs); page 187 fs/ufs/dir.c SetPageChecked(page); page 188 fs/ufs/dir.c SetPageError(page); page 191 fs/ufs/dir.c static struct page *ufs_get_page(struct inode *dir, unsigned long n) page 194 fs/ufs/dir.c struct page *page = read_mapping_page(mapping, n, NULL); page 195 fs/ufs/dir.c if (!IS_ERR(page)) { page 196 fs/ufs/dir.c kmap(page); page 197 fs/ufs/dir.c if (!PageChecked(page)) page 198 fs/ufs/dir.c ufs_check_page(page); page 199 fs/ufs/dir.c if (PageError(page)) page 202 fs/ufs/dir.c return page; page 205 fs/ufs/dir.c ufs_put_page(page); page 233 fs/ufs/dir.c struct page *page = ufs_get_page(dir, 0); page 236 fs/ufs/dir.c if (!IS_ERR(page)) { page 238 fs/ufs/dir.c (struct ufs_dir_entry *)page_address(page)); page 239 fs/ufs/dir.c *p = page; page 261 fs/ufs/dir.c struct page *page = NULL; page 280 fs/ufs/dir.c page = ufs_get_page(dir, n); page 281 fs/ufs/dir.c if (!IS_ERR(page)) { page 282 fs/ufs/dir.c kaddr = page_address(page); page 289 fs/ufs/dir.c ufs_put_page(page); page 296 fs/ufs/dir.c ufs_put_page(page); page 305 fs/ufs/dir.c *res_page = page; page 322 fs/ufs/dir.c struct page *page = NULL; page 340 fs/ufs/dir.c page = ufs_get_page(dir, n); page 341 fs/ufs/dir.c err = PTR_ERR(page); page 342 fs/ufs/dir.c if (IS_ERR(page)) page 344 fs/ufs/dir.c lock_page(page); page 345 fs/ufs/dir.c kaddr = page_address(page); page 375 fs/ufs/dir.c unlock_page(page); page 376 fs/ufs/dir.c ufs_put_page(page); page 382 fs/ufs/dir.c pos = page_offset(page) + page 383 fs/ufs/dir.c (char*)de - (char*)page_address(page); page 384 fs/ufs/dir.c err = __ufs_write_begin(NULL, page->mapping, pos, rec_len, page 385 fs/ufs/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 402 fs/ufs/dir.c err = ufs_commit_chunk(page, pos, rec_len); page 408 fs/ufs/dir.c ufs_put_page(page); page 412 fs/ufs/dir.c unlock_page(page); page 456 fs/ufs/dir.c struct page *page = ufs_get_page(inode, n); page 458 fs/ufs/dir.c if (IS_ERR(page)) { page 465 fs/ufs/dir.c kaddr = page_address(page); page 480 fs/ufs/dir.c ufs_put_page(page); page 501 fs/ufs/dir.c ufs_put_page(page); page 507 fs/ufs/dir.c ufs_put_page(page); page 521 fs/ufs/dir.c struct address_space *mapping = page->mapping; page 522 fs/ufs/dir.c char *kaddr = page_address(page); page 548 fs/ufs/dir.c from = (char*)pde - (char*)page_address(page); page 550 fs/ufs/dir.c pos = page_offset(page) + from; page 551 fs/ufs/dir.c lock_page(page); page 553 fs/ufs/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 558 fs/ufs/dir.c err = ufs_commit_chunk(page, pos, to - from); page 562 fs/ufs/dir.c ufs_put_page(page); page 571 fs/ufs/dir.c struct page *page = grab_cache_page(mapping, 0); page 577 fs/ufs/dir.c if (!page) page 581 fs/ufs/dir.c AOP_FLAG_UNINTERRUPTIBLE, &page, NULL); page 583 fs/ufs/dir.c unlock_page(page); page 587 fs/ufs/dir.c kmap(page); page 588 fs/ufs/dir.c base = (char*)page_address(page); page 605 fs/ufs/dir.c kunmap(page); page 607 fs/ufs/dir.c err = ufs_commit_chunk(page, 0, chunk_size); page 609 fs/ufs/dir.c page_cache_release(page); page 619 fs/ufs/dir.c struct page *page = NULL; page 625 fs/ufs/dir.c page = ufs_get_page(inode, i); page 627 fs/ufs/dir.c if (IS_ERR(page)) page 630 fs/ufs/dir.c kaddr = page_address(page); page 657 fs/ufs/dir.c ufs_put_page(page); page 662 fs/ufs/dir.c ufs_put_page(page); page 560 fs/ufs/inode.c return block_write_full_page(page,ufs_getfrag_block,wbc); page 565 fs/ufs/inode.c return block_read_full_page(page,ufs_getfrag_block); page 237 fs/ufs/namei.c struct page *page; page 240 fs/ufs/namei.c de = ufs_find_entry(dir, dentry, &page); page 244 fs/ufs/namei.c err = ufs_delete_entry(dir, de, page); page 278 fs/ufs/namei.c struct page *dir_page = NULL; page 280 fs/ufs/namei.c struct page *old_page; page 296 fs/ufs/namei.c struct page *new_page; page 388 fs/ufs/truncate.c struct page *lastpage; page 248 fs/ufs/util.c struct page *ufs_get_locked_page(struct address_space *mapping, page 251 fs/ufs/util.c struct page *page; page 253 fs/ufs/util.c page = find_lock_page(mapping, index); page 254 fs/ufs/util.c if (!page) { page 255 fs/ufs/util.c page = read_mapping_page(mapping, index, NULL); page 257 fs/ufs/util.c if (IS_ERR(page)) { page 264 fs/ufs/util.c lock_page(page); page 266 fs/ufs/util.c if (unlikely(page->mapping == NULL)) { page 268 fs/ufs/util.c unlock_page(page); page 269 fs/ufs/util.c page_cache_release(page); page 270 fs/ufs/util.c page = NULL; page 274 fs/ufs/util.c if (!PageUptodate(page) || PageError(page)) { page 275 fs/ufs/util.c unlock_page(page); page 276 fs/ufs/util.c page_cache_release(page); page 282 fs/ufs/util.c page = ERR_PTR(-EIO); page 286 fs/ufs/util.c return page; page 284 fs/ufs/util.h extern struct page *ufs_get_locked_page(struct address_space *mapping, page 288 fs/ufs/util.h unlock_page(page); page 289 fs/ufs/util.h page_cache_release(page); page 56 fs/xfs/linux-2.6/xfs_aops.c bh = head = page_buffers(page); page 77 fs/xfs/linux-2.6/xfs_aops.c loff_t offset = page_offset(page); page 80 fs/xfs/linux-2.6/xfs_aops.c if (page_has_buffers(page)) page 81 fs/xfs/linux-2.6/xfs_aops.c xfs_count_page_state(page, &delalloc, &unmapped, &unwritten); page 91 fs/xfs/linux-2.6/xfs_aops.c (void *)page, page 414 fs/xfs/linux-2.6/xfs_aops.c ASSERT(PageLocked(page)); page 415 fs/xfs/linux-2.6/xfs_aops.c ASSERT(!PageWriteback(page)); page 417 fs/xfs/linux-2.6/xfs_aops.c clear_page_dirty_for_io(page); page 418 fs/xfs/linux-2.6/xfs_aops.c set_page_writeback(page); page 419 fs/xfs/linux-2.6/xfs_aops.c unlock_page(page); page 422 fs/xfs/linux-2.6/xfs_aops.c end_page_writeback(page); page 605 fs/xfs/linux-2.6/xfs_aops.c if (PageWriteback(page)) page 608 fs/xfs/linux-2.6/xfs_aops.c if (page->mapping && PageDirty(page)) { page 609 fs/xfs/linux-2.6/xfs_aops.c if (page_has_buffers(page)) { page 612 fs/xfs/linux-2.6/xfs_aops.c bh = head = page_buffers(page); page 664 fs/xfs/linux-2.6/xfs_aops.c struct page *page = pvec.pages[i]; page 677 fs/xfs/linux-2.6/xfs_aops.c if (page->index == tindex && trylock_page(page)) { page 678 fs/xfs/linux-2.6/xfs_aops.c pg_len = xfs_probe_page(page, pg_offset, mapped); page 679 fs/xfs/linux-2.6/xfs_aops.c unlock_page(page); page 707 fs/xfs/linux-2.6/xfs_aops.c if (PageWriteback(page)) page 710 fs/xfs/linux-2.6/xfs_aops.c if (page->mapping && page_has_buffers(page)) { page 714 fs/xfs/linux-2.6/xfs_aops.c bh = head = page_buffers(page); page 757 fs/xfs/linux-2.6/xfs_aops.c xfs_off_t offset = page_offset(page); page 759 fs/xfs/linux-2.6/xfs_aops.c if (page->index != tindex) page 761 fs/xfs/linux-2.6/xfs_aops.c if (!trylock_page(page)) page 763 fs/xfs/linux-2.6/xfs_aops.c if (PageWriteback(page)) page 765 fs/xfs/linux-2.6/xfs_aops.c if (page->mapping != inode->i_mapping) page 767 fs/xfs/linux-2.6/xfs_aops.c if (!xfs_is_delayed_page(page, (*ioendp)->io_type)) page 784 fs/xfs/linux-2.6/xfs_aops.c (xfs_off_t)(page->index + 1) << PAGE_CACHE_SHIFT, page 793 fs/xfs/linux-2.6/xfs_aops.c bh = head = page_buffers(page); page 799 fs/xfs/linux-2.6/xfs_aops.c if (!(PageUptodate(page) || buffer_uptodate(bh))) { page 844 fs/xfs/linux-2.6/xfs_aops.c SetPageUptodate(page); page 859 fs/xfs/linux-2.6/xfs_aops.c xfs_start_page_writeback(page, !page_dirty, count); page 864 fs/xfs/linux-2.6/xfs_aops.c unlock_page(page); page 956 fs/xfs/linux-2.6/xfs_aops.c if (page->index >= end_index) { page 957 fs/xfs/linux-2.6/xfs_aops.c if ((page->index >= end_index + 1) || page 960 fs/xfs/linux-2.6/xfs_aops.c unlock_page(page); page 979 fs/xfs/linux-2.6/xfs_aops.c (xfs_off_t)(page->index + 1) << PAGE_CACHE_SHIFT, offset); page 986 fs/xfs/linux-2.6/xfs_aops.c bh = head = page_buffers(page); page 987 fs/xfs/linux-2.6/xfs_aops.c offset = page_offset(page); page 998 fs/xfs/linux-2.6/xfs_aops.c if (!(PageUptodate(page) || buffer_uptodate(bh)) && !startio) { page 1021 fs/xfs/linux-2.6/xfs_aops.c ((buffer_uptodate(bh) || PageUptodate(page)) && page 1054 fs/xfs/linux-2.6/xfs_aops.c page, bh, head, 0); page 1088 fs/xfs/linux-2.6/xfs_aops.c size = xfs_probe_cluster(inode, page, bh, page 1117 fs/xfs/linux-2.6/xfs_aops.c } else if ((buffer_uptodate(bh) || PageUptodate(page)) && page 1128 fs/xfs/linux-2.6/xfs_aops.c SetPageUptodate(page); page 1131 fs/xfs/linux-2.6/xfs_aops.c xfs_start_page_writeback(page, 1, count); page 1137 fs/xfs/linux-2.6/xfs_aops.c xfs_cluster_write(inode, page->index + 1, &iomap, &ioend, page 1157 fs/xfs/linux-2.6/xfs_aops.c block_invalidatepage(page, 0); page 1158 fs/xfs/linux-2.6/xfs_aops.c ClearPageUptodate(page); page 1191 fs/xfs/linux-2.6/xfs_aops.c struct inode *inode = page->mapping->host; page 1193 fs/xfs/linux-2.6/xfs_aops.c xfs_page_trace(XFS_WRITEPAGE_ENTER, inode, page, 0); page 1203 fs/xfs/linux-2.6/xfs_aops.c if (!page_has_buffers(page)) { page 1207 fs/xfs/linux-2.6/xfs_aops.c xfs_count_page_state(page, &delalloc, &unmapped, &unwritten); page 1208 fs/xfs/linux-2.6/xfs_aops.c if (!PageUptodate(page)) page 1226 fs/xfs/linux-2.6/xfs_aops.c if (!page_has_buffers(page)) page 1227 fs/xfs/linux-2.6/xfs_aops.c create_empty_buffers(page, 1 << inode->i_blkbits, 0); page 1233 fs/xfs/linux-2.6/xfs_aops.c error = xfs_page_state_convert(inode, page, wbc, 1, unmapped); page 1242 fs/xfs/linux-2.6/xfs_aops.c redirty_page_for_writepage(wbc, page); page 1243 fs/xfs/linux-2.6/xfs_aops.c unlock_page(page); page 1246 fs/xfs/linux-2.6/xfs_aops.c unlock_page(page); page 1283 fs/xfs/linux-2.6/xfs_aops.c struct inode *inode = page->mapping->host; page 1290 fs/xfs/linux-2.6/xfs_aops.c xfs_page_trace(XFS_RELEASEPAGE_ENTER, inode, page, 0); page 1292 fs/xfs/linux-2.6/xfs_aops.c if (!page_has_buffers(page)) page 1295 fs/xfs/linux-2.6/xfs_aops.c xfs_count_page_state(page, &delalloc, &unmapped, &unwritten); page 1314 fs/xfs/linux-2.6/xfs_aops.c dirty = xfs_page_state_convert(inode, page, &wbc, 0, 0); page 1320 fs/xfs/linux-2.6/xfs_aops.c return try_to_free_buffers(page); page 1552 fs/xfs/linux-2.6/xfs_aops.c return mpage_readpage(page, xfs_get_blocks); page 1571 fs/xfs/linux-2.6/xfs_aops.c page->mapping->host, page, offset); page 1572 fs/xfs/linux-2.6/xfs_aops.c block_invalidatepage(page, offset); page 151 fs/xfs/linux-2.6/xfs_buf.c set_page_private(page, page 152 fs/xfs/linux-2.6/xfs_buf.c page_private(page) | page_region_mask(offset, length)); page 153 fs/xfs/linux-2.6/xfs_buf.c if (page_private(page) == ~0UL) page 154 fs/xfs/linux-2.6/xfs_buf.c SetPageUptodate(page); page 165 fs/xfs/linux-2.6/xfs_buf.c return (mask && (page_private(page) & mask) == mask); page 295 fs/xfs/linux-2.6/xfs_buf.c bp->b_pages = kmem_alloc(sizeof(struct page *) * page 300 fs/xfs/linux-2.6/xfs_buf.c memset(bp->b_pages, 0, sizeof(struct page *) * page_count); page 339 fs/xfs/linux-2.6/xfs_buf.c struct page *page = bp->b_pages[i]; page 342 fs/xfs/linux-2.6/xfs_buf.c ASSERT(!PagePrivate(page)); page 343 fs/xfs/linux-2.6/xfs_buf.c page_cache_release(page); page 381 fs/xfs/linux-2.6/xfs_buf.c struct page *page; page 385 fs/xfs/linux-2.6/xfs_buf.c page = find_or_create_page(mapping, first + i, gfp_mask); page 386 fs/xfs/linux-2.6/xfs_buf.c if (unlikely(page == NULL)) { page 417 fs/xfs/linux-2.6/xfs_buf.c ASSERT(!PagePrivate(page)); page 418 fs/xfs/linux-2.6/xfs_buf.c if (!PageUptodate(page)) { page 423 fs/xfs/linux-2.6/xfs_buf.c } else if (!PagePrivate(page)) { page 424 fs/xfs/linux-2.6/xfs_buf.c if (test_page_region(page, offset, nbytes)) page 429 fs/xfs/linux-2.6/xfs_buf.c bp->b_pages[i] = page; page 707 fs/xfs/linux-2.6/xfs_buf.c static inline struct page * page 1121 fs/xfs/linux-2.6/xfs_buf.c struct page *page = bvec->bv_page; page 1123 fs/xfs/linux-2.6/xfs_buf.c ASSERT(!PagePrivate(page)); page 1126 fs/xfs/linux-2.6/xfs_buf.c ClearPageUptodate(page); page 1128 fs/xfs/linux-2.6/xfs_buf.c SetPageUptodate(page); page 1129 fs/xfs/linux-2.6/xfs_buf.c } else if (!PagePrivate(page) && page 1131 fs/xfs/linux-2.6/xfs_buf.c set_page_region(page, bvec->bv_offset, bvec->bv_len); page 1138 fs/xfs/linux-2.6/xfs_buf.c unlock_page(page); page 1285 fs/xfs/linux-2.6/xfs_buf.c struct page *page; page 1291 fs/xfs/linux-2.6/xfs_buf.c page = bp->b_pages[offset >> PAGE_CACHE_SHIFT]; page 1292 fs/xfs/linux-2.6/xfs_buf.c return (xfs_caddr_t)page_address(page) + (offset & (PAGE_CACHE_SIZE-1)); page 1307 fs/xfs/linux-2.6/xfs_buf.c struct page *page; page 1311 fs/xfs/linux-2.6/xfs_buf.c page = bp->b_pages[xfs_buf_btoct(boff + bp->b_offset)]; page 1320 fs/xfs/linux-2.6/xfs_buf.c memset(page_address(page) + cpoff, 0, csize); page 1323 fs/xfs/linux-2.6/xfs_buf.c memcpy(data, page_address(page) + cpoff, csize); page 1326 fs/xfs/linux-2.6/xfs_buf.c memcpy(page_address(page) + cpoff, data, csize); page 175 fs/xfs/linux-2.6/xfs_buf.h struct page **b_pages; /* array of page pointers */ page 176 fs/xfs/linux-2.6/xfs_buf.h struct page *b_page_array[XB_PAGES]; /* inline pages */ page 432 fs/xfs/linux-2.6/xfs_file.c return block_page_mkwrite(vma, page, xfs_get_blocks); page 136 fs/xfs/linux-2.6/xfs_lrw.c struct page *page; page 152 fs/xfs/linux-2.6/xfs_lrw.c &page, &fsdata); page 156 fs/xfs/linux-2.6/xfs_lrw.c zero_user(page, offset, bytes); page 159 fs/xfs/linux-2.6/xfs_lrw.c page, fsdata); page 68 include/asm-cris/dma-mapping.h return page_to_phys(page) + offset; page 12 include/asm-cris/page.h #define clear_page(page) memset((void *)(page), 0, PAGE_SIZE) page 15 include/asm-cris/page.h #define clear_user_page(page, vaddr, pg) clear_page(page) page 29 include/asm-cris/page.h typedef struct page *pgtable_t; page 54 include/asm-cris/page.h #define VALID_PAGE(page) (((page) - mem_map) < max_mapnr) page 61 include/asm-cris/page.h #define page_to_phys(page) __pa((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET) page 33 include/asm-cris/pgalloc.h struct page *pte; page 183 include/asm-cris/pgtable.h pte_val(pte) = __pa(page) | pgprot_val(pgprot); page 187 include/asm-cris/pgtable.h #define mk_pte(page, pgprot) __mk_pte(page_address(page), (pgprot)) page 216 include/asm-cris/pgtable.h #define __page_address(page) (PAGE_OFFSET + (((page) - mem_map) << PAGE_SHIFT)) page 55 include/asm-frv/cacheflush.h unsigned long addr = page_to_phys(page); page 62 include/asm-frv/cacheflush.h flush_dcache_page(page); page 88 include/asm-frv/cacheflush.h flush_icache_user_range(vma, page, page_to_phys(page), PAGE_SIZE); page 98 include/asm-frv/cacheflush.h flush_icache_user_range((vma), (page), (vaddr), (len)); \ page 64 include/asm-frv/highmem.h extern struct page *kmap_atomic_to_page(void *ptr); page 119 include/asm-frv/highmem.h paddr = page_to_phys(page); page 28 include/asm-frv/page.h typedef struct page *pgtable_t; page 423 include/asm-frv/pgtable.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) page 37 include/asm-frv/virtconvert.h #define __page_address(page) (PAGE_OFFSET + (((page) - mem_map) << PAGE_SHIFT)) page 38 include/asm-frv/virtconvert.h #define page_to_phys(page) virt_to_phys((void *)__page_address(page)) page 78 include/asm-generic/dma-mapping.h return pci_map_page(to_pci_dev(dev), page, offset, size, (int)direction); page 31 include/asm-generic/memory_model.h #define __page_to_pfn(page) ((unsigned long)((page) - mem_map) + \ page 42 include/asm-generic/memory_model.h ({ struct page *__pg = (pg); \ page 52 include/asm-generic/memory_model.h #define __page_to_pfn(page) ((page) - vmemmap) page 60 include/asm-generic/memory_model.h ({ struct page *__pg = (pg); \ page 73 include/asm-generic/memory_model.h struct page; page 75 include/asm-generic/memory_model.h extern struct page *pfn_to_page(unsigned long pfn); page 49 include/asm-generic/pci-dma-compat.h return dma_map_page(hwdev == NULL ? NULL : &hwdev->dev, page, offset, size, (enum dma_data_direction)direction); page 44 include/asm-generic/tlb.h struct page * pages[FREE_PTE_NR]; page 105 include/asm-generic/tlb.h free_page_and_swap_cache(page); page 108 include/asm-generic/tlb.h tlb->pages[tlb->nr++] = page; page 64 include/asm-m32r/cacheflush.h flush_icache_user_range(vma, page, vaddr, len); \ page 74 include/asm-m32r/io.h #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT) page 14 include/asm-m32r/page.h #define clear_user_page(page, vaddr, pg) clear_page(page) page 31 include/asm-m32r/page.h typedef struct page *pgtable_t; page 44 include/asm-m32r/pgalloc.h struct page *pte = alloc_page(GFP_KERNEL|__GFP_ZERO); page 290 include/asm-m32r/pgtable.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), pgprot) page 27 include/asm-m32r/tlbflush.h #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) page 46 include/asm-m32r/tlbflush.h #define flush_tlb_page(vma, page) smp_flush_tlb_page(vma, page) page 71 include/asm-m32r/tlbflush.h : "r" (page), "i" (MMU_REG_BASE), "i" (MSVA_offset), page 131 include/asm-m68k/cacheflush.h #define flush_dcache_page(page) __flush_page_to_ram(page_address(page)) page 134 include/asm-m68k/cacheflush.h #define flush_icache_page(vma, page) __flush_page_to_ram(page_address(page)) page 144 include/asm-m68k/cacheflush.h flush_cache_page(vma, vaddr, page_to_pfn(page)); page 146 include/asm-m68k/cacheflush.h flush_icache_user_range(vma, page, vaddr, len); page 152 include/asm-m68k/cacheflush.h flush_cache_page(vma, vaddr, page_to_pfn(page)); page 32 include/asm-m68k/motorola_pgalloc.h struct page *page = alloc_pages(GFP_KERNEL|__GFP_REPEAT|__GFP_ZERO, 0); page 35 include/asm-m68k/motorola_pgalloc.h if(!page) page 38 include/asm-m68k/motorola_pgalloc.h pte = kmap(page); page 45 include/asm-m68k/motorola_pgalloc.h pgtable_page_ctor(page); page 46 include/asm-m68k/motorola_pgalloc.h return page; page 51 include/asm-m68k/motorola_pgalloc.h pgtable_page_dtor(page); page 52 include/asm-m68k/motorola_pgalloc.h cache_page(kmap(page)); page 53 include/asm-m68k/motorola_pgalloc.h kunmap(page); page 54 include/asm-m68k/motorola_pgalloc.h __free_page(page); page 59 include/asm-m68k/motorola_pgalloc.h pgtable_page_dtor(page); page 60 include/asm-m68k/motorola_pgalloc.h cache_page(kmap(page)); page 61 include/asm-m68k/motorola_pgalloc.h kunmap(page); page 62 include/asm-m68k/motorola_pgalloc.h __free_page(page); page 100 include/asm-m68k/motorola_pgalloc.h pmd_set(pmd, page_address(page)); page 100 include/asm-m68k/motorola_pgtable.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) page 54 include/asm-m68k/page.h unsigned long *sp = page; page 69 include/asm-m68k/page.h : "a" (page), "0" (sp), page 74 include/asm-m68k/page.h #define clear_page(page) memset((page), 0, PAGE_SIZE) page 80 include/asm-m68k/page.h flush_dcache_page(page); \ page 84 include/asm-m68k/page.h flush_dcache_page(page); \ page 94 include/asm-m68k/page.h typedef struct page *pgtable_t; page 202 include/asm-m68k/page.h pfn_to_virt(page_to_pfn(page)); \ page 212 include/asm-m68k/page.h struct page *__p = (_page); \ page 31 include/asm-m68k/sun3_pgalloc.h pgtable_page_dtor(page); page 32 include/asm-m68k/sun3_pgalloc.h __free_page(page); page 44 include/asm-m68k/sun3_pgalloc.h unsigned long page = __get_free_page(GFP_KERNEL|__GFP_REPEAT); page 46 include/asm-m68k/sun3_pgalloc.h if (!page) page 49 include/asm-m68k/sun3_pgalloc.h memset((void *)page, 0, PAGE_SIZE); page 50 include/asm-m68k/sun3_pgalloc.h return (pte_t *) (page); page 56 include/asm-m68k/sun3_pgalloc.h struct page *page = alloc_pages(GFP_KERNEL|__GFP_REPEAT, 0); page 58 include/asm-m68k/sun3_pgalloc.h if (page == NULL) page 61 include/asm-m68k/sun3_pgalloc.h clear_highpage(page); page 62 include/asm-m68k/sun3_pgalloc.h pgtable_page_ctor(page); page 63 include/asm-m68k/sun3_pgalloc.h return page; page 74 include/asm-m68k/sun3_pgalloc.h pmd_val(*pmd) = __pa((unsigned long)page_address(page)); page 104 include/asm-m68k/sun3_pgtable.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) page 31 include/asm-m68k/virtconvert.h __pa(PAGE_OFFSET + (((page) - pg_data_map[0].node_mem_map) << PAGE_SHIFT)) page 34 include/asm-m68k/virtconvert.h struct page *__page = _page; \ page 54 include/asm-mn10300/cacheflush.h flush_icache_page(vma, page); \ page 116 include/asm-mn10300/dma-mapping.h return page_to_bus(page) + offset; page 52 include/asm-mn10300/highmem.h if (page < highmem_start_page) page 53 include/asm-mn10300/highmem.h return page_address(page); page 54 include/asm-mn10300/highmem.h return kmap_high(page); page 61 include/asm-mn10300/highmem.h if (page < highmem_start_page) page 63 include/asm-mn10300/highmem.h kunmap_high(page); page 77 include/asm-mn10300/highmem.h if (page < highmem_start_page) page 78 include/asm-mn10300/highmem.h return page_address(page); page 86 include/asm-mn10300/highmem.h set_pte(kmap_pte - idx, mk_pte(page, kmap_prot)); page 29 include/asm-mn10300/page.h #define clear_page(page) memset((void *)(page), 0, PAGE_SIZE) page 41 include/asm-mn10300/page.h typedef struct page *pgtable_t; page 109 include/asm-mn10300/page.h #define page_to_pfn(page) ((unsigned long)((page) - mem_map) + __pfn_disp) page 119 include/asm-mn10300/page.h #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT) page 20 include/asm-mn10300/pgalloc.h struct page; page 41 include/asm-mn10300/pgalloc.h extern struct page *pte_alloc_one(struct mm_struct *, unsigned long); page 44 include/asm-mn10300/pgtable.h extern struct page *pgd_list; page 384 include/asm-mn10300/pgtable.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) page 395 include/asm-mn10300/pgtable.h #define page_pte(page) page_pte_prot((page), __pgprot(0)) page 53 include/asm-parisc/cacheflush.h flush_kernel_dcache_page(page); \ page 54 include/asm-parisc/cacheflush.h flush_kernel_icache_page(page_address(page)); \ page 64 include/asm-parisc/cacheflush.h flush_cache_page(vma, vaddr, page_to_pfn(page)); \ page 71 include/asm-parisc/cacheflush.h flush_cache_page(vma, vaddr, page_to_pfn(page)); \ page 83 include/asm-parisc/cacheflush.h if (PageAnon(page)) page 91 include/asm-parisc/cacheflush.h flush_kernel_dcache_page_addr(page_address(page)); page 107 include/asm-parisc/cacheflush.h return page_address(page); page 110 include/asm-parisc/cacheflush.h #define kunmap(page) kunmap_parisc(page_address(page)) page 112 include/asm-parisc/cacheflush.h #define kmap_atomic(page, idx) page_address(page) page 109 include/asm-parisc/dma-mapping.h return dma_map_single(dev, (page_address(page) + (offset)), size, direction); page 24 include/asm-parisc/page.h #define clear_page(page) memset((void *)(page), 0, PAGE_SIZE) page 27 include/asm-parisc/page.h struct page; page 94 include/asm-parisc/page.h typedef struct page *pgtable_t; page 164 include/asm-parisc/page.h #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT) page 123 include/asm-parisc/pgalloc.h struct page *page = alloc_page(GFP_KERNEL|__GFP_REPEAT|__GFP_ZERO); page 124 include/asm-parisc/pgalloc.h if (page) page 125 include/asm-parisc/pgalloc.h pgtable_page_ctor(page); page 126 include/asm-parisc/pgalloc.h return page; page 349 include/asm-parisc/pgtable.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) page 19 include/asm-um/page.h struct page; page 28 include/asm-um/page.h #define clear_page(page) memset((void *)(page), 0, PAGE_SIZE) page 31 include/asm-um/page.h #define clear_user_page(page, vaddr, pg) clear_page(page) page 86 include/asm-um/page.h typedef struct page *pgtable_t; page 275 include/asm-um/pgtable.h #define page_to_phys(page) pfn_to_phys((pfn_t) page_to_pfn(page)) page 281 include/asm-um/pgtable.h pte_set_val(pte, page_to_phys(page), (pgprot)); \ page 102 include/asm-um/tlb.h free_page_and_swap_cache(page); page 15 include/asm-x86/agp.h #define map_page_into_agp(page) set_pages_uc(page, 1) page 16 include/asm-x86/agp.h #define unmap_page_from_agp(page) set_pages_wb(page, 1) page 220 include/asm-x86/dma-mapping.h return ops->map_single(dev, page_to_phys(page) + offset, page 69 include/asm-x86/highmem.h struct page *kmap_atomic_to_page(void *ptr); page 72 include/asm-x86/highmem.h #define kmap_atomic_pte(page, type) kmap_atomic(page, type) page 95 include/asm-x86/io_32.h #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT) page 164 include/asm-x86/io_64.h #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT) page 209 include/asm-x86/kvm_host.h struct kvm_mmu_page *page); page 294 include/asm-x86/kvm_host.h struct page *time_page; page 325 include/asm-x86/kvm_host.h struct page *apic_access_page; page 329 include/asm-x86/kvm_host.h struct page *ept_identity_pagetable; page 568 include/asm-x86/kvm_host.h struct page *page = pfn_to_page(shadow_page >> PAGE_SHIFT); page 570 include/asm-x86/kvm_host.h return (struct kvm_mmu_page *)page_private(page); page 70 include/asm-x86/page.h struct page; page 75 include/asm-x86/page.h clear_page(page); page 74 include/asm-x86/page_32.h typedef struct page *pgtable_t; page 116 include/asm-x86/page_32.h mmx_clear_page(page); page 128 include/asm-x86/page_32.h memset(page, 0, PAGE_SIZE); page 84 include/asm-x86/page_64.h typedef struct page *pgtable_t; page 88 include/asm-x86/page_64.h #define vmemmap ((struct page *)VMEMMAP_START) page 36 include/asm-x86/paravirt.h struct page; page 313 include/asm-x86/paravirt.h void *(*kmap_atomic_pte)(struct page *page, enum km_type type); page 1040 include/asm-x86/paravirt.h ret = PVOP_CALL2(unsigned long, pv_mmu_ops.kmap_atomic_pte, page, type); page 123 include/asm-x86/pgtable-3level.h #define pud_page(pud) ((struct page *) __va(pud_val(pud) & PTE_PFN_MASK)) page 116 include/asm-x86/pgtable_32.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) page 236 include/asm-x86/pgtable_64.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn((page)), (pgprot)) page 75 include/asm-xtensa/dma-mapping.h return (dma_addr_t)(page_to_pfn(page)) * PAGE_SIZE + offset; page 99 include/asm-xtensa/page.h typedef struct page *pgtable_t; page 129 include/asm-xtensa/page.h struct page; page 142 include/asm-xtensa/page.h # define clear_user_page(page, vaddr, pg) clear_page(page) page 161 include/asm-xtensa/page.h #define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT) page 163 include/asm-xtensa/page.h #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT) page 26 include/asm-xtensa/pgalloc.h (pmd_val(*(pmdp)) = ((unsigned long)page_to_virt(page))) page 53 include/asm-xtensa/pgalloc.h struct page *page; page 55 include/asm-xtensa/pgalloc.h page = virt_to_page(pte_alloc_one_kernel(mm, addr)); page 56 include/asm-xtensa/pgalloc.h pgtable_page_ctor(page); page 57 include/asm-xtensa/pgalloc.h return page; page 239 include/asm-xtensa/pgtable.h #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) page 73 include/crypto/algapi.h struct page *page; page 78 include/crypto/algapi.h u8 *page; page 89 include/crypto/algapi.h void *page; page 28 include/crypto/internal/hash.h struct page *pg; page 42 include/crypto/scatterwalk.h return kmap_atomic(page, crypto_kmap_type(out)); page 104 include/crypto/scatterwalk.h static inline struct page *scatterwalk_page(struct scatter_walk *walk) page 483 include/drm/drmP.h struct page **pagelist; page 170 include/linux/aio.h struct page **ring_pages; page 176 include/linux/aio.h struct page *internal_pages[AIO_RING_PAGES]; page 392 include/linux/atmdev.h int (*proc_read)(struct atm_dev *dev,loff_t *pos,char *page); page 18 include/linux/backing-dev.h struct page; page 48 include/linux/backing-dev.h void (*unplug_io_fn)(struct backing_dev_info *, struct page *); page 33 include/linux/binfmts.h struct page *page[MAX_ARG_PAGES]; page 47 include/linux/bio.h struct page *bv_page; page 671 include/linux/blkdev.h struct page **pages; page 771 include/linux/blkdev.h bdi->unplug_io_fn(bdi, page); page 960 include/linux/blkdev.h typedef struct {struct page *v;} Sector; page 46 include/linux/buffer_head.h struct page; page 63 include/linux/buffer_head.h struct page *b_page; /* the page this bh is mapped to */ page 137 include/linux/buffer_head.h BUG_ON(!PagePrivate(page)); \ page 138 include/linux/buffer_head.h ((struct buffer_head *)page_private(page)); \ page 140 include/linux/buffer_head.h #define page_has_buffers(page) PagePrivate(page) page 250 include/linux/buffer_head.h page_cache_get(page); page 251 include/linux/buffer_head.h SetPagePrivate(page); page 252 include/linux/buffer_head.h set_page_private(page, (unsigned long)head); page 185 include/linux/configfs.h char *page) \ page 193 include/linux/configfs.h ret = _item##_attr->show(_item, page); \ page 198 include/linux/configfs.h const char *page, size_t count) \ page 206 include/linux/configfs.h ret = _item##_attr->store(_item, page, count); \ page 25 include/linux/dm-io.h struct page *page; page 489 include/linux/dmaengine.h struct page **pages; page 415 include/linux/fs.h struct page; page 474 include/linux/fs.h int (*writepage)(struct page *page, struct writeback_control *wbc); page 475 include/linux/fs.h int (*readpage)(struct file *, struct page *); page 476 include/linux/fs.h void (*sync_page)(struct page *); page 482 include/linux/fs.h int (*set_page_dirty)(struct page *page); page 491 include/linux/fs.h int (*prepare_write)(struct file *, struct page *, unsigned, unsigned); page 492 include/linux/fs.h int (*commit_write)(struct file *, struct page *, unsigned, unsigned); page 496 include/linux/fs.h struct page **pagep, void **fsdata); page 499 include/linux/fs.h struct page *page, void *fsdata); page 503 include/linux/fs.h void (*invalidatepage) (struct page *, unsigned long); page 504 include/linux/fs.h int (*releasepage) (struct page *, gfp_t); page 511 include/linux/fs.h struct page *, struct page *); page 512 include/linux/fs.h int (*launder_page) (struct page *); page 513 include/linux/fs.h int (*is_partially_uptodate) (struct page *, read_descriptor_t *, page 1315 include/linux/fs.h ssize_t (*sendpage) (struct file *, struct page *, int, size_t, loff_t *, int); page 176 include/linux/gfp.h struct page * page 180 include/linux/gfp.h static inline struct page * page 187 include/linux/gfp.h static inline struct page * page 195 include/linux/gfp.h static inline struct page *alloc_pages_node(int nid, gfp_t gfp_mask, page 209 include/linux/gfp.h extern struct page *alloc_pages_current(gfp_t gfp_mask, unsigned order); page 211 include/linux/gfp.h static inline struct page * page 219 include/linux/gfp.h extern struct page *alloc_page_vma(gfp_t gfp_mask, page 245 include/linux/gfp.h #define __free_page(page) __free_pages((page), 0) page 42 include/linux/highmem.h return page_address(page); page 45 include/linux/highmem.h #define kunmap(page) do { (void) (page); } while (0) page 52 include/linux/highmem.h return page_address(page); page 54 include/linux/highmem.h #define kmap_atomic_prot(page, idx, prot) kmap_atomic(page, idx) page 68 include/linux/highmem.h void *addr = kmap_atomic(page, KM_USER0); page 69 include/linux/highmem.h clear_user_page(addr, vaddr, page); page 88 include/linux/highmem.h static inline struct page * page 93 include/linux/highmem.h struct page *page = alloc_page_vma(GFP_HIGHUSER | movableflags, page 96 include/linux/highmem.h if (page) page 97 include/linux/highmem.h clear_user_highpage(page, vaddr); page 99 include/linux/highmem.h return page; page 111 include/linux/highmem.h static inline struct page * page 120 include/linux/highmem.h void *kaddr = kmap_atomic(page, KM_USER0); page 129 include/linux/highmem.h void *kaddr = kmap_atomic(page, KM_USER0); page 140 include/linux/highmem.h flush_dcache_page(page); page 146 include/linux/highmem.h zero_user_segments(page, start, end, 0, 0); page 152 include/linux/highmem.h zero_user_segments(page, start, start + size, 0, 0); page 158 include/linux/highmem.h zero_user(page, offset, size); page 50 include/linux/hugetlb.h struct page *follow_huge_addr(struct mm_struct *mm, unsigned long address, page 52 include/linux/hugetlb.h struct page *follow_huge_pmd(struct mm_struct *mm, unsigned long address, page 54 include/linux/hugetlb.h struct page *follow_huge_pud(struct mm_struct *mm, unsigned long address, page 263 include/linux/hugetlb.h return size_to_hstate(PAGE_SIZE << compound_order(page)); page 87 include/linux/i2o-dev.h unsigned int page; /* HTML page */ page 1043 include/linux/ide.h *start = page + off; \ page 53 include/linux/isdn/capilli.h int (*ctr_read_proc)(char *page, char **start, off_t off, page 85 include/linux/kexec.h struct page *control_code_page; page 86 include/linux/kexec.h struct page *swap_page; page 122 include/linux/kexec.h extern struct page *kimage_alloc_control_pages(struct kimage *image, page 159 include/linux/kvm_host.h struct page *gva_to_page(struct kvm_vcpu *vcpu, gva_t gva); page 161 include/linux/kvm_host.h extern struct page *bad_page; page 179 include/linux/kvm_host.h struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn); page 46 include/linux/kvm_types.h struct page *guest_pages[2]; page 38 include/linux/loop.h struct page *raw_page, unsigned raw_off, page 39 include/linux/loop.h struct page *loop_page, unsigned loop_off, page 138 include/linux/loop.h struct page *raw_page, unsigned raw_off, page 139 include/linux/loop.h struct page *loop_page, unsigned loop_off, page 25 include/linux/memcontrol.h struct page; page 30 include/linux/memcontrol.h #define page_reset_bad_cgroup(page) ((page)->page_cgroup = 0) page 8 include/linux/memory_hotplug.h struct page; page 220 include/linux/memory_hotplug.h extern struct page *sparse_decode_mem_map(unsigned long coded_mem_map, page 7 include/linux/migrate.h typedef struct page *new_page_t(struct page *, unsigned long private, int **); page 43 include/linux/mm.h #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n)) page 158 include/linux/mm.h struct page *page; /* ->fault handlers should return a page 177 include/linux/mm.h int (*page_mkwrite)(struct vm_area_struct *vma, struct page *page); page 214 include/linux/mm.h #define page_private(page) ((page)->private) page 215 include/linux/mm.h #define set_page_private(page, v) ((page)->private = (v)) page 241 include/linux/mm.h VM_BUG_ON(atomic_read(&page->_count) == 0); page 242 include/linux/mm.h return atomic_dec_and_test(&page->_count); page 251 include/linux/mm.h VM_BUG_ON(PageTail(page)); page 252 include/linux/mm.h return atomic_inc_not_zero(&page->_count); page 256 include/linux/mm.h struct page *vmalloc_to_page(const void *addr); page 276 include/linux/mm.h static inline struct page *compound_head(struct page *page) page 278 include/linux/mm.h if (unlikely(PageTail(page))) page 279 include/linux/mm.h return page->first_page; page 280 include/linux/mm.h return page; page 285 include/linux/mm.h return atomic_read(&compound_head(page)->_count); page 290 include/linux/mm.h page = compound_head(page); page 291 include/linux/mm.h VM_BUG_ON(atomic_read(&page->_count) == 0); page 292 include/linux/mm.h atomic_inc(&page->_count); page 295 include/linux/mm.h static inline struct page *virt_to_head_page(const void *x) page 297 include/linux/mm.h struct page *page = virt_to_page(x); page 298 include/linux/mm.h return compound_head(page); page 307 include/linux/mm.h atomic_set(&page->_count, 1); page 325 include/linux/mm.h page[1].lru.next = (void *)dtor; page 330 include/linux/mm.h return (compound_page_dtor *)page[1].lru.next; page 335 include/linux/mm.h if (!PageHead(page)) page 337 include/linux/mm.h return (unsigned long)page[1].lru.prev; page 342 include/linux/mm.h page[1].lru.prev = (void *)order; page 492 include/linux/mm.h return (page->flags >> ZONES_PGSHIFT) & ZONES_MASK; page 505 include/linux/mm.h return (page->flags >> ZONEID_PGSHIFT) & ZONEID_MASK; page 522 include/linux/mm.h return (page->flags >> NODES_PGSHIFT) & NODES_MASK; page 528 include/linux/mm.h return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; page 534 include/linux/mm.h return (page->flags >> SECTIONS_PGSHIFT) & SECTIONS_MASK; page 540 include/linux/mm.h page->flags &= ~(ZONES_MASK << ZONES_PGSHIFT); page 541 include/linux/mm.h page->flags |= (zone & ZONES_MASK) << ZONES_PGSHIFT; page 546 include/linux/mm.h page->flags &= ~(NODES_MASK << NODES_PGSHIFT); page 547 include/linux/mm.h page->flags |= (node & NODES_MASK) << NODES_PGSHIFT; page 552 include/linux/mm.h page->flags &= ~(SECTIONS_MASK << SECTIONS_PGSHIFT); page 553 include/linux/mm.h page->flags |= (section & SECTIONS_MASK) << SECTIONS_PGSHIFT; page 559 include/linux/mm.h set_page_zone(page, zone); page 560 include/linux/mm.h set_page_node(page, node); page 561 include/linux/mm.h set_page_section(page, pfn_to_section_nr(pfn)); page 586 include/linux/mm.h return __va(page_to_pfn(page) << PAGE_SHIFT); page 594 include/linux/mm.h #define page_address(page) ((page)->virtual) page 597 include/linux/mm.h (page)->virtual = (address); \ page 609 include/linux/mm.h #define page_address(page) lowmem_page_address(page) page 628 include/linux/mm.h struct address_space *mapping = page->mapping; page 630 include/linux/mm.h VM_BUG_ON(PageSlab(page)); page 632 include/linux/mm.h if (unlikely(PageSwapCache(page))) page 643 include/linux/mm.h return ((unsigned long)page->mapping & PAGE_MAPPING_ANON) != 0; page 652 include/linux/mm.h if (unlikely(PageSwapCache(page))) page 653 include/linux/mm.h return page_private(page); page 654 include/linux/mm.h return page->index; page 664 include/linux/mm.h atomic_set(&(page)->_mapcount, -1); page 669 include/linux/mm.h return atomic_read(&(page)->_mapcount) + 1; page 677 include/linux/mm.h return atomic_read(&(page)->_mapcount) >= 0; page 739 include/linux/mm.h struct page *vm_normal_page(struct vm_area_struct *vma, unsigned long addr, page 924 include/linux/mm.h #define __pte_lockptr(page) &((page)->ptl) page 928 include/linux/mm.h #define pte_lock_deinit(page) ((page)->mapping = NULL) page 941 include/linux/mm.h pte_lock_init(page); page 942 include/linux/mm.h inc_zone_page_state(page, NR_PAGETABLE); page 947 include/linux/mm.h pte_lock_deinit(page); page 948 include/linux/mm.h dec_zone_page_state(page, NR_PAGETABLE); page 1205 include/linux/mm.h struct page *follow_page(struct vm_area_struct *, unsigned long address, page 1272 include/linux/mm.h struct page *sparse_mem_map_populate(unsigned long pnum, int nid); page 4 include/linux/mm_inline.h list_add(&page->lru, &zone->active_list); page 11 include/linux/mm_inline.h list_add(&page->lru, &zone->inactive_list); page 18 include/linux/mm_inline.h list_del(&page->lru); page 25 include/linux/mm_inline.h list_del(&page->lru); page 32 include/linux/mm_inline.h list_del(&page->lru); page 33 include/linux/mm_inline.h if (PageActive(page)) { page 34 include/linux/mm_inline.h __ClearPageActive(page); page 74 include/linux/mm_types.h struct page *first_page; /* Compound tail pages */ page 56 include/linux/mmzone.h return get_pageblock_flags_group(page, PB_migrate, PB_migrate_end); page 508 include/linux/mmzone.h extern struct page *mem_map; page 528 include/linux/mmzone.h struct page *node_mem_map; page 856 include/linux/mmzone.h struct page; page 912 include/linux/mmzone.h static inline struct page *__section_mem_map_addr(struct mem_section *section) page 916 include/linux/mmzone.h return (struct page *)map; page 289 include/linux/mtd/nand.h int page, page 293 include/linux/mtd/nand.h int page); page 387 include/linux/mtd/nand.h void (*erase_cmd)(struct mtd_info *mtd, int page); page 389 include/linux/mtd/nand.h int (*errstat)(struct mtd_info *mtd, struct nand_chip *this, int state, int status, int page); page 391 include/linux/mtd/nand.h const uint8_t *buf, int page, int cached, int raw); page 140 include/linux/net.h struct page; page 186 include/linux/net.h ssize_t (*sendpage) (struct socket *sock, struct page *page, page 40 include/linux/nfs_page.h struct page *wb_page; /* page to read in/write out */ page 252 include/linux/nfs_xdr.h struct page ** pages; page 271 include/linux/nfs_xdr.h struct page ** pages; page 358 include/linux/nfs_xdr.h struct page ** acl_pages; page 365 include/linux/nfs_xdr.h struct page ** acl_pages; page 384 include/linux/nfs_xdr.h struct page ** pages; page 393 include/linux/nfs_xdr.h struct page ** pages; page 399 include/linux/nfs_xdr.h struct page ** pages; page 407 include/linux/nfs_xdr.h struct page ** pages; page 419 include/linux/nfs_xdr.h struct page ** pages; page 460 include/linux/nfs_xdr.h struct page ** pages; page 496 include/linux/nfs_xdr.h struct page ** pages; page 514 include/linux/nfs_xdr.h struct page ** pages; page 563 include/linux/nfs_xdr.h struct page ** pages; page 642 include/linux/nfs_xdr.h struct page ** pages; /* zero-copy data */ page 656 include/linux/nfs_xdr.h struct page ** pages; /* zero-copy data */ page 730 include/linux/nfs_xdr.h struct page *page; page 748 include/linux/nfs_xdr.h struct page **pagevec; page 755 include/linux/nfs_xdr.h struct page *page_array[NFS_PAGEVEC_SIZE]; page 767 include/linux/nfs_xdr.h struct page **pagevec; page 774 include/linux/nfs_xdr.h struct page *page_array[NFS_PAGEVEC_SIZE]; page 800 include/linux/nfs_xdr.h int (*readlink)(struct inode *, struct page *, unsigned int, page 810 include/linux/nfs_xdr.h int (*symlink) (struct inode *, struct dentry *, struct page *, page 815 include/linux/nfs_xdr.h u64, struct page *, unsigned int, int); page 389 include/linux/nfsd/xdr4.h struct page ** pagelist; page 123 include/linux/page-flags.h static inline int Page##uname(struct page *page) \ page 124 include/linux/page-flags.h { return test_bit(PG_##lname, &page->flags); } page 127 include/linux/page-flags.h static inline void SetPage##uname(struct page *page) \ page 128 include/linux/page-flags.h { set_bit(PG_##lname, &page->flags); } page 131 include/linux/page-flags.h static inline void ClearPage##uname(struct page *page) \ page 132 include/linux/page-flags.h { clear_bit(PG_##lname, &page->flags); } page 135 include/linux/page-flags.h static inline void __SetPage##uname(struct page *page) \ page 136 include/linux/page-flags.h { __set_bit(PG_##lname, &page->flags); } page 139 include/linux/page-flags.h static inline void __ClearPage##uname(struct page *page) \ page 140 include/linux/page-flags.h { __clear_bit(PG_##lname, &page->flags); } page 143 include/linux/page-flags.h static inline int TestSetPage##uname(struct page *page) \ page 144 include/linux/page-flags.h { return test_and_set_bit(PG_##lname, &page->flags); } page 147 include/linux/page-flags.h static inline int TestClearPage##uname(struct page *page) \ page 148 include/linux/page-flags.h { return test_and_clear_bit(PG_##lname, &page->flags); } page 158 include/linux/page-flags.h static inline int Page##uname(struct page *page) \ page 164 include/linux/page-flags.h struct page; /* forward declaration */ page 222 include/linux/page-flags.h int ret = test_bit(PG_uptodate, &(page)->flags); page 241 include/linux/page-flags.h __set_bit(PG_uptodate, &(page)->flags); page 247 include/linux/page-flags.h if (!test_and_set_bit(PG_uptodate, &page->flags)) page 248 include/linux/page-flags.h page_clear_dirty(page); page 259 include/linux/page-flags.h set_bit(PG_uptodate, &(page)->flags); page 272 include/linux/page-flags.h test_set_page_writeback(page); page 287 include/linux/page-flags.h return page->flags & ((1L << PG_head) | (1L << PG_tail)); page 314 include/linux/page-flags.h return ((page->flags & PG_head_tail_mask) == PG_head_tail_mask); page 319 include/linux/page-flags.h page->flags |= PG_head_tail_mask; page 324 include/linux/page-flags.h page->flags &= ~PG_head_tail_mask; page 60 include/linux/pageblock-flags.h struct page; page 69 include/linux/pageblock-flags.h get_pageblock_flags_group(page, 0, NR_PAGEBLOCK_BITS-1) page 71 include/linux/pageblock-flags.h set_pageblock_flags_group(page, 0, NR_PAGEBLOCK_BITS-1) page 63 include/linux/pagemap.h #define page_cache_get(page) get_page(page) page 64 include/linux/pagemap.h #define page_cache_release(page) put_page(page) page 128 include/linux/pagemap.h VM_BUG_ON(page_count(page) == 0); page 129 include/linux/pagemap.h atomic_inc(&page->_count); page 132 include/linux/pagemap.h if (unlikely(!get_page_unless_zero(page))) { page 141 include/linux/pagemap.h VM_BUG_ON(PageTail(page)); page 157 include/linux/pagemap.h VM_BUG_ON(page_count(page) == 0); page 158 include/linux/pagemap.h atomic_add(count, &page->_count); page 161 include/linux/pagemap.h if (unlikely(!atomic_add_unless(&page->_count, count, 0))) page 164 include/linux/pagemap.h VM_BUG_ON(PageCompound(page) && page != compound_head(page)); page 171 include/linux/pagemap.h return likely(atomic_cmpxchg(&page->_count, count, 0) == count); page 176 include/linux/pagemap.h VM_BUG_ON(page_count(page) != 0); page 179 include/linux/pagemap.h atomic_set(&page->_count, count); page 183 include/linux/pagemap.h extern struct page *__page_cache_alloc(gfp_t gfp); page 185 include/linux/pagemap.h static inline struct page *__page_cache_alloc(gfp_t gfp) page 191 include/linux/pagemap.h static inline struct page *page_cache_alloc(struct address_space *x) page 196 include/linux/pagemap.h static inline struct page *page_cache_alloc_cold(struct address_space *x) page 203 include/linux/pagemap.h extern struct page * find_get_page(struct address_space *mapping, page 205 include/linux/pagemap.h extern struct page * find_lock_page(struct address_space *mapping, page 207 include/linux/pagemap.h extern struct page * find_or_create_page(struct address_space *mapping, page 216 include/linux/pagemap.h struct page *__grab_cache_page(struct address_space *mapping, pgoff_t index); page 221 include/linux/pagemap.h static inline struct page *grab_cache_page(struct address_space *mapping, page 227 include/linux/pagemap.h extern struct page * grab_cache_page_nowait(struct address_space *mapping, page 229 include/linux/pagemap.h extern struct page * read_cache_page_async(struct address_space *mapping, page 232 include/linux/pagemap.h extern struct page * read_cache_page(struct address_space *mapping, page 238 include/linux/pagemap.h static inline struct page *read_mapping_page_async( page 246 include/linux/pagemap.h static inline struct page *read_mapping_page(struct address_space *mapping, page 258 include/linux/pagemap.h return ((loff_t)page->index) << PAGE_CACHE_SHIFT; page 276 include/linux/pagemap.h set_bit(PG_locked, &page->flags); page 281 include/linux/pagemap.h clear_bit(PG_locked, &page->flags); page 286 include/linux/pagemap.h return !test_and_set_bit(PG_locked, &page->flags); page 295 include/linux/pagemap.h if (!trylock_page(page)) page 296 include/linux/pagemap.h __lock_page(page); page 307 include/linux/pagemap.h if (!trylock_page(page)) page 308 include/linux/pagemap.h return __lock_page_killable(page); page 319 include/linux/pagemap.h if (!trylock_page(page)) page 320 include/linux/pagemap.h __lock_page_nosync(page); page 338 include/linux/pagemap.h if (PageLocked(page)) page 339 include/linux/pagemap.h wait_on_page_bit(page, PG_locked); page 347 include/linux/pagemap.h if (PageWriteback(page)) page 348 include/linux/pagemap.h wait_on_page_bit(page, PG_writeback); page 420 include/linux/pagemap.h set_page_locked(page); page 421 include/linux/pagemap.h error = add_to_page_cache_locked(page, mapping, offset, gfp_mask); page 423 include/linux/pagemap.h clear_page_locked(page); page 14 include/linux/pagevec.h struct page; page 20 include/linux/pagevec.h struct page *pages[PAGEVEC_SIZE]; page 61 include/linux/pagevec.h pvec->pages[pvec->nr++] = page; page 12 include/linux/pid_namespace.h void *page; page 22 include/linux/pipe_fs_i.h struct page *page; page 48 include/linux/pipe_fs_i.h struct page *tmp_page; page 229 include/linux/pktcdvd.h struct page *pages[PACKET_MAX_SIZE / FRAMES_PER_PAGE]; page 278 include/linux/proc_fs.h int (*proc_read)(struct task_struct *task, char *page); page 17 include/linux/quicklist.h void *page; page 39 include/linux/quicklist.h p = q->page; page 41 include/linux/quicklist.h q->page = p[0]; page 61 include/linux/quicklist.h *(void **)p = q->page; page 62 include/linux/quicklist.h q->page = p; page 75 include/linux/quicklist.h __quicklist_free(nr, dtor, page_address(page), page); page 198 include/linux/raid/bitmap.h struct page *page; page 231 include/linux/raid/bitmap.h struct page *sb_page; /* cached copy of the bitmap file superblock */ page 232 include/linux/raid/bitmap.h struct page **filemap; /* list of cache pages for the file */ page 58 include/linux/raid/md_k.h struct page *sb_page; page 60 include/linux/raid/raid1.h struct page *tmppage; page 61 include/linux/raid/raid10.h struct page *tmppage; page 221 include/linux/raid/raid5.h struct page *page; page 373 include/linux/raid/raid5.h struct page *spare_page; /* Used when checking P/Q in raid6 */ page 45 include/linux/relay.h struct page **page_array; /* array of current buffer pages */ page 91 include/linux/rmap.h atomic_inc(&page->_mapcount); page 126 include/linux/rmap.h #define page_referenced(page,l,cnt) TestClearPageReferenced(page) page 63 include/linux/scatterlist.h BUG_ON((unsigned long) page & 0x03); page 68 include/linux/scatterlist.h sg->page_link = page_link | (unsigned long) page; page 88 include/linux/scatterlist.h sg_assign_page(sg, page); page 93 include/linux/scatterlist.h static inline struct page *sg_page(struct scatterlist *sg) page 99 include/linux/scatterlist.h return (struct page *)((sg)->page_link & ~0x3); page 248 include/linux/scatterlist.h struct page *page; /* currently mapped page */ page 18 include/linux/shmem_fs.h struct page *i_indirect; /* top indirect blocks page */ page 133 include/linux/skbuff.h struct page *page; page 965 include/linux/skbuff.h frag->page = page; page 1388 include/linux/skbuff.h extern struct page *__netdev_alloc_page(struct net_device *dev, gfp_t gfp_mask); page 1398 include/linux/skbuff.h static inline struct page *netdev_alloc_page(struct net_device *dev) page 1405 include/linux/skbuff.h __free_page(page); page 1515 include/linux/skbuff.h return page == frag->page && page 37 include/linux/slub_def.h struct page *page; /* The slab from which we are allocating */ page 114 include/linux/smb_fs.h struct page *page; page 51 include/linux/splice.h struct page **pages; /* page map */ page 103 include/linux/sunrpc/gss_api.h struct page **inpages); page 223 include/linux/sunrpc/svc.h struct page * rq_pages[RPCSVC_MAXPAGES]; page 224 include/linux/sunrpc/svc.h struct page * *rq_respages; /* points into rq_pages */ page 310 include/linux/sunrpc/svc.h struct page **pp = (rqstp->rq_respages + page 87 include/linux/sunrpc/svc_rdma.h struct page *pages[RPCSVC_MAXPAGES]; page 210 include/linux/sunrpc/svc_rdma.h struct page *svc_rdma_get_page(void); page 71 include/linux/sunrpc/xdr.h struct page ** pages; /* Array of contiguous pages */ page 73 include/linux/sunrpc/xprt.h struct page **rq_enc_pages; /* scratch pages for use by page 232 include/linux/swap.h extern struct page *lookup_swap_cache(swp_entry_t); page 233 include/linux/swap.h extern struct page *read_swap_cache_async(swp_entry_t, gfp_t, page 235 include/linux/swap.h extern struct page *swapin_readahead(swp_entry_t, gfp_t, page 287 include/linux/swap.h page_cache_release(page) page 308 include/linux/swap.h static inline struct page *swapin_readahead(swp_entry_t swp, gfp_t gfp_mask, page 314 include/linux/swap.h static inline struct page *lookup_swap_cache(swp_entry_t swp) page 82 include/linux/swapops.h BUG_ON(!PageLocked(page)); page 84 include/linux/swapops.h page_to_pfn(page)); page 98 include/linux/swapops.h static inline struct page *migration_entry_to_page(swp_entry_t entry) page 100 include/linux/swapops.h struct page *p = pfn_to_page(swp_offset(entry)); page 256 include/linux/tty_driver.h int (*read_proc)(char *page, char **start, off_t off, page 86 include/linux/videotext.h int page; /* number of requested page (hexadecimal) */ page 31 include/linux/vmalloc.h struct page **pages; page 247 include/linux/vmstat.h __inc_zone_state(page_zone(page), item); page 259 include/linux/vmstat.h __dec_zone_state(page_zone(page), item); page 66 include/media/videobuf-dma-sg.h struct page **pages; page 328 include/net/bluetooth/hci.h __u8 page; page 560 include/net/bluetooth/hci.h __u8 page; page 766 include/net/bluetooth/hci.h __u8 page; page 268 include/net/sock.h struct page *sk_sndmsg_page; page 556 include/net/sock.h int (*sendpage)(struct sock *sk, struct page *page, page 1107 include/net/sock.h page_address(page) + off, page 1112 include/net/sock.h } else if (copy_from_user(page_address(page) + off, from, copy)) page 1214 include/net/sock.h static inline struct page *sk_stream_alloc_page(struct sock *sk) page 1216 include/net/sock.h struct page *page = NULL; page 1218 include/net/sock.h page = alloc_pages(sk->sk_allocation, 0); page 1219 include/net/sock.h if (!page) { page 1223 include/net/sock.h return page; page 944 include/rdma/ib_verbs.h struct page *page, unsigned long offset, page 1662 include/rdma/ib_verbs.h return dev->dma_ops->map_page(dev, page, offset, size, direction); page 1663 include/rdma/ib_verbs.h return dma_map_page(dev->dma_device, page, offset, size, direction); page 81 include/sound/memalloc.h struct page **page_table; /* page table (for vmap/vunmap) */ page 78 include/sound/pcm.h struct page *(*page)(struct snd_pcm_substream *substream, page 997 include/sound/pcm.h struct page *snd_pcm_sgbuf_ops_page(struct snd_pcm_substream *substream, page 96 include/xen/interface/io/fbif.h ((union xenfb_in_event *)((char *)(page) + XENFB_IN_RING_OFFS)) page 98 include/xen/interface/io/fbif.h (XENFB_IN_RING((page))[(idx) % XENFB_IN_RING_LEN]) page 104 include/xen/interface/io/fbif.h ((union xenfb_out_event *)((char *)(page) + XENFB_OUT_RING_OFFS)) page 106 include/xen/interface/io/fbif.h (XENFB_OUT_RING((page))[(idx) % XENFB_OUT_RING_LEN]) page 99 include/xen/interface/io/kbdif.h ((union xenkbd_in_event *)((char *)(page) + XENKBD_IN_RING_OFFS)) page 101 include/xen/interface/io/kbdif.h (XENKBD_IN_RING((page))[(idx) % XENKBD_IN_RING_LEN]) page 107 include/xen/interface/io/kbdif.h ((union xenkbd_out_event *)((char *)(page) + XENKBD_OUT_RING_OFFS)) page 109 include/xen/interface/io/kbdif.h (XENKBD_OUT_RING((page))[(idx) % XENKBD_OUT_RING_LEN]) page 190 init/do_mounts.c char *s = page; page 193 init/do_mounts.c strcpy(page, root_fs_names); page 199 init/do_mounts.c int len = get_filesystem_list(page); page 202 init/do_mounts.c page[len] = '\0'; page 203 init/do_mounts.c for (p = page-1; p; p = next) { page 1481 kernel/cpuset.c return cpulist_scnprintf(page, PAGE_SIZE, mask); page 1492 kernel/cpuset.c return nodelist_scnprintf(page, PAGE_SIZE, mask); page 1503 kernel/cpuset.c char *page; page 1507 kernel/cpuset.c if (!(page = (char *)__get_free_page(GFP_TEMPORARY))) page 1510 kernel/cpuset.c s = page; page 1525 kernel/cpuset.c retval = simple_read_from_buffer(buf, nbytes, ppos, page, s - page); page 1527 kernel/cpuset.c free_page((unsigned long)page); page 116 kernel/dma-coherent.c int page = bitmap_find_free_region(mem->bitmap, mem->size, page 118 kernel/dma-coherent.c if (page >= 0) { page 119 kernel/dma-coherent.c *dma_handle = mem->device_base + (page << PAGE_SHIFT); page 120 kernel/dma-coherent.c *ret = mem->virt_base + (page << PAGE_SHIFT); page 148 kernel/dma-coherent.c int page = (vaddr - mem->virt_base) >> PAGE_SHIFT; page 150 kernel/dma-coherent.c bitmap_release_region(mem->bitmap, page, order); page 184 kernel/exec_domain.c len += sprintf(page + len, "%d-%d\t%-16s\t[%s]\n", page 188 kernel/futex.c struct page *page; page 260 kernel/futex.c err = get_user_pages(current, mm, address, 1, 0, 0, &page, NULL); page 263 kernel/futex.c page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); page 264 kernel/futex.c put_page(page); page 136 kernel/irq/proc.c return sprintf(page, "count %u\n" page 112 kernel/kexec.c static struct page *kimage_alloc_page(struct kimage *image, page 354 kernel/kexec.c static struct page *kimage_alloc_pages(gfp_t gfp_mask, unsigned int order) page 356 kernel/kexec.c struct page *pages; page 375 kernel/kexec.c order = page_private(page); page 378 kernel/kexec.c ClearPageReserved(page + i); page 379 kernel/kexec.c __free_pages(page, order); page 387 kernel/kexec.c struct page *page; page 389 kernel/kexec.c page = list_entry(pos, struct page, lru); page 390 kernel/kexec.c list_del(&page->lru); page 391 kernel/kexec.c kimage_free_pages(page); page 395 kernel/kexec.c static struct page *kimage_alloc_normal_control_pages(struct kimage *image, page 412 kernel/kexec.c struct page *pages; page 461 kernel/kexec.c static struct page *kimage_alloc_crash_control_pages(struct kimage *image, page 486 kernel/kexec.c struct page *pages; page 525 kernel/kexec.c struct page *kimage_alloc_control_pages(struct kimage *image, page 528 kernel/kexec.c struct page *pages = NULL; page 549 kernel/kexec.c struct page *page; page 551 kernel/kexec.c page = kimage_alloc_page(image, GFP_KERNEL, KIMAGE_NO_DEST); page 552 kernel/kexec.c if (!page) page 555 kernel/kexec.c ind_page = page_address(page); page 586 kernel/kexec.c page &= PAGE_MASK; page 587 kernel/kexec.c result = kimage_add_entry(image, page | IND_SOURCE); page 619 kernel/kexec.c struct page *page; page 621 kernel/kexec.c page = pfn_to_page(entry >> PAGE_SHIFT); page 622 kernel/kexec.c kimage_free_pages(page); page 669 kernel/kexec.c if (page == destination) page 678 kernel/kexec.c static struct page *kimage_alloc_page(struct kimage *image, page 700 kernel/kexec.c struct page *page; page 707 kernel/kexec.c list_for_each_entry(page, &image->dest_pages, lru) { page 708 kernel/kexec.c addr = page_to_pfn(page) << PAGE_SHIFT; page 710 kernel/kexec.c list_del(&page->lru); page 711 kernel/kexec.c return page; page 714 kernel/kexec.c page = NULL; page 719 kernel/kexec.c page = kimage_alloc_pages(gfp_mask, 0); page 720 kernel/kexec.c if (!page) page 723 kernel/kexec.c if (page_to_pfn(page) > page 725 kernel/kexec.c list_add(&page->lru, &image->unuseable_pages); page 728 kernel/kexec.c addr = page_to_pfn(page) << PAGE_SHIFT; page 748 kernel/kexec.c struct page *old_page; page 752 kernel/kexec.c copy_highpage(page, old_page); page 765 kernel/kexec.c page = old_page; page 772 kernel/kexec.c list_add(&page->lru, &image->dest_pages); page 776 kernel/kexec.c return page; page 798 kernel/kexec.c struct page *page; page 802 kernel/kexec.c page = kimage_alloc_page(image, GFP_HIGHUSER, maddr); page 803 kernel/kexec.c if (!page) { page 807 kernel/kexec.c result = kimage_add_page(image, page_to_pfn(page) page 812 kernel/kexec.c ptr = kmap(page); page 825 kernel/kexec.c kunmap(page); page 857 kernel/kexec.c struct page *page; page 861 kernel/kexec.c page = pfn_to_page(maddr >> PAGE_SHIFT); page 862 kernel/kexec.c if (!page) { page 866 kernel/kexec.c ptr = kmap(page); page 879 kernel/kexec.c kexec_flush_icache_page(page); page 880 kernel/kexec.c kunmap(page); page 1385 kernel/kexec.c VMCOREINFO_STRUCT_SIZE(page); page 1391 kernel/kexec.c VMCOREINFO_OFFSET(page, flags); page 1392 kernel/kexec.c VMCOREINFO_OFFSET(page, _count); page 1393 kernel/kexec.c VMCOREINFO_OFFSET(page, mapping); page 1394 kernel/kexec.c VMCOREINFO_OFFSET(page, lru); page 63 kernel/pid.c find_next_zero_bit((map)->page, BITS_PER_PAGE, off) page 121 kernel/pid.c clear_bit(offset, map->page); page 137 kernel/pid.c if (unlikely(!map->page)) { page 138 kernel/pid.c void *page = kzalloc(PAGE_SIZE, GFP_KERNEL); page 144 kernel/pid.c if (map->page) page 145 kernel/pid.c kfree(page); page 147 kernel/pid.c map->page = page; page 149 kernel/pid.c if (unlikely(!map->page)) page 154 kernel/pid.c if (!test_and_set_bit(offset, map->page)) { page 194 kernel/pid.c if (unlikely(!map->page)) page 196 kernel/pid.c offset = find_next_bit((map)->page, BITS_PER_PAGE, offset); page 523 kernel/pid.c init_pid_ns.pidmap[0].page = kzalloc(PAGE_SIZE, GFP_KERNEL); page 525 kernel/pid.c set_bit(0, init_pid_ns.pidmap[0].page); page 79 kernel/pid_namespace.c ns->pidmap[0].page = kzalloc(PAGE_SIZE, GFP_KERNEL); page 80 kernel/pid_namespace.c if (!ns->pidmap[0].page) page 90 kernel/pid_namespace.c set_bit(0, ns->pidmap[0].page); page 99 kernel/pid_namespace.c kfree(ns->pidmap[0].page); page 111 kernel/pid_namespace.c kfree(ns->pidmap[i].page); page 92 kernel/power/snapshot.c static struct page *alloc_image_page(gfp_t gfp_mask) page 94 kernel/power/snapshot.c struct page *page; page 96 kernel/power/snapshot.c page = alloc_page(gfp_mask); page 97 kernel/power/snapshot.c if (page) { page 98 kernel/power/snapshot.c swsusp_set_page_forbidden(page); page 99 kernel/power/snapshot.c swsusp_set_page_free(page); page 101 kernel/power/snapshot.c return page; page 111 kernel/power/snapshot.c struct page *page; page 115 kernel/power/snapshot.c page = virt_to_page(addr); page 117 kernel/power/snapshot.c swsusp_unset_page_forbidden(page); page 119 kernel/power/snapshot.c swsusp_unset_page_free(page); page 121 kernel/power/snapshot.c __free_page(page); page 633 kernel/power/snapshot.c memory_bm_set_bit(free_pages_map, page_to_pfn(page)); page 639 kernel/power/snapshot.c memory_bm_test_bit(free_pages_map, page_to_pfn(page)) : 0; page 645 kernel/power/snapshot.c memory_bm_clear_bit(free_pages_map, page_to_pfn(page)); page 651 kernel/power/snapshot.c memory_bm_set_bit(forbidden_pages_map, page_to_pfn(page)); page 657 kernel/power/snapshot.c memory_bm_test_bit(forbidden_pages_map, page_to_pfn(page)) : 0; page 663 kernel/power/snapshot.c memory_bm_clear_bit(forbidden_pages_map, page_to_pfn(page)); page 812 kernel/power/snapshot.c static struct page *saveable_highmem_page(unsigned long pfn) page 814 kernel/power/snapshot.c struct page *page; page 819 kernel/power/snapshot.c page = pfn_to_page(pfn); page 821 kernel/power/snapshot.c BUG_ON(!PageHighMem(page)); page 823 kernel/power/snapshot.c if (swsusp_page_is_forbidden(page) || swsusp_page_is_free(page) || page 824 kernel/power/snapshot.c PageReserved(page)) page 827 kernel/power/snapshot.c return page; page 867 kernel/power/snapshot.c static struct page *saveable_page(unsigned long pfn) page 869 kernel/power/snapshot.c struct page *page; page 874 kernel/power/snapshot.c page = pfn_to_page(pfn); page 876 kernel/power/snapshot.c BUG_ON(PageHighMem(page)); page 878 kernel/power/snapshot.c if (swsusp_page_is_forbidden(page) || swsusp_page_is_free(page)) page 881 kernel/power/snapshot.c if (PageReserved(page) page 882 kernel/power/snapshot.c && (!kernel_page_present(page) || pfn_is_nosave(pfn))) page 885 kernel/power/snapshot.c return page; page 943 kernel/power/snapshot.c static inline struct page * page 952 kernel/power/snapshot.c struct page *s_page, *d_page; page 1033 kernel/power/snapshot.c struct page *page = pfn_to_page(pfn); page 1035 kernel/power/snapshot.c if (swsusp_page_is_forbidden(page) && page 1036 kernel/power/snapshot.c swsusp_page_is_free(page)) { page 1037 kernel/power/snapshot.c swsusp_unset_page_forbidden(page); page 1038 kernel/power/snapshot.c swsusp_unset_page_free(page); page 1039 kernel/power/snapshot.c __free_page(page); page 1122 kernel/power/snapshot.c struct page *page; page 1124 kernel/power/snapshot.c page = alloc_image_page(__GFP_HIGHMEM); page 1125 kernel/power/snapshot.c memory_bm_set_bit(bm, page_to_pfn(page)); page 1170 kernel/power/snapshot.c struct page *page = alloc_image_page(GFP_ATOMIC | __GFP_COLD); page 1172 kernel/power/snapshot.c if (!page) page 1175 kernel/power/snapshot.c memory_bm_set_bit(copy_bm, page_to_pfn(page)); page 1345 kernel/power/snapshot.c struct page *page; page 1347 kernel/power/snapshot.c page = pfn_to_page(memory_bm_next_pfn(©_bm)); page 1348 kernel/power/snapshot.c if (PageHighMem(page)) { page 1355 kernel/power/snapshot.c kaddr = kmap_atomic(page, KM_USER0); page 1360 kernel/power/snapshot.c handle->buffer = page_address(page); page 1487 kernel/power/snapshot.c struct page *copy_page; /* data is here now */ page 1488 kernel/power/snapshot.c struct page *orig_page; /* data was here before the suspend */ page 1556 kernel/power/snapshot.c struct page *page; page 1558 kernel/power/snapshot.c page = alloc_page(__GFP_HIGHMEM); page 1559 kernel/power/snapshot.c if (!swsusp_page_is_free(page)) { page 1561 kernel/power/snapshot.c memory_bm_set_bit(bm, page_to_pfn(page)); page 1565 kernel/power/snapshot.c swsusp_set_page_forbidden(page); page 1566 kernel/power/snapshot.c swsusp_set_page_free(page); page 1590 kernel/power/snapshot.c static struct page *last_highmem_page; page 1598 kernel/power/snapshot.c if (swsusp_page_is_forbidden(page) && swsusp_page_is_free(page)) { page 1602 kernel/power/snapshot.c last_highmem_page = page; page 1613 kernel/power/snapshot.c pbe->orig_page = page; page 1615 kernel/power/snapshot.c struct page *tmp; page 1791 kernel/power/snapshot.c struct page *page = pfn_to_page(memory_bm_next_pfn(bm)); page 1793 kernel/power/snapshot.c if (PageHighMem(page)) page 1794 kernel/power/snapshot.c return get_highmem_page_buffer(page, ca); page 1796 kernel/power/snapshot.c if (swsusp_page_is_forbidden(page) && swsusp_page_is_free(page)) page 1800 kernel/power/snapshot.c return page_address(page); page 1810 kernel/power/snapshot.c pbe->orig_address = page_address(page); page 72 kernel/power/swap.c if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) { page 79 kernel/power/swap.c lock_page(page); page 84 kernel/power/swap.c wait_on_page_locked(page); page 90 kernel/power/swap.c get_page(page); /* These pages are freed later */ page 121 kernel/power/swap.c struct page *page; page 124 kernel/power/swap.c page = bio->bi_io_vec[0].bv_page; page 125 kernel/power/swap.c wait_on_page_locked(page); page 126 kernel/power/swap.c if (!PageUptodate(page) || PageError(page)) page 128 kernel/power/swap.c put_page(page); page 337 kernel/profile.c struct page *page; page 345 kernel/profile.c page = alloc_pages_node(node, page 348 kernel/profile.c if (!page) page 350 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[1] = page_address(page); page 353 kernel/profile.c page = alloc_pages_node(node, page 356 kernel/profile.c if (!page) page 358 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[0] = page_address(page); page 362 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[1]); page 364 kernel/profile.c __free_page(page); page 376 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[0]); page 378 kernel/profile.c __free_page(page); page 381 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[1]); page 383 kernel/profile.c __free_page(page); page 424 kernel/profile.c int len = cpumask_scnprintf(page, count, *(cpumask_t *)data); page 427 kernel/profile.c len += sprintf(page + len, "\n"); page 536 kernel/profile.c struct page *page; page 538 kernel/profile.c page = alloc_pages_node(node, page 541 kernel/profile.c if (!page) page 544 kernel/profile.c = (struct profile_hit *)page_address(page); page 545 kernel/profile.c page = alloc_pages_node(node, page 548 kernel/profile.c if (!page) page 551 kernel/profile.c = (struct profile_hit *)page_address(page); page 559 kernel/profile.c struct page *page; page 562 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[0]); page 564 kernel/profile.c __free_page(page); page 567 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[1]); page 569 kernel/profile.c __free_page(page); page 223 kernel/rcutorture.c int (*stats)(char *page); page 478 kernel/rcutorture.c cnt += sprintf(&page[cnt], "%s%s per-CPU(idx=%d):", page 481 kernel/rcutorture.c cnt += sprintf(&page[cnt], " %d(%d,%d)", cpu, page 485 kernel/rcutorture.c cnt += sprintf(&page[cnt], "\n"); page 764 kernel/rcutorture.c cnt += sprintf(&page[cnt], "%s%s ", torture_type, TORTURE_FLAG); page 765 kernel/rcutorture.c cnt += sprintf(&page[cnt], page 777 kernel/rcutorture.c cnt += sprintf(&page[cnt], " !!!"); page 778 kernel/rcutorture.c cnt += sprintf(&page[cnt], "\n%s%s ", torture_type, TORTURE_FLAG); page 780 kernel/rcutorture.c cnt += sprintf(&page[cnt], "!!! "); page 784 kernel/rcutorture.c cnt += sprintf(&page[cnt], "Reader Pipe: "); page 786 kernel/rcutorture.c cnt += sprintf(&page[cnt], " %ld", pipesummary[i]); page 787 kernel/rcutorture.c cnt += sprintf(&page[cnt], "\n%s%s ", torture_type, TORTURE_FLAG); page 788 kernel/rcutorture.c cnt += sprintf(&page[cnt], "Reader Batch: "); page 790 kernel/rcutorture.c cnt += sprintf(&page[cnt], " %ld", batchsummary[i]); page 791 kernel/rcutorture.c cnt += sprintf(&page[cnt], "\n%s%s ", torture_type, TORTURE_FLAG); page 792 kernel/rcutorture.c cnt += sprintf(&page[cnt], "Free-Block Circulation: "); page 794 kernel/rcutorture.c cnt += sprintf(&page[cnt], " %d", page 797 kernel/rcutorture.c cnt += sprintf(&page[cnt], "\n"); page 799 kernel/rcutorture.c cnt += cur_ops->stats(&page[cnt]); page 44 kernel/relay.c struct page *page; page 51 kernel/relay.c page = vmalloc_to_page(buf->start + (pgoff << PAGE_SHIFT)); page 52 kernel/relay.c if (!page) page 54 kernel/relay.c get_page(page); page 55 kernel/relay.c vmf->page = page; page 71 kernel/relay.c static struct page **relay_alloc_page_array(unsigned int n_pages) page 73 kernel/relay.c struct page **array; page 74 kernel/relay.c size_t pa_size = n_pages * sizeof(struct page *); page 1196 kernel/relay.c rbuf = (struct rchan_buf *)page_private(buf->page); page 1233 kernel/relay.c struct page *pages[PIPE_BUFFERS]; page 7878 kernel/sched.c return sprintf(page, "%u\n", sched_mc_power_savings); page 7894 kernel/sched.c return sprintf(page, "%u\n", sched_smt_power_savings); page 284 kernel/trace/trace.c struct page *page, *tmp; page 289 kernel/trace/trace.c list_for_each_entry_safe(page, tmp, &data->trace_pages, lru) { page 290 kernel/trace/trace.c CHECK_COND(page->lru.next->prev != &page->lru); page 291 kernel/trace/trace.c CHECK_COND(page->lru.prev->next != &page->lru); page 306 kernel/trace/trace.c struct page *page; page 311 kernel/trace/trace.c page = list_entry(data->trace_pages.next, struct page, lru); page 312 kernel/trace/trace.c BUG_ON(&page->lru == &data->trace_pages); page 314 kernel/trace/trace.c return page_address(page); page 766 kernel/trace/trace.c struct page *page; page 768 kernel/trace/trace.c page = virt_to_page(addr); page 770 kernel/trace/trace.c next = trace_next_list(data, &page->lru); page 771 kernel/trace/trace.c page = list_entry(next, struct page, lru); page 773 kernel/trace/trace.c return page_address(page); page 1080 kernel/trace/trace.c struct page *page; page 1092 kernel/trace/trace.c page = virt_to_page(data->trace_tail); page 1093 kernel/trace/trace.c iter->next_page[cpu] = &page->lru; page 1097 kernel/trace/trace.c page = list_entry(iter->next_page[cpu], struct page, lru); page 1098 kernel/trace/trace.c BUG_ON(&data->trace_pages == &page->lru); page 1100 kernel/trace/trace.c array = page_address(page); page 2936 kernel/trace/trace.c struct page *page, *tmp; page 2952 kernel/trace/trace.c page = virt_to_page(array); page 2953 kernel/trace/trace.c list_add(&page->lru, &pages); page 2964 kernel/trace/trace.c page = virt_to_page(array); page 2965 kernel/trace/trace.c list_add(&page->lru, &pages); page 2972 kernel/trace/trace.c page = list_entry(pages.next, struct page, lru); page 2973 kernel/trace/trace.c list_del_init(&page->lru); page 2974 kernel/trace/trace.c list_add_tail(&page->lru, &data->trace_pages); page 2975 kernel/trace/trace.c ClearPageLRU(page); page 2979 kernel/trace/trace.c page = list_entry(pages.next, struct page, lru); page 2980 kernel/trace/trace.c list_del_init(&page->lru); page 2981 kernel/trace/trace.c list_add_tail(&page->lru, &data->trace_pages); page 2982 kernel/trace/trace.c SetPageLRU(page); page 2991 kernel/trace/trace.c list_for_each_entry_safe(page, tmp, &pages, lru) { page 2992 kernel/trace/trace.c list_del_init(&page->lru); page 2993 kernel/trace/trace.c __free_page(page); page 3001 kernel/trace/trace.c struct page *page; page 3017 kernel/trace/trace.c page = list_entry(p, struct page, lru); page 3018 kernel/trace/trace.c ClearPageLRU(page); page 3019 kernel/trace/trace.c list_del(&page->lru); page 3022 kernel/trace/trace.c __free_page(page); page 3036 kernel/trace/trace.c page = list_entry(p, struct page, lru); page 3037 kernel/trace/trace.c ClearPageLRU(page); page 3038 kernel/trace/trace.c list_del(&page->lru); page 3039 kernel/trace/trace.c __free_page(page); page 3053 kernel/trace/trace.c struct page *page; page 3076 kernel/trace/trace.c page = virt_to_page(array); page 3077 kernel/trace/trace.c list_add(&page->lru, &data->trace_pages); page 3079 kernel/trace/trace.c ClearPageLRU(page); page 3094 kernel/trace/trace.c page = virt_to_page(array); page 3095 kernel/trace/trace.c list_add(&page->lru, &max_tr.data[i]->trace_pages); page 3096 kernel/trace/trace.c SetPageLRU(page); page 3133 kernel/trace/trace.c struct page *page, *tmp; page 3137 kernel/trace/trace.c list_for_each_entry_safe(page, tmp, page 3139 kernel/trace/trace.c list_del_init(&page->lru); page 3140 kernel/trace/trace.c __free_page(page); page 3147 kernel/trace/trace.c list_for_each_entry_safe(page, tmp, page 3149 kernel/trace/trace.c list_del_init(&page->lru); page 3150 kernel/trace/trace.c __free_page(page); page 23 kernel/trace/trace_selftest.c struct page *page; page 28 kernel/trace/trace_selftest.c page = list_entry(data->trace_pages.next, struct page, lru); page 29 kernel/trace/trace_selftest.c entries = page_address(page); page 51 kernel/trace/trace_selftest.c page = virt_to_page(entries); page 52 kernel/trace/trace_selftest.c if (page->lru.next == &data->trace_pages) { page 58 kernel/trace/trace_selftest.c page = list_entry(page->lru.next, struct page, lru); page 59 kernel/trace/trace_selftest.c entries = page_address(page); page 65 kernel/trace/trace_selftest.c page = virt_to_page(entries); page 66 kernel/trace/trace_selftest.c if (page->lru.next != &data->trace_pages) { page 359 lib/scatterlist.c miter->page = nth_page(sg_page(miter->__sg), off >> PAGE_SHIFT); page 365 lib/scatterlist.c miter->addr = kmap_atomic(miter->page, KM_BIO_SRC_IRQ) + off; page 367 lib/scatterlist.c miter->addr = kmap(miter->page) + off; page 400 lib/scatterlist.c miter->page = NULL; page 440 lib/scatterlist.c flush_kernel_dcache_page(miter.page); page 26 lib/show_mem.c struct page *page; page 35 lib/show_mem.c page = pfn_to_page(pfn); page 37 lib/show_mem.c if (PageHighMem(page)) page 40 lib/show_mem.c if (PageReserved(page)) page 42 lib/show_mem.c else if (page_count(page) == 1) page 44 lib/show_mem.c else if (page_count(page) > 1) page 45 lib/show_mem.c shared += page_count(page) - 1; page 108 mm/backing-dev.c struct device_attribute *attr, char *page) \ page 112 mm/backing-dev.c return snprintf(page, PAGE_SIZE-1, "%lld\n", (long long)expr); \ page 148 mm/bootmem.c struct page *page; page 183 mm/bootmem.c page = pfn_to_page(start + off); page 184 mm/bootmem.c __free_pages_bootmem(page, 0); page 194 mm/bootmem.c page = virt_to_page(bdata->node_bootmem_map); page 199 mm/bootmem.c __free_pages_bootmem(page++, 0); page 181 mm/bounce.c struct page *page; page 187 mm/bounce.c page = from->bv_page; page 192 mm/bounce.c if (page_to_pfn(page) <= q->bounce_pfn) page 74 mm/dmapool.c struct dma_page *page; page 89 mm/dmapool.c list_for_each_entry(page, &pool->page_list, page_list) { page 91 mm/dmapool.c blocks += page->in_use; page 208 mm/dmapool.c *(int *)(page->vaddr + offset) = next; page 215 mm/dmapool.c struct dma_page *page; page 217 mm/dmapool.c page = kmalloc(sizeof(*page), mem_flags); page 218 mm/dmapool.c if (!page) page 220 mm/dmapool.c page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, page 221 mm/dmapool.c &page->dma, mem_flags); page 222 mm/dmapool.c if (page->vaddr) { page 224 mm/dmapool.c memset(page->vaddr, POOL_POISON_FREED, pool->allocation); page 226 mm/dmapool.c pool_initialise_page(pool, page); page 227 mm/dmapool.c list_add(&page->page_list, &pool->page_list); page 228 mm/dmapool.c page->in_use = 0; page 229 mm/dmapool.c page->offset = 0; page 231 mm/dmapool.c kfree(page); page 232 mm/dmapool.c page = NULL; page 234 mm/dmapool.c return page; page 239 mm/dmapool.c return page->in_use != 0; page 244 mm/dmapool.c dma_addr_t dma = page->dma; page 247 mm/dmapool.c memset(page->vaddr, POOL_POISON_FREED, pool->allocation); page 249 mm/dmapool.c dma_free_coherent(pool->dev, pool->allocation, page->vaddr, dma); page 250 mm/dmapool.c list_del(&page->page_list); page 251 mm/dmapool.c kfree(page); page 271 mm/dmapool.c struct dma_page *page; page 272 mm/dmapool.c page = list_entry(pool->page_list.next, page 274 mm/dmapool.c if (is_page_busy(page)) { page 278 mm/dmapool.c pool->name, page->vaddr); page 282 mm/dmapool.c pool->name, page->vaddr); page 284 mm/dmapool.c list_del(&page->page_list); page 285 mm/dmapool.c kfree(page); page 287 mm/dmapool.c pool_free_page(pool, page); page 308 mm/dmapool.c struct dma_page *page; page 314 mm/dmapool.c list_for_each_entry(page, &pool->page_list, page_list) { page 315 mm/dmapool.c if (page->offset < pool->allocation) page 318 mm/dmapool.c page = pool_alloc_page(pool, GFP_ATOMIC); page 319 mm/dmapool.c if (!page) { page 338 mm/dmapool.c page->in_use++; page 339 mm/dmapool.c offset = page->offset; page 340 mm/dmapool.c page->offset = *(int *)(page->vaddr + offset); page 341 mm/dmapool.c retval = offset + page->vaddr; page 342 mm/dmapool.c *handle = offset + page->dma; page 355 mm/dmapool.c struct dma_page *page; page 358 mm/dmapool.c list_for_each_entry(page, &pool->page_list, page_list) { page 359 mm/dmapool.c if (dma < page->dma) page 361 mm/dmapool.c if (dma < (page->dma + pool->allocation)) page 364 mm/dmapool.c page = NULL; page 367 mm/dmapool.c return page; page 381 mm/dmapool.c struct dma_page *page; page 385 mm/dmapool.c page = pool_find_page(pool, dma); page 386 mm/dmapool.c if (!page) { page 397 mm/dmapool.c offset = vaddr - page->vaddr; page 399 mm/dmapool.c if ((dma - page->dma) != offset) { page 411 mm/dmapool.c unsigned int chain = page->offset; page 414 mm/dmapool.c chain = *(int *)(page->vaddr + chain); page 432 mm/dmapool.c page->in_use--; page 433 mm/dmapool.c *(int *)vaddr = page->offset; page 434 mm/dmapool.c page->offset = offset; page 116 mm/filemap.c struct address_space *mapping = page->mapping; page 118 mm/filemap.c mem_cgroup_uncharge_cache_page(page); page 119 mm/filemap.c radix_tree_delete(&mapping->page_tree, page->index); page 120 mm/filemap.c page->mapping = NULL; page 122 mm/filemap.c __dec_zone_page_state(page, NR_FILE_PAGES); page 123 mm/filemap.c BUG_ON(page_mapped(page)); page 132 mm/filemap.c if (PageDirty(page) && mapping_cap_account_dirty(mapping)) { page 133 mm/filemap.c dec_zone_page_state(page, NR_FILE_DIRTY); page 140 mm/filemap.c struct address_space *mapping = page->mapping; page 142 mm/filemap.c BUG_ON(!PageLocked(page)); page 145 mm/filemap.c __remove_from_page_cache(page); page 152 mm/filemap.c struct page *page; page 154 mm/filemap.c page = container_of((unsigned long *)word, struct page, flags); page 178 mm/filemap.c mapping = page_mapping(page); page 180 mm/filemap.c mapping->a_ops->sync_page(page); page 285 mm/filemap.c struct page *page = pvec.pages[i]; page 288 mm/filemap.c if (page->index > end) page 291 mm/filemap.c wait_on_page_writeback(page); page 292 mm/filemap.c if (PageError(page)) page 459 mm/filemap.c VM_BUG_ON(!PageLocked(page)); page 461 mm/filemap.c error = mem_cgroup_cache_charge(page, current->mm, page 468 mm/filemap.c page_cache_get(page); page 469 mm/filemap.c page->mapping = mapping; page 470 mm/filemap.c page->index = offset; page 473 mm/filemap.c error = radix_tree_insert(&mapping->page_tree, offset, page); page 476 mm/filemap.c __inc_zone_page_state(page, NR_FILE_PAGES); page 478 mm/filemap.c page->mapping = NULL; page 479 mm/filemap.c mem_cgroup_uncharge_cache_page(page); page 480 mm/filemap.c page_cache_release(page); page 486 mm/filemap.c mem_cgroup_uncharge_cache_page(page); page 495 mm/filemap.c int ret = add_to_page_cache(page, mapping, offset, gfp_mask); page 497 mm/filemap.c lru_cache_add(page); page 502 mm/filemap.c struct page *__page_cache_alloc(gfp_t gfp) page 531 mm/filemap.c const struct zone *zone = page_zone(page); page 533 mm/filemap.c return &zone->wait_table[hash_ptr(page, zone->wait_table_bits)]; page 538 mm/filemap.c __wake_up_bit(page_waitqueue(page), &page->flags, bit); page 543 mm/filemap.c DEFINE_WAIT_BIT(wait, &page->flags, bit_nr); page 545 mm/filemap.c if (test_bit(bit_nr, &page->flags)) page 546 mm/filemap.c __wait_on_bit(page_waitqueue(page), &wait, sync_page, page 568 mm/filemap.c if (!test_and_clear_bit(PG_locked, &page->flags)) page 571 mm/filemap.c wake_up_page(page, PG_locked); page 581 mm/filemap.c if (TestClearPageReclaim(page)) page 582 mm/filemap.c rotate_reclaimable_page(page); page 584 mm/filemap.c if (!test_clear_page_writeback(page)) page 588 mm/filemap.c wake_up_page(page, PG_writeback); page 603 mm/filemap.c DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); page 605 mm/filemap.c __wait_on_bit_lock(page_waitqueue(page), &wait, sync_page, page 612 mm/filemap.c DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); page 614 mm/filemap.c return __wait_on_bit_lock(page_waitqueue(page), &wait, page 627 mm/filemap.c DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); page 628 mm/filemap.c __wait_on_bit_lock(page_waitqueue(page), &wait, __sleep_on_page_lock, page 640 mm/filemap.c struct page *find_get_page(struct address_space *mapping, pgoff_t offset) page 643 mm/filemap.c struct page *page; page 647 mm/filemap.c page = NULL; page 650 mm/filemap.c page = radix_tree_deref_slot(pagep); page 651 mm/filemap.c if (unlikely(!page || page == RADIX_TREE_RETRY)) page 654 mm/filemap.c if (!page_cache_get_speculative(page)) page 662 mm/filemap.c if (unlikely(page != *pagep)) { page 663 mm/filemap.c page_cache_release(page); page 669 mm/filemap.c return page; page 683 mm/filemap.c struct page *find_lock_page(struct address_space *mapping, pgoff_t offset) page 685 mm/filemap.c struct page *page; page 688 mm/filemap.c page = find_get_page(mapping, offset); page 689 mm/filemap.c if (page) { page 690 mm/filemap.c lock_page(page); page 692 mm/filemap.c if (unlikely(page->mapping != mapping)) { page 693 mm/filemap.c unlock_page(page); page 694 mm/filemap.c page_cache_release(page); page 697 mm/filemap.c VM_BUG_ON(page->index != offset); page 699 mm/filemap.c return page; page 720 mm/filemap.c struct page *find_or_create_page(struct address_space *mapping, page 723 mm/filemap.c struct page *page; page 726 mm/filemap.c page = find_lock_page(mapping, index); page 727 mm/filemap.c if (!page) { page 728 mm/filemap.c page = __page_cache_alloc(gfp_mask); page 729 mm/filemap.c if (!page) page 731 mm/filemap.c err = add_to_page_cache_lru(page, mapping, index, gfp_mask); page 733 mm/filemap.c page_cache_release(page); page 734 mm/filemap.c page = NULL; page 739 mm/filemap.c return page; page 772 mm/filemap.c struct page *page; page 774 mm/filemap.c page = radix_tree_deref_slot((void **)pages[i]); page 775 mm/filemap.c if (unlikely(!page)) page 781 mm/filemap.c if (unlikely(page == RADIX_TREE_RETRY)) page 784 mm/filemap.c if (!page_cache_get_speculative(page)) page 788 mm/filemap.c if (unlikely(page != *((void **)pages[i]))) { page 789 mm/filemap.c page_cache_release(page); page 793 mm/filemap.c pages[ret] = page; page 825 mm/filemap.c struct page *page; page 827 mm/filemap.c page = radix_tree_deref_slot((void **)pages[i]); page 828 mm/filemap.c if (unlikely(!page)) page 834 mm/filemap.c if (unlikely(page == RADIX_TREE_RETRY)) page 837 mm/filemap.c if (page->mapping == NULL || page->index != index) page 840 mm/filemap.c if (!page_cache_get_speculative(page)) page 844 mm/filemap.c if (unlikely(page != *((void **)pages[i]))) { page 845 mm/filemap.c page_cache_release(page); page 849 mm/filemap.c pages[ret] = page; page 882 mm/filemap.c struct page *page; page 884 mm/filemap.c page = radix_tree_deref_slot((void **)pages[i]); page 885 mm/filemap.c if (unlikely(!page)) page 891 mm/filemap.c if (unlikely(page == RADIX_TREE_RETRY)) page 894 mm/filemap.c if (!page_cache_get_speculative(page)) page 898 mm/filemap.c if (unlikely(page != *((void **)pages[i]))) { page 899 mm/filemap.c page_cache_release(page); page 903 mm/filemap.c pages[ret] = page; page 928 mm/filemap.c struct page * page 931 mm/filemap.c struct page *page = find_get_page(mapping, index); page 933 mm/filemap.c if (page) { page 934 mm/filemap.c if (trylock_page(page)) page 935 mm/filemap.c return page; page 936 mm/filemap.c page_cache_release(page); page 939 mm/filemap.c page = __page_cache_alloc(mapping_gfp_mask(mapping) & ~__GFP_FS); page 940 mm/filemap.c if (page && add_to_page_cache_lru(page, mapping, index, GFP_KERNEL)) { page 941 mm/filemap.c page_cache_release(page); page 942 mm/filemap.c page = NULL; page 944 mm/filemap.c return page; page 1005 mm/filemap.c struct page *page; page 1012 mm/filemap.c page = find_get_page(mapping, index); page 1013 mm/filemap.c if (!page) { page 1017 mm/filemap.c page = find_get_page(mapping, index); page 1018 mm/filemap.c if (unlikely(page == NULL)) page 1021 mm/filemap.c if (PageReadahead(page)) { page 1023 mm/filemap.c ra, filp, page, page 1026 mm/filemap.c if (!PageUptodate(page)) { page 1030 mm/filemap.c if (!trylock_page(page)) page 1032 mm/filemap.c if (!mapping->a_ops->is_partially_uptodate(page, page 1035 mm/filemap.c unlock_page(page); page 1050 mm/filemap.c page_cache_release(page); page 1059 mm/filemap.c page_cache_release(page); page 1070 mm/filemap.c flush_dcache_page(page); page 1077 mm/filemap.c mark_page_accessed(page); page 1090 mm/filemap.c ret = actor(desc, page, offset, nr); page 1096 mm/filemap.c page_cache_release(page); page 1103 mm/filemap.c if (lock_page_killable(page)) page 1108 mm/filemap.c if (!page->mapping) { page 1109 mm/filemap.c unlock_page(page); page 1110 mm/filemap.c page_cache_release(page); page 1115 mm/filemap.c if (PageUptodate(page)) { page 1116 mm/filemap.c unlock_page(page); page 1122 mm/filemap.c error = mapping->a_ops->readpage(filp, page); page 1126 mm/filemap.c page_cache_release(page); page 1132 mm/filemap.c if (!PageUptodate(page)) { page 1133 mm/filemap.c if (lock_page_killable(page)) page 1135 mm/filemap.c if (!PageUptodate(page)) { page 1136 mm/filemap.c if (page->mapping == NULL) { page 1140 mm/filemap.c unlock_page(page); page 1141 mm/filemap.c page_cache_release(page); page 1144 mm/filemap.c unlock_page(page); page 1148 mm/filemap.c unlock_page(page); page 1158 mm/filemap.c page_cache_release(page); page 1166 mm/filemap.c page = page_cache_alloc_cold(mapping); page 1167 mm/filemap.c if (!page) { page 1171 mm/filemap.c error = add_to_page_cache_lru(page, mapping, page 1174 mm/filemap.c page_cache_release(page); page 1207 mm/filemap.c kaddr = kmap_atomic(page, KM_USER0); page 1216 mm/filemap.c kaddr = kmap(page); page 1218 mm/filemap.c kunmap(page); page 1388 mm/filemap.c struct page *page; page 1392 mm/filemap.c page = page_cache_alloc_cold(mapping); page 1393 mm/filemap.c if (!page) page 1396 mm/filemap.c ret = add_to_page_cache_lru(page, mapping, offset, GFP_KERNEL); page 1398 mm/filemap.c ret = mapping->a_ops->readpage(file, page); page 1402 mm/filemap.c page_cache_release(page); page 1430 mm/filemap.c struct page *page; page 1447 mm/filemap.c page = find_lock_page(mapping, vmf->pgoff); page 1452 mm/filemap.c if (!page) { page 1455 mm/filemap.c page = find_lock_page(mapping, vmf->pgoff); page 1456 mm/filemap.c if (!page) page 1459 mm/filemap.c if (PageReadahead(page)) { page 1460 mm/filemap.c page_cache_async_readahead(mapping, ra, file, page, page 1465 mm/filemap.c if (!page) { page 1494 mm/filemap.c page = find_lock_page(mapping, vmf->pgoff); page 1495 mm/filemap.c if (!page) page 1506 mm/filemap.c if (unlikely(!PageUptodate(page))) page 1512 mm/filemap.c unlock_page(page); page 1513 mm/filemap.c page_cache_release(page); page 1520 mm/filemap.c mark_page_accessed(page); page 1521 mm/filemap.c ra->prev_pos = (loff_t)page->index << PAGE_CACHE_SHIFT; page 1522 mm/filemap.c vmf->page = page; page 1562 mm/filemap.c ClearPageError(page); page 1563 mm/filemap.c error = mapping->a_ops->readpage(file, page); page 1565 mm/filemap.c wait_on_page_locked(page); page 1566 mm/filemap.c if (!PageUptodate(page)) page 1569 mm/filemap.c page_cache_release(page); page 1621 mm/filemap.c static struct page *__read_cache_page(struct address_space *mapping, page 1626 mm/filemap.c struct page *page; page 1629 mm/filemap.c page = find_get_page(mapping, index); page 1630 mm/filemap.c if (!page) { page 1631 mm/filemap.c page = page_cache_alloc_cold(mapping); page 1632 mm/filemap.c if (!page) page 1634 mm/filemap.c err = add_to_page_cache_lru(page, mapping, index, GFP_KERNEL); page 1636 mm/filemap.c page_cache_release(page); page 1642 mm/filemap.c err = filler(data, page); page 1644 mm/filemap.c page_cache_release(page); page 1645 mm/filemap.c page = ERR_PTR(err); page 1648 mm/filemap.c return page; page 1666 mm/filemap.c struct page *read_cache_page_async(struct address_space *mapping, page 1671 mm/filemap.c struct page *page; page 1675 mm/filemap.c page = __read_cache_page(mapping, index, filler, data); page 1676 mm/filemap.c if (IS_ERR(page)) page 1677 mm/filemap.c return page; page 1678 mm/filemap.c if (PageUptodate(page)) page 1681 mm/filemap.c lock_page(page); page 1682 mm/filemap.c if (!page->mapping) { page 1683 mm/filemap.c unlock_page(page); page 1684 mm/filemap.c page_cache_release(page); page 1687 mm/filemap.c if (PageUptodate(page)) { page 1688 mm/filemap.c unlock_page(page); page 1691 mm/filemap.c err = filler(data, page); page 1693 mm/filemap.c page_cache_release(page); page 1697 mm/filemap.c mark_page_accessed(page); page 1698 mm/filemap.c return page; page 1714 mm/filemap.c struct page *read_cache_page(struct address_space *mapping, page 1719 mm/filemap.c struct page *page; page 1721 mm/filemap.c page = read_cache_page_async(mapping, index, filler, data); page 1722 mm/filemap.c if (IS_ERR(page)) page 1724 mm/filemap.c wait_on_page_locked(page); page 1725 mm/filemap.c if (!PageUptodate(page)) { page 1726 mm/filemap.c page_cache_release(page); page 1727 mm/filemap.c page = ERR_PTR(-EIO); page 1730 mm/filemap.c return page; page 1823 mm/filemap.c kaddr = kmap_atomic(page, KM_USER0); page 1852 mm/filemap.c kaddr = kmap(page); page 1862 mm/filemap.c kunmap(page); page 2027 mm/filemap.c struct page *page; page 2029 mm/filemap.c page = __grab_cache_page(mapping, index); page 2030 mm/filemap.c *pagep = page; page 2031 mm/filemap.c if (!page) page 2034 mm/filemap.c if (flags & AOP_FLAG_UNINTERRUPTIBLE && !PageUptodate(page)) { page 2042 mm/filemap.c ret = aops->readpage(file, page); page 2043 mm/filemap.c page_cache_release(page); page 2052 mm/filemap.c ret = aops->prepare_write(file, page, offset, offset+len); page 2054 mm/filemap.c unlock_page(page); page 2055 mm/filemap.c page_cache_release(page); page 2072 mm/filemap.c mark_page_accessed(page); page 2074 mm/filemap.c page, fsdata); page 2079 mm/filemap.c flush_dcache_page(page); page 2080 mm/filemap.c ret = aops->commit_write(file, page, offset, offset+len); page 2081 mm/filemap.c unlock_page(page); page 2082 mm/filemap.c mark_page_accessed(page); page 2083 mm/filemap.c page_cache_release(page); page 2193 mm/filemap.c struct page *__grab_cache_page(struct address_space *mapping, pgoff_t index) page 2196 mm/filemap.c struct page *page; page 2198 mm/filemap.c page = find_lock_page(mapping, index); page 2199 mm/filemap.c if (likely(page)) page 2200 mm/filemap.c return page; page 2202 mm/filemap.c page = page_cache_alloc(mapping); page 2203 mm/filemap.c if (!page) page 2205 mm/filemap.c status = add_to_page_cache_lru(page, mapping, index, GFP_KERNEL); page 2207 mm/filemap.c page_cache_release(page); page 2212 mm/filemap.c return page; page 2226 mm/filemap.c struct page *src_page; page 2227 mm/filemap.c struct page *page; page 2259 mm/filemap.c page = __grab_cache_page(mapping, index); page 2260 mm/filemap.c if (!page) { page 2270 mm/filemap.c if (!PageUptodate(page) && !segment_eq(get_fs(), KERNEL_DS)) { page 2271 mm/filemap.c unlock_page(page); page 2275 mm/filemap.c page_cache_release(page); page 2289 mm/filemap.c page_cache_release(page); page 2295 mm/filemap.c lock_page(page); page 2304 mm/filemap.c if (unlikely(!page->mapping || PageUptodate(page))) { page 2305 mm/filemap.c unlock_page(page); page 2306 mm/filemap.c page_cache_release(page); page 2312 mm/filemap.c status = a_ops->prepare_write(file, page, offset, offset+bytes); page 2331 mm/filemap.c copied = iov_iter_copy_from_user_atomic(page, i, page 2337 mm/filemap.c dst = kmap_atomic(page, KM_USER1); page 2343 mm/filemap.c flush_dcache_page(page); page 2345 mm/filemap.c status = a_ops->commit_write(file, page, offset, offset+bytes); page 2351 mm/filemap.c unlock_page(page); page 2352 mm/filemap.c mark_page_accessed(page); page 2353 mm/filemap.c page_cache_release(page); page 2366 mm/filemap.c unlock_page(page); page 2367 mm/filemap.c page_cache_release(page); page 2400 mm/filemap.c struct page *page; page 2430 mm/filemap.c &page, &fsdata); page 2435 mm/filemap.c copied = iov_iter_copy_from_user_atomic(page, i, offset, bytes); page 2437 mm/filemap.c flush_dcache_page(page); page 2440 mm/filemap.c page, fsdata); page 2682 mm/filemap.c struct address_space * const mapping = page->mapping; page 2684 mm/filemap.c BUG_ON(!PageLocked(page)); page 2685 mm/filemap.c if (PageWriteback(page)) page 2689 mm/filemap.c return mapping->a_ops->releasepage(page, gfp_mask); page 2690 mm/filemap.c return try_to_free_buffers(page); page 29 mm/filemap_xip.c static struct page *__xip_sparse_page; page 32 mm/filemap_xip.c static struct page *xip_sparse_page(void) page 35 mm/filemap_xip.c struct page *page = alloc_page(GFP_HIGHUSER | __GFP_ZERO); page 37 mm/filemap_xip.c if (page) page 38 mm/filemap_xip.c __xip_sparse_page = page; page 174 mm/filemap_xip.c struct page *page; page 180 mm/filemap_xip.c page = __xip_sparse_page; page 181 mm/filemap_xip.c if (!page) page 191 mm/filemap_xip.c pte = page_check_address(page, mm, address, &ptl, 1); page 196 mm/filemap_xip.c page_remove_rmap(page, vma); page 200 mm/filemap_xip.c page_cache_release(page); page 228 mm/filemap_xip.c struct page *page; page 282 mm/filemap_xip.c page = xip_sparse_page(); page 283 mm/filemap_xip.c if (!page) page 286 mm/filemap_xip.c page); page 30 mm/fremap.c struct page *page; page 34 mm/fremap.c page = vm_normal_page(vma, addr, pte); page 35 mm/fremap.c if (page) { page 37 mm/fremap.c set_page_dirty(page); page 38 mm/fremap.c page_remove_rmap(page, vma); page 39 mm/fremap.c page_cache_release(page); page 78 mm/highmem.c struct page *page; page 100 mm/highmem.c page = pte_page(pkmap_page_table[i]); page 101 mm/highmem.c pte_clear(&init_mm, (unsigned long)page_address(page), page 104 mm/highmem.c set_page_address(page, NULL); page 154 mm/highmem.c if (page_address(page)) page 155 mm/highmem.c return (unsigned long)page_address(page); page 163 mm/highmem.c &(pkmap_page_table[last_pkmap_nr]), mk_pte(page, kmap_prot)); page 166 mm/highmem.c set_page_address(page, (void *)vaddr); page 188 mm/highmem.c vaddr = (unsigned long)page_address(page); page 190 mm/highmem.c vaddr = map_new_virtual(page); page 210 mm/highmem.c vaddr = (unsigned long)page_address(page); page 253 mm/highmem.c struct page *page; page 274 mm/highmem.c return &page_address_htable[hash_ptr(page, PA_HASH_ORDER)]; page 289 mm/highmem.c if (!PageHighMem(page)) page 290 mm/highmem.c return lowmem_page_address(page); page 292 mm/highmem.c pas = page_slot(page); page 299 mm/highmem.c if (pam->page == page) { page 323 mm/highmem.c BUG_ON(!PageHighMem(page)); page 325 mm/highmem.c pas = page_slot(page); page 335 mm/highmem.c pam->page = page; page 344 mm/highmem.c if (pam->page == page) { page 364 mm/hugetlb.c clear_user_highpage(page + i, addr + i * PAGE_SIZE); page 383 mm/hugetlb.c int nid = page_to_nid(page); page 384 mm/hugetlb.c list_add(&page->lru, &h->hugepage_freelists[nid]); page 389 mm/hugetlb.c static struct page *dequeue_huge_page(struct hstate *h) page 392 mm/hugetlb.c struct page *page = NULL; page 396 mm/hugetlb.c page = list_entry(h->hugepage_freelists[nid].next, page 397 mm/hugetlb.c struct page, lru); page 398 mm/hugetlb.c list_del(&page->lru); page 404 mm/hugetlb.c return page; page 407 mm/hugetlb.c static struct page *dequeue_huge_page_vma(struct hstate *h, page 412 mm/hugetlb.c struct page *page = NULL; page 438 mm/hugetlb.c page = list_entry(h->hugepage_freelists[nid].next, page 439 mm/hugetlb.c struct page, lru); page 440 mm/hugetlb.c list_del(&page->lru); page 451 mm/hugetlb.c return page; page 459 mm/hugetlb.c h->nr_huge_pages_node[page_to_nid(page)]--; page 461 mm/hugetlb.c page[i].flags &= ~(1 << PG_locked | 1 << PG_error | 1 << PG_referenced | page 465 mm/hugetlb.c set_compound_page_dtor(page, NULL); page 466 mm/hugetlb.c set_page_refcounted(page); page 467 mm/hugetlb.c arch_release_hugepage(page); page 468 mm/hugetlb.c __free_pages(page, huge_page_order(h)); page 488 mm/hugetlb.c struct hstate *h = page_hstate(page); page 489 mm/hugetlb.c int nid = page_to_nid(page); page 492 mm/hugetlb.c mapping = (struct address_space *) page_private(page); page 493 mm/hugetlb.c set_page_private(page, 0); page 494 mm/hugetlb.c BUG_ON(page_count(page)); page 495 mm/hugetlb.c INIT_LIST_HEAD(&page->lru); page 499 mm/hugetlb.c update_and_free_page(h, page); page 503 mm/hugetlb.c enqueue_huge_page(h, page); page 547 mm/hugetlb.c set_compound_page_dtor(page, free_huge_page); page 552 mm/hugetlb.c put_page(page); /* free it into the hugepage allocator */ page 555 mm/hugetlb.c static struct page *alloc_fresh_huge_page_node(struct hstate *h, int nid) page 557 mm/hugetlb.c struct page *page; page 562 mm/hugetlb.c page = alloc_pages_node(nid, page 566 mm/hugetlb.c if (page) { page 567 mm/hugetlb.c if (arch_prepare_hugepage(page)) { page 568 mm/hugetlb.c __free_pages(page, huge_page_order(h)); page 571 mm/hugetlb.c prep_new_huge_page(h, page, nid); page 574 mm/hugetlb.c return page; page 600 mm/hugetlb.c struct page *page; page 608 mm/hugetlb.c page = alloc_fresh_huge_page_node(h, h->hugetlb_next_nid); page 609 mm/hugetlb.c if (page) page 612 mm/hugetlb.c } while (!page && h->hugetlb_next_nid != start_nid); page 622 mm/hugetlb.c static struct page *alloc_buddy_huge_page(struct hstate *h, page 625 mm/hugetlb.c struct page *page; page 664 mm/hugetlb.c page = alloc_pages(htlb_alloc_mask|__GFP_COMP| page 668 mm/hugetlb.c if (page && arch_prepare_hugepage(page)) { page 669 mm/hugetlb.c __free_pages(page, huge_page_order(h)); page 674 mm/hugetlb.c if (page) { page 679 mm/hugetlb.c put_page_testzero(page); page 680 mm/hugetlb.c VM_BUG_ON(page_count(page)); page 681 mm/hugetlb.c nid = page_to_nid(page); page 682 mm/hugetlb.c set_compound_page_dtor(page, free_huge_page); page 696 mm/hugetlb.c return page; page 706 mm/hugetlb.c struct page *page, *tmp; page 723 mm/hugetlb.c page = alloc_buddy_huge_page(h, NULL, 0); page 724 mm/hugetlb.c if (!page) { page 735 mm/hugetlb.c list_add(&page->lru, &surplus_list); page 762 mm/hugetlb.c list_for_each_entry_safe(page, tmp, &surplus_list, lru) { page 765 mm/hugetlb.c list_del(&page->lru); page 766 mm/hugetlb.c enqueue_huge_page(h, page); page 772 mm/hugetlb.c list_for_each_entry_safe(page, tmp, &surplus_list, lru) { page 773 mm/hugetlb.c list_del(&page->lru); page 781 mm/hugetlb.c free_huge_page(page); page 798 mm/hugetlb.c struct page *page; page 827 mm/hugetlb.c page = list_entry(h->hugepage_freelists[nid].next, page 828 mm/hugetlb.c struct page, lru); page 829 mm/hugetlb.c list_del(&page->lru); page 830 mm/hugetlb.c update_and_free_page(h, page); page 894 mm/hugetlb.c static struct page *alloc_huge_page(struct vm_area_struct *vma, page 898 mm/hugetlb.c struct page *page; page 918 mm/hugetlb.c page = dequeue_huge_page_vma(h, vma, addr, avoid_reserve); page 921 mm/hugetlb.c if (!page) { page 922 mm/hugetlb.c page = alloc_buddy_huge_page(h, vma, addr); page 923 mm/hugetlb.c if (!page) { page 929 mm/hugetlb.c set_page_refcounted(page); page 930 mm/hugetlb.c set_page_private(page, (unsigned long) mapping); page 934 mm/hugetlb.c return page; page 978 mm/hugetlb.c struct page *page = virt_to_page(m); page 980 mm/hugetlb.c __ClearPageReserved(page); page 981 mm/hugetlb.c WARN_ON(page_count(page) != 1); page 982 mm/hugetlb.c prep_compound_page(page, h->order); page 983 mm/hugetlb.c prep_new_huge_page(h, page, page_to_nid(page)); page 1045 mm/hugetlb.c struct page *page, *next; page 1047 mm/hugetlb.c list_for_each_entry_safe(page, next, freel, lru) { page 1050 mm/hugetlb.c if (PageHighMem(page)) page 1052 mm/hugetlb.c list_del(&page->lru); page 1053 mm/hugetlb.c update_and_free_page(h, page); page 1055 mm/hugetlb.c h->free_huge_pages_node[page_to_nid(page)]--; page 1123 mm/hugetlb.c struct page *page = dequeue_huge_page(h); page 1124 mm/hugetlb.c if (!page) page 1126 mm/hugetlb.c update_and_free_page(h, page); page 1599 mm/hugetlb.c pte_mkwrite(pte_mkdirty(mk_pte(page, vma->vm_page_prot))); page 1601 mm/hugetlb.c entry = huge_pte_wrprotect(mk_pte(page, vma->vm_page_prot)); page 1625 mm/hugetlb.c struct page *ptepage; page 1671 mm/hugetlb.c struct page *page; page 1672 mm/hugetlb.c struct page *tmp; page 1706 mm/hugetlb.c page = pte_page(pte); page 1707 mm/hugetlb.c if (page != ref_page) page 1722 mm/hugetlb.c page = pte_page(pte); page 1724 mm/hugetlb.c set_page_dirty(page); page 1725 mm/hugetlb.c list_add(&page->lru, &page_list); page 1730 mm/hugetlb.c list_for_each_entry_safe(page, tmp, &page_list, lru) { page 1731 mm/hugetlb.c list_del(&page->lru); page 1732 mm/hugetlb.c put_page(page); page 1767 mm/hugetlb.c mapping = (struct address_space *)page_private(page); page 1784 mm/hugetlb.c page); page 1795 mm/hugetlb.c struct page *old_page, *new_page; page 1870 mm/hugetlb.c static struct page *hugetlbfs_pagecache_page(struct hstate *h, page 1889 mm/hugetlb.c struct page *page; page 1913 mm/hugetlb.c page = find_lock_page(mapping, idx); page 1914 mm/hugetlb.c if (!page) { page 1918 mm/hugetlb.c page = alloc_huge_page(vma, address, 0); page 1919 mm/hugetlb.c if (IS_ERR(page)) { page 1920 mm/hugetlb.c ret = -PTR_ERR(page); page 1923 mm/hugetlb.c clear_huge_page(page, address, huge_page_size(h)); page 1924 mm/hugetlb.c __SetPageUptodate(page); page 1930 mm/hugetlb.c err = add_to_page_cache(page, mapping, idx, GFP_KERNEL); page 1932 mm/hugetlb.c put_page(page); page 1942 mm/hugetlb.c lock_page(page); page 1966 mm/hugetlb.c new_pte = make_huge_pte(vma, page, ((vma->vm_flags & VM_WRITE) page 1972 mm/hugetlb.c ret = hugetlb_cow(mm, vma, address, ptep, new_pte, page); page 1976 mm/hugetlb.c unlock_page(page); page 1983 mm/hugetlb.c unlock_page(page); page 1984 mm/hugetlb.c put_page(page); page 1994 mm/hugetlb.c struct page *pagecache_page = NULL; page 2055 mm/hugetlb.c __attribute__((weak)) struct page * page 2076 mm/hugetlb.c struct page *page; page 2102 mm/hugetlb.c page = pte_page(huge_ptep_get(pte)); page 2105 mm/hugetlb.c get_page(page); page 2106 mm/hugetlb.c pages[i] = page + pfn_offset; page 23 mm/internal.h atomic_set(&page->_count, v); page 32 mm/internal.h VM_BUG_ON(PageTail(page)); page 33 mm/internal.h VM_BUG_ON(atomic_read(&page->_count)); page 34 mm/internal.h set_page_count(page, 1); page 39 mm/internal.h atomic_dec(&page->_count); page 51 mm/internal.h VM_BUG_ON(!PageBuddy(page)); page 52 mm/internal.h return page_private(page); page 167 mm/memcontrol.c struct page *page; page 176 mm/memcontrol.c return page_to_nid(pc->page); page 181 mm/memcontrol.c return page_zonenum(pc->page); page 267 mm/memcontrol.c return bit_spin_is_locked(PAGE_CGROUP_LOCK_BIT, &page->page_cgroup); page 272 mm/memcontrol.c VM_BUG_ON(!page_cgroup_locked(page)); page 273 mm/memcontrol.c page->page_cgroup = ((unsigned long)pc | PAGE_CGROUP_LOCK); page 278 mm/memcontrol.c return (struct page_cgroup *) (page->page_cgroup & ~PAGE_CGROUP_LOCK); page 283 mm/memcontrol.c bit_spin_lock(PAGE_CGROUP_LOCK_BIT, &page->page_cgroup); page 288 mm/memcontrol.c return bit_spin_trylock(PAGE_CGROUP_LOCK_BIT, &page->page_cgroup); page 293 mm/memcontrol.c bit_spin_unlock(PAGE_CGROUP_LOCK_BIT, &page->page_cgroup); page 375 mm/memcontrol.c if (!try_lock_page_cgroup(page)) page 378 mm/memcontrol.c pc = page_get_page_cgroup(page); page 385 mm/memcontrol.c unlock_page_cgroup(page); page 479 mm/memcontrol.c struct page *page; page 501 mm/memcontrol.c page = pc->page; page 503 mm/memcontrol.c if (unlikely(!PageLRU(page))) page 506 mm/memcontrol.c if (PageActive(page) && !active) { page 510 mm/memcontrol.c if (!PageActive(page) && active) { page 518 mm/memcontrol.c if (__isolate_lru_page(page, mode) == 0) { page 519 mm/memcontrol.c list_move(&page->lru, dst); page 599 mm/memcontrol.c pc->page = page; page 609 mm/memcontrol.c lock_page_cgroup(page); page 610 mm/memcontrol.c if (unlikely(page_get_page_cgroup(page))) { page 611 mm/memcontrol.c unlock_page_cgroup(page); page 617 mm/memcontrol.c page_assign_page_cgroup(page, pc); page 624 mm/memcontrol.c unlock_page_cgroup(page); page 646 mm/memcontrol.c if (page_mapped(page) || (page->mapping && !PageAnon(page))) page 650 mm/memcontrol.c return mem_cgroup_charge_common(page, mm, gfp_mask, page 672 mm/memcontrol.c lock_page_cgroup(page); page 673 mm/memcontrol.c pc = page_get_page_cgroup(page); page 675 mm/memcontrol.c VM_BUG_ON(pc->page != page); page 677 mm/memcontrol.c unlock_page_cgroup(page); page 680 mm/memcontrol.c unlock_page_cgroup(page); page 686 mm/memcontrol.c return mem_cgroup_charge_common(page, mm, gfp_mask, page 707 mm/memcontrol.c lock_page_cgroup(page); page 708 mm/memcontrol.c pc = page_get_page_cgroup(page); page 712 mm/memcontrol.c VM_BUG_ON(pc->page != page); page 716 mm/memcontrol.c || page_mapped(page))) page 724 mm/memcontrol.c page_assign_page_cgroup(page, NULL); page 725 mm/memcontrol.c unlock_page_cgroup(page); page 734 mm/memcontrol.c unlock_page_cgroup(page); page 739 mm/memcontrol.c __mem_cgroup_uncharge_common(page, MEM_CGROUP_CHARGE_TYPE_MAPPED); page 744 mm/memcontrol.c VM_BUG_ON(page_mapped(page)); page 745 mm/memcontrol.c __mem_cgroup_uncharge_common(page, MEM_CGROUP_CHARGE_TYPE_CACHE); page 761 mm/memcontrol.c lock_page_cgroup(page); page 762 mm/memcontrol.c pc = page_get_page_cgroup(page); page 769 mm/memcontrol.c unlock_page_cgroup(page); page 869 mm/memcontrol.c struct page *page; page 882 mm/memcontrol.c page = pc->page; page 883 mm/memcontrol.c get_page(page); page 889 mm/memcontrol.c if (PageLRU(page)) { page 890 mm/memcontrol.c __mem_cgroup_uncharge_common(page, page 892 mm/memcontrol.c put_page(page); page 70 mm/memory.c struct page *mem_map; page 441 mm/memory.c struct page *vm_normal_page(struct vm_area_struct *vma, unsigned long addr, page 498 mm/memory.c struct page *page; page 545 mm/memory.c page = vm_normal_page(vma, addr, pte); page 546 mm/memory.c if (page) { page 547 mm/memory.c get_page(page); page 548 mm/memory.c page_dup_rmap(page, vma, addr); page 549 mm/memory.c rss[!!PageAnon(page)]++; page 724 mm/memory.c struct page *page; page 726 mm/memory.c page = vm_normal_page(vma, addr, ptent); page 727 mm/memory.c if (unlikely(details) && page) { page 734 mm/memory.c details->check_mapping != page->mapping) page 741 mm/memory.c (page->index < details->first_index || page 742 mm/memory.c page->index > details->last_index)) page 748 mm/memory.c if (unlikely(!page)) page 752 mm/memory.c addr) != page->index) page 754 mm/memory.c pgoff_to_pte(page->index)); page 755 mm/memory.c if (PageAnon(page)) page 759 mm/memory.c set_page_dirty(page); page 761 mm/memory.c SetPageReferenced(page); page 764 mm/memory.c page_remove_rmap(page, vma); page 765 mm/memory.c tlb_remove_page(tlb, page); page 1023 mm/memory.c struct page *follow_page(struct vm_area_struct *vma, unsigned long address, page 1031 mm/memory.c struct page *page; page 1034 mm/memory.c page = follow_huge_addr(mm, address, flags & FOLL_WRITE); page 1035 mm/memory.c if (!IS_ERR(page)) { page 1040 mm/memory.c page = NULL; page 1050 mm/memory.c page = follow_huge_pud(mm, address, pud, flags & FOLL_WRITE); page 1061 mm/memory.c page = follow_huge_pmd(mm, address, pmd, flags & FOLL_WRITE); page 1074 mm/memory.c page = vm_normal_page(vma, address, pte); page 1075 mm/memory.c if (unlikely(!page)) page 1079 mm/memory.c get_page(page); page 1082 mm/memory.c !pte_dirty(pte) && !PageDirty(page)) page 1083 mm/memory.c set_page_dirty(page); page 1084 mm/memory.c mark_page_accessed(page); page 1089 mm/memory.c return page; page 1098 mm/memory.c return page; page 1106 mm/memory.c page = ZERO_PAGE(0); page 1108 mm/memory.c get_page(page); page 1111 mm/memory.c return page; page 1179 mm/memory.c struct page *page = vm_normal_page(gate_vma, start, *pte); page 1180 mm/memory.c pages[i] = page; page 1181 mm/memory.c if (page) page 1182 mm/memory.c get_page(page); page 1210 mm/memory.c struct page *page; page 1224 mm/memory.c while (!(page = follow_page(vma, start, foll_flags))) { page 1252 mm/memory.c if (IS_ERR(page)) page 1253 mm/memory.c return i ? i : PTR_ERR(page); page 1255 mm/memory.c pages[i] = page; page 1257 mm/memory.c flush_anon_page(vma, page, start); page 1258 mm/memory.c flush_dcache_page(page); page 1299 mm/memory.c retval = mem_cgroup_charge(page, mm, GFP_KERNEL); page 1304 mm/memory.c if (PageAnon(page)) page 1307 mm/memory.c flush_dcache_page(page); page 1316 mm/memory.c get_page(page); page 1318 mm/memory.c page_add_file_rmap(page); page 1319 mm/memory.c set_pte_at(mm, addr, pte, mk_pte(page, prot)); page 1327 mm/memory.c mem_cgroup_uncharge_page(page); page 1359 mm/memory.c if (!page_count(page)) page 1362 mm/memory.c return insert_page(vma, addr, page, vma->vm_page_prot); page 1447 mm/memory.c struct page *page; page 1449 mm/memory.c page = pfn_to_page(pfn); page 1450 mm/memory.c return insert_page(vma, addr, page, vma->vm_page_prot); page 1766 mm/memory.c struct page *old_page, *new_page; page 1770 mm/memory.c struct page *dirty_page = NULL; page 2255 mm/memory.c struct page *page; page 2269 mm/memory.c page = lookup_swap_cache(entry); page 2270 mm/memory.c if (!page) { page 2272 mm/memory.c page = swapin_readahead(entry, page 2274 mm/memory.c if (!page) { page 2291 mm/memory.c if (mem_cgroup_charge(page, mm, GFP_KERNEL)) { page 2297 mm/memory.c mark_page_accessed(page); page 2298 mm/memory.c lock_page(page); page 2308 mm/memory.c if (unlikely(!PageUptodate(page))) { page 2316 mm/memory.c pte = mk_pte(page, vma->vm_page_prot); page 2317 mm/memory.c if (write_access && can_share_swap_page(page)) { page 2322 mm/memory.c flush_icache_page(vma, page); page 2324 mm/memory.c page_add_anon_rmap(page, vma, address); page 2328 mm/memory.c remove_exclusive_swap_page(page); page 2329 mm/memory.c unlock_page(page); page 2345 mm/memory.c mem_cgroup_uncharge_page(page); page 2347 mm/memory.c unlock_page(page); page 2348 mm/memory.c page_cache_release(page); page 2361 mm/memory.c struct page *page; page 2370 mm/memory.c page = alloc_zeroed_user_highpage_movable(vma, address); page 2371 mm/memory.c if (!page) page 2373 mm/memory.c __SetPageUptodate(page); page 2375 mm/memory.c if (mem_cgroup_charge(page, mm, GFP_KERNEL)) page 2378 mm/memory.c entry = mk_pte(page, vma->vm_page_prot); page 2385 mm/memory.c lru_cache_add_active(page); page 2386 mm/memory.c page_add_new_anon_rmap(page, vma, address); page 2395 mm/memory.c mem_cgroup_uncharge_page(page); page 2396 mm/memory.c page_cache_release(page); page 2399 mm/memory.c page_cache_release(page); page 2423 mm/memory.c struct page *page; page 2426 mm/memory.c struct page *dirty_page = NULL; page 2434 mm/memory.c vmf.page = NULL; page 2445 mm/memory.c lock_page(vmf.page); page 2447 mm/memory.c VM_BUG_ON(!PageLocked(vmf.page)); page 2452 mm/memory.c page = vmf.page; page 2460 mm/memory.c page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, page 2462 mm/memory.c if (!page) { page 2466 mm/memory.c copy_user_highpage(page, vmf.page, address, vma); page 2467 mm/memory.c __SetPageUptodate(page); page 2475 mm/memory.c unlock_page(page); page 2476 mm/memory.c if (vma->vm_ops->page_mkwrite(vma, page) < 0) { page 2481 mm/memory.c lock_page(page); page 2489 mm/memory.c if (!page->mapping) { page 2500 mm/memory.c if (mem_cgroup_charge(page, mm, GFP_KERNEL)) { page 2519 mm/memory.c flush_icache_page(vma, page); page 2520 mm/memory.c entry = mk_pte(page, vma->vm_page_prot); page 2526 mm/memory.c lru_cache_add_active(page); page 2527 mm/memory.c page_add_new_anon_rmap(page, vma, address); page 2530 mm/memory.c page_add_file_rmap(page); page 2532 mm/memory.c dirty_page = page; page 2540 mm/memory.c mem_cgroup_uncharge_page(page); page 2542 mm/memory.c page_cache_release(page); page 2550 mm/memory.c unlock_page(vmf.page); page 2553 mm/memory.c page_cache_release(vmf.page); page 2934 mm/memory.c struct page *page = NULL; page 2937 mm/memory.c write, 1, &page, &vma); page 2960 mm/memory.c maddr = kmap(page); page 2962 mm/memory.c copy_to_user_page(vma, page, addr, page 2964 mm/memory.c set_page_dirty_lock(page); page 2966 mm/memory.c copy_from_user_page(vma, page, addr, page 2969 mm/memory.c kunmap(page); page 2970 mm/memory.c page_cache_release(page); page 67 mm/memory_hotplug.c atomic_set(&page->_mapcount, type); page 68 mm/memory_hotplug.c SetPagePrivate(page); page 69 mm/memory_hotplug.c set_page_private(page, info); page 70 mm/memory_hotplug.c atomic_inc(&page->_count); page 77 mm/memory_hotplug.c type = atomic_read(&page->_mapcount); page 80 mm/memory_hotplug.c if (atomic_dec_return(&page->_count) == 1) { page 81 mm/memory_hotplug.c ClearPagePrivate(page); page 82 mm/memory_hotplug.c set_page_private(page, 0); page 83 mm/memory_hotplug.c reset_page_mapcount(page); page 84 mm/memory_hotplug.c __free_pages_bootmem(page, 0); page 93 mm/memory_hotplug.c struct page *page, *memmap; page 108 mm/memory_hotplug.c page = virt_to_page(memmap); page 109 mm/memory_hotplug.c mapsize = sizeof(struct page) * PAGES_PER_SECTION; page 113 mm/memory_hotplug.c for (i = 0; i < mapsize; i++, page++) page 114 mm/memory_hotplug.c get_page_bootmem(section_nr, page, SECTION_INFO); page 117 mm/memory_hotplug.c page = virt_to_page(usemap); page 121 mm/memory_hotplug.c for (i = 0; i < mapsize; i++, page++) page 122 mm/memory_hotplug.c get_page_bootmem(section_nr, page, MIX_SECTION_INFO); page 130 mm/memory_hotplug.c struct page *page; page 134 mm/memory_hotplug.c page = virt_to_page(pgdat); page 136 mm/memory_hotplug.c for (i = 0; i < nr_pages; i++, page++) page 137 mm/memory_hotplug.c get_page_bootmem(node, page, NODE_INFO); page 145 mm/memory_hotplug.c page = virt_to_page(zone->wait_table); page 147 mm/memory_hotplug.c for (i = 0; i < nr_pages; i++, page++) page 148 mm/memory_hotplug.c get_page_bootmem(node, page, NODE_INFO); page 345 mm/memory_hotplug.c if (PageHighMem(page)) page 350 mm/memory_hotplug.c max_mapnr = max(page_to_pfn(page), max_mapnr); page 353 mm/memory_hotplug.c ClearPageReserved(page); page 354 mm/memory_hotplug.c init_page_count(page); page 355 mm/memory_hotplug.c __free_page(page); page 363 mm/memory_hotplug.c struct page *page; page 366 mm/memory_hotplug.c page = pfn_to_page(start_pfn + i); page 367 mm/memory_hotplug.c online_page(page); page 534 mm/memory_hotplug.c return PageBuddy(page) && page_order(page) >= pageblock_order; page 538 mm/memory_hotplug.c static struct page *next_active_pageblock(struct page *page) page 543 mm/memory_hotplug.c BUG_ON(page_to_pfn(page) & (pageblock_nr_pages - 1)); page 549 mm/memory_hotplug.c if (pageblock_free(page)) page 550 mm/memory_hotplug.c pageblocks_stride += page_order(page) - pageblock_order; page 552 mm/memory_hotplug.c return page + (pageblocks_stride * pageblock_nr_pages); page 559 mm/memory_hotplug.c struct page *page = pfn_to_page(start_pfn); page 560 mm/memory_hotplug.c struct page *end_page = page + nr_pages; page 563 mm/memory_hotplug.c for (; page < end_page; page = next_active_pageblock(page)) { page 564 mm/memory_hotplug.c type = get_pageblock_migratetype(page); page 570 mm/memory_hotplug.c if (type != MIGRATE_MOVABLE && !pageblock_free(page)) page 577 mm/memory_hotplug.c if (PageReserved(page)) page 592 mm/memory_hotplug.c struct page *page; page 603 mm/memory_hotplug.c page = pfn_to_page(pfn + i); page 604 mm/memory_hotplug.c if (zone && page_zone(page) != zone) page 606 mm/memory_hotplug.c zone = page_zone(page); page 618 mm/memory_hotplug.c struct page *page; page 621 mm/memory_hotplug.c page = pfn_to_page(pfn); page 622 mm/memory_hotplug.c if (PageLRU(page)) page 629 mm/memory_hotplug.c static struct page * page 644 mm/memory_hotplug.c struct page *page; page 653 mm/memory_hotplug.c page = pfn_to_page(pfn); page 654 mm/memory_hotplug.c if (!page_count(page)) page 660 mm/memory_hotplug.c ret = isolate_lru_page(page, &source); page 666 mm/memory_hotplug.c if (page_count(page)) page 671 mm/memory_hotplug.c pfn, page_count(page), page->flags); page 382 mm/mempolicy.c struct page *page; page 387 mm/mempolicy.c page = vm_normal_page(vma, addr, *pte); page 388 mm/mempolicy.c if (!page) page 401 mm/mempolicy.c if (PageReserved(page)) page 403 mm/mempolicy.c nid = page_to_nid(page); page 408 mm/mempolicy.c gather_stats(page, private, pte_dirty(*pte)); page 410 mm/mempolicy.c migrate_page_add(page, private, flags); page 661 mm/mempolicy.c struct page *p; page 765 mm/mempolicy.c if ((flags & MPOL_MF_MOVE_ALL) || page_mapcount(page) == 1) page 766 mm/mempolicy.c isolate_lru_page(page, pagelist); page 769 mm/mempolicy.c static struct page *new_node_page(struct page *page, unsigned long node, int **x) page 890 mm/mempolicy.c static struct page *new_vma_page(struct page *page, unsigned long private, int **x) page 896 mm/mempolicy.c address = page_address_in_vma(page, vma); page 920 mm/mempolicy.c static struct page *new_vma_page(struct page *page, unsigned long private, int **x) page 1495 mm/mempolicy.c static struct page *alloc_page_interleave(gfp_t gfp, unsigned order, page 1499 mm/mempolicy.c struct page *page; page 1502 mm/mempolicy.c page = __alloc_pages(gfp, order, zl); page 1503 mm/mempolicy.c if (page && page_zone(page) == zonelist_zone(&zl->_zonerefs[0])) page 1504 mm/mempolicy.c inc_zone_page_state(page, NUMA_INTERLEAVE_HIT); page 1505 mm/mempolicy.c return page; page 1530 mm/mempolicy.c struct page * page 1550 mm/mempolicy.c struct page *page = __alloc_pages_nodemask(gfp, 0, page 1553 mm/mempolicy.c return page; page 1580 mm/mempolicy.c struct page *alloc_pages_current(gfp_t gfp, unsigned order) page 2191 mm/mempolicy.c int count = page_mapcount(page); page 2194 mm/mempolicy.c if (pte_dirty || PageDirty(page)) page 2197 mm/mempolicy.c if (PageSwapCache(page)) page 2200 mm/mempolicy.c if (PageActive(page)) page 2203 mm/mempolicy.c if (PageWriteback(page)) page 2206 mm/mempolicy.c if (PageAnon(page)) page 2212 mm/mempolicy.c md->node[page_to_nid(page)]++; page 2221 mm/mempolicy.c struct page *page; page 2237 mm/mempolicy.c page = pte_page(pte); page 2238 mm/mempolicy.c if (!page) page 2241 mm/mempolicy.c gather_stats(page, md, pte_dirty(*ptep)); page 37 mm/migrate.c #define lru_to_page(_head) (list_entry((_head)->prev, struct page, lru)) page 51 mm/migrate.c if (PageLRU(page)) { page 52 mm/migrate.c struct zone *zone = page_zone(page); page 55 mm/migrate.c if (PageLRU(page) && get_page_unless_zero(page)) { page 57 mm/migrate.c ClearPageLRU(page); page 58 mm/migrate.c if (PageActive(page)) page 59 mm/migrate.c del_page_from_active_list(zone, page); page 61 mm/migrate.c del_page_from_inactive_list(zone, page); page 62 mm/migrate.c list_add_tail(&page->lru, pagelist); page 88 mm/migrate.c if (PageActive(page)) { page 93 mm/migrate.c ClearPageActive(page); page 94 mm/migrate.c lru_cache_add_active(page); page 96 mm/migrate.c lru_cache_add(page); page 98 mm/migrate.c put_page(page); page 108 mm/migrate.c struct page *page; page 109 mm/migrate.c struct page *page2; page 112 mm/migrate.c list_for_each_entry_safe(page, page2, l, lru) { page 113 mm/migrate.c list_del(&page->lru); page 114 mm/migrate.c move_to_lru(page); page 275 mm/migrate.c struct page *page; page 286 mm/migrate.c page = migration_entry_to_page(entry); page 295 mm/migrate.c if (!get_page_unless_zero(page)) page 298 mm/migrate.c wait_on_page_locked(page); page 299 mm/migrate.c put_page(page); page 321 mm/migrate.c if (page_count(page) != 1) page 329 mm/migrate.c page_index(page)); page 331 mm/migrate.c expected_count = 2 + !!PagePrivate(page); page 332 mm/migrate.c if (page_count(page) != expected_count || page 333 mm/migrate.c (struct page *)radix_tree_deref_slot(pslot) != page) { page 338 mm/migrate.c if (!page_freeze_refs(page, expected_count)) { page 348 mm/migrate.c if (PageSwapCache(page)) { page 350 mm/migrate.c set_page_private(newpage, page_private(page)); page 356 mm/migrate.c page_unfreeze_refs(page, expected_count); page 361 mm/migrate.c __put_page(page); page 373 mm/migrate.c __dec_zone_page_state(page, NR_FILE_PAGES); page 378 mm/migrate.c mem_cgroup_uncharge_cache_page(page); page 388 mm/migrate.c copy_highpage(newpage, page); page 390 mm/migrate.c if (PageError(page)) page 392 mm/migrate.c if (PageReferenced(page)) page 394 mm/migrate.c if (PageUptodate(page)) page 396 mm/migrate.c if (PageActive(page)) page 398 mm/migrate.c if (PageChecked(page)) page 400 mm/migrate.c if (PageMappedToDisk(page)) page 403 mm/migrate.c if (PageDirty(page)) { page 404 mm/migrate.c clear_page_dirty_for_io(page); page 416 mm/migrate.c ClearPageSwapCache(page); page 418 mm/migrate.c ClearPageActive(page); page 419 mm/migrate.c ClearPagePrivate(page); page 420 mm/migrate.c set_page_private(page, 0); page 421 mm/migrate.c page->mapping = NULL; page 454 mm/migrate.c BUG_ON(PageWriteback(page)); /* Writeback must be complete */ page 456 mm/migrate.c rc = migrate_page_move_mapping(mapping, newpage, page); page 461 mm/migrate.c migrate_page_copy(newpage, page); page 478 mm/migrate.c if (!page_has_buffers(page)) page 479 mm/migrate.c return migrate_page(mapping, newpage, page); page 481 mm/migrate.c head = page_buffers(page); page 483 mm/migrate.c rc = migrate_page_move_mapping(mapping, newpage, page); page 496 mm/migrate.c ClearPagePrivate(page); page 497 mm/migrate.c set_page_private(newpage, page_private(page)); page 498 mm/migrate.c set_page_private(page, 0); page 499 mm/migrate.c put_page(page); page 511 mm/migrate.c migrate_page_copy(newpage, page); page 545 mm/migrate.c if (!clear_page_dirty_for_io(page)) page 557 mm/migrate.c remove_migration_ptes(page, page); page 559 mm/migrate.c rc = mapping->a_ops->writepage(page, &wbc); page 566 mm/migrate.c lock_page(page); page 577 mm/migrate.c if (PageDirty(page)) page 578 mm/migrate.c return writeout(mapping, page); page 584 mm/migrate.c if (PagePrivate(page) && page 585 mm/migrate.c !try_to_release_page(page, GFP_KERNEL)) page 588 mm/migrate.c return migrate_page(mapping, newpage, page); page 612 mm/migrate.c newpage->index = page->index; page 613 mm/migrate.c newpage->mapping = page->mapping; page 615 mm/migrate.c mapping = page_mapping(page); page 617 mm/migrate.c rc = migrate_page(mapping, newpage, page); page 627 mm/migrate.c newpage, page); page 629 mm/migrate.c rc = fallback_migrate_page(mapping, newpage, page); page 632 mm/migrate.c remove_migration_ptes(page, newpage); page 650 mm/migrate.c struct page *newpage = get_new_page(page, private, &result); page 657 mm/migrate.c if (page_count(page) == 1) page 661 mm/migrate.c charge = mem_cgroup_prepare_migration(page, newpage); page 670 mm/migrate.c if (!trylock_page(page)) { page 673 mm/migrate.c lock_page(page); page 676 mm/migrate.c if (PageWriteback(page)) { page 679 mm/migrate.c wait_on_page_writeback(page); page 689 mm/migrate.c if (PageAnon(page)) { page 706 mm/migrate.c if (!page->mapping) { page 707 mm/migrate.c if (!PageAnon(page) && PagePrivate(page)) { page 715 mm/migrate.c try_to_free_buffers(page); page 721 mm/migrate.c try_to_unmap(page, 1); page 723 mm/migrate.c if (!page_mapped(page)) page 724 mm/migrate.c rc = move_to_new_page(newpage, page); page 727 mm/migrate.c remove_migration_ptes(page, page); page 734 mm/migrate.c unlock_page(page); page 743 mm/migrate.c list_del(&page->lru); page 744 mm/migrate.c move_to_lru(page); page 784 mm/migrate.c struct page *page; page 785 mm/migrate.c struct page *page2; page 795 mm/migrate.c list_for_each_entry_safe(page, page2, from, lru) { page 799 mm/migrate.c page, pass > 2); page 835 mm/migrate.c struct page *page; page 840 mm/migrate.c static struct page *new_page_node(struct page *p, unsigned long private, page 845 mm/migrate.c while (pm->node != MAX_NUMNODES && pm->page != p) page 877 mm/migrate.c struct page *page; page 883 mm/migrate.c pp->page = ZERO_PAGE(0); page 890 mm/migrate.c page = follow_page(vma, pp->addr, FOLL_GET); page 892 mm/migrate.c err = PTR_ERR(page); page 893 mm/migrate.c if (IS_ERR(page)) page 897 mm/migrate.c if (!page) page 900 mm/migrate.c if (PageReserved(page)) /* Check for zero page */ page 903 mm/migrate.c pp->page = page; page 904 mm/migrate.c err = page_to_nid(page); page 913 mm/migrate.c if (page_mapcount(page) > 1 && page 917 mm/migrate.c err = isolate_lru_page(page, &pagelist); page 924 mm/migrate.c put_page(page); page 950 mm/migrate.c struct page *page; page 958 mm/migrate.c page = follow_page(vma, pm->addr, 0); page 960 mm/migrate.c err = PTR_ERR(page); page 961 mm/migrate.c if (IS_ERR(page)) page 966 mm/migrate.c if (!page || PageReserved(page)) page 969 mm/migrate.c err = page_to_nid(page); page 30 mm/mincore.c struct page *page; page 42 mm/mincore.c page = find_get_page(mapping, pgoff); page 43 mm/mincore.c if (page) { page 44 mm/mincore.c present = PageUptodate(page); page 45 mm/mincore.c page_cache_release(page); page 127 mm/mm_init.c BUG_ON(page_to_nid(page) != nid); page 128 mm/mm_init.c BUG_ON(page_zonenum(page) != zone); page 129 mm/mm_init.c BUG_ON(page_to_pfn(page) != pfn); page 2204 mm/mmap.c struct page **pages; page 2218 mm/mmap.c struct page *page = *pages; page 2219 mm/mmap.c get_page(page); page 2220 mm/mmap.c vmf->page = page; page 38 mm/nommu.c struct page *mem_map; page 106 mm/nommu.c struct page *page; page 115 mm/nommu.c page = virt_to_head_page(objp); page 121 mm/nommu.c if (PageSlab(page)) page 128 mm/nommu.c return PAGE_SIZE << compound_order(page); page 218 mm/nommu.c struct page *vmalloc_to_page(const void *addr) page 1276 mm/nommu.c struct page *follow_page(struct vm_area_struct *vma, unsigned long address, page 535 mm/page-writeback.c if (set_page_dirty(page) || page_mkwrite) { page 536 mm/page-writeback.c struct address_space *mapping = page_mapping(page); page 905 mm/page-writeback.c struct page *page = pvec.pages[i]; page 914 mm/page-writeback.c lock_page(page); page 916 mm/page-writeback.c if (unlikely(page->mapping != mapping)) { page 917 mm/page-writeback.c unlock_page(page); page 921 mm/page-writeback.c if (!wbc->range_cyclic && page->index > end) { page 923 mm/page-writeback.c unlock_page(page); page 928 mm/page-writeback.c wait_on_page_writeback(page); page 930 mm/page-writeback.c if (PageWriteback(page) || page 931 mm/page-writeback.c !clear_page_dirty_for_io(page)) { page 932 mm/page-writeback.c unlock_page(page); page 936 mm/page-writeback.c ret = (*writepage)(page, wbc, data); page 939 mm/page-writeback.c unlock_page(page); page 978 mm/page-writeback.c int ret = mapping->a_ops->writepage(page, wbc); page 1029 mm/page-writeback.c struct address_space *mapping = page->mapping; page 1036 mm/page-writeback.c BUG_ON(!PageLocked(page)); page 1039 mm/page-writeback.c wait_on_page_writeback(page); page 1041 mm/page-writeback.c if (clear_page_dirty_for_io(page)) { page 1042 mm/page-writeback.c page_cache_get(page); page 1043 mm/page-writeback.c ret = mapping->a_ops->writepage(page, &wbc); page 1045 mm/page-writeback.c wait_on_page_writeback(page); page 1046 mm/page-writeback.c if (PageError(page)) page 1049 mm/page-writeback.c page_cache_release(page); page 1051 mm/page-writeback.c unlock_page(page); page 1062 mm/page-writeback.c if (!PageDirty(page)) page 1063 mm/page-writeback.c SetPageDirty(page); page 1084 mm/page-writeback.c if (!TestSetPageDirty(page)) { page 1085 mm/page-writeback.c struct address_space *mapping = page_mapping(page); page 1092 mm/page-writeback.c mapping2 = page_mapping(page); page 1095 mm/page-writeback.c WARN_ON_ONCE(!PagePrivate(page) && !PageUptodate(page)); page 1097 mm/page-writeback.c __inc_zone_page_state(page, NR_FILE_DIRTY); page 1103 mm/page-writeback.c page_index(page), PAGECACHE_TAG_DIRTY); page 1124 mm/page-writeback.c return __set_page_dirty_nobuffers(page); page 1134 mm/page-writeback.c struct address_space *mapping = page_mapping(page); page 1137 mm/page-writeback.c int (*spd)(struct page *) = mapping->a_ops->set_page_dirty; page 1142 mm/page-writeback.c return (*spd)(page); page 1144 mm/page-writeback.c if (!PageDirty(page)) { page 1145 mm/page-writeback.c if (!TestSetPageDirty(page)) page 1153 mm/page-writeback.c int ret = __set_page_dirty(page); page 1174 mm/page-writeback.c lock_page_nosync(page); page 1175 mm/page-writeback.c ret = set_page_dirty(page); page 1176 mm/page-writeback.c unlock_page(page); page 1197 mm/page-writeback.c struct address_space *mapping = page_mapping(page); page 1199 mm/page-writeback.c BUG_ON(!PageLocked(page)); page 1201 mm/page-writeback.c ClearPageReclaim(page); page 1228 mm/page-writeback.c if (page_mkclean(page)) page 1229 mm/page-writeback.c set_page_dirty(page); page 1240 mm/page-writeback.c if (TestClearPageDirty(page)) { page 1241 mm/page-writeback.c dec_zone_page_state(page, NR_FILE_DIRTY); page 1248 mm/page-writeback.c return TestClearPageDirty(page); page 1254 mm/page-writeback.c struct address_space *mapping = page_mapping(page); page 1262 mm/page-writeback.c ret = TestClearPageWriteback(page); page 1265 mm/page-writeback.c page_index(page), page 1274 mm/page-writeback.c ret = TestClearPageWriteback(page); page 1277 mm/page-writeback.c dec_zone_page_state(page, NR_WRITEBACK); page 1283 mm/page-writeback.c struct address_space *mapping = page_mapping(page); page 1291 mm/page-writeback.c ret = TestSetPageWriteback(page); page 1294 mm/page-writeback.c page_index(page), page 1299 mm/page-writeback.c if (!PageDirty(page)) page 1301 mm/page-writeback.c page_index(page), page 1305 mm/page-writeback.c ret = TestSetPageWriteback(page); page 1308 mm/page-writeback.c inc_zone_page_state(page, NR_WRITEBACK); page 174 mm/page_alloc.c set_pageblock_flags_group(page, (unsigned long)migratetype, page 183 mm/page_alloc.c unsigned long pfn = page_to_pfn(page); page 198 mm/page_alloc.c if (!pfn_valid_within(page_to_pfn(page))) page 200 mm/page_alloc.c if (zone != page_zone(page)) page 210 mm/page_alloc.c if (page_outside_zone_boundaries(zone, page)) page 212 mm/page_alloc.c if (!page_is_consistent(zone, page)) page 226 mm/page_alloc.c void *pc = page_get_page_cgroup(page); page 230 mm/page_alloc.c current->comm, page, (int)(2*sizeof(unsigned long)), page 231 mm/page_alloc.c (unsigned long)page->flags, page->mapping, page 232 mm/page_alloc.c page_mapcount(page), page_count(page)); page 235 mm/page_alloc.c page_reset_bad_cgroup(page); page 240 mm/page_alloc.c page->flags &= ~PAGE_FLAGS_CLEAR_WHEN_BAD; page 241 mm/page_alloc.c set_page_count(page, 0); page 242 mm/page_alloc.c reset_page_mapcount(page); page 243 mm/page_alloc.c page->mapping = NULL; page 264 mm/page_alloc.c __free_pages_ok(page, compound_order(page)); page 271 mm/page_alloc.c struct page *p = page + 1; page 273 mm/page_alloc.c set_compound_page_dtor(page, free_compound_page); page 274 mm/page_alloc.c set_compound_order(page, order); page 275 mm/page_alloc.c __SetPageHead(page); page 278 mm/page_alloc.c p = pfn_to_page(page_to_pfn(page) + i); page 280 mm/page_alloc.c p->first_page = page; page 288 mm/page_alloc.c struct page *p = page + 1; page 290 mm/page_alloc.c if (unlikely(compound_order(page) != order)) page 291 mm/page_alloc.c bad_page(page); page 293 mm/page_alloc.c if (unlikely(!PageHead(page))) page 294 mm/page_alloc.c bad_page(page); page 295 mm/page_alloc.c __ClearPageHead(page); page 298 mm/page_alloc.c p = pfn_to_page(page_to_pfn(page) + i); page 301 mm/page_alloc.c (p->first_page != page))) page 302 mm/page_alloc.c bad_page(page); page 317 mm/page_alloc.c clear_highpage(page + i); page 322 mm/page_alloc.c set_page_private(page, order); page 323 mm/page_alloc.c __SetPageBuddy(page); page 328 mm/page_alloc.c __ClearPageBuddy(page); page 329 mm/page_alloc.c set_page_private(page, 0); page 349 mm/page_alloc.c static inline struct page * page 354 mm/page_alloc.c return page + (buddy_idx - page_idx); page 382 mm/page_alloc.c if (page_zone_id(page) != page_zone_id(buddy)) page 421 mm/page_alloc.c int migratetype = get_pageblock_migratetype(page); page 423 mm/page_alloc.c if (unlikely(PageCompound(page))) page 424 mm/page_alloc.c destroy_compound_page(page, order); page 426 mm/page_alloc.c page_idx = page_to_pfn(page) & ((1 << MAX_ORDER) - 1); page 429 mm/page_alloc.c VM_BUG_ON(bad_range(zone, page)); page 434 mm/page_alloc.c struct page *buddy; page 436 mm/page_alloc.c buddy = __page_find_buddy(page, page_idx, order); page 437 mm/page_alloc.c if (!page_is_buddy(page, buddy, order)) page 445 mm/page_alloc.c page = page + (combined_idx - page_idx); page 449 mm/page_alloc.c set_page_order(page, order); page 450 mm/page_alloc.c list_add(&page->lru, page 457 mm/page_alloc.c if (unlikely(page_mapcount(page) | page 458 mm/page_alloc.c (page->mapping != NULL) | page 459 mm/page_alloc.c (page_get_page_cgroup(page) != NULL) | page 460 mm/page_alloc.c (page_count(page) != 0) | page 461 mm/page_alloc.c (page->flags & PAGE_FLAGS_CHECK_AT_FREE))) page 462 mm/page_alloc.c bad_page(page); page 463 mm/page_alloc.c if (PageDirty(page)) page 464 mm/page_alloc.c __ClearPageDirty(page); page 470 mm/page_alloc.c return PageReserved(page); page 491 mm/page_alloc.c struct page *page; page 494 mm/page_alloc.c page = list_entry(list->prev, struct page, lru); page 496 mm/page_alloc.c list_del(&page->lru); page 497 mm/page_alloc.c __free_one_page(page, zone, order); page 507 mm/page_alloc.c __free_one_page(page, zone, order); page 518 mm/page_alloc.c reserved += free_pages_check(page + i); page 522 mm/page_alloc.c if (!PageHighMem(page)) { page 523 mm/page_alloc.c debug_check_no_locks_freed(page_address(page),PAGE_SIZE<<order); page 524 mm/page_alloc.c debug_check_no_obj_freed(page_address(page), page 527 mm/page_alloc.c arch_free_page(page, order); page 528 mm/page_alloc.c kernel_map_pages(page, 1 << order, 0); page 532 mm/page_alloc.c free_one_page(page_zone(page), page, order); page 542 mm/page_alloc.c __ClearPageReserved(page); page 543 mm/page_alloc.c set_page_count(page, 0); page 544 mm/page_alloc.c set_page_refcounted(page); page 545 mm/page_alloc.c __free_page(page); page 549 mm/page_alloc.c prefetchw(page); page 551 mm/page_alloc.c struct page *p = &page[loop]; page 559 mm/page_alloc.c set_page_refcounted(page); page 560 mm/page_alloc.c __free_pages(page, order); page 589 mm/page_alloc.c VM_BUG_ON(bad_range(zone, &page[size])); page 590 mm/page_alloc.c list_add(&page[size].lru, &area->free_list[migratetype]); page 592 mm/page_alloc.c set_page_order(&page[size], high); page 601 mm/page_alloc.c if (unlikely(page_mapcount(page) | page 602 mm/page_alloc.c (page->mapping != NULL) | page 603 mm/page_alloc.c (page_get_page_cgroup(page) != NULL) | page 604 mm/page_alloc.c (page_count(page) != 0) | page 605 mm/page_alloc.c (page->flags & PAGE_FLAGS_CHECK_AT_PREP))) page 606 mm/page_alloc.c bad_page(page); page 612 mm/page_alloc.c if (PageReserved(page)) page 615 mm/page_alloc.c page->flags &= ~(1 << PG_uptodate | 1 << PG_error | 1 << PG_reclaim | page 618 mm/page_alloc.c set_page_private(page, 0); page 619 mm/page_alloc.c set_page_refcounted(page); page 621 mm/page_alloc.c arch_alloc_page(page, order); page 622 mm/page_alloc.c kernel_map_pages(page, 1 << order, 1); page 625 mm/page_alloc.c prep_zero_page(page, order, gfp_flags); page 628 mm/page_alloc.c prep_compound_page(page, order); page 637 mm/page_alloc.c static struct page *__rmqueue_smallest(struct zone *zone, unsigned int order, page 642 mm/page_alloc.c struct page *page; page 650 mm/page_alloc.c page = list_entry(area->free_list[migratetype].next, page 651 mm/page_alloc.c struct page, lru); page 652 mm/page_alloc.c list_del(&page->lru); page 653 mm/page_alloc.c rmv_page_order(page); page 656 mm/page_alloc.c expand(zone, page, order, current_order, area, migratetype); page 657 mm/page_alloc.c return page; page 684 mm/page_alloc.c struct page *page; page 699 mm/page_alloc.c for (page = start_page; page <= end_page;) { page 701 mm/page_alloc.c VM_BUG_ON(page_to_nid(page) != zone_to_nid(zone)); page 703 mm/page_alloc.c if (!pfn_valid_within(page_to_pfn(page))) { page 704 mm/page_alloc.c page++; page 708 mm/page_alloc.c if (!PageBuddy(page)) { page 709 mm/page_alloc.c page++; page 713 mm/page_alloc.c order = page_order(page); page 714 mm/page_alloc.c list_del(&page->lru); page 715 mm/page_alloc.c list_add(&page->lru, page 717 mm/page_alloc.c page += 1 << order; page 728 mm/page_alloc.c struct page *start_page, *end_page; page 730 mm/page_alloc.c start_pfn = page_to_pfn(page); page 738 mm/page_alloc.c start_page = page; page 746 mm/page_alloc.c static struct page *__rmqueue_fallback(struct zone *zone, int order, page 751 mm/page_alloc.c struct page *page; page 768 mm/page_alloc.c page = list_entry(area->free_list[migratetype].next, page 769 mm/page_alloc.c struct page, lru); page 781 mm/page_alloc.c pages = move_freepages_block(zone, page, page 786 mm/page_alloc.c set_pageblock_migratetype(page, page 793 mm/page_alloc.c list_del(&page->lru); page 794 mm/page_alloc.c rmv_page_order(page); page 799 mm/page_alloc.c set_pageblock_migratetype(page, page 802 mm/page_alloc.c expand(zone, page, order, current_order, area, migratetype); page 803 mm/page_alloc.c return page; page 815 mm/page_alloc.c static struct page *__rmqueue(struct zone *zone, unsigned int order, page 818 mm/page_alloc.c struct page *page; page 820 mm/page_alloc.c page = __rmqueue_smallest(zone, order, migratetype); page 822 mm/page_alloc.c if (unlikely(!page)) page 823 mm/page_alloc.c page = __rmqueue_fallback(zone, order, migratetype); page 825 mm/page_alloc.c return page; page 841 mm/page_alloc.c struct page *page = __rmqueue(zone, order, migratetype); page 842 mm/page_alloc.c if (unlikely(page == NULL)) page 854 mm/page_alloc.c list_add(&page->lru, list); page 855 mm/page_alloc.c set_page_private(page, migratetype); page 856 mm/page_alloc.c list = &page->lru; page 949 mm/page_alloc.c struct page *page = pfn_to_page(pfn); page 951 mm/page_alloc.c if (!swsusp_page_is_forbidden(page)) page 952 mm/page_alloc.c swsusp_unset_page_free(page); page 959 mm/page_alloc.c pfn = page_to_pfn(list_entry(curr, struct page, lru)); page 973 mm/page_alloc.c struct zone *zone = page_zone(page); page 977 mm/page_alloc.c if (PageAnon(page)) page 978 mm/page_alloc.c page->mapping = NULL; page 979 mm/page_alloc.c if (free_pages_check(page)) page 982 mm/page_alloc.c if (!PageHighMem(page)) { page 983 mm/page_alloc.c debug_check_no_locks_freed(page_address(page), PAGE_SIZE); page 984 mm/page_alloc.c debug_check_no_obj_freed(page_address(page), PAGE_SIZE); page 986 mm/page_alloc.c arch_free_page(page, 0); page 987 mm/page_alloc.c kernel_map_pages(page, 1, 0); page 993 mm/page_alloc.c list_add_tail(&page->lru, &pcp->list); page 995 mm/page_alloc.c list_add(&page->lru, &pcp->list); page 996 mm/page_alloc.c set_page_private(page, get_pageblock_migratetype(page)); page 1008 mm/page_alloc.c free_hot_cold_page(page, 0); page 1013 mm/page_alloc.c free_hot_cold_page(page, 1); page 1028 mm/page_alloc.c VM_BUG_ON(PageCompound(page)); page 1029 mm/page_alloc.c VM_BUG_ON(!page_count(page)); page 1031 mm/page_alloc.c set_page_refcounted(page + i); page 1039 mm/page_alloc.c static struct page *buffered_rmqueue(struct zone *preferred_zone, page 1043 mm/page_alloc.c struct page *page; page 1064 mm/page_alloc.c list_for_each_entry_reverse(page, &pcp->list, lru) page 1065 mm/page_alloc.c if (page_private(page) == migratetype) page 1068 mm/page_alloc.c list_for_each_entry(page, &pcp->list, lru) page 1069 mm/page_alloc.c if (page_private(page) == migratetype) page 1074 mm/page_alloc.c if (unlikely(&page->lru == &pcp->list)) { page 1077 mm/page_alloc.c page = list_entry(pcp->list.next, struct page, lru); page 1080 mm/page_alloc.c list_del(&page->lru); page 1084 mm/page_alloc.c page = __rmqueue(zone, order, migratetype); page 1086 mm/page_alloc.c if (!page) page 1095 mm/page_alloc.c VM_BUG_ON(bad_range(zone, page)); page 1096 mm/page_alloc.c if (prep_new_page(page, order, gfp_flags)) page 1098 mm/page_alloc.c return page; page 1365 mm/page_alloc.c static struct page * page 1370 mm/page_alloc.c struct page *page = NULL; page 1414 mm/page_alloc.c page = buffered_rmqueue(preferred_zone, zone, order, gfp_mask); page 1415 mm/page_alloc.c if (page) page 1429 mm/page_alloc.c if (unlikely(NUMA_BUILD && page == NULL && zlc_active)) { page 1434 mm/page_alloc.c return page; page 1440 mm/page_alloc.c struct page * page 1448 mm/page_alloc.c struct page *page; page 1472 mm/page_alloc.c page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask, order, page 1474 mm/page_alloc.c if (page) page 1517 mm/page_alloc.c page = get_page_from_freelist(gfp_mask, nodemask, order, zonelist, page 1519 mm/page_alloc.c if (page) page 1530 mm/page_alloc.c page = get_page_from_freelist(gfp_mask, nodemask, order, page 1532 mm/page_alloc.c if (page) page 1565 mm/page_alloc.c page = get_page_from_freelist(gfp_mask, nodemask, order, page 1567 mm/page_alloc.c if (page) page 1581 mm/page_alloc.c page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask, page 1584 mm/page_alloc.c if (page) { page 1641 mm/page_alloc.c return page; page 1650 mm/page_alloc.c struct page * page; page 1651 mm/page_alloc.c page = alloc_pages(gfp_mask, order); page 1652 mm/page_alloc.c if (!page) page 1654 mm/page_alloc.c return (unsigned long) page_address(page); page 1661 mm/page_alloc.c struct page * page; page 1669 mm/page_alloc.c page = alloc_pages(gfp_mask | __GFP_ZERO, 0); page 1670 mm/page_alloc.c if (page) page 1671 mm/page_alloc.c return (unsigned long) page_address(page); page 1687 mm/page_alloc.c if (put_page_testzero(page)) { page 1689 mm/page_alloc.c free_hot_page(page); page 1691 mm/page_alloc.c __free_pages_ok(page, order); page 2511 mm/page_alloc.c struct page *page; page 2523 mm/page_alloc.c page = pfn_to_page(pfn); page 2526 mm/page_alloc.c if (page_to_nid(page) != zone_to_nid(zone)) page 2530 mm/page_alloc.c if (PageReserved(page)) page 2533 mm/page_alloc.c block_migratetype = get_pageblock_migratetype(page); page 2543 mm/page_alloc.c set_pageblock_migratetype(page, MIGRATE_RESERVE); page 2544 mm/page_alloc.c move_freepages_block(zone, page, MIGRATE_RESERVE); page 2554 mm/page_alloc.c set_pageblock_migratetype(page, MIGRATE_MOVABLE); page 2555 mm/page_alloc.c move_freepages_block(zone, page, MIGRATE_MOVABLE); page 2568 mm/page_alloc.c struct page *page; page 2586 mm/page_alloc.c page = pfn_to_page(pfn); page 2587 mm/page_alloc.c set_page_links(page, zone, nid, pfn); page 2588 mm/page_alloc.c mminit_verify_page_links(page, zone, nid, pfn); page 2589 mm/page_alloc.c init_page_count(page); page 2590 mm/page_alloc.c reset_page_mapcount(page); page 2591 mm/page_alloc.c SetPageReserved(page); page 2609 mm/page_alloc.c set_pageblock_migratetype(page, MIGRATE_MOVABLE); page 2611 mm/page_alloc.c INIT_LIST_HEAD(&page->lru); page 2615 mm/page_alloc.c set_page_address(page, __va(pfn << PAGE_SHIFT)); page 3428 mm/page_alloc.c PAGE_ALIGN(size * sizeof(struct page)) >> PAGE_SHIFT; page 3497 mm/page_alloc.c struct page *map; page 3507 mm/page_alloc.c size = (end - start) * sizeof(struct page); page 4492 mm/page_alloc.c struct page *pfn_to_page(unsigned long pfn) page 4498 mm/page_alloc.c return __page_to_pfn(page); page 4542 mm/page_alloc.c zone = page_zone(page); page 4543 mm/page_alloc.c pfn = page_to_pfn(page); page 4569 mm/page_alloc.c zone = page_zone(page); page 4570 mm/page_alloc.c pfn = page_to_pfn(page); page 4595 mm/page_alloc.c zone = page_zone(page); page 4600 mm/page_alloc.c if (get_pageblock_migratetype(page) != MIGRATE_MOVABLE) page 4602 mm/page_alloc.c set_pageblock_migratetype(page, MIGRATE_ISOLATE); page 4603 mm/page_alloc.c move_freepages_block(zone, page, MIGRATE_ISOLATE); page 4616 mm/page_alloc.c zone = page_zone(page); page 4618 mm/page_alloc.c if (get_pageblock_migratetype(page) != MIGRATE_ISOLATE) page 4620 mm/page_alloc.c set_pageblock_migratetype(page, MIGRATE_MOVABLE); page 4621 mm/page_alloc.c move_freepages_block(zone, page, MIGRATE_MOVABLE); page 4633 mm/page_alloc.c struct page *page; page 4652 mm/page_alloc.c page = pfn_to_page(pfn); page 4653 mm/page_alloc.c BUG_ON(page_count(page)); page 4654 mm/page_alloc.c BUG_ON(!PageBuddy(page)); page 4655 mm/page_alloc.c order = page_order(page); page 4660 mm/page_alloc.c list_del(&page->lru); page 4661 mm/page_alloc.c rmv_page_order(page); page 4666 mm/page_alloc.c SetPageReserved((page+i)); page 36 mm/page_io.c bio->bi_io_vec[0].bv_page = page; page 50 mm/page_io.c struct page *page = bio->bi_io_vec[0].bv_page; page 53 mm/page_io.c SetPageError(page); page 62 mm/page_io.c set_page_dirty(page); page 67 mm/page_io.c ClearPageReclaim(page); page 69 mm/page_io.c end_page_writeback(page); page 76 mm/page_io.c struct page *page = bio->bi_io_vec[0].bv_page; page 79 mm/page_io.c SetPageError(page); page 80 mm/page_io.c ClearPageUptodate(page); page 86 mm/page_io.c SetPageUptodate(page); page 88 mm/page_io.c unlock_page(page); page 101 mm/page_io.c if (remove_exclusive_swap_page(page)) { page 102 mm/page_io.c unlock_page(page); page 105 mm/page_io.c bio = get_swap_bio(GFP_NOIO, page_private(page), page, page 108 mm/page_io.c set_page_dirty(page); page 109 mm/page_io.c unlock_page(page); page 116 mm/page_io.c set_page_writeback(page); page 117 mm/page_io.c unlock_page(page); page 128 mm/page_io.c BUG_ON(!PageLocked(page)); page 129 mm/page_io.c BUG_ON(PageUptodate(page)); page 130 mm/page_io.c bio = get_swap_bio(GFP_KERNEL, page_private(page), page, page 133 mm/page_io.c unlock_page(page); page 10 mm/page_isolation.c static inline struct page * page 40 mm/page_isolation.c struct page *page; page 48 mm/page_isolation.c page = __first_valid_page(pfn, pageblock_nr_pages); page 49 mm/page_isolation.c if (page && set_migratetype_isolate(page)) { page 71 mm/page_isolation.c struct page *page; page 77 mm/page_isolation.c page = __first_valid_page(pfn, pageblock_nr_pages); page 78 mm/page_isolation.c if (!page || get_pageblock_migratetype(page) != MIGRATE_ISOLATE) page 80 mm/page_isolation.c unset_migratetype_isolate(page); page 94 mm/page_isolation.c struct page *page; page 101 mm/page_isolation.c page = pfn_to_page(pfn); page 102 mm/page_isolation.c if (PageBuddy(page)) page 103 mm/page_isolation.c pfn += 1 << page_order(page); page 104 mm/page_isolation.c else if (page_count(page) == 0 && page 105 mm/page_isolation.c page_private(page) == MIGRATE_ISOLATE) page 118 mm/page_isolation.c struct page *page; page 129 mm/page_isolation.c page = __first_valid_page(pfn, pageblock_nr_pages); page 130 mm/page_isolation.c if (page && get_pageblock_migratetype(page) != MIGRATE_ISOLATE) page 45 mm/readahead.c #define list_to_page(head) (list_entry((head)->prev, struct page, lru)) page 60 mm/readahead.c struct page *page; page 64 mm/readahead.c page = list_to_page(pages); page 65 mm/readahead.c list_del(&page->lru); page 66 mm/readahead.c if (add_to_page_cache_lru(page, mapping, page 67 mm/readahead.c page->index, GFP_KERNEL)) { page 68 mm/readahead.c page_cache_release(page); page 71 mm/readahead.c page_cache_release(page); page 73 mm/readahead.c ret = filler(data, page); page 99 mm/readahead.c struct page *page = list_to_page(pages); page 100 mm/readahead.c list_del(&page->lru); page 101 mm/readahead.c if (!add_to_page_cache_lru(page, mapping, page 102 mm/readahead.c page->index, GFP_KERNEL)) { page 103 mm/readahead.c mapping->a_ops->readpage(filp, page); page 105 mm/readahead.c page_cache_release(page); page 129 mm/readahead.c struct page *page; page 151 mm/readahead.c page = radix_tree_lookup(&mapping->page_tree, page_offset); page 153 mm/readahead.c if (page) page 156 mm/readahead.c page = page_cache_alloc_cold(mapping); page 157 mm/readahead.c if (!page) page 159 mm/readahead.c page->index = page_offset; page 160 mm/readahead.c list_add(&page->lru, &page_pool); page 162 mm/readahead.c SetPageReadahead(page); page 469 mm/readahead.c if (PageWriteback(page)) page 472 mm/readahead.c ClearPageReadahead(page); page 166 mm/rmap.c anon_mapping = (unsigned long) page->mapping; page 169 mm/rmap.c if (!page_mapped(page)) page 194 mm/rmap.c pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); page 211 mm/rmap.c if (PageAnon(page)) { page 213 mm/rmap.c (void *)page->mapping - PAGE_MAPPING_ANON) page 215 mm/rmap.c } else if (page->mapping && !(vma->vm_flags & VM_NONLINEAR)) { page 217 mm/rmap.c vma->vm_file->f_mapping != page->mapping) page 221 mm/rmap.c return vma_address(page, vma); page 263 mm/rmap.c if (pte_present(*pte) && page_to_pfn(page) == pte_pfn(*pte)) { page 284 mm/rmap.c address = vma_address(page, vma); page 288 mm/rmap.c pte = page_check_address(page, mm, address, &ptl, 0); page 318 mm/rmap.c anon_vma = page_lock_anon_vma(page); page 322 mm/rmap.c mapcount = page_mapcount(page); page 331 mm/rmap.c referenced += page_referenced_one(page, vma, &mapcount); page 356 mm/rmap.c struct address_space *mapping = page->mapping; page 357 mm/rmap.c pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); page 367 mm/rmap.c BUG_ON(PageAnon(page)); page 375 mm/rmap.c BUG_ON(!PageLocked(page)); page 383 mm/rmap.c mapcount = page_mapcount(page); page 398 mm/rmap.c referenced += page_referenced_one(page, vma, &mapcount); page 421 mm/rmap.c if (TestClearPageReferenced(page)) page 424 mm/rmap.c if (page_mapped(page) && page->mapping) { page 425 mm/rmap.c if (PageAnon(page)) page 426 mm/rmap.c referenced += page_referenced_anon(page, mem_cont); page 428 mm/rmap.c referenced += page_referenced_file(page, mem_cont); page 429 mm/rmap.c else if (!trylock_page(page)) page 432 mm/rmap.c if (page->mapping) page 434 mm/rmap.c page_referenced_file(page, mem_cont); page 435 mm/rmap.c unlock_page(page); page 439 mm/rmap.c if (page_test_and_clear_young(page)) page 453 mm/rmap.c address = vma_address(page, vma); page 457 mm/rmap.c pte = page_check_address(page, mm, address, &ptl, 1); page 479 mm/rmap.c pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); page 484 mm/rmap.c BUG_ON(PageAnon(page)); page 489 mm/rmap.c ret += page_mkclean_one(page, vma); page 499 mm/rmap.c BUG_ON(!PageLocked(page)); page 501 mm/rmap.c if (page_mapped(page)) { page 502 mm/rmap.c struct address_space *mapping = page_mapping(page); page 504 mm/rmap.c ret = page_mkclean_file(mapping, page); page 505 mm/rmap.c if (page_test_dirty(page)) { page 506 mm/rmap.c page_clear_dirty(page); page 529 mm/rmap.c page->mapping = (struct address_space *) anon_vma; page 531 mm/rmap.c page->index = linear_page_index(vma, address); page 537 mm/rmap.c __inc_zone_page_state(page, NR_ANON_PAGES); page 564 mm/rmap.c BUG_ON(page->mapping != (struct address_space *)anon_vma); page 565 mm/rmap.c BUG_ON(page->index != linear_page_index(vma, address)); page 580 mm/rmap.c VM_BUG_ON(!PageLocked(page)); page 582 mm/rmap.c if (atomic_inc_and_test(&page->_mapcount)) page 583 mm/rmap.c __page_set_anon_rmap(page, vma, address); page 585 mm/rmap.c __page_check_anon_rmap(page, vma, address); page 602 mm/rmap.c atomic_set(&page->_mapcount, 0); /* elevate count by 1 (starts at -1) */ page 603 mm/rmap.c __page_set_anon_rmap(page, vma, address); page 614 mm/rmap.c if (atomic_inc_and_test(&page->_mapcount)) page 615 mm/rmap.c __inc_zone_page_state(page, NR_FILE_MAPPED); page 633 mm/rmap.c BUG_ON(page_mapcount(page) == 0); page 634 mm/rmap.c if (PageAnon(page)) page 635 mm/rmap.c __page_check_anon_rmap(page, vma, address); page 636 mm/rmap.c atomic_inc(&page->_mapcount); page 649 mm/rmap.c if (atomic_add_negative(-1, &page->_mapcount)) { page 650 mm/rmap.c if (unlikely(page_mapcount(page) < 0)) { page 651 mm/rmap.c printk (KERN_EMERG "Eeek! page_mapcount(page) went negative! (%d)\n", page_mapcount(page)); page 652 mm/rmap.c printk (KERN_EMERG " page pfn = %lx\n", page_to_pfn(page)); page 653 mm/rmap.c printk (KERN_EMERG " page->flags = %lx\n", page->flags); page 654 mm/rmap.c printk (KERN_EMERG " page->count = %x\n", page_count(page)); page 655 mm/rmap.c printk (KERN_EMERG " page->mapping = %p\n", page->mapping); page 672 mm/rmap.c if ((!PageAnon(page) || PageSwapCache(page)) && page 673 mm/rmap.c page_test_dirty(page)) { page 674 mm/rmap.c page_clear_dirty(page); page 675 mm/rmap.c set_page_dirty(page); page 678 mm/rmap.c mem_cgroup_uncharge_page(page); page 679 mm/rmap.c __dec_zone_page_state(page, page 680 mm/rmap.c PageAnon(page) ? NR_ANON_PAGES : NR_FILE_MAPPED); page 707 mm/rmap.c address = vma_address(page, vma); page 711 mm/rmap.c pte = page_check_address(page, mm, address, &ptl, 0); page 727 mm/rmap.c flush_cache_page(vma, address, page_to_pfn(page)); page 732 mm/rmap.c set_page_dirty(page); page 737 mm/rmap.c if (PageAnon(page)) { page 738 mm/rmap.c swp_entry_t entry = { .val = page_private(page) }; page 740 mm/rmap.c if (PageSwapCache(page)) { page 761 mm/rmap.c entry = make_migration_entry(page, pte_write(pteval)); page 771 mm/rmap.c entry = make_migration_entry(page, pte_write(pteval)); page 778 mm/rmap.c page_remove_rmap(page, vma); page 779 mm/rmap.c page_cache_release(page); page 819 mm/rmap.c struct page *page; page 850 mm/rmap.c page = vm_normal_page(vma, address, *pte); page 851 mm/rmap.c BUG_ON(!page || PageAnon(page)); page 861 mm/rmap.c if (page->index != linear_page_index(vma, address)) page 862 mm/rmap.c set_pte_at(mm, address, pte, pgoff_to_pte(page->index)); page 866 mm/rmap.c set_page_dirty(page); page 868 mm/rmap.c page_remove_rmap(page, vma); page 869 mm/rmap.c page_cache_release(page); page 882 mm/rmap.c anon_vma = page_lock_anon_vma(page); page 887 mm/rmap.c ret = try_to_unmap_one(page, vma, migration); page 888 mm/rmap.c if (ret == SWAP_FAIL || !page_mapped(page)) page 908 mm/rmap.c struct address_space *mapping = page->mapping; page 909 mm/rmap.c pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); page 920 mm/rmap.c ret = try_to_unmap_one(page, vma, migration); page 921 mm/rmap.c if (ret == SWAP_FAIL || !page_mapped(page)) page 952 mm/rmap.c mapcount = page_mapcount(page); page 1010 mm/rmap.c BUG_ON(!PageLocked(page)); page 1012 mm/rmap.c if (PageAnon(page)) page 1013 mm/rmap.c ret = try_to_unmap_anon(page, migration); page 1015 mm/rmap.c ret = try_to_unmap_file(page, migration); page 1017 mm/rmap.c if (!page_mapped(page)) page 101 mm/shmem.c static inline struct page *shmem_dir_alloc(gfp_t gfp_mask) page 116 mm/shmem.c __free_pages(page, PAGE_CACHE_SHIFT-PAGE_SHIFT); page 119 mm/shmem.c static struct page **shmem_dir_map(struct page *page) page 121 mm/shmem.c return (struct page **)kmap_atomic(page, KM_USER0); page 131 mm/shmem.c return (swp_entry_t *)kmap_atomic(page, KM_USER1); page 310 mm/shmem.c struct page **dir; page 311 mm/shmem.c struct page *subdir; page 318 mm/shmem.c if (page) { page 319 mm/shmem.c info->i_indirect = *page; page 320 mm/shmem.c *page = NULL; page 336 mm/shmem.c if (page) { page 337 mm/shmem.c *dir = *page; page 338 mm/shmem.c *page = NULL; page 350 mm/shmem.c if (!page || !(subdir = *page)) { page 355 mm/shmem.c *page = NULL; page 368 mm/shmem.c struct page *page = kmap_atomic_to_page(entry); page 369 mm/shmem.c set_page_private(page, page_private(page) + incdec); page 385 mm/shmem.c struct page *page = NULL; page 392 mm/shmem.c while (!(entry = shmem_swp_entry(info, index, &page))) { page 412 mm/shmem.c page = shmem_dir_alloc(mapping_gfp_mask(inode->i_mapping)); page 413 mm/shmem.c if (page) page 414 mm/shmem.c set_page_private(page, 0); page 417 mm/shmem.c if (!page) { page 429 mm/shmem.c if (page) { page 432 mm/shmem.c shmem_dir_free(page); page 500 mm/shmem.c struct page *page; page 504 mm/shmem.c page = container_of(next, struct page, lru); page 506 mm/shmem.c shmem_dir_free(page); page 523 mm/shmem.c struct page **dir; page 524 mm/shmem.c struct page *topdir; page 525 mm/shmem.c struct page *middir; page 526 mm/shmem.c struct page *subdir; page 750 mm/shmem.c struct page *page = NULL; page 765 mm/shmem.c &page, SGP_READ, NULL); page 766 mm/shmem.c if (page) page 767 mm/shmem.c unlock_page(page); page 792 mm/shmem.c if (page) page 793 mm/shmem.c page_cache_release(page); page 835 mm/shmem.c struct page **dir; page 836 mm/shmem.c struct page *subdir; page 924 mm/shmem.c error = mem_cgroup_cache_charge(page, current->mm, GFP_KERNEL); page 929 mm/shmem.c mem_cgroup_uncharge_cache_page(page); page 937 mm/shmem.c error = add_to_page_cache_locked(page, inode->i_mapping, page 941 mm/shmem.c mem_cgroup_uncharge_cache_page(page); page 944 mm/shmem.c struct page *filepage = find_get_page(inode->i_mapping, idx); page 957 mm/shmem.c delete_from_swap_cache(page); page 958 mm/shmem.c set_page_dirty(page); page 969 mm/shmem.c unlock_page(page); page 970 mm/shmem.c page_cache_release(page); page 987 mm/shmem.c found = shmem_unuse_inode(info, entry, page); page 1007 mm/shmem.c BUG_ON(!PageLocked(page)); page 1008 mm/shmem.c mapping = page->mapping; page 1009 mm/shmem.c index = page->index; page 1046 mm/shmem.c if (swap.val && add_to_swap_cache(page, swap, GFP_ATOMIC) == 0) { page 1047 mm/shmem.c remove_from_page_cache(page); page 1056 mm/shmem.c BUG_ON(page_mapped(page)); page 1057 mm/shmem.c page_cache_release(page); /* pagecache ref */ page 1058 mm/shmem.c set_page_dirty(page); page 1059 mm/shmem.c unlock_page(page); page 1075 mm/shmem.c set_page_dirty(page); page 1078 mm/shmem.c unlock_page(page); page 1109 mm/shmem.c static struct page *shmem_swapin(swp_entry_t entry, gfp_t gfp, page 1114 mm/shmem.c struct page *page; page 1124 mm/shmem.c page = swapin_readahead(entry, gfp, &pvma, 0); page 1125 mm/shmem.c return page; page 1128 mm/shmem.c static struct page *shmem_alloc_page(gfp_t gfp, page 1151 mm/shmem.c static inline struct page *shmem_swapin(swp_entry_t entry, gfp_t gfp, page 1157 mm/shmem.c static inline struct page *shmem_alloc_page(gfp_t gfp, page 1184 mm/shmem.c struct page *filepage = *pagep; page 1185 mm/shmem.c struct page *swappage; page 1442 mm/shmem.c error = shmem_getpage(inode, vmf->pgoff, &vmf->page, SGP_CACHE, &ret); page 1446 mm/shmem.c mark_page_accessed(vmf->page); page 1565 mm/shmem.c struct inode *inode = page->mapping->host; page 1566 mm/shmem.c int error = shmem_getpage(inode, page->index, &page, SGP_CACHE, NULL); page 1567 mm/shmem.c unlock_page(page); page 1592 mm/shmem.c unlock_page(page); page 1593 mm/shmem.c set_page_dirty(page); page 1594 mm/shmem.c page_cache_release(page); page 1618 mm/shmem.c struct page *page = NULL; page 1631 mm/shmem.c desc->error = shmem_getpage(inode, index, &page, sgp, NULL); page 1637 mm/shmem.c if (page) page 1638 mm/shmem.c unlock_page(page); page 1650 mm/shmem.c if (page) page 1651 mm/shmem.c page_cache_release(page); page 1657 mm/shmem.c if (page) { page 1664 mm/shmem.c flush_dcache_page(page); page 1669 mm/shmem.c mark_page_accessed(page); page 1671 mm/shmem.c page = ZERO_PAGE(0); page 1672 mm/shmem.c page_cache_get(page); page 1685 mm/shmem.c ret = actor(desc, page, offset, nr); page 1690 mm/shmem.c page_cache_release(page); page 1895 mm/shmem.c struct page *page = NULL; page 1924 mm/shmem.c error = shmem_getpage(inode, 0, &page, SGP_WRITE, NULL); page 1929 mm/shmem.c unlock_page(page); page 1932 mm/shmem.c kaddr = kmap_atomic(page, KM_USER0); page 1935 mm/shmem.c set_page_dirty(page); page 1936 mm/shmem.c page_cache_release(page); page 1955 mm/shmem.c struct page *page = NULL; page 1956 mm/shmem.c int res = shmem_getpage(dentry->d_inode, 0, &page, SGP_READ, NULL); page 1957 mm/shmem.c nd_set_link(nd, res ? ERR_PTR(res) : kmap(page)); page 1958 mm/shmem.c if (page) page 1959 mm/shmem.c unlock_page(page); page 1960 mm/shmem.c return page; page 1966 mm/shmem.c struct page *page = cookie; page 1967 mm/shmem.c kunmap(page); page 1968 mm/shmem.c mark_page_accessed(page); page 1969 mm/shmem.c page_cache_release(page); page 584 mm/slab.c page->lru.next = (struct list_head *)cache; page 589 mm/slab.c page = compound_head(page); page 590 mm/slab.c BUG_ON(!PageSlab(page)); page 591 mm/slab.c return (struct kmem_cache *)page->lru.next; page 596 mm/slab.c page->lru.prev = (struct list_head *)slab; page 601 mm/slab.c BUG_ON(!PageSlab(page)); page 602 mm/slab.c return (struct slab *)page->lru.prev; page 607 mm/slab.c struct page *page = virt_to_head_page(obj); page 608 mm/slab.c return page_get_cache(page); page 613 mm/slab.c struct page *page = virt_to_head_page(obj); page 614 mm/slab.c return page_get_slab(page); page 1666 mm/slab.c struct page *page; page 1682 mm/slab.c page = alloc_pages_node(nodeid, flags, cachep->gfporder); page 1683 mm/slab.c if (!page) page 1688 mm/slab.c add_zone_page_state(page_zone(page), page 1691 mm/slab.c add_zone_page_state(page_zone(page), page 1694 mm/slab.c __SetPageSlab(page + i); page 1695 mm/slab.c return page_address(page); page 1704 mm/slab.c struct page *page = virt_to_page(addr); page 1708 mm/slab.c sub_zone_page_state(page_zone(page), page 1711 mm/slab.c sub_zone_page_state(page_zone(page), page 1714 mm/slab.c BUG_ON(!PageSlab(page)); page 1715 mm/slab.c __ClearPageSlab(page); page 1716 mm/slab.c page++; page 2734 mm/slab.c struct page *page; page 2736 mm/slab.c page = virt_to_page(addr); page 2739 mm/slab.c if (likely(!PageCompound(page))) page 2743 mm/slab.c page_set_cache(page, cache); page 2744 mm/slab.c page_set_slab(page, slab); page 2745 mm/slab.c page++; page 2872 mm/slab.c struct page *page; page 2880 mm/slab.c page = virt_to_head_page(objp); page 2882 mm/slab.c slabp = page_get_slab(page); page 3638 mm/slab.c struct page *page; page 3650 mm/slab.c page = virt_to_page(ptr); page 3651 mm/slab.c if (unlikely(!PageSlab(page))) page 3653 mm/slab.c if (unlikely(page_get_cache(page) != cachep)) page 104 mm/slob.c struct page page; page 108 mm/slob.c { BUILD_BUG_ON(sizeof(struct slob_page) != sizeof(struct page)); } page 115 mm/slob.c reset_page_mapcount(&sp->page); page 116 mm/slob.c sp->page.mapping = NULL; page 133 mm/slob.c return PageSlobPage((struct page *)sp); page 138 mm/slob.c __SetPageSlobPage((struct page *)sp); page 143 mm/slob.c __ClearPageSlobPage((struct page *)sp); page 151 mm/slob.c return PageSlobFree((struct page *)sp); page 157 mm/slob.c __SetPageSlobFree((struct page *)sp); page 163 mm/slob.c __ClearPageSlobFree((struct page *)sp); page 235 mm/slob.c void *page; page 239 mm/slob.c page = alloc_pages_node(node, gfp, order); page 242 mm/slob.c page = alloc_pages(gfp, order); page 244 mm/slob.c if (!page) page 247 mm/slob.c return page_address(page); page 327 mm/slob.c if (node != -1 && page_to_nid(&sp->page) != node) page 481 mm/slob.c struct page *page; page 482 mm/slob.c page = virt_to_page(ret); page 483 mm/slob.c page->private = size; page 503 mm/slob.c put_page(&sp->page); page 522 mm/slob.c return sp->page.private; page 247 mm/slub.c base = page_address(page); page 248 mm/slub.c if (object < base || object >= base + page->objects * s->size || page 418 mm/slub.c page, page->objects, page->inuse, page->freelist, page->flags); page 451 mm/slub.c u8 *addr = page_address(page); page 455 mm/slub.c print_page_info(page); page 488 mm/slub.c print_trailer(s, page, object); page 500 mm/slub.c print_page_info(page); page 555 mm/slub.c print_trailer(s, page, object); page 614 mm/slub.c return check_bytes_and_report(s, page, p, "Object padding", page 630 mm/slub.c start = page_address(page); page 631 mm/slub.c length = (PAGE_SIZE << compound_order(page)); page 643 mm/slub.c slab_err(s, page, "Padding overwritten. 0x%p-0x%p", fault, end - 1); page 660 mm/slub.c if (!check_bytes_and_report(s, page, object, "Redzone", page 665 mm/slub.c check_bytes_and_report(s, page, p, "Alignment padding", page 672 mm/slub.c (!check_bytes_and_report(s, page, p, "Poison", p, page 674 mm/slub.c !check_bytes_and_report(s, page, p, "Poison", page 680 mm/slub.c check_pad_bytes(s, page, p); page 691 mm/slub.c if (!check_valid_pointer(s, page, get_freepointer(s, p))) { page 692 mm/slub.c object_err(s, page, p, "Freepointer corrupt"); page 710 mm/slub.c if (!PageSlab(page)) { page 711 mm/slub.c slab_err(s, page, "Not a valid slab page"); page 715 mm/slub.c maxobj = (PAGE_SIZE << compound_order(page)) / s->size; page 716 mm/slub.c if (page->objects > maxobj) { page 717 mm/slub.c slab_err(s, page, "objects %u > max %u", page 718 mm/slub.c s->name, page->objects, maxobj); page 721 mm/slub.c if (page->inuse > page->objects) { page 722 mm/slub.c slab_err(s, page, "inuse %u > max %u", page 723 mm/slub.c s->name, page->inuse, page->objects); page 727 mm/slub.c slab_pad_check(s, page); page 738 mm/slub.c void *fp = page->freelist; page 742 mm/slub.c while (fp && nr <= page->objects) { page 745 mm/slub.c if (!check_valid_pointer(s, page, fp)) { page 747 mm/slub.c object_err(s, page, object, page 752 mm/slub.c slab_err(s, page, "Freepointer corrupt"); page 753 mm/slub.c page->freelist = NULL; page 754 mm/slub.c page->inuse = page->objects; page 765 mm/slub.c max_objects = (PAGE_SIZE << compound_order(page)) / s->size; page 769 mm/slub.c if (page->objects != max_objects) { page 770 mm/slub.c slab_err(s, page, "Wrong number of objects. Found %d but " page 771 mm/slub.c "should be %d", page->objects, max_objects); page 772 mm/slub.c page->objects = max_objects; page 775 mm/slub.c if (page->inuse != page->objects - nr) { page 776 mm/slub.c slab_err(s, page, "Wrong object count. Counter is %d but " page 777 mm/slub.c "counted were %d", page->inuse, page->objects - nr); page 778 mm/slub.c page->inuse = page->objects - nr; page 791 mm/slub.c object, page->inuse, page 792 mm/slub.c page->freelist); page 807 mm/slub.c list_add(&page->lru, &n->full); page 818 mm/slub.c n = get_node(s, page_to_nid(page)); page 821 mm/slub.c list_del(&page->lru); page 870 mm/slub.c if (!check_slab(s, page)) page 873 mm/slub.c if (!on_freelist(s, page, object)) { page 874 mm/slub.c object_err(s, page, object, "Object already allocated"); page 878 mm/slub.c if (!check_valid_pointer(s, page, object)) { page 879 mm/slub.c object_err(s, page, object, "Freelist Pointer check fails"); page 883 mm/slub.c if (!check_object(s, page, object, 0)) page 889 mm/slub.c trace(s, page, object, 1); page 894 mm/slub.c if (PageSlab(page)) { page 901 mm/slub.c page->inuse = page->objects; page 902 mm/slub.c page->freelist = NULL; page 910 mm/slub.c if (!check_slab(s, page)) page 913 mm/slub.c if (!check_valid_pointer(s, page, object)) { page 914 mm/slub.c slab_err(s, page, "Invalid object pointer 0x%p", object); page 918 mm/slub.c if (on_freelist(s, page, object)) { page 919 mm/slub.c object_err(s, page, object, "Object already free"); page 923 mm/slub.c if (!check_object(s, page, object, 1)) page 926 mm/slub.c if (unlikely(s != page->slab)) { page 927 mm/slub.c if (!PageSlab(page)) { page 928 mm/slub.c slab_err(s, page, "Attempt to free object(0x%p) " page 930 mm/slub.c } else if (!page->slab) { page 936 mm/slub.c object_err(s, page, object, page 942 mm/slub.c if (!PageSlubFrozen(page) && !page->freelist) page 943 mm/slub.c remove_full(s, page); page 946 mm/slub.c trace(s, page, object, 0); page 1060 mm/slub.c static inline struct page *alloc_slab_page(gfp_t flags, int node, page 1071 mm/slub.c static struct page *allocate_slab(struct kmem_cache *s, gfp_t flags, int node) page 1073 mm/slub.c struct page *page; page 1078 mm/slub.c page = alloc_slab_page(flags | __GFP_NOWARN | __GFP_NORETRY, node, page 1080 mm/slub.c if (unlikely(!page)) { page 1086 mm/slub.c page = alloc_slab_page(flags, node, oo); page 1087 mm/slub.c if (!page) page 1092 mm/slub.c page->objects = oo_objects(oo); page 1093 mm/slub.c mod_zone_page_state(page_zone(page), page 1098 mm/slub.c return page; page 1104 mm/slub.c setup_object_debug(s, page, object); page 1109 mm/slub.c static struct page *new_slab(struct kmem_cache *s, gfp_t flags, int node) page 1111 mm/slub.c struct page *page; page 1118 mm/slub.c page = allocate_slab(s, page 1120 mm/slub.c if (!page) page 1123 mm/slub.c inc_slabs_node(s, page_to_nid(page), page->objects); page 1124 mm/slub.c page->slab = s; page 1125 mm/slub.c page->flags |= 1 << PG_slab; page 1128 mm/slub.c __SetPageSlubDebug(page); page 1130 mm/slub.c start = page_address(page); page 1133 mm/slub.c memset(start, POISON_INUSE, PAGE_SIZE << compound_order(page)); page 1136 mm/slub.c for_each_object(p, s, start, page->objects) { page 1137 mm/slub.c setup_object(s, page, last); page 1141 mm/slub.c setup_object(s, page, last); page 1144 mm/slub.c page->freelist = start; page 1145 mm/slub.c page->inuse = 0; page 1147 mm/slub.c return page; page 1152 mm/slub.c int order = compound_order(page); page 1155 mm/slub.c if (unlikely(SLABDEBUG && PageSlubDebug(page))) { page 1158 mm/slub.c slab_pad_check(s, page); page 1159 mm/slub.c for_each_object(p, s, page_address(page), page 1160 mm/slub.c page->objects) page 1161 mm/slub.c check_object(s, page, p, 0); page 1162 mm/slub.c __ClearPageSlubDebug(page); page 1165 mm/slub.c mod_zone_page_state(page_zone(page), page 1170 mm/slub.c __ClearPageSlab(page); page 1171 mm/slub.c reset_page_mapcount(page); page 1172 mm/slub.c __free_pages(page, order); page 1177 mm/slub.c struct page *page; page 1179 mm/slub.c page = container_of((struct list_head *)h, struct page, lru); page 1180 mm/slub.c __free_slab(page->slab, page); page 1189 mm/slub.c struct rcu_head *head = (void *)&page->lru; page 1193 mm/slub.c __free_slab(s, page); page 1198 mm/slub.c dec_slabs_node(s, page_to_nid(page), page->objects); page 1199 mm/slub.c free_slab(s, page); page 1207 mm/slub.c bit_spin_lock(PG_locked, &page->flags); page 1212 mm/slub.c __bit_spin_unlock(PG_locked, &page->flags); page 1219 mm/slub.c rc = bit_spin_trylock(PG_locked, &page->flags); page 1232 mm/slub.c list_add_tail(&page->lru, &n->partial); page 1234 mm/slub.c list_add(&page->lru, &n->partial); page 1240 mm/slub.c struct kmem_cache_node *n = get_node(s, page_to_nid(page)); page 1243 mm/slub.c list_del(&page->lru); page 1256 mm/slub.c if (slab_trylock(page)) { page 1257 mm/slub.c list_del(&page->lru); page 1259 mm/slub.c __SetPageSlubFrozen(page); page 1268 mm/slub.c static struct page *get_partial_node(struct kmem_cache_node *n) page 1270 mm/slub.c struct page *page; page 1282 mm/slub.c list_for_each_entry(page, &n->partial, lru) page 1283 mm/slub.c if (lock_and_freeze_slab(n, page)) page 1285 mm/slub.c page = NULL; page 1288 mm/slub.c return page; page 1294 mm/slub.c static struct page *get_any_partial(struct kmem_cache *s, gfp_t flags) page 1301 mm/slub.c struct page *page; page 1333 mm/slub.c page = get_partial_node(n); page 1334 mm/slub.c if (page) page 1335 mm/slub.c return page; page 1345 mm/slub.c static struct page *get_partial(struct kmem_cache *s, gfp_t flags, int node) page 1347 mm/slub.c struct page *page; page 1350 mm/slub.c page = get_partial_node(get_node(s, searchnode)); page 1351 mm/slub.c if (page || (flags & __GFP_THISNODE)) page 1352 mm/slub.c return page; page 1366 mm/slub.c struct kmem_cache_node *n = get_node(s, page_to_nid(page)); page 1369 mm/slub.c __ClearPageSlubFrozen(page); page 1370 mm/slub.c if (page->inuse) { page 1372 mm/slub.c if (page->freelist) { page 1373 mm/slub.c add_partial(n, page, tail); page 1377 mm/slub.c if (SLABDEBUG && PageSlubDebug(page) && page 1379 mm/slub.c add_full(n, page); page 1381 mm/slub.c slab_unlock(page); page 1395 mm/slub.c add_partial(n, page, 1); page 1396 mm/slub.c slab_unlock(page); page 1398 mm/slub.c slab_unlock(page); page 1400 mm/slub.c discard_slab(s, page); page 1410 mm/slub.c struct page *page = c->page; page 1413 mm/slub.c if (page->freelist) page 1430 mm/slub.c object[c->offset] = page->freelist; page 1431 mm/slub.c page->freelist = object; page 1432 mm/slub.c page->inuse--; page 1434 mm/slub.c c->page = NULL; page 1435 mm/slub.c unfreeze_slab(s, page, tail); page 1441 mm/slub.c slab_lock(c->page); page 1454 mm/slub.c if (likely(c && c->page)) page 1505 mm/slub.c struct page *new; page 1510 mm/slub.c if (!c->page) page 1513 mm/slub.c slab_lock(c->page); page 1520 mm/slub.c object = c->page->freelist; page 1523 mm/slub.c if (unlikely(SLABDEBUG && PageSlubDebug(c->page))) page 1527 mm/slub.c c->page->inuse = c->page->objects; page 1528 mm/slub.c c->page->freelist = NULL; page 1529 mm/slub.c c->node = page_to_nid(c->page); page 1531 mm/slub.c slab_unlock(c->page); page 1541 mm/slub.c c->page = new; page 1557 mm/slub.c if (c->page) page 1561 mm/slub.c c->page = new; page 1566 mm/slub.c if (!alloc_debug_processing(s, c->page, object, addr)) page 1569 mm/slub.c c->page->inuse++; page 1570 mm/slub.c c->page->freelist = object[c->offset]; page 1644 mm/slub.c slab_lock(page); page 1646 mm/slub.c if (unlikely(SLABDEBUG && PageSlubDebug(page))) page 1650 mm/slub.c prior = object[offset] = page->freelist; page 1651 mm/slub.c page->freelist = object; page 1652 mm/slub.c page->inuse--; page 1654 mm/slub.c if (unlikely(PageSlubFrozen(page))) { page 1659 mm/slub.c if (unlikely(!page->inuse)) page 1667 mm/slub.c add_partial(get_node(s, page_to_nid(page)), page, 1); page 1672 mm/slub.c slab_unlock(page); page 1680 mm/slub.c remove_partial(s, page); page 1683 mm/slub.c slab_unlock(page); page 1685 mm/slub.c discard_slab(s, page); page 1689 mm/slub.c if (!free_debug_processing(s, page, x, addr)) page 1717 mm/slub.c if (likely(page == c->page && c->node >= 0)) { page 1722 mm/slub.c __slab_free(s, page, x, addr, c->offset); page 1729 mm/slub.c struct page *page; page 1731 mm/slub.c page = virt_to_head_page(x); page 1733 mm/slub.c slab_free(s, page, x, __builtin_return_address(0)); page 1738 mm/slub.c static struct page *get_object_page(const void *x) page 1740 mm/slub.c struct page *page = virt_to_head_page(x); page 1742 mm/slub.c if (!PageSlab(page)) page 1745 mm/slub.c return page; page 1906 mm/slub.c c->page = NULL; page 2078 mm/slub.c struct page *page; page 2084 mm/slub.c page = new_slab(kmalloc_caches, gfpflags, node); page 2086 mm/slub.c BUG_ON(!page); page 2087 mm/slub.c if (page_to_nid(page) != node) { page 2094 mm/slub.c n = page->freelist; page 2096 mm/slub.c page->freelist = get_freepointer(kmalloc_caches, n); page 2097 mm/slub.c page->inuse++; page 2104 mm/slub.c inc_slabs_node(kmalloc_caches, node, page->objects); page 2112 mm/slub.c add_partial(n, page, 0); page 2338 mm/slub.c struct page *page; page 2340 mm/slub.c page = get_object_page(object); page 2342 mm/slub.c if (!page || s != page->slab) page 2346 mm/slub.c if (!check_valid_pointer(s, page, object)) page 2378 mm/slub.c void *addr = page_address(page); page 2380 mm/slub.c DECLARE_BITMAP(map, page->objects); page 2382 mm/slub.c bitmap_zero(map, page->objects); page 2383 mm/slub.c slab_err(s, page, "%s", text); page 2384 mm/slub.c slab_lock(page); page 2385 mm/slub.c for_each_free_object(p, s, page->freelist) page 2388 mm/slub.c for_each_object(p, s, addr, page->objects) { page 2396 mm/slub.c slab_unlock(page); page 2406 mm/slub.c struct page *page, *h; page 2409 mm/slub.c list_for_each_entry_safe(page, h, &n->partial, lru) { page 2410 mm/slub.c if (!page->inuse) { page 2411 mm/slub.c list_del(&page->lru); page 2412 mm/slub.c discard_slab(s, page); page 2415 mm/slub.c list_slab_objects(s, page, page 2667 mm/slub.c struct page *page = alloc_pages_node(node, flags | __GFP_COMP, page 2670 mm/slub.c if (page) page 2671 mm/slub.c return page_address(page); page 2696 mm/slub.c struct page *page; page 2702 mm/slub.c page = virt_to_head_page(object); page 2704 mm/slub.c if (unlikely(!PageSlab(page))) { page 2705 mm/slub.c WARN_ON(!PageCompound(page)); page 2706 mm/slub.c return PAGE_SIZE << compound_order(page); page 2708 mm/slub.c s = page->slab; page 2734 mm/slub.c struct page *page; page 2740 mm/slub.c page = virt_to_head_page(x); page 2741 mm/slub.c if (unlikely(!PageSlab(page))) { page 2742 mm/slub.c BUG_ON(!PageCompound(page)); page 2743 mm/slub.c put_page(page); page 2746 mm/slub.c slab_free(page->slab, page, object, __builtin_return_address(0)); page 2765 mm/slub.c struct page *page; page 2766 mm/slub.c struct page *t; page 2793 mm/slub.c list_for_each_entry_safe(page, t, &n->partial, lru) { page 2794 mm/slub.c if (!page->inuse && slab_trylock(page)) { page 2800 mm/slub.c list_del(&page->lru); page 2802 mm/slub.c slab_unlock(page); page 2803 mm/slub.c discard_slab(s, page); page 2805 mm/slub.c list_move(&page->lru, page 2806 mm/slub.c slabs_by_inuse + page->inuse); page 3239 mm/slub.c struct page *page; page 3242 mm/slub.c list_for_each_entry(page, &n->partial, lru) page 3243 mm/slub.c x += get_count(page); page 3250 mm/slub.c return page->inuse; page 3255 mm/slub.c return page->objects; page 3260 mm/slub.c return page->objects - page->inuse; page 3267 mm/slub.c void *addr = page_address(page); page 3269 mm/slub.c if (!check_slab(s, page) || page 3270 mm/slub.c !on_freelist(s, page, NULL)) page 3274 mm/slub.c bitmap_zero(map, page->objects); page 3276 mm/slub.c for_each_free_object(p, s, page->freelist) { page 3278 mm/slub.c if (!check_object(s, page, p, 0)) page 3282 mm/slub.c for_each_object(p, s, addr, page->objects) page 3284 mm/slub.c if (!check_object(s, page, p, 1)) page 3292 mm/slub.c if (slab_trylock(page)) { page 3293 mm/slub.c validate_slab(s, page, map); page 3294 mm/slub.c slab_unlock(page); page 3297 mm/slub.c s->name, page); page 3300 mm/slub.c if (!PageSlubDebug(page)) page 3302 mm/slub.c "on slab 0x%p\n", s->name, page); page 3304 mm/slub.c if (PageSlubDebug(page)) page 3306 mm/slub.c "slab 0x%p\n", s->name, page); page 3314 mm/slub.c struct page *page; page 3319 mm/slub.c list_for_each_entry(page, &n->partial, lru) { page 3320 mm/slub.c validate_slab_slab(s, page, map); page 3330 mm/slub.c list_for_each_entry(page, &n->full, lru) { page 3331 mm/slub.c validate_slab_slab(s, page, map); page 3550 mm/slub.c void *addr = page_address(page); page 3551 mm/slub.c DECLARE_BITMAP(map, page->objects); page 3554 mm/slub.c bitmap_zero(map, page->objects); page 3555 mm/slub.c for_each_free_object(p, s, page->freelist) page 3558 mm/slub.c for_each_object(p, s, addr, page->objects) page 3581 mm/slub.c struct page *page; page 3587 mm/slub.c list_for_each_entry(page, &n->partial, lru) page 3588 mm/slub.c process_slab(&t, s, page, alloc); page 3589 mm/slub.c list_for_each_entry(page, &n->full, lru) page 3590 mm/slub.c process_slab(&t, s, page, alloc); page 3682 mm/slub.c if (c->page) { page 3684 mm/slub.c x = c->page->objects; page 3686 mm/slub.c x = c->page->inuse; page 51 mm/sparse-vmemmap.c struct page *page = alloc_pages_node(node, page 53 mm/sparse-vmemmap.c if (page) page 54 mm/sparse-vmemmap.c return page_address(page); page 151 mm/sparse-vmemmap.c struct page * __meminit sparse_mem_map_populate(unsigned long pnum, int nid) page 153 mm/sparse-vmemmap.c struct page *map = pfn_to_page(pnum * PAGES_PER_SECTION); page 44 mm/sparse.c return section_to_node_table[page_to_section(page)]; page 218 mm/sparse.c return nr_pages * sizeof(struct page); page 234 mm/sparse.c struct page *sparse_decode_mem_map(unsigned long coded_mem_map, unsigned long pnum) page 238 mm/sparse.c return ((struct page *)coded_mem_map) + section_nr_to_pfn(pnum); page 365 mm/sparse.c struct page __init *sparse_mem_map_populate(unsigned long pnum, int nid) page 367 mm/sparse.c struct page *map; page 369 mm/sparse.c map = alloc_remap(nid, sizeof(struct page) * PAGES_PER_SECTION); page 374 mm/sparse.c PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION)); page 379 mm/sparse.c static struct page __init *sparse_early_mem_map_alloc(unsigned long pnum) page 381 mm/sparse.c struct page *map; page 405 mm/sparse.c struct page *map; page 455 mm/sparse.c static inline struct page *kmalloc_section_memmap(unsigned long pnum, int nid, page 469 mm/sparse.c static struct page *__kmalloc_section_memmap(unsigned long nr_pages) page 471 mm/sparse.c struct page *page, *ret; page 472 mm/sparse.c unsigned long memmap_size = sizeof(struct page) * nr_pages; page 474 mm/sparse.c page = alloc_pages(GFP_KERNEL|__GFP_NOWARN, get_order(memmap_size)); page 475 mm/sparse.c if (page) page 484 mm/sparse.c ret = (struct page *)pfn_to_kaddr(page_to_pfn(page)); page 491 mm/sparse.c static inline struct page *kmalloc_section_memmap(unsigned long pnum, int nid, page 503 mm/sparse.c get_order(sizeof(struct page) * nr_pages)); page 511 mm/sparse.c for (i = 0; i < nr_pages; i++, page++) { page 512 mm/sparse.c magic = atomic_read(&page->_mapcount); page 516 mm/sparse.c maps_section_nr = pfn_to_section_nr(page_to_pfn(page)); page 517 mm/sparse.c removing_section_nr = page->private; page 528 mm/sparse.c put_page_bootmem(page); page 535 mm/sparse.c struct page *usemap_page; page 558 mm/sparse.c struct page *memmap_page; page 561 mm/sparse.c nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page)) page 579 mm/sparse.c struct page *memmap; page 623 mm/sparse.c struct page *memmap = NULL; page 47 mm/swap.c if (PageLRU(page)) { page 49 mm/swap.c struct zone *zone = page_zone(page); page 52 mm/swap.c VM_BUG_ON(!PageLRU(page)); page 53 mm/swap.c __ClearPageLRU(page); page 54 mm/swap.c del_page_from_lru(zone, page); page 57 mm/swap.c free_hot_page(page); page 62 mm/swap.c page = compound_head(page); page 63 mm/swap.c if (put_page_testzero(page)) { page 66 mm/swap.c dtor = get_compound_page_dtor(page); page 67 mm/swap.c (*dtor)(page); page 73 mm/swap.c if (unlikely(PageCompound(page))) page 74 mm/swap.c put_compound_page(page); page 75 mm/swap.c else if (put_page_testzero(page)) page 76 mm/swap.c __page_cache_release(page); page 90 mm/swap.c struct page *victim; page 92 mm/swap.c victim = list_entry(pages->prev, struct page, lru); page 110 mm/swap.c struct page *page = pvec->pages[i]; page 111 mm/swap.c struct zone *pagezone = page_zone(page); page 119 mm/swap.c if (PageLRU(page) && !PageActive(page)) { page 120 mm/swap.c list_move_tail(&page->lru, &zone->inactive_list); page 138 mm/swap.c if (!PageLocked(page) && !PageDirty(page) && !PageActive(page) && page 139 mm/swap.c PageLRU(page)) { page 143 mm/swap.c page_cache_get(page); page 146 mm/swap.c if (!pagevec_add(pvec, page)) page 157 mm/swap.c struct zone *zone = page_zone(page); page 160 mm/swap.c if (PageLRU(page) && !PageActive(page)) { page 161 mm/swap.c del_page_from_inactive_list(zone, page); page 162 mm/swap.c SetPageActive(page); page 163 mm/swap.c add_page_to_active_list(zone, page); page 165 mm/swap.c mem_cgroup_move_lists(page, true); page 179 mm/swap.c if (!PageActive(page) && PageReferenced(page) && PageLRU(page)) { page 180 mm/swap.c activate_page(page); page 181 mm/swap.c ClearPageReferenced(page); page 182 mm/swap.c } else if (!PageReferenced(page)) { page 183 mm/swap.c SetPageReferenced(page); page 197 mm/swap.c page_cache_get(page); page 198 mm/swap.c if (!pagevec_add(pvec, page)) page 207 mm/swap.c page_cache_get(page); page 208 mm/swap.c if (!pagevec_add(pvec, page)) page 295 mm/swap.c struct page *page = pages[i]; page 297 mm/swap.c if (unlikely(PageCompound(page))) { page 302 mm/swap.c put_compound_page(page); page 306 mm/swap.c if (!put_page_testzero(page)) page 309 mm/swap.c if (PageLRU(page)) { page 310 mm/swap.c struct zone *pagezone = page_zone(page); page 318 mm/swap.c VM_BUG_ON(!PageLRU(page)); page 319 mm/swap.c __ClearPageLRU(page); page 320 mm/swap.c del_page_from_lru(zone, page); page 323 mm/swap.c if (!pagevec_add(&pages_to_free, page)) { page 369 mm/swap.c struct page *page = pvec->pages[i]; page 371 mm/swap.c VM_BUG_ON(PageLRU(page)); page 372 mm/swap.c if (put_page_testzero(page)) page 373 mm/swap.c pagevec_add(&pages_to_free, page); page 389 mm/swap.c struct page *page = pvec->pages[i]; page 390 mm/swap.c struct zone *pagezone = page_zone(page); page 398 mm/swap.c VM_BUG_ON(PageLRU(page)); page 399 mm/swap.c SetPageLRU(page); page 400 mm/swap.c add_page_to_inactive_list(zone, page); page 416 mm/swap.c struct page *page = pvec->pages[i]; page 417 mm/swap.c struct zone *pagezone = page_zone(page); page 425 mm/swap.c VM_BUG_ON(PageLRU(page)); page 426 mm/swap.c SetPageLRU(page); page 427 mm/swap.c VM_BUG_ON(PageActive(page)); page 428 mm/swap.c SetPageActive(page); page 429 mm/swap.c add_page_to_active_list(zone, page); page 445 mm/swap.c struct page *page = pvec->pages[i]; page 447 mm/swap.c if (PagePrivate(page) && trylock_page(page)) { page 448 mm/swap.c if (PagePrivate(page)) page 449 mm/swap.c try_to_release_page(page, 0); page 450 mm/swap.c unlock_page(page); page 75 mm/swap_state.c BUG_ON(!PageLocked(page)); page 76 mm/swap_state.c BUG_ON(PageSwapCache(page)); page 77 mm/swap_state.c BUG_ON(PagePrivate(page)); page 80 mm/swap_state.c page_cache_get(page); page 81 mm/swap_state.c SetPageSwapCache(page); page 82 mm/swap_state.c set_page_private(page, entry.val); page 86 mm/swap_state.c entry.val, page); page 89 mm/swap_state.c __inc_zone_page_state(page, NR_FILE_PAGES); page 96 mm/swap_state.c set_page_private(page, 0UL); page 97 mm/swap_state.c ClearPageSwapCache(page); page 98 mm/swap_state.c page_cache_release(page); page 110 mm/swap_state.c BUG_ON(!PageLocked(page)); page 111 mm/swap_state.c BUG_ON(!PageSwapCache(page)); page 112 mm/swap_state.c BUG_ON(PageWriteback(page)); page 113 mm/swap_state.c BUG_ON(PagePrivate(page)); page 115 mm/swap_state.c radix_tree_delete(&swapper_space.page_tree, page_private(page)); page 116 mm/swap_state.c set_page_private(page, 0); page 117 mm/swap_state.c ClearPageSwapCache(page); page 119 mm/swap_state.c __dec_zone_page_state(page, NR_FILE_PAGES); page 136 mm/swap_state.c BUG_ON(!PageLocked(page)); page 137 mm/swap_state.c BUG_ON(!PageUptodate(page)); page 155 mm/swap_state.c err = add_to_swap_cache(page, entry, page 160 mm/swap_state.c SetPageDirty(page); page 184 mm/swap_state.c entry.val = page_private(page); page 187 mm/swap_state.c __delete_from_swap_cache(page); page 191 mm/swap_state.c page_cache_release(page); page 204 mm/swap_state.c if (PageSwapCache(page) && trylock_page(page)) { page 205 mm/swap_state.c remove_exclusive_swap_page(page); page 206 mm/swap_state.c unlock_page(page); page 216 mm/swap_state.c free_swap_cache(page); page 217 mm/swap_state.c page_cache_release(page); page 226 mm/swap_state.c struct page **pagep = pages; page 247 mm/swap_state.c struct page * lookup_swap_cache(swp_entry_t entry) page 249 mm/swap_state.c struct page *page; page 251 mm/swap_state.c page = find_get_page(&swapper_space, entry.val); page 253 mm/swap_state.c if (page) page 257 mm/swap_state.c return page; page 266 mm/swap_state.c struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, page 269 mm/swap_state.c struct page *found_page, *new_page = NULL; page 343 mm/swap_state.c struct page *swapin_readahead(swp_entry_t entry, gfp_t gfp_mask, page 347 mm/swap_state.c struct page *page; page 361 mm/swap_state.c page = read_swap_cache_async(swp_entry(swp_type(entry), offset), page 363 mm/swap_state.c if (!page) page 365 mm/swap_state.c page_cache_release(page); page 65 mm/swapfile.c entry.val = page_private(page); page 66 mm/swapfile.c if (PageSwapCache(page)) { page 78 mm/swapfile.c WARN_ON(page_count(page) <= 1); page 81 mm/swapfile.c blk_run_backing_dev(bdi, page); page 318 mm/swapfile.c entry.val = page_private(page); page 336 mm/swapfile.c BUG_ON(!PageLocked(page)); page 337 mm/swapfile.c count = page_mapcount(page); page 338 mm/swapfile.c if (count <= 1 && PageSwapCache(page)) page 339 mm/swapfile.c count += page_swapcount(page); page 353 mm/swapfile.c BUG_ON(PagePrivate(page)); page 354 mm/swapfile.c BUG_ON(!PageLocked(page)); page 356 mm/swapfile.c if (!PageSwapCache(page)) page 358 mm/swapfile.c if (PageWriteback(page)) page 360 mm/swapfile.c if (page_count(page) != 2) /* 2: us + cache */ page 363 mm/swapfile.c entry.val = page_private(page); page 373 mm/swapfile.c if ((page_count(page) == 2) && !PageWriteback(page)) { page 374 mm/swapfile.c __delete_from_swap_cache(page); page 375 mm/swapfile.c SetPageDirty(page); page 384 mm/swapfile.c page_cache_release(page); page 397 mm/swapfile.c struct page *page = NULL; page 405 mm/swapfile.c page = find_get_page(&swapper_space, entry.val); page 406 mm/swapfile.c if (page && unlikely(!trylock_page(page))) { page 407 mm/swapfile.c page_cache_release(page); page 408 mm/swapfile.c page = NULL; page 413 mm/swapfile.c if (page) { page 416 mm/swapfile.c BUG_ON(PagePrivate(page)); page 417 mm/swapfile.c one_user = (page_count(page) == 2); page 420 mm/swapfile.c if (PageSwapCache(page) && !PageWriteback(page) && page 422 mm/swapfile.c delete_from_swap_cache(page); page 423 mm/swapfile.c SetPageDirty(page); page 425 mm/swapfile.c unlock_page(page); page 426 mm/swapfile.c page_cache_release(page); page 518 mm/swapfile.c if (mem_cgroup_charge(page, vma->vm_mm, GFP_KERNEL)) page 524 mm/swapfile.c mem_cgroup_uncharge_page(page); page 530 mm/swapfile.c get_page(page); page 532 mm/swapfile.c pte_mkold(mk_pte(page, vma->vm_page_prot))); page 533 mm/swapfile.c page_add_anon_rmap(page, vma, addr); page 539 mm/swapfile.c activate_page(page); page 570 mm/swapfile.c ret = unuse_pte(vma, pmd, addr, entry, page); page 594 mm/swapfile.c ret = unuse_pte_range(vma, pmd, addr, next, entry, page); page 614 mm/swapfile.c ret = unuse_pmd_range(vma, pud, addr, next, entry, page); page 628 mm/swapfile.c if (page->mapping) { page 629 mm/swapfile.c addr = page_address_in_vma(page, vma); page 644 mm/swapfile.c ret = unuse_pud_range(vma, pgd, addr, next, entry, page); page 662 mm/swapfile.c activate_page(page); page 663 mm/swapfile.c unlock_page(page); page 665 mm/swapfile.c lock_page(page); page 668 mm/swapfile.c if (vma->anon_vma && (ret = unuse_vma(vma, entry, page))) page 724 mm/swapfile.c struct page *page; page 767 mm/swapfile.c page = read_swap_cache_async(entry, page 769 mm/swapfile.c if (!page) { page 799 mm/swapfile.c wait_on_page_locked(page); page 800 mm/swapfile.c wait_on_page_writeback(page); page 801 mm/swapfile.c lock_page(page); page 802 mm/swapfile.c wait_on_page_writeback(page); page 813 mm/swapfile.c shmem = shmem_unuse(entry, page); page 815 mm/swapfile.c retval = unuse_mm(start_mm, entry, page); page 843 mm/swapfile.c shmem = shmem_unuse(entry, page); page 845 mm/swapfile.c retval = unuse_mm(mm, entry, page); page 867 mm/swapfile.c unlock_page(page); page 868 mm/swapfile.c page_cache_release(page); page 905 mm/swapfile.c if ((*swap_map > 1) && PageDirty(page) && PageSwapCache(page)) { page 910 mm/swapfile.c swap_writepage(page, &wbc); page 911 mm/swapfile.c lock_page(page); page 912 mm/swapfile.c wait_on_page_writeback(page); page 914 mm/swapfile.c if (PageSwapCache(page)) page 915 mm/swapfile.c delete_from_swap_cache(page); page 922 mm/swapfile.c SetPageDirty(page); page 923 mm/swapfile.c unlock_page(page); page 924 mm/swapfile.c page_cache_release(page); page 1193 mm/swapfile.c BUG_ON(!PageLocked(page)); /* It pins the swap_info_struct */ page 1195 mm/swapfile.c if (PageSwapCache(page)) { page 1196 mm/swapfile.c swp_entry_t entry = { .val = page_private(page) }; page 1202 mm/swapfile.c bdi = page->mapping->backing_dev_info; page 1469 mm/swapfile.c struct page *page = NULL; page 1554 mm/swapfile.c page = read_mapping_page(mapping, 0, swap_file); page 1555 mm/swapfile.c if (IS_ERR(page)) { page 1556 mm/swapfile.c error = PTR_ERR(page); page 1559 mm/swapfile.c kmap(page); page 1560 mm/swapfile.c swap_header = page_address(page); page 1722 mm/swapfile.c if (page && !IS_ERR(page)) { page 1723 mm/swapfile.c kunmap(page); page 1724 mm/swapfile.c page_cache_release(page); page 39 mm/truncate.c void (*invalidatepage)(struct page *, unsigned long); page 40 mm/truncate.c invalidatepage = page->mapping->a_ops->invalidatepage; page 46 mm/truncate.c (*invalidatepage)(page, offset); page 51 mm/truncate.c zero_user_segment(page, partial, PAGE_CACHE_SIZE); page 52 mm/truncate.c if (PagePrivate(page)) page 53 mm/truncate.c do_invalidatepage(page, partial); page 72 mm/truncate.c if (TestClearPageDirty(page)) { page 73 mm/truncate.c struct address_space *mapping = page->mapping; page 75 mm/truncate.c dec_zone_page_state(page, NR_FILE_DIRTY); page 98 mm/truncate.c if (page->mapping != mapping) page 101 mm/truncate.c if (PagePrivate(page)) page 102 mm/truncate.c do_invalidatepage(page, 0); page 104 mm/truncate.c cancel_dirty_page(page, PAGE_CACHE_SIZE); page 106 mm/truncate.c remove_from_page_cache(page); page 107 mm/truncate.c ClearPageMappedToDisk(page); page 108 mm/truncate.c page_cache_release(page); /* pagecache ref */ page 124 mm/truncate.c if (page->mapping != mapping) page 127 mm/truncate.c if (PagePrivate(page) && !try_to_release_page(page, 0)) page 130 mm/truncate.c ret = remove_mapping(mapping, page); page 179 mm/truncate.c struct page *page = pvec.pages[i]; page 180 mm/truncate.c pgoff_t page_index = page->index; page 190 mm/truncate.c if (!trylock_page(page)) page 192 mm/truncate.c if (PageWriteback(page)) { page 193 mm/truncate.c unlock_page(page); page 196 mm/truncate.c if (page_mapped(page)) { page 201 mm/truncate.c truncate_complete_page(mapping, page); page 202 mm/truncate.c unlock_page(page); page 209 mm/truncate.c struct page *page = find_lock_page(mapping, start - 1); page 210 mm/truncate.c if (page) { page 211 mm/truncate.c wait_on_page_writeback(page); page 212 mm/truncate.c truncate_partial_page(page, partial); page 213 mm/truncate.c unlock_page(page); page 214 mm/truncate.c page_cache_release(page); page 232 mm/truncate.c struct page *page = pvec.pages[i]; page 234 mm/truncate.c if (page->index > end) page 236 mm/truncate.c lock_page(page); page 237 mm/truncate.c wait_on_page_writeback(page); page 238 mm/truncate.c if (page_mapped(page)) { page 240 mm/truncate.c (loff_t)page->index<<PAGE_CACHE_SHIFT, page 243 mm/truncate.c if (page->index > next) page 244 mm/truncate.c next = page->index; page 246 mm/truncate.c truncate_complete_page(mapping, page); page 247 mm/truncate.c unlock_page(page); page 279 mm/truncate.c struct page *page = pvec.pages[i]; page 283 mm/truncate.c lock_failed = !trylock_page(page); page 291 mm/truncate.c index = page->index; page 298 mm/truncate.c if (PageDirty(page) || PageWriteback(page)) page 300 mm/truncate.c if (page_mapped(page)) page 302 mm/truncate.c ret += invalidate_complete_page(mapping, page); page 304 mm/truncate.c unlock_page(page); page 345 mm/truncate.c if (page->mapping != mapping) page 348 mm/truncate.c if (PagePrivate(page) && !try_to_release_page(page, GFP_KERNEL)) page 352 mm/truncate.c if (PageDirty(page)) page 355 mm/truncate.c BUG_ON(PagePrivate(page)); page 356 mm/truncate.c __remove_from_page_cache(page); page 358 mm/truncate.c page_cache_release(page); /* pagecache ref */ page 367 mm/truncate.c if (!PageDirty(page)) page 369 mm/truncate.c if (page->mapping != mapping || mapping->a_ops->launder_page == NULL) page 371 mm/truncate.c return mapping->a_ops->launder_page(page); page 402 mm/truncate.c struct page *page = pvec.pages[i]; page 405 mm/truncate.c lock_page(page); page 406 mm/truncate.c if (page->mapping != mapping) { page 407 mm/truncate.c unlock_page(page); page 410 mm/truncate.c page_index = page->index; page 415 mm/truncate.c unlock_page(page); page 418 mm/truncate.c wait_on_page_writeback(page); page 419 mm/truncate.c if (page_mapped(page)) { page 439 mm/truncate.c BUG_ON(page_mapped(page)); page 440 mm/truncate.c ret2 = do_launder_page(mapping, page); page 442 mm/truncate.c if (!invalidate_complete_page2(mapping, page)) page 447 mm/truncate.c unlock_page(page); page 106 mm/vmalloc.c struct page *page = **pages; page 108 mm/vmalloc.c if (!page) page 110 mm/vmalloc.c set_pte_at(&init_mm, addr, pte, mk_pte(page, prot)); page 174 mm/vmalloc.c struct page *vmalloc_to_page(const void *vmalloc_addr) page 177 mm/vmalloc.c struct page *page = NULL; page 198 mm/vmalloc.c page = pte_page(pte); page 203 mm/vmalloc.c return page; page 409 mm/vmalloc.c struct page *page = area->pages[i]; page 411 mm/vmalloc.c BUG_ON(!page); page 412 mm/vmalloc.c __free_page(page); page 493 mm/vmalloc.c struct page **pages; page 497 mm/vmalloc.c array_size = (nr_pages * sizeof(struct page *)); page 519 mm/vmalloc.c struct page *page; page 522 mm/vmalloc.c page = alloc_page(gfp_mask); page 524 mm/vmalloc.c page = alloc_pages_node(node, gfp_mask, 0); page 526 mm/vmalloc.c if (unlikely(!page)) { page 531 mm/vmalloc.c area->pages[i] = page; page 822 mm/vmalloc.c struct page *page = vmalloc_to_page(addr); page 823 mm/vmalloc.c ret = vm_insert_page(vma, uaddr, page); page 84 mm/vmscan.c #define lru_to_page(_head) (list_entry((_head)->prev, struct page, lru)) page 90 mm/vmscan.c struct page *prev; \ page 104 mm/vmscan.c struct page *prev; \ page 239 mm/vmscan.c if (page_mapped(page)) page 243 mm/vmscan.c if (PageSwapCache(page)) page 246 mm/vmscan.c mapping = page_mapping(page); page 256 mm/vmscan.c return page_count(page) - !!PagePrivate(page) == 2; page 285 mm/vmscan.c lock_page(page); page 286 mm/vmscan.c if (page_mapping(page) == mapping) page 288 mm/vmscan.c unlock_page(page); page 333 mm/vmscan.c if (!is_page_cache_freeable(page)) page 340 mm/vmscan.c if (PagePrivate(page)) { page 341 mm/vmscan.c if (try_to_free_buffers(page)) { page 342 mm/vmscan.c ClearPageDirty(page); page 354 mm/vmscan.c if (clear_page_dirty_for_io(page)) { page 365 mm/vmscan.c SetPageReclaim(page); page 366 mm/vmscan.c res = mapping->a_ops->writepage(page, &wbc); page 368 mm/vmscan.c handle_write_error(mapping, page, res); page 370 mm/vmscan.c ClearPageReclaim(page); page 379 mm/vmscan.c if (PageWriteback(page) && sync_writeback == PAGEOUT_IO_SYNC) page 380 mm/vmscan.c wait_on_page_writeback(page); page 382 mm/vmscan.c if (!PageWriteback(page)) { page 384 mm/vmscan.c ClearPageReclaim(page); page 386 mm/vmscan.c inc_zone_page_state(page, NR_VMSCAN_WRITE); page 399 mm/vmscan.c BUG_ON(!PageLocked(page)); page 400 mm/vmscan.c BUG_ON(mapping != page_mapping(page)); page 428 mm/vmscan.c if (!page_freeze_refs(page, 2)) page 431 mm/vmscan.c if (unlikely(PageDirty(page))) { page 432 mm/vmscan.c page_unfreeze_refs(page, 2); page 436 mm/vmscan.c if (PageSwapCache(page)) { page 437 mm/vmscan.c swp_entry_t swap = { .val = page_private(page) }; page 438 mm/vmscan.c __delete_from_swap_cache(page); page 442 mm/vmscan.c __remove_from_page_cache(page); page 461 mm/vmscan.c if (__remove_mapping(mapping, page)) { page 467 mm/vmscan.c page_unfreeze_refs(page, 1); page 490 mm/vmscan.c struct page *page; page 496 mm/vmscan.c page = lru_to_page(page_list); page 497 mm/vmscan.c list_del(&page->lru); page 499 mm/vmscan.c if (!trylock_page(page)) page 502 mm/vmscan.c VM_BUG_ON(PageActive(page)); page 506 mm/vmscan.c if (!sc->may_swap && page_mapped(page)) page 510 mm/vmscan.c if (page_mapped(page) || PageSwapCache(page)) page 514 mm/vmscan.c (PageSwapCache(page) && (sc->gfp_mask & __GFP_IO)); page 516 mm/vmscan.c if (PageWriteback(page)) { page 526 mm/vmscan.c wait_on_page_writeback(page); page 531 mm/vmscan.c referenced = page_referenced(page, 1, sc->mem_cgroup); page 534 mm/vmscan.c referenced && page_mapping_inuse(page)) page 542 mm/vmscan.c if (PageAnon(page) && !PageSwapCache(page)) page 543 mm/vmscan.c if (!add_to_swap(page, GFP_ATOMIC)) page 547 mm/vmscan.c mapping = page_mapping(page); page 553 mm/vmscan.c if (page_mapped(page) && mapping) { page 554 mm/vmscan.c switch (try_to_unmap(page, 0)) { page 564 mm/vmscan.c if (PageDirty(page)) { page 573 mm/vmscan.c switch (pageout(page, mapping, sync_writeback)) { page 579 mm/vmscan.c if (PageWriteback(page) || PageDirty(page)) page 585 mm/vmscan.c if (!trylock_page(page)) page 587 mm/vmscan.c if (PageDirty(page) || PageWriteback(page)) page 589 mm/vmscan.c mapping = page_mapping(page); page 616 mm/vmscan.c if (PagePrivate(page)) { page 617 mm/vmscan.c if (!try_to_release_page(page, sc->gfp_mask)) page 619 mm/vmscan.c if (!mapping && page_count(page) == 1) { page 620 mm/vmscan.c unlock_page(page); page 621 mm/vmscan.c if (put_page_testzero(page)) page 637 mm/vmscan.c if (!mapping || !__remove_mapping(mapping, page)) page 640 mm/vmscan.c unlock_page(page); page 643 mm/vmscan.c if (!pagevec_add(&freed_pvec, page)) { page 650 mm/vmscan.c SetPageActive(page); page 653 mm/vmscan.c unlock_page(page); page 655 mm/vmscan.c list_add(&page->lru, &ret_pages); page 656 mm/vmscan.c VM_BUG_ON(PageLRU(page)); page 685 mm/vmscan.c if (!PageLRU(page)) page 693 mm/vmscan.c if (mode != ISOLATE_BOTH && (!PageActive(page) != !mode)) page 697 mm/vmscan.c if (likely(get_page_unless_zero(page))) { page 703 mm/vmscan.c ClearPageLRU(page); page 737 mm/vmscan.c struct page *page; page 743 mm/vmscan.c page = lru_to_page(src); page 744 mm/vmscan.c prefetchw_prev_lru_page(page, src, flags); page 746 mm/vmscan.c VM_BUG_ON(!PageLRU(page)); page 748 mm/vmscan.c switch (__isolate_lru_page(page, mode)) { page 750 mm/vmscan.c list_move(&page->lru, dst); page 756 mm/vmscan.c list_move(&page->lru, src); page 775 mm/vmscan.c zone_id = page_zone_id(page); page 776 mm/vmscan.c page_pfn = page_to_pfn(page); page 780 mm/vmscan.c struct page *cursor_page; page 836 mm/vmscan.c struct page *page; page 838 mm/vmscan.c list_for_each_entry(page, page_list, lru) page 839 mm/vmscan.c if (PageActive(page)) { page 840 mm/vmscan.c ClearPageActive(page); page 864 mm/vmscan.c struct page *page; page 927 mm/vmscan.c page = lru_to_page(&page_list); page 928 mm/vmscan.c VM_BUG_ON(PageLRU(page)); page 929 mm/vmscan.c SetPageLRU(page); page 930 mm/vmscan.c list_del(&page->lru); page 931 mm/vmscan.c if (PageActive(page)) page 932 mm/vmscan.c add_page_to_active_list(zone, page); page 934 mm/vmscan.c add_page_to_inactive_list(zone, page); page 935 mm/vmscan.c if (!pagevec_add(&pvec, page)) { page 1104 mm/vmscan.c struct page *page; page 1128 mm/vmscan.c page = lru_to_page(&l_hold); page 1129 mm/vmscan.c list_del(&page->lru); page 1130 mm/vmscan.c if (page_mapped(page)) { page 1132 mm/vmscan.c (total_swap_pages == 0 && PageAnon(page)) || page 1133 mm/vmscan.c page_referenced(page, 0, sc->mem_cgroup)) { page 1134 mm/vmscan.c list_add(&page->lru, &l_active); page 1138 mm/vmscan.c list_add(&page->lru, &l_inactive); page 1145 mm/vmscan.c page = lru_to_page(&l_inactive); page 1146 mm/vmscan.c prefetchw_prev_lru_page(page, &l_inactive, flags); page 1147 mm/vmscan.c VM_BUG_ON(PageLRU(page)); page 1148 mm/vmscan.c SetPageLRU(page); page 1149 mm/vmscan.c VM_BUG_ON(!PageActive(page)); page 1150 mm/vmscan.c ClearPageActive(page); page 1152 mm/vmscan.c list_move(&page->lru, &zone->inactive_list); page 1153 mm/vmscan.c mem_cgroup_move_lists(page, false); page 1155 mm/vmscan.c if (!pagevec_add(&pvec, page)) { page 1176 mm/vmscan.c page = lru_to_page(&l_active); page 1177 mm/vmscan.c prefetchw_prev_lru_page(page, &l_active, flags); page 1178 mm/vmscan.c VM_BUG_ON(PageLRU(page)); page 1179 mm/vmscan.c SetPageLRU(page); page 1180 mm/vmscan.c VM_BUG_ON(!PageActive(page)); page 1182 mm/vmscan.c list_move(&page->lru, &zone->active_list); page 1183 mm/vmscan.c mem_cgroup_move_lists(page, true); page 1185 mm/vmscan.c if (!pagevec_add(&pvec, page)) { page 224 mm/vmstat.c __inc_zone_state(page_zone(page), item); page 245 mm/vmstat.c __dec_zone_state(page_zone(page), item); page 263 mm/vmstat.c zone = page_zone(page); page 275 mm/vmstat.c __dec_zone_page_state(page, item); page 513 mm/vmstat.c struct page *page; page 518 mm/vmstat.c page = pfn_to_page(pfn); page 532 mm/vmstat.c if (page_zone(page) != zone) page 535 mm/vmstat.c mtype = get_pageblock_migratetype(page); page 195 net/atm/mpoa_proc.c char *page, *p; page 204 net/atm/mpoa_proc.c page = (char *)__get_free_page(GFP_KERNEL); page 205 net/atm/mpoa_proc.c if (!page) page 208 net/atm/mpoa_proc.c for (p = page, len = 0; len < nbytes; p++, len++) { page 210 net/atm/mpoa_proc.c free_page((unsigned long)page); page 219 net/atm/mpoa_proc.c if (!parse_qos(page)) page 220 net/atm/mpoa_proc.c printk("mpoa: proc_mpc_write: could not parse '%s'\n", page); page 222 net/atm/mpoa_proc.c free_page((unsigned long)page); page 375 net/atm/proc.c unsigned long page; page 379 net/atm/proc.c page = get_zeroed_page(GFP_KERNEL); page 380 net/atm/proc.c if (!page) return -ENOMEM; page 385 net/atm/proc.c length = dev->ops->proc_read(dev,pos,(char *) page); page 389 net/atm/proc.c if (copy_to_user(buf,(char *) page,length)) length = -EFAULT; page 392 net/atm/proc.c free_page(page); page 531 net/bluetooth/cmtp/capi.c len += sprintf(page + len, "%s\n\n", cmtp_procinfo(ctrl)); page 532 net/bluetooth/cmtp/capi.c len += sprintf(page + len, "addr %s\n", session->name); page 533 net/bluetooth/cmtp/capi.c len += sprintf(page + len, "ctrl %d\n", session->num); page 537 net/bluetooth/cmtp/capi.c len += sprintf(page + len, "appl %d -> %d\n", app->appl, app->mapping); page 546 net/bluetooth/cmtp/capi.c *start = page + off; page 1158 net/bluetooth/hci_event.c cp.page = 0x01; page 1597 net/bluetooth/hci_event.c if (!ev->status && ev->page == 0x01) { page 154 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, ">>> socket %p", page 156 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, " / sk %p", sk); page 157 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, " / bo %p", bo); page 158 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, " / dropped %lu", page 160 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, " / bound %s", page 162 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, " <<<\n"); page 172 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, page 175 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, "[%d]%c ", page 179 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, page 185 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, page 190 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, page 196 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, "%s%ld%%\n", page 201 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, "(..)\n"); page 208 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, page 214 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, "t1=%lld ", page 218 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, "t2=%lld ", page 221 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, "# sent %ld\n", page 226 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, "(..)\n"); page 231 net/can/bcm.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 211 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, fmt, page 221 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 237 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 249 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 250 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 253 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 255 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 258 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 261 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 265 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 268 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 272 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 274 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 278 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 281 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 285 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 287 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 291 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 294 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 298 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 301 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 304 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 309 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 314 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 318 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 332 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 340 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 354 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "%s\n", page 369 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 376 net/can/proc.c len = can_print_recv_banner(page, len); page 377 net/can/proc.c len = can_print_rcvlist(page, len, &d->rx[idx], d->dev); page 379 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 388 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 402 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 416 net/can/proc.c len = can_print_recv_banner(page, len); page 420 net/can/proc.c len = can_print_rcvlist(page, len, page 425 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, page 434 net/can/proc.c len += snprintf(page + len, PAGE_SIZE - len, "\n"); page 295 net/core/datagram.c struct page *page = frag->page; page 299 net/core/datagram.c vaddr = kmap(page); page 302 net/core/datagram.c kunmap(page); page 380 net/core/datagram.c struct page *page = frag->page; page 384 net/core/datagram.c vaddr = kmap(page); page 387 net/core/datagram.c kunmap(page); page 464 net/core/datagram.c struct page *page = frag->page; page 468 net/core/datagram.c vaddr = kmap(page); page 473 net/core/datagram.c kunmap(page); page 1594 net/core/dev.c if (PageHighMem(skb_shinfo(skb)->frags[i].page)) page 10 net/core/kmap_skb.h return kmap_atomic(frag->page, KM_SKB_DATA_SOFTIRQ); page 2636 net/core/pktgen.c struct page *page = alloc_pages(GFP_KERNEL, 0); page 2637 net/core/pktgen.c skb_shinfo(skb)->frags[i].page = page; page 2662 net/core/pktgen.c get_page(skb_shinfo(skb)->frags[i].page); page 2663 net/core/pktgen.c skb_shinfo(skb)->frags[i].page = page 2664 net/core/pktgen.c skb_shinfo(skb)->frags[i - 1].page; page 2984 net/core/pktgen.c struct page *page = alloc_pages(GFP_KERNEL, 0); page 2985 net/core/pktgen.c skb_shinfo(skb)->frags[i].page = page; page 3010 net/core/pktgen.c get_page(skb_shinfo(skb)->frags[i].page); page 3011 net/core/pktgen.c skb_shinfo(skb)->frags[i].page = page 3012 net/core/pktgen.c skb_shinfo(skb)->frags[i - 1].page; page 27 net/core/skb_dma_map.c map = dma_map_page(dev, fp->page, fp->page_offset, page 266 net/core/skbuff.c struct page *__netdev_alloc_page(struct net_device *dev, gfp_t gfp_mask) page 269 net/core/skbuff.c struct page *page; page 271 net/core/skbuff.c page = alloc_pages_node(node, gfp_mask, 0); page 272 net/core/skbuff.c return page; page 279 net/core/skbuff.c skb_fill_page_desc(skb, i, page, off, size); page 342 net/core/skbuff.c put_page(skb_shinfo(skb)->frags[i].page); page 718 net/core/skbuff.c get_page(skb_shinfo(n)->frags[i].page); page 783 net/core/skbuff.c get_page(skb_shinfo(skb)->frags[i].page); page 1052 net/core/skbuff.c put_page(skb_shinfo(skb)->frags[i].page); page 1221 net/core/skbuff.c put_page(skb_shinfo(skb)->frags[i].page); page 1339 net/core/skbuff.c spd->pages[spd->nr_pages] = page; page 1351 net/core/skbuff.c *page += *poff / PAGE_SIZE; page 1372 net/core/skbuff.c __segment_seek(&page, &poff, &plen, *off); page 1382 net/core/skbuff.c if (spd_fill_page(spd, page, flen, poff, skb)) page 1385 net/core/skbuff.c __segment_seek(&page, &poff, &plen, flen); page 1418 net/core/skbuff.c if (__splice_segment(f->page, f->page_offset, f->size, page 1437 net/core/skbuff.c struct page *pages[PIPE_BUFFERS]; page 1978 net/core/skbuff.c get_page(skb_shinfo(skb)->frags[i].page); page 2187 net/core/skbuff.c struct page *page = NULL; page 2199 net/core/skbuff.c page = alloc_pages(sk->sk_allocation, 0); page 2204 net/core/skbuff.c if (page == NULL) page 2208 net/core/skbuff.c sk->sk_sndmsg_page = page; page 2210 net/core/skbuff.c skb_fill_page_desc(skb, frg_cnt, page, 0, 0); page 2222 net/core/skbuff.c ret = getfrag(from, (page_address(frag->page) + page 2346 net/core/skbuff.c get_page(frag->page); page 2438 net/core/skbuff.c sg_set_page(&sg[elt], frag->page, copy, page 1297 net/core/sock.c struct page *page; page 1300 net/core/sock.c page = alloc_pages(sk->sk_allocation, 0); page 1301 net/core/sock.c if (!page) { page 1309 net/core/sock.c frag->page = page; page 1605 net/core/sock.c char *kaddr = kmap(page); page 1609 net/core/sock.c kunmap(page); page 80 net/core/user_dma.c struct page *page = frag->page; page 85 net/core/user_dma.c cookie = dma_memcpy_pg_to_iovec(chan, to, pinned_list, page, page 238 net/ieee80211/ieee80211_module.c return snprintf(page, count, "0x%08X\n", ieee80211_debug_level); page 726 net/ipv4/af_inet.c return sk->sk_prot->sendpage(sk, page, offset, size, flags); page 727 net/ipv4/af_inet.c return sock_no_sendpage(sock, page, offset, size, flags); page 452 net/ipv4/inet_lro.c mac_hdr = page_address(frags->page) + frags->page_offset; page 713 net/ipv4/ip_output.c kaddr = kmap(page); page 715 net/ipv4/ip_output.c kunmap(page); page 999 net/ipv4/ip_output.c struct page *page = sk->sk_sndmsg_page; page 1003 net/ipv4/ip_output.c if (page && (left = PAGE_SIZE - off) > 0) { page 1006 net/ipv4/ip_output.c if (page != frag->page) { page 1011 net/ipv4/ip_output.c get_page(page); page 1012 net/ipv4/ip_output.c skb_fill_page_desc(skb, i, page, sk->sk_sndmsg_off, 0); page 1018 net/ipv4/ip_output.c page = alloc_pages(sk->sk_allocation, 0); page 1019 net/ipv4/ip_output.c if (page == NULL) { page 1023 net/ipv4/ip_output.c sk->sk_sndmsg_page = page; page 1026 net/ipv4/ip_output.c skb_fill_page_desc(skb, i, page, 0, 0); page 1032 net/ipv4/ip_output.c if (getfrag(from, page_address(frag->page)+frag->page_offset+frag->size, offset, copy, skb->len, skb) < 0) { page 1166 net/ipv4/ip_output.c if (skb_can_coalesce(skb, i, page, offset)) { page 1169 net/ipv4/ip_output.c get_page(page); page 1170 net/ipv4/ip_output.c skb_fill_page_desc(skb, i, page, offset, len); page 1178 net/ipv4/ip_output.c csum = csum_page(page, offset, len); page 684 net/ipv4/tcp.c struct page *page = pages[poffset / PAGE_SIZE]; page 706 net/ipv4/tcp.c can_coalesce = skb_can_coalesce(skb, i, page, offset); page 717 net/ipv4/tcp.c get_page(page); page 718 net/ipv4/tcp.c skb_fill_page_desc(skb, i, page, offset, copy); page 782 net/ipv4/tcp.c return sock_no_sendpage(sock, page, offset, size, flags); page 786 net/ipv4/tcp.c res = do_tcp_sendpages(sk, &page, offset, size, flags); page 903 net/ipv4/tcp.c struct page *page = TCP_PAGE(sk); page 906 net/ipv4/tcp.c if (skb_can_coalesce(skb, i, page, off) && page 920 net/ipv4/tcp.c } else if (page) { page 922 net/ipv4/tcp.c put_page(page); page 923 net/ipv4/tcp.c TCP_PAGE(sk) = page = NULL; page 935 net/ipv4/tcp.c if (!page) { page 937 net/ipv4/tcp.c if (!(page = sk_stream_alloc_page(sk))) page 943 net/ipv4/tcp.c err = skb_copy_to_page(sk, from, skb, page, page 950 net/ipv4/tcp.c TCP_PAGE(sk) = page; page 961 net/ipv4/tcp.c skb_fill_page_desc(skb, i, page, off, copy); page 963 net/ipv4/tcp.c get_page(page); page 965 net/ipv4/tcp.c get_page(page); page 966 net/ipv4/tcp.c TCP_PAGE(sk) = page; page 2630 net/ipv4/tcp.c sg_set_page(&sg, f->page, f->size, f->page_offset); page 863 net/ipv4/tcp_output.c put_page(skb_shinfo(skb)->frags[i].page); page 751 net/ipv4/udp.c ret = ip_append_page(sk, page, offset, size, flags); page 754 net/ipv4/udp.c return sock_no_sendpage(sk->sk_socket, page, offset, page 1353 net/ipv6/ip6_output.c struct page *page = sk->sk_sndmsg_page; page 1357 net/ipv6/ip6_output.c if (page && (left = PAGE_SIZE - off) > 0) { page 1360 net/ipv6/ip6_output.c if (page != frag->page) { page 1365 net/ipv6/ip6_output.c get_page(page); page 1366 net/ipv6/ip6_output.c skb_fill_page_desc(skb, i, page, sk->sk_sndmsg_off, 0); page 1372 net/ipv6/ip6_output.c page = alloc_pages(sk->sk_allocation, 0); page 1373 net/ipv6/ip6_output.c if (page == NULL) { page 1377 net/ipv6/ip6_output.c sk->sk_sndmsg_page = page; page 1380 net/ipv6/ip6_output.c skb_fill_page_desc(skb, i, page, 0, 0); page 1386 net/ipv6/ip6_output.c if (getfrag(from, page_address(frag->page)+frag->page_offset+frag->size, offset, copy, skb->len, skb) < 0) { page 730 net/packet/af_packet.c struct page *p_start, *p_end; page 1927 net/packet/af_packet.c struct page *page = virt_to_page(po->pg_vec[i]); page 1930 net/packet/af_packet.c for (pg_num = 0; pg_num < po->pg_vec_pages; pg_num++, page++) { page 1931 net/packet/af_packet.c err = vm_insert_page(vma, start, page); page 698 net/socket.c return sock->ops->sendpage(sock, page, offset, size, flags); page 2430 net/socket.c return sock->ops->sendpage(sock, page, offset, size, flags); page 2432 net/socket.c return sock_no_sendpage(sock, page, offset, size, flags); page 1076 net/sunrpc/auth_gss/auth_gss.c = kmalloc(rqstp->rq_enc_pages_num * sizeof(struct page *), page 1104 net/sunrpc/auth_gss/auth_gss.c struct page **inpages; page 165 net/sunrpc/auth_gss/gss_krb5_crypto.c struct page **pages; page 177 net/sunrpc/auth_gss/gss_krb5_crypto.c struct page *in_page; page 142 net/sunrpc/auth_gss/gss_krb5_wrap.c struct page **tmp_pages; page 1274 net/sunrpc/auth_gss/svcauth_gss.c struct page **inpages = NULL; page 72 net/sunrpc/socklib.c struct page **ppage = xdr->pages; page 512 net/sunrpc/svc.c struct page *p = alloc_page(GFP_KERNEL); page 592 net/sunrpc/svc_xprt.c struct page *p = alloc_page(GFP_KERNEL); page 176 net/sunrpc/svcsock.c struct page **ppage = xdr->pages; page 183 net/sunrpc/xdr.c struct page **pgfrom, **pgto; page 240 net/sunrpc/xdr.c struct page **pgto; page 284 net/sunrpc/xdr.c struct page **pgfrom; page 833 net/sunrpc/xdr.c struct page **ppages = NULL; page 61 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c struct page *page; page 66 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c page = ctxt->pages[0]; page 68 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c rqstp->rq_pages[0] = page; page 71 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c rqstp->rq_arg.head[0].iov_base = page_address(page); page 85 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c page = ctxt->pages[sge_no]; page 87 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c rqstp->rq_pages[sge_no] = page; page 105 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c page = ctxt->pages[sge_no++]; page 106 net/sunrpc/xprtrdma/svc_rdma_recvfrom.c put_page(page); page 124 net/sunrpc/xprtrdma/svc_rdma_sendto.c struct page *page; page 126 net/sunrpc/xprtrdma/svc_rdma_sendto.c page = xdr->pages[page_no++]; page 131 net/sunrpc/xprtrdma/svc_rdma_sendto.c ib_dma_map_page(xprt->sc_cm_id->device, page, 0, page 521 net/sunrpc/xprtrdma/svc_rdma_sendto.c ctxt->pages[0] = page; page 532 net/sunrpc/xprtrdma/svc_rdma_sendto.c page, 0, PAGE_SIZE, DMA_TO_DEVICE); page 639 net/sunrpc/xprtrdma/svc_rdma_sendto.c struct page *res_page; page 473 net/sunrpc/xprtrdma/svc_rdma_transport.c struct page *svc_rdma_get_page(void) page 475 net/sunrpc/xprtrdma/svc_rdma_transport.c struct page *page; page 477 net/sunrpc/xprtrdma/svc_rdma_transport.c while ((page = alloc_page(GFP_KERNEL)) == NULL) { page 483 net/sunrpc/xprtrdma/svc_rdma_transport.c return page; page 490 net/sunrpc/xprtrdma/svc_rdma_transport.c struct page *page; page 501 net/sunrpc/xprtrdma/svc_rdma_transport.c page = svc_rdma_get_page(); page 502 net/sunrpc/xprtrdma/svc_rdma_transport.c ctxt->pages[sge_no] = page; page 504 net/sunrpc/xprtrdma/svc_rdma_transport.c page, 0, PAGE_SIZE, page 1304 net/sunrpc/xprtrdma/svc_rdma_transport.c struct page *p; page 170 net/sunrpc/xprtrdma/xprt_rdma.h struct page *mr_page; /* owning page, if any */ page 428 net/sunrpc/xprtsock.c struct page **ppage; page 731 net/xfrm/xfrm_algo.c sg_set_page(&sg, frag->page, copy, page 80 net/xfrm/xfrm_ipcomp.c frag->page = alloc_page(GFP_ATOMIC); page 83 net/xfrm/xfrm_ipcomp.c if (!frag->page) page 90 net/xfrm/xfrm_ipcomp.c memcpy(page_address(frag->page), scratch, len); page 33 scripts/kconfig/lxdialog/textbox.c static const char *page; page 64 scripts/kconfig/lxdialog/textbox.c page = buf; /* page is pointer to start of page to be displayed */ page 139 scripts/kconfig/lxdialog/textbox.c page = buf; page 149 scripts/kconfig/lxdialog/textbox.c page = buf + strlen(buf); page 278 scripts/kconfig/lxdialog/textbox.c if (*page == '\0') { page 284 scripts/kconfig/lxdialog/textbox.c if (page == buf) { page 288 scripts/kconfig/lxdialog/textbox.c page--; page 290 scripts/kconfig/lxdialog/textbox.c if (page == buf) { page 294 scripts/kconfig/lxdialog/textbox.c page--; page 295 scripts/kconfig/lxdialog/textbox.c } while (*page != '\n'); page 296 scripts/kconfig/lxdialog/textbox.c page++; page 356 scripts/kconfig/lxdialog/textbox.c while (*page != '\n') { page 357 scripts/kconfig/lxdialog/textbox.c if (*page == '\0') { page 363 scripts/kconfig/lxdialog/textbox.c line[i++] = *(page++); page 368 scripts/kconfig/lxdialog/textbox.c page++; page 374 scripts/kconfig/lxdialog/textbox.c page++; /* move pass '\n' */ page 388 scripts/kconfig/lxdialog/textbox.c percent = (page - buf) * 100 / strlen(buf); page 269 security/selinux/avc.c return scnprintf(page, PAGE_SIZE, "entries: %d\nbuckets used: %d/%d\n" page 153 security/selinux/selinuxfs.c char *page; page 163 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 164 security/selinux/selinuxfs.c if (!page) page 167 security/selinux/selinuxfs.c if (copy_from_user(page, buf, count)) page 171 security/selinux/selinuxfs.c if (sscanf(page, "%d", &new_value) != 1) page 190 security/selinux/selinuxfs.c free_page((unsigned long) page); page 224 security/selinux/selinuxfs.c char *page; page 235 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 236 security/selinux/selinuxfs.c if (!page) page 239 security/selinux/selinuxfs.c if (copy_from_user(page, buf, count)) page 243 security/selinux/selinuxfs.c if (sscanf(page, "%d", &new_value) != 1) page 258 security/selinux/selinuxfs.c free_page((unsigned long) page); page 418 security/selinux/selinuxfs.c char *page; page 432 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 433 security/selinux/selinuxfs.c if (!page) page 436 security/selinux/selinuxfs.c if (copy_from_user(page, buf, count)) page 440 security/selinux/selinuxfs.c if (sscanf(page, "%u", &new_value) != 1) page 446 security/selinux/selinuxfs.c free_page((unsigned long) page); page 467 security/selinux/selinuxfs.c char *page; page 481 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 482 security/selinux/selinuxfs.c if (!page) page 485 security/selinux/selinuxfs.c if (copy_from_user(page, buf, count)) page 489 security/selinux/selinuxfs.c if (sscanf(page, "%d", &new_value) != 1) page 495 security/selinux/selinuxfs.c free_page((unsigned long) page); page 855 security/selinux/selinuxfs.c char *page = NULL; page 874 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 875 security/selinux/selinuxfs.c if (!page) { page 885 security/selinux/selinuxfs.c length = scnprintf(page, PAGE_SIZE, "%d %d", cur_enforcing, page 887 security/selinux/selinuxfs.c ret = simple_read_from_buffer(buf, count, ppos, page, length); page 890 security/selinux/selinuxfs.c if (page) page 891 security/selinux/selinuxfs.c free_page((unsigned long)page); page 898 security/selinux/selinuxfs.c char *page = NULL; page 926 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 927 security/selinux/selinuxfs.c if (!page) { page 933 security/selinux/selinuxfs.c if (copy_from_user(page, buf, count)) page 937 security/selinux/selinuxfs.c if (sscanf(page, "%d", &new_value) != 1) page 948 security/selinux/selinuxfs.c if (page) page 949 security/selinux/selinuxfs.c free_page((unsigned long) page); page 962 security/selinux/selinuxfs.c char *page = NULL; page 980 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 981 security/selinux/selinuxfs.c if (!page) { page 987 security/selinux/selinuxfs.c if (copy_from_user(page, buf, count)) page 991 security/selinux/selinuxfs.c if (sscanf(page, "%d", &new_value) != 1) page 1001 security/selinux/selinuxfs.c if (page) page 1002 security/selinux/selinuxfs.c free_page((unsigned long) page); page 1044 security/selinux/selinuxfs.c char **names = NULL, *page; page 1057 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 1058 security/selinux/selinuxfs.c if (!page) page 1077 security/selinux/selinuxfs.c len = snprintf(page, PAGE_SIZE, "/%s/%s", BOOL_DIR_NAME, names[i]); page 1086 security/selinux/selinuxfs.c ret = security_genfs_sid("selinuxfs", page, SECCLASS_FILE, &sid); page 1099 security/selinux/selinuxfs.c free_page((unsigned long)page); page 1132 security/selinux/selinuxfs.c char *page; page 1147 security/selinux/selinuxfs.c page = (char *)get_zeroed_page(GFP_KERNEL); page 1148 security/selinux/selinuxfs.c if (!page) { page 1153 security/selinux/selinuxfs.c if (copy_from_user(page, buf, count)) { page 1158 security/selinux/selinuxfs.c if (sscanf(page, "%u", &new_value) != 1) { page 1171 security/selinux/selinuxfs.c free_page((unsigned long)page); page 1179 security/selinux/selinuxfs.c char *page; page 1182 security/selinux/selinuxfs.c page = (char *)__get_free_page(GFP_KERNEL); page 1183 security/selinux/selinuxfs.c if (!page) { page 1187 security/selinux/selinuxfs.c ret = avc_get_hash_stats(page); page 1189 security/selinux/selinuxfs.c ret = simple_read_from_buffer(buf, count, ppos, page, ret); page 1190 security/selinux/selinuxfs.c free_page((unsigned long)page); page 1383 security/selinux/selinuxfs.c char *page; page 1386 security/selinux/selinuxfs.c page = (char *)__get_free_page(GFP_KERNEL); page 1387 security/selinux/selinuxfs.c if (!page) { page 1392 security/selinux/selinuxfs.c len = snprintf(page, PAGE_SIZE, "%d", sel_ino_to_class(ino)); page 1393 security/selinux/selinuxfs.c rc = simple_read_from_buffer(buf, count, ppos, page, len); page 1394 security/selinux/selinuxfs.c free_page((unsigned long)page); page 1407 security/selinux/selinuxfs.c char *page; page 1410 security/selinux/selinuxfs.c page = (char *)__get_free_page(GFP_KERNEL); page 1411 security/selinux/selinuxfs.c if (!page) { page 1416 security/selinux/selinuxfs.c len = snprintf(page, PAGE_SIZE, "%d", sel_ino_to_perm(ino)); page 1417 security/selinux/selinuxfs.c rc = simple_read_from_buffer(buf, count, ppos, page, len); page 1418 security/selinux/selinuxfs.c free_page((unsigned long)page); page 315 sound/core/pcm_memory.c struct page *snd_pcm_sgbuf_ops_page(struct snd_pcm_substream *substream, unsigned long offset) page 3004 sound/core/pcm_native.c vmf->page = virt_to_page(runtime->status); page 3005 sound/core/pcm_native.c get_page(vmf->page); page 3043 sound/core/pcm_native.c vmf->page = virt_to_page(runtime->control); page 3044 sound/core/pcm_native.c get_page(vmf->page); page 3094 sound/core/pcm_native.c struct page * page; page 3105 sound/core/pcm_native.c if (substream->ops->page) { page 3106 sound/core/pcm_native.c page = substream->ops->page(substream, offset); page 3107 sound/core/pcm_native.c if (!page) page 3111 sound/core/pcm_native.c page = virt_to_page(vaddr); page 3113 sound/core/pcm_native.c get_page(page); page 3114 sound/core/pcm_native.c vmf->page = page; page 73 sound/core/sgbuf.c struct page **pgtable; page 63 sound/drivers/vx/vx_pcm.c static struct page *snd_pcm_get_vmalloc_page(struct snd_pcm_substream *subs, page 956 sound/drivers/vx/vx_pcm.c .page = snd_pcm_get_vmalloc_page, page 1176 sound/drivers/vx/vx_pcm.c .page = snd_pcm_get_vmalloc_page, page 85 sound/isa/wavefront/wavefront_fx.c if (page < 0 || page > 7) { page 100 sound/isa/wavefront/wavefront_fx.c outb (page, dev->fx_dsp_page); page 106 sound/isa/wavefront/wavefront_fx.c page, addr, data[0]); page 112 sound/isa/wavefront/wavefront_fx.c outb (page, dev->fx_dsp_page); page 126 sound/isa/wavefront/wavefront_fx.c page, addr, (unsigned long) data, cnt); page 237 sound/mips/hal2.h u32 page; /* DOC Page register */ page 697 sound/mips/sgio2audio.c static struct page *snd_sgio2audio_page(struct snd_pcm_substream *substream, page 713 sound/mips/sgio2audio.c .page = snd_sgio2audio_page, page 725 sound/mips/sgio2audio.c .page = snd_sgio2audio_page, page 737 sound/mips/sgio2audio.c .page = snd_sgio2audio_page, page 62 sound/oss/dmabuf.c struct page *page; page 118 sound/oss/dmabuf.c for (page = virt_to_page(start_addr); page <= virt_to_page(end_addr); page++) page 119 sound/oss/dmabuf.c SetPageReserved(page); page 126 sound/oss/dmabuf.c struct page *page; page 138 sound/oss/dmabuf.c for (page = virt_to_page(start_addr); page <= virt_to_page(end_addr); page++) page 139 sound/oss/dmabuf.c ClearPageReserved(page); page 889 sound/oss/msnd_pinnacle.c char *page = (char *)__get_free_page(GFP_KERNEL); page 891 sound/oss/msnd_pinnacle.c if (!page) page 904 sound/oss/msnd_pinnacle.c n = msnd_fifo_read(&dev.DARF, page, k); page 906 sound/oss/msnd_pinnacle.c if (copy_to_user(buf, page, n)) { page 907 sound/oss/msnd_pinnacle.c free_page((unsigned long)page); page 923 sound/oss/msnd_pinnacle.c free_page((unsigned long)page); page 935 sound/oss/msnd_pinnacle.c free_page((unsigned long)page); page 940 sound/oss/msnd_pinnacle.c free_page((unsigned long)page); page 947 sound/oss/msnd_pinnacle.c char *page = (char *)__get_free_page(GFP_KERNEL); page 949 sound/oss/msnd_pinnacle.c if (!page) page 960 sound/oss/msnd_pinnacle.c if (copy_from_user(page, buf, k)) { page 961 sound/oss/msnd_pinnacle.c free_page((unsigned long)page); page 967 sound/oss/msnd_pinnacle.c n = msnd_fifo_write(&dev.DAPF, page, k); page 982 sound/oss/msnd_pinnacle.c free_page((unsigned long)page); page 993 sound/oss/msnd_pinnacle.c free_page((unsigned long)page); page 999 sound/oss/msnd_pinnacle.c free_page((unsigned long)page); page 789 sound/oss/sscape.c struct page *page; page 826 sound/oss/sscape.c for (page = virt_to_page(start_addr); page <= virt_to_page(end_addr); page++) page 827 sound/oss/sscape.c SetPageReserved(page); page 835 sound/oss/sscape.c struct page *page; page 842 sound/oss/sscape.c for (page = virt_to_page(start_addr); page <= virt_to_page(end_addr); page++) page 843 sound/oss/sscape.c ClearPageReserved(page); page 498 sound/pci/ac97/ac97_codec.c unsigned short page = (kcontrol->private_value >> 26) & 0x0f; page 501 sound/pci/ac97/ac97_codec.c snd_ac97_update_bits(ac97, AC97_INT_PAGING, AC97_PAGE_MASK, page); page 65 sound/pci/ac97/ac97_patch.c snd_ac97_update_bits(ac97, AC97_INT_PAGING, AC97_PAGE_MASK, page); page 29 sound/pci/ac97/ac97_patch.h (AC97_SINGLE_VALUE(reg,shift,mask,invert) | (1<<25) | ((page) << 26)) page 39 sound/pci/ac97/ac97_patch.h .private_value = AC97_PAGE_SINGLE_VALUE(reg, shift, mask, invert, page) } page 1172 sound/pci/au88x0/au88x0_core.c int page, p, pp, delta, i; page 1174 sound/pci/au88x0/au88x0_core.c page = page 1178 sound/pci/au88x0/au88x0_core.c delta = (page - dma->period_real) & 3; page 1180 sound/pci/au88x0/au88x0_core.c delta = (page - dma->period_real); page 1209 sound/pci/au88x0/au88x0_core.c dma->period_real = page; page 1433 sound/pci/au88x0/au88x0_core.c int page, p, pp, delta, i; page 1435 sound/pci/au88x0/au88x0_core.c page = page 1440 sound/pci/au88x0/au88x0_core.c delta = (page - dma->period_real) & 3; page 1442 sound/pci/au88x0/au88x0_core.c delta = (page - dma->period_real); page 1473 sound/pci/au88x0/au88x0_core.c dma->period_real = page; page 402 sound/pci/au88x0/au88x0_pcm.c .page = snd_pcm_sgbuf_ops_page, page 552 sound/pci/bt87x.c .page = snd_pcm_sgbuf_ops_page, page 492 sound/pci/echoaudio/echoaudio.c int err, per, rest, page, edge, offs; page 535 sound/pci/echoaudio/echoaudio.c for (offs = page = per = 0; offs < params_buffer_bytes(hw_params); page 556 sound/pci/echoaudio/echoaudio.c page++; page 800 sound/pci/echoaudio/echoaudio.c .page = snd_pcm_sgbuf_ops_page, page 811 sound/pci/echoaudio/echoaudio.c .page = snd_pcm_sgbuf_ops_page, page 824 sound/pci/echoaudio/echoaudio.c .page = snd_pcm_sgbuf_ops_page, page 836 sound/pci/echoaudio/echoaudio.c .page = snd_pcm_sgbuf_ops_page, page 1326 sound/pci/emu10k1/emupcm.c .page = snd_pcm_sgbuf_ops_page, page 1350 sound/pci/emu10k1/emupcm.c .page = snd_pcm_sgbuf_ops_page, page 35 sound/pci/emu10k1/memory.c (((u32 *)(emu)->ptb_pages.area)[page] = cpu_to_le32(((addr) << 1) | (page))) page 42 sound/pci/emu10k1/memory.c #define aligned_page_offset(page) ((page) << PAGE_SHIFT) page 47 sound/pci/emu10k1/memory.c #define set_ptb_entry(emu,page,addr) __set_ptb_entry(emu,page,addr) page 49 sound/pci/emu10k1/memory.c #define set_silent_ptb(emu,page) __set_ptb_entry(emu,page,emu->silent_page.addr) page 55 sound/pci/emu10k1/memory.c page *= UNIT_PAGES; page 56 sound/pci/emu10k1/memory.c for (i = 0; i < UNIT_PAGES; i++, page++) { page 57 sound/pci/emu10k1/memory.c __set_ptb_entry(emu, page, addr); page 64 sound/pci/emu10k1/memory.c page *= UNIT_PAGES; page 65 sound/pci/emu10k1/memory.c for (i = 0; i < UNIT_PAGES; i++, page++) page 67 sound/pci/emu10k1/memory.c __set_ptb_entry(emu, page, emu->silent_page.addr); page 102 sound/pci/emu10k1/memory.c int page = 0, found_page = -ENOMEM; page 112 sound/pci/emu10k1/memory.c size = blk->mapped_page - page; page 115 sound/pci/emu10k1/memory.c return page; page 121 sound/pci/emu10k1/memory.c found_page = page; page 123 sound/pci/emu10k1/memory.c page = blk->mapped_page + blk->pages; page 125 sound/pci/emu10k1/memory.c size = MAX_ALIGN_PAGES - page; page 128 sound/pci/emu10k1/memory.c return page; page 141 sound/pci/emu10k1/memory.c int page, pg; page 144 sound/pci/emu10k1/memory.c page = search_empty_map_area(emu, blk->pages, &next); page 145 sound/pci/emu10k1/memory.c if (page < 0) /* not found */ page 146 sound/pci/emu10k1/memory.c return page; page 151 sound/pci/emu10k1/memory.c blk->mapped_page = page; page 154 sound/pci/emu10k1/memory.c set_ptb_entry(emu, page, emu->page_addr_table[pg]); page 155 sound/pci/emu10k1/memory.c page++; page 207 sound/pci/emu10k1/memory.c int page, psize; page 210 sound/pci/emu10k1/memory.c page = 0; page 213 sound/pci/emu10k1/memory.c if (page + psize <= blk->first_page) page 215 sound/pci/emu10k1/memory.c page = blk->last_page + 1; page 217 sound/pci/emu10k1/memory.c if (page + psize > emu->max_cache_pages) page 225 sound/pci/emu10k1/memory.c blk->mem.offset = aligned_page_offset(page); /* set aligned offset */ page 301 sound/pci/emu10k1/memory.c int page, err, idx; page 322 sound/pci/emu10k1/memory.c for (page = blk->first_page; page <= blk->last_page; page++, idx++) { page 331 sound/pci/emu10k1/memory.c emu->page_addr_table[page] = addr; page 332 sound/pci/emu10k1/memory.c emu->page_ptr_table[page] = NULL; page 442 sound/pci/emu10k1/memory.c int page; page 444 sound/pci/emu10k1/memory.c for (page = first_page; page <= last_page; page++) { page 445 sound/pci/emu10k1/memory.c free_page((unsigned long)emu->page_ptr_table[page]); page 446 sound/pci/emu10k1/memory.c emu->page_addr_table[page] = 0; page 447 sound/pci/emu10k1/memory.c emu->page_ptr_table[page] = NULL; page 456 sound/pci/emu10k1/memory.c int page, first_page, last_page; page 461 sound/pci/emu10k1/memory.c for (page = first_page; page <= last_page; page++) { page 463 sound/pci/emu10k1/memory.c struct page *p = alloc_page(GFP_KERNEL | GFP_DMA32 | page 474 sound/pci/emu10k1/memory.c __synth_free_pages(emu, first_page, page - 1); page 477 sound/pci/emu10k1/memory.c emu->page_addr_table[page] = page_to_phys(p); page 478 sound/pci/emu10k1/memory.c emu->page_ptr_table[page] = page_address(p); page 499 sound/pci/emu10k1/memory.c if (snd_BUG_ON(page < 0 || page >= emu->max_cache_pages)) page 501 sound/pci/emu10k1/memory.c ptr = emu->page_ptr_table[page]; page 503 sound/pci/emu10k1/memory.c printk(KERN_ERR "emu10k1: access to NULL ptr: page = %d\n", page); page 516 sound/pci/emu10k1/memory.c int page, nextofs, end_offset, temp, temp1; page 522 sound/pci/emu10k1/memory.c page = get_aligned_page(offset); page 524 sound/pci/emu10k1/memory.c nextofs = aligned_page_offset(page + 1); page 529 sound/pci/emu10k1/memory.c ptr = offset_ptr(emu, page + p->first_page, offset); page 533 sound/pci/emu10k1/memory.c page++; page 546 sound/pci/emu10k1/memory.c int page, nextofs, end_offset, temp, temp1; page 552 sound/pci/emu10k1/memory.c page = get_aligned_page(offset); page 554 sound/pci/emu10k1/memory.c nextofs = aligned_page_offset(page + 1); page 559 sound/pci/emu10k1/memory.c ptr = offset_ptr(emu, page + p->first_page, offset); page 564 sound/pci/emu10k1/memory.c page++; page 1716 sound/pci/hda/hda_intel.c .page = snd_pcm_sgbuf_ops_page, page 1702 sound/pci/riptide/riptide.c .page = snd_pcm_sgbuf_ops_page, page 1713 sound/pci/riptide/riptide.c .page = snd_pcm_sgbuf_ops_page, page 4186 sound/pci/rme9652/hdspm.c .page = snd_pcm_sgbuf_ops_page, page 4199 sound/pci/rme9652/hdspm.c .page = snd_pcm_sgbuf_ops_page, page 2087 sound/pci/trident/trident_main.c .page = snd_pcm_sgbuf_ops_page, page 2132 sound/pci/trident/trident_main.c .page = snd_pcm_sgbuf_ops_page, page 38 sound/pci/trident/trident_memory.c do { (trident)->tlb.entries[page] = cpu_to_le32((addr) & ~(SNDRV_TRIDENT_PAGE_SIZE-1)); \ page 39 sound/pci/trident/trident_memory.c (trident)->tlb.shadow_entries[page] = (ptr); } while (0) page 41 sound/pci/trident/trident_memory.c (void*)((trident)->tlb.shadow_entries[page]) page 43 sound/pci/trident/trident_memory.c (dma_addr_t)le32_to_cpu((trident->tlb.entries[page]) & ~(SNDRV_TRIDENT_PAGE_SIZE - 1)) page 50 sound/pci/trident/trident_memory.c #define set_tlb_bus(trident,page,ptr,addr) __set_tlb_bus(trident,page,ptr,addr) page 52 sound/pci/trident/trident_memory.c #define set_silent_tlb(trident,page) __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr) page 56 sound/pci/trident/trident_memory.c #define aligned_page_offset(page) ((page) << 12) page 58 sound/pci/trident/trident_memory.c #define page_to_ptr(trident,page) __tlb_to_ptr(trident, page) page 60 sound/pci/trident/trident_memory.c #define page_to_addr(trident,page) __tlb_to_addr(trident, page) page 67 sound/pci/trident/trident_memory.c #define aligned_page_offset(page) ((page) << 13) page 68 sound/pci/trident/trident_memory.c #define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) << 1) page 69 sound/pci/trident/trident_memory.c #define page_to_addr(trident,page) __tlb_to_addr(trident, (page) << 1) page 75 sound/pci/trident/trident_memory.c page <<= 1; page 76 sound/pci/trident/trident_memory.c __set_tlb_bus(trident, page, ptr, addr); page 77 sound/pci/trident/trident_memory.c __set_tlb_bus(trident, page+1, ptr + SNDRV_TRIDENT_PAGE_SIZE, addr + SNDRV_TRIDENT_PAGE_SIZE); page 81 sound/pci/trident/trident_memory.c page <<= 1; page 82 sound/pci/trident/trident_memory.c __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); page 83 sound/pci/trident/trident_memory.c __set_tlb_bus(trident, page+1, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); page 97 sound/pci/trident/trident_memory.c #define aligned_page_offset(page) ((page) * ALIGN_PAGE_SIZE) page 98 sound/pci/trident/trident_memory.c #define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) * UNIT_PAGES) page 99 sound/pci/trident/trident_memory.c #define page_to_addr(trident,page) __tlb_to_addr(trident, (page) * UNIT_PAGES) page 106 sound/pci/trident/trident_memory.c page *= UNIT_PAGES; page 107 sound/pci/trident/trident_memory.c for (i = 0; i < UNIT_PAGES; i++, page++) { page 108 sound/pci/trident/trident_memory.c __set_tlb_bus(trident, page, ptr, addr); page 116 sound/pci/trident/trident_memory.c page *= UNIT_PAGES; page 117 sound/pci/trident/trident_memory.c for (i = 0; i < UNIT_PAGES; i++, page++) page 118 sound/pci/trident/trident_memory.c __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); page 143 sound/pci/trident/trident_memory.c int page, psize; page 148 sound/pci/trident/trident_memory.c page = 0; page 151 sound/pci/trident/trident_memory.c if (page + psize <= firstpg(blk)) page 153 sound/pci/trident/trident_memory.c page = lastpg(blk) + 1; page 155 sound/pci/trident/trident_memory.c if (page + psize > MAX_ALIGN_PAGES) page 163 sound/pci/trident/trident_memory.c blk->offset = aligned_page_offset(page); /* set aligned offset */ page 164 sound/pci/trident/trident_memory.c firstpg(blk) = page; page 165 sound/pci/trident/trident_memory.c lastpg(blk) = page + psize - 1; page 196 sound/pci/trident/trident_memory.c int idx, page; page 217 sound/pci/trident/trident_memory.c for (page = firstpg(blk); page <= lastpg(blk); page++, idx++) { page 227 sound/pci/trident/trident_memory.c set_tlb_bus(trident, page, ptr, addr); page 242 sound/pci/trident/trident_memory.c int page; page 265 sound/pci/trident/trident_memory.c for (page = firstpg(blk); page <= lastpg(blk); page++, page 272 sound/pci/trident/trident_memory.c set_tlb_bus(trident, page, ptr, addr); page 301 sound/pci/trident/trident_memory.c int page; page 309 sound/pci/trident/trident_memory.c for (page = firstpg(blk); page <= lastpg(blk); page++) page 310 sound/pci/trident/trident_memory.c set_silent_tlb(trident, page); page 1309 sound/pci/via82xx.c .page = snd_pcm_sgbuf_ops_page, page 1322 sound/pci/via82xx.c .page = snd_pcm_sgbuf_ops_page, page 1335 sound/pci/via82xx.c .page = snd_pcm_sgbuf_ops_page, page 1348 sound/pci/via82xx.c .page = snd_pcm_sgbuf_ops_page, page 1361 sound/pci/via82xx.c .page = snd_pcm_sgbuf_ops_page, page 808 sound/pci/via82xx_modem.c .page = snd_pcm_sgbuf_ops_page, page 821 sound/pci/via82xx_modem.c .page = snd_pcm_sgbuf_ops_page, page 36 sound/pcmcia/pdaudiocf/pdaudiocf_pcm.c static struct page *snd_pcm_get_vmalloc_page(struct snd_pcm_substream *subs, unsigned long offset) page 321 sound/pcmcia/pdaudiocf/pdaudiocf_pcm.c .page = snd_pcm_get_vmalloc_page, page 14 sound/soc/codecs/tlv320aic26.h #define AIC26_PAGE_ADDR(page, offset) ((page << 6) | offset) page 942 sound/soc/soc-core.c soc_pcm_ops.page = socdev->platform->pcm_ops->page; page 730 sound/usb/usbaudio.c static struct page *snd_pcm_get_vmalloc_page(struct snd_pcm_substream *subs, page 2010 sound/usb/usbaudio.c .page = snd_pcm_get_vmalloc_page, page 2022 sound/usb/usbaudio.c .page = snd_pcm_get_vmalloc_page, page 117 sound/usb/usx2y/us122l.c struct page *page; page 138 sound/usb/usx2y/us122l.c page = virt_to_page(vaddr); page 140 sound/usb/usx2y/us122l.c get_page(page); page 143 sound/usb/usx2y/us122l.c vmf->page = page; page 40 sound/usb/usx2y/usX2Yhwdep.c struct page * page; page 49 sound/usb/usx2y/usX2Yhwdep.c page = virt_to_page(vaddr); page 50 sound/usb/usx2y/usX2Yhwdep.c get_page(page); page 51 sound/usb/usx2y/usX2Yhwdep.c vmf->page = page; page 54 sound/usb/usx2y/usX2Yhwdep.c vaddr, page); page 694 sound/usb/usx2y/usx2yhwdeppcm.c vmf->page = virt_to_page(vaddr); page 695 sound/usb/usx2y/usx2yhwdeppcm.c get_page(vmf->page); page 160 virt/kvm/kvm_main.c struct page *page; page 169 virt/kvm/kvm_main.c page = alloc_page(GFP_KERNEL | __GFP_ZERO); page 170 virt/kvm/kvm_main.c if (!page) { page 174 virt/kvm/kvm_main.c vcpu->run = page_address(page); page 316 virt/kvm/kvm_main.c struct page *page; page 323 virt/kvm/kvm_main.c page = alloc_page(GFP_KERNEL | __GFP_ZERO); page 324 virt/kvm/kvm_main.c if (!page) { page 329 virt/kvm/kvm_main.c (struct kvm_coalesced_mmio_ring *)page_address(page); page 339 virt/kvm/kvm_main.c put_page(page); page 505 virt/kvm/kvm_main.c new.rmap = vmalloc(npages * sizeof(struct page *)); page 642 virt/kvm/kvm_main.c return page == bad_page; page 716 virt/kvm/kvm_main.c struct page *page[1]; page 729 virt/kvm/kvm_main.c npages = get_user_pages(current, current->mm, addr, 1, 1, 1, page, page 745 virt/kvm/kvm_main.c pfn = page_to_pfn(page[0]); page 752 virt/kvm/kvm_main.c struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn) page 770 virt/kvm/kvm_main.c kvm_release_pfn_clean(page_to_pfn(page)); page 783 virt/kvm/kvm_main.c kvm_release_pfn_dirty(page_to_pfn(page)); page 796 virt/kvm/kvm_main.c kvm_set_pfn_dirty(page_to_pfn(page)); page 803 virt/kvm/kvm_main.c struct page *page = pfn_to_page(pfn); page 804 virt/kvm/kvm_main.c if (!PageReserved(page)) page 805 virt/kvm/kvm_main.c SetPageDirty(page); page 1003 virt/kvm/kvm_main.c struct page *page; page 1006 virt/kvm/kvm_main.c page = virt_to_page(vcpu->run); page 1009 virt/kvm/kvm_main.c page = virt_to_page(vcpu->arch.pio_data); page 1013 virt/kvm/kvm_main.c page = virt_to_page(vcpu->kvm->coalesced_mmio_ring); page 1017 virt/kvm/kvm_main.c get_page(page); page 1018 virt/kvm/kvm_main.c vmf->page = page; page 1373 virt/kvm/kvm_main.c struct page *page; page 1377 virt/kvm/kvm_main.c page = gfn_to_page(kvm, vmf->pgoff); page 1378 virt/kvm/kvm_main.c if (is_error_page(page)) { page 1379 virt/kvm/kvm_main.c kvm_release_page_clean(page); page 1382 virt/kvm/kvm_main.c vmf->page = page; page 1679 virt/kvm/kvm_main.c struct page *bad_page;