zone 1062 arch/x86/mm/init_32.c struct zone *zone = pgdata->node_zones + ZONE_HIGHMEM; zone 1066 arch/x86/mm/init_32.c return __add_pages(zone, start_pfn, nr_pages); zone 829 arch/x86/mm/init_64.c struct zone *zone = pgdat->node_zones + ZONE_NORMAL; zone 838 arch/x86/mm/init_64.c ret = __add_pages(zone, start_pfn, nr_pages); zone 394 arch/x86/mm/numa_32.c struct zone *zone; zone 397 arch/x86/mm/numa_32.c for_each_zone(zone) { zone 400 arch/x86/mm/numa_32.c if (!is_highmem(zone)) zone 403 arch/x86/mm/numa_32.c zone_start_pfn = zone->zone_start_pfn; zone 404 arch/x86/mm/numa_32.c zone_end_pfn = zone_start_pfn + zone->spanned_pages; zone 406 arch/x86/mm/numa_32.c nid = zone_to_nid(zone); zone 408 arch/x86/mm/numa_32.c zone->name, nid, zone_start_pfn, zone_end_pfn); zone 205 fs/adfs/map.c dm = asb->s_map + zone; zone 206 fs/adfs/map.c zone = asb->s_map_size; zone 207 fs/adfs/map.c dm_end = asb->s_map + zone; zone 218 fs/adfs/map.c } while (--zone > 0); zone 241 fs/adfs/map.c unsigned int zone; zone 244 fs/adfs/map.c zone = asb->s_map_size; zone 248 fs/adfs/map.c } while (--zone > 0); zone 258 fs/adfs/map.c unsigned int zone, mapoff; zone 266 fs/adfs/map.c zone = asb->s_map_size >> 1; zone 268 fs/adfs/map.c zone = frag_id / asb->s_ids_per_zone; zone 270 fs/adfs/map.c if (zone >= asb->s_map_size) zone 277 fs/adfs/map.c result = scan_map(asb, zone, frag_id, mapoff); zone 294 fs/adfs/map.c frag_id, zone, asb->s_map_size); zone 290 fs/adfs/super.c int i, zone; zone 307 fs/adfs/super.c for (zone = 0; zone < nzones; zone++, map_addr++) { zone 308 fs/adfs/super.c dm[zone].dm_startbit = 0; zone 309 fs/adfs/super.c dm[zone].dm_endbit = zone_size; zone 310 fs/adfs/super.c dm[zone].dm_startblk = zone * zone_size - ADFS_DR_SIZE_BITS; zone 311 fs/adfs/super.c dm[zone].dm_bh = sb_bread(sb, map_addr); zone 313 fs/adfs/super.c if (!dm[zone].dm_bh) { zone 320 fs/adfs/super.c i = zone - 1; zone 333 fs/adfs/super.c while (--zone >= 0) zone 334 fs/adfs/super.c brelse(dm[zone].dm_bh); zone 363 fs/buffer.c struct zone *zone; zone 372 fs/buffer.c &zone); zone 373 fs/buffer.c if (zone) zone 58 fs/minix/bitmap.c unsigned long bit, zone; zone 64 fs/minix/bitmap.c zone = block - sbi->s_firstdatazone + 1; zone 65 fs/minix/bitmap.c bit = zone & ((1<<k) - 1); zone 66 fs/minix/bitmap.c zone >>= k; zone 67 fs/minix/bitmap.c if (zone >= sbi->s_zmap_blocks) { zone 71 fs/minix/bitmap.c bh = sbi->s_zmap[zone]; zone 167 fs/ntfs/lcnalloc.c zone == MFT_ZONE ? "MFT" : "DATA"); zone 174 fs/ntfs/lcnalloc.c BUG_ON(zone < FIRST_ZONE); zone 175 fs/ntfs/lcnalloc.c BUG_ON(zone > LAST_ZONE); zone 202 fs/ntfs/lcnalloc.c if (zone == DATA_ZONE) zone 213 fs/ntfs/lcnalloc.c } else if (zone == DATA_ZONE && zone_start >= vol->mft_zone_start && zone 221 fs/ntfs/lcnalloc.c } else if (zone == MFT_ZONE && (zone_start < vol->mft_zone_start || zone 232 fs/ntfs/lcnalloc.c if (zone == MFT_ZONE) { zone 696 fs/ntfs/lcnalloc.c if (zone == MFT_ZONE || mft_zone_size <= 0) { zone 187 fs/sysv/balloc.c sysv_zone_t zone; zone 190 fs/sysv/balloc.c zone = 0; zone 191 fs/sysv/balloc.c while (n && (zone = blocks[--n]) != 0) zone 193 fs/sysv/balloc.c if (zone == 0) zone 196 fs/sysv/balloc.c block = fs32_to_cpu(sbi, zone); zone 126 fs/xfs/linux-2.6/kmem.c ptr = kmem_cache_alloc(zone, lflags); zone 142 fs/xfs/linux-2.6/kmem.c ptr = kmem_zone_alloc(zone, flags); zone 144 fs/xfs/linux-2.6/kmem.c memset((char *)ptr, 0, kmem_cache_size(zone)); zone 90 fs/xfs/linux-2.6/kmem.h kmem_cache_free(zone, ptr); zone 96 fs/xfs/linux-2.6/kmem.h if (zone) zone 97 fs/xfs/linux-2.6/kmem.h kmem_cache_destroy(zone); zone 59 fs/xfs/xfs_acl.h (zone) = kmem_zone_init(sizeof(xfs_acl_t), (name)) zone 60 fs/xfs/xfs_acl.h #define xfs_acl_zone_destroy(zone) kmem_zone_destroy(zone) zone 26 fs/xfs/xfs_da_btree.h struct zone; zone 9 include/linux/memory_hotplug.h struct zone; zone 46 include/linux/memory_hotplug.h return read_seqbegin(&zone->span_seqlock); zone 50 include/linux/memory_hotplug.h return read_seqretry(&zone->span_seqlock, iv); zone 54 include/linux/memory_hotplug.h write_seqlock(&zone->span_seqlock); zone 58 include/linux/memory_hotplug.h write_sequnlock(&zone->span_seqlock); zone 62 include/linux/memory_hotplug.h seqlock_init(&zone->span_seqlock); zone 511 include/linux/mm.h return zone->node; zone 526 include/linux/mm.h static inline struct zone *page_zone(struct page *page) zone 541 include/linux/mm.h page->flags |= (zone & ZONES_MASK) << ZONES_PGSHIFT; zone 559 include/linux/mm.h set_page_zone(page, zone); zone 4 include/linux/mm_inline.h list_add(&page->lru, &zone->active_list); zone 5 include/linux/mm_inline.h __inc_zone_state(zone, NR_ACTIVE); zone 11 include/linux/mm_inline.h list_add(&page->lru, &zone->inactive_list); zone 12 include/linux/mm_inline.h __inc_zone_state(zone, NR_INACTIVE); zone 19 include/linux/mm_inline.h __dec_zone_state(zone, NR_ACTIVE); zone 26 include/linux/mm_inline.h __dec_zone_state(zone, NR_INACTIVE); zone 35 include/linux/mm_inline.h __dec_zone_state(zone, NR_ACTIVE); zone 37 include/linux/mm_inline.h __dec_zone_state(zone, NR_INACTIVE); zone 345 include/linux/mmzone.h set_bit(flag, &zone->flags); zone 350 include/linux/mmzone.h return test_and_set_bit(flag, &zone->flags); zone 355 include/linux/mmzone.h clear_bit(flag, &zone->flags); zone 360 include/linux/mmzone.h return test_bit(ZONE_ALL_UNRECLAIMABLE, &zone->flags); zone 365 include/linux/mmzone.h return test_bit(ZONE_RECLAIM_LOCKED, &zone->flags); zone 370 include/linux/mmzone.h return test_bit(ZONE_OOM_LOCKED, &zone->flags); zone 469 include/linux/mmzone.h struct zone *zone; /* Pointer to actual zone */ zone 524 include/linux/mmzone.h struct zone node_zones[MAX_NR_ZONES]; zone 589 include/linux/mmzone.h #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) zone 593 include/linux/mmzone.h return (!!zone->present_pages); zone 631 include/linux/mmzone.h int zone_off = (char *)zone - (char *)zone->zone_pgdat->node_zones; zone 632 include/linux/mmzone.h return zone_off == ZONE_HIGHMEM * sizeof(*zone) || zone 633 include/linux/mmzone.h (zone_off == ZONE_MOVABLE * sizeof(*zone) && zone 642 include/linux/mmzone.h return zone == zone->zone_pgdat->node_zones + ZONE_NORMAL; zone 648 include/linux/mmzone.h return zone == zone->zone_pgdat->node_zones + ZONE_DMA32; zone 657 include/linux/mmzone.h return zone == zone->zone_pgdat->node_zones + ZONE_DMA; zone 703 include/linux/mmzone.h extern struct zone *next_zone(struct zone *zone); zone 721 include/linux/mmzone.h for (zone = (first_online_pgdat())->node_zones; \ zone 722 include/linux/mmzone.h zone; \ zone 723 include/linux/mmzone.h zone = next_zone(zone)) zone 725 include/linux/mmzone.h static inline struct zone *zonelist_zone(struct zoneref *zoneref) zone 727 include/linux/mmzone.h return zoneref->zone; zone 739 include/linux/mmzone.h return zoneref->zone->node; zone 781 include/linux/mmzone.h zone); zone 796 include/linux/mmzone.h for (z = first_zones_zonelist(zlist, highidx, nodemask, &zone); \ zone 797 include/linux/mmzone.h zone; \ zone 798 include/linux/mmzone.h z = next_zones_zonelist(++z, highidx, nodemask, &zone)) \ zone 810 include/linux/mmzone.h for_each_zone_zonelist_nodemask(zone, z, zlist, highidx, NULL) zone 95 include/linux/swap.h struct zone; zone 66 include/linux/tipc.h return (zone << 24) | (cluster << 12) | node; zone 127 include/linux/vmstat.h zone_idx(zone), delta) zone 137 include/linux/vmstat.h atomic_long_add(x, &zone->vm_stat[item]); zone 154 include/linux/vmstat.h long x = atomic_long_read(&zone->vm_stat[item]); zone 171 include/linux/vmstat.h struct zone *zones = NODE_DATA(node)->node_zones; zone 206 include/linux/vmstat.h memset(zone->vm_stat, 0, sizeof(zone->vm_stat)); zone 235 include/linux/vmstat.h zone_page_state_add(delta, zone, item); zone 240 include/linux/vmstat.h atomic_long_inc(&zone->vm_stat[item]); zone 252 include/linux/vmstat.h atomic_long_dec(&zone->vm_stat[item]); zone 183 include/sound/emux_synth.h struct snd_sf_zone *zone; /* Zone assigned to this note */ zone 1387 kernel/kexec.c VMCOREINFO_STRUCT_SIZE(zone); zone 1403 kernel/kexec.c VMCOREINFO_OFFSET(zone, free_area); zone 1404 kernel/kexec.c VMCOREINFO_OFFSET(zone, vm_stat); zone 1405 kernel/kexec.c VMCOREINFO_OFFSET(zone, spanned_pages); zone 1409 kernel/kexec.c VMCOREINFO_LENGTH(zone.free_area, MAX_ORDER); zone 338 kernel/power/snapshot.c struct zone *zone; zone 347 kernel/power/snapshot.c for_each_zone(zone) zone 348 kernel/power/snapshot.c if (populated_zone(zone)) zone 360 kernel/power/snapshot.c for_each_zone(zone) { zone 363 kernel/power/snapshot.c if (!populated_zone(zone)) zone 366 kernel/power/snapshot.c zone_bm->start_pfn = zone->zone_start_pfn; zone 367 kernel/power/snapshot.c zone_bm->end_pfn = zone->zone_start_pfn + zone->spanned_pages; zone 369 kernel/power/snapshot.c nr = DIV_ROUND_UP(zone->spanned_pages, BM_BITS_PER_BLOCK); zone 376 kernel/power/snapshot.c nr = zone->spanned_pages; zone 377 kernel/power/snapshot.c pfn = zone->zone_start_pfn; zone 781 kernel/power/snapshot.c res = DIV_ROUND_UP(zone->spanned_pages, BM_BITS_PER_BLOCK); zone 794 kernel/power/snapshot.c struct zone *zone; zone 797 kernel/power/snapshot.c for_each_zone(zone) zone 798 kernel/power/snapshot.c if (populated_zone(zone) && is_highmem(zone)) zone 799 kernel/power/snapshot.c cnt += zone_page_state(zone, NR_FREE_PAGES); zone 837 kernel/power/snapshot.c struct zone *zone; zone 840 kernel/power/snapshot.c for_each_zone(zone) { zone 843 kernel/power/snapshot.c if (!is_highmem(zone)) zone 846 kernel/power/snapshot.c mark_free_pages(zone); zone 847 kernel/power/snapshot.c max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; zone 848 kernel/power/snapshot.c for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) zone 895 kernel/power/snapshot.c struct zone *zone; zone 899 kernel/power/snapshot.c for_each_zone(zone) { zone 900 kernel/power/snapshot.c if (is_highmem(zone)) zone 903 kernel/power/snapshot.c mark_free_pages(zone); zone 904 kernel/power/snapshot.c max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; zone 905 kernel/power/snapshot.c for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) zone 946 kernel/power/snapshot.c return is_highmem(zone) ? zone 990 kernel/power/snapshot.c struct zone *zone; zone 993 kernel/power/snapshot.c for_each_zone(zone) { zone 996 kernel/power/snapshot.c mark_free_pages(zone); zone 997 kernel/power/snapshot.c max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; zone 998 kernel/power/snapshot.c for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) zone 999 kernel/power/snapshot.c if (page_is_saveable(zone, pfn)) zone 1026 kernel/power/snapshot.c struct zone *zone; zone 1029 kernel/power/snapshot.c for_each_zone(zone) { zone 1030 kernel/power/snapshot.c max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; zone 1031 kernel/power/snapshot.c for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) zone 1078 kernel/power/snapshot.c struct zone *zone; zone 1081 kernel/power/snapshot.c for_each_zone(zone) { zone 1082 kernel/power/snapshot.c meta += snapshot_additional_pages(zone); zone 1083 kernel/power/snapshot.c if (!is_highmem(zone)) zone 1084 kernel/power/snapshot.c free += zone_page_state(zone, NR_FREE_PAGES); zone 1385 kernel/power/snapshot.c struct zone *zone; zone 1389 kernel/power/snapshot.c for_each_zone(zone) { zone 1390 kernel/power/snapshot.c max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; zone 1391 kernel/power/snapshot.c for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) zone 217 kernel/power/swsusp.c struct zone *zone; zone 232 kernel/power/swsusp.c for_each_zone (zone) zone 233 kernel/power/swsusp.c if (populated_zone(zone)) { zone 234 kernel/power/swsusp.c tmp += snapshot_additional_pages(zone); zone 235 kernel/power/swsusp.c if (is_highmem(zone)) { zone 237 kernel/power/swsusp.c zone_page_state(zone, NR_FREE_PAGES); zone 239 kernel/power/swsusp.c tmp -= zone_page_state(zone, NR_FREE_PAGES); zone 240 kernel/power/swsusp.c tmp += zone->lowmem_reserve[ZONE_NORMAL]; zone 248 kernel/wait.c const struct zone *zone = page_zone(virt_to_page(word)); zone 251 kernel/wait.c return &zone->wait_table[hash_long(val, zone->wait_table_bits)]; zone 531 mm/filemap.c const struct zone *zone = page_zone(page); zone 533 mm/filemap.c return &zone->wait_table[hash_ptr(page, zone->wait_table_bits)]; zone 417 mm/hugetlb.c struct zone *zone; zone 433 mm/hugetlb.c for_each_zone_zonelist_nodemask(zone, z, zonelist, zone 435 mm/hugetlb.c nid = zone_to_nid(zone); zone 436 mm/hugetlb.c if (cpuset_zone_allowed_softwall(zone, htlb_alloc_mask) && zone 451 mm/memcontrol.c int nid = zone->zone_pgdat->node_id; zone 452 mm/memcontrol.c int zid = zone_idx(zone); zone 463 mm/memcontrol.c int nid = zone->zone_pgdat->node_id; zone 464 mm/memcontrol.c int zid = zone_idx(zone); zone 1065 mm/memcontrol.c int zone, tmp = node; zone 1083 mm/memcontrol.c for (zone = 0; zone < MAX_NR_ZONES; zone++) { zone 1084 mm/memcontrol.c mz = &pn->zoneinfo[zone]; zone 131 mm/memory_hotplug.c struct zone *zone; zone 139 mm/memory_hotplug.c zone = &pgdat->node_zones[0]; zone 140 mm/memory_hotplug.c for (; zone < pgdat->node_zones + MAX_NR_ZONES - 1; zone++) { zone 141 mm/memory_hotplug.c if (zone->wait_table) { zone 142 mm/memory_hotplug.c nr_pages = zone->wait_table_hash_nr_entries zone 145 mm/memory_hotplug.c page = virt_to_page(zone->wait_table); zone 167 mm/memory_hotplug.c zone_span_writelock(zone); zone 169 mm/memory_hotplug.c old_zone_end_pfn = zone->zone_start_pfn + zone->spanned_pages; zone 170 mm/memory_hotplug.c if (start_pfn < zone->zone_start_pfn) zone 171 mm/memory_hotplug.c zone->zone_start_pfn = start_pfn; zone 173 mm/memory_hotplug.c zone->spanned_pages = max(old_zone_end_pfn, end_pfn) - zone 174 mm/memory_hotplug.c zone->zone_start_pfn; zone 176 mm/memory_hotplug.c zone_span_writeunlock(zone); zone 194 mm/memory_hotplug.c struct pglist_data *pgdat = zone->zone_pgdat; zone 200 mm/memory_hotplug.c zone_type = zone - pgdat->node_zones; zone 201 mm/memory_hotplug.c if (!zone->wait_table) { zone 204 mm/memory_hotplug.c ret = init_currently_empty_zone(zone, phys_start_pfn, zone 209 mm/memory_hotplug.c pgdat_resize_lock(zone->zone_pgdat, &flags); zone 210 mm/memory_hotplug.c grow_zone_span(zone, phys_start_pfn, phys_start_pfn + nr_pages); zone 211 mm/memory_hotplug.c grow_pgdat_span(zone->zone_pgdat, phys_start_pfn, zone 213 mm/memory_hotplug.c pgdat_resize_unlock(zone->zone_pgdat, &flags); zone 227 mm/memory_hotplug.c ret = sparse_add_one_section(zone, phys_start_pfn, nr_pages); zone 232 mm/memory_hotplug.c ret = __add_zone(zone, phys_start_pfn); zone 253 mm/memory_hotplug.c struct pglist_data *pgdat = zone->zone_pgdat; zone 264 mm/memory_hotplug.c sparse_remove_one_section(zone, ms); zone 287 mm/memory_hotplug.c err = __add_section(zone, i << PFN_SECTION_SHIFT); zone 331 mm/memory_hotplug.c ret = __remove_section(zone, __pfn_to_section(pfn)); zone 378 mm/memory_hotplug.c struct zone *zone; zone 403 mm/memory_hotplug.c zone = page_zone(pfn_to_page(pfn)); zone 409 mm/memory_hotplug.c if (!populated_zone(zone)) zone 421 mm/memory_hotplug.c zone->present_pages += onlined_pages; zone 422 mm/memory_hotplug.c zone->zone_pgdat->node_present_pages += onlined_pages; zone 426 mm/memory_hotplug.c kswapd_run(zone_to_nid(zone)); zone 427 mm/memory_hotplug.c node_set_state(zone_to_nid(zone), N_HIGH_MEMORY); zone 591 mm/memory_hotplug.c struct zone *zone = NULL; zone 604 mm/memory_hotplug.c if (zone && page_zone(page) != zone) zone 606 mm/memory_hotplug.c zone = page_zone(page); zone 744 mm/memory_hotplug.c struct zone *zone; zone 758 mm/memory_hotplug.c zone = page_zone(pfn_to_page(start_pfn)); zone 759 mm/memory_hotplug.c node = zone_to_nid(zone); zone 832 mm/memory_hotplug.c zone->present_pages -= offlined_pages; zone 833 mm/memory_hotplug.c zone->zone_pgdat->node_present_pages -= offlined_pages; zone 131 mm/mempolicy.c struct zone *z; zone 1402 mm/mempolicy.c struct zone *zone; zone 1407 mm/mempolicy.c &zone); zone 1408 mm/mempolicy.c return zone->node; zone 52 mm/migrate.c struct zone *zone = page_zone(page); zone 54 mm/migrate.c spin_lock_irq(&zone->lru_lock); zone 59 mm/migrate.c del_page_from_active_list(zone, page); zone 61 mm/migrate.c del_page_from_inactive_list(zone, page); zone 64 mm/migrate.c spin_unlock_irq(&zone->lru_lock); zone 31 mm/mm_init.c struct zone *zone; zone 43 mm/mm_init.c zone = &pgdat->node_zones[zoneid]; zone 44 mm/mm_init.c if (!populated_zone(zone)) zone 50 mm/mm_init.c zone->name); zone 53 mm/mm_init.c for_each_zone_zonelist(zone, z, zonelist, zoneid) { zone 56 mm/mm_init.c zone->node, zone->name); zone 58 mm/mm_init.c printk(KERN_CONT "0:%s ", zone->name); zone 128 mm/mm_init.c BUG_ON(page_zonenum(page) != zone); zone 29 mm/mmzone.c struct zone *next_zone(struct zone *zone) zone 31 mm/mmzone.c pg_data_t *pgdat = zone->zone_pgdat; zone 33 mm/mmzone.c if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) zone 34 mm/mmzone.c zone++; zone 38 mm/mmzone.c zone = pgdat->node_zones; zone 40 mm/mmzone.c zone = NULL; zone 42 mm/mmzone.c return zone; zone 69 mm/mmzone.c (z->zone && !zref_in_nodemask(z, nodes))) zone 72 mm/mmzone.c *zone = zonelist_zone(z); zone 179 mm/oom_kill.c struct zone *zone; zone 184 mm/oom_kill.c for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) zone 185 mm/oom_kill.c if (cpuset_zone_allowed_softwall(zone, gfp_mask)) zone 186 mm/oom_kill.c node_clear(zone_to_nid(zone), nodes); zone 469 mm/oom_kill.c struct zone *zone; zone 473 mm/oom_kill.c for_each_zone_zonelist(zone, z, zonelist, gfp_zone(gfp_mask)) { zone 474 mm/oom_kill.c if (zone_is_oom_locked(zone)) { zone 480 mm/oom_kill.c for_each_zone_zonelist(zone, z, zonelist, gfp_zone(gfp_mask)) { zone 486 mm/oom_kill.c zone_set_flag(zone, ZONE_OOM_LOCKED); zone 502 mm/oom_kill.c struct zone *zone; zone 505 mm/oom_kill.c for_each_zone_zonelist(zone, z, zonelist, gfp_zone(gfp_mask)) { zone 506 mm/oom_kill.c zone_clear_flag(zone, ZONE_OOM_LOCKED); zone 329 mm/page-writeback.c struct zone *z = zone 186 mm/page_alloc.c seq = zone_span_seqbegin(zone); zone 187 mm/page_alloc.c if (pfn >= zone->zone_start_pfn + zone->spanned_pages) zone 189 mm/page_alloc.c else if (pfn < zone->zone_start_pfn) zone 191 mm/page_alloc.c } while (zone_span_seqretry(zone, seq)); zone 200 mm/page_alloc.c if (zone != page_zone(page)) zone 210 mm/page_alloc.c if (page_outside_zone_boundaries(zone, page)) zone 212 mm/page_alloc.c if (!page_is_consistent(zone, page)) zone 429 mm/page_alloc.c VM_BUG_ON(bad_range(zone, page)); zone 431 mm/page_alloc.c __mod_zone_page_state(zone, NR_FREE_PAGES, order_size); zone 442 mm/page_alloc.c zone->free_area[order].nr_free--; zone 451 mm/page_alloc.c &zone->free_area[order].free_list[migratetype]); zone 452 mm/page_alloc.c zone->free_area[order].nr_free++; zone 487 mm/page_alloc.c spin_lock(&zone->lock); zone 488 mm/page_alloc.c zone_clear_flag(zone, ZONE_ALL_UNRECLAIMABLE); zone 489 mm/page_alloc.c zone->pages_scanned = 0; zone 497 mm/page_alloc.c __free_one_page(page, zone, order); zone 499 mm/page_alloc.c spin_unlock(&zone->lock); zone 504 mm/page_alloc.c spin_lock(&zone->lock); zone 505 mm/page_alloc.c zone_clear_flag(zone, ZONE_ALL_UNRECLAIMABLE); zone 506 mm/page_alloc.c zone->pages_scanned = 0; zone 507 mm/page_alloc.c __free_one_page(page, zone, order); zone 508 mm/page_alloc.c spin_unlock(&zone->lock); zone 589 mm/page_alloc.c VM_BUG_ON(bad_range(zone, &page[size])); zone 646 mm/page_alloc.c area = &(zone->free_area[current_order]); zone 655 mm/page_alloc.c __mod_zone_page_state(zone, NR_FREE_PAGES, - (1UL << order)); zone 656 mm/page_alloc.c expand(zone, page, order, current_order, area, migratetype); zone 701 mm/page_alloc.c VM_BUG_ON(page_to_nid(page) != zone_to_nid(zone)); zone 716 mm/page_alloc.c &zone->free_area[order].free_list[migratetype]); zone 737 mm/page_alloc.c if (start_pfn < zone->zone_start_pfn) zone 739 mm/page_alloc.c if (end_pfn >= zone->zone_start_pfn + zone->spanned_pages) zone 742 mm/page_alloc.c return move_freepages(zone, start_page, end_page, migratetype); zone 764 mm/page_alloc.c area = &(zone->free_area[current_order]); zone 781 mm/page_alloc.c pages = move_freepages_block(zone, page, zone 795 mm/page_alloc.c __mod_zone_page_state(zone, NR_FREE_PAGES, zone 802 mm/page_alloc.c expand(zone, page, order, current_order, area, migratetype); zone 808 mm/page_alloc.c return __rmqueue_smallest(zone, order, MIGRATE_RESERVE); zone 820 mm/page_alloc.c page = __rmqueue_smallest(zone, order, migratetype); zone 823 mm/page_alloc.c page = __rmqueue_fallback(zone, order, migratetype); zone 839 mm/page_alloc.c spin_lock(&zone->lock); zone 841 mm/page_alloc.c struct page *page = __rmqueue(zone, order, migratetype); zone 858 mm/page_alloc.c spin_unlock(&zone->lock); zone 881 mm/page_alloc.c free_pages_bulk(zone, to_drain, &pcp->list, 0); zone 897 mm/page_alloc.c struct zone *zone; zone 899 mm/page_alloc.c for_each_zone(zone) { zone 903 mm/page_alloc.c if (!populated_zone(zone)) zone 906 mm/page_alloc.c pset = zone_pcp(zone, cpu); zone 910 mm/page_alloc.c free_pages_bulk(zone, pcp->count, &pcp->list, 0); zone 941 mm/page_alloc.c if (!zone->spanned_pages) zone 944 mm/page_alloc.c spin_lock_irqsave(&zone->lock, flags); zone 946 mm/page_alloc.c max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; zone 947 mm/page_alloc.c for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) zone 956 mm/page_alloc.c list_for_each(curr, &zone->free_area[order].free_list[t]) { zone 964 mm/page_alloc.c spin_unlock_irqrestore(&zone->lock, flags); zone 973 mm/page_alloc.c struct zone *zone = page_zone(page); zone 989 mm/page_alloc.c pcp = &zone_pcp(zone, get_cpu())->pcp; zone 999 mm/page_alloc.c free_pages_bulk(zone, pcp->batch, &pcp->list, 0); zone 1053 mm/page_alloc.c pcp = &zone_pcp(zone, cpu)->pcp; zone 1056 mm/page_alloc.c pcp->count = rmqueue_bulk(zone, 0, zone 1075 mm/page_alloc.c pcp->count += rmqueue_bulk(zone, 0, zone 1083 mm/page_alloc.c spin_lock_irqsave(&zone->lock, flags); zone 1084 mm/page_alloc.c page = __rmqueue(zone, order, migratetype); zone 1085 mm/page_alloc.c spin_unlock(&zone->lock); zone 1090 mm/page_alloc.c __count_zone_vm_events(PGALLOC, zone, 1 << order); zone 1091 mm/page_alloc.c zone_statistics(preferred_zone, zone); zone 1095 mm/page_alloc.c VM_BUG_ON(bad_range(zone, page)); zone 1372 mm/page_alloc.c struct zone *zone, *preferred_zone; zone 1389 mm/page_alloc.c for_each_zone_zonelist_nodemask(zone, z, zonelist, zone 1395 mm/page_alloc.c !cpuset_zone_allowed_softwall(zone, gfp_mask)) zone 1401 mm/page_alloc.c mark = zone->pages_min; zone 1403 mm/page_alloc.c mark = zone->pages_low; zone 1405 mm/page_alloc.c mark = zone->pages_high; zone 1406 mm/page_alloc.c if (!zone_watermark_ok(zone, order, mark, zone 1409 mm/page_alloc.c !zone_reclaim(zone, gfp_mask, order)) zone 1414 mm/page_alloc.c page = buffered_rmqueue(preferred_zone, zone, order, gfp_mask); zone 1447 mm/page_alloc.c struct zone *zone; zone 1464 mm/page_alloc.c if (unlikely(!z->zone)) { zone 1488 mm/page_alloc.c for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) zone 1489 mm/page_alloc.c wakeup_kswapd(zone, order); zone 1763 mm/page_alloc.c struct zone *zone; zone 1770 mm/page_alloc.c for_each_zone_zonelist(zone, z, zonelist, offset) { zone 1771 mm/page_alloc.c unsigned long size = zone->present_pages; zone 1772 mm/page_alloc.c unsigned long high = zone->pages_high; zone 1800 mm/page_alloc.c printk("Node %d ", zone_to_nid(zone)); zone 1845 mm/page_alloc.c struct zone *zone; zone 1847 mm/page_alloc.c for_each_zone(zone) { zone 1848 mm/page_alloc.c if (!populated_zone(zone)) zone 1851 mm/page_alloc.c show_node(zone); zone 1852 mm/page_alloc.c printk("%s per-cpu:\n", zone->name); zone 1857 mm/page_alloc.c pageset = zone_pcp(zone, cpu); zone 1879 mm/page_alloc.c for_each_zone(zone) { zone 1882 mm/page_alloc.c if (!populated_zone(zone)) zone 1885 mm/page_alloc.c show_node(zone); zone 1897 mm/page_alloc.c zone->name, zone 1898 mm/page_alloc.c K(zone_page_state(zone, NR_FREE_PAGES)), zone 1899 mm/page_alloc.c K(zone->pages_min), zone 1900 mm/page_alloc.c K(zone->pages_low), zone 1901 mm/page_alloc.c K(zone->pages_high), zone 1902 mm/page_alloc.c K(zone_page_state(zone, NR_ACTIVE)), zone 1903 mm/page_alloc.c K(zone_page_state(zone, NR_INACTIVE)), zone 1904 mm/page_alloc.c K(zone->present_pages), zone 1905 mm/page_alloc.c zone->pages_scanned, zone 1906 mm/page_alloc.c (zone_is_all_unreclaimable(zone) ? "yes" : "no") zone 1910 mm/page_alloc.c printk(" %lu", zone->lowmem_reserve[i]); zone 1914 mm/page_alloc.c for_each_zone(zone) { zone 1917 mm/page_alloc.c if (!populated_zone(zone)) zone 1920 mm/page_alloc.c show_node(zone); zone 1921 mm/page_alloc.c printk("%s: ", zone->name); zone 1923 mm/page_alloc.c spin_lock_irqsave(&zone->lock, flags); zone 1925 mm/page_alloc.c nr[order] = zone->free_area[order].nr_free; zone 1928 mm/page_alloc.c spin_unlock_irqrestore(&zone->lock, flags); zone 1941 mm/page_alloc.c zoneref->zone = zone; zone 1942 mm/page_alloc.c zoneref->zone_idx = zone_idx(zone); zone 1953 mm/page_alloc.c struct zone *zone; zone 1960 mm/page_alloc.c zone = pgdat->node_zones + zone_type; zone 1961 mm/page_alloc.c if (populated_zone(zone)) { zone 1962 mm/page_alloc.c zoneref_set_zone(zone, zone 2139 mm/page_alloc.c for (j = 0; zonelist->_zonerefs[j].zone != NULL; j++) zone 2143 mm/page_alloc.c zonelist->_zonerefs[j].zone = NULL; zone 2157 mm/page_alloc.c zonelist->_zonerefs[j].zone = NULL; zone 2173 mm/page_alloc.c struct zone *z; zone 2189 mm/page_alloc.c zonelist->_zonerefs[pos].zone = NULL; zone 2197 mm/page_alloc.c struct zone *z; zone 2267 mm/page_alloc.c zonelist->_zonerefs[0].zone = NULL; zone 2325 mm/page_alloc.c for (z = zonelist->_zonerefs; z->zone; z++) zone 2369 mm/page_alloc.c zonelist->_zonerefs[j].zone = NULL; zone 2515 mm/page_alloc.c start_pfn = zone->zone_start_pfn; zone 2516 mm/page_alloc.c end_pfn = start_pfn + zone->spanned_pages; zone 2517 mm/page_alloc.c reserve = roundup(zone->pages_min, pageblock_nr_pages) >> zone 2526 mm/page_alloc.c if (page_to_nid(page) != zone_to_nid(zone)) zone 2544 mm/page_alloc.c move_freepages_block(zone, page, MIGRATE_RESERVE); zone 2555 mm/page_alloc.c move_freepages_block(zone, page, MIGRATE_MOVABLE); zone 2571 mm/page_alloc.c struct zone *z; zone 2573 mm/page_alloc.c z = &NODE_DATA(nid)->node_zones[zone]; zone 2587 mm/page_alloc.c set_page_links(page, zone, nid, pfn); zone 2588 mm/page_alloc.c mminit_verify_page_links(page, zone, nid, pfn); zone 2614 mm/page_alloc.c if (!is_highmem_idx(zone)) zone 2624 mm/page_alloc.c INIT_LIST_HEAD(&zone->free_area[order].free_list[t]); zone 2625 mm/page_alloc.c zone->free_area[order].nr_free = 0; zone 2631 mm/page_alloc.c memmap_init_zone((size), (nid), (zone), (start_pfn), MEMMAP_EARLY) zone 2644 mm/page_alloc.c batch = zone->present_pages / 1024; zone 2723 mm/page_alloc.c struct zone *zone, *dzone; zone 2728 mm/page_alloc.c for_each_zone(zone) { zone 2730 mm/page_alloc.c if (!populated_zone(zone)) zone 2733 mm/page_alloc.c zone_pcp(zone, cpu) = kmalloc_node(sizeof(struct per_cpu_pageset), zone 2735 mm/page_alloc.c if (!zone_pcp(zone, cpu)) zone 2738 mm/page_alloc.c setup_pageset(zone_pcp(zone, cpu), zone_batchsize(zone)); zone 2741 mm/page_alloc.c setup_pagelist_highmark(zone_pcp(zone, cpu), zone 2742 mm/page_alloc.c (zone->present_pages / percpu_pagelist_fraction)); zone 2750 mm/page_alloc.c if (dzone == zone) zone 2760 mm/page_alloc.c struct zone *zone; zone 2762 mm/page_alloc.c for_each_zone(zone) { zone 2763 mm/page_alloc.c struct per_cpu_pageset *pset = zone_pcp(zone, cpu); zone 2768 mm/page_alloc.c zone_pcp(zone, cpu) = NULL; zone 2819 mm/page_alloc.c struct pglist_data *pgdat = zone->zone_pgdat; zone 2826 mm/page_alloc.c zone->wait_table_hash_nr_entries = zone 2828 mm/page_alloc.c zone->wait_table_bits = zone 2829 mm/page_alloc.c wait_table_bits(zone->wait_table_hash_nr_entries); zone 2830 mm/page_alloc.c alloc_size = zone->wait_table_hash_nr_entries zone 2834 mm/page_alloc.c zone->wait_table = (wait_queue_head_t *) zone 2847 mm/page_alloc.c zone->wait_table = vmalloc(alloc_size); zone 2849 mm/page_alloc.c if (!zone->wait_table) zone 2852 mm/page_alloc.c for(i = 0; i < zone->wait_table_hash_nr_entries; ++i) zone 2853 mm/page_alloc.c init_waitqueue_head(zone->wait_table + i); zone 2861 mm/page_alloc.c unsigned long batch = zone_batchsize(zone); zone 2866 mm/page_alloc.c zone_pcp(zone, cpu) = &boot_pageset[cpu]; zone 2869 mm/page_alloc.c setup_pageset(zone_pcp(zone,cpu), batch); zone 2872 mm/page_alloc.c if (zone->present_pages) zone 2874 mm/page_alloc.c zone->name, zone->present_pages, batch); zone 2882 mm/page_alloc.c struct pglist_data *pgdat = zone->zone_pgdat; zone 2884 mm/page_alloc.c ret = zone_wait_table_init(zone, size); zone 2887 mm/page_alloc.c pgdat->nr_zones = zone_idx(zone) + 1; zone 2889 mm/page_alloc.c zone->zone_start_pfn = zone_start_pfn; zone 2894 mm/page_alloc.c (unsigned long)zone_idx(zone), zone 2897 mm/page_alloc.c zone_init_free_lists(zone); zone 3344 mm/page_alloc.c zone->pageblock_flags = NULL; zone 3346 mm/page_alloc.c zone->pageblock_flags = alloc_bootmem_node(pgdat, usemapsize); zone 3347 mm/page_alloc.c memset(zone->pageblock_flags, 0, usemapsize); zone 3415 mm/page_alloc.c struct zone *zone = pgdat->node_zones + j; zone 3451 mm/page_alloc.c zone->spanned_pages = size; zone 3452 mm/page_alloc.c zone->present_pages = realsize; zone 3454 mm/page_alloc.c zone->node = nid; zone 3455 mm/page_alloc.c zone->min_unmapped_pages = (realsize*sysctl_min_unmapped_ratio) zone 3457 mm/page_alloc.c zone->min_slab_pages = (realsize * sysctl_min_slab_ratio) / 100; zone 3459 mm/page_alloc.c zone->name = zone_names[j]; zone 3460 mm/page_alloc.c spin_lock_init(&zone->lock); zone 3461 mm/page_alloc.c spin_lock_init(&zone->lru_lock); zone 3462 mm/page_alloc.c zone_seqlock_init(zone); zone 3463 mm/page_alloc.c zone->zone_pgdat = pgdat; zone 3465 mm/page_alloc.c zone->prev_priority = DEF_PRIORITY; zone 3467 mm/page_alloc.c zone_pcp_init(zone); zone 3468 mm/page_alloc.c INIT_LIST_HEAD(&zone->active_list); zone 3469 mm/page_alloc.c INIT_LIST_HEAD(&zone->inactive_list); zone 3470 mm/page_alloc.c zone->nr_scan_active = 0; zone 3471 mm/page_alloc.c zone->nr_scan_inactive = 0; zone 3472 mm/page_alloc.c zap_zone_vm_stats(zone); zone 3473 mm/page_alloc.c zone->flags = 0; zone 3478 mm/page_alloc.c setup_usemap(pgdat, zone, size); zone 3479 mm/page_alloc.c ret = init_currently_empty_zone(zone, zone_start_pfn, zone 3932 mm/page_alloc.c struct zone *zone = &pgdat->node_zones[zone_type]; zone 3933 mm/page_alloc.c if (zone->present_pages) zone 3934 mm/page_alloc.c node_set_state(zone_to_nid(zone), N_NORMAL_MEMORY); zone 4132 mm/page_alloc.c struct zone *zone = pgdat->node_zones + i; zone 4137 mm/page_alloc.c if (zone->lowmem_reserve[j] > max) zone 4138 mm/page_alloc.c max = zone->lowmem_reserve[j]; zone 4142 mm/page_alloc.c max += zone->pages_high; zone 4144 mm/page_alloc.c if (max > zone->present_pages) zone 4145 mm/page_alloc.c max = zone->present_pages; zone 4165 mm/page_alloc.c struct zone *zone = pgdat->node_zones + j; zone 4166 mm/page_alloc.c unsigned long present_pages = zone->present_pages; zone 4168 mm/page_alloc.c zone->lowmem_reserve[j] = 0; zone 4172 mm/page_alloc.c struct zone *lower_zone; zone 4201 mm/page_alloc.c struct zone *zone; zone 4205 mm/page_alloc.c for_each_zone(zone) { zone 4206 mm/page_alloc.c if (!is_highmem(zone)) zone 4207 mm/page_alloc.c lowmem_pages += zone->present_pages; zone 4210 mm/page_alloc.c for_each_zone(zone) { zone 4213 mm/page_alloc.c spin_lock_irqsave(&zone->lru_lock, flags); zone 4214 mm/page_alloc.c tmp = (u64)pages_min * zone->present_pages; zone 4216 mm/page_alloc.c if (is_highmem(zone)) { zone 4228 mm/page_alloc.c min_pages = zone->present_pages / 1024; zone 4233 mm/page_alloc.c zone->pages_min = min_pages; zone 4239 mm/page_alloc.c zone->pages_min = tmp; zone 4242 mm/page_alloc.c zone->pages_low = zone->pages_min + (tmp >> 2); zone 4243 mm/page_alloc.c zone->pages_high = zone->pages_min + (tmp >> 1); zone 4244 mm/page_alloc.c setup_zone_migrate_reserve(zone); zone 4245 mm/page_alloc.c spin_unlock_irqrestore(&zone->lru_lock, flags); zone 4311 mm/page_alloc.c struct zone *zone; zone 4318 mm/page_alloc.c for_each_zone(zone) zone 4319 mm/page_alloc.c zone->min_unmapped_pages = (zone->present_pages * zone 4327 mm/page_alloc.c struct zone *zone; zone 4334 mm/page_alloc.c for_each_zone(zone) zone 4335 mm/page_alloc.c zone->min_slab_pages = (zone->present_pages * zone 4367 mm/page_alloc.c struct zone *zone; zone 4374 mm/page_alloc.c for_each_zone(zone) { zone 4377 mm/page_alloc.c high = zone->present_pages / percpu_pagelist_fraction; zone 4378 mm/page_alloc.c setup_pagelist_highmark(zone_pcp(zone, cpu), high); zone 4511 mm/page_alloc.c return zone->pageblock_flags; zone 4521 mm/page_alloc.c pfn = pfn - zone->zone_start_pfn; zone 4536 mm/page_alloc.c struct zone *zone; zone 4542 mm/page_alloc.c zone = page_zone(page); zone 4544 mm/page_alloc.c bitmap = get_pageblock_bitmap(zone, pfn); zone 4545 mm/page_alloc.c bitidx = pfn_to_bitidx(zone, pfn); zone 4564 mm/page_alloc.c struct zone *zone; zone 4569 mm/page_alloc.c zone = page_zone(page); zone 4571 mm/page_alloc.c bitmap = get_pageblock_bitmap(zone, pfn); zone 4572 mm/page_alloc.c bitidx = pfn_to_bitidx(zone, pfn); zone 4573 mm/page_alloc.c VM_BUG_ON(pfn < zone->zone_start_pfn); zone 4574 mm/page_alloc.c VM_BUG_ON(pfn >= zone->zone_start_pfn + zone->spanned_pages); zone 4591 mm/page_alloc.c struct zone *zone; zone 4595 mm/page_alloc.c zone = page_zone(page); zone 4596 mm/page_alloc.c spin_lock_irqsave(&zone->lock, flags); zone 4603 mm/page_alloc.c move_freepages_block(zone, page, MIGRATE_ISOLATE); zone 4606 mm/page_alloc.c spin_unlock_irqrestore(&zone->lock, flags); zone 4614 mm/page_alloc.c struct zone *zone; zone 4616 mm/page_alloc.c zone = page_zone(page); zone 4617 mm/page_alloc.c spin_lock_irqsave(&zone->lock, flags); zone 4621 mm/page_alloc.c move_freepages_block(zone, page, MIGRATE_MOVABLE); zone 4623 mm/page_alloc.c spin_unlock_irqrestore(&zone->lock, flags); zone 4634 mm/page_alloc.c struct zone *zone; zone 4644 mm/page_alloc.c zone = page_zone(pfn_to_page(pfn)); zone 4645 mm/page_alloc.c spin_lock_irqsave(&zone->lock, flags); zone 4662 mm/page_alloc.c zone->free_area[order].nr_free--; zone 4663 mm/page_alloc.c __mod_zone_page_state(zone, NR_FREE_PAGES, zone 4669 mm/page_alloc.c spin_unlock_irqrestore(&zone->lock, flags); zone 119 mm/page_isolation.c struct zone *zone; zone 136 mm/page_isolation.c zone = page_zone(pfn_to_page(pfn)); zone 137 mm/page_isolation.c spin_lock_irqsave(&zone->lock, flags); zone 139 mm/page_isolation.c spin_unlock_irqrestore(&zone->lock, flags); zone 30 mm/quicklist.c struct zone *zones = NODE_DATA(node)->node_zones; zone 3236 mm/slab.c struct zone *zone; zone 3252 mm/slab.c for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { zone 3253 mm/slab.c nid = zone_to_nid(zone); zone 3255 mm/slab.c if (cpuset_zone_allowed_hardwall(zone, flags) && zone 1299 mm/slub.c struct zone *zone; zone 1326 mm/slub.c for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { zone 1329 mm/slub.c n = get_node(s, zone_to_nid(zone)); zone 1331 mm/slub.c if (n && cpuset_zone_allowed_hardwall(zone, flags) && zone 577 mm/sparse.c struct pglist_data *pgdat = zone->zone_pgdat; zone 49 mm/swap.c struct zone *zone = page_zone(page); zone 51 mm/swap.c spin_lock_irqsave(&zone->lru_lock, flags); zone 54 mm/swap.c del_page_from_lru(zone, page); zone 55 mm/swap.c spin_unlock_irqrestore(&zone->lru_lock, flags); zone 107 mm/swap.c struct zone *zone = NULL; zone 111 mm/swap.c struct zone *pagezone = page_zone(page); zone 113 mm/swap.c if (pagezone != zone) { zone 114 mm/swap.c if (zone) zone 115 mm/swap.c spin_unlock(&zone->lru_lock); zone 116 mm/swap.c zone = pagezone; zone 117 mm/swap.c spin_lock(&zone->lru_lock); zone 120 mm/swap.c list_move_tail(&page->lru, &zone->inactive_list); zone 124 mm/swap.c if (zone) zone 125 mm/swap.c spin_unlock(&zone->lru_lock); zone 157 mm/swap.c struct zone *zone = page_zone(page); zone 159 mm/swap.c spin_lock_irq(&zone->lru_lock); zone 161 mm/swap.c del_page_from_inactive_list(zone, page); zone 163 mm/swap.c add_page_to_active_list(zone, page); zone 167 mm/swap.c spin_unlock_irq(&zone->lru_lock); zone 290 mm/swap.c struct zone *zone = NULL; zone 298 mm/swap.c if (zone) { zone 299 mm/swap.c spin_unlock_irqrestore(&zone->lru_lock, flags); zone 300 mm/swap.c zone = NULL; zone 310 mm/swap.c struct zone *pagezone = page_zone(page); zone 311 mm/swap.c if (pagezone != zone) { zone 312 mm/swap.c if (zone) zone 313 mm/swap.c spin_unlock_irqrestore(&zone->lru_lock, zone 315 mm/swap.c zone = pagezone; zone 316 mm/swap.c spin_lock_irqsave(&zone->lru_lock, flags); zone 320 mm/swap.c del_page_from_lru(zone, page); zone 324 mm/swap.c if (zone) { zone 325 mm/swap.c spin_unlock_irqrestore(&zone->lru_lock, flags); zone 326 mm/swap.c zone = NULL; zone 332 mm/swap.c if (zone) zone 333 mm/swap.c spin_unlock_irqrestore(&zone->lru_lock, flags); zone 386 mm/swap.c struct zone *zone = NULL; zone 390 mm/swap.c struct zone *pagezone = page_zone(page); zone 392 mm/swap.c if (pagezone != zone) { zone 393 mm/swap.c if (zone) zone 394 mm/swap.c spin_unlock_irq(&zone->lru_lock); zone 395 mm/swap.c zone = pagezone; zone 396 mm/swap.c spin_lock_irq(&zone->lru_lock); zone 400 mm/swap.c add_page_to_inactive_list(zone, page); zone 402 mm/swap.c if (zone) zone 403 mm/swap.c spin_unlock_irq(&zone->lru_lock); zone 413 mm/swap.c struct zone *zone = NULL; zone 417 mm/swap.c struct zone *pagezone = page_zone(page); zone 419 mm/swap.c if (pagezone != zone) { zone 420 mm/swap.c if (zone) zone 421 mm/swap.c spin_unlock_irq(&zone->lru_lock); zone 422 mm/swap.c zone = pagezone; zone 423 mm/swap.c spin_lock_irq(&zone->lru_lock); zone 429 mm/swap.c add_page_to_active_list(zone, page); zone 431 mm/swap.c if (zone) zone 432 mm/swap.c spin_unlock_irq(&zone->lru_lock); zone 80 mm/vmscan.c struct zone *z, struct mem_cgroup *mem_cont, zone 862 mm/vmscan.c spin_lock_irq(&zone->lru_lock); zone 874 mm/vmscan.c zone, sc->mem_cgroup, 0); zone 878 mm/vmscan.c __mod_zone_page_state(zone, NR_ACTIVE, -nr_active); zone 879 mm/vmscan.c __mod_zone_page_state(zone, NR_INACTIVE, zone 882 mm/vmscan.c zone->pages_scanned += nr_scan; zone 883 mm/vmscan.c spin_unlock_irq(&zone->lru_lock); zone 912 mm/vmscan.c __count_zone_vm_events(PGSCAN_KSWAPD, zone, nr_scan); zone 915 mm/vmscan.c __count_zone_vm_events(PGSCAN_DIRECT, zone, nr_scan); zone 917 mm/vmscan.c __count_zone_vm_events(PGSTEAL, zone, nr_freed); zone 922 mm/vmscan.c spin_lock(&zone->lru_lock); zone 932 mm/vmscan.c add_page_to_active_list(zone, page); zone 934 mm/vmscan.c add_page_to_inactive_list(zone, page); zone 936 mm/vmscan.c spin_unlock_irq(&zone->lru_lock); zone 938 mm/vmscan.c spin_lock_irq(&zone->lru_lock); zone 942 mm/vmscan.c spin_unlock(&zone->lru_lock); zone 959 mm/vmscan.c if (priority < zone->prev_priority) zone 960 mm/vmscan.c zone->prev_priority = priority; zone 965 mm/vmscan.c return zone->pages_scanned >= (zone_page_state(zone, NR_ACTIVE) zone 966 mm/vmscan.c + zone_page_state(zone, NR_INACTIVE))*3; zone 983 mm/vmscan.c if (scan_global_lru(sc) && zone_is_near_oom(zone)) zone 990 mm/vmscan.c prev_priority = zone->prev_priority; zone 1036 mm/vmscan.c imbalance = zone_page_state(zone, NR_ACTIVE); zone 1037 mm/vmscan.c imbalance /= zone_page_state(zone, NR_INACTIVE) + 1; zone 1109 mm/vmscan.c reclaim_mapped = calc_reclaim_mapped(sc, zone, priority); zone 1112 mm/vmscan.c spin_lock_irq(&zone->lru_lock); zone 1114 mm/vmscan.c ISOLATE_ACTIVE, zone, zone 1121 mm/vmscan.c zone->pages_scanned += pgscanned; zone 1123 mm/vmscan.c __mod_zone_page_state(zone, NR_ACTIVE, -pgmoved); zone 1124 mm/vmscan.c spin_unlock_irq(&zone->lru_lock); zone 1143 mm/vmscan.c spin_lock_irq(&zone->lru_lock); zone 1152 mm/vmscan.c list_move(&page->lru, &zone->inactive_list); zone 1156 mm/vmscan.c __mod_zone_page_state(zone, NR_INACTIVE, pgmoved); zone 1157 mm/vmscan.c spin_unlock_irq(&zone->lru_lock); zone 1163 mm/vmscan.c spin_lock_irq(&zone->lru_lock); zone 1166 mm/vmscan.c __mod_zone_page_state(zone, NR_INACTIVE, pgmoved); zone 1169 mm/vmscan.c spin_unlock_irq(&zone->lru_lock); zone 1171 mm/vmscan.c spin_lock_irq(&zone->lru_lock); zone 1182 mm/vmscan.c list_move(&page->lru, &zone->active_list); zone 1186 mm/vmscan.c __mod_zone_page_state(zone, NR_ACTIVE, pgmoved); zone 1188 mm/vmscan.c spin_unlock_irq(&zone->lru_lock); zone 1190 mm/vmscan.c spin_lock_irq(&zone->lru_lock); zone 1193 mm/vmscan.c __mod_zone_page_state(zone, NR_ACTIVE, pgmoved); zone 1195 mm/vmscan.c __count_zone_vm_events(PGREFILL, zone, pgscanned); zone 1197 mm/vmscan.c spin_unlock_irq(&zone->lru_lock); zone 1218 mm/vmscan.c zone->nr_scan_active += zone 1219 mm/vmscan.c (zone_page_state(zone, NR_ACTIVE) >> priority) + 1; zone 1220 mm/vmscan.c nr_active = zone->nr_scan_active; zone 1221 mm/vmscan.c zone->nr_scan_inactive += zone 1222 mm/vmscan.c (zone_page_state(zone, NR_INACTIVE) >> priority) + 1; zone 1223 mm/vmscan.c nr_inactive = zone->nr_scan_inactive; zone 1225 mm/vmscan.c zone->nr_scan_inactive = 0; zone 1230 mm/vmscan.c zone->nr_scan_active = 0; zone 1240 mm/vmscan.c zone, priority); zone 1243 mm/vmscan.c zone, priority); zone 1252 mm/vmscan.c shrink_active_list(nr_to_scan, zone, sc, priority); zone 1259 mm/vmscan.c nr_reclaimed += shrink_inactive_list(nr_to_scan, zone, zone 1290 mm/vmscan.c struct zone *zone; zone 1293 mm/vmscan.c for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { zone 1294 mm/vmscan.c if (!populated_zone(zone)) zone 1301 mm/vmscan.c if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL)) zone 1303 mm/vmscan.c note_zone_scanning_priority(zone, priority); zone 1305 mm/vmscan.c if (zone_is_all_unreclaimable(zone) && zone 1319 mm/vmscan.c nr_reclaimed += shrink_zone(priority, zone, sc); zone 1351 mm/vmscan.c struct zone *zone; zone 1362 mm/vmscan.c for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { zone 1364 mm/vmscan.c if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL)) zone 1367 mm/vmscan.c lru_pages += zone_page_state(zone, NR_ACTIVE) zone 1368 mm/vmscan.c + zone_page_state(zone, NR_INACTIVE); zone 1426 mm/vmscan.c for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { zone 1428 mm/vmscan.c if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL)) zone 1431 mm/vmscan.c zone->prev_priority = priority; zone 1549 mm/vmscan.c struct zone *zone = pgdat->node_zones + i; zone 1551 mm/vmscan.c if (!populated_zone(zone)) zone 1554 mm/vmscan.c if (zone_is_all_unreclaimable(zone) && zone 1558 mm/vmscan.c if (!zone_watermark_ok(zone, order, zone->pages_high, zone 1568 mm/vmscan.c struct zone *zone = pgdat->node_zones + i; zone 1570 mm/vmscan.c lru_pages += zone_page_state(zone, NR_ACTIVE) zone 1571 mm/vmscan.c + zone_page_state(zone, NR_INACTIVE); zone 1584 mm/vmscan.c struct zone *zone = pgdat->node_zones + i; zone 1587 mm/vmscan.c if (!populated_zone(zone)) zone 1590 mm/vmscan.c if (zone_is_all_unreclaimable(zone) && zone 1594 mm/vmscan.c if (!zone_watermark_ok(zone, order, zone->pages_high, zone 1599 mm/vmscan.c note_zone_scanning_priority(zone, priority); zone 1604 mm/vmscan.c if (!zone_watermark_ok(zone, order, 8*zone->pages_high, zone 1606 mm/vmscan.c nr_reclaimed += shrink_zone(priority, zone, &sc); zone 1612 mm/vmscan.c if (zone_is_all_unreclaimable(zone)) zone 1614 mm/vmscan.c if (nr_slab == 0 && zone->pages_scanned >= zone 1615 mm/vmscan.c (zone_page_state(zone, NR_ACTIVE) zone 1616 mm/vmscan.c + zone_page_state(zone, NR_INACTIVE)) * 6) zone 1617 mm/vmscan.c zone_set_flag(zone, zone 1653 mm/vmscan.c struct zone *zone = pgdat->node_zones + i; zone 1655 mm/vmscan.c zone->prev_priority = temp_priority[i]; zone 1749 mm/vmscan.c if (!populated_zone(zone)) zone 1752 mm/vmscan.c pgdat = zone->zone_pgdat; zone 1753 mm/vmscan.c if (zone_watermark_ok(zone, order, zone->pages_low, 0, 0)) zone 1757 mm/vmscan.c if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL)) zone 1775 mm/vmscan.c struct zone *zone; zone 1778 mm/vmscan.c for_each_zone(zone) { zone 1780 mm/vmscan.c if (!populated_zone(zone)) zone 1783 mm/vmscan.c if (zone_is_all_unreclaimable(zone) && prio != DEF_PRIORITY) zone 1788 mm/vmscan.c zone->nr_scan_active += zone 1789 mm/vmscan.c (zone_page_state(zone, NR_ACTIVE) >> prio) + 1; zone 1790 mm/vmscan.c if (zone->nr_scan_active >= nr_pages || pass > 3) { zone 1791 mm/vmscan.c zone->nr_scan_active = 0; zone 1793 mm/vmscan.c zone_page_state(zone, NR_ACTIVE)); zone 1794 mm/vmscan.c shrink_active_list(nr_to_scan, zone, sc, prio); zone 1798 mm/vmscan.c zone->nr_scan_inactive += zone 1799 mm/vmscan.c (zone_page_state(zone, NR_INACTIVE) >> prio) + 1; zone 1800 mm/vmscan.c if (zone->nr_scan_inactive >= nr_pages || pass > 3) { zone 1801 mm/vmscan.c zone->nr_scan_inactive = 0; zone 1803 mm/vmscan.c zone_page_state(zone, NR_INACTIVE)); zone 1804 mm/vmscan.c ret += shrink_inactive_list(nr_to_scan, zone, sc); zone 2038 mm/vmscan.c if (zone_page_state(zone, NR_FILE_PAGES) - zone 2039 mm/vmscan.c zone_page_state(zone, NR_FILE_MAPPED) > zone 2040 mm/vmscan.c zone->min_unmapped_pages) { zone 2047 mm/vmscan.c note_zone_scanning_priority(zone, priority); zone 2048 mm/vmscan.c nr_reclaimed += shrink_zone(priority, zone, &sc); zone 2053 mm/vmscan.c slab_reclaimable = zone_page_state(zone, NR_SLAB_RECLAIMABLE); zone 2054 mm/vmscan.c if (slab_reclaimable > zone->min_slab_pages) { zone 2066 mm/vmscan.c zone_page_state(zone, NR_SLAB_RECLAIMABLE) > zone 2075 mm/vmscan.c zone_page_state(zone, NR_SLAB_RECLAIMABLE); zone 2098 mm/vmscan.c if (zone_page_state(zone, NR_FILE_PAGES) - zone 2099 mm/vmscan.c zone_page_state(zone, NR_FILE_MAPPED) <= zone->min_unmapped_pages zone 2100 mm/vmscan.c && zone_page_state(zone, NR_SLAB_RECLAIMABLE) zone 2101 mm/vmscan.c <= zone->min_slab_pages) zone 2104 mm/vmscan.c if (zone_is_all_unreclaimable(zone)) zone 2119 mm/vmscan.c node_id = zone_to_nid(zone); zone 2123 mm/vmscan.c if (zone_test_and_set_flag(zone, ZONE_RECLAIM_LOCKED)) zone 2125 mm/vmscan.c ret = __zone_reclaim(zone, gfp_mask, order); zone 2126 mm/vmscan.c zone_clear_flag(zone, ZONE_RECLAIM_LOCKED); zone 117 mm/vmstat.c mem = zone->present_pages >> (27 - PAGE_SHIFT); zone 134 mm/vmstat.c struct zone *zone; zone 138 mm/vmstat.c for_each_zone(zone) { zone 140 mm/vmstat.c if (!zone->present_pages) zone 143 mm/vmstat.c threshold = calculate_threshold(zone); zone 146 mm/vmstat.c zone_pcp(zone, cpu)->stat_threshold = threshold; zone 156 mm/vmstat.c struct per_cpu_pageset *pcp = zone_pcp(zone, smp_processor_id()); zone 163 mm/vmstat.c zone_page_state_add(x, zone, item); zone 179 mm/vmstat.c __mod_zone_page_state(zone, item, delta); zone 209 mm/vmstat.c struct per_cpu_pageset *pcp = zone_pcp(zone, smp_processor_id()); zone 217 mm/vmstat.c zone_page_state_add(*p + overstep, zone, item); zone 230 mm/vmstat.c struct per_cpu_pageset *pcp = zone_pcp(zone, smp_processor_id()); zone 238 mm/vmstat.c zone_page_state_add(*p - overstep, zone, item); zone 254 mm/vmstat.c __inc_zone_state(zone, item); zone 261 mm/vmstat.c struct zone *zone; zone 263 mm/vmstat.c zone = page_zone(page); zone 265 mm/vmstat.c __inc_zone_state(zone, item); zone 300 mm/vmstat.c struct zone *zone; zone 304 mm/vmstat.c for_each_zone(zone) { zone 307 mm/vmstat.c if (!populated_zone(zone)) zone 310 mm/vmstat.c p = zone_pcp(zone, cpu); zone 321 mm/vmstat.c atomic_long_add(v, &zone->vm_stat[i]); zone 343 mm/vmstat.c if (zone_to_nid(zone) == numa_node_id()) { zone 353 mm/vmstat.c drain_zone_pages(zone, &p->pcp); zone 426 mm/vmstat.c struct zone *zone; zone 427 mm/vmstat.c struct zone *node_zones = pgdat->node_zones; zone 430 mm/vmstat.c for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { zone 431 mm/vmstat.c if (!populated_zone(zone)) zone 434 mm/vmstat.c spin_lock_irqsave(&zone->lock, flags); zone 435 mm/vmstat.c print(m, pgdat, zone); zone 436 mm/vmstat.c spin_unlock_irqrestore(&zone->lock, flags); zone 445 mm/vmstat.c seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); zone 447 mm/vmstat.c seq_printf(m, "%6lu ", zone->free_area[order].nr_free); zone 469 mm/vmstat.c zone->name, zone 476 mm/vmstat.c area = &(zone->free_area[order]); zone 508 mm/vmstat.c unsigned long start_pfn = zone->zone_start_pfn; zone 509 mm/vmstat.c unsigned long end_pfn = start_pfn + zone->spanned_pages; zone 532 mm/vmstat.c if (page_zone(page) != zone) zone 542 mm/vmstat.c seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); zone 685 mm/vmstat.c seq_printf(m, "Node %d, zone %8s", pgdat->node_id, zone->name); zone 694 mm/vmstat.c zone_page_state(zone, NR_FREE_PAGES), zone 695 mm/vmstat.c zone->pages_min, zone 696 mm/vmstat.c zone->pages_low, zone 697 mm/vmstat.c zone->pages_high, zone 698 mm/vmstat.c zone->pages_scanned, zone 699 mm/vmstat.c zone->nr_scan_active, zone->nr_scan_inactive, zone 700 mm/vmstat.c zone->spanned_pages, zone 701 mm/vmstat.c zone->present_pages); zone 705 mm/vmstat.c zone_page_state(zone, i)); zone 709 mm/vmstat.c zone->lowmem_reserve[0]); zone 710 mm/vmstat.c for (i = 1; i < ARRAY_SIZE(zone->lowmem_reserve); i++) zone 711 mm/vmstat.c seq_printf(m, ", %lu", zone->lowmem_reserve[i]); zone 718 mm/vmstat.c pageset = zone_pcp(zone, i); zone 737 mm/vmstat.c zone_is_all_unreclaimable(zone), zone 738 mm/vmstat.c zone->prev_priority, zone 739 mm/vmstat.c zone->zone_start_pfn); zone 806 net/ipv4/fib_hash.c struct fn_zone *zone; zone 833 net/ipv4/fib_hash.c for (iter->zone = table->fn_zone_list; iter->zone; zone 834 net/ipv4/fib_hash.c iter->zone = iter->zone->fz_next) { zone 837 net/ipv4/fib_hash.c if (!iter->zone->fz_nent) zone 840 net/ipv4/fib_hash.c iter->hash_head = iter->zone->fz_hash; zone 841 net/ipv4/fib_hash.c maxslot = iter->zone->fz_divisor; zone 898 net/ipv4/fib_hash.c if (!iter->zone) zone 905 net/ipv4/fib_hash.c maxslot = iter->zone->fz_divisor; zone 919 net/ipv4/fib_hash.c iter->zone = iter->zone->fz_next; zone 921 net/ipv4/fib_hash.c if (!iter->zone) zone 925 net/ipv4/fib_hash.c iter->hash_head = iter->zone->fz_hash; zone 1023 net/ipv4/fib_hash.c mask = FZ_MASK(iter->zone); zone 215 sound/synth/emux/emux_effect.c origp = (unsigned char*)&vp->zone->v.parm + offset; zone 114 sound/synth/emux/emux_synth.c vp->zone = table[i]; zone 115 sound/synth/emux/emux_synth.c if (vp->zone->sample) zone 116 sound/synth/emux/emux_synth.c vp->block = vp->zone->sample->block; zone 514 sound/synth/emux/emux_synth.c vp->zone = NULL; zone 567 sound/synth/emux/emux_synth.c vp->reg = vp->zone->v; zone 522 sound/synth/emux/soundfont.c struct snd_sf_zone *zone; zone 558 sound/synth/emux/soundfont.c for (zone = sf->zones; zone; zone = zone->next) { zone 559 sound/synth/emux/soundfont.c if (!zone->mapped && zone 560 sound/synth/emux/soundfont.c zone->bank == hdr.bank && zone 561 sound/synth/emux/soundfont.c zone->instr == hdr.instr) zone 590 sound/synth/emux/soundfont.c if ((zone = sf_zone_new(sflist, sf)) == NULL) { zone 595 sound/synth/emux/soundfont.c zone->bank = tmpzone.bank; zone 596 sound/synth/emux/soundfont.c zone->instr = tmpzone.instr; zone 597 sound/synth/emux/soundfont.c zone->v = tmpzone.v; zone 600 sound/synth/emux/soundfont.c zone->sample = set_sample(sf, &zone->v); zone 949 sound/synth/emux/soundfont.c struct snd_sf_zone *zone; zone 1005 sound/synth/emux/soundfont.c if ((zone = sf_zone_new(sflist, sf)) == NULL) { zone 1027 sound/synth/emux/soundfont.c zone->v.sample = sample_id; /* the last sample */ zone 1028 sound/synth/emux/soundfont.c zone->v.rate_offset = calc_rate_offset(patch.base_freq); zone 1030 sound/synth/emux/soundfont.c zone->v.root = note / 100; zone 1031 sound/synth/emux/soundfont.c zone->v.tune = -(note % 100); zone 1032 sound/synth/emux/soundfont.c zone->v.low = (freq_to_note(patch.low_note) + 99) / 100; zone 1033 sound/synth/emux/soundfont.c zone->v.high = freq_to_note(patch.high_note) / 100; zone 1035 sound/synth/emux/soundfont.c zone->v.pan = (patch.panning + 128) / 2; zone 1038 sound/synth/emux/soundfont.c (int)patch.base_freq, zone->v.rate_offset, zone 1039 sound/synth/emux/soundfont.c zone->v.root, zone->v.tune, zone->v.low, zone->v.high); zone 1063 sound/synth/emux/soundfont.c zone->v.parm.volatkhld = zone 1066 sound/synth/emux/soundfont.c zone->v.parm.voldcysus = (calc_gus_sustain(patch.env_offset[2]) << 8) | zone 1068 sound/synth/emux/soundfont.c zone->v.parm.volrelease = 0x8000 | snd_sf_calc_parm_decay(release); zone 1069 sound/synth/emux/soundfont.c zone->v.attenuation = calc_gus_attenuation(patch.env_offset[0]); zone 1072 sound/synth/emux/soundfont.c zone->v.parm.volatkhld, zone 1073 sound/synth/emux/soundfont.c zone->v.parm.voldcysus, zone 1074 sound/synth/emux/soundfont.c zone->v.parm.volrelease, zone 1075 sound/synth/emux/soundfont.c zone->v.attenuation); zone 1081 sound/synth/emux/soundfont.c zone->v.parm.volrelease = 0x807f; zone 1087 sound/synth/emux/soundfont.c zone->v.parm.tremfrq = ((patch.tremolo_depth / 2) << 8) | rate; zone 1092 sound/synth/emux/soundfont.c zone->v.parm.fm2frq2 = ((patch.vibrato_depth / 6) << 8) | rate; zone 1098 sound/synth/emux/soundfont.c zone->v.mode = SNDRV_SFNT_MODE_LOOPING; zone 1100 sound/synth/emux/soundfont.c zone->v.mode = 0; zone 1104 sound/synth/emux/soundfont.c zone->bank = 0; zone 1105 sound/synth/emux/soundfont.c zone->instr = patch.instr_no; zone 1106 sound/synth/emux/soundfont.c zone->mapped = 0; zone 1107 sound/synth/emux/soundfont.c zone->v.sf_id = sf->id; zone 1109 sound/synth/emux/soundfont.c zone->sample = set_sample(sf, &zone->v); zone 1112 sound/synth/emux/soundfont.c add_preset(sflist, zone); zone 1168 sound/synth/emux/soundfont.c struct snd_sf_zone *zone; zone 1171 sound/synth/emux/soundfont.c zone = search_first_zone(sflist, cur->bank, cur->instr, cur->v.low); zone 1172 sound/synth/emux/soundfont.c if (zone && zone->v.sf_id != cur->v.sf_id) { zone 1176 sound/synth/emux/soundfont.c for (p = zone; p; p = p->next_zone) { zone 1182 sound/synth/emux/soundfont.c delete_preset(sflist, zone); zone 1183 sound/synth/emux/soundfont.c zone = NULL; /* do not forget to clear this! */ zone 1189 sound/synth/emux/soundfont.c cur->next_zone = zone; /* zone link */ zone 22 virt/kvm/coalesced_mmio.c struct kvm_coalesced_mmio_zone *zone; zone 50 virt/kvm/coalesced_mmio.c zone = &dev->zone[i]; zone 56 virt/kvm/coalesced_mmio.c if (zone->addr <= addr && zone 57 virt/kvm/coalesced_mmio.c addr + len <= zone->addr + zone->size) zone 118 virt/kvm/coalesced_mmio.c dev->zone[dev->nb_zones] = *zone; zone 139 virt/kvm/coalesced_mmio.c z = &dev->zone[i - 1]; zone 145 virt/kvm/coalesced_mmio.c if (zone->addr <= z->addr && zone 146 virt/kvm/coalesced_mmio.c z->addr + z->size <= zone->addr + zone->size) { zone 148 virt/kvm/coalesced_mmio.c *z = dev->zone[dev->nb_zones]; zone 16 virt/kvm/coalesced_mmio.h struct kvm_coalesced_mmio_zone zone[KVM_COALESCED_MMIO_ZONE_MAX]; zone 1339 virt/kvm/kvm_main.c struct kvm_coalesced_mmio_zone zone; zone 1341 virt/kvm/kvm_main.c if (copy_from_user(&zone, argp, sizeof zone)) zone 1344 virt/kvm/kvm_main.c r = kvm_vm_ioctl_register_coalesced_mmio(kvm, &zone); zone 1351 virt/kvm/kvm_main.c struct kvm_coalesced_mmio_zone zone; zone 1353 virt/kvm/kvm_main.c if (copy_from_user(&zone, argp, sizeof zone)) zone 1356 virt/kvm/kvm_main.c r = kvm_vm_ioctl_unregister_coalesced_mmio(kvm, &zone);