kmem_cache_cpu 106 include/linux/slub_def.h struct kmem_cache_cpu *cpu_slab[NR_CPUS]; kmem_cache_cpu 108 include/linux/slub_def.h struct kmem_cache_cpu cpu_slab; kmem_cache_cpu 229 mm/slub.c static inline struct kmem_cache_cpu *get_cpu_slab(struct kmem_cache *s, int cpu) kmem_cache_cpu 1367 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, smp_processor_id()); kmem_cache_cpu 1452 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); kmem_cache_cpu 1589 mm/slub.c struct kmem_cache_cpu *c; kmem_cache_cpu 1640 mm/slub.c struct kmem_cache_cpu *c; kmem_cache_cpu 1709 mm/slub.c struct kmem_cache_cpu *c; kmem_cache_cpu 1964 mm/slub.c static struct kmem_cache_cpu *alloc_kmem_cache_cpu(struct kmem_cache *s, kmem_cache_cpu 1967 mm/slub.c struct kmem_cache_cpu *c = per_cpu(kmem_cache_cpu_free, cpu); kmem_cache_cpu 1975 mm/slub.c ALIGN(sizeof(struct kmem_cache_cpu), cache_line_size()), kmem_cache_cpu 1987 mm/slub.c if (c < per_cpu(kmem_cache_cpu, cpu) || kmem_cache_cpu 1988 mm/slub.c c > per_cpu(kmem_cache_cpu, cpu) + NR_KMEM_CACHE_CPU) { kmem_cache_cpu 2001 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); kmem_cache_cpu 2015 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); kmem_cache_cpu 2041 mm/slub.c free_kmem_cache_cpu(&per_cpu(kmem_cache_cpu, cpu)[i], cpu); kmem_cache_cpu 3022 mm/slub.c nr_cpu_ids * sizeof(struct kmem_cache_cpu *); kmem_cache_cpu 3180 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); kmem_cache_cpu 3677 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu);