kmem_cache_cpu    106 include/linux/slub_def.h 	struct kmem_cache_cpu *cpu_slab[NR_CPUS];
kmem_cache_cpu    108 include/linux/slub_def.h 	struct kmem_cache_cpu cpu_slab;
kmem_cache_cpu    229 mm/slub.c      static inline struct kmem_cache_cpu *get_cpu_slab(struct kmem_cache *s, int cpu)
kmem_cache_cpu   1367 mm/slub.c      	struct kmem_cache_cpu *c = get_cpu_slab(s, smp_processor_id());
kmem_cache_cpu   1452 mm/slub.c      	struct kmem_cache_cpu *c = get_cpu_slab(s, cpu);
kmem_cache_cpu   1589 mm/slub.c      	struct kmem_cache_cpu *c;
kmem_cache_cpu   1640 mm/slub.c      	struct kmem_cache_cpu *c;
kmem_cache_cpu   1709 mm/slub.c      	struct kmem_cache_cpu *c;
kmem_cache_cpu   1964 mm/slub.c      static struct kmem_cache_cpu *alloc_kmem_cache_cpu(struct kmem_cache *s,
kmem_cache_cpu   1967 mm/slub.c      	struct kmem_cache_cpu *c = per_cpu(kmem_cache_cpu_free, cpu);
kmem_cache_cpu   1975 mm/slub.c      			ALIGN(sizeof(struct kmem_cache_cpu), cache_line_size()),
kmem_cache_cpu   1987 mm/slub.c      	if (c < per_cpu(kmem_cache_cpu, cpu) ||
kmem_cache_cpu   1988 mm/slub.c      			c > per_cpu(kmem_cache_cpu, cpu) + NR_KMEM_CACHE_CPU) {
kmem_cache_cpu   2001 mm/slub.c      		struct kmem_cache_cpu *c = get_cpu_slab(s, cpu);
kmem_cache_cpu   2015 mm/slub.c      		struct kmem_cache_cpu *c = get_cpu_slab(s, cpu);
kmem_cache_cpu   2041 mm/slub.c      		free_kmem_cache_cpu(&per_cpu(kmem_cache_cpu, cpu)[i], cpu);
kmem_cache_cpu   3022 mm/slub.c      				nr_cpu_ids * sizeof(struct kmem_cache_cpu *);
kmem_cache_cpu   3180 mm/slub.c      			struct kmem_cache_cpu *c = get_cpu_slab(s, cpu);
kmem_cache_cpu   3677 mm/slub.c      			struct kmem_cache_cpu *c = get_cpu_slab(s, cpu);