per_cpu_offset    176 arch/x86/kernel/setup_percpu.c 		per_cpu_offset(cpu) = ptr - __per_cpu_start;
per_cpu_offset    233 arch/x86/kernel/setup_percpu.c 	else if (per_cpu_offset(cpu))
per_cpu_offset    316 arch/x86/kernel/setup_percpu.c 	if (!per_cpu_offset(cpu)) {
per_cpu_offset     34 include/asm-generic/percpu.h #define __my_cpu_offset per_cpu_offset(raw_smp_processor_id())
per_cpu_offset     37 include/asm-generic/percpu.h #define my_cpu_offset per_cpu_offset(smp_processor_id())
per_cpu_offset     57 include/asm-generic/percpu.h 	(*SHIFT_PERCPU_PTR(&per_cpu_var(var), per_cpu_offset(cpu)))
per_cpu_offset    641 kernel/lockdep.c 		start = (unsigned long) &__per_cpu_start + per_cpu_offset(i);
per_cpu_offset    643 kernel/lockdep.c 					+ per_cpu_offset(i);
per_cpu_offset    467 kernel/module.c 		memcpy(pcpudest + per_cpu_offset(cpu), from, size);