__get_cpu_var     388 arch/x86/kernel/apic_32.c 	struct clock_event_device *levt = &__get_cpu_var(lapic_events);
__get_cpu_var     458 arch/x86/kernel/apic_32.c 	struct clock_event_device *levt = &__get_cpu_var(lapic_events);
__get_cpu_var    1366 arch/x86/kernel/apic_32.c 	__get_cpu_var(irq_stat).irq_spurious_count++;
__get_cpu_var     400 arch/x86/kernel/apic_64.c 	struct clock_event_device *levt = &__get_cpu_var(lapic_events);
__get_cpu_var      64 arch/x86/kernel/cpu/mcheck/p4.c 	__get_cpu_var(irq_stat).irq_thermal_count++;
__get_cpu_var      91 arch/x86/kernel/cpu/mcheck/therm_throt.c 		__get_cpu_var(thermal_throttle_count)++;
__get_cpu_var      93 arch/x86/kernel/cpu/mcheck/therm_throt.c 	if (time_before64(tmp_jiffs, __get_cpu_var(next_check)))
__get_cpu_var      96 arch/x86/kernel/cpu/mcheck/therm_throt.c 	__get_cpu_var(next_check) = tmp_jiffs + CHECK_INTERVAL;
__get_cpu_var     102 arch/x86/kernel/cpu/mcheck/therm_throt.c 		       __get_cpu_var(thermal_throttle_count));
__get_cpu_var     283 arch/x86/kernel/cpu/perfctr-watchdog.c 	struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var     316 arch/x86/kernel/cpu/perfctr-watchdog.c 	struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var     370 arch/x86/kernel/cpu/perfctr-watchdog.c 	struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var     479 arch/x86/kernel/cpu/perfctr-watchdog.c 	struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var     571 arch/x86/kernel/cpu/perfctr-watchdog.c 	struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var     654 arch/x86/kernel/cpu/perfctr-watchdog.c 	struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var     780 arch/x86/kernel/cpu/perfctr-watchdog.c 	struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var     789 arch/x86/kernel/cpu/perfctr-watchdog.c 	struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var     180 arch/x86/kernel/genx2apic_uv_x.c 	id = x | __get_cpu_var(x2apic_extra_bits);
__get_cpu_var     232 arch/x86/kernel/genx2apic_uv_x.c 	__get_cpu_var(x2apic_extra_bits) = (pnode << 6);
__get_cpu_var    1709 arch/x86/kernel/io_apic_64.c 		irq = __get_cpu_var(vector_irq)[vector];
__get_cpu_var    1722 arch/x86/kernel/io_apic_64.c 		__get_cpu_var(vector_irq)[vector] = -1;
__get_cpu_var     198 arch/x86/kernel/irq_64.c 	irq = __get_cpu_var(vector_irq)[vector];
__get_cpu_var     394 arch/x86/kernel/kprobes.c 	__get_cpu_var(current_kprobe) = kcb->prev_kprobe.kp;
__get_cpu_var     403 arch/x86/kernel/kprobes.c 	__get_cpu_var(current_kprobe) = p;
__get_cpu_var     568 arch/x86/kernel/kprobes.c 		p = __get_cpu_var(current_kprobe);
__get_cpu_var     713 arch/x86/kernel/kprobes.c 			__get_cpu_var(current_kprobe) = &ri->rp->kp;
__get_cpu_var     716 arch/x86/kernel/kprobes.c 			__get_cpu_var(current_kprobe) = NULL;
__get_cpu_var     308 arch/x86/kernel/nmi.c 	__get_cpu_var(wd_enabled) = 1;
__get_cpu_var     313 arch/x86/kernel/nmi.c 	if (__get_cpu_var(wd_enabled))
__get_cpu_var     324 arch/x86/kernel/nmi.c 			__get_cpu_var(wd_enabled) = 0;
__get_cpu_var     329 arch/x86/kernel/nmi.c 		__get_cpu_var(wd_enabled) = 1;
__get_cpu_var     339 arch/x86/kernel/nmi.c 	if (__get_cpu_var(wd_enabled) == 0)
__get_cpu_var     343 arch/x86/kernel/nmi.c 	__get_cpu_var(wd_enabled) = 0;
__get_cpu_var     410 arch/x86/kernel/nmi.c 	if (__get_cpu_var(nmi_touch)) {
__get_cpu_var     411 arch/x86/kernel/nmi.c 		__get_cpu_var(nmi_touch) = 0;
__get_cpu_var     430 arch/x86/kernel/nmi.c 	if (!touched && __get_cpu_var(last_irq_sum) == sum) {
__get_cpu_var     435 arch/x86/kernel/nmi.c 		local_inc(&__get_cpu_var(alert_counter));
__get_cpu_var     436 arch/x86/kernel/nmi.c 		if (local_read(&__get_cpu_var(alert_counter)) == 5 * nmi_hz)
__get_cpu_var     443 arch/x86/kernel/nmi.c 		__get_cpu_var(last_irq_sum) = sum;
__get_cpu_var     444 arch/x86/kernel/nmi.c 		local_set(&__get_cpu_var(alert_counter), 0);
__get_cpu_var     448 arch/x86/kernel/nmi.c 	if (!__get_cpu_var(wd_enabled))
__get_cpu_var     232 arch/x86/kernel/paravirt.c 	BUG_ON(__get_cpu_var(paravirt_lazy_mode) != PARAVIRT_LAZY_NONE);
__get_cpu_var     235 arch/x86/kernel/paravirt.c 	__get_cpu_var(paravirt_lazy_mode) = mode;
__get_cpu_var     240 arch/x86/kernel/paravirt.c 	BUG_ON(__get_cpu_var(paravirt_lazy_mode) != mode);
__get_cpu_var     243 arch/x86/kernel/paravirt.c 	__get_cpu_var(paravirt_lazy_mode) = PARAVIRT_LAZY_NONE;
__get_cpu_var     268 arch/x86/kernel/paravirt.c 	return __get_cpu_var(paravirt_lazy_mode);
__get_cpu_var     113 arch/x86/kernel/process_32.c 			__get_cpu_var(irq_stat).idle_timestamp = jiffies;
__get_cpu_var     182 arch/x86/kernel/smp.c 	__get_cpu_var(irq_stat).irq_resched_count++;
__get_cpu_var     194 arch/x86/kernel/smp.c 	__get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var     207 arch/x86/kernel/smp.c 	__get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var    1409 arch/x86/kernel/smpboot.c 	__get_cpu_var(cpu_state) = CPU_DEAD;
__get_cpu_var     122 arch/x86/kernel/tlb_32.c 	__get_cpu_var(irq_stat).irq_tlb_count++;
__get_cpu_var      68 arch/x86/kernel/tlb_uv.c 	msp = __get_cpu_var(bau_control).msg_statuses + msg_slot;
__get_cpu_var      82 arch/x86/kernel/tlb_uv.c 		__get_cpu_var(ptcstats).alltlb++;
__get_cpu_var      85 arch/x86/kernel/tlb_uv.c 		__get_cpu_var(ptcstats).onetlb++;
__get_cpu_var      88 arch/x86/kernel/tlb_uv.c 	__get_cpu_var(ptcstats).requestee++;
__get_cpu_var     171 arch/x86/kernel/tlb_uv.c 			__get_cpu_var(ptcstats).s_retry++;
__get_cpu_var     185 arch/x86/kernel/tlb_uv.c 					__get_cpu_var(ptcstats).d_retry++;
__get_cpu_var     249 arch/x86/kernel/tlb_uv.c 	__get_cpu_var(ptcstats).sflush += (time2 - time1);
__get_cpu_var     251 arch/x86/kernel/tlb_uv.c 		__get_cpu_var(ptcstats).retriesok++;
__get_cpu_var     258 arch/x86/kernel/tlb_uv.c 		__get_cpu_var(ptcstats).ptc_i++;
__get_cpu_var     312 arch/x86/kernel/tlb_uv.c 	bau_desc = __get_cpu_var(bau_control).descriptor_base;
__get_cpu_var     337 arch/x86/kernel/tlb_uv.c 	__get_cpu_var(ptcstats).requestor++;
__get_cpu_var     338 arch/x86/kernel/tlb_uv.c 	__get_cpu_var(ptcstats).ntargeted += i;
__get_cpu_var     382 arch/x86/kernel/tlb_uv.c 	va_queue_first = __get_cpu_var(bau_control).va_queue_first;
__get_cpu_var     383 arch/x86/kernel/tlb_uv.c 	va_queue_last = __get_cpu_var(bau_control).va_queue_last;
__get_cpu_var     385 arch/x86/kernel/tlb_uv.c 	msg = __get_cpu_var(bau_control).bau_msg_head;
__get_cpu_var     397 arch/x86/kernel/tlb_uv.c 		__get_cpu_var(bau_control).bau_msg_head = msg;
__get_cpu_var     400 arch/x86/kernel/tlb_uv.c 		__get_cpu_var(ptcstats).nomsg++;
__get_cpu_var     402 arch/x86/kernel/tlb_uv.c 		__get_cpu_var(ptcstats).multmsg++;
__get_cpu_var     405 arch/x86/kernel/tlb_uv.c 	__get_cpu_var(ptcstats).dflush += (time2 - time1);
__get_cpu_var     197 arch/x86/kernel/vmiclock_32.c 	struct clock_event_device *evt = &__get_cpu_var(local_events);
__get_cpu_var     215 arch/x86/kernel/vmiclock_32.c 	evt = &__get_cpu_var(local_events);
__get_cpu_var     963 arch/x86/mach-voyager/voyager_smp.c 	__get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var     971 arch/x86/mach-voyager/voyager_smp.c 	__get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var     283 arch/x86/oprofile/nmi_int.c 	struct op_msrs *msrs = &__get_cpu_var(cpu_msrs);
__get_cpu_var     312 arch/x86/oprofile/nmi_int.c 	struct op_msrs const *msrs = &__get_cpu_var(cpu_msrs);
__get_cpu_var     324 arch/x86/oprofile/nmi_int.c 	struct op_msrs const *msrs = &__get_cpu_var(cpu_msrs);
__get_cpu_var     444 arch/x86/xen/enlighten.c 	start = __get_cpu_var(idt_desc).address;
__get_cpu_var     445 arch/x86/xen/enlighten.c 	end = start + __get_cpu_var(idt_desc).size + 1;
__get_cpu_var     483 arch/x86/xen/enlighten.c 	const struct desc_ptr *desc = &__get_cpu_var(idt_desc);
__get_cpu_var     498 arch/x86/xen/enlighten.c 	__get_cpu_var(idt_desc) = *desc;
__get_cpu_var    1045 arch/x86/xen/mmu.c 	active_mm = __get_cpu_var(cpu_tlbstate).active_mm;
__get_cpu_var     126 arch/x86/xen/multicalls.c 	struct mc_buffer *b = &__get_cpu_var(mc_buffer);
__get_cpu_var     185 arch/x86/xen/multicalls.c 	struct mc_buffer *b = &__get_cpu_var(mc_buffer);
__get_cpu_var     210 arch/x86/xen/multicalls.c 	struct mc_buffer *b = &__get_cpu_var(mc_buffer);
__get_cpu_var     235 arch/x86/xen/multicalls.c 	struct mc_buffer *b = &__get_cpu_var(mc_buffer);
__get_cpu_var      23 arch/x86/xen/multicalls.h 	local_irq_save(__get_cpu_var(xen_mc_irq_flags));
__get_cpu_var      54 arch/x86/xen/smp.c 	__get_cpu_var(irq_stat).irq_resched_count++;
__get_cpu_var     446 arch/x86/xen/smp.c 	__get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var     460 arch/x86/xen/smp.c 	__get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var     161 arch/x86/xen/spinlock.c 	prev = __get_cpu_var(lock_spinners);
__get_cpu_var     162 arch/x86/xen/spinlock.c 	__get_cpu_var(lock_spinners) = xl;
__get_cpu_var     181 arch/x86/xen/spinlock.c 	__get_cpu_var(lock_spinners) = prev;
__get_cpu_var     188 arch/x86/xen/spinlock.c 	int irq = __get_cpu_var(lock_kicker_irq);
__get_cpu_var      82 arch/x86/xen/time.c 	state = &__get_cpu_var(runstate);
__get_cpu_var     125 arch/x86/xen/time.c 	snap = &__get_cpu_var(runstate_snapshot);
__get_cpu_var     137 arch/x86/xen/time.c 	stolen = runnable + offline + __get_cpu_var(residual_stolen);
__get_cpu_var     143 arch/x86/xen/time.c 	__get_cpu_var(residual_stolen) = stolen;
__get_cpu_var     149 arch/x86/xen/time.c 	blocked += __get_cpu_var(residual_blocked);
__get_cpu_var     155 arch/x86/xen/time.c 	__get_cpu_var(residual_blocked) = blocked;
__get_cpu_var     414 arch/x86/xen/time.c 	struct clock_event_device *evt = &__get_cpu_var(xen_clock_events);
__get_cpu_var     465 arch/x86/xen/time.c 	clockevents_register_device(&__get_cpu_var(xen_clock_events));
__get_cpu_var      25 block/blk-softirq.c 	cpu_list = &__get_cpu_var(blk_cpu_done);
__get_cpu_var      46 block/blk-softirq.c 	list = &__get_cpu_var(blk_cpu_done);
__get_cpu_var      92 block/blk-softirq.c 				 &__get_cpu_var(blk_cpu_done));
__get_cpu_var     127 block/blk-softirq.c 		list = &__get_cpu_var(blk_cpu_done);
__get_cpu_var    1303 fs/buffer.c    	lru = &__get_cpu_var(bh_lrus);
__get_cpu_var    1347 fs/buffer.c    	lru = &__get_cpu_var(bh_lrus);
__get_cpu_var    3225 fs/buffer.c    	if (__get_cpu_var(bh_accounting).ratelimit++ < 4096)
__get_cpu_var    3227 fs/buffer.c    	__get_cpu_var(bh_accounting).ratelimit = 0;
__get_cpu_var      25 include/asm-generic/irq_regs.h 	return __get_cpu_var(__irq_regs);
__get_cpu_var      30 include/asm-generic/irq_regs.h 	struct pt_regs *old_regs, **pp_regs = &__get_cpu_var(__irq_regs);
__get_cpu_var      60 include/asm-generic/local.h #define cpu_local_read(l)	local_read(&__get_cpu_var(l))
__get_cpu_var      61 include/asm-generic/local.h #define cpu_local_set(l, i)	local_set(&__get_cpu_var(l), (i))
__get_cpu_var      62 include/asm-generic/local.h #define cpu_local_inc(l)	local_inc(&__get_cpu_var(l))
__get_cpu_var      63 include/asm-generic/local.h #define cpu_local_dec(l)	local_dec(&__get_cpu_var(l))
__get_cpu_var      64 include/asm-generic/local.h #define cpu_local_add(i, l)	local_add((i), &__get_cpu_var(l))
__get_cpu_var      65 include/asm-generic/local.h #define cpu_local_sub(i, l)	local_sub((i), &__get_cpu_var(l))
__get_cpu_var      70 include/asm-generic/local.h #define __cpu_local_inc(l)	__local_inc(&__get_cpu_var(l))
__get_cpu_var      71 include/asm-generic/local.h #define __cpu_local_dec(l)	__local_dec(&__get_cpu_var(l))
__get_cpu_var      72 include/asm-generic/local.h #define __cpu_local_add(i, l)	__local_add((i), &__get_cpu_var(l))
__get_cpu_var      73 include/asm-generic/local.h #define __cpu_local_sub(i, l)	__local_sub((i), &__get_cpu_var(l))
__get_cpu_var     354 include/asm-m32r/local.h #define cpu_local_read(l)    cpu_local_wrap_v(local_read(&__get_cpu_var(l)))
__get_cpu_var     355 include/asm-m32r/local.h #define cpu_local_set(l, i)  cpu_local_wrap(local_set(&__get_cpu_var(l), (i)))
__get_cpu_var     356 include/asm-m32r/local.h #define cpu_local_inc(l)     cpu_local_wrap(local_inc(&__get_cpu_var(l)))
__get_cpu_var     357 include/asm-m32r/local.h #define cpu_local_dec(l)     cpu_local_wrap(local_dec(&__get_cpu_var(l)))
__get_cpu_var     358 include/asm-m32r/local.h #define cpu_local_add(i, l)  cpu_local_wrap(local_add((i), &__get_cpu_var(l)))
__get_cpu_var     359 include/asm-m32r/local.h #define cpu_local_sub(i, l)  cpu_local_wrap(local_sub((i), &__get_cpu_var(l)))
__get_cpu_var     223 include/asm-x86/local.h #define cpu_local_read(l)    cpu_local_wrap_v(local_read(&__get_cpu_var((l))))
__get_cpu_var     224 include/asm-x86/local.h #define cpu_local_set(l, i)  cpu_local_wrap(local_set(&__get_cpu_var((l)), (i)))
__get_cpu_var     225 include/asm-x86/local.h #define cpu_local_inc(l)     cpu_local_wrap(local_inc(&__get_cpu_var((l))))
__get_cpu_var     226 include/asm-x86/local.h #define cpu_local_dec(l)     cpu_local_wrap(local_dec(&__get_cpu_var((l))))
__get_cpu_var     227 include/asm-x86/local.h #define cpu_local_add(i, l)  cpu_local_wrap(local_add((i), &__get_cpu_var((l))))
__get_cpu_var     228 include/asm-x86/local.h #define cpu_local_sub(i, l)  cpu_local_wrap(local_sub((i), &__get_cpu_var((l))))
__get_cpu_var      39 include/asm-x86/percpu.h 		__tmp = __get_cpu_var(var);				\
__get_cpu_var      47 include/asm-x86/percpu.h 		__get_cpu_var(var) = (val);				\
__get_cpu_var     138 include/asm-x86/processor.h #define current_cpu_data	__get_cpu_var(cpu_info)
__get_cpu_var     135 include/asm-x86/uv/uv_hub.h #define uv_hub_info 		(&__get_cpu_var(__uv_hub_info))
__get_cpu_var     194 include/linux/elevator.h 		__get_cpu_var(name) += (__val);			\
__get_cpu_var      38 include/linux/kernel_stat.h #define kstat_this_cpu	__get_cpu_var(kstat)
__get_cpu_var     223 include/linux/kprobes.h 	return (__get_cpu_var(current_kprobe));
__get_cpu_var     228 include/linux/kprobes.h 	__get_cpu_var(current_kprobe) = NULL;
__get_cpu_var     233 include/linux/kprobes.h 	return (&__get_cpu_var(kprobe_ctlblk));
__get_cpu_var      63 include/linux/percpu.h 	&__get_cpu_var(var); }))
__get_cpu_var     126 include/linux/rcupreempt.h 	__get_cpu_var(rcu_dyntick_sched).dynticks++;
__get_cpu_var     127 include/linux/rcupreempt.h 	WARN_ON_RATELIMIT(__get_cpu_var(rcu_dyntick_sched).dynticks & 0x1, &rs);
__get_cpu_var     134 include/linux/rcupreempt.h 	__get_cpu_var(rcu_dyntick_sched).dynticks++;
__get_cpu_var     136 include/linux/rcupreempt.h 	WARN_ON_RATELIMIT(!(__get_cpu_var(rcu_dyntick_sched).dynticks & 0x1),
__get_cpu_var      72 include/linux/vmstat.h 	__get_cpu_var(vm_event_states).event[item]++;
__get_cpu_var      83 include/linux/vmstat.h 	__get_cpu_var(vm_event_states).event[item] += delta;
__get_cpu_var      31 include/net/netdma.h 	chan = rcu_dereference(__get_cpu_var(softnet_data).net_dma);
__get_cpu_var      72 kernel/exit.c  		__get_cpu_var(process_counts)--;
__get_cpu_var    1239 kernel/fork.c  			__get_cpu_var(process_counts)++;
__get_cpu_var     202 kernel/hrtimer.c 	new_cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var     498 kernel/hrtimer.c 	return __get_cpu_var(hrtimer_bases).hres_active;
__get_cpu_var     541 kernel/hrtimer.c 	ktime_t *expires_next = &__get_cpu_var(hrtimer_bases).expires_next;
__get_cpu_var     600 kernel/hrtimer.c 	base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var     939 kernel/hrtimer.c 		reprogram = base->cpu_base == &__get_cpu_var(hrtimer_bases);
__get_cpu_var     995 kernel/hrtimer.c 			new_base->cpu_base == &__get_cpu_var(hrtimer_bases));
__get_cpu_var    1096 kernel/hrtimer.c 	struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var    1284 kernel/hrtimer.c 	struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var    1354 kernel/hrtimer.c 	run_hrtimer_pending(&__get_cpu_var(hrtimer_bases));
__get_cpu_var    1368 kernel/hrtimer.c 	struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var    1393 kernel/hrtimer.c 	struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var     276 kernel/kprobes.c 	__get_cpu_var(kprobe_instance) = kp;
__get_cpu_var     281 kernel/kprobes.c 	__get_cpu_var(kprobe_instance) = NULL;
__get_cpu_var     340 kernel/kprobes.c 	struct kprobe *cur = __get_cpu_var(kprobe_instance);
__get_cpu_var     355 kernel/kprobes.c 	struct kprobe *cur = __get_cpu_var(kprobe_instance);
__get_cpu_var     266 kernel/rcuclassic.c 	__call_rcu(head, &rcu_ctrlblk, &__get_cpu_var(rcu_data));
__get_cpu_var     294 kernel/rcuclassic.c 	__call_rcu(head, &rcu_bh_ctrlblk, &__get_cpu_var(rcu_bh_data));
__get_cpu_var     594 kernel/rcuclassic.c 	__rcu_process_callbacks(&rcu_ctrlblk, &__get_cpu_var(rcu_data));
__get_cpu_var     595 kernel/rcuclassic.c 	__rcu_process_callbacks(&rcu_bh_ctrlblk, &__get_cpu_var(rcu_bh_data));
__get_cpu_var     202 kernel/rcupreempt.c #define RCU_DATA_ME()		(&__get_cpu_var(rcu_data))
__get_cpu_var     667 kernel/rcutorture.c 	++__get_cpu_var(rcu_torture_count)[pipe_count];
__get_cpu_var     673 kernel/rcutorture.c 	++__get_cpu_var(rcu_torture_batch)[completed];
__get_cpu_var     722 kernel/rcutorture.c 		++__get_cpu_var(rcu_torture_count)[pipe_count];
__get_cpu_var     728 kernel/rcutorture.c 		++__get_cpu_var(rcu_torture_batch)[completed];
__get_cpu_var     637 kernel/sched.c #define this_rq()		(&__get_cpu_var(runqueues))
__get_cpu_var      64 kernel/sched_clock.c 	return &__get_cpu_var(sched_clock_data);
__get_cpu_var     977 kernel/sched_rt.c 	cpumask_t *lowest_mask = &__get_cpu_var(local_cpu_mask);
__get_cpu_var     153 kernel/smp.c   	struct call_single_queue *q = &__get_cpu_var(call_single_queue);
__get_cpu_var      62 kernel/softirq.c 	struct task_struct *tsk = __get_cpu_var(ksoftirqd);
__get_cpu_var     345 kernel/softirq.c 	*__get_cpu_var(tasklet_vec).tail = t;
__get_cpu_var     346 kernel/softirq.c 	__get_cpu_var(tasklet_vec).tail = &(t->next);
__get_cpu_var     359 kernel/softirq.c 	*__get_cpu_var(tasklet_hi_vec).tail = t;
__get_cpu_var     360 kernel/softirq.c 	__get_cpu_var(tasklet_hi_vec).tail = &(t->next);
__get_cpu_var     372 kernel/softirq.c 	list = __get_cpu_var(tasklet_vec).head;
__get_cpu_var     373 kernel/softirq.c 	__get_cpu_var(tasklet_vec).head = NULL;
__get_cpu_var     374 kernel/softirq.c 	__get_cpu_var(tasklet_vec).tail = &__get_cpu_var(tasklet_vec).head;
__get_cpu_var     395 kernel/softirq.c 		*__get_cpu_var(tasklet_vec).tail = t;
__get_cpu_var     396 kernel/softirq.c 		__get_cpu_var(tasklet_vec).tail = &(t->next);
__get_cpu_var     407 kernel/softirq.c 	list = __get_cpu_var(tasklet_hi_vec).head;
__get_cpu_var     408 kernel/softirq.c 	__get_cpu_var(tasklet_hi_vec).head = NULL;
__get_cpu_var     409 kernel/softirq.c 	__get_cpu_var(tasklet_hi_vec).tail = &__get_cpu_var(tasklet_hi_vec).head;
__get_cpu_var     430 kernel/softirq.c 		*__get_cpu_var(tasklet_hi_vec).tail = t;
__get_cpu_var     431 kernel/softirq.c 		__get_cpu_var(tasklet_hi_vec).tail = &(t->next);
__get_cpu_var     564 kernel/softirq.c 		*(__get_cpu_var(tasklet_vec).tail) = per_cpu(tasklet_vec, cpu).head;
__get_cpu_var     565 kernel/softirq.c 		__get_cpu_var(tasklet_vec).tail = per_cpu(tasklet_vec, cpu).tail;
__get_cpu_var     572 kernel/softirq.c 		*__get_cpu_var(tasklet_hi_vec).tail = per_cpu(tasklet_hi_vec, cpu).head;
__get_cpu_var     573 kernel/softirq.c 		__get_cpu_var(tasklet_hi_vec).tail = per_cpu(tasklet_hi_vec, cpu).tail;
__get_cpu_var      52 kernel/time/tick-common.c 	struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev;
__get_cpu_var     312 kernel/time/tick-common.c 	struct tick_device *td = &__get_cpu_var(tick_cpu_device);
__get_cpu_var     322 kernel/time/tick-common.c 	struct tick_device *td = &__get_cpu_var(tick_cpu_device);
__get_cpu_var      70 kernel/time/tick-oneshot.c 	struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev;
__get_cpu_var      80 kernel/time/tick-oneshot.c 	struct tick_device *td = &__get_cpu_var(tick_cpu_device);
__get_cpu_var     104 kernel/time/tick-oneshot.c 	struct tick_device *td = &__get_cpu_var(tick_cpu_device);
__get_cpu_var     218 kernel/time/tick-sched.c 	struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev;
__get_cpu_var     375 kernel/time/tick-sched.c 	struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var     469 kernel/time/tick-sched.c 	struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var     521 kernel/time/tick-sched.c 	struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var     628 kernel/time/tick-sched.c 	struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var     693 kernel/time/tick-sched.c 	struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var     708 kernel/time/tick-sched.c 	struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var     541 kernel/timer.c 	new_base = __get_cpu_var(tvec_bases);
__get_cpu_var     938 kernel/timer.c 	struct tvec_base *base = __get_cpu_var(tvec_bases);
__get_cpu_var    1031 kernel/timer.c 	struct tvec_base *base = __get_cpu_var(tvec_bases);
__get_cpu_var     160 lib/radix-tree.c 		rtp = &__get_cpu_var(radix_tree_preloads);
__get_cpu_var     211 lib/radix-tree.c 	rtp = &__get_cpu_var(radix_tree_preloads);
__get_cpu_var     218 lib/radix-tree.c 		rtp = &__get_cpu_var(radix_tree_preloads);
__get_cpu_var     573 mm/page-writeback.c 	p =  &__get_cpu_var(ratelimits);
__get_cpu_var     913 mm/slab.c      	int node = __get_cpu_var(reap_node);
__get_cpu_var     918 mm/slab.c      	__get_cpu_var(reap_node) = node;
__get_cpu_var    1091 mm/slab.c      	int node = __get_cpu_var(reap_node);
__get_cpu_var     145 mm/swap.c      		pvec = &__get_cpu_var(lru_rotate_pvecs);
__get_cpu_var     504 mm/swap.c      	local = &__get_cpu_var(committed_space);
__get_cpu_var     832 mm/vmstat.c    	schedule_delayed_work(&__get_cpu_var(vmstat_work),
__get_cpu_var    1380 net/core/dev.c 	sd = &__get_cpu_var(softnet_data);
__get_cpu_var    1401 net/core/dev.c 		sd = &__get_cpu_var(softnet_data);
__get_cpu_var    1946 net/core/dev.c 	queue = &__get_cpu_var(softnet_data);
__get_cpu_var    1948 net/core/dev.c 	__get_cpu_var(netdev_rx_stat).total++;
__get_cpu_var    1961 net/core/dev.c 	__get_cpu_var(netdev_rx_stat).dropped++;
__get_cpu_var    1985 net/core/dev.c 	struct softnet_data *sd = &__get_cpu_var(softnet_data);
__get_cpu_var    2240 net/core/dev.c 	__get_cpu_var(netdev_rx_stat).total++;
__get_cpu_var    2315 net/core/dev.c 	struct softnet_data *queue = &__get_cpu_var(softnet_data);
__get_cpu_var    2328 net/core/dev.c 	struct softnet_data *queue = &__get_cpu_var(softnet_data);
__get_cpu_var    2361 net/core/dev.c 	list_add_tail(&n->poll_list, &__get_cpu_var(softnet_data).poll_list);
__get_cpu_var    2370 net/core/dev.c 	struct list_head *list = &__get_cpu_var(softnet_data).poll_list;
__get_cpu_var    2455 net/core/dev.c 	__get_cpu_var(netdev_rx_stat).time_squeeze++;
__get_cpu_var    1994 net/core/sock.c 	__get_cpu_var(prot_inuse).val[prot->inuse_idx] += val;
__get_cpu_var      45 net/ipv4/syncookies.c 	__u32 *tmp = __get_cpu_var(cookie_scratch);
__get_cpu_var    1318 net/ipv4/tcp.c 		    __get_cpu_var(softnet_data).net_dma) {
__get_cpu_var      82 net/ipv6/syncookies.c 	__u32 *tmp = __get_cpu_var(cookie_scratch);
__get_cpu_var      98 net/sched/sch_generic.c 		__get_cpu_var(netdev_rx_stat).cpu_collision++;