__get_cpu_var 388 arch/x86/kernel/apic_32.c struct clock_event_device *levt = &__get_cpu_var(lapic_events);
__get_cpu_var 458 arch/x86/kernel/apic_32.c struct clock_event_device *levt = &__get_cpu_var(lapic_events);
__get_cpu_var 1366 arch/x86/kernel/apic_32.c __get_cpu_var(irq_stat).irq_spurious_count++;
__get_cpu_var 400 arch/x86/kernel/apic_64.c struct clock_event_device *levt = &__get_cpu_var(lapic_events);
__get_cpu_var 64 arch/x86/kernel/cpu/mcheck/p4.c __get_cpu_var(irq_stat).irq_thermal_count++;
__get_cpu_var 91 arch/x86/kernel/cpu/mcheck/therm_throt.c __get_cpu_var(thermal_throttle_count)++;
__get_cpu_var 93 arch/x86/kernel/cpu/mcheck/therm_throt.c if (time_before64(tmp_jiffs, __get_cpu_var(next_check)))
__get_cpu_var 96 arch/x86/kernel/cpu/mcheck/therm_throt.c __get_cpu_var(next_check) = tmp_jiffs + CHECK_INTERVAL;
__get_cpu_var 102 arch/x86/kernel/cpu/mcheck/therm_throt.c __get_cpu_var(thermal_throttle_count));
__get_cpu_var 283 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var 316 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var 370 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var 479 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var 571 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var 654 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var 780 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var 789 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk);
__get_cpu_var 180 arch/x86/kernel/genx2apic_uv_x.c id = x | __get_cpu_var(x2apic_extra_bits);
__get_cpu_var 232 arch/x86/kernel/genx2apic_uv_x.c __get_cpu_var(x2apic_extra_bits) = (pnode << 6);
__get_cpu_var 1709 arch/x86/kernel/io_apic_64.c irq = __get_cpu_var(vector_irq)[vector];
__get_cpu_var 1722 arch/x86/kernel/io_apic_64.c __get_cpu_var(vector_irq)[vector] = -1;
__get_cpu_var 198 arch/x86/kernel/irq_64.c irq = __get_cpu_var(vector_irq)[vector];
__get_cpu_var 394 arch/x86/kernel/kprobes.c __get_cpu_var(current_kprobe) = kcb->prev_kprobe.kp;
__get_cpu_var 403 arch/x86/kernel/kprobes.c __get_cpu_var(current_kprobe) = p;
__get_cpu_var 568 arch/x86/kernel/kprobes.c p = __get_cpu_var(current_kprobe);
__get_cpu_var 713 arch/x86/kernel/kprobes.c __get_cpu_var(current_kprobe) = &ri->rp->kp;
__get_cpu_var 716 arch/x86/kernel/kprobes.c __get_cpu_var(current_kprobe) = NULL;
__get_cpu_var 308 arch/x86/kernel/nmi.c __get_cpu_var(wd_enabled) = 1;
__get_cpu_var 313 arch/x86/kernel/nmi.c if (__get_cpu_var(wd_enabled))
__get_cpu_var 324 arch/x86/kernel/nmi.c __get_cpu_var(wd_enabled) = 0;
__get_cpu_var 329 arch/x86/kernel/nmi.c __get_cpu_var(wd_enabled) = 1;
__get_cpu_var 339 arch/x86/kernel/nmi.c if (__get_cpu_var(wd_enabled) == 0)
__get_cpu_var 343 arch/x86/kernel/nmi.c __get_cpu_var(wd_enabled) = 0;
__get_cpu_var 410 arch/x86/kernel/nmi.c if (__get_cpu_var(nmi_touch)) {
__get_cpu_var 411 arch/x86/kernel/nmi.c __get_cpu_var(nmi_touch) = 0;
__get_cpu_var 430 arch/x86/kernel/nmi.c if (!touched && __get_cpu_var(last_irq_sum) == sum) {
__get_cpu_var 435 arch/x86/kernel/nmi.c local_inc(&__get_cpu_var(alert_counter));
__get_cpu_var 436 arch/x86/kernel/nmi.c if (local_read(&__get_cpu_var(alert_counter)) == 5 * nmi_hz)
__get_cpu_var 443 arch/x86/kernel/nmi.c __get_cpu_var(last_irq_sum) = sum;
__get_cpu_var 444 arch/x86/kernel/nmi.c local_set(&__get_cpu_var(alert_counter), 0);
__get_cpu_var 448 arch/x86/kernel/nmi.c if (!__get_cpu_var(wd_enabled))
__get_cpu_var 232 arch/x86/kernel/paravirt.c BUG_ON(__get_cpu_var(paravirt_lazy_mode) != PARAVIRT_LAZY_NONE);
__get_cpu_var 235 arch/x86/kernel/paravirt.c __get_cpu_var(paravirt_lazy_mode) = mode;
__get_cpu_var 240 arch/x86/kernel/paravirt.c BUG_ON(__get_cpu_var(paravirt_lazy_mode) != mode);
__get_cpu_var 243 arch/x86/kernel/paravirt.c __get_cpu_var(paravirt_lazy_mode) = PARAVIRT_LAZY_NONE;
__get_cpu_var 268 arch/x86/kernel/paravirt.c return __get_cpu_var(paravirt_lazy_mode);
__get_cpu_var 113 arch/x86/kernel/process_32.c __get_cpu_var(irq_stat).idle_timestamp = jiffies;
__get_cpu_var 182 arch/x86/kernel/smp.c __get_cpu_var(irq_stat).irq_resched_count++;
__get_cpu_var 194 arch/x86/kernel/smp.c __get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var 207 arch/x86/kernel/smp.c __get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var 1409 arch/x86/kernel/smpboot.c __get_cpu_var(cpu_state) = CPU_DEAD;
__get_cpu_var 122 arch/x86/kernel/tlb_32.c __get_cpu_var(irq_stat).irq_tlb_count++;
__get_cpu_var 68 arch/x86/kernel/tlb_uv.c msp = __get_cpu_var(bau_control).msg_statuses + msg_slot;
__get_cpu_var 82 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).alltlb++;
__get_cpu_var 85 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).onetlb++;
__get_cpu_var 88 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).requestee++;
__get_cpu_var 171 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).s_retry++;
__get_cpu_var 185 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).d_retry++;
__get_cpu_var 249 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).sflush += (time2 - time1);
__get_cpu_var 251 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).retriesok++;
__get_cpu_var 258 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).ptc_i++;
__get_cpu_var 312 arch/x86/kernel/tlb_uv.c bau_desc = __get_cpu_var(bau_control).descriptor_base;
__get_cpu_var 337 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).requestor++;
__get_cpu_var 338 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).ntargeted += i;
__get_cpu_var 382 arch/x86/kernel/tlb_uv.c va_queue_first = __get_cpu_var(bau_control).va_queue_first;
__get_cpu_var 383 arch/x86/kernel/tlb_uv.c va_queue_last = __get_cpu_var(bau_control).va_queue_last;
__get_cpu_var 385 arch/x86/kernel/tlb_uv.c msg = __get_cpu_var(bau_control).bau_msg_head;
__get_cpu_var 397 arch/x86/kernel/tlb_uv.c __get_cpu_var(bau_control).bau_msg_head = msg;
__get_cpu_var 400 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).nomsg++;
__get_cpu_var 402 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).multmsg++;
__get_cpu_var 405 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).dflush += (time2 - time1);
__get_cpu_var 197 arch/x86/kernel/vmiclock_32.c struct clock_event_device *evt = &__get_cpu_var(local_events);
__get_cpu_var 215 arch/x86/kernel/vmiclock_32.c evt = &__get_cpu_var(local_events);
__get_cpu_var 963 arch/x86/mach-voyager/voyager_smp.c __get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var 971 arch/x86/mach-voyager/voyager_smp.c __get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var 283 arch/x86/oprofile/nmi_int.c struct op_msrs *msrs = &__get_cpu_var(cpu_msrs);
__get_cpu_var 312 arch/x86/oprofile/nmi_int.c struct op_msrs const *msrs = &__get_cpu_var(cpu_msrs);
__get_cpu_var 324 arch/x86/oprofile/nmi_int.c struct op_msrs const *msrs = &__get_cpu_var(cpu_msrs);
__get_cpu_var 444 arch/x86/xen/enlighten.c start = __get_cpu_var(idt_desc).address;
__get_cpu_var 445 arch/x86/xen/enlighten.c end = start + __get_cpu_var(idt_desc).size + 1;
__get_cpu_var 483 arch/x86/xen/enlighten.c const struct desc_ptr *desc = &__get_cpu_var(idt_desc);
__get_cpu_var 498 arch/x86/xen/enlighten.c __get_cpu_var(idt_desc) = *desc;
__get_cpu_var 1045 arch/x86/xen/mmu.c active_mm = __get_cpu_var(cpu_tlbstate).active_mm;
__get_cpu_var 126 arch/x86/xen/multicalls.c struct mc_buffer *b = &__get_cpu_var(mc_buffer);
__get_cpu_var 185 arch/x86/xen/multicalls.c struct mc_buffer *b = &__get_cpu_var(mc_buffer);
__get_cpu_var 210 arch/x86/xen/multicalls.c struct mc_buffer *b = &__get_cpu_var(mc_buffer);
__get_cpu_var 235 arch/x86/xen/multicalls.c struct mc_buffer *b = &__get_cpu_var(mc_buffer);
__get_cpu_var 23 arch/x86/xen/multicalls.h local_irq_save(__get_cpu_var(xen_mc_irq_flags));
__get_cpu_var 54 arch/x86/xen/smp.c __get_cpu_var(irq_stat).irq_resched_count++;
__get_cpu_var 446 arch/x86/xen/smp.c __get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var 460 arch/x86/xen/smp.c __get_cpu_var(irq_stat).irq_call_count++;
__get_cpu_var 161 arch/x86/xen/spinlock.c prev = __get_cpu_var(lock_spinners);
__get_cpu_var 162 arch/x86/xen/spinlock.c __get_cpu_var(lock_spinners) = xl;
__get_cpu_var 181 arch/x86/xen/spinlock.c __get_cpu_var(lock_spinners) = prev;
__get_cpu_var 188 arch/x86/xen/spinlock.c int irq = __get_cpu_var(lock_kicker_irq);
__get_cpu_var 82 arch/x86/xen/time.c state = &__get_cpu_var(runstate);
__get_cpu_var 125 arch/x86/xen/time.c snap = &__get_cpu_var(runstate_snapshot);
__get_cpu_var 137 arch/x86/xen/time.c stolen = runnable + offline + __get_cpu_var(residual_stolen);
__get_cpu_var 143 arch/x86/xen/time.c __get_cpu_var(residual_stolen) = stolen;
__get_cpu_var 149 arch/x86/xen/time.c blocked += __get_cpu_var(residual_blocked);
__get_cpu_var 155 arch/x86/xen/time.c __get_cpu_var(residual_blocked) = blocked;
__get_cpu_var 414 arch/x86/xen/time.c struct clock_event_device *evt = &__get_cpu_var(xen_clock_events);
__get_cpu_var 465 arch/x86/xen/time.c clockevents_register_device(&__get_cpu_var(xen_clock_events));
__get_cpu_var 25 block/blk-softirq.c cpu_list = &__get_cpu_var(blk_cpu_done);
__get_cpu_var 46 block/blk-softirq.c list = &__get_cpu_var(blk_cpu_done);
__get_cpu_var 92 block/blk-softirq.c &__get_cpu_var(blk_cpu_done));
__get_cpu_var 127 block/blk-softirq.c list = &__get_cpu_var(blk_cpu_done);
__get_cpu_var 1303 fs/buffer.c lru = &__get_cpu_var(bh_lrus);
__get_cpu_var 1347 fs/buffer.c lru = &__get_cpu_var(bh_lrus);
__get_cpu_var 3225 fs/buffer.c if (__get_cpu_var(bh_accounting).ratelimit++ < 4096)
__get_cpu_var 3227 fs/buffer.c __get_cpu_var(bh_accounting).ratelimit = 0;
__get_cpu_var 25 include/asm-generic/irq_regs.h return __get_cpu_var(__irq_regs);
__get_cpu_var 30 include/asm-generic/irq_regs.h struct pt_regs *old_regs, **pp_regs = &__get_cpu_var(__irq_regs);
__get_cpu_var 60 include/asm-generic/local.h #define cpu_local_read(l) local_read(&__get_cpu_var(l))
__get_cpu_var 61 include/asm-generic/local.h #define cpu_local_set(l, i) local_set(&__get_cpu_var(l), (i))
__get_cpu_var 62 include/asm-generic/local.h #define cpu_local_inc(l) local_inc(&__get_cpu_var(l))
__get_cpu_var 63 include/asm-generic/local.h #define cpu_local_dec(l) local_dec(&__get_cpu_var(l))
__get_cpu_var 64 include/asm-generic/local.h #define cpu_local_add(i, l) local_add((i), &__get_cpu_var(l))
__get_cpu_var 65 include/asm-generic/local.h #define cpu_local_sub(i, l) local_sub((i), &__get_cpu_var(l))
__get_cpu_var 70 include/asm-generic/local.h #define __cpu_local_inc(l) __local_inc(&__get_cpu_var(l))
__get_cpu_var 71 include/asm-generic/local.h #define __cpu_local_dec(l) __local_dec(&__get_cpu_var(l))
__get_cpu_var 72 include/asm-generic/local.h #define __cpu_local_add(i, l) __local_add((i), &__get_cpu_var(l))
__get_cpu_var 73 include/asm-generic/local.h #define __cpu_local_sub(i, l) __local_sub((i), &__get_cpu_var(l))
__get_cpu_var 354 include/asm-m32r/local.h #define cpu_local_read(l) cpu_local_wrap_v(local_read(&__get_cpu_var(l)))
__get_cpu_var 355 include/asm-m32r/local.h #define cpu_local_set(l, i) cpu_local_wrap(local_set(&__get_cpu_var(l), (i)))
__get_cpu_var 356 include/asm-m32r/local.h #define cpu_local_inc(l) cpu_local_wrap(local_inc(&__get_cpu_var(l)))
__get_cpu_var 357 include/asm-m32r/local.h #define cpu_local_dec(l) cpu_local_wrap(local_dec(&__get_cpu_var(l)))
__get_cpu_var 358 include/asm-m32r/local.h #define cpu_local_add(i, l) cpu_local_wrap(local_add((i), &__get_cpu_var(l)))
__get_cpu_var 359 include/asm-m32r/local.h #define cpu_local_sub(i, l) cpu_local_wrap(local_sub((i), &__get_cpu_var(l)))
__get_cpu_var 223 include/asm-x86/local.h #define cpu_local_read(l) cpu_local_wrap_v(local_read(&__get_cpu_var((l))))
__get_cpu_var 224 include/asm-x86/local.h #define cpu_local_set(l, i) cpu_local_wrap(local_set(&__get_cpu_var((l)), (i)))
__get_cpu_var 225 include/asm-x86/local.h #define cpu_local_inc(l) cpu_local_wrap(local_inc(&__get_cpu_var((l))))
__get_cpu_var 226 include/asm-x86/local.h #define cpu_local_dec(l) cpu_local_wrap(local_dec(&__get_cpu_var((l))))
__get_cpu_var 227 include/asm-x86/local.h #define cpu_local_add(i, l) cpu_local_wrap(local_add((i), &__get_cpu_var((l))))
__get_cpu_var 228 include/asm-x86/local.h #define cpu_local_sub(i, l) cpu_local_wrap(local_sub((i), &__get_cpu_var((l))))
__get_cpu_var 39 include/asm-x86/percpu.h __tmp = __get_cpu_var(var); \
__get_cpu_var 47 include/asm-x86/percpu.h __get_cpu_var(var) = (val); \
__get_cpu_var 138 include/asm-x86/processor.h #define current_cpu_data __get_cpu_var(cpu_info)
__get_cpu_var 135 include/asm-x86/uv/uv_hub.h #define uv_hub_info (&__get_cpu_var(__uv_hub_info))
__get_cpu_var 194 include/linux/elevator.h __get_cpu_var(name) += (__val); \
__get_cpu_var 38 include/linux/kernel_stat.h #define kstat_this_cpu __get_cpu_var(kstat)
__get_cpu_var 223 include/linux/kprobes.h return (__get_cpu_var(current_kprobe));
__get_cpu_var 228 include/linux/kprobes.h __get_cpu_var(current_kprobe) = NULL;
__get_cpu_var 233 include/linux/kprobes.h return (&__get_cpu_var(kprobe_ctlblk));
__get_cpu_var 63 include/linux/percpu.h &__get_cpu_var(var); }))
__get_cpu_var 126 include/linux/rcupreempt.h __get_cpu_var(rcu_dyntick_sched).dynticks++;
__get_cpu_var 127 include/linux/rcupreempt.h WARN_ON_RATELIMIT(__get_cpu_var(rcu_dyntick_sched).dynticks & 0x1, &rs);
__get_cpu_var 134 include/linux/rcupreempt.h __get_cpu_var(rcu_dyntick_sched).dynticks++;
__get_cpu_var 136 include/linux/rcupreempt.h WARN_ON_RATELIMIT(!(__get_cpu_var(rcu_dyntick_sched).dynticks & 0x1),
__get_cpu_var 72 include/linux/vmstat.h __get_cpu_var(vm_event_states).event[item]++;
__get_cpu_var 83 include/linux/vmstat.h __get_cpu_var(vm_event_states).event[item] += delta;
__get_cpu_var 31 include/net/netdma.h chan = rcu_dereference(__get_cpu_var(softnet_data).net_dma);
__get_cpu_var 72 kernel/exit.c __get_cpu_var(process_counts)--;
__get_cpu_var 1239 kernel/fork.c __get_cpu_var(process_counts)++;
__get_cpu_var 202 kernel/hrtimer.c new_cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var 498 kernel/hrtimer.c return __get_cpu_var(hrtimer_bases).hres_active;
__get_cpu_var 541 kernel/hrtimer.c ktime_t *expires_next = &__get_cpu_var(hrtimer_bases).expires_next;
__get_cpu_var 600 kernel/hrtimer.c base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var 939 kernel/hrtimer.c reprogram = base->cpu_base == &__get_cpu_var(hrtimer_bases);
__get_cpu_var 995 kernel/hrtimer.c new_base->cpu_base == &__get_cpu_var(hrtimer_bases));
__get_cpu_var 1096 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var 1284 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var 1354 kernel/hrtimer.c run_hrtimer_pending(&__get_cpu_var(hrtimer_bases));
__get_cpu_var 1368 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var 1393 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases);
__get_cpu_var 276 kernel/kprobes.c __get_cpu_var(kprobe_instance) = kp;
__get_cpu_var 281 kernel/kprobes.c __get_cpu_var(kprobe_instance) = NULL;
__get_cpu_var 340 kernel/kprobes.c struct kprobe *cur = __get_cpu_var(kprobe_instance);
__get_cpu_var 355 kernel/kprobes.c struct kprobe *cur = __get_cpu_var(kprobe_instance);
__get_cpu_var 266 kernel/rcuclassic.c __call_rcu(head, &rcu_ctrlblk, &__get_cpu_var(rcu_data));
__get_cpu_var 294 kernel/rcuclassic.c __call_rcu(head, &rcu_bh_ctrlblk, &__get_cpu_var(rcu_bh_data));
__get_cpu_var 594 kernel/rcuclassic.c __rcu_process_callbacks(&rcu_ctrlblk, &__get_cpu_var(rcu_data));
__get_cpu_var 595 kernel/rcuclassic.c __rcu_process_callbacks(&rcu_bh_ctrlblk, &__get_cpu_var(rcu_bh_data));
__get_cpu_var 202 kernel/rcupreempt.c #define RCU_DATA_ME() (&__get_cpu_var(rcu_data))
__get_cpu_var 667 kernel/rcutorture.c ++__get_cpu_var(rcu_torture_count)[pipe_count];
__get_cpu_var 673 kernel/rcutorture.c ++__get_cpu_var(rcu_torture_batch)[completed];
__get_cpu_var 722 kernel/rcutorture.c ++__get_cpu_var(rcu_torture_count)[pipe_count];
__get_cpu_var 728 kernel/rcutorture.c ++__get_cpu_var(rcu_torture_batch)[completed];
__get_cpu_var 637 kernel/sched.c #define this_rq() (&__get_cpu_var(runqueues))
__get_cpu_var 64 kernel/sched_clock.c return &__get_cpu_var(sched_clock_data);
__get_cpu_var 977 kernel/sched_rt.c cpumask_t *lowest_mask = &__get_cpu_var(local_cpu_mask);
__get_cpu_var 153 kernel/smp.c struct call_single_queue *q = &__get_cpu_var(call_single_queue);
__get_cpu_var 62 kernel/softirq.c struct task_struct *tsk = __get_cpu_var(ksoftirqd);
__get_cpu_var 345 kernel/softirq.c *__get_cpu_var(tasklet_vec).tail = t;
__get_cpu_var 346 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = &(t->next);
__get_cpu_var 359 kernel/softirq.c *__get_cpu_var(tasklet_hi_vec).tail = t;
__get_cpu_var 360 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = &(t->next);
__get_cpu_var 372 kernel/softirq.c list = __get_cpu_var(tasklet_vec).head;
__get_cpu_var 373 kernel/softirq.c __get_cpu_var(tasklet_vec).head = NULL;
__get_cpu_var 374 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = &__get_cpu_var(tasklet_vec).head;
__get_cpu_var 395 kernel/softirq.c *__get_cpu_var(tasklet_vec).tail = t;
__get_cpu_var 396 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = &(t->next);
__get_cpu_var 407 kernel/softirq.c list = __get_cpu_var(tasklet_hi_vec).head;
__get_cpu_var 408 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).head = NULL;
__get_cpu_var 409 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = &__get_cpu_var(tasklet_hi_vec).head;
__get_cpu_var 430 kernel/softirq.c *__get_cpu_var(tasklet_hi_vec).tail = t;
__get_cpu_var 431 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = &(t->next);
__get_cpu_var 564 kernel/softirq.c *(__get_cpu_var(tasklet_vec).tail) = per_cpu(tasklet_vec, cpu).head;
__get_cpu_var 565 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = per_cpu(tasklet_vec, cpu).tail;
__get_cpu_var 572 kernel/softirq.c *__get_cpu_var(tasklet_hi_vec).tail = per_cpu(tasklet_hi_vec, cpu).head;
__get_cpu_var 573 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = per_cpu(tasklet_hi_vec, cpu).tail;
__get_cpu_var 52 kernel/time/tick-common.c struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev;
__get_cpu_var 312 kernel/time/tick-common.c struct tick_device *td = &__get_cpu_var(tick_cpu_device);
__get_cpu_var 322 kernel/time/tick-common.c struct tick_device *td = &__get_cpu_var(tick_cpu_device);
__get_cpu_var 70 kernel/time/tick-oneshot.c struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev;
__get_cpu_var 80 kernel/time/tick-oneshot.c struct tick_device *td = &__get_cpu_var(tick_cpu_device);
__get_cpu_var 104 kernel/time/tick-oneshot.c struct tick_device *td = &__get_cpu_var(tick_cpu_device);
__get_cpu_var 218 kernel/time/tick-sched.c struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev;
__get_cpu_var 375 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var 469 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var 521 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var 628 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var 693 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var 708 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched);
__get_cpu_var 541 kernel/timer.c new_base = __get_cpu_var(tvec_bases);
__get_cpu_var 938 kernel/timer.c struct tvec_base *base = __get_cpu_var(tvec_bases);
__get_cpu_var 1031 kernel/timer.c struct tvec_base *base = __get_cpu_var(tvec_bases);
__get_cpu_var 160 lib/radix-tree.c rtp = &__get_cpu_var(radix_tree_preloads);
__get_cpu_var 211 lib/radix-tree.c rtp = &__get_cpu_var(radix_tree_preloads);
__get_cpu_var 218 lib/radix-tree.c rtp = &__get_cpu_var(radix_tree_preloads);
__get_cpu_var 573 mm/page-writeback.c p = &__get_cpu_var(ratelimits);
__get_cpu_var 913 mm/slab.c int node = __get_cpu_var(reap_node);
__get_cpu_var 918 mm/slab.c __get_cpu_var(reap_node) = node;
__get_cpu_var 1091 mm/slab.c int node = __get_cpu_var(reap_node);
__get_cpu_var 145 mm/swap.c pvec = &__get_cpu_var(lru_rotate_pvecs);
__get_cpu_var 504 mm/swap.c local = &__get_cpu_var(committed_space);
__get_cpu_var 832 mm/vmstat.c schedule_delayed_work(&__get_cpu_var(vmstat_work),
__get_cpu_var 1380 net/core/dev.c sd = &__get_cpu_var(softnet_data);
__get_cpu_var 1401 net/core/dev.c sd = &__get_cpu_var(softnet_data);
__get_cpu_var 1946 net/core/dev.c queue = &__get_cpu_var(softnet_data);
__get_cpu_var 1948 net/core/dev.c __get_cpu_var(netdev_rx_stat).total++;
__get_cpu_var 1961 net/core/dev.c __get_cpu_var(netdev_rx_stat).dropped++;
__get_cpu_var 1985 net/core/dev.c struct softnet_data *sd = &__get_cpu_var(softnet_data);
__get_cpu_var 2240 net/core/dev.c __get_cpu_var(netdev_rx_stat).total++;
__get_cpu_var 2315 net/core/dev.c struct softnet_data *queue = &__get_cpu_var(softnet_data);
__get_cpu_var 2328 net/core/dev.c struct softnet_data *queue = &__get_cpu_var(softnet_data);
__get_cpu_var 2361 net/core/dev.c list_add_tail(&n->poll_list, &__get_cpu_var(softnet_data).poll_list);
__get_cpu_var 2370 net/core/dev.c struct list_head *list = &__get_cpu_var(softnet_data).poll_list;
__get_cpu_var 2455 net/core/dev.c __get_cpu_var(netdev_rx_stat).time_squeeze++;
__get_cpu_var 1994 net/core/sock.c __get_cpu_var(prot_inuse).val[prot->inuse_idx] += val;
__get_cpu_var 45 net/ipv4/syncookies.c __u32 *tmp = __get_cpu_var(cookie_scratch);
__get_cpu_var 1318 net/ipv4/tcp.c __get_cpu_var(softnet_data).net_dma) {
__get_cpu_var 82 net/ipv6/syncookies.c __u32 *tmp = __get_cpu_var(cookie_scratch);
__get_cpu_var 98 net/sched/sch_generic.c __get_cpu_var(netdev_rx_stat).cpu_collision++;