__get_cpu_var 388 arch/x86/kernel/apic_32.c struct clock_event_device *levt = &__get_cpu_var(lapic_events); __get_cpu_var 458 arch/x86/kernel/apic_32.c struct clock_event_device *levt = &__get_cpu_var(lapic_events); __get_cpu_var 1366 arch/x86/kernel/apic_32.c __get_cpu_var(irq_stat).irq_spurious_count++; __get_cpu_var 400 arch/x86/kernel/apic_64.c struct clock_event_device *levt = &__get_cpu_var(lapic_events); __get_cpu_var 64 arch/x86/kernel/cpu/mcheck/p4.c __get_cpu_var(irq_stat).irq_thermal_count++; __get_cpu_var 91 arch/x86/kernel/cpu/mcheck/therm_throt.c __get_cpu_var(thermal_throttle_count)++; __get_cpu_var 93 arch/x86/kernel/cpu/mcheck/therm_throt.c if (time_before64(tmp_jiffs, __get_cpu_var(next_check))) __get_cpu_var 96 arch/x86/kernel/cpu/mcheck/therm_throt.c __get_cpu_var(next_check) = tmp_jiffs + CHECK_INTERVAL; __get_cpu_var 102 arch/x86/kernel/cpu/mcheck/therm_throt.c __get_cpu_var(thermal_throttle_count)); __get_cpu_var 283 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk); __get_cpu_var 316 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk); __get_cpu_var 370 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk); __get_cpu_var 479 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk); __get_cpu_var 571 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk); __get_cpu_var 654 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk); __get_cpu_var 780 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk); __get_cpu_var 789 arch/x86/kernel/cpu/perfctr-watchdog.c struct nmi_watchdog_ctlblk *wd = &__get_cpu_var(nmi_watchdog_ctlblk); __get_cpu_var 180 arch/x86/kernel/genx2apic_uv_x.c id = x | __get_cpu_var(x2apic_extra_bits); __get_cpu_var 232 arch/x86/kernel/genx2apic_uv_x.c __get_cpu_var(x2apic_extra_bits) = (pnode << 6); __get_cpu_var 1709 arch/x86/kernel/io_apic_64.c irq = __get_cpu_var(vector_irq)[vector]; __get_cpu_var 1722 arch/x86/kernel/io_apic_64.c __get_cpu_var(vector_irq)[vector] = -1; __get_cpu_var 198 arch/x86/kernel/irq_64.c irq = __get_cpu_var(vector_irq)[vector]; __get_cpu_var 394 arch/x86/kernel/kprobes.c __get_cpu_var(current_kprobe) = kcb->prev_kprobe.kp; __get_cpu_var 403 arch/x86/kernel/kprobes.c __get_cpu_var(current_kprobe) = p; __get_cpu_var 568 arch/x86/kernel/kprobes.c p = __get_cpu_var(current_kprobe); __get_cpu_var 713 arch/x86/kernel/kprobes.c __get_cpu_var(current_kprobe) = &ri->rp->kp; __get_cpu_var 716 arch/x86/kernel/kprobes.c __get_cpu_var(current_kprobe) = NULL; __get_cpu_var 308 arch/x86/kernel/nmi.c __get_cpu_var(wd_enabled) = 1; __get_cpu_var 313 arch/x86/kernel/nmi.c if (__get_cpu_var(wd_enabled)) __get_cpu_var 324 arch/x86/kernel/nmi.c __get_cpu_var(wd_enabled) = 0; __get_cpu_var 329 arch/x86/kernel/nmi.c __get_cpu_var(wd_enabled) = 1; __get_cpu_var 339 arch/x86/kernel/nmi.c if (__get_cpu_var(wd_enabled) == 0) __get_cpu_var 343 arch/x86/kernel/nmi.c __get_cpu_var(wd_enabled) = 0; __get_cpu_var 410 arch/x86/kernel/nmi.c if (__get_cpu_var(nmi_touch)) { __get_cpu_var 411 arch/x86/kernel/nmi.c __get_cpu_var(nmi_touch) = 0; __get_cpu_var 430 arch/x86/kernel/nmi.c if (!touched && __get_cpu_var(last_irq_sum) == sum) { __get_cpu_var 435 arch/x86/kernel/nmi.c local_inc(&__get_cpu_var(alert_counter)); __get_cpu_var 436 arch/x86/kernel/nmi.c if (local_read(&__get_cpu_var(alert_counter)) == 5 * nmi_hz) __get_cpu_var 443 arch/x86/kernel/nmi.c __get_cpu_var(last_irq_sum) = sum; __get_cpu_var 444 arch/x86/kernel/nmi.c local_set(&__get_cpu_var(alert_counter), 0); __get_cpu_var 448 arch/x86/kernel/nmi.c if (!__get_cpu_var(wd_enabled)) __get_cpu_var 232 arch/x86/kernel/paravirt.c BUG_ON(__get_cpu_var(paravirt_lazy_mode) != PARAVIRT_LAZY_NONE); __get_cpu_var 235 arch/x86/kernel/paravirt.c __get_cpu_var(paravirt_lazy_mode) = mode; __get_cpu_var 240 arch/x86/kernel/paravirt.c BUG_ON(__get_cpu_var(paravirt_lazy_mode) != mode); __get_cpu_var 243 arch/x86/kernel/paravirt.c __get_cpu_var(paravirt_lazy_mode) = PARAVIRT_LAZY_NONE; __get_cpu_var 268 arch/x86/kernel/paravirt.c return __get_cpu_var(paravirt_lazy_mode); __get_cpu_var 113 arch/x86/kernel/process_32.c __get_cpu_var(irq_stat).idle_timestamp = jiffies; __get_cpu_var 182 arch/x86/kernel/smp.c __get_cpu_var(irq_stat).irq_resched_count++; __get_cpu_var 194 arch/x86/kernel/smp.c __get_cpu_var(irq_stat).irq_call_count++; __get_cpu_var 207 arch/x86/kernel/smp.c __get_cpu_var(irq_stat).irq_call_count++; __get_cpu_var 1409 arch/x86/kernel/smpboot.c __get_cpu_var(cpu_state) = CPU_DEAD; __get_cpu_var 122 arch/x86/kernel/tlb_32.c __get_cpu_var(irq_stat).irq_tlb_count++; __get_cpu_var 68 arch/x86/kernel/tlb_uv.c msp = __get_cpu_var(bau_control).msg_statuses + msg_slot; __get_cpu_var 82 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).alltlb++; __get_cpu_var 85 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).onetlb++; __get_cpu_var 88 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).requestee++; __get_cpu_var 171 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).s_retry++; __get_cpu_var 185 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).d_retry++; __get_cpu_var 249 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).sflush += (time2 - time1); __get_cpu_var 251 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).retriesok++; __get_cpu_var 258 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).ptc_i++; __get_cpu_var 312 arch/x86/kernel/tlb_uv.c bau_desc = __get_cpu_var(bau_control).descriptor_base; __get_cpu_var 337 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).requestor++; __get_cpu_var 338 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).ntargeted += i; __get_cpu_var 382 arch/x86/kernel/tlb_uv.c va_queue_first = __get_cpu_var(bau_control).va_queue_first; __get_cpu_var 383 arch/x86/kernel/tlb_uv.c va_queue_last = __get_cpu_var(bau_control).va_queue_last; __get_cpu_var 385 arch/x86/kernel/tlb_uv.c msg = __get_cpu_var(bau_control).bau_msg_head; __get_cpu_var 397 arch/x86/kernel/tlb_uv.c __get_cpu_var(bau_control).bau_msg_head = msg; __get_cpu_var 400 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).nomsg++; __get_cpu_var 402 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).multmsg++; __get_cpu_var 405 arch/x86/kernel/tlb_uv.c __get_cpu_var(ptcstats).dflush += (time2 - time1); __get_cpu_var 197 arch/x86/kernel/vmiclock_32.c struct clock_event_device *evt = &__get_cpu_var(local_events); __get_cpu_var 215 arch/x86/kernel/vmiclock_32.c evt = &__get_cpu_var(local_events); __get_cpu_var 963 arch/x86/mach-voyager/voyager_smp.c __get_cpu_var(irq_stat).irq_call_count++; __get_cpu_var 971 arch/x86/mach-voyager/voyager_smp.c __get_cpu_var(irq_stat).irq_call_count++; __get_cpu_var 283 arch/x86/oprofile/nmi_int.c struct op_msrs *msrs = &__get_cpu_var(cpu_msrs); __get_cpu_var 312 arch/x86/oprofile/nmi_int.c struct op_msrs const *msrs = &__get_cpu_var(cpu_msrs); __get_cpu_var 324 arch/x86/oprofile/nmi_int.c struct op_msrs const *msrs = &__get_cpu_var(cpu_msrs); __get_cpu_var 444 arch/x86/xen/enlighten.c start = __get_cpu_var(idt_desc).address; __get_cpu_var 445 arch/x86/xen/enlighten.c end = start + __get_cpu_var(idt_desc).size + 1; __get_cpu_var 483 arch/x86/xen/enlighten.c const struct desc_ptr *desc = &__get_cpu_var(idt_desc); __get_cpu_var 498 arch/x86/xen/enlighten.c __get_cpu_var(idt_desc) = *desc; __get_cpu_var 1045 arch/x86/xen/mmu.c active_mm = __get_cpu_var(cpu_tlbstate).active_mm; __get_cpu_var 126 arch/x86/xen/multicalls.c struct mc_buffer *b = &__get_cpu_var(mc_buffer); __get_cpu_var 185 arch/x86/xen/multicalls.c struct mc_buffer *b = &__get_cpu_var(mc_buffer); __get_cpu_var 210 arch/x86/xen/multicalls.c struct mc_buffer *b = &__get_cpu_var(mc_buffer); __get_cpu_var 235 arch/x86/xen/multicalls.c struct mc_buffer *b = &__get_cpu_var(mc_buffer); __get_cpu_var 23 arch/x86/xen/multicalls.h local_irq_save(__get_cpu_var(xen_mc_irq_flags)); __get_cpu_var 54 arch/x86/xen/smp.c __get_cpu_var(irq_stat).irq_resched_count++; __get_cpu_var 446 arch/x86/xen/smp.c __get_cpu_var(irq_stat).irq_call_count++; __get_cpu_var 460 arch/x86/xen/smp.c __get_cpu_var(irq_stat).irq_call_count++; __get_cpu_var 161 arch/x86/xen/spinlock.c prev = __get_cpu_var(lock_spinners); __get_cpu_var 162 arch/x86/xen/spinlock.c __get_cpu_var(lock_spinners) = xl; __get_cpu_var 181 arch/x86/xen/spinlock.c __get_cpu_var(lock_spinners) = prev; __get_cpu_var 188 arch/x86/xen/spinlock.c int irq = __get_cpu_var(lock_kicker_irq); __get_cpu_var 82 arch/x86/xen/time.c state = &__get_cpu_var(runstate); __get_cpu_var 125 arch/x86/xen/time.c snap = &__get_cpu_var(runstate_snapshot); __get_cpu_var 137 arch/x86/xen/time.c stolen = runnable + offline + __get_cpu_var(residual_stolen); __get_cpu_var 143 arch/x86/xen/time.c __get_cpu_var(residual_stolen) = stolen; __get_cpu_var 149 arch/x86/xen/time.c blocked += __get_cpu_var(residual_blocked); __get_cpu_var 155 arch/x86/xen/time.c __get_cpu_var(residual_blocked) = blocked; __get_cpu_var 414 arch/x86/xen/time.c struct clock_event_device *evt = &__get_cpu_var(xen_clock_events); __get_cpu_var 465 arch/x86/xen/time.c clockevents_register_device(&__get_cpu_var(xen_clock_events)); __get_cpu_var 25 block/blk-softirq.c cpu_list = &__get_cpu_var(blk_cpu_done); __get_cpu_var 46 block/blk-softirq.c list = &__get_cpu_var(blk_cpu_done); __get_cpu_var 92 block/blk-softirq.c &__get_cpu_var(blk_cpu_done)); __get_cpu_var 127 block/blk-softirq.c list = &__get_cpu_var(blk_cpu_done); __get_cpu_var 1303 fs/buffer.c lru = &__get_cpu_var(bh_lrus); __get_cpu_var 1347 fs/buffer.c lru = &__get_cpu_var(bh_lrus); __get_cpu_var 3225 fs/buffer.c if (__get_cpu_var(bh_accounting).ratelimit++ < 4096) __get_cpu_var 3227 fs/buffer.c __get_cpu_var(bh_accounting).ratelimit = 0; __get_cpu_var 25 include/asm-generic/irq_regs.h return __get_cpu_var(__irq_regs); __get_cpu_var 30 include/asm-generic/irq_regs.h struct pt_regs *old_regs, **pp_regs = &__get_cpu_var(__irq_regs); __get_cpu_var 60 include/asm-generic/local.h #define cpu_local_read(l) local_read(&__get_cpu_var(l)) __get_cpu_var 61 include/asm-generic/local.h #define cpu_local_set(l, i) local_set(&__get_cpu_var(l), (i)) __get_cpu_var 62 include/asm-generic/local.h #define cpu_local_inc(l) local_inc(&__get_cpu_var(l)) __get_cpu_var 63 include/asm-generic/local.h #define cpu_local_dec(l) local_dec(&__get_cpu_var(l)) __get_cpu_var 64 include/asm-generic/local.h #define cpu_local_add(i, l) local_add((i), &__get_cpu_var(l)) __get_cpu_var 65 include/asm-generic/local.h #define cpu_local_sub(i, l) local_sub((i), &__get_cpu_var(l)) __get_cpu_var 70 include/asm-generic/local.h #define __cpu_local_inc(l) __local_inc(&__get_cpu_var(l)) __get_cpu_var 71 include/asm-generic/local.h #define __cpu_local_dec(l) __local_dec(&__get_cpu_var(l)) __get_cpu_var 72 include/asm-generic/local.h #define __cpu_local_add(i, l) __local_add((i), &__get_cpu_var(l)) __get_cpu_var 73 include/asm-generic/local.h #define __cpu_local_sub(i, l) __local_sub((i), &__get_cpu_var(l)) __get_cpu_var 354 include/asm-m32r/local.h #define cpu_local_read(l) cpu_local_wrap_v(local_read(&__get_cpu_var(l))) __get_cpu_var 355 include/asm-m32r/local.h #define cpu_local_set(l, i) cpu_local_wrap(local_set(&__get_cpu_var(l), (i))) __get_cpu_var 356 include/asm-m32r/local.h #define cpu_local_inc(l) cpu_local_wrap(local_inc(&__get_cpu_var(l))) __get_cpu_var 357 include/asm-m32r/local.h #define cpu_local_dec(l) cpu_local_wrap(local_dec(&__get_cpu_var(l))) __get_cpu_var 358 include/asm-m32r/local.h #define cpu_local_add(i, l) cpu_local_wrap(local_add((i), &__get_cpu_var(l))) __get_cpu_var 359 include/asm-m32r/local.h #define cpu_local_sub(i, l) cpu_local_wrap(local_sub((i), &__get_cpu_var(l))) __get_cpu_var 223 include/asm-x86/local.h #define cpu_local_read(l) cpu_local_wrap_v(local_read(&__get_cpu_var((l)))) __get_cpu_var 224 include/asm-x86/local.h #define cpu_local_set(l, i) cpu_local_wrap(local_set(&__get_cpu_var((l)), (i))) __get_cpu_var 225 include/asm-x86/local.h #define cpu_local_inc(l) cpu_local_wrap(local_inc(&__get_cpu_var((l)))) __get_cpu_var 226 include/asm-x86/local.h #define cpu_local_dec(l) cpu_local_wrap(local_dec(&__get_cpu_var((l)))) __get_cpu_var 227 include/asm-x86/local.h #define cpu_local_add(i, l) cpu_local_wrap(local_add((i), &__get_cpu_var((l)))) __get_cpu_var 228 include/asm-x86/local.h #define cpu_local_sub(i, l) cpu_local_wrap(local_sub((i), &__get_cpu_var((l)))) __get_cpu_var 39 include/asm-x86/percpu.h __tmp = __get_cpu_var(var); \ __get_cpu_var 47 include/asm-x86/percpu.h __get_cpu_var(var) = (val); \ __get_cpu_var 138 include/asm-x86/processor.h #define current_cpu_data __get_cpu_var(cpu_info) __get_cpu_var 135 include/asm-x86/uv/uv_hub.h #define uv_hub_info (&__get_cpu_var(__uv_hub_info)) __get_cpu_var 194 include/linux/elevator.h __get_cpu_var(name) += (__val); \ __get_cpu_var 38 include/linux/kernel_stat.h #define kstat_this_cpu __get_cpu_var(kstat) __get_cpu_var 223 include/linux/kprobes.h return (__get_cpu_var(current_kprobe)); __get_cpu_var 228 include/linux/kprobes.h __get_cpu_var(current_kprobe) = NULL; __get_cpu_var 233 include/linux/kprobes.h return (&__get_cpu_var(kprobe_ctlblk)); __get_cpu_var 63 include/linux/percpu.h &__get_cpu_var(var); })) __get_cpu_var 126 include/linux/rcupreempt.h __get_cpu_var(rcu_dyntick_sched).dynticks++; __get_cpu_var 127 include/linux/rcupreempt.h WARN_ON_RATELIMIT(__get_cpu_var(rcu_dyntick_sched).dynticks & 0x1, &rs); __get_cpu_var 134 include/linux/rcupreempt.h __get_cpu_var(rcu_dyntick_sched).dynticks++; __get_cpu_var 136 include/linux/rcupreempt.h WARN_ON_RATELIMIT(!(__get_cpu_var(rcu_dyntick_sched).dynticks & 0x1), __get_cpu_var 72 include/linux/vmstat.h __get_cpu_var(vm_event_states).event[item]++; __get_cpu_var 83 include/linux/vmstat.h __get_cpu_var(vm_event_states).event[item] += delta; __get_cpu_var 31 include/net/netdma.h chan = rcu_dereference(__get_cpu_var(softnet_data).net_dma); __get_cpu_var 72 kernel/exit.c __get_cpu_var(process_counts)--; __get_cpu_var 1239 kernel/fork.c __get_cpu_var(process_counts)++; __get_cpu_var 202 kernel/hrtimer.c new_cpu_base = &__get_cpu_var(hrtimer_bases); __get_cpu_var 498 kernel/hrtimer.c return __get_cpu_var(hrtimer_bases).hres_active; __get_cpu_var 541 kernel/hrtimer.c ktime_t *expires_next = &__get_cpu_var(hrtimer_bases).expires_next; __get_cpu_var 600 kernel/hrtimer.c base = &__get_cpu_var(hrtimer_bases); __get_cpu_var 939 kernel/hrtimer.c reprogram = base->cpu_base == &__get_cpu_var(hrtimer_bases); __get_cpu_var 995 kernel/hrtimer.c new_base->cpu_base == &__get_cpu_var(hrtimer_bases)); __get_cpu_var 1096 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases); __get_cpu_var 1284 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases); __get_cpu_var 1354 kernel/hrtimer.c run_hrtimer_pending(&__get_cpu_var(hrtimer_bases)); __get_cpu_var 1368 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases); __get_cpu_var 1393 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases); __get_cpu_var 276 kernel/kprobes.c __get_cpu_var(kprobe_instance) = kp; __get_cpu_var 281 kernel/kprobes.c __get_cpu_var(kprobe_instance) = NULL; __get_cpu_var 340 kernel/kprobes.c struct kprobe *cur = __get_cpu_var(kprobe_instance); __get_cpu_var 355 kernel/kprobes.c struct kprobe *cur = __get_cpu_var(kprobe_instance); __get_cpu_var 266 kernel/rcuclassic.c __call_rcu(head, &rcu_ctrlblk, &__get_cpu_var(rcu_data)); __get_cpu_var 294 kernel/rcuclassic.c __call_rcu(head, &rcu_bh_ctrlblk, &__get_cpu_var(rcu_bh_data)); __get_cpu_var 594 kernel/rcuclassic.c __rcu_process_callbacks(&rcu_ctrlblk, &__get_cpu_var(rcu_data)); __get_cpu_var 595 kernel/rcuclassic.c __rcu_process_callbacks(&rcu_bh_ctrlblk, &__get_cpu_var(rcu_bh_data)); __get_cpu_var 202 kernel/rcupreempt.c #define RCU_DATA_ME() (&__get_cpu_var(rcu_data)) __get_cpu_var 667 kernel/rcutorture.c ++__get_cpu_var(rcu_torture_count)[pipe_count]; __get_cpu_var 673 kernel/rcutorture.c ++__get_cpu_var(rcu_torture_batch)[completed]; __get_cpu_var 722 kernel/rcutorture.c ++__get_cpu_var(rcu_torture_count)[pipe_count]; __get_cpu_var 728 kernel/rcutorture.c ++__get_cpu_var(rcu_torture_batch)[completed]; __get_cpu_var 637 kernel/sched.c #define this_rq() (&__get_cpu_var(runqueues)) __get_cpu_var 64 kernel/sched_clock.c return &__get_cpu_var(sched_clock_data); __get_cpu_var 977 kernel/sched_rt.c cpumask_t *lowest_mask = &__get_cpu_var(local_cpu_mask); __get_cpu_var 153 kernel/smp.c struct call_single_queue *q = &__get_cpu_var(call_single_queue); __get_cpu_var 62 kernel/softirq.c struct task_struct *tsk = __get_cpu_var(ksoftirqd); __get_cpu_var 345 kernel/softirq.c *__get_cpu_var(tasklet_vec).tail = t; __get_cpu_var 346 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = &(t->next); __get_cpu_var 359 kernel/softirq.c *__get_cpu_var(tasklet_hi_vec).tail = t; __get_cpu_var 360 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = &(t->next); __get_cpu_var 372 kernel/softirq.c list = __get_cpu_var(tasklet_vec).head; __get_cpu_var 373 kernel/softirq.c __get_cpu_var(tasklet_vec).head = NULL; __get_cpu_var 374 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = &__get_cpu_var(tasklet_vec).head; __get_cpu_var 395 kernel/softirq.c *__get_cpu_var(tasklet_vec).tail = t; __get_cpu_var 396 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = &(t->next); __get_cpu_var 407 kernel/softirq.c list = __get_cpu_var(tasklet_hi_vec).head; __get_cpu_var 408 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).head = NULL; __get_cpu_var 409 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = &__get_cpu_var(tasklet_hi_vec).head; __get_cpu_var 430 kernel/softirq.c *__get_cpu_var(tasklet_hi_vec).tail = t; __get_cpu_var 431 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = &(t->next); __get_cpu_var 564 kernel/softirq.c *(__get_cpu_var(tasklet_vec).tail) = per_cpu(tasklet_vec, cpu).head; __get_cpu_var 565 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = per_cpu(tasklet_vec, cpu).tail; __get_cpu_var 572 kernel/softirq.c *__get_cpu_var(tasklet_hi_vec).tail = per_cpu(tasklet_hi_vec, cpu).head; __get_cpu_var 573 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = per_cpu(tasklet_hi_vec, cpu).tail; __get_cpu_var 52 kernel/time/tick-common.c struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev; __get_cpu_var 312 kernel/time/tick-common.c struct tick_device *td = &__get_cpu_var(tick_cpu_device); __get_cpu_var 322 kernel/time/tick-common.c struct tick_device *td = &__get_cpu_var(tick_cpu_device); __get_cpu_var 70 kernel/time/tick-oneshot.c struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev; __get_cpu_var 80 kernel/time/tick-oneshot.c struct tick_device *td = &__get_cpu_var(tick_cpu_device); __get_cpu_var 104 kernel/time/tick-oneshot.c struct tick_device *td = &__get_cpu_var(tick_cpu_device); __get_cpu_var 218 kernel/time/tick-sched.c struct clock_event_device *dev = __get_cpu_var(tick_cpu_device).evtdev; __get_cpu_var 375 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched); __get_cpu_var 469 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched); __get_cpu_var 521 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched); __get_cpu_var 628 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched); __get_cpu_var 693 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched); __get_cpu_var 708 kernel/time/tick-sched.c struct tick_sched *ts = &__get_cpu_var(tick_cpu_sched); __get_cpu_var 541 kernel/timer.c new_base = __get_cpu_var(tvec_bases); __get_cpu_var 938 kernel/timer.c struct tvec_base *base = __get_cpu_var(tvec_bases); __get_cpu_var 1031 kernel/timer.c struct tvec_base *base = __get_cpu_var(tvec_bases); __get_cpu_var 160 lib/radix-tree.c rtp = &__get_cpu_var(radix_tree_preloads); __get_cpu_var 211 lib/radix-tree.c rtp = &__get_cpu_var(radix_tree_preloads); __get_cpu_var 218 lib/radix-tree.c rtp = &__get_cpu_var(radix_tree_preloads); __get_cpu_var 573 mm/page-writeback.c p = &__get_cpu_var(ratelimits); __get_cpu_var 913 mm/slab.c int node = __get_cpu_var(reap_node); __get_cpu_var 918 mm/slab.c __get_cpu_var(reap_node) = node; __get_cpu_var 1091 mm/slab.c int node = __get_cpu_var(reap_node); __get_cpu_var 145 mm/swap.c pvec = &__get_cpu_var(lru_rotate_pvecs); __get_cpu_var 504 mm/swap.c local = &__get_cpu_var(committed_space); __get_cpu_var 832 mm/vmstat.c schedule_delayed_work(&__get_cpu_var(vmstat_work), __get_cpu_var 1380 net/core/dev.c sd = &__get_cpu_var(softnet_data); __get_cpu_var 1401 net/core/dev.c sd = &__get_cpu_var(softnet_data); __get_cpu_var 1946 net/core/dev.c queue = &__get_cpu_var(softnet_data); __get_cpu_var 1948 net/core/dev.c __get_cpu_var(netdev_rx_stat).total++; __get_cpu_var 1961 net/core/dev.c __get_cpu_var(netdev_rx_stat).dropped++; __get_cpu_var 1985 net/core/dev.c struct softnet_data *sd = &__get_cpu_var(softnet_data); __get_cpu_var 2240 net/core/dev.c __get_cpu_var(netdev_rx_stat).total++; __get_cpu_var 2315 net/core/dev.c struct softnet_data *queue = &__get_cpu_var(softnet_data); __get_cpu_var 2328 net/core/dev.c struct softnet_data *queue = &__get_cpu_var(softnet_data); __get_cpu_var 2361 net/core/dev.c list_add_tail(&n->poll_list, &__get_cpu_var(softnet_data).poll_list); __get_cpu_var 2370 net/core/dev.c struct list_head *list = &__get_cpu_var(softnet_data).poll_list; __get_cpu_var 2455 net/core/dev.c __get_cpu_var(netdev_rx_stat).time_squeeze++; __get_cpu_var 1994 net/core/sock.c __get_cpu_var(prot_inuse).val[prot->inuse_idx] += val; __get_cpu_var 45 net/ipv4/syncookies.c __u32 *tmp = __get_cpu_var(cookie_scratch); __get_cpu_var 1318 net/ipv4/tcp.c __get_cpu_var(softnet_data).net_dma) { __get_cpu_var 82 net/ipv6/syncookies.c __u32 *tmp = __get_cpu_var(cookie_scratch); __get_cpu_var 98 net/sched/sch_generic.c __get_cpu_var(netdev_rx_stat).cpu_collision++;