preempt_count 57 arch/x86/kernel/asm-offsets_32.c OFFSET(TI_preempt_count, thread_info, preempt_count); preempt_count 42 arch/x86/kernel/asm-offsets_64.c ENTRY(preempt_count); preempt_count 127 arch/x86/kernel/irq_32.c irqctx->tinfo.preempt_count = preempt_count 128 arch/x86/kernel/irq_32.c (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | preempt_count 129 arch/x86/kernel/irq_32.c (curctx->tinfo.preempt_count & SOFTIRQ_MASK); preempt_count 158 arch/x86/kernel/irq_32.c irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; preempt_count 167 arch/x86/kernel/irq_32.c irqctx->tinfo.preempt_count = 0; preempt_count 36 include/asm-cris/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 63 include/asm-cris/thread_info.h .preempt_count = 1, \ preempt_count 38 include/asm-frv/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 70 include/asm-frv/thread_info.h .preempt_count = 1, \ preempt_count 31 include/asm-m32r/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 71 include/asm-m32r/thread_info.h .preempt_count = 1, \ preempt_count 11 include/asm-m68k/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 47 include/asm-mn10300/thread_info.h __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 79 include/asm-mn10300/thread_info.h .preempt_count = 1, \ preempt_count 15 include/asm-parisc/thread_info.h int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ preempt_count 26 include/asm-parisc/thread_info.h .preempt_count = 1, \ preempt_count 20 include/asm-um/thread_info.h int preempt_count; /* 0 => preemptable, preempt_count 35 include/asm-um/thread_info.h .preempt_count = 1, \ preempt_count 30 include/asm-x86/thread_info.h int preempt_count; /* 0 => preemptable, preempt_count 49 include/asm-x86/thread_info.h .preempt_count = 1, \ preempt_count 51 include/asm-xtensa/thread_info.h __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ preempt_count 95 include/asm-xtensa/thread_info.h .preempt_count = 1, \ preempt_count 88 include/linux/bit_spinlock.h return preempt_count(); preempt_count 63 include/linux/hardirq.h #define hardirq_count() (preempt_count() & HARDIRQ_MASK) preempt_count 64 include/linux/hardirq.h #define softirq_count() (preempt_count() & SOFTIRQ_MASK) preempt_count 65 include/linux/hardirq.h #define irq_count() (preempt_count() & (HARDIRQ_MASK | SOFTIRQ_MASK)) preempt_count 90 include/linux/hardirq.h #define in_atomic() ((preempt_count() & ~PREEMPT_ACTIVE) != PREEMPT_INATOMIC_BASE) preempt_count 97 include/linux/hardirq.h ((preempt_count() & ~PREEMPT_ACTIVE) != PREEMPT_CHECK_OFFSET) preempt_count 100 include/linux/hardirq.h # define preemptible() (preempt_count() == 0 && !irqs_disabled()) preempt_count 17 include/linux/preempt.h # define add_preempt_count(val) do { preempt_count() += (val); } while (0) preempt_count 18 include/linux/preempt.h # define sub_preempt_count(val) do { preempt_count() -= (val); } while (0) preempt_count 24 include/linux/preempt.h #define preempt_count() (current_thread_info()->preempt_count) preempt_count 57 include/linux/preempt.h do { preempt_count() += (val); } while (0) preempt_count 59 include/linux/preempt.h do { preempt_count() -= (val); } while (0) preempt_count 705 init/main.c int count = preempt_count(); preempt_count 731 init/main.c if (preempt_count() != count) { preempt_count 733 init/main.c preempt_count() = count; preempt_count 1044 kernel/exit.c preempt_count()); preempt_count 2402 kernel/sched.c task_thread_info(p)->preempt_count = 1; preempt_count 4301 kernel/sched.c if (DEBUG_LOCKS_WARN_ON((preempt_count() < 0))) preempt_count 4304 kernel/sched.c preempt_count() += val; preempt_count 4309 kernel/sched.c DEBUG_LOCKS_WARN_ON((preempt_count() & PREEMPT_MASK) >= preempt_count 4312 kernel/sched.c if (preempt_count() == val) preempt_count 4323 kernel/sched.c if (DEBUG_LOCKS_WARN_ON(val > preempt_count())) preempt_count 4329 kernel/sched.c !(preempt_count() & PREEMPT_MASK))) preempt_count 4333 kernel/sched.c if (preempt_count() == val) preempt_count 4335 kernel/sched.c preempt_count() -= val; preempt_count 4349 kernel/sched.c prev->comm, prev->pid, preempt_count()); preempt_count 4452 kernel/sched.c if (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) { preempt_count 4511 kernel/sched.c if (likely(ti->preempt_count || irqs_disabled())) preempt_count 4539 kernel/sched.c BUG_ON(ti->preempt_count || !irqs_disabled()); preempt_count 5572 kernel/sched.c if (need_resched() && !(preempt_count() & PREEMPT_ACTIVE) && preempt_count 5875 kernel/sched.c task_thread_info(idle)->preempt_count = (idle->lock_depth >= 0); preempt_count 5877 kernel/sched.c task_thread_info(idle)->preempt_count = 0; preempt_count 797 kernel/timer.c int preempt_count = preempt_count(); preempt_count 799 kernel/timer.c if (preempt_count != preempt_count()) { preempt_count 803 kernel/timer.c fn, preempt_count, preempt_count 804 kernel/timer.c preempt_count()); preempt_count 818 kernel/trace/trace.c pc = preempt_count(); preempt_count 820 kernel/trace/trace.c entry->preempt_count = pc & 0xff; preempt_count 1451 kernel/trace/trace.c if (entry->preempt_count) preempt_count 1452 kernel/trace/trace.c trace_seq_printf(s, "%x", entry->preempt_count); preempt_count 1500 kernel/trace/trace.c entry->preempt_count, trace_idx, preempt_count 73 kernel/trace/trace.h char preempt_count; preempt_count 39 kernel/trace/trace_irqsoff.c return ((trace_type & TRACER_PREEMPT_OFF) && preempt_count()); preempt_count 302 kernel/workqueue.c current->comm, preempt_count(), preempt_count 71 lib/kernel_lock.c if (preempt_count() > 1) { preempt_count 941 lib/locking-selftest.c unsigned long saved_preempt_count = preempt_count(); preempt_count 983 lib/locking-selftest.c preempt_count() = saved_preempt_count; preempt_count 12 lib/smp_processor_id.c unsigned long preempt_count = preempt_count(); preempt_count 15 lib/smp_processor_id.c if (likely(preempt_count)) preempt_count 44 lib/smp_processor_id.c preempt_count() - 1, current->comm, current->pid); preempt_count 2994 mm/memory.c if (preempt_count())