trace 319 arch/x86/kernel/cpu/intel_cacheinfo.c unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0; /* Cache sizes */ trace 424 arch/x86/kernel/cpu/intel_cacheinfo.c trace += cache_table[k].size; trace 457 arch/x86/kernel/cpu/intel_cacheinfo.c if (trace) trace 458 arch/x86/kernel/cpu/intel_cacheinfo.c printk (KERN_INFO "CPU: Trace cache: %dK uops", trace); trace 27 arch/x86/kernel/stacktrace.c struct stack_trace *trace = data; trace 30 arch/x86/kernel/stacktrace.c if (trace->skip > 0) { trace 31 arch/x86/kernel/stacktrace.c trace->skip--; trace 34 arch/x86/kernel/stacktrace.c if (trace->nr_entries < trace->max_entries) trace 35 arch/x86/kernel/stacktrace.c trace->entries[trace->nr_entries++] = addr; trace 41 arch/x86/kernel/stacktrace.c struct stack_trace *trace = (struct stack_trace *)data; trace 46 arch/x86/kernel/stacktrace.c if (trace->skip > 0) { trace 47 arch/x86/kernel/stacktrace.c trace->skip--; trace 50 arch/x86/kernel/stacktrace.c if (trace->nr_entries < trace->max_entries) trace 51 arch/x86/kernel/stacktrace.c trace->entries[trace->nr_entries++] = addr; trace 73 arch/x86/kernel/stacktrace.c dump_trace(current, NULL, NULL, 0, &save_stack_ops, trace); trace 74 arch/x86/kernel/stacktrace.c if (trace->nr_entries < trace->max_entries) trace 75 arch/x86/kernel/stacktrace.c trace->entries[trace->nr_entries++] = ULONG_MAX; trace 81 arch/x86/kernel/stacktrace.c dump_trace(tsk, NULL, NULL, 0, &save_stack_ops_nosched, trace); trace 82 arch/x86/kernel/stacktrace.c if (trace->nr_entries < trace->max_entries) trace 83 arch/x86/kernel/stacktrace.c trace->entries[trace->nr_entries++] = ULONG_MAX; trace 194 arch/x86/mm/mmio-mod.c struct remap_trace *trace = p->private; trace 206 arch/x86/mm/mmio-mod.c my_trace->phys = addr - trace->probe.addr + trace->phys; trace 207 arch/x86/mm/mmio-mod.c my_trace->map_id = trace->id; trace 282 arch/x86/mm/mmio-mod.c struct remap_trace *trace = kmalloc(sizeof(*trace), GFP_KERNEL); trace 291 arch/x86/mm/mmio-mod.c if (!trace) { trace 296 arch/x86/mm/mmio-mod.c *trace = (struct remap_trace) { trace 302 arch/x86/mm/mmio-mod.c .private = trace trace 307 arch/x86/mm/mmio-mod.c map.map_id = trace->id; trace 314 arch/x86/mm/mmio-mod.c list_add_tail(&trace->list, &trace_list); trace 316 arch/x86/mm/mmio-mod.c register_kmmio_probe(&trace->probe); trace 343 arch/x86/mm/mmio-mod.c struct remap_trace *trace; trace 353 arch/x86/mm/mmio-mod.c list_for_each_entry_safe(trace, tmp, &trace_list, list) { trace 354 arch/x86/mm/mmio-mod.c if ((unsigned long)addr == trace->probe.addr) { trace 356 arch/x86/mm/mmio-mod.c unregister_kmmio_probe(&trace->probe); trace 357 arch/x86/mm/mmio-mod.c list_del(&trace->list); trace 358 arch/x86/mm/mmio-mod.c found_trace = trace; trace 382 arch/x86/mm/mmio-mod.c struct remap_trace *trace; trace 391 arch/x86/mm/mmio-mod.c list_for_each_entry(trace, &trace_list, list) { trace 394 arch/x86/mm/mmio-mod.c trace->probe.addr, trace->probe.len); trace 396 arch/x86/mm/mmio-mod.c unregister_kmmio_probe(&trace->probe); trace 400 arch/x86/mm/mmio-mod.c list_for_each_entry_safe(trace, tmp, &trace_list, list) { trace 401 arch/x86/mm/mmio-mod.c list_del(&trace->list); trace 402 arch/x86/mm/mmio-mod.c kfree(trace); trace 218 include/asm-x86/processor.h unsigned short trace; trace 178 include/linux/lockdep.h struct stack_trace trace; trace 257 include/linux/tracehook.h ptrace_init_task(child, (clone_flags & CLONE_PTRACE) || trace); trace 283 include/linux/tracehook.h if (unlikely(trace) || unlikely(clone_flags & CLONE_PTRACE)) { trace 314 include/linux/tracehook.h if (unlikely(trace)) trace 315 include/linux/tracehook.h ptrace_event(0, trace, pid); trace 51 kernel/backtracetest.c struct stack_trace trace; trace 57 kernel/backtracetest.c trace.nr_entries = 0; trace 58 kernel/backtracetest.c trace.max_entries = ARRAY_SIZE(entries); trace 59 kernel/backtracetest.c trace.entries = entries; trace 60 kernel/backtracetest.c trace.skip = 0; trace 62 kernel/backtracetest.c save_stack_trace(&trace); trace 63 kernel/backtracetest.c print_stack_trace(&trace, 0); trace 1225 kernel/fork.c tracehook_finish_clone(p, clone_flags, trace); trace 1332 kernel/fork.c int trace = 0; trace 1356 kernel/fork.c trace = tracehook_prepare_clone(clone_flags); trace 1359 kernel/fork.c child_tidptr, NULL, trace); trace 1377 kernel/fork.c tracehook_report_clone(trace, regs, clone_flags, nr, p); trace 1398 kernel/fork.c tracehook_report_clone_complete(trace, regs, trace 106 kernel/latencytop.c struct stack_trace trace; trace 108 kernel/latencytop.c memset(&trace, 0, sizeof(trace)); trace 109 kernel/latencytop.c trace.max_entries = LT_BACKTRACEDEPTH; trace 110 kernel/latencytop.c trace.entries = &lat->backtrace[0]; trace 111 kernel/latencytop.c trace.skip = 0; trace 112 kernel/latencytop.c save_stack_trace_tsk(tsk, &trace); trace 352 kernel/lockdep.c trace->nr_entries = 0; trace 353 kernel/lockdep.c trace->max_entries = MAX_STACK_TRACE_ENTRIES - nr_stack_trace_entries; trace 354 kernel/lockdep.c trace->entries = stack_trace + nr_stack_trace_entries; trace 356 kernel/lockdep.c trace->skip = 3; trace 358 kernel/lockdep.c save_stack_trace(trace); trace 360 kernel/lockdep.c trace->max_entries = trace->nr_entries; trace 362 kernel/lockdep.c nr_stack_trace_entries += trace->nr_entries; trace 598 kernel/lockdep.c print_stack_trace(&entry->trace, 2); trace 878 kernel/lockdep.c if (!save_trace(&entry->trace)) trace 918 kernel/lockdep.c print_stack_trace(&target->trace, 6); trace 961 kernel/lockdep.c if (!save_trace(&this.trace)) trace 90 kernel/rcupreempt.c struct rcupreempt_trace trace; trace 209 kernel/rcupreempt.c #define RCU_TRACE_CPU(f, cpu) RCU_TRACE(f, &(RCU_DATA_CPU(cpu)->trace)); trace 215 kernel/rcupreempt.c #define RCU_TRACE_ME(f) RCU_TRACE(f, &(RCU_DATA_ME()->trace)); trace 221 kernel/rcupreempt.c #define RCU_TRACE_RDP(f, rdp) RCU_TRACE(f, &((rdp)->trace)); trace 1478 kernel/rcupreempt.c return &rdp->trace; trace 52 kernel/rcupreempt_trace.c trace->done_length += trace->wait_length; trace 53 kernel/rcupreempt_trace.c trace->done_add += trace->wait_length; trace 54 kernel/rcupreempt_trace.c trace->wait_length = 0; trace 58 kernel/rcupreempt_trace.c trace->wait_length += trace->next_length; trace 59 kernel/rcupreempt_trace.c trace->wait_add += trace->next_length; trace 60 kernel/rcupreempt_trace.c trace->next_length = 0; trace 64 kernel/rcupreempt_trace.c atomic_inc(&trace->rcu_try_flip_1); trace 68 kernel/rcupreempt_trace.c atomic_inc(&trace->rcu_try_flip_e1); trace 72 kernel/rcupreempt_trace.c trace->rcu_try_flip_i1++; trace 76 kernel/rcupreempt_trace.c trace->rcu_try_flip_ie1++; trace 80 kernel/rcupreempt_trace.c trace->rcu_try_flip_g1++; trace 84 kernel/rcupreempt_trace.c trace->rcu_try_flip_a1++; trace 88 kernel/rcupreempt_trace.c trace->rcu_try_flip_ae1++; trace 92 kernel/rcupreempt_trace.c trace->rcu_try_flip_a2++; trace 96 kernel/rcupreempt_trace.c trace->rcu_try_flip_z1++; trace 100 kernel/rcupreempt_trace.c trace->rcu_try_flip_ze1++; trace 104 kernel/rcupreempt_trace.c trace->rcu_try_flip_z2++; trace 108 kernel/rcupreempt_trace.c trace->rcu_try_flip_m1++; trace 112 kernel/rcupreempt_trace.c trace->rcu_try_flip_me1++; trace 116 kernel/rcupreempt_trace.c trace->rcu_try_flip_m2++; trace 120 kernel/rcupreempt_trace.c trace->rcu_check_callbacks++; trace 124 kernel/rcupreempt_trace.c trace->done_remove += trace->done_length; trace 125 kernel/rcupreempt_trace.c trace->done_length = 0; trace 129 kernel/rcupreempt_trace.c atomic_inc(&trace->done_invoked); trace 133 kernel/rcupreempt_trace.c trace->next_add++; trace 134 kernel/rcupreempt_trace.c trace->next_length++; trace 176 kernel/rcupreempt_trace.c struct rcupreempt_trace trace; trace 180 kernel/rcupreempt_trace.c rcupreempt_trace_sum(&trace); trace 185 kernel/rcupreempt_trace.c trace.rcu_check_callbacks); trace 191 kernel/rcupreempt_trace.c trace.next_add, trace.next_length, trace 192 kernel/rcupreempt_trace.c trace.wait_add, trace.wait_length, trace 193 kernel/rcupreempt_trace.c trace.done_add, trace.done_length, trace 194 kernel/rcupreempt_trace.c trace.done_remove, atomic_read(&trace.done_invoked), trace 195 kernel/rcupreempt_trace.c atomic_read(&trace.rcu_try_flip_1), trace 196 kernel/rcupreempt_trace.c atomic_read(&trace.rcu_try_flip_e1), trace 197 kernel/rcupreempt_trace.c trace.rcu_try_flip_i1, trace.rcu_try_flip_ie1, trace 198 kernel/rcupreempt_trace.c trace.rcu_try_flip_g1, trace 199 kernel/rcupreempt_trace.c trace.rcu_try_flip_a1, trace.rcu_try_flip_ae1, trace 200 kernel/rcupreempt_trace.c trace.rcu_try_flip_a2, trace 201 kernel/rcupreempt_trace.c trace.rcu_try_flip_z1, trace.rcu_try_flip_ze1, trace 202 kernel/rcupreempt_trace.c trace.rcu_try_flip_z2, trace 203 kernel/rcupreempt_trace.c trace.rcu_try_flip_m1, trace.rcu_try_flip_me1, trace 204 kernel/rcupreempt_trace.c trace.rcu_try_flip_m2); trace 17 kernel/stacktrace.c if (WARN_ON(!trace->entries)) trace 20 kernel/stacktrace.c for (i = 0; i < trace->nr_entries; i++) { trace 22 kernel/stacktrace.c print_ip_sym(trace->entries[i]); trace 903 kernel/trace/trace.c struct stack_trace trace; trace 914 kernel/trace/trace.c trace.nr_entries = 0; trace 915 kernel/trace/trace.c trace.max_entries = FTRACE_STACK_ENTRIES; trace 916 kernel/trace/trace.c trace.skip = skip; trace 917 kernel/trace/trace.c trace.entries = entry->stack.caller; trace 919 kernel/trace/trace.c save_stack_trace(&trace); trace 1217 kernel/trace/trace.c if (!current_trace || current_trace != iter->trace) { trace 1258 kernel/trace/trace.c if (current_trace && current_trace == iter->trace && iter->trace->stop) trace 1259 kernel/trace/trace.c iter->trace->stop(iter); trace 1815 kernel/trace/trace.c if (iter->trace && iter->trace->print_line) trace 1816 kernel/trace/trace.c return iter->trace->print_line(iter); trace 1839 kernel/trace/trace.c seq_printf(m, "# tracer: %s\n", iter->trace->name); trace 1889 kernel/trace/trace.c iter->trace = current_trace; trace 1904 kernel/trace/trace.c if (iter->trace && iter->trace->open) trace 1905 kernel/trace/trace.c iter->trace->open(iter); trace 1931 kernel/trace/trace.c if (iter->trace && iter->trace->close) trace 1932 kernel/trace/trace.c iter->trace->close(iter); trace 2454 kernel/trace/trace.c iter->trace = current_trace; trace 2457 kernel/trace/trace.c if (iter->trace->pipe_open) trace 2458 kernel/trace/trace.c iter->trace->pipe_open(iter); trace 2521 kernel/trace/trace.c if (iter->trace->read) { trace 2522 kernel/trace/trace.c sret = iter->trace->read(iter, filp, ubuf, cnt, ppos); trace 2560 kernel/trace/trace.c if (iter->trace != current_trace) trace 152 kernel/trace/trace.h int (*selftest)(struct tracer *trace, trace 172 kernel/trace/trace.h struct tracer *trace; trace 422 kernel/trace/trace_irqsoff.c # define register_irqsoff(trace) register_tracer(&trace) trace 448 kernel/trace/trace_irqsoff.c # define register_preemptoff(trace) register_tracer(&trace) trace 477 kernel/trace/trace_irqsoff.c # define register_preemptirqsoff(trace) register_tracer(&trace) trace 159 kernel/trace/trace_selftest.c trace->init(tr); trace 182 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 187 kernel/trace/trace_selftest.c trace->reset(tr); trace 235 kernel/trace/trace_selftest.c trace->init(tr); trace 240 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 245 kernel/trace/trace_selftest.c trace->reset(tr); trace 253 kernel/trace/trace_selftest.c ret = trace_selftest_startup_dynamic_tracing(trace, tr, trace 278 kernel/trace/trace_selftest.c trace->init(tr); trace 287 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 292 kernel/trace/trace_selftest.c trace->reset(tr); trace 315 kernel/trace/trace_selftest.c trace->init(tr); trace 324 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 329 kernel/trace/trace_selftest.c trace->reset(tr); trace 352 kernel/trace/trace_selftest.c trace->init(tr); trace 367 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 386 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 396 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 411 kernel/trace/trace_selftest.c trace->reset(tr); trace 469 kernel/trace/trace_selftest.c trace->init(tr); trace 491 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 498 kernel/trace/trace_selftest.c trace->reset(tr); trace 523 kernel/trace/trace_selftest.c trace->init(tr); trace 528 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 531 kernel/trace/trace_selftest.c trace->reset(tr); trace 551 kernel/trace/trace_selftest.c trace->init(tr); trace 556 kernel/trace/trace_selftest.c trace->ctrl_update(tr); trace 559 kernel/trace/trace_selftest.c trace->reset(tr); trace 62 lib/fault-inject.c struct stack_trace trace; trace 71 lib/fault-inject.c trace.nr_entries = 0; trace 72 lib/fault-inject.c trace.entries = entries; trace 73 lib/fault-inject.c trace.max_entries = depth; trace 74 lib/fault-inject.c trace.skip = 1; trace 76 lib/fault-inject.c save_stack_trace(&trace); trace 77 lib/fault-inject.c for (n = 0; n < trace.nr_entries; n++) { trace 889 mm/slub.c trace(s, page, object, 1); trace 946 mm/slub.c trace(s, page, object, 0);