trace 319 arch/x86/kernel/cpu/intel_cacheinfo.c unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0; /* Cache sizes */
trace 424 arch/x86/kernel/cpu/intel_cacheinfo.c trace += cache_table[k].size;
trace 457 arch/x86/kernel/cpu/intel_cacheinfo.c if (trace)
trace 458 arch/x86/kernel/cpu/intel_cacheinfo.c printk (KERN_INFO "CPU: Trace cache: %dK uops", trace);
trace 27 arch/x86/kernel/stacktrace.c struct stack_trace *trace = data;
trace 30 arch/x86/kernel/stacktrace.c if (trace->skip > 0) {
trace 31 arch/x86/kernel/stacktrace.c trace->skip--;
trace 34 arch/x86/kernel/stacktrace.c if (trace->nr_entries < trace->max_entries)
trace 35 arch/x86/kernel/stacktrace.c trace->entries[trace->nr_entries++] = addr;
trace 41 arch/x86/kernel/stacktrace.c struct stack_trace *trace = (struct stack_trace *)data;
trace 46 arch/x86/kernel/stacktrace.c if (trace->skip > 0) {
trace 47 arch/x86/kernel/stacktrace.c trace->skip--;
trace 50 arch/x86/kernel/stacktrace.c if (trace->nr_entries < trace->max_entries)
trace 51 arch/x86/kernel/stacktrace.c trace->entries[trace->nr_entries++] = addr;
trace 73 arch/x86/kernel/stacktrace.c dump_trace(current, NULL, NULL, 0, &save_stack_ops, trace);
trace 74 arch/x86/kernel/stacktrace.c if (trace->nr_entries < trace->max_entries)
trace 75 arch/x86/kernel/stacktrace.c trace->entries[trace->nr_entries++] = ULONG_MAX;
trace 81 arch/x86/kernel/stacktrace.c dump_trace(tsk, NULL, NULL, 0, &save_stack_ops_nosched, trace);
trace 82 arch/x86/kernel/stacktrace.c if (trace->nr_entries < trace->max_entries)
trace 83 arch/x86/kernel/stacktrace.c trace->entries[trace->nr_entries++] = ULONG_MAX;
trace 194 arch/x86/mm/mmio-mod.c struct remap_trace *trace = p->private;
trace 206 arch/x86/mm/mmio-mod.c my_trace->phys = addr - trace->probe.addr + trace->phys;
trace 207 arch/x86/mm/mmio-mod.c my_trace->map_id = trace->id;
trace 282 arch/x86/mm/mmio-mod.c struct remap_trace *trace = kmalloc(sizeof(*trace), GFP_KERNEL);
trace 291 arch/x86/mm/mmio-mod.c if (!trace) {
trace 296 arch/x86/mm/mmio-mod.c *trace = (struct remap_trace) {
trace 302 arch/x86/mm/mmio-mod.c .private = trace
trace 307 arch/x86/mm/mmio-mod.c map.map_id = trace->id;
trace 314 arch/x86/mm/mmio-mod.c list_add_tail(&trace->list, &trace_list);
trace 316 arch/x86/mm/mmio-mod.c register_kmmio_probe(&trace->probe);
trace 343 arch/x86/mm/mmio-mod.c struct remap_trace *trace;
trace 353 arch/x86/mm/mmio-mod.c list_for_each_entry_safe(trace, tmp, &trace_list, list) {
trace 354 arch/x86/mm/mmio-mod.c if ((unsigned long)addr == trace->probe.addr) {
trace 356 arch/x86/mm/mmio-mod.c unregister_kmmio_probe(&trace->probe);
trace 357 arch/x86/mm/mmio-mod.c list_del(&trace->list);
trace 358 arch/x86/mm/mmio-mod.c found_trace = trace;
trace 382 arch/x86/mm/mmio-mod.c struct remap_trace *trace;
trace 391 arch/x86/mm/mmio-mod.c list_for_each_entry(trace, &trace_list, list) {
trace 394 arch/x86/mm/mmio-mod.c trace->probe.addr, trace->probe.len);
trace 396 arch/x86/mm/mmio-mod.c unregister_kmmio_probe(&trace->probe);
trace 400 arch/x86/mm/mmio-mod.c list_for_each_entry_safe(trace, tmp, &trace_list, list) {
trace 401 arch/x86/mm/mmio-mod.c list_del(&trace->list);
trace 402 arch/x86/mm/mmio-mod.c kfree(trace);
trace 218 include/asm-x86/processor.h unsigned short trace;
trace 178 include/linux/lockdep.h struct stack_trace trace;
trace 257 include/linux/tracehook.h ptrace_init_task(child, (clone_flags & CLONE_PTRACE) || trace);
trace 283 include/linux/tracehook.h if (unlikely(trace) || unlikely(clone_flags & CLONE_PTRACE)) {
trace 314 include/linux/tracehook.h if (unlikely(trace))
trace 315 include/linux/tracehook.h ptrace_event(0, trace, pid);
trace 51 kernel/backtracetest.c struct stack_trace trace;
trace 57 kernel/backtracetest.c trace.nr_entries = 0;
trace 58 kernel/backtracetest.c trace.max_entries = ARRAY_SIZE(entries);
trace 59 kernel/backtracetest.c trace.entries = entries;
trace 60 kernel/backtracetest.c trace.skip = 0;
trace 62 kernel/backtracetest.c save_stack_trace(&trace);
trace 63 kernel/backtracetest.c print_stack_trace(&trace, 0);
trace 1225 kernel/fork.c tracehook_finish_clone(p, clone_flags, trace);
trace 1332 kernel/fork.c int trace = 0;
trace 1356 kernel/fork.c trace = tracehook_prepare_clone(clone_flags);
trace 1359 kernel/fork.c child_tidptr, NULL, trace);
trace 1377 kernel/fork.c tracehook_report_clone(trace, regs, clone_flags, nr, p);
trace 1398 kernel/fork.c tracehook_report_clone_complete(trace, regs,
trace 106 kernel/latencytop.c struct stack_trace trace;
trace 108 kernel/latencytop.c memset(&trace, 0, sizeof(trace));
trace 109 kernel/latencytop.c trace.max_entries = LT_BACKTRACEDEPTH;
trace 110 kernel/latencytop.c trace.entries = &lat->backtrace[0];
trace 111 kernel/latencytop.c trace.skip = 0;
trace 112 kernel/latencytop.c save_stack_trace_tsk(tsk, &trace);
trace 352 kernel/lockdep.c trace->nr_entries = 0;
trace 353 kernel/lockdep.c trace->max_entries = MAX_STACK_TRACE_ENTRIES - nr_stack_trace_entries;
trace 354 kernel/lockdep.c trace->entries = stack_trace + nr_stack_trace_entries;
trace 356 kernel/lockdep.c trace->skip = 3;
trace 358 kernel/lockdep.c save_stack_trace(trace);
trace 360 kernel/lockdep.c trace->max_entries = trace->nr_entries;
trace 362 kernel/lockdep.c nr_stack_trace_entries += trace->nr_entries;
trace 598 kernel/lockdep.c print_stack_trace(&entry->trace, 2);
trace 878 kernel/lockdep.c if (!save_trace(&entry->trace))
trace 918 kernel/lockdep.c print_stack_trace(&target->trace, 6);
trace 961 kernel/lockdep.c if (!save_trace(&this.trace))
trace 90 kernel/rcupreempt.c struct rcupreempt_trace trace;
trace 209 kernel/rcupreempt.c #define RCU_TRACE_CPU(f, cpu) RCU_TRACE(f, &(RCU_DATA_CPU(cpu)->trace));
trace 215 kernel/rcupreempt.c #define RCU_TRACE_ME(f) RCU_TRACE(f, &(RCU_DATA_ME()->trace));
trace 221 kernel/rcupreempt.c #define RCU_TRACE_RDP(f, rdp) RCU_TRACE(f, &((rdp)->trace));
trace 1478 kernel/rcupreempt.c return &rdp->trace;
trace 52 kernel/rcupreempt_trace.c trace->done_length += trace->wait_length;
trace 53 kernel/rcupreempt_trace.c trace->done_add += trace->wait_length;
trace 54 kernel/rcupreempt_trace.c trace->wait_length = 0;
trace 58 kernel/rcupreempt_trace.c trace->wait_length += trace->next_length;
trace 59 kernel/rcupreempt_trace.c trace->wait_add += trace->next_length;
trace 60 kernel/rcupreempt_trace.c trace->next_length = 0;
trace 64 kernel/rcupreempt_trace.c atomic_inc(&trace->rcu_try_flip_1);
trace 68 kernel/rcupreempt_trace.c atomic_inc(&trace->rcu_try_flip_e1);
trace 72 kernel/rcupreempt_trace.c trace->rcu_try_flip_i1++;
trace 76 kernel/rcupreempt_trace.c trace->rcu_try_flip_ie1++;
trace 80 kernel/rcupreempt_trace.c trace->rcu_try_flip_g1++;
trace 84 kernel/rcupreempt_trace.c trace->rcu_try_flip_a1++;
trace 88 kernel/rcupreempt_trace.c trace->rcu_try_flip_ae1++;
trace 92 kernel/rcupreempt_trace.c trace->rcu_try_flip_a2++;
trace 96 kernel/rcupreempt_trace.c trace->rcu_try_flip_z1++;
trace 100 kernel/rcupreempt_trace.c trace->rcu_try_flip_ze1++;
trace 104 kernel/rcupreempt_trace.c trace->rcu_try_flip_z2++;
trace 108 kernel/rcupreempt_trace.c trace->rcu_try_flip_m1++;
trace 112 kernel/rcupreempt_trace.c trace->rcu_try_flip_me1++;
trace 116 kernel/rcupreempt_trace.c trace->rcu_try_flip_m2++;
trace 120 kernel/rcupreempt_trace.c trace->rcu_check_callbacks++;
trace 124 kernel/rcupreempt_trace.c trace->done_remove += trace->done_length;
trace 125 kernel/rcupreempt_trace.c trace->done_length = 0;
trace 129 kernel/rcupreempt_trace.c atomic_inc(&trace->done_invoked);
trace 133 kernel/rcupreempt_trace.c trace->next_add++;
trace 134 kernel/rcupreempt_trace.c trace->next_length++;
trace 176 kernel/rcupreempt_trace.c struct rcupreempt_trace trace;
trace 180 kernel/rcupreempt_trace.c rcupreempt_trace_sum(&trace);
trace 185 kernel/rcupreempt_trace.c trace.rcu_check_callbacks);
trace 191 kernel/rcupreempt_trace.c trace.next_add, trace.next_length,
trace 192 kernel/rcupreempt_trace.c trace.wait_add, trace.wait_length,
trace 193 kernel/rcupreempt_trace.c trace.done_add, trace.done_length,
trace 194 kernel/rcupreempt_trace.c trace.done_remove, atomic_read(&trace.done_invoked),
trace 195 kernel/rcupreempt_trace.c atomic_read(&trace.rcu_try_flip_1),
trace 196 kernel/rcupreempt_trace.c atomic_read(&trace.rcu_try_flip_e1),
trace 197 kernel/rcupreempt_trace.c trace.rcu_try_flip_i1, trace.rcu_try_flip_ie1,
trace 198 kernel/rcupreempt_trace.c trace.rcu_try_flip_g1,
trace 199 kernel/rcupreempt_trace.c trace.rcu_try_flip_a1, trace.rcu_try_flip_ae1,
trace 200 kernel/rcupreempt_trace.c trace.rcu_try_flip_a2,
trace 201 kernel/rcupreempt_trace.c trace.rcu_try_flip_z1, trace.rcu_try_flip_ze1,
trace 202 kernel/rcupreempt_trace.c trace.rcu_try_flip_z2,
trace 203 kernel/rcupreempt_trace.c trace.rcu_try_flip_m1, trace.rcu_try_flip_me1,
trace 204 kernel/rcupreempt_trace.c trace.rcu_try_flip_m2);
trace 17 kernel/stacktrace.c if (WARN_ON(!trace->entries))
trace 20 kernel/stacktrace.c for (i = 0; i < trace->nr_entries; i++) {
trace 22 kernel/stacktrace.c print_ip_sym(trace->entries[i]);
trace 903 kernel/trace/trace.c struct stack_trace trace;
trace 914 kernel/trace/trace.c trace.nr_entries = 0;
trace 915 kernel/trace/trace.c trace.max_entries = FTRACE_STACK_ENTRIES;
trace 916 kernel/trace/trace.c trace.skip = skip;
trace 917 kernel/trace/trace.c trace.entries = entry->stack.caller;
trace 919 kernel/trace/trace.c save_stack_trace(&trace);
trace 1217 kernel/trace/trace.c if (!current_trace || current_trace != iter->trace) {
trace 1258 kernel/trace/trace.c if (current_trace && current_trace == iter->trace && iter->trace->stop)
trace 1259 kernel/trace/trace.c iter->trace->stop(iter);
trace 1815 kernel/trace/trace.c if (iter->trace && iter->trace->print_line)
trace 1816 kernel/trace/trace.c return iter->trace->print_line(iter);
trace 1839 kernel/trace/trace.c seq_printf(m, "# tracer: %s\n", iter->trace->name);
trace 1889 kernel/trace/trace.c iter->trace = current_trace;
trace 1904 kernel/trace/trace.c if (iter->trace && iter->trace->open)
trace 1905 kernel/trace/trace.c iter->trace->open(iter);
trace 1931 kernel/trace/trace.c if (iter->trace && iter->trace->close)
trace 1932 kernel/trace/trace.c iter->trace->close(iter);
trace 2454 kernel/trace/trace.c iter->trace = current_trace;
trace 2457 kernel/trace/trace.c if (iter->trace->pipe_open)
trace 2458 kernel/trace/trace.c iter->trace->pipe_open(iter);
trace 2521 kernel/trace/trace.c if (iter->trace->read) {
trace 2522 kernel/trace/trace.c sret = iter->trace->read(iter, filp, ubuf, cnt, ppos);
trace 2560 kernel/trace/trace.c if (iter->trace != current_trace)
trace 152 kernel/trace/trace.h int (*selftest)(struct tracer *trace,
trace 172 kernel/trace/trace.h struct tracer *trace;
trace 422 kernel/trace/trace_irqsoff.c # define register_irqsoff(trace) register_tracer(&trace)
trace 448 kernel/trace/trace_irqsoff.c # define register_preemptoff(trace) register_tracer(&trace)
trace 477 kernel/trace/trace_irqsoff.c # define register_preemptirqsoff(trace) register_tracer(&trace)
trace 159 kernel/trace/trace_selftest.c trace->init(tr);
trace 182 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 187 kernel/trace/trace_selftest.c trace->reset(tr);
trace 235 kernel/trace/trace_selftest.c trace->init(tr);
trace 240 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 245 kernel/trace/trace_selftest.c trace->reset(tr);
trace 253 kernel/trace/trace_selftest.c ret = trace_selftest_startup_dynamic_tracing(trace, tr,
trace 278 kernel/trace/trace_selftest.c trace->init(tr);
trace 287 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 292 kernel/trace/trace_selftest.c trace->reset(tr);
trace 315 kernel/trace/trace_selftest.c trace->init(tr);
trace 324 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 329 kernel/trace/trace_selftest.c trace->reset(tr);
trace 352 kernel/trace/trace_selftest.c trace->init(tr);
trace 367 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 386 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 396 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 411 kernel/trace/trace_selftest.c trace->reset(tr);
trace 469 kernel/trace/trace_selftest.c trace->init(tr);
trace 491 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 498 kernel/trace/trace_selftest.c trace->reset(tr);
trace 523 kernel/trace/trace_selftest.c trace->init(tr);
trace 528 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 531 kernel/trace/trace_selftest.c trace->reset(tr);
trace 551 kernel/trace/trace_selftest.c trace->init(tr);
trace 556 kernel/trace/trace_selftest.c trace->ctrl_update(tr);
trace 559 kernel/trace/trace_selftest.c trace->reset(tr);
trace 62 lib/fault-inject.c struct stack_trace trace;
trace 71 lib/fault-inject.c trace.nr_entries = 0;
trace 72 lib/fault-inject.c trace.entries = entries;
trace 73 lib/fault-inject.c trace.max_entries = depth;
trace 74 lib/fault-inject.c trace.skip = 1;
trace 76 lib/fault-inject.c save_stack_trace(&trace);
trace 77 lib/fault-inject.c for (n = 0; n < trace.nr_entries; n++) {
trace 889 mm/slub.c trace(s, page, object, 1);
trace 946 mm/slub.c trace(s, page, object, 0);