trace             319 arch/x86/kernel/cpu/intel_cacheinfo.c 	unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0; /* Cache sizes */
trace             424 arch/x86/kernel/cpu/intel_cacheinfo.c 							trace += cache_table[k].size;
trace             457 arch/x86/kernel/cpu/intel_cacheinfo.c 	if (trace)
trace             458 arch/x86/kernel/cpu/intel_cacheinfo.c 		printk (KERN_INFO "CPU: Trace cache: %dK uops", trace);
trace              27 arch/x86/kernel/stacktrace.c 	struct stack_trace *trace = data;
trace              30 arch/x86/kernel/stacktrace.c 	if (trace->skip > 0) {
trace              31 arch/x86/kernel/stacktrace.c 		trace->skip--;
trace              34 arch/x86/kernel/stacktrace.c 	if (trace->nr_entries < trace->max_entries)
trace              35 arch/x86/kernel/stacktrace.c 		trace->entries[trace->nr_entries++] = addr;
trace              41 arch/x86/kernel/stacktrace.c 	struct stack_trace *trace = (struct stack_trace *)data;
trace              46 arch/x86/kernel/stacktrace.c 	if (trace->skip > 0) {
trace              47 arch/x86/kernel/stacktrace.c 		trace->skip--;
trace              50 arch/x86/kernel/stacktrace.c 	if (trace->nr_entries < trace->max_entries)
trace              51 arch/x86/kernel/stacktrace.c 		trace->entries[trace->nr_entries++] = addr;
trace              73 arch/x86/kernel/stacktrace.c 	dump_trace(current, NULL, NULL, 0, &save_stack_ops, trace);
trace              74 arch/x86/kernel/stacktrace.c 	if (trace->nr_entries < trace->max_entries)
trace              75 arch/x86/kernel/stacktrace.c 		trace->entries[trace->nr_entries++] = ULONG_MAX;
trace              81 arch/x86/kernel/stacktrace.c 	dump_trace(tsk, NULL, NULL, 0, &save_stack_ops_nosched, trace);
trace              82 arch/x86/kernel/stacktrace.c 	if (trace->nr_entries < trace->max_entries)
trace              83 arch/x86/kernel/stacktrace.c 		trace->entries[trace->nr_entries++] = ULONG_MAX;
trace             194 arch/x86/mm/mmio-mod.c 	struct remap_trace *trace = p->private;
trace             206 arch/x86/mm/mmio-mod.c 	my_trace->phys = addr - trace->probe.addr + trace->phys;
trace             207 arch/x86/mm/mmio-mod.c 	my_trace->map_id = trace->id;
trace             282 arch/x86/mm/mmio-mod.c 	struct remap_trace *trace = kmalloc(sizeof(*trace), GFP_KERNEL);
trace             291 arch/x86/mm/mmio-mod.c 	if (!trace) {
trace             296 arch/x86/mm/mmio-mod.c 	*trace = (struct remap_trace) {
trace             302 arch/x86/mm/mmio-mod.c 			.private = trace
trace             307 arch/x86/mm/mmio-mod.c 	map.map_id = trace->id;
trace             314 arch/x86/mm/mmio-mod.c 	list_add_tail(&trace->list, &trace_list);
trace             316 arch/x86/mm/mmio-mod.c 		register_kmmio_probe(&trace->probe);
trace             343 arch/x86/mm/mmio-mod.c 	struct remap_trace *trace;
trace             353 arch/x86/mm/mmio-mod.c 	list_for_each_entry_safe(trace, tmp, &trace_list, list) {
trace             354 arch/x86/mm/mmio-mod.c 		if ((unsigned long)addr == trace->probe.addr) {
trace             356 arch/x86/mm/mmio-mod.c 				unregister_kmmio_probe(&trace->probe);
trace             357 arch/x86/mm/mmio-mod.c 			list_del(&trace->list);
trace             358 arch/x86/mm/mmio-mod.c 			found_trace = trace;
trace             382 arch/x86/mm/mmio-mod.c 	struct remap_trace *trace;
trace             391 arch/x86/mm/mmio-mod.c 	list_for_each_entry(trace, &trace_list, list) {
trace             394 arch/x86/mm/mmio-mod.c 					trace->probe.addr, trace->probe.len);
trace             396 arch/x86/mm/mmio-mod.c 			unregister_kmmio_probe(&trace->probe);
trace             400 arch/x86/mm/mmio-mod.c 	list_for_each_entry_safe(trace, tmp, &trace_list, list) {
trace             401 arch/x86/mm/mmio-mod.c 		list_del(&trace->list);
trace             402 arch/x86/mm/mmio-mod.c 		kfree(trace);
trace             218 include/asm-x86/processor.h 	unsigned short		trace;
trace             178 include/linux/lockdep.h 	struct stack_trace		trace;
trace             257 include/linux/tracehook.h 	ptrace_init_task(child, (clone_flags & CLONE_PTRACE) || trace);
trace             283 include/linux/tracehook.h 	if (unlikely(trace) || unlikely(clone_flags & CLONE_PTRACE)) {
trace             314 include/linux/tracehook.h 	if (unlikely(trace))
trace             315 include/linux/tracehook.h 		ptrace_event(0, trace, pid);
trace              51 kernel/backtracetest.c 	struct stack_trace trace;
trace              57 kernel/backtracetest.c 	trace.nr_entries = 0;
trace              58 kernel/backtracetest.c 	trace.max_entries = ARRAY_SIZE(entries);
trace              59 kernel/backtracetest.c 	trace.entries = entries;
trace              60 kernel/backtracetest.c 	trace.skip = 0;
trace              62 kernel/backtracetest.c 	save_stack_trace(&trace);
trace              63 kernel/backtracetest.c 	print_stack_trace(&trace, 0);
trace            1225 kernel/fork.c  		tracehook_finish_clone(p, clone_flags, trace);
trace            1332 kernel/fork.c  	int trace = 0;
trace            1356 kernel/fork.c  		trace = tracehook_prepare_clone(clone_flags);
trace            1359 kernel/fork.c  			 child_tidptr, NULL, trace);
trace            1377 kernel/fork.c  		tracehook_report_clone(trace, regs, clone_flags, nr, p);
trace            1398 kernel/fork.c  		tracehook_report_clone_complete(trace, regs,
trace             106 kernel/latencytop.c 	struct stack_trace trace;
trace             108 kernel/latencytop.c 	memset(&trace, 0, sizeof(trace));
trace             109 kernel/latencytop.c 	trace.max_entries = LT_BACKTRACEDEPTH;
trace             110 kernel/latencytop.c 	trace.entries = &lat->backtrace[0];
trace             111 kernel/latencytop.c 	trace.skip = 0;
trace             112 kernel/latencytop.c 	save_stack_trace_tsk(tsk, &trace);
trace             352 kernel/lockdep.c 	trace->nr_entries = 0;
trace             353 kernel/lockdep.c 	trace->max_entries = MAX_STACK_TRACE_ENTRIES - nr_stack_trace_entries;
trace             354 kernel/lockdep.c 	trace->entries = stack_trace + nr_stack_trace_entries;
trace             356 kernel/lockdep.c 	trace->skip = 3;
trace             358 kernel/lockdep.c 	save_stack_trace(trace);
trace             360 kernel/lockdep.c 	trace->max_entries = trace->nr_entries;
trace             362 kernel/lockdep.c 	nr_stack_trace_entries += trace->nr_entries;
trace             598 kernel/lockdep.c 		print_stack_trace(&entry->trace, 2);
trace             878 kernel/lockdep.c 	if (!save_trace(&entry->trace))
trace             918 kernel/lockdep.c 	print_stack_trace(&target->trace, 6);
trace             961 kernel/lockdep.c 	if (!save_trace(&this.trace))
trace              90 kernel/rcupreempt.c 	struct rcupreempt_trace trace;
trace             209 kernel/rcupreempt.c #define RCU_TRACE_CPU(f, cpu) RCU_TRACE(f, &(RCU_DATA_CPU(cpu)->trace));
trace             215 kernel/rcupreempt.c #define RCU_TRACE_ME(f) RCU_TRACE(f, &(RCU_DATA_ME()->trace));
trace             221 kernel/rcupreempt.c #define RCU_TRACE_RDP(f, rdp) RCU_TRACE(f, &((rdp)->trace));
trace            1478 kernel/rcupreempt.c 	return &rdp->trace;
trace              52 kernel/rcupreempt_trace.c 	trace->done_length += trace->wait_length;
trace              53 kernel/rcupreempt_trace.c 	trace->done_add += trace->wait_length;
trace              54 kernel/rcupreempt_trace.c 	trace->wait_length = 0;
trace              58 kernel/rcupreempt_trace.c 	trace->wait_length += trace->next_length;
trace              59 kernel/rcupreempt_trace.c 	trace->wait_add += trace->next_length;
trace              60 kernel/rcupreempt_trace.c 	trace->next_length = 0;
trace              64 kernel/rcupreempt_trace.c 	atomic_inc(&trace->rcu_try_flip_1);
trace              68 kernel/rcupreempt_trace.c 	atomic_inc(&trace->rcu_try_flip_e1);
trace              72 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_i1++;
trace              76 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_ie1++;
trace              80 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_g1++;
trace              84 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_a1++;
trace              88 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_ae1++;
trace              92 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_a2++;
trace              96 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_z1++;
trace             100 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_ze1++;
trace             104 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_z2++;
trace             108 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_m1++;
trace             112 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_me1++;
trace             116 kernel/rcupreempt_trace.c 	trace->rcu_try_flip_m2++;
trace             120 kernel/rcupreempt_trace.c 	trace->rcu_check_callbacks++;
trace             124 kernel/rcupreempt_trace.c 	trace->done_remove += trace->done_length;
trace             125 kernel/rcupreempt_trace.c 	trace->done_length = 0;
trace             129 kernel/rcupreempt_trace.c 	atomic_inc(&trace->done_invoked);
trace             133 kernel/rcupreempt_trace.c 	trace->next_add++;
trace             134 kernel/rcupreempt_trace.c 	trace->next_length++;
trace             176 kernel/rcupreempt_trace.c 	struct rcupreempt_trace trace;
trace             180 kernel/rcupreempt_trace.c 	rcupreempt_trace_sum(&trace);
trace             185 kernel/rcupreempt_trace.c 		 trace.rcu_check_callbacks);
trace             191 kernel/rcupreempt_trace.c 		 trace.next_add, trace.next_length,
trace             192 kernel/rcupreempt_trace.c 		 trace.wait_add, trace.wait_length,
trace             193 kernel/rcupreempt_trace.c 		 trace.done_add, trace.done_length,
trace             194 kernel/rcupreempt_trace.c 		 trace.done_remove, atomic_read(&trace.done_invoked),
trace             195 kernel/rcupreempt_trace.c 		 atomic_read(&trace.rcu_try_flip_1),
trace             196 kernel/rcupreempt_trace.c 		 atomic_read(&trace.rcu_try_flip_e1),
trace             197 kernel/rcupreempt_trace.c 		 trace.rcu_try_flip_i1, trace.rcu_try_flip_ie1,
trace             198 kernel/rcupreempt_trace.c 		 trace.rcu_try_flip_g1,
trace             199 kernel/rcupreempt_trace.c 		 trace.rcu_try_flip_a1, trace.rcu_try_flip_ae1,
trace             200 kernel/rcupreempt_trace.c 			 trace.rcu_try_flip_a2,
trace             201 kernel/rcupreempt_trace.c 		 trace.rcu_try_flip_z1, trace.rcu_try_flip_ze1,
trace             202 kernel/rcupreempt_trace.c 			 trace.rcu_try_flip_z2,
trace             203 kernel/rcupreempt_trace.c 		 trace.rcu_try_flip_m1, trace.rcu_try_flip_me1,
trace             204 kernel/rcupreempt_trace.c 			trace.rcu_try_flip_m2);
trace              17 kernel/stacktrace.c 	if (WARN_ON(!trace->entries))
trace              20 kernel/stacktrace.c 	for (i = 0; i < trace->nr_entries; i++) {
trace              22 kernel/stacktrace.c 		print_ip_sym(trace->entries[i]);
trace             903 kernel/trace/trace.c 	struct stack_trace trace;
trace             914 kernel/trace/trace.c 	trace.nr_entries	= 0;
trace             915 kernel/trace/trace.c 	trace.max_entries	= FTRACE_STACK_ENTRIES;
trace             916 kernel/trace/trace.c 	trace.skip		= skip;
trace             917 kernel/trace/trace.c 	trace.entries		= entry->stack.caller;
trace             919 kernel/trace/trace.c 	save_stack_trace(&trace);
trace            1217 kernel/trace/trace.c 	if (!current_trace || current_trace != iter->trace) {
trace            1258 kernel/trace/trace.c 	if (current_trace && current_trace == iter->trace && iter->trace->stop)
trace            1259 kernel/trace/trace.c 		iter->trace->stop(iter);
trace            1815 kernel/trace/trace.c 	if (iter->trace && iter->trace->print_line)
trace            1816 kernel/trace/trace.c 		return iter->trace->print_line(iter);
trace            1839 kernel/trace/trace.c 			seq_printf(m, "# tracer: %s\n", iter->trace->name);
trace            1889 kernel/trace/trace.c 	iter->trace = current_trace;
trace            1904 kernel/trace/trace.c 		if (iter->trace && iter->trace->open)
trace            1905 kernel/trace/trace.c 			iter->trace->open(iter);
trace            1931 kernel/trace/trace.c 	if (iter->trace && iter->trace->close)
trace            1932 kernel/trace/trace.c 		iter->trace->close(iter);
trace            2454 kernel/trace/trace.c 	iter->trace = current_trace;
trace            2457 kernel/trace/trace.c 	if (iter->trace->pipe_open)
trace            2458 kernel/trace/trace.c 		iter->trace->pipe_open(iter);
trace            2521 kernel/trace/trace.c 	if (iter->trace->read) {
trace            2522 kernel/trace/trace.c 		sret = iter->trace->read(iter, filp, ubuf, cnt, ppos);
trace            2560 kernel/trace/trace.c 		if (iter->trace != current_trace)
trace             152 kernel/trace/trace.h 	int			(*selftest)(struct tracer *trace,
trace             172 kernel/trace/trace.h 	struct tracer		*trace;
trace             422 kernel/trace/trace_irqsoff.c # define register_irqsoff(trace) register_tracer(&trace)
trace             448 kernel/trace/trace_irqsoff.c # define register_preemptoff(trace) register_tracer(&trace)
trace             477 kernel/trace/trace_irqsoff.c # define register_preemptirqsoff(trace) register_tracer(&trace)
trace             159 kernel/trace/trace_selftest.c 	trace->init(tr);
trace             182 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             187 kernel/trace/trace_selftest.c 	trace->reset(tr);
trace             235 kernel/trace/trace_selftest.c 	trace->init(tr);
trace             240 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             245 kernel/trace/trace_selftest.c 	trace->reset(tr);
trace             253 kernel/trace/trace_selftest.c 	ret = trace_selftest_startup_dynamic_tracing(trace, tr,
trace             278 kernel/trace/trace_selftest.c 	trace->init(tr);
trace             287 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             292 kernel/trace/trace_selftest.c 	trace->reset(tr);
trace             315 kernel/trace/trace_selftest.c 	trace->init(tr);
trace             324 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             329 kernel/trace/trace_selftest.c 	trace->reset(tr);
trace             352 kernel/trace/trace_selftest.c 	trace->init(tr);
trace             367 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             386 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             396 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             411 kernel/trace/trace_selftest.c 	trace->reset(tr);
trace             469 kernel/trace/trace_selftest.c 	trace->init(tr);
trace             491 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             498 kernel/trace/trace_selftest.c 	trace->reset(tr);
trace             523 kernel/trace/trace_selftest.c 	trace->init(tr);
trace             528 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             531 kernel/trace/trace_selftest.c 	trace->reset(tr);
trace             551 kernel/trace/trace_selftest.c 	trace->init(tr);
trace             556 kernel/trace/trace_selftest.c 	trace->ctrl_update(tr);
trace             559 kernel/trace/trace_selftest.c 	trace->reset(tr);
trace              62 lib/fault-inject.c 	struct stack_trace trace;
trace              71 lib/fault-inject.c 	trace.nr_entries = 0;
trace              72 lib/fault-inject.c 	trace.entries = entries;
trace              73 lib/fault-inject.c 	trace.max_entries = depth;
trace              74 lib/fault-inject.c 	trace.skip = 1;
trace              76 lib/fault-inject.c 	save_stack_trace(&trace);
trace              77 lib/fault-inject.c 	for (n = 0; n < trace.nr_entries; n++) {
trace             889 mm/slub.c      	trace(s, page, object, 1);
trace             946 mm/slub.c      	trace(s, page, object, 0);