/Linux-v4.19/tools/testing/selftests/powerpc/pmu/ebb/ |
D | trace.h | 24 struct trace_buffer struct 32 struct trace_buffer *trace_buffer_allocate(u64 size); argument 33 int trace_log_reg(struct trace_buffer *tb, u64 reg, u64 value); 34 int trace_log_counter(struct trace_buffer *tb, u64 value); 35 int trace_log_string(struct trace_buffer *tb, char *str); 36 int trace_log_indent(struct trace_buffer *tb); 37 int trace_log_outdent(struct trace_buffer *tb); 38 void trace_buffer_print(struct trace_buffer *tb); 39 void trace_print_location(struct trace_buffer *tb);
|
D | trace.c | 15 struct trace_buffer *trace_buffer_allocate(u64 size) in trace_buffer_allocate() 17 struct trace_buffer *tb; in trace_buffer_allocate() 38 static bool trace_check_bounds(struct trace_buffer *tb, void *p) in trace_check_bounds() 43 static bool trace_check_alloc(struct trace_buffer *tb, void *p) in trace_check_alloc() 62 static void *trace_alloc(struct trace_buffer *tb, int bytes) in trace_alloc() 76 static struct trace_entry *trace_alloc_entry(struct trace_buffer *tb, int payload_size) in trace_alloc_entry() 87 int trace_log_reg(struct trace_buffer *tb, u64 reg, u64 value) in trace_log_reg() 104 int trace_log_counter(struct trace_buffer *tb, u64 value) in trace_log_counter() 120 int trace_log_string(struct trace_buffer *tb, char *str) in trace_log_string() 142 int trace_log_indent(struct trace_buffer *tb) in trace_log_indent() [all …]
|
D | ebb.h | 29 struct trace_buffer *trace;
|
/Linux-v4.19/arch/powerpc/oprofile/cell/ |
D | spu_profiler.c | 80 u64 trace_buffer[2]; in spu_pc_extract() local 94 cbe_read_trace_buffer(cpu, trace_buffer); in spu_pc_extract() 101 = (spu_mask & trace_buffer[0]) << 2; in spu_pc_extract() 103 = (spu_mask & trace_buffer[1]) << 2; in spu_pc_extract() 105 trace_buffer[0] = trace_buffer[0] >> NUM_SPU_BITS_TRBUF; in spu_pc_extract() 106 trace_buffer[1] = trace_buffer[1] >> NUM_SPU_BITS_TRBUF; in spu_pc_extract()
|
/Linux-v4.19/kernel/trace/ |
D | trace.c | 568 static u64 buffer_ftrace_now(struct trace_buffer *buf, int cpu) in buffer_ftrace_now() 584 return buffer_ftrace_now(&global_trace.trace_buffer, cpu); in ftrace_now() 762 if (tr->trace_buffer.buffer) in tracer_tracing_on() 763 ring_buffer_record_on(tr->trace_buffer.buffer); in tracer_tracing_on() 831 buffer = global_trace.trace_buffer.buffer; in __trace_puts() 879 buffer = global_trace.trace_buffer.buffer; in __trace_bputs() 949 static int resize_buffer_duplicate_size(struct trace_buffer *trace_buf, 950 struct trace_buffer *size_buf, int cpu_id); 951 static void set_buffer_entries(struct trace_buffer *buf, unsigned long val); 961 &tr->trace_buffer, RING_BUFFER_ALL_CPUS); in tracing_alloc_snapshot_instance() [all …]
|
D | trace_kdb.c | 34 atomic_inc(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump_buf() 54 ring_buffer_read_prepare(iter.trace_buffer->buffer, cpu); in ftrace_dump_buf() 61 ring_buffer_read_prepare(iter.trace_buffer->buffer, cpu_file); in ftrace_dump_buf() 91 atomic_dec(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump_buf()
|
D | trace_selftest.c | 26 static int trace_test_buffer_cpu(struct trace_buffer *buf, int cpu) in trace_test_buffer_cpu() 63 static int __maybe_unused trace_test_buffer(struct trace_buffer *buf, unsigned long *count) in trace_test_buffer() 365 ret = trace_test_buffer(&tr->trace_buffer, &count); in trace_selftest_startup_dynamic_tracing() 386 ret = trace_test_buffer(&tr->trace_buffer, &count); in trace_selftest_startup_dynamic_tracing() 685 ret = trace_test_buffer(&tr->trace_buffer, &count); in trace_selftest_startup_function() 766 tracing_reset_online_cpus(&tr->trace_buffer); in trace_selftest_startup_function_graph() 789 ret = trace_test_buffer(&tr->trace_buffer, &count); in trace_selftest_startup_function_graph() 844 ret = trace_test_buffer(&tr->trace_buffer, NULL); in trace_selftest_startup_irqsoff() 906 ret = trace_test_buffer(&tr->trace_buffer, NULL); in trace_selftest_startup_preemptoff() 972 ret = trace_test_buffer(&tr->trace_buffer, NULL); in trace_selftest_startup_preemptirqsoff() [all …]
|
D | trace_sched_wakeup.c | 92 *data = per_cpu_ptr(tr->trace_buffer.data, cpu); in func_prolog_preempt_disable() 390 struct ring_buffer *buffer = tr->trace_buffer.buffer; in tracing_sched_switch_trace() 420 struct ring_buffer *buffer = tr->trace_buffer.buffer; in tracing_sched_wakeup_trace() 471 disabled = atomic_inc_return(&per_cpu_ptr(wakeup_trace->trace_buffer.data, cpu)->disabled); in probe_wakeup_sched_switch() 483 data = per_cpu_ptr(wakeup_trace->trace_buffer.data, wakeup_cpu); in probe_wakeup_sched_switch() 505 atomic_dec(&per_cpu_ptr(wakeup_trace->trace_buffer.data, cpu)->disabled); in probe_wakeup_sched_switch() 524 tracing_reset_online_cpus(&tr->trace_buffer); in wakeup_reset() 562 disabled = atomic_inc_return(&per_cpu_ptr(wakeup_trace->trace_buffer.data, cpu)->disabled); in probe_wakeup() 595 data = per_cpu_ptr(wakeup_trace->trace_buffer.data, wakeup_cpu); in probe_wakeup() 609 atomic_dec(&per_cpu_ptr(wakeup_trace->trace_buffer.data, cpu)->disabled); in probe_wakeup()
|
D | trace_mmiotrace.c | 35 tracing_reset_online_cpus(&tr->trace_buffer); in mmio_reset_data() 125 unsigned long over = ring_buffer_overruns(iter->trace_buffer->buffer); in count_overruns() 300 struct ring_buffer *buffer = tr->trace_buffer.buffer; in __trace_mmiotrace_rw() 321 struct trace_array_cpu *data = per_cpu_ptr(tr->trace_buffer.data, smp_processor_id()); in mmio_trace_rw() 330 struct ring_buffer *buffer = tr->trace_buffer.buffer; in __trace_mmiotrace_map() 354 data = per_cpu_ptr(tr->trace_buffer.data, smp_processor_id()); in mmio_trace_mapping()
|
D | trace_functions.c | 104 tr->trace_buffer.cpu = get_cpu(); in function_trace_init() 121 tracing_reset_online_cpus(&tr->trace_buffer); in function_trace_start() 146 data = per_cpu_ptr(tr->trace_buffer.data, cpu); in function_trace_call() 195 data = per_cpu_ptr(tr->trace_buffer.data, cpu); in function_stack_trace_call()
|
D | trace_irqsoff.c | 124 *data = per_cpu_ptr(tr->trace_buffer.data, cpu); in func_prolog_dec() 169 tracing_reset_online_cpus(&irqsoff_trace->trace_buffer); in irqsoff_display_graph() 385 data = per_cpu_ptr(tr->trace_buffer.data, cpu); in start_critical_timing() 423 data = per_cpu_ptr(tr->trace_buffer.data, cpu); in stop_critical_timing()
|
D | trace.h | 177 struct trace_buffer { struct 205 struct trace_buffer trace_buffer; argument 218 struct trace_buffer max_buffer; 591 void tracing_reset(struct trace_buffer *buf, int cpu); 592 void tracing_reset_online_cpus(struct trace_buffer *buf); 931 return !this_cpu_read(tr->trace_buffer.data->ftrace_ignore_pid); in ftrace_trace_task()
|
D | trace_branch.c | 58 data = this_cpu_ptr(tr->trace_buffer.data); in probe_likely_condition() 63 buffer = tr->trace_buffer.buffer; in probe_likely_condition()
|
D | trace_functions_graph.c | 352 struct ring_buffer *buffer = tr->trace_buffer.buffer; in __trace_graph_entry() 413 data = per_cpu_ptr(tr->trace_buffer.data, cpu); in trace_graph_entry() 463 struct ring_buffer *buffer = tr->trace_buffer.buffer; in __trace_graph_return() 487 data = per_cpu_ptr(tr->trace_buffer.data, cpu); in trace_graph_return() 660 ring_buffer_consume(iter->trace_buffer->buffer, iter->cpu, in get_return_for_leaf() 662 event = ring_buffer_peek(iter->trace_buffer->buffer, iter->cpu, in get_return_for_leaf()
|
D | trace_events.c | 247 data = this_cpu_ptr(tr->trace_buffer.data); in trace_event_ignore_this_pid() 556 this_cpu_write(tr->trace_buffer.data->ignore_pid, in event_filter_pid_sched_switch_probe_pre() 570 this_cpu_write(tr->trace_buffer.data->ignore_pid, in event_filter_pid_sched_switch_probe_post() 581 if (!this_cpu_read(tr->trace_buffer.data->ignore_pid)) in event_filter_pid_sched_wakeup_probe_pre() 586 this_cpu_write(tr->trace_buffer.data->ignore_pid, in event_filter_pid_sched_wakeup_probe_pre() 597 if (this_cpu_read(tr->trace_buffer.data->ignore_pid)) in event_filter_pid_sched_wakeup_probe_post() 603 this_cpu_write(tr->trace_buffer.data->ignore_pid, in event_filter_pid_sched_wakeup_probe_post() 635 per_cpu_ptr(tr->trace_buffer.data, cpu)->ignore_pid = false; in __ftrace_clear_event_pids() 1588 this_cpu_write(tr->trace_buffer.data->ignore_pid, in ignore_task_cpu()
|
D | trace_syscalls.c | 342 buffer = tr->trace_buffer.buffer; in ftrace_syscall_enter() 387 buffer = tr->trace_buffer.buffer; in ftrace_syscall_exit()
|
D | trace_hwlat.c | 107 struct ring_buffer *buffer = tr->trace_buffer.buffer; in trace_hwlat_sample()
|
D | blktrace.c | 79 buffer = blk_tr->trace_buffer.buffer; in trace_note() 252 buffer = blk_tr->trace_buffer.buffer; in __blk_add_trace()
|
D | ftrace.c | 166 if (tr && this_cpu_read(tr->trace_buffer.data->ftrace_ignore_pid)) in ftrace_pid_func() 6377 this_cpu_write(tr->trace_buffer.data->ftrace_ignore_pid, in ftrace_filter_pid_sched_switch_probe() 6431 per_cpu_ptr(tr->trace_buffer.data, cpu)->ftrace_ignore_pid = false; in clear_ftrace_pids() 6554 this_cpu_write(tr->trace_buffer.data->ftrace_ignore_pid, in ignore_task_cpu()
|
D | trace_output.c | 548 unsigned long long abs_ts = iter->ts - iter->trace_buffer->time_start; in lat_print_timestamp()
|
D | trace_events_hist.c | 650 buffer = trace_file->tr->trace_buffer.buffer; in trace_event_raw_event_synth()
|
/Linux-v4.19/include/linux/ |
D | trace_events.h | 14 struct trace_buffer; 77 struct trace_buffer *trace_buffer; member
|
/Linux-v4.19/arch/powerpc/oprofile/ |
D | op_model_cell.c | 1488 u64 trace_buffer[2]; in cell_handle_interrupt_spu() local 1532 cbe_read_trace_buffer(cpu, trace_buffer); in cell_handle_interrupt_spu() 1551 trace_entry = trace_buffer[0] in cell_handle_interrupt_spu() 1558 last_trace_buffer = trace_buffer[0]; in cell_handle_interrupt_spu()
|
/Linux-v4.19/drivers/net/ethernet/qlogic/qed/ |
D | qed_hsi.h | 13145 u8 trace_buffer[MCP_TRACE_SIZE]; member 13146 #define MCP_TRACE_BUF ((u8 *)(STRUCT_OFFSET(trace_buffer)))
|