Lines Matching defs:cpu_buffer

524 	struct ring_buffer_per_cpu	*cpu_buffer;  member
666 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_poll_wait() local
845 rb_is_head_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_is_head_page()
875 static void rb_set_list_to_head(struct ring_buffer_per_cpu *cpu_buffer, in rb_set_list_to_head()
888 static void rb_head_page_activate(struct ring_buffer_per_cpu *cpu_buffer) in rb_head_page_activate()
913 rb_head_page_deactivate(struct ring_buffer_per_cpu *cpu_buffer) in rb_head_page_deactivate()
924 static int rb_head_page_set(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set()
947 static int rb_head_page_set_update(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_update()
956 static int rb_head_page_set_head(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_head()
965 static int rb_head_page_set_normal(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_normal()
974 static inline void rb_inc_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_inc_page()
983 rb_set_head_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_set_head_page()
1038 static void rb_tail_page_update(struct ring_buffer_per_cpu *cpu_buffer, in rb_tail_page_update()
1098 static int rb_check_bpage(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_bpage()
1112 static int rb_check_list(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_list()
1129 static int rb_check_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_check_pages()
1235 static int rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, in rb_allocate_pages()
1263 struct ring_buffer_per_cpu *cpu_buffer; in rb_allocate_cpu_buffer() local
1321 static void rb_free_cpu_buffer(struct ring_buffer_per_cpu *cpu_buffer) in rb_free_cpu_buffer()
1470 rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned long nr_pages) in rb_remove_pages()
1582 rb_insert_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_insert_pages()
1658 static void rb_update_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_update_pages()
1674 struct ring_buffer_per_cpu *cpu_buffer = container_of(work, in update_pages_handler() local
1693 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_resize() local
1886 rb_reader_event(struct ring_buffer_per_cpu *cpu_buffer) in rb_reader_event()
1910 rb_commit_index(struct ring_buffer_per_cpu *cpu_buffer) in rb_commit_index()
1925 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_inc_iter() local
1950 rb_handle_head_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_handle_head_page()
2104 rb_reset_tail(struct ring_buffer_per_cpu *cpu_buffer, in rb_reset_tail()
2179 rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, in rb_move_tail()
2318 rb_update_event(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_event()
2390 rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer, in rb_try_to_discard()
2429 static void rb_start_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_start_commit()
2436 rb_set_commit_to_write(struct ring_buffer_per_cpu *cpu_buffer) in rb_set_commit_to_write()
2490 static __always_inline void rb_end_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_end_commit()
2536 rb_event_is_commit(struct ring_buffer_per_cpu *cpu_buffer, in rb_event_is_commit()
2550 rb_update_write_stamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_write_stamp()
2578 static void rb_commit(struct ring_buffer_per_cpu *cpu_buffer, in rb_commit()
2587 rb_wakeups(struct ring_buffer *buffer, struct ring_buffer_per_cpu *cpu_buffer) in rb_wakeups()
2652 trace_recursive_lock(struct ring_buffer_per_cpu *cpu_buffer) in trace_recursive_lock()
2674 trace_recursive_unlock(struct ring_buffer_per_cpu *cpu_buffer) in trace_recursive_unlock()
2698 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_nest_start() local
2718 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_nest_end() local
2741 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_unlock_commit() local
2759 rb_handle_timestamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_handle_timestamp()
2776 __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer, in __rb_reserve_next()
2832 struct ring_buffer_per_cpu *cpu_buffer, in rb_reserve_next_event()
2926 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_lock_reserve() local
2973 rb_decrement_entry(struct ring_buffer_per_cpu *cpu_buffer, in rb_decrement_entry()
3028 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_discard_commit() local
3080 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_write() local
3131 static bool rb_per_cpu_empty(struct ring_buffer_per_cpu *cpu_buffer) in rb_per_cpu_empty()
3260 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_record_disable_cpu() local
3280 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_record_enable_cpu() local
3297 rb_num_of_entries(struct ring_buffer_per_cpu *cpu_buffer) in rb_num_of_entries()
3311 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_oldest_event_ts() local
3343 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_bytes_cpu() local
3363 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_entries_cpu() local
3382 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_overrun_cpu() local
3405 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_commit_overrun_cpu() local
3427 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_dropped_events_cpu() local
3448 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_events_cpu() local
3467 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_entries() local
3490 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_overruns() local
3506 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_iter_reset() local
3530 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_iter_reset() local
3550 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_iter_empty() local
3572 rb_update_read_stamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_read_stamp()
3632 rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_get_reader_page()
3764 static void rb_advance_reader(struct ring_buffer_per_cpu *cpu_buffer) in rb_advance_reader()
3789 struct ring_buffer_per_cpu *cpu_buffer; in rb_advance_iter() local
3829 static int rb_lost_events(struct ring_buffer_per_cpu *cpu_buffer) in rb_lost_events()
3835 rb_buffer_peek(struct ring_buffer_per_cpu *cpu_buffer, u64 *ts, in rb_buffer_peek()
3911 struct ring_buffer_per_cpu *cpu_buffer; in rb_iter_peek() local
3995 static inline bool rb_reader_lock(struct ring_buffer_per_cpu *cpu_buffer) in rb_reader_lock()
4020 rb_reader_unlock(struct ring_buffer_per_cpu *cpu_buffer, bool locked) in rb_reader_unlock()
4041 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_peek() local
4075 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_iter_peek() local
4105 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_consume() local
4163 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_prepare() local
4212 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_start() local
4238 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_read_finish() local
4268 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_read() local
4308 rb_reset_cpu(struct ring_buffer_per_cpu *cpu_buffer) in rb_reset_cpu()
4356 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_reset_cpu() local
4406 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_empty() local
4436 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_empty_cpu() local
4548 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_alloc_read_page() local
4595 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_free_read_page() local
4656 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_read_page() local