Lines Matching refs:ring_buffer_per_cpu

483 struct ring_buffer_per_cpu {  struct
537 struct ring_buffer_per_cpu **buffers; argument
547 struct ring_buffer_per_cpu *cpu_buffer;
806 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_wait()
918 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_poll_wait()
959 if (__same_type(*(b), struct ring_buffer_per_cpu)) { \
960 struct ring_buffer_per_cpu *__b = \
1105 rb_is_head_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_is_head_page()
1135 static void rb_set_list_to_head(struct ring_buffer_per_cpu *cpu_buffer, in rb_set_list_to_head()
1148 static void rb_head_page_activate(struct ring_buffer_per_cpu *cpu_buffer) in rb_head_page_activate()
1173 rb_head_page_deactivate(struct ring_buffer_per_cpu *cpu_buffer) in rb_head_page_deactivate()
1184 static int rb_head_page_set(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set()
1207 static int rb_head_page_set_update(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_update()
1216 static int rb_head_page_set_head(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_head()
1225 static int rb_head_page_set_normal(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_normal()
1234 static inline void rb_inc_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_inc_page()
1243 rb_set_head_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_set_head_page()
1298 static void rb_tail_page_update(struct ring_buffer_per_cpu *cpu_buffer, in rb_tail_page_update()
1359 static int rb_check_bpage(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_bpage()
1373 static int rb_check_list(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_list()
1390 static int rb_check_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_check_pages()
1496 static int rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, in rb_allocate_pages()
1521 static struct ring_buffer_per_cpu *
1524 struct ring_buffer_per_cpu *cpu_buffer; in rb_allocate_cpu_buffer()
1582 static void rb_free_cpu_buffer(struct ring_buffer_per_cpu *cpu_buffer) in rb_free_cpu_buffer()
1719 static void rb_reset_cpu(struct ring_buffer_per_cpu *cpu_buffer);
1732 rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned long nr_pages) in rb_remove_pages()
1844 rb_insert_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_insert_pages()
1920 static void rb_update_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_update_pages()
1936 struct ring_buffer_per_cpu *cpu_buffer = container_of(work, in update_pages_handler()
1937 struct ring_buffer_per_cpu, update_pages_work); in update_pages_handler()
1955 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_resize()
2165 rb_reader_event(struct ring_buffer_per_cpu *cpu_buffer) in rb_reader_event()
2237 rb_commit_index(struct ring_buffer_per_cpu *cpu_buffer) in rb_commit_index()
2252 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_inc_iter()
2278 rb_handle_head_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_handle_head_page()
2432 rb_reset_tail(struct ring_buffer_per_cpu *cpu_buffer, in rb_reset_tail()
2501 static inline void rb_end_commit(struct ring_buffer_per_cpu *cpu_buffer);
2507 rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, in rb_move_tail()
2631 static inline bool rb_event_is_commit(struct ring_buffer_per_cpu *cpu_buffer,
2642 rb_check_timestamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_timestamp()
2660 static void rb_add_timestamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_add_timestamp()
2706 rb_update_event(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_event()
2762 rb_event_is_commit(struct ring_buffer_per_cpu *cpu_buffer, in rb_event_is_commit()
2795 rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer, in rb_try_to_discard()
2858 static void rb_start_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_start_commit()
2865 rb_set_commit_to_write(struct ring_buffer_per_cpu *cpu_buffer) in rb_set_commit_to_write()
2915 static __always_inline void rb_end_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_end_commit()
2960 static void rb_commit(struct ring_buffer_per_cpu *cpu_buffer, in rb_commit()
2968 rb_wakeups(struct trace_buffer *buffer, struct ring_buffer_per_cpu *cpu_buffer) in rb_wakeups()
3072 trace_recursive_lock(struct ring_buffer_per_cpu *cpu_buffer) in trace_recursive_lock()
3102 trace_recursive_unlock(struct ring_buffer_per_cpu *cpu_buffer) in trace_recursive_unlock()
3126 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_nest_start()
3146 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_nest_end()
3169 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_unlock_commit()
3187 __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer, in __rb_reserve_next()
3338 struct ring_buffer_per_cpu *cpu_buffer, in rb_reserve_next_event()
3422 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_lock_reserve()
3469 rb_decrement_entry(struct ring_buffer_per_cpu *cpu_buffer, in rb_decrement_entry()
3524 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_discard_commit()
3571 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_write()
3622 static bool rb_per_cpu_empty(struct ring_buffer_per_cpu *cpu_buffer) in rb_per_cpu_empty()
3751 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_record_disable_cpu()
3771 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_record_enable_cpu()
3788 rb_num_of_entries(struct ring_buffer_per_cpu *cpu_buffer) in rb_num_of_entries()
3802 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_oldest_event_ts()
3834 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_bytes_cpu()
3854 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_entries_cpu()
3873 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_overrun_cpu()
3896 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_commit_overrun_cpu()
3918 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_dropped_events_cpu()
3939 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_events_cpu()
3958 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_entries()
3981 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_overruns()
3997 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_iter_reset()
4025 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_iter_reset()
4045 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_iter_empty()
4088 rb_update_read_stamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_read_stamp()
4148 rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_get_reader_page()
4282 static void rb_advance_reader(struct ring_buffer_per_cpu *cpu_buffer) in rb_advance_reader()
4307 struct ring_buffer_per_cpu *cpu_buffer; in rb_advance_iter()
4334 static int rb_lost_events(struct ring_buffer_per_cpu *cpu_buffer) in rb_lost_events()
4340 rb_buffer_peek(struct ring_buffer_per_cpu *cpu_buffer, u64 *ts, in rb_buffer_peek()
4416 struct ring_buffer_per_cpu *cpu_buffer; in rb_iter_peek()
4501 static inline bool rb_reader_lock(struct ring_buffer_per_cpu *cpu_buffer) in rb_reader_lock()
4526 rb_reader_unlock(struct ring_buffer_per_cpu *cpu_buffer, bool locked) in rb_reader_unlock()
4547 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_peek()
4595 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_iter_peek()
4625 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_consume()
4684 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_prepare()
4738 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_start()
4764 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_read_finish()
4792 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_iter_advance()
4824 rb_reset_cpu(struct ring_buffer_per_cpu *cpu_buffer) in rb_reset_cpu()
4870 static void reset_disabled_cpu_buffer(struct ring_buffer_per_cpu *cpu_buffer) in reset_disabled_cpu_buffer()
4896 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_reset_cpu()
4926 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_reset_online_cpus()
4960 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_reset()
4990 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_empty()
5020 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_empty_cpu()
5054 struct ring_buffer_per_cpu *cpu_buffer_a; in ring_buffer_swap_cpu()
5055 struct ring_buffer_per_cpu *cpu_buffer_b; in ring_buffer_swap_cpu()
5133 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_alloc_read_page()
5180 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_free_read_page()
5241 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_read_page()