Lines Matching defs:cpu_buffer

520 	struct ring_buffer_per_cpu	*cpu_buffer;  member
581 static void verify_event(struct ring_buffer_per_cpu *cpu_buffer, in verify_event()
608 static inline void verify_event(struct ring_buffer_per_cpu *cpu_buffer, in verify_event()
655 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[smp_processor_id()]; in ring_buffer_event_time_stamp() local
728 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in full_hit() local
759 struct ring_buffer_per_cpu *cpu_buffer = in rb_wake_up_waiters() local
785 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_wake_waiters() local
817 struct ring_buffer_per_cpu *cpu_buffer; in rb_watermark_hit() local
857 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_wait() local
947 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_poll_wait() local
1193 static void rb_head_page_activate(struct ring_buffer_per_cpu *cpu_buffer) in rb_head_page_activate()
1218 rb_head_page_deactivate(struct ring_buffer_per_cpu *cpu_buffer) in rb_head_page_deactivate()
1229 static int rb_head_page_set(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set()
1252 static int rb_head_page_set_update(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_update()
1261 static int rb_head_page_set_head(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_head()
1270 static int rb_head_page_set_normal(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_normal()
1287 rb_set_head_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_set_head_page()
1339 static void rb_tail_page_update(struct ring_buffer_per_cpu *cpu_buffer, in rb_tail_page_update()
1400 static void rb_check_bpage(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_bpage()
1415 static void rb_check_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_check_pages()
1439 static int __rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, in __rb_allocate_pages()
1515 static int rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, in rb_allocate_pages()
1543 struct ring_buffer_per_cpu *cpu_buffer; in rb_allocate_cpu_buffer() local
1602 static void rb_free_cpu_buffer(struct ring_buffer_per_cpu *cpu_buffer) in rb_free_cpu_buffer()
1765 rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned long nr_pages) in rb_remove_pages()
1874 rb_insert_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_insert_pages()
1954 static void rb_update_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_update_pages()
1970 struct ring_buffer_per_cpu *cpu_buffer = container_of(work, in update_pages_handler() local
1989 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_resize() local
2213 rb_reader_event(struct ring_buffer_per_cpu *cpu_buffer) in rb_reader_event()
2285 rb_commit_index(struct ring_buffer_per_cpu *cpu_buffer) in rb_commit_index()
2291 rb_event_index(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) in rb_event_index()
2302 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_inc_iter() local
2328 rb_handle_head_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_handle_head_page()
2483 rb_reset_tail(struct ring_buffer_per_cpu *cpu_buffer, in rb_reset_tail()
2566 rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, in rb_move_tail()
2670 rb_add_time_stamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_add_time_stamp()
2699 rb_check_timestamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_timestamp()
2717 static void rb_add_timestamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_add_timestamp()
2770 rb_update_event(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_event()
2830 rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer, in rb_try_to_discard()
2895 static void rb_start_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_start_commit()
2902 rb_set_commit_to_write(struct ring_buffer_per_cpu *cpu_buffer) in rb_set_commit_to_write()
2958 static __always_inline void rb_end_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_end_commit()
3003 static void rb_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_commit()
3010 rb_wakeups(struct trace_buffer *buffer, struct ring_buffer_per_cpu *cpu_buffer) in rb_wakeups()
3114 trace_recursive_lock(struct ring_buffer_per_cpu *cpu_buffer) in trace_recursive_lock()
3141 trace_recursive_unlock(struct ring_buffer_per_cpu *cpu_buffer) in trace_recursive_unlock()
3165 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_nest_start() local
3185 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_nest_end() local
3206 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_unlock_commit() local
3372 static void check_buffer(struct ring_buffer_per_cpu *cpu_buffer, in check_buffer()
3453 static inline void check_buffer(struct ring_buffer_per_cpu *cpu_buffer, in check_buffer()
3461 __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer, in __rb_reserve_next()
3605 struct ring_buffer_per_cpu *cpu_buffer, in rb_reserve_next_event()
3697 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_lock_reserve() local
3744 rb_decrement_entry(struct ring_buffer_per_cpu *cpu_buffer, in rb_decrement_entry()
3799 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_discard_commit() local
3846 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_write() local
3897 static bool rb_per_cpu_empty(struct ring_buffer_per_cpu *cpu_buffer) in rb_per_cpu_empty()
4046 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_record_disable_cpu() local
4066 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_record_enable_cpu() local
4083 rb_num_of_entries(struct ring_buffer_per_cpu *cpu_buffer) in rb_num_of_entries()
4097 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_oldest_event_ts() local
4129 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_bytes_cpu() local
4149 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_entries_cpu() local
4168 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_overrun_cpu() local
4191 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_commit_overrun_cpu() local
4213 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_dropped_events_cpu() local
4234 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_events_cpu() local
4253 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_entries() local
4276 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_overruns() local
4292 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_iter_reset() local
4321 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_iter_reset() local
4341 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_iter_empty() local
4384 rb_update_read_stamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_read_stamp()
4444 rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_get_reader_page()
4611 static void rb_advance_reader(struct ring_buffer_per_cpu *cpu_buffer) in rb_advance_reader()
4637 struct ring_buffer_per_cpu *cpu_buffer; in rb_advance_iter() local
4664 static int rb_lost_events(struct ring_buffer_per_cpu *cpu_buffer) in rb_lost_events()
4670 rb_buffer_peek(struct ring_buffer_per_cpu *cpu_buffer, u64 *ts, in rb_buffer_peek()
4747 struct ring_buffer_per_cpu *cpu_buffer; in rb_iter_peek() local
4834 static inline bool rb_reader_lock(struct ring_buffer_per_cpu *cpu_buffer) in rb_reader_lock()
4859 rb_reader_unlock(struct ring_buffer_per_cpu *cpu_buffer, bool locked) in rb_reader_unlock()
4879 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_peek() local
4927 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_iter_peek() local
4957 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_consume() local
5016 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_prepare() local
5072 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_start() local
5098 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_read_finish() local
5126 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_iter_advance() local
5175 rb_reset_cpu(struct ring_buffer_per_cpu *cpu_buffer) in rb_reset_cpu()
5223 static void reset_disabled_cpu_buffer(struct ring_buffer_per_cpu *cpu_buffer) in reset_disabled_cpu_buffer()
5249 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_reset_cpu() local
5281 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_reset_online_cpus() local
5322 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_reset() local
5357 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_empty() local
5387 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_empty_cpu() local
5513 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_alloc_read_page() local
5568 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_free_read_page() local
5641 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_read_page() local
5875 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_subbuf_order_set() local