Home
last modified time | relevance | path

Searched refs:rb (Results 1 – 25 of 359) sorted by relevance

12345678910>>...15

/Linux-v6.1/kernel/events/
Dring_buffer.c22 atomic_set(&handle->rb->poll, EPOLLIN); in perf_output_wakeup()
38 struct perf_buffer *rb = handle->rb; in perf_output_get_handle() local
46 (*(volatile unsigned int *)&rb->nest)++; in perf_output_get_handle()
47 handle->wakeup = local_read(&rb->wakeup); in perf_output_get_handle()
52 struct perf_buffer *rb = handle->rb; in perf_output_put_handle() local
60 nest = READ_ONCE(rb->nest); in perf_output_put_handle()
62 WRITE_ONCE(rb->nest, nest - 1); in perf_output_put_handle()
76 head = local_read(&rb->head); in perf_output_put_handle()
110 WRITE_ONCE(rb->user_page->data_head, head); in perf_output_put_handle()
118 WRITE_ONCE(rb->nest, 0); in perf_output_put_handle()
[all …]
Dinternal.h61 extern void rb_free(struct perf_buffer *rb);
65 struct perf_buffer *rb; in rb_free_rcu() local
67 rb = container_of(rcu_head, struct perf_buffer, rcu_head); in rb_free_rcu()
68 rb_free(rb); in rb_free_rcu()
71 static inline void rb_toggle_paused(struct perf_buffer *rb, bool pause) in rb_toggle_paused() argument
73 if (!pause && rb->nr_pages) in rb_toggle_paused()
74 rb->paused = 0; in rb_toggle_paused()
76 rb->paused = 1; in rb_toggle_paused()
82 extern int rb_alloc_aux(struct perf_buffer *rb, struct perf_event *event,
84 extern void rb_free_aux(struct perf_buffer *rb);
[all …]
/Linux-v6.1/tools/lib/bpf/
Dringbuf.c60 static void ringbuf_unmap_ring(struct ring_buffer *rb, struct ring *r) in ringbuf_unmap_ring() argument
63 munmap(r->consumer_pos, rb->page_size); in ringbuf_unmap_ring()
67 munmap(r->producer_pos, rb->page_size + 2 * (r->mask + 1)); in ringbuf_unmap_ring()
73 int ring_buffer__add(struct ring_buffer *rb, int map_fd, in ring_buffer__add() argument
100 tmp = libbpf_reallocarray(rb->rings, rb->ring_cnt + 1, sizeof(*rb->rings)); in ring_buffer__add()
103 rb->rings = tmp; in ring_buffer__add()
105 tmp = libbpf_reallocarray(rb->events, rb->ring_cnt + 1, sizeof(*rb->events)); in ring_buffer__add()
108 rb->events = tmp; in ring_buffer__add()
110 r = &rb->rings[rb->ring_cnt]; in ring_buffer__add()
119 tmp = mmap(NULL, rb->page_size, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0); in ring_buffer__add()
[all …]
/Linux-v6.1/drivers/scsi/bfa/
Dbfa_ioc_ct.c185 void __iomem *rb; in bfa_ioc_ct_reg_init() local
188 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init()
190 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init()
191 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init()
192 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init()
195 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init()
196 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init()
197 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init()
198 ioc->ioc_regs.hfn_mbox_cmd = rb + ct_p0reg[pcifn].hfn; in bfa_ioc_ct_reg_init()
199 ioc->ioc_regs.lpu_mbox_cmd = rb + ct_p0reg[pcifn].lpu; in bfa_ioc_ct_reg_init()
[all …]
Dbfa_ioc_cb.c138 void __iomem *rb; in bfa_ioc_cb_reg_init() local
141 rb = bfa_ioc_bar0(ioc); in bfa_ioc_cb_reg_init()
143 ioc->ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox; in bfa_ioc_cb_reg_init()
144 ioc->ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox; in bfa_ioc_cb_reg_init()
145 ioc->ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn; in bfa_ioc_cb_reg_init()
148 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_cb_reg_init()
149 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_cb_reg_init()
150 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_cb_reg_init()
152 ioc->ioc_regs.heartbeat = (rb + BFA_IOC1_HBEAT_REG); in bfa_ioc_cb_reg_init()
153 ioc->ioc_regs.ioc_fwstate = (rb + BFA_IOC1_STATE_REG); in bfa_ioc_cb_reg_init()
[all …]
/Linux-v6.1/drivers/net/ethernet/brocade/bna/
Dbfa_ioc_ct.c49 static enum bfa_status bfa_ioc_ct_pll_init(void __iomem *rb,
51 static enum bfa_status bfa_ioc_ct2_pll_init(void __iomem *rb,
251 void __iomem *rb; in bfa_ioc_ct_reg_init() local
254 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init()
256 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init()
257 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init()
258 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init()
261 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init()
262 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init()
263 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init()
[all …]
/Linux-v6.1/kernel/bpf/
Dringbuf.c86 struct bpf_ringbuf *rb; member
103 struct bpf_ringbuf *rb; in bpf_ringbuf_area_alloc() local
140 rb = vmap(pages, nr_meta_pages + 2 * nr_data_pages, in bpf_ringbuf_area_alloc()
142 if (rb) { in bpf_ringbuf_area_alloc()
144 rb->pages = pages; in bpf_ringbuf_area_alloc()
145 rb->nr_pages = nr_pages; in bpf_ringbuf_area_alloc()
146 return rb; in bpf_ringbuf_area_alloc()
158 struct bpf_ringbuf *rb = container_of(work, struct bpf_ringbuf, work); in bpf_ringbuf_notify() local
160 wake_up_all(&rb->waitq); in bpf_ringbuf_notify()
165 struct bpf_ringbuf *rb; in bpf_ringbuf_alloc() local
[all …]
/Linux-v6.1/drivers/hid/intel-ish-hid/ishtp/
Dclient-buffers.c23 struct ishtp_cl_rb *rb; in ishtp_cl_alloc_rx_ring() local
28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring()
29 if (!rb) { in ishtp_cl_alloc_rx_ring()
33 ret = ishtp_io_rb_alloc_buf(rb, len); in ishtp_cl_alloc_rx_ring()
37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring()
99 struct ishtp_cl_rb *rb; in ishtp_cl_free_rx_ring() local
105 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, in ishtp_cl_free_rx_ring()
107 list_del(&rb->list); in ishtp_cl_free_rx_ring()
108 kfree(rb->buffer.data); in ishtp_cl_free_rx_ring()
109 kfree(rb); in ishtp_cl_free_rx_ring()
[all …]
Dclient.c44 struct ishtp_cl_rb *rb; in ishtp_read_list_flush() local
49 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush()
50 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush()
51 list_del(&rb->list); in ishtp_read_list_flush()
52 ishtp_io_rb_free(rb); in ishtp_read_list_flush()
447 struct ishtp_cl_rb *rb; in ishtp_cl_read_start() local
477 rb = NULL; in ishtp_cl_read_start()
481 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list); in ishtp_cl_read_start()
482 list_del_init(&rb->list); in ishtp_cl_read_start()
485 rb->cl = cl; in ishtp_cl_read_start()
[all …]
/Linux-v6.1/lib/
Drbtree_test.c20 struct rb_node rb; member
39 if (key < rb_entry(parent, struct test_node, rb)->key) in insert()
45 rb_link_node(&node->rb, parent, new); in insert()
46 rb_insert_color(&node->rb, &root->rb_root); in insert()
57 if (key < rb_entry(parent, struct test_node, rb)->key) in insert_cached()
65 rb_link_node(&node->rb, parent, new); in insert_cached()
66 rb_insert_color_cached(&node->rb, root, leftmost); in insert_cached()
71 rb_erase(&node->rb, &root->rb_root); in erase()
76 rb_erase_cached(&node->rb, root); in erase_cached()
83 struct test_node, rb, u32, augmented, NODE_VAL) in RB_DECLARE_CALLBACKS_MAX() argument
[all …]
/Linux-v6.1/kernel/printk/
Dprintk_ringbuffer.c564 static bool data_make_reusable(struct printk_ringbuffer *rb, in data_make_reusable() argument
570 struct prb_data_ring *data_ring = &rb->text_data_ring; in data_make_reusable()
571 struct prb_desc_ring *desc_ring = &rb->desc_ring; in data_make_reusable()
632 static bool data_push_tail(struct printk_ringbuffer *rb, unsigned long lpos) in data_push_tail() argument
634 struct prb_data_ring *data_ring = &rb->text_data_ring; in data_push_tail()
675 if (!data_make_reusable(rb, tail_lpos, lpos, &next_lpos)) { in data_push_tail()
767 static bool desc_push_tail(struct printk_ringbuffer *rb, in desc_push_tail() argument
770 struct prb_desc_ring *desc_ring = &rb->desc_ring; in desc_push_tail()
812 if (!data_push_tail(rb, desc.text_blk_lpos.next)) in desc_push_tail()
875 static bool desc_reserve(struct printk_ringbuffer *rb, unsigned long *id_out) in desc_reserve() argument
[all …]
Dprintk_ringbuffer.h106 struct printk_ringbuffer *rb; member
305 bool prb_reserve(struct prb_reserved_entry *e, struct printk_ringbuffer *rb,
307 bool prb_reserve_in_last(struct prb_reserved_entry *e, struct printk_ringbuffer *rb,
312 void prb_init(struct printk_ringbuffer *rb,
355 #define prb_for_each_record(from, rb, s, r) \ argument
356 for ((s) = from; prb_read_valid(rb, s, r); (s) = (r)->info->seq + 1)
373 #define prb_for_each_info(from, rb, s, i, lc) \ argument
374 for ((s) = from; prb_read_valid_info(rb, s, i, lc); (s) = (i)->seq + 1)
376 bool prb_read_valid(struct printk_ringbuffer *rb, u64 seq,
378 bool prb_read_valid_info(struct printk_ringbuffer *rb, u64 seq,
[all …]
/Linux-v6.1/drivers/gpu/drm/
Ddrm_mm.c152 INTERVAL_TREE_DEFINE(struct drm_mm_node, rb, in INTERVAL_TREE_DEFINE() argument
168 struct rb_node **link, *rb; in drm_mm_interval_tree_add_node() local
175 rb = &hole_node->rb; in drm_mm_interval_tree_add_node()
176 while (rb) { in drm_mm_interval_tree_add_node()
177 parent = rb_entry(rb, struct drm_mm_node, rb); in drm_mm_interval_tree_add_node()
182 rb = rb_parent(rb); in drm_mm_interval_tree_add_node()
185 rb = &hole_node->rb; in drm_mm_interval_tree_add_node()
186 link = &hole_node->rb.rb_right; in drm_mm_interval_tree_add_node()
189 rb = NULL; in drm_mm_interval_tree_add_node()
195 rb = *link; in drm_mm_interval_tree_add_node()
[all …]
Ddrm_prime.c105 struct rb_node **p, *rb; in drm_prime_add_buf_handle() local
115 rb = NULL; in drm_prime_add_buf_handle()
120 rb = *p; in drm_prime_add_buf_handle()
121 pos = rb_entry(rb, struct drm_prime_member, dmabuf_rb); in drm_prime_add_buf_handle()
123 p = &rb->rb_right; in drm_prime_add_buf_handle()
125 p = &rb->rb_left; in drm_prime_add_buf_handle()
127 rb_link_node(&member->dmabuf_rb, rb, p); in drm_prime_add_buf_handle()
130 rb = NULL; in drm_prime_add_buf_handle()
135 rb = *p; in drm_prime_add_buf_handle()
136 pos = rb_entry(rb, struct drm_prime_member, handle_rb); in drm_prime_add_buf_handle()
[all …]
/Linux-v6.1/drivers/gpu/drm/amd/display/dmub/inc/
Ddmub_cmd.h3421 static inline bool dmub_rb_empty(struct dmub_rb *rb) in dmub_rb_empty() argument
3423 return (rb->wrpt == rb->rptr); in dmub_rb_empty()
3433 static inline bool dmub_rb_full(struct dmub_rb *rb) in dmub_rb_full() argument
3437 if (rb->wrpt >= rb->rptr) in dmub_rb_full()
3438 data_count = rb->wrpt - rb->rptr; in dmub_rb_full()
3440 data_count = rb->capacity - (rb->rptr - rb->wrpt); in dmub_rb_full()
3442 return (data_count == (rb->capacity - DMUB_RB_CMD_SIZE)); in dmub_rb_full()
3453 static inline bool dmub_rb_push_front(struct dmub_rb *rb, in dmub_rb_push_front() argument
3456 uint64_t volatile *dst = (uint64_t volatile *)((uint8_t *)(rb->base_address) + rb->wrpt); in dmub_rb_push_front()
3460 if (dmub_rb_full(rb)) in dmub_rb_push_front()
[all …]
/Linux-v6.1/mm/
Dinterval_tree.c23 INTERVAL_TREE_DEFINE(struct vm_area_struct, shared.rb,
38 if (!prev->shared.rb.rb_right) { in vma_interval_tree_insert_after()
40 link = &prev->shared.rb.rb_right; in vma_interval_tree_insert_after()
42 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after()
43 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after()
46 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after()
47 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after()
48 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after()
52 link = &parent->shared.rb.rb_left; in vma_interval_tree_insert_after()
56 rb_link_node(&node->shared.rb, &parent->shared.rb, link); in vma_interval_tree_insert_after()
[all …]
/Linux-v6.1/Documentation/translations/zh_CN/core-api/
Drbtree.rst271 node = rb_entry(root->rb_node, struct interval_tree_node, rb);
274 if (node->rb.rb_left) {
276 rb_entry(node->rb.rb_left,
277 struct interval_tree_node, rb);
294 if (node->rb.rb_right) {
295 node = rb_entry(node->rb.rb_right,
296 struct interval_tree_node, rb);
311 if (node->rb.rb_left) {
312 subtree_last = rb_entry(node->rb.rb_left,
313 struct interval_tree_node, rb)->__subtree_last;
[all …]
/Linux-v6.1/drivers/target/iscsi/
Discsi_target_configfs.c44 ssize_t rb; in lio_target_np_driver_show() local
48 rb = sprintf(page, "1\n"); in lio_target_np_driver_show()
50 rb = sprintf(page, "0\n"); in lio_target_np_driver_show()
52 return rb; in lio_target_np_driver_show()
474 ssize_t rb; \
479 rb = snprintf(page, PAGE_SIZE, \
483 rb = snprintf(page, PAGE_SIZE, "%u\n", \
488 return rb; \
530 ssize_t rb = 0; in lio_target_nacl_info_show() local
536 rb += sprintf(page+rb, "No active iSCSI Session for Initiator" in lio_target_nacl_info_show()
[all …]
/Linux-v6.1/fs/jffs2/
Dnodelist.h230 struct rb_node rb; member
271 struct rb_node rb; member
334 return rb_entry(node, struct jffs2_node_frag, rb); in frag_first()
344 return rb_entry(node, struct jffs2_node_frag, rb); in frag_last()
347 #define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb)
348 #define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb)
349 #define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb)
350 #define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb)
351 #define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb)
352 #define frag_erase(frag, list) rb_erase(&frag->rb, list)
[all …]
/Linux-v6.1/tools/testing/selftests/bpf/benchs/
Drun_bench_ringbufs.sh8 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
13 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
18 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
41 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
/Linux-v6.1/net/sunrpc/xprtrdma/
Dverbs.c80 static void rpcrdma_regbuf_dma_unmap(struct rpcrdma_regbuf *rb);
81 static void rpcrdma_regbuf_free(struct rpcrdma_regbuf *rb);
850 struct rpcrdma_regbuf *rb; in rpcrdma_req_setup() local
857 rb = rpcrdma_regbuf_alloc(__roundup_pow_of_two(maxhdrsize), in rpcrdma_req_setup()
859 if (!rb) in rpcrdma_req_setup()
862 if (!__rpcrdma_regbuf_dma_map(r_xprt, rb)) in rpcrdma_req_setup()
865 req->rl_rdmabuf = rb; in rpcrdma_req_setup()
866 xdr_buf_init(&req->rl_hdrbuf, rdmab_data(rb), rdmab_length(rb)); in rpcrdma_req_setup()
870 rpcrdma_regbuf_free(rb); in rpcrdma_req_setup()
1237 struct rpcrdma_regbuf *rb; in rpcrdma_regbuf_alloc() local
[all …]
Dxprt_rdma.h127 static inline u64 rdmab_addr(struct rpcrdma_regbuf *rb) in rdmab_addr() argument
129 return rb->rg_iov.addr; in rdmab_addr()
132 static inline u32 rdmab_length(struct rpcrdma_regbuf *rb) in rdmab_length() argument
134 return rb->rg_iov.length; in rdmab_length()
137 static inline u32 rdmab_lkey(struct rpcrdma_regbuf *rb) in rdmab_lkey() argument
139 return rb->rg_iov.lkey; in rdmab_lkey()
142 static inline struct ib_device *rdmab_device(struct rpcrdma_regbuf *rb) in rdmab_device() argument
144 return rb->rg_device; in rdmab_device()
147 static inline void *rdmab_data(const struct rpcrdma_regbuf *rb) in rdmab_data() argument
149 return rb->rg_data; in rdmab_data()
[all …]
/Linux-v6.1/arch/arm/lib/
Dgetuser.S49 rb .req ip label
51 3: ldrbt rb, [r0], #0
53 rb .req r0 label
55 3: ldrb rb, [r0, #1]
58 orr r2, r2, rb, lsl #8
60 orr r2, rb, r2, lsl #8
117 rb .req ip label
119 10: ldrbt rb, [r0], #0
121 rb .req r0 label
123 10: ldrb rb, [r0, #1]
[all …]
/Linux-v6.1/drivers/xen/xenbus/
Dxenbus_dev_frontend.c130 struct read_buffer *rb; in xenbus_file_read() local
148 rb = list_entry(u->read_buffers.next, struct read_buffer, list); in xenbus_file_read()
151 size_t sz = min_t(size_t, len - i, rb->len - rb->cons); in xenbus_file_read()
153 ret = copy_to_user(ubuf + i, &rb->msg[rb->cons], sz); in xenbus_file_read()
156 rb->cons += sz - ret; in xenbus_file_read()
165 if (rb->cons == rb->len) { in xenbus_file_read()
166 list_del(&rb->list); in xenbus_file_read()
167 kfree(rb); in xenbus_file_read()
170 rb = list_entry(u->read_buffers.next, in xenbus_file_read()
191 struct read_buffer *rb; in queue_reply() local
[all …]
/Linux-v6.1/arch/powerpc/mm/book3s64/
Dradix_tlb.c32 unsigned long rb; in tlbiel_radix_set_isa300() local
35 rb = (set << PPC_BITLSHIFT(51)) | (is << PPC_BITLSHIFT(53)); in tlbiel_radix_set_isa300()
39 : : "r"(rb), "r"(rs), "i"(ric), "i"(prs) in tlbiel_radix_set_isa300()
103 unsigned long rb,rs,prs,r; in __tlbiel_pid() local
105 rb = PPC_BIT(53); /* IS = 1 */ in __tlbiel_pid()
106 rb |= set << PPC_BITLSHIFT(51); in __tlbiel_pid()
112 : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory"); in __tlbiel_pid()
113 trace_tlbie(0, 1, rb, rs, ric, prs, r); in __tlbiel_pid()
118 unsigned long rb,rs,prs,r; in __tlbie_pid() local
120 rb = PPC_BIT(53); /* IS = 1 */ in __tlbie_pid()
[all …]

12345678910>>...15