Home
last modified time | relevance | path

Searched full:rb (Results 1 – 25 of 582) sorted by relevance

12345678910>>...24

/Linux-v5.10/kernel/events/
Dring_buffer.c22 atomic_set(&handle->rb->poll, EPOLLIN); in perf_output_wakeup()
38 struct perf_buffer *rb = handle->rb; in perf_output_get_handle() local
46 (*(volatile unsigned int *)&rb->nest)++; in perf_output_get_handle()
47 handle->wakeup = local_read(&rb->wakeup); in perf_output_get_handle()
52 struct perf_buffer *rb = handle->rb; in perf_output_put_handle() local
58 * @rb->user_page->data_head. in perf_output_put_handle()
60 nest = READ_ONCE(rb->nest); in perf_output_put_handle()
62 WRITE_ONCE(rb->nest, nest - 1); in perf_output_put_handle()
69 * we must ensure the load of @rb->head happens after we've in perf_output_put_handle()
70 * incremented @rb->nest. in perf_output_put_handle()
[all …]
Dinternal.h61 extern void rb_free(struct perf_buffer *rb);
65 struct perf_buffer *rb; in rb_free_rcu() local
67 rb = container_of(rcu_head, struct perf_buffer, rcu_head); in rb_free_rcu()
68 rb_free(rb); in rb_free_rcu()
71 static inline void rb_toggle_paused(struct perf_buffer *rb, bool pause) in rb_toggle_paused() argument
73 if (!pause && rb->nr_pages) in rb_toggle_paused()
74 rb->paused = 0; in rb_toggle_paused()
76 rb->paused = 1; in rb_toggle_paused()
82 extern int rb_alloc_aux(struct perf_buffer *rb, struct perf_event *event,
84 extern void rb_free_aux(struct perf_buffer *rb);
[all …]
/Linux-v5.10/drivers/lightnvm/
Dpblk-rb.c17 * pblk-rb.c - pblk's write buffer
26 static void pblk_rb_data_free(struct pblk_rb *rb) in pblk_rb_data_free() argument
31 list_for_each_entry_safe(p, t, &rb->pages, list) { in pblk_rb_data_free()
39 void pblk_rb_free(struct pblk_rb *rb) in pblk_rb_free() argument
41 pblk_rb_data_free(rb); in pblk_rb_free()
42 vfree(rb->entries); in pblk_rb_free()
72 int pblk_rb_init(struct pblk_rb *rb, unsigned int size, unsigned int threshold, in pblk_rb_init() argument
75 struct pblk *pblk = container_of(rb, struct pblk, rwb); in pblk_rb_init()
92 rb->entries = entries; in pblk_rb_init()
93 rb->seg_size = (1 << power_seg_sz); in pblk_rb_init()
[all …]
/Linux-v5.10/drivers/net/ethernet/brocade/bna/
Dbfa_ioc_ct.c49 static enum bfa_status bfa_ioc_ct_pll_init(void __iomem *rb,
51 static enum bfa_status bfa_ioc_ct2_pll_init(void __iomem *rb,
251 void __iomem *rb; in bfa_ioc_ct_reg_init() local
254 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init()
256 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init()
257 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init()
258 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init()
261 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init()
262 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init()
263 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init()
[all …]
/Linux-v5.10/drivers/scsi/bfa/
Dbfa_ioc_ct.c185 void __iomem *rb; in bfa_ioc_ct_reg_init() local
188 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init()
190 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init()
191 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init()
192 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init()
195 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init()
196 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init()
197 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init()
198 ioc->ioc_regs.hfn_mbox_cmd = rb + ct_p0reg[pcifn].hfn; in bfa_ioc_ct_reg_init()
199 ioc->ioc_regs.lpu_mbox_cmd = rb + ct_p0reg[pcifn].lpu; in bfa_ioc_ct_reg_init()
[all …]
Dbfa_ioc_cb.c138 void __iomem *rb; in bfa_ioc_cb_reg_init() local
141 rb = bfa_ioc_bar0(ioc); in bfa_ioc_cb_reg_init()
143 ioc->ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox; in bfa_ioc_cb_reg_init()
144 ioc->ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox; in bfa_ioc_cb_reg_init()
145 ioc->ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn; in bfa_ioc_cb_reg_init()
148 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_cb_reg_init()
149 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_cb_reg_init()
150 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_cb_reg_init()
152 ioc->ioc_regs.heartbeat = (rb + BFA_IOC1_HBEAT_REG); in bfa_ioc_cb_reg_init()
153 ioc->ioc_regs.ioc_fwstate = (rb + BFA_IOC1_STATE_REG); in bfa_ioc_cb_reg_init()
[all …]
/Linux-v5.10/tools/testing/selftests/bpf/benchs/
Drun_bench_ringbufs.sh34 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
39 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
40 summarize $b "$($RUN_BENCH --rb-sampled $b)"
44 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
45 summarize $b "$($RUN_BENCH --rb-b2b $b)"
46 summarize $b-sampled "$($RUN_BENCH --rb-sampled --rb-b2b $b)"
51 …summarize "rb-sampled-$b" "$($RUN_BENCH --rb-b2b --rb-batch-cnt $b --rb-sampled --rb-sample-rate $…
55 …summarize "pb-sampled-$b" "$($RUN_BENCH --rb-b2b --rb-batch-cnt $b --rb-sampled --rb-sample-rate $…
59 summarize "reserve" "$($RUN_BENCH --rb-b2b rb-custom)"
60 summarize "output" "$($RUN_BENCH --rb-b2b --rb-use-output rb-custom)"
[all …]
/Linux-v5.10/arch/x86/crypto/
Dserpent-sse2-i586-asm_32.S28 #define RB %xmm1 macro
513 read_blocks(%eax, RA, RB, RC, RD, RT0, RT1, RE);
515 K(RA, RB, RC, RD, RE, 0);
516 S0(RA, RB, RC, RD, RE); LK(RC, RB, RD, RA, RE, 1);
517 S1(RC, RB, RD, RA, RE); LK(RE, RD, RA, RC, RB, 2);
518 S2(RE, RD, RA, RC, RB); LK(RB, RD, RE, RC, RA, 3);
519 S3(RB, RD, RE, RC, RA); LK(RC, RA, RD, RB, RE, 4);
520 S4(RC, RA, RD, RB, RE); LK(RA, RD, RB, RE, RC, 5);
521 S5(RA, RD, RB, RE, RC); LK(RC, RA, RD, RE, RB, 6);
522 S6(RC, RA, RD, RE, RB); LK(RD, RB, RA, RE, RC, 7);
[all …]
Dserpent-sse2-x86_64-asm_64.S636 K2(RA, RB, RC, RD, RE, 0);
637 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1);
638 S(S1, RC, RB, RD, RA, RE); LK2(RE, RD, RA, RC, RB, 2);
639 S(S2, RE, RD, RA, RC, RB); LK2(RB, RD, RE, RC, RA, 3);
640 S(S3, RB, RD, RE, RC, RA); LK2(RC, RA, RD, RB, RE, 4);
641 S(S4, RC, RA, RD, RB, RE); LK2(RA, RD, RB, RE, RC, 5);
642 S(S5, RA, RD, RB, RE, RC); LK2(RC, RA, RD, RE, RB, 6);
643 S(S6, RC, RA, RD, RE, RB); LK2(RD, RB, RA, RE, RC, 7);
644 S(S7, RD, RB, RA, RE, RC); LK2(RC, RA, RE, RD, RB, 8);
645 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9);
[all …]
Dserpent-avx2-asm_64.S577 K2(RA, RB, RC, RD, RE, 0);
578 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1);
579 S(S1, RC, RB, RD, RA, RE); LK2(RE, RD, RA, RC, RB, 2);
580 S(S2, RE, RD, RA, RC, RB); LK2(RB, RD, RE, RC, RA, 3);
581 S(S3, RB, RD, RE, RC, RA); LK2(RC, RA, RD, RB, RE, 4);
582 S(S4, RC, RA, RD, RB, RE); LK2(RA, RD, RB, RE, RC, 5);
583 S(S5, RA, RD, RB, RE, RC); LK2(RC, RA, RD, RE, RB, 6);
584 S(S6, RC, RA, RD, RE, RB); LK2(RD, RB, RA, RE, RC, 7);
585 S(S7, RD, RB, RA, RE, RC); LK2(RC, RA, RE, RD, RB, 8);
586 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9);
[all …]
Dserpent-avx-x86_64-asm_64.S571 K2(RA, RB, RC, RD, RE, 0);
572 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1);
573 S(S1, RC, RB, RD, RA, RE); LK2(RE, RD, RA, RC, RB, 2);
574 S(S2, RE, RD, RA, RC, RB); LK2(RB, RD, RE, RC, RA, 3);
575 S(S3, RB, RD, RE, RC, RA); LK2(RC, RA, RD, RB, RE, 4);
576 S(S4, RC, RA, RD, RB, RE); LK2(RA, RD, RB, RE, RC, 5);
577 S(S5, RA, RD, RB, RE, RC); LK2(RC, RA, RD, RE, RB, 6);
578 S(S6, RC, RA, RD, RE, RB); LK2(RD, RB, RA, RE, RC, 7);
579 S(S7, RD, RB, RA, RE, RC); LK2(RC, RA, RE, RD, RB, 8);
580 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9);
[all …]
/Linux-v5.10/tools/lib/bpf/
Dringbuf.c42 static void ringbuf_unmap_ring(struct ring_buffer *rb, struct ring *r) in ringbuf_unmap_ring() argument
45 munmap(r->consumer_pos, rb->page_size); in ringbuf_unmap_ring()
49 munmap(r->producer_pos, rb->page_size + 2 * (r->mask + 1)); in ringbuf_unmap_ring()
55 int ring_buffer__add(struct ring_buffer *rb, int map_fd, in ring_buffer__add() argument
81 tmp = libbpf_reallocarray(rb->rings, rb->ring_cnt + 1, sizeof(*rb->rings)); in ring_buffer__add()
84 rb->rings = tmp; in ring_buffer__add()
86 tmp = libbpf_reallocarray(rb->events, rb->ring_cnt + 1, sizeof(*rb->events)); in ring_buffer__add()
89 rb->events = tmp; in ring_buffer__add()
91 r = &rb->rings[rb->ring_cnt]; in ring_buffer__add()
100 tmp = mmap(NULL, rb->page_size, PROT_READ | PROT_WRITE, MAP_SHARED, in ring_buffer__add()
[all …]
/Linux-v5.10/lib/
Drbtree_test.c14 __param(int, nnodes, 100, "Number of nodes in the rb-tree");
15 __param(int, perf_loops, 1000, "Number of iterations modifying the rb-tree");
16 __param(int, check_loops, 100, "Number of iterations modifying and verifying the rb-tree");
20 struct rb_node rb; member
39 if (key < rb_entry(parent, struct test_node, rb)->key) in insert()
45 rb_link_node(&node->rb, parent, new); in insert()
46 rb_insert_color(&node->rb, &root->rb_root); in insert()
57 if (key < rb_entry(parent, struct test_node, rb)->key) in insert_cached()
65 rb_link_node(&node->rb, parent, new); in insert_cached()
66 rb_insert_color_cached(&node->rb, root, leftmost); in insert_cached()
[all …]
/Linux-v5.10/drivers/hid/intel-ish-hid/ishtp/
Dclient-buffers.c23 struct ishtp_cl_rb *rb; in ishtp_cl_alloc_rx_ring() local
28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring()
29 if (!rb) { in ishtp_cl_alloc_rx_ring()
33 ret = ishtp_io_rb_alloc_buf(rb, len); in ishtp_cl_alloc_rx_ring()
37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring()
99 struct ishtp_cl_rb *rb; in ishtp_cl_free_rx_ring() local
105 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, in ishtp_cl_free_rx_ring()
107 list_del(&rb->list); in ishtp_cl_free_rx_ring()
108 kfree(rb->buffer.data); in ishtp_cl_free_rx_ring()
109 kfree(rb); in ishtp_cl_free_rx_ring()
[all …]
Dclient.c43 struct ishtp_cl_rb *rb; in ishtp_read_list_flush() local
48 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush()
49 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush()
50 list_del(&rb->list); in ishtp_read_list_flush()
51 ishtp_io_rb_free(rb); in ishtp_read_list_flush()
447 struct ishtp_cl_rb *rb; in ishtp_cl_read_start() local
471 /* The current rb is the head of the free rb list */ in ishtp_cl_read_start()
477 rb = NULL; in ishtp_cl_read_start()
481 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list); in ishtp_cl_read_start()
482 list_del_init(&rb->list); in ishtp_cl_read_start()
[all …]
/Linux-v5.10/kernel/bpf/
Dringbuf.c52 struct bpf_ringbuf *rb; member
69 struct bpf_ringbuf *rb; in bpf_ringbuf_area_alloc() local
109 rb = vmap(pages, nr_meta_pages + 2 * nr_data_pages, in bpf_ringbuf_area_alloc()
111 if (rb) { in bpf_ringbuf_area_alloc()
112 rb->pages = pages; in bpf_ringbuf_area_alloc()
113 rb->nr_pages = nr_pages; in bpf_ringbuf_area_alloc()
114 return rb; in bpf_ringbuf_area_alloc()
126 struct bpf_ringbuf *rb = container_of(work, struct bpf_ringbuf, work); in bpf_ringbuf_notify() local
128 wake_up_all(&rb->waitq); in bpf_ringbuf_notify()
133 struct bpf_ringbuf *rb; in bpf_ringbuf_alloc() local
[all …]
/Linux-v5.10/tools/testing/selftests/powerpc/stringloops/
Dmemcmp_64.S15 #define rB r10 macro
121 lbz rB,0(r4)
122 subf. rC,rB,rA
127 lbz rB,1(r4)
128 subf. rC,rB,rA
133 lbz rB,2(r4)
134 subf. rC,rB,rA
139 lbz rB,3(r4)
140 subf. rC,rB,rA
173 LD rB,0,r4
[all …]
/Linux-v5.10/arch/powerpc/lib/
Dmemcmp_64.S15 #define rB r10 macro
121 lbz rB,0(r4)
122 subf. rC,rB,rA
127 lbz rB,1(r4)
128 subf. rC,rB,rA
133 lbz rB,2(r4)
134 subf. rC,rB,rA
139 lbz rB,3(r4)
140 subf. rC,rB,rA
173 LD rB,0,r4
[all …]
/Linux-v5.10/mm/
Dinterval_tree.c23 INTERVAL_TREE_DEFINE(struct vm_area_struct, shared.rb,
38 if (!prev->shared.rb.rb_right) { in vma_interval_tree_insert_after()
40 link = &prev->shared.rb.rb_right; in vma_interval_tree_insert_after()
42 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after()
43 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after()
46 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after()
47 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after()
48 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after()
52 link = &parent->shared.rb.rb_left; in vma_interval_tree_insert_after()
56 rb_link_node(&node->shared.rb, &parent->shared.rb, link); in vma_interval_tree_insert_after()
[all …]
/Linux-v5.10/arch/powerpc/crypto/
Dsha1-powerpc-asm.S28 #define RB(t) ((((t)+3)%6)+7) macro
40 andc r0,RD(t),RB(t); \
41 and r6,RB(t),RC(t); \
48 rotlwi RB(t),RB(t),30; \
52 and r6,RB(t),RC(t); \
53 andc r0,RD(t),RB(t); \
55 rotlwi RB(t),RB(t),30; \
67 xor r6,RB(t),RC(t); \
69 rotlwi RB(t),RB(t),30; \
77 xor r6,RB(t),RC(t); \
[all …]
/Linux-v5.10/arch/powerpc/xmon/
Dspu-insns.h15 RRR | op | RC | RB | RA | RT | RI7 | op | I7 | RA | RT |
33 RR | op | RB | RA | RT | LBT | op |RO| I16 | RO |
53 ASM_RRR mnemonic RC, RA, RB, RT ASM_RI4 mnemonic RT, RA, I4
71 ASM_RAB mnemonic RA, RB
73 ASM_RR mnemonic RT, RA, RB
85 The first(most significant) digit is always 0. Then it is followd by RC, RB, RA and RT digits.
90 …ction has 00113 as the DEPENDENCY field. This means RC is not used in this operation, RB and RA are
122 /* 0[RC][RB][RA][RT] */
181 APUOP(M_CBX, RR, 0x1d4, "cbx", _A3(A_T,A_A,A_B), 00112, SHUF) /* genCtl%%insX RT<-sta(Ra+Rb,siz…
182 APUOP(M_CHX, RR, 0x1d5, "chx", _A3(A_T,A_A,A_B), 00112, SHUF) /* genCtl%%insX RT<-sta(Ra+Rb,siz…
[all …]
Dppc-opc.c538 /* The RB field in an X, XO, M, or MDS form instruction. */
539 #define RB RAOPT + 1 macro
543 /* The RB field in an X form instruction when it must be the same as
546 #define RBS RB + 1
549 /* The RB field in an lswx instruction, which has special value
554 /* The RB field of the dccci and iccci instructions, which are optional. */
1783 /* The RB field in an X form instruction when it must be the same as
1808 /* The RB field in an lswx instruction, which has special value
2629 /* An XX1_MASK with the RB field fixed. */
2681 /* An X_MASK with the RB field fixed. */
[all …]
/Linux-v5.10/drivers/gpu/drm/amd/display/dmub/inc/
Ddmub_cmd.h812 static inline bool dmub_rb_empty(struct dmub_rb *rb) in dmub_rb_empty() argument
814 return (rb->wrpt == rb->rptr); in dmub_rb_empty()
817 static inline bool dmub_rb_full(struct dmub_rb *rb) in dmub_rb_full() argument
821 if (rb->wrpt >= rb->rptr) in dmub_rb_full()
822 data_count = rb->wrpt - rb->rptr; in dmub_rb_full()
824 data_count = rb->capacity - (rb->rptr - rb->wrpt); in dmub_rb_full()
826 return (data_count == (rb->capacity - DMUB_RB_CMD_SIZE)); in dmub_rb_full()
829 static inline bool dmub_rb_push_front(struct dmub_rb *rb, in dmub_rb_push_front() argument
832 uint64_t volatile *dst = (uint64_t volatile *)(rb->base_address) + rb->wrpt / sizeof(uint64_t); in dmub_rb_push_front()
836 if (dmub_rb_full(rb)) in dmub_rb_push_front()
[all …]
/Linux-v5.10/drivers/target/iscsi/
Discsi_target_configfs.c44 ssize_t rb; in lio_target_np_driver_show() local
48 rb = sprintf(page, "1\n"); in lio_target_np_driver_show()
50 rb = sprintf(page, "0\n"); in lio_target_np_driver_show()
52 return rb; in lio_target_np_driver_show()
449 ssize_t rb; \
454 rb = snprintf(page, PAGE_SIZE, \
458 rb = snprintf(page, PAGE_SIZE, "%u\n", \
463 return rb; \
505 ssize_t rb = 0; in lio_target_nacl_info_show() local
511 rb += sprintf(page+rb, "No active iSCSI Session for Initiator" in lio_target_nacl_info_show()
[all …]
/Linux-v5.10/drivers/gpu/drm/
Ddrm_mm.c37 * an RB-tree is used instead. At least if we expect heavy fragmentation.
155 INTERVAL_TREE_DEFINE(struct drm_mm_node, rb, in INTERVAL_TREE_DEFINE() argument
171 struct rb_node **link, *rb; in drm_mm_interval_tree_add_node() local
178 rb = &hole_node->rb; in drm_mm_interval_tree_add_node()
179 while (rb) { in drm_mm_interval_tree_add_node()
180 parent = rb_entry(rb, struct drm_mm_node, rb); in drm_mm_interval_tree_add_node()
185 rb = rb_parent(rb); in drm_mm_interval_tree_add_node()
188 rb = &hole_node->rb; in drm_mm_interval_tree_add_node()
189 link = &hole_node->rb.rb_right; in drm_mm_interval_tree_add_node()
192 rb = NULL; in drm_mm_interval_tree_add_node()
[all …]

12345678910>>...24