/Linux-v5.4/drivers/infiniband/hw/cxgb3/ |
D | iwch_provider.c | 67 struct iwch_dev *rhp = to_iwch_dev(context->device); in iwch_dealloc_ucontext() local 74 cxio_release_ucontext(&rhp->rdev, &ucontext->uctx); in iwch_dealloc_ucontext() 82 struct iwch_dev *rhp = to_iwch_dev(ibdev); in iwch_alloc_ucontext() local 85 cxio_init_ucontext(&rhp->rdev, &context->uctx); in iwch_alloc_ucontext() 98 xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); in iwch_destroy_cq() 102 cxio_destroy_cq(&chp->rhp->rdev, &chp->cq); in iwch_destroy_cq() 111 struct iwch_dev *rhp = to_iwch_dev(ibcq->device); in iwch_create_cq() local 123 if (!t3a_device(rhp)) { in iwch_create_cq() 131 if (t3a_device(rhp)) { in iwch_create_cq() 147 if (cxio_create_cq(&rhp->rdev, &chp->cq, !udata)) in iwch_create_cq() [all …]
|
D | iwch.h | 126 static inline int t3b_device(const struct iwch_dev *rhp) in t3b_device() argument 128 return rhp->rdev.t3cdev_p->type == T3B; in t3b_device() 131 static inline int t3a_device(const struct iwch_dev *rhp) in t3a_device() argument 133 return rhp->rdev.t3cdev_p->type == T3A; in t3a_device() 136 static inline struct iwch_cq *get_chp(struct iwch_dev *rhp, u32 cqid) in get_chp() argument 138 return xa_load(&rhp->cqs, cqid); in get_chp() 141 static inline struct iwch_qp *get_qhp(struct iwch_dev *rhp, u32 qpid) in get_qhp() argument 143 return xa_load(&rhp->qps, qpid); in get_qhp() 146 static inline struct iwch_mr *get_mhp(struct iwch_dev *rhp, u32 mmid) in get_mhp() argument 148 return xa_load(&rhp->mrs, mmid); in get_mhp()
|
D | iwch_cq.c | 35 static int __iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, in __iwch_poll_cq_one() argument 47 if (t3a_device(chp->rhp) && credit) { in __iwch_poll_cq_one() 50 cxio_hal_cq_op(&rhp->rdev, &chp->cq, CQ_CREDIT_UPDATE, credit); in __iwch_poll_cq_one() 175 static int iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, in iwch_poll_cq_one() argument 187 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe)); in iwch_poll_cq_one() 190 ret = __iwch_poll_cq_one(rhp, chp, qhp, wc); in iwch_poll_cq_one() 193 ret = __iwch_poll_cq_one(rhp, chp, NULL, wc); in iwch_poll_cq_one() 200 struct iwch_dev *rhp; in iwch_poll_cq() local 207 rhp = chp->rhp; in iwch_poll_cq() 218 err = iwch_poll_cq_one(rhp, chp, wc + npolled); in iwch_poll_cq()
|
D | iwch_mem.c | 52 return xa_insert_irq(&mhp->rhp->mrs, mmid, mhp, GFP_KERNEL); in iwch_finish_mem_reg() 55 int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php, in iwch_register_mem() argument 61 if (cxio_register_phys_mem(&rhp->rdev, in iwch_register_mem() 73 cxio_dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in iwch_register_mem() 80 mhp->attr.pbl_addr = cxio_hal_pblpool_alloc(&mhp->rhp->rdev, in iwch_alloc_pbl() 93 cxio_hal_pblpool_free(&mhp->rhp->rdev, mhp->attr.pbl_addr, in iwch_free_pbl() 99 return cxio_write_pbl(&mhp->rhp->rdev, pages, in iwch_write_pbl()
|
D | iwch_provider.h | 47 struct iwch_dev *rhp; member 77 struct iwch_dev *rhp; member 93 struct iwch_dev *rhp; member 105 struct iwch_dev *rhp; member 163 struct iwch_dev *rhp; member 260 int iwch_modify_qp(struct iwch_dev *rhp, 339 int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php,
|
D | iwch_qp.c | 201 static int iwch_sgl2pbl_map(struct iwch_dev *rhp, struct ib_sge *sg_list, in iwch_sgl2pbl_map() argument 209 mhp = get_mhp(rhp, (sg_list[i].lkey) >> 8); in iwch_sgl2pbl_map() 241 rhp->rdev.rnic_info.pbl_base) >> 3) + in iwch_sgl2pbl_map() 255 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, in build_rdma_recv() 302 pbl_addr = cxio_hal_pblpool_alloc(&qhp->rhp->rdev, T3_STAG0_PBL_SIZE); in build_zero_stag_recv() 309 pbl_offset = (pbl_addr - qhp->rhp->rdev.rnic_info.pbl_base) >> 3; in build_zero_stag_recv() 686 return iwch_cxgb3_ofld_send(ep->com.qp->rhp->rdev.t3cdev_p, skb); in iwch_post_zb_read() 717 return iwch_cxgb3_ofld_send(qhp->rhp->rdev.t3cdev_p, skb); in iwch_post_terminate() 777 rchp = get_chp(qhp->rhp, qhp->attr.rcq); in flush_qp() 778 schp = get_chp(qhp->rhp, qhp->attr.scq); in flush_qp() [all …]
|
/Linux-v5.4/drivers/infiniband/hw/cxgb4/ |
D | mem.c | 405 return xa_insert_irq(&mhp->rhp->mrs, mmid, mhp, GFP_KERNEL); in finish_mem_reg() 408 static int register_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, in register_mem() argument 414 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, in register_mem() 427 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in register_mem() 436 mhp->attr.pbl_addr = c4iw_pblpool_alloc(&mhp->rhp->rdev, in alloc_pbl() 449 struct c4iw_dev *rhp; in c4iw_get_dma_mr() local 457 rhp = php->rhp; in c4iw_get_dma_mr() 475 mhp->rhp = rhp; in c4iw_get_dma_mr() 485 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, php->pdid, in c4iw_get_dma_mr() 497 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in c4iw_get_dma_mr() [all …]
|
D | provider.c | 64 struct c4iw_dev *rhp; in c4iw_dealloc_ucontext() local 68 rhp = to_c4iw_dev(ucontext->ibucontext.device); in c4iw_dealloc_ucontext() 72 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in c4iw_dealloc_ucontext() 80 struct c4iw_dev *rhp = to_c4iw_dev(ibdev); in c4iw_alloc_ucontext() local 86 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext() 92 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; in c4iw_alloc_ucontext() 113 mm->addr = virt_to_phys(rhp->rdev.status_page); in c4iw_alloc_ucontext() 195 struct c4iw_dev *rhp; in c4iw_deallocate_pd() local 199 rhp = php->rhp; in c4iw_deallocate_pd() 201 c4iw_put_resource(&rhp->rdev.resource.pdid_table, php->pdid); in c4iw_deallocate_pd() [all …]
|
D | qp.c | 719 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl() 743 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl() 806 PBL_OFF(&mhp->rhp->rdev, mhp->attr.pbl_addr)>>3)); in build_tpte_memreg() 915 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_sq_db() 917 if (qhp->rhp->db_state == NORMAL) in ring_kernel_sq_db() 920 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); in ring_kernel_sq_db() 924 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_sq_db() 932 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_rq_db() 934 if (qhp->rhp->db_state == NORMAL) in ring_kernel_rq_db() 937 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); in ring_kernel_rq_db() [all …]
|
D | cq.c | 351 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); in c4iw_flush_hw_cq() 803 c4iw_invalidate_mr(qhp->rhp, wc->ex.invalidate_rkey); in __c4iw_poll_cq_one() 844 c4iw_invalidate_mr(qhp->rhp, in __c4iw_poll_cq_one() 933 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); in c4iw_poll_cq_one() 978 xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); in c4iw_destroy_cq() 984 destroy_cq(&chp->rhp->rdev, &chp->cq, in c4iw_destroy_cq() 996 struct c4iw_dev *rhp = to_c4iw_dev(ibcq->device); in c4iw_create_cq() local 1010 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq() 1046 hwentries = min(entries * 2, rhp->rdev.hw_queue.t4_max_iq_size); in c4iw_create_cq() 1068 ret = create_cq(&rhp->rdev, &chp->cq, in c4iw_create_cq() [all …]
|
D | iw_cxgb4.h | 349 static inline struct c4iw_cq *get_chp(struct c4iw_dev *rhp, u32 cqid) in get_chp() argument 351 return xa_load(&rhp->cqs, cqid); in get_chp() 354 static inline struct c4iw_qp *get_qhp(struct c4iw_dev *rhp, u32 qpid) in get_qhp() argument 356 return xa_load(&rhp->qps, qpid); in get_qhp() 369 struct c4iw_dev *rhp; member 398 struct c4iw_dev *rhp; member 416 struct c4iw_dev *rhp; member 430 struct c4iw_dev *rhp; member 487 struct c4iw_dev *rhp; member 510 struct c4iw_dev *rhp; member [all …]
|
/Linux-v5.4/include/trace/events/ |
D | rcu.h | 477 TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen_lazy, 480 TP_ARGS(rcuname, rhp, qlen_lazy, qlen), 484 __field(void *, rhp) 492 __entry->rhp = rhp; 493 __entry->func = rhp->func; 499 __entry->rcuname, __entry->rhp, __entry->func, 513 TP_PROTO(const char *rcuname, struct rcu_head *rhp, unsigned long offset, 516 TP_ARGS(rcuname, rhp, offset, qlen_lazy, qlen), 520 __field(void *, rhp) 528 __entry->rhp = rhp; [all …]
|
/Linux-v5.4/kernel/rcu/ |
D | rcu_segcblist.c | 32 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp) in rcu_cblist_enqueue() argument 34 *rclp->tail = rhp; in rcu_cblist_enqueue() 35 rclp->tail = &rhp->next; in rcu_cblist_enqueue() 49 struct rcu_head *rhp) in rcu_cblist_flush_enqueue() argument 58 if (!rhp) { in rcu_cblist_flush_enqueue() 61 rhp->next = NULL; in rcu_cblist_flush_enqueue() 62 srclp->head = rhp; in rcu_cblist_flush_enqueue() 63 srclp->tail = &rhp->next; in rcu_cblist_flush_enqueue() 78 struct rcu_head *rhp; in rcu_cblist_dequeue() local 80 rhp = rclp->head; in rcu_cblist_dequeue() [all …]
|
D | srcutiny.c | 113 struct rcu_head *rhp; in srcu_drive_gp() local 135 rhp = lh; in srcu_drive_gp() 138 rhp->func(rhp); in srcu_drive_gp() 158 void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in call_srcu() argument 163 rhp->func = func; in call_srcu() 164 rhp->next = NULL; in call_srcu() 166 *ssp->srcu_cb_tail = rhp; in call_srcu() 167 ssp->srcu_cb_tail = &rhp->next; in call_srcu()
|
D | rcu_segcblist.h | 28 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp); 31 struct rcu_head *rhp); 109 struct rcu_head *rhp, bool lazy); 111 struct rcu_head *rhp, bool lazy);
|
D | rcuperf.c | 347 static void rcu_perf_async_cb(struct rcu_head *rhp) in rcu_perf_async_cb() argument 350 kfree(rhp); in rcu_perf_async_cb() 362 struct rcu_head *rhp = NULL; in rcu_perf_writer() local 404 if (!rhp) in rcu_perf_writer() 405 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_perf_writer() 406 if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { in rcu_perf_writer() 409 cur_ops->async(rhp, rcu_perf_async_cb); in rcu_perf_writer() 410 rhp = NULL; in rcu_perf_writer() 416 kfree(rhp); /* Because we are stopping. */ in rcu_perf_writer()
|
D | update.c | 439 void do_trace_rcu_torture_read(const char *rcutorturename, struct rcu_head *rhp, in do_trace_rcu_torture_read() argument 443 trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c); in do_trace_rcu_torture_read() 447 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument 522 void call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func) in call_rcu_tasks() argument 527 rhp->next = NULL; in call_rcu_tasks() 528 rhp->func = func; in call_rcu_tasks() 531 *rcu_tasks_cbs_tail = rhp; in call_rcu_tasks() 532 rcu_tasks_cbs_tail = &rhp->next; in call_rcu_tasks()
|
D | srcutree.c | 802 static void srcu_leak_callback(struct rcu_head *rhp) in srcu_leak_callback() argument 834 static void __call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in __call_srcu() argument 845 if (debug_rcu_head_queue(rhp)) { in __call_srcu() 847 WRITE_ONCE(rhp->func, srcu_leak_callback); in __call_srcu() 851 rhp->func = func; in __call_srcu() 856 rcu_segcblist_enqueue(&sdp->srcu_cblist, rhp, false); in __call_srcu() 894 void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in call_srcu() argument 897 __call_srcu(ssp, rhp, func, true); in call_srcu() 1006 static void srcu_barrier_cb(struct rcu_head *rhp) in srcu_barrier_cb() argument 1011 sdp = container_of(rhp, struct srcu_data, srcu_barrier_head); in srcu_barrier_cb() [all …]
|
D | sync.c | 43 static void rcu_sync_func(struct rcu_head *rhp); 73 static void rcu_sync_func(struct rcu_head *rhp) in rcu_sync_func() argument 75 struct rcu_sync *rsp = container_of(rhp, struct rcu_sync, cb_head); in rcu_sync_func()
|
D | rcu.h | 460 struct rcu_head *rhp, 474 struct rcu_head *rhp, 479 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
|
D | rcutorture.c | 1128 static void rcu_torture_timer_cb(struct rcu_head *rhp) in rcu_torture_timer_cb() argument 1130 kfree(rhp); in rcu_torture_timer_cb() 1340 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_NOWAIT); in rcu_torture_timer() local 1342 if (rhp) in rcu_torture_timer() 1343 cur_ops->call(rhp, rcu_torture_timer_cb); in rcu_torture_timer() 1644 static void rcu_torture_fwd_prog_cb(struct rcu_head *rhp) in rcu_torture_fwd_prog_cb() argument 1646 struct fwd_cb_state *fcsp = container_of(rhp, struct fwd_cb_state, rh); in rcu_torture_fwd_prog_cb() 1703 static void rcu_torture_fwd_cb_cr(struct rcu_head *rhp) in rcu_torture_fwd_cb_cr() argument 1707 struct rcu_fwd_cb *rfcp = container_of(rhp, struct rcu_fwd_cb, rh); in rcu_torture_fwd_cb_cr() 2260 static void rcu_torture_leak_cb(struct rcu_head *rhp) in rcu_torture_leak_cb() argument [all …]
|
D | tree_plugin.h | 1690 static bool rcu_nocb_do_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_do_flush_bypass() argument 1698 if (rhp && !rcu_cblist_n_cbs(&rdp->nocb_bypass)) { in rcu_nocb_do_flush_bypass() 1703 if (rhp) in rcu_nocb_do_flush_bypass() 1705 rcu_cblist_flush_enqueue(&rcl, &rdp->nocb_bypass, rhp); in rcu_nocb_do_flush_bypass() 1720 static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_flush_bypass() argument 1727 return rcu_nocb_do_flush_bypass(rdp, rhp, j); in rcu_nocb_flush_bypass() 1761 static bool rcu_nocb_try_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_try_bypass() argument 1817 if (!rcu_nocb_flush_bypass(rdp, rhp, j)) { in rcu_nocb_try_bypass() 1840 rcu_cblist_enqueue(&rdp->nocb_bypass, rhp); in rcu_nocb_try_bypass() 2485 static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_flush_bypass() argument [all …]
|
/Linux-v5.4/drivers/media/usb/pvrusb2/ |
D | pvrusb2-v4l2.c | 42 struct pvr2_ioread *rhp; member 915 if (fhp->rhp) { in pvr2_v4l2_release() 918 sp = pvr2_ioread_get_stream(fhp->rhp); in pvr2_v4l2_release() 920 pvr2_ioread_destroy(fhp->rhp); in pvr2_v4l2_release() 921 fhp->rhp = NULL; in pvr2_v4l2_release() 1045 if (fh->rhp) return 0; in pvr2_v4l2_iosetup() 1061 fh->rhp = pvr2_channel_create_mpeg_stream(fh->pdi->stream); in pvr2_v4l2_iosetup() 1062 if (!fh->rhp) { in pvr2_v4l2_iosetup() 1072 return pvr2_ioread_set_enabled(fh->rhp,!0); in pvr2_v4l2_iosetup() 1115 if (!fh->rhp) { in pvr2_v4l2_read() [all …]
|
/Linux-v5.4/include/linux/ |
D | rcupdate.h | 868 static inline void rcu_head_init(struct rcu_head *rhp) in rcu_head_init() argument 870 rhp->func = (rcu_callback_t)~0L; in rcu_head_init() 887 rcu_head_after_call_rcu(struct rcu_head *rhp, rcu_callback_t f) in rcu_head_after_call_rcu() argument 889 rcu_callback_t func = READ_ONCE(rhp->func); in rcu_head_after_call_rcu()
|
/Linux-v5.4/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ |
D | misc.h | 33 #define trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
|