/Linux-v4.19/drivers/infiniband/hw/cxgb3/ |
D | iwch.h | 127 static inline int t3b_device(const struct iwch_dev *rhp) in t3b_device() argument 129 return rhp->rdev.t3cdev_p->type == T3B; in t3b_device() 132 static inline int t3a_device(const struct iwch_dev *rhp) in t3a_device() argument 134 return rhp->rdev.t3cdev_p->type == T3A; in t3a_device() 137 static inline struct iwch_cq *get_chp(struct iwch_dev *rhp, u32 cqid) in get_chp() argument 139 return idr_find(&rhp->cqidr, cqid); in get_chp() 142 static inline struct iwch_qp *get_qhp(struct iwch_dev *rhp, u32 qpid) in get_qhp() argument 144 return idr_find(&rhp->qpidr, qpid); in get_qhp() 147 static inline struct iwch_mr *get_mhp(struct iwch_dev *rhp, u32 mmid) in get_mhp() argument 149 return idr_find(&rhp->mmidr, mmid); in get_mhp() [all …]
|
D | iwch_provider.c | 66 struct iwch_dev *rhp = to_iwch_dev(context->device); in iwch_dealloc_ucontext() local 73 cxio_release_ucontext(&rhp->rdev, &ucontext->uctx); in iwch_dealloc_ucontext() 82 struct iwch_dev *rhp = to_iwch_dev(ibdev); in iwch_alloc_ucontext() local 88 cxio_init_ucontext(&rhp->rdev, &context->uctx); in iwch_alloc_ucontext() 101 remove_handle(chp->rhp, &chp->rhp->cqidr, chp->cq.cqid); in iwch_destroy_cq() 105 cxio_destroy_cq(&chp->rhp->rdev, &chp->cq); in iwch_destroy_cq() 116 struct iwch_dev *rhp; in iwch_create_cq() local 128 rhp = to_iwch_dev(ibdev); in iwch_create_cq() 135 if (!t3a_device(rhp)) { in iwch_create_cq() 144 if (t3a_device(rhp)) { in iwch_create_cq() [all …]
|
D | iwch_mem.c | 52 return insert_handle(mhp->rhp, &mhp->rhp->mmidr, mhp, mmid); in iwch_finish_mem_reg() 55 int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php, in iwch_register_mem() argument 61 if (cxio_register_phys_mem(&rhp->rdev, in iwch_register_mem() 73 cxio_dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in iwch_register_mem() 80 mhp->attr.pbl_addr = cxio_hal_pblpool_alloc(&mhp->rhp->rdev, in iwch_alloc_pbl() 93 cxio_hal_pblpool_free(&mhp->rhp->rdev, mhp->attr.pbl_addr, in iwch_free_pbl() 99 return cxio_write_pbl(&mhp->rhp->rdev, pages, in iwch_write_pbl()
|
D | iwch_cq.c | 35 static int __iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, in __iwch_poll_cq_one() argument 47 if (t3a_device(chp->rhp) && credit) { in __iwch_poll_cq_one() 50 cxio_hal_cq_op(&rhp->rdev, &chp->cq, CQ_CREDIT_UPDATE, credit); in __iwch_poll_cq_one() 175 static int iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, in iwch_poll_cq_one() argument 187 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe)); in iwch_poll_cq_one() 190 ret = __iwch_poll_cq_one(rhp, chp, qhp, wc); in iwch_poll_cq_one() 193 ret = __iwch_poll_cq_one(rhp, chp, NULL, wc); in iwch_poll_cq_one() 200 struct iwch_dev *rhp; in iwch_poll_cq() local 207 rhp = chp->rhp; in iwch_poll_cq() 218 err = iwch_poll_cq_one(rhp, chp, wc + npolled); in iwch_poll_cq()
|
D | iwch_provider.h | 47 struct iwch_dev *rhp; member 77 struct iwch_dev *rhp; member 93 struct iwch_dev *rhp; member 105 struct iwch_dev *rhp; member 163 struct iwch_dev *rhp; member 260 int iwch_modify_qp(struct iwch_dev *rhp, 339 int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php,
|
D | iwch_qp.c | 201 static int iwch_sgl2pbl_map(struct iwch_dev *rhp, struct ib_sge *sg_list, in iwch_sgl2pbl_map() argument 209 mhp = get_mhp(rhp, (sg_list[i].lkey) >> 8); in iwch_sgl2pbl_map() 241 rhp->rdev.rnic_info.pbl_base) >> 3) + in iwch_sgl2pbl_map() 255 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, in build_rdma_recv() 302 pbl_addr = cxio_hal_pblpool_alloc(&qhp->rhp->rdev, T3_STAG0_PBL_SIZE); in build_zero_stag_recv() 309 pbl_offset = (pbl_addr - qhp->rhp->rdev.rnic_info.pbl_base) >> 3; in build_zero_stag_recv() 686 return iwch_cxgb3_ofld_send(ep->com.qp->rhp->rdev.t3cdev_p, skb); in iwch_post_zb_read() 717 return iwch_cxgb3_ofld_send(qhp->rhp->rdev.t3cdev_p, skb); in iwch_post_terminate() 777 rchp = get_chp(qhp->rhp, qhp->attr.rcq); in flush_qp() 778 schp = get_chp(qhp->rhp, qhp->attr.scq); in flush_qp() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/cxgb4/ |
D | mem.c | 398 return insert_handle(mhp->rhp, &mhp->rhp->mmidr, mhp, mmid); in finish_mem_reg() 401 static int register_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, in register_mem() argument 407 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, in register_mem() 420 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in register_mem() 429 mhp->attr.pbl_addr = c4iw_pblpool_alloc(&mhp->rhp->rdev, in alloc_pbl() 442 struct c4iw_dev *rhp; in c4iw_get_dma_mr() local 450 rhp = php->rhp; in c4iw_get_dma_mr() 468 mhp->rhp = rhp; in c4iw_get_dma_mr() 478 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, php->pdid, in c4iw_get_dma_mr() 490 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in c4iw_get_dma_mr() [all …]
|
D | qp.c | 715 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl() 736 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl() 799 PBL_OFF(&mhp->rhp->rdev, mhp->attr.pbl_addr)>>3)); in build_tpte_memreg() 889 struct c4iw_dev *rhp; in free_qp_work() local 893 rhp = qhp->rhp; in free_qp_work() 896 destroy_qp(&rhp->rdev, &qhp->wq, in free_qp_work() 897 ucontext ? &ucontext->uctx : &rhp->rdev.uctx, !qhp->srq); in free_qp_work() 911 queue_work(qhp->rhp->rdev.free_workq, &qhp->free_work); in queue_qp_free() 936 spin_lock_irqsave(&qhp->rhp->lock, flags); in ring_kernel_sq_db() 938 if (qhp->rhp->db_state == NORMAL) in ring_kernel_sq_db() [all …]
|
D | provider.c | 64 struct c4iw_dev *rhp; in _c4iw_free_ucontext() local 68 rhp = to_c4iw_dev(ucontext->ibucontext.device); in _c4iw_free_ucontext() 73 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in _c4iw_free_ucontext() 90 struct c4iw_dev *rhp = to_c4iw_dev(ibdev); in c4iw_alloc_ucontext() local 102 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext() 109 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; in c4iw_alloc_ucontext() 130 mm->addr = virt_to_phys(rhp->rdev.status_page); in c4iw_alloc_ucontext() 214 struct c4iw_dev *rhp; in c4iw_deallocate_pd() local 218 rhp = php->rhp; in c4iw_deallocate_pd() 220 c4iw_put_resource(&rhp->rdev.resource.pdid_table, php->pdid); in c4iw_deallocate_pd() [all …]
|
D | iw_cxgb4.h | 350 static inline struct c4iw_cq *get_chp(struct c4iw_dev *rhp, u32 cqid) in get_chp() argument 352 return idr_find(&rhp->cqidr, cqid); in get_chp() 355 static inline struct c4iw_qp *get_qhp(struct c4iw_dev *rhp, u32 qpid) in get_qhp() argument 357 return idr_find(&rhp->qpidr, qpid); in get_qhp() 360 static inline struct c4iw_mr *get_mhp(struct c4iw_dev *rhp, u32 mmid) in get_mhp() argument 362 return idr_find(&rhp->mmidr, mmid); in get_mhp() 365 static inline int _insert_handle(struct c4iw_dev *rhp, struct idr *idr, in _insert_handle() argument 372 spin_lock_irq(&rhp->lock); in _insert_handle() 378 spin_unlock_irq(&rhp->lock); in _insert_handle() 385 static inline int insert_handle(struct c4iw_dev *rhp, struct idr *idr, in insert_handle() argument [all …]
|
D | cq.c | 352 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); in c4iw_flush_hw_cq() 804 c4iw_invalidate_mr(qhp->rhp, wc->ex.invalidate_rkey); in __c4iw_poll_cq_one() 845 c4iw_invalidate_mr(qhp->rhp, in __c4iw_poll_cq_one() 934 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); in c4iw_poll_cq_one() 979 remove_handle(chp->rhp, &chp->rhp->cqidr, chp->cq.cqid); in c4iw_destroy_cq() 985 destroy_cq(&chp->rhp->rdev, &chp->cq, in c4iw_destroy_cq() 1000 struct c4iw_dev *rhp; in c4iw_create_cq() local 1013 rhp = to_c4iw_dev(ibdev); in c4iw_create_cq() 1015 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq() 1056 hwentries = min(entries * 2, rhp->rdev.hw_queue.t4_max_iq_size); in c4iw_create_cq() [all …]
|
/Linux-v4.19/include/trace/events/ |
D | rcu.h | 474 TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen_lazy, 477 TP_ARGS(rcuname, rhp, qlen_lazy, qlen), 481 __field(void *, rhp) 489 __entry->rhp = rhp; 490 __entry->func = rhp->func; 496 __entry->rcuname, __entry->rhp, __entry->func, 510 TP_PROTO(const char *rcuname, struct rcu_head *rhp, unsigned long offset, 513 TP_ARGS(rcuname, rhp, offset, qlen_lazy, qlen), 517 __field(void *, rhp) 525 __entry->rhp = rhp; [all …]
|
/Linux-v4.19/kernel/rcu/ |
D | srcutiny.c | 126 struct rcu_head *rhp; in srcu_drive_gp() local 148 rhp = lh; in srcu_drive_gp() 151 rhp->func(rhp); in srcu_drive_gp() 171 void call_srcu(struct srcu_struct *sp, struct rcu_head *rhp, in call_srcu() argument 176 rhp->func = func; in call_srcu() 177 rhp->next = NULL; in call_srcu() 179 *sp->srcu_cb_tail = rhp; in call_srcu() 180 sp->srcu_cb_tail = &rhp->next; in call_srcu()
|
D | rcu_segcblist.c | 48 struct rcu_head *rhp; in rcu_cblist_dequeue() local 50 rhp = rclp->head; in rcu_cblist_dequeue() 51 if (!rhp) in rcu_cblist_dequeue() 54 rclp->head = rhp->next; in rcu_cblist_dequeue() 57 return rhp; in rcu_cblist_dequeue() 143 struct rcu_head *rhp, bool lazy) in rcu_segcblist_enqueue() argument 149 rhp->next = NULL; in rcu_segcblist_enqueue() 150 *rsclp->tails[RCU_NEXT_TAIL] = rhp; in rcu_segcblist_enqueue() 151 rsclp->tails[RCU_NEXT_TAIL] = &rhp->next; in rcu_segcblist_enqueue() 165 struct rcu_head *rhp, bool lazy) in rcu_segcblist_entrain() argument [all …]
|
D | srcutree.c | 818 static void srcu_leak_callback(struct rcu_head *rhp) in srcu_leak_callback() argument 850 void __call_srcu(struct srcu_struct *sp, struct rcu_head *rhp, in __call_srcu() argument 860 if (debug_rcu_head_queue(rhp)) { in __call_srcu() 862 WRITE_ONCE(rhp->func, srcu_leak_callback); in __call_srcu() 866 rhp->func = func; in __call_srcu() 870 rcu_segcblist_enqueue(&sdp->srcu_cblist, rhp, false); in __call_srcu() 907 void call_srcu(struct srcu_struct *sp, struct rcu_head *rhp, in call_srcu() argument 910 __call_srcu(sp, rhp, func, true); in call_srcu() 1019 static void srcu_barrier_cb(struct rcu_head *rhp) in srcu_barrier_cb() argument 1024 sdp = container_of(rhp, struct srcu_data, srcu_barrier_head); in srcu_barrier_cb() [all …]
|
D | rcuperf.c | 414 static void rcu_perf_async_cb(struct rcu_head *rhp) in rcu_perf_async_cb() argument 417 kfree(rhp); in rcu_perf_async_cb() 429 struct rcu_head *rhp = NULL; in rcu_perf_writer() local 463 if (!rhp) in rcu_perf_writer() 464 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_perf_writer() 465 if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { in rcu_perf_writer() 468 cur_ops->async(rhp, rcu_perf_async_cb); in rcu_perf_writer() 469 rhp = NULL; in rcu_perf_writer() 475 kfree(rhp); /* Because we are stopping. */ in rcu_perf_writer()
|
D | update.c | 431 void do_trace_rcu_torture_read(const char *rcutorturename, struct rcu_head *rhp, in do_trace_rcu_torture_read() argument 435 trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c); in do_trace_rcu_torture_read() 439 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument 555 void call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func) in call_rcu_tasks() argument 560 rhp->next = NULL; in call_rcu_tasks() 561 rhp->func = func; in call_rcu_tasks() 564 *rcu_tasks_cbs_tail = rhp; in call_rcu_tasks() 565 rcu_tasks_cbs_tail = &rhp->next; in call_rcu_tasks()
|
D | tree_plugin.h | 1666 static void rcu_oom_callback(struct rcu_head *rhp) in rcu_oom_callback() argument 1974 struct rcu_head *rhp; in rcu_nocb_cpu_needs_barrier() local 1992 rhp = READ_ONCE(rdp->nocb_head); in rcu_nocb_cpu_needs_barrier() 1993 if (!rhp) in rcu_nocb_cpu_needs_barrier() 1994 rhp = READ_ONCE(rdp->nocb_gp_head); in rcu_nocb_cpu_needs_barrier() 1995 if (!rhp) in rcu_nocb_cpu_needs_barrier() 1996 rhp = READ_ONCE(rdp->nocb_follower_head); in rcu_nocb_cpu_needs_barrier() 1999 if (!READ_ONCE(rdp->nocb_kthread) && rhp && in rcu_nocb_cpu_needs_barrier() 2003 cpu, rhp->func); in rcu_nocb_cpu_needs_barrier() 2020 struct rcu_head *rhp, in __call_rcu_nocb_enqueue() argument [all …]
|
D | rcutorture.c | 918 static void rcu_torture_cbflood_cb(struct rcu_head *rhp) in rcu_torture_cbflood_cb() argument 933 struct rcu_head *rhp; in rcu_torture_cbflood() local 940 rhp = vmalloc(array3_size(cbflood_n_burst, in rcu_torture_cbflood() 942 sizeof(*rhp))); in rcu_torture_cbflood() 943 err = !rhp; in rcu_torture_cbflood() 956 cur_ops->call(&rhp[i * cbflood_n_per_burst + j], in rcu_torture_cbflood() 965 vfree(rhp); in rcu_torture_cbflood() 1187 static void rcu_torture_timer_cb(struct rcu_head *rhp) in rcu_torture_timer_cb() argument 1189 kfree(rhp); in rcu_torture_timer_cb() 1374 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_NOWAIT); in rcu_torture_timer() local [all …]
|
D | rcu_segcblist.h | 120 struct rcu_head *rhp, bool lazy); 122 struct rcu_head *rhp, bool lazy);
|
D | sync.c | 164 static void rcu_sync_func(struct rcu_head *rhp) in rcu_sync_func() argument 166 struct rcu_sync *rsp = container_of(rhp, struct rcu_sync, cb_head); in rcu_sync_func()
|
D | rcu.h | 472 struct rcu_head *rhp, 486 struct rcu_head *rhp, 491 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
|
/Linux-v4.19/drivers/media/usb/pvrusb2/ |
D | pvrusb2-v4l2.c | 52 struct pvr2_ioread *rhp; member 933 if (fhp->rhp) { in pvr2_v4l2_release() 936 sp = pvr2_ioread_get_stream(fhp->rhp); in pvr2_v4l2_release() 938 pvr2_ioread_destroy(fhp->rhp); in pvr2_v4l2_release() 939 fhp->rhp = NULL; in pvr2_v4l2_release() 1063 if (fh->rhp) return 0; in pvr2_v4l2_iosetup() 1079 fh->rhp = pvr2_channel_create_mpeg_stream(fh->pdi->stream); in pvr2_v4l2_iosetup() 1080 if (!fh->rhp) { in pvr2_v4l2_iosetup() 1090 return pvr2_ioread_set_enabled(fh->rhp,!0); in pvr2_v4l2_iosetup() 1133 if (!fh->rhp) { in pvr2_v4l2_read() [all …]
|
/Linux-v4.19/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ |
D | misc.h | 33 #define trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
|
/Linux-v4.19/kernel/ |
D | pid.c | 117 static void delayed_put_pid(struct rcu_head *rhp) in delayed_put_pid() argument 119 struct pid *pid = container_of(rhp, struct pid, rcu); in delayed_put_pid()
|