Home
last modified time | relevance | path

Searched refs:rhp (Results 1 – 25 of 40) sorted by relevance

12

/Linux-v4.19/drivers/infiniband/hw/cxgb3/
Diwch.h127 static inline int t3b_device(const struct iwch_dev *rhp) in t3b_device() argument
129 return rhp->rdev.t3cdev_p->type == T3B; in t3b_device()
132 static inline int t3a_device(const struct iwch_dev *rhp) in t3a_device() argument
134 return rhp->rdev.t3cdev_p->type == T3A; in t3a_device()
137 static inline struct iwch_cq *get_chp(struct iwch_dev *rhp, u32 cqid) in get_chp() argument
139 return idr_find(&rhp->cqidr, cqid); in get_chp()
142 static inline struct iwch_qp *get_qhp(struct iwch_dev *rhp, u32 qpid) in get_qhp() argument
144 return idr_find(&rhp->qpidr, qpid); in get_qhp()
147 static inline struct iwch_mr *get_mhp(struct iwch_dev *rhp, u32 mmid) in get_mhp() argument
149 return idr_find(&rhp->mmidr, mmid); in get_mhp()
[all …]
Diwch_provider.c66 struct iwch_dev *rhp = to_iwch_dev(context->device); in iwch_dealloc_ucontext() local
73 cxio_release_ucontext(&rhp->rdev, &ucontext->uctx); in iwch_dealloc_ucontext()
82 struct iwch_dev *rhp = to_iwch_dev(ibdev); in iwch_alloc_ucontext() local
88 cxio_init_ucontext(&rhp->rdev, &context->uctx); in iwch_alloc_ucontext()
101 remove_handle(chp->rhp, &chp->rhp->cqidr, chp->cq.cqid); in iwch_destroy_cq()
105 cxio_destroy_cq(&chp->rhp->rdev, &chp->cq); in iwch_destroy_cq()
116 struct iwch_dev *rhp; in iwch_create_cq() local
128 rhp = to_iwch_dev(ibdev); in iwch_create_cq()
135 if (!t3a_device(rhp)) { in iwch_create_cq()
144 if (t3a_device(rhp)) { in iwch_create_cq()
[all …]
Diwch_mem.c52 return insert_handle(mhp->rhp, &mhp->rhp->mmidr, mhp, mmid); in iwch_finish_mem_reg()
55 int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php, in iwch_register_mem() argument
61 if (cxio_register_phys_mem(&rhp->rdev, in iwch_register_mem()
73 cxio_dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in iwch_register_mem()
80 mhp->attr.pbl_addr = cxio_hal_pblpool_alloc(&mhp->rhp->rdev, in iwch_alloc_pbl()
93 cxio_hal_pblpool_free(&mhp->rhp->rdev, mhp->attr.pbl_addr, in iwch_free_pbl()
99 return cxio_write_pbl(&mhp->rhp->rdev, pages, in iwch_write_pbl()
Diwch_cq.c35 static int __iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, in __iwch_poll_cq_one() argument
47 if (t3a_device(chp->rhp) && credit) { in __iwch_poll_cq_one()
50 cxio_hal_cq_op(&rhp->rdev, &chp->cq, CQ_CREDIT_UPDATE, credit); in __iwch_poll_cq_one()
175 static int iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, in iwch_poll_cq_one() argument
187 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe)); in iwch_poll_cq_one()
190 ret = __iwch_poll_cq_one(rhp, chp, qhp, wc); in iwch_poll_cq_one()
193 ret = __iwch_poll_cq_one(rhp, chp, NULL, wc); in iwch_poll_cq_one()
200 struct iwch_dev *rhp; in iwch_poll_cq() local
207 rhp = chp->rhp; in iwch_poll_cq()
218 err = iwch_poll_cq_one(rhp, chp, wc + npolled); in iwch_poll_cq()
Diwch_provider.h47 struct iwch_dev *rhp; member
77 struct iwch_dev *rhp; member
93 struct iwch_dev *rhp; member
105 struct iwch_dev *rhp; member
163 struct iwch_dev *rhp; member
260 int iwch_modify_qp(struct iwch_dev *rhp,
339 int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php,
Diwch_qp.c201 static int iwch_sgl2pbl_map(struct iwch_dev *rhp, struct ib_sge *sg_list, in iwch_sgl2pbl_map() argument
209 mhp = get_mhp(rhp, (sg_list[i].lkey) >> 8); in iwch_sgl2pbl_map()
241 rhp->rdev.rnic_info.pbl_base) >> 3) + in iwch_sgl2pbl_map()
255 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, in build_rdma_recv()
302 pbl_addr = cxio_hal_pblpool_alloc(&qhp->rhp->rdev, T3_STAG0_PBL_SIZE); in build_zero_stag_recv()
309 pbl_offset = (pbl_addr - qhp->rhp->rdev.rnic_info.pbl_base) >> 3; in build_zero_stag_recv()
686 return iwch_cxgb3_ofld_send(ep->com.qp->rhp->rdev.t3cdev_p, skb); in iwch_post_zb_read()
717 return iwch_cxgb3_ofld_send(qhp->rhp->rdev.t3cdev_p, skb); in iwch_post_terminate()
777 rchp = get_chp(qhp->rhp, qhp->attr.rcq); in flush_qp()
778 schp = get_chp(qhp->rhp, qhp->attr.scq); in flush_qp()
[all …]
/Linux-v4.19/drivers/infiniband/hw/cxgb4/
Dmem.c398 return insert_handle(mhp->rhp, &mhp->rhp->mmidr, mhp, mmid); in finish_mem_reg()
401 static int register_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, in register_mem() argument
407 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, in register_mem()
420 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in register_mem()
429 mhp->attr.pbl_addr = c4iw_pblpool_alloc(&mhp->rhp->rdev, in alloc_pbl()
442 struct c4iw_dev *rhp; in c4iw_get_dma_mr() local
450 rhp = php->rhp; in c4iw_get_dma_mr()
468 mhp->rhp = rhp; in c4iw_get_dma_mr()
478 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, php->pdid, in c4iw_get_dma_mr()
490 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in c4iw_get_dma_mr()
[all …]
Dqp.c715 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl()
736 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl()
799 PBL_OFF(&mhp->rhp->rdev, mhp->attr.pbl_addr)>>3)); in build_tpte_memreg()
889 struct c4iw_dev *rhp; in free_qp_work() local
893 rhp = qhp->rhp; in free_qp_work()
896 destroy_qp(&rhp->rdev, &qhp->wq, in free_qp_work()
897 ucontext ? &ucontext->uctx : &rhp->rdev.uctx, !qhp->srq); in free_qp_work()
911 queue_work(qhp->rhp->rdev.free_workq, &qhp->free_work); in queue_qp_free()
936 spin_lock_irqsave(&qhp->rhp->lock, flags); in ring_kernel_sq_db()
938 if (qhp->rhp->db_state == NORMAL) in ring_kernel_sq_db()
[all …]
Dprovider.c64 struct c4iw_dev *rhp; in _c4iw_free_ucontext() local
68 rhp = to_c4iw_dev(ucontext->ibucontext.device); in _c4iw_free_ucontext()
73 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in _c4iw_free_ucontext()
90 struct c4iw_dev *rhp = to_c4iw_dev(ibdev); in c4iw_alloc_ucontext() local
102 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext()
109 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; in c4iw_alloc_ucontext()
130 mm->addr = virt_to_phys(rhp->rdev.status_page); in c4iw_alloc_ucontext()
214 struct c4iw_dev *rhp; in c4iw_deallocate_pd() local
218 rhp = php->rhp; in c4iw_deallocate_pd()
220 c4iw_put_resource(&rhp->rdev.resource.pdid_table, php->pdid); in c4iw_deallocate_pd()
[all …]
Diw_cxgb4.h350 static inline struct c4iw_cq *get_chp(struct c4iw_dev *rhp, u32 cqid) in get_chp() argument
352 return idr_find(&rhp->cqidr, cqid); in get_chp()
355 static inline struct c4iw_qp *get_qhp(struct c4iw_dev *rhp, u32 qpid) in get_qhp() argument
357 return idr_find(&rhp->qpidr, qpid); in get_qhp()
360 static inline struct c4iw_mr *get_mhp(struct c4iw_dev *rhp, u32 mmid) in get_mhp() argument
362 return idr_find(&rhp->mmidr, mmid); in get_mhp()
365 static inline int _insert_handle(struct c4iw_dev *rhp, struct idr *idr, in _insert_handle() argument
372 spin_lock_irq(&rhp->lock); in _insert_handle()
378 spin_unlock_irq(&rhp->lock); in _insert_handle()
385 static inline int insert_handle(struct c4iw_dev *rhp, struct idr *idr, in insert_handle() argument
[all …]
Dcq.c352 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); in c4iw_flush_hw_cq()
804 c4iw_invalidate_mr(qhp->rhp, wc->ex.invalidate_rkey); in __c4iw_poll_cq_one()
845 c4iw_invalidate_mr(qhp->rhp, in __c4iw_poll_cq_one()
934 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); in c4iw_poll_cq_one()
979 remove_handle(chp->rhp, &chp->rhp->cqidr, chp->cq.cqid); in c4iw_destroy_cq()
985 destroy_cq(&chp->rhp->rdev, &chp->cq, in c4iw_destroy_cq()
1000 struct c4iw_dev *rhp; in c4iw_create_cq() local
1013 rhp = to_c4iw_dev(ibdev); in c4iw_create_cq()
1015 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq()
1056 hwentries = min(entries * 2, rhp->rdev.hw_queue.t4_max_iq_size); in c4iw_create_cq()
[all …]
/Linux-v4.19/include/trace/events/
Drcu.h474 TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen_lazy,
477 TP_ARGS(rcuname, rhp, qlen_lazy, qlen),
481 __field(void *, rhp)
489 __entry->rhp = rhp;
490 __entry->func = rhp->func;
496 __entry->rcuname, __entry->rhp, __entry->func,
510 TP_PROTO(const char *rcuname, struct rcu_head *rhp, unsigned long offset,
513 TP_ARGS(rcuname, rhp, offset, qlen_lazy, qlen),
517 __field(void *, rhp)
525 __entry->rhp = rhp;
[all …]
/Linux-v4.19/kernel/rcu/
Dsrcutiny.c126 struct rcu_head *rhp; in srcu_drive_gp() local
148 rhp = lh; in srcu_drive_gp()
151 rhp->func(rhp); in srcu_drive_gp()
171 void call_srcu(struct srcu_struct *sp, struct rcu_head *rhp, in call_srcu() argument
176 rhp->func = func; in call_srcu()
177 rhp->next = NULL; in call_srcu()
179 *sp->srcu_cb_tail = rhp; in call_srcu()
180 sp->srcu_cb_tail = &rhp->next; in call_srcu()
Drcu_segcblist.c48 struct rcu_head *rhp; in rcu_cblist_dequeue() local
50 rhp = rclp->head; in rcu_cblist_dequeue()
51 if (!rhp) in rcu_cblist_dequeue()
54 rclp->head = rhp->next; in rcu_cblist_dequeue()
57 return rhp; in rcu_cblist_dequeue()
143 struct rcu_head *rhp, bool lazy) in rcu_segcblist_enqueue() argument
149 rhp->next = NULL; in rcu_segcblist_enqueue()
150 *rsclp->tails[RCU_NEXT_TAIL] = rhp; in rcu_segcblist_enqueue()
151 rsclp->tails[RCU_NEXT_TAIL] = &rhp->next; in rcu_segcblist_enqueue()
165 struct rcu_head *rhp, bool lazy) in rcu_segcblist_entrain() argument
[all …]
Dsrcutree.c818 static void srcu_leak_callback(struct rcu_head *rhp) in srcu_leak_callback() argument
850 void __call_srcu(struct srcu_struct *sp, struct rcu_head *rhp, in __call_srcu() argument
860 if (debug_rcu_head_queue(rhp)) { in __call_srcu()
862 WRITE_ONCE(rhp->func, srcu_leak_callback); in __call_srcu()
866 rhp->func = func; in __call_srcu()
870 rcu_segcblist_enqueue(&sdp->srcu_cblist, rhp, false); in __call_srcu()
907 void call_srcu(struct srcu_struct *sp, struct rcu_head *rhp, in call_srcu() argument
910 __call_srcu(sp, rhp, func, true); in call_srcu()
1019 static void srcu_barrier_cb(struct rcu_head *rhp) in srcu_barrier_cb() argument
1024 sdp = container_of(rhp, struct srcu_data, srcu_barrier_head); in srcu_barrier_cb()
[all …]
Drcuperf.c414 static void rcu_perf_async_cb(struct rcu_head *rhp) in rcu_perf_async_cb() argument
417 kfree(rhp); in rcu_perf_async_cb()
429 struct rcu_head *rhp = NULL; in rcu_perf_writer() local
463 if (!rhp) in rcu_perf_writer()
464 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_perf_writer()
465 if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { in rcu_perf_writer()
468 cur_ops->async(rhp, rcu_perf_async_cb); in rcu_perf_writer()
469 rhp = NULL; in rcu_perf_writer()
475 kfree(rhp); /* Because we are stopping. */ in rcu_perf_writer()
Dupdate.c431 void do_trace_rcu_torture_read(const char *rcutorturename, struct rcu_head *rhp, in do_trace_rcu_torture_read() argument
435 trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c); in do_trace_rcu_torture_read()
439 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
555 void call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func) in call_rcu_tasks() argument
560 rhp->next = NULL; in call_rcu_tasks()
561 rhp->func = func; in call_rcu_tasks()
564 *rcu_tasks_cbs_tail = rhp; in call_rcu_tasks()
565 rcu_tasks_cbs_tail = &rhp->next; in call_rcu_tasks()
Dtree_plugin.h1666 static void rcu_oom_callback(struct rcu_head *rhp) in rcu_oom_callback() argument
1974 struct rcu_head *rhp; in rcu_nocb_cpu_needs_barrier() local
1992 rhp = READ_ONCE(rdp->nocb_head); in rcu_nocb_cpu_needs_barrier()
1993 if (!rhp) in rcu_nocb_cpu_needs_barrier()
1994 rhp = READ_ONCE(rdp->nocb_gp_head); in rcu_nocb_cpu_needs_barrier()
1995 if (!rhp) in rcu_nocb_cpu_needs_barrier()
1996 rhp = READ_ONCE(rdp->nocb_follower_head); in rcu_nocb_cpu_needs_barrier()
1999 if (!READ_ONCE(rdp->nocb_kthread) && rhp && in rcu_nocb_cpu_needs_barrier()
2003 cpu, rhp->func); in rcu_nocb_cpu_needs_barrier()
2020 struct rcu_head *rhp, in __call_rcu_nocb_enqueue() argument
[all …]
Drcutorture.c918 static void rcu_torture_cbflood_cb(struct rcu_head *rhp) in rcu_torture_cbflood_cb() argument
933 struct rcu_head *rhp; in rcu_torture_cbflood() local
940 rhp = vmalloc(array3_size(cbflood_n_burst, in rcu_torture_cbflood()
942 sizeof(*rhp))); in rcu_torture_cbflood()
943 err = !rhp; in rcu_torture_cbflood()
956 cur_ops->call(&rhp[i * cbflood_n_per_burst + j], in rcu_torture_cbflood()
965 vfree(rhp); in rcu_torture_cbflood()
1187 static void rcu_torture_timer_cb(struct rcu_head *rhp) in rcu_torture_timer_cb() argument
1189 kfree(rhp); in rcu_torture_timer_cb()
1374 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_NOWAIT); in rcu_torture_timer() local
[all …]
Drcu_segcblist.h120 struct rcu_head *rhp, bool lazy);
122 struct rcu_head *rhp, bool lazy);
Dsync.c164 static void rcu_sync_func(struct rcu_head *rhp) in rcu_sync_func() argument
166 struct rcu_sync *rsp = container_of(rhp, struct rcu_sync, cb_head); in rcu_sync_func()
Drcu.h472 struct rcu_head *rhp,
486 struct rcu_head *rhp,
491 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
/Linux-v4.19/drivers/media/usb/pvrusb2/
Dpvrusb2-v4l2.c52 struct pvr2_ioread *rhp; member
933 if (fhp->rhp) { in pvr2_v4l2_release()
936 sp = pvr2_ioread_get_stream(fhp->rhp); in pvr2_v4l2_release()
938 pvr2_ioread_destroy(fhp->rhp); in pvr2_v4l2_release()
939 fhp->rhp = NULL; in pvr2_v4l2_release()
1063 if (fh->rhp) return 0; in pvr2_v4l2_iosetup()
1079 fh->rhp = pvr2_channel_create_mpeg_stream(fh->pdi->stream); in pvr2_v4l2_iosetup()
1080 if (!fh->rhp) { in pvr2_v4l2_iosetup()
1090 return pvr2_ioread_set_enabled(fh->rhp,!0); in pvr2_v4l2_iosetup()
1133 if (!fh->rhp) { in pvr2_v4l2_read()
[all …]
/Linux-v4.19/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/
Dmisc.h33 #define trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
/Linux-v4.19/kernel/
Dpid.c117 static void delayed_put_pid(struct rcu_head *rhp) in delayed_put_pid() argument
119 struct pid *pid = container_of(rhp, struct pid, rcu); in delayed_put_pid()

12