Home
last modified time | relevance | path

Searched refs:rcd (Results 1 – 25 of 48) sorted by relevance

12

/Linux-v4.19/drivers/infiniband/hw/hfi1/
Dinit.c130 struct hfi1_ctxtdata *rcd; in hfi1_create_kctxt() local
136 ret = hfi1_create_ctxtdata(ppd, dd->node, &rcd); in hfi1_create_kctxt()
147 rcd->flags = HFI1_CAP_KGET(MULTI_PKT_EGR) | in hfi1_create_kctxt()
153 if (rcd->ctxt == HFI1_CTRL_CTXT) in hfi1_create_kctxt()
154 rcd->flags |= HFI1_CAP_DMA_RTAIL; in hfi1_create_kctxt()
155 rcd->seq_cnt = 1; in hfi1_create_kctxt()
157 rcd->sc = sc_alloc(dd, SC_ACK, rcd->rcvhdrqentsize, dd->node); in hfi1_create_kctxt()
158 if (!rcd->sc) { in hfi1_create_kctxt()
162 hfi1_init_ctxt(rcd->sc); in hfi1_create_kctxt()
175 dd->rcd = kcalloc_node(dd->num_rcv_contexts, sizeof(*dd->rcd), in hfi1_create_kctxts()
[all …]
Daspm.h176 static inline void aspm_ctx_disable(struct hfi1_ctxtdata *rcd) in aspm_ctx_disable() argument
184 if (!rcd->aspm_intr_supported) in aspm_ctx_disable()
187 spin_lock_irqsave(&rcd->aspm_lock, flags); in aspm_ctx_disable()
189 if (!rcd->aspm_intr_enable) in aspm_ctx_disable()
192 prev = rcd->aspm_ts_last_intr; in aspm_ctx_disable()
194 rcd->aspm_ts_last_intr = now; in aspm_ctx_disable()
200 restart_timer = ktime_to_ns(ktime_sub(now, rcd->aspm_ts_timer_sched)) > in aspm_ctx_disable()
205 if (rcd->aspm_enabled && close_interrupts) { in aspm_ctx_disable()
206 aspm_disable_inc(rcd->dd); in aspm_ctx_disable()
207 rcd->aspm_enabled = false; in aspm_ctx_disable()
[all …]
Dexp_rcv.c65 void hfi1_exp_tid_group_init(struct hfi1_ctxtdata *rcd) in hfi1_exp_tid_group_init() argument
67 hfi1_exp_tid_set_init(&rcd->tid_group_list); in hfi1_exp_tid_group_init()
68 hfi1_exp_tid_set_init(&rcd->tid_used_list); in hfi1_exp_tid_group_init()
69 hfi1_exp_tid_set_init(&rcd->tid_full_list); in hfi1_exp_tid_group_init()
76 int hfi1_alloc_ctxt_rcv_groups(struct hfi1_ctxtdata *rcd) in hfi1_alloc_ctxt_rcv_groups() argument
78 struct hfi1_devdata *dd = rcd->dd; in hfi1_alloc_ctxt_rcv_groups()
84 ngroups = rcd->expected_count / dd->rcv_entries.group_size; in hfi1_alloc_ctxt_rcv_groups()
85 rcd->groups = in hfi1_alloc_ctxt_rcv_groups()
86 kcalloc_node(ngroups, sizeof(*rcd->groups), in hfi1_alloc_ctxt_rcv_groups()
87 GFP_KERNEL, rcd->numa_id); in hfi1_alloc_ctxt_rcv_groups()
[all …]
Ddriver.c201 static inline void *get_egrbuf(const struct hfi1_ctxtdata *rcd, u64 rhf, in get_egrbuf() argument
206 *update |= !(idx & (rcd->egrbufs.threshold - 1)) && !offset; in get_egrbuf()
207 return (void *)(((u64)(rcd->egrbufs.rcvtids[idx].addr)) + in get_egrbuf()
211 static inline void *hfi1_get_header(struct hfi1_ctxtdata *rcd, in hfi1_get_header() argument
216 return (void *)(rhf_addr - rcd->rhf_offset + offset); in hfi1_get_header()
219 static inline struct ib_header *hfi1_get_msgheader(struct hfi1_ctxtdata *rcd, in hfi1_get_msgheader() argument
222 return (struct ib_header *)hfi1_get_header(rcd, rhf_addr); in hfi1_get_msgheader()
226 *hfi1_get_16B_header(struct hfi1_ctxtdata *rcd, in hfi1_get_16B_header() argument
229 return (struct hfi1_16b_header *)hfi1_get_header(rcd, rhf_addr); in hfi1_get_16B_header()
252 static void rcv_hdrerr(struct hfi1_ctxtdata *rcd, struct hfi1_pportdata *ppd, in rcv_hdrerr() argument
[all …]
Dintr.c244 void handle_user_interrupt(struct hfi1_ctxtdata *rcd) in handle_user_interrupt() argument
246 struct hfi1_devdata *dd = rcd->dd; in handle_user_interrupt()
250 if (bitmap_empty(rcd->in_use_ctxts, HFI1_MAX_SHARED_CTXTS)) in handle_user_interrupt()
253 if (test_and_clear_bit(HFI1_CTXT_WAITING_RCV, &rcd->event_flags)) { in handle_user_interrupt()
254 wake_up_interruptible(&rcd->wait); in handle_user_interrupt()
255 hfi1_rcvctrl(dd, HFI1_RCVCTRL_INTRAVAIL_DIS, rcd); in handle_user_interrupt()
257 &rcd->event_flags)) { in handle_user_interrupt()
258 rcd->urgent++; in handle_user_interrupt()
259 wake_up_interruptible(&rcd->wait); in handle_user_interrupt()
Dexp_rcv.h192 hfi1_tid_group_to_idx(struct hfi1_ctxtdata *rcd, struct tid_group *grp) in hfi1_tid_group_to_idx() argument
194 return grp - &rcd->groups[0]; in hfi1_tid_group_to_idx()
203 hfi1_idx_to_tid_group(struct hfi1_ctxtdata *rcd, u16 idx) in hfi1_idx_to_tid_group() argument
205 return &rcd->groups[idx]; in hfi1_idx_to_tid_group()
208 int hfi1_alloc_ctxt_rcv_groups(struct hfi1_ctxtdata *rcd);
209 void hfi1_free_ctxt_rcv_groups(struct hfi1_ctxtdata *rcd);
210 void hfi1_exp_tid_group_init(struct hfi1_ctxtdata *rcd);
Dtrace_rx.h68 TP_STRUCT__entry(DD_DEV_ENTRY(packet->rcd->dd)
77 TP_fast_assign(DD_DEV_ASSIGN(packet->rcd->dd);
79 __entry->ctxt = packet->rcd->ctxt;
100 TP_PROTO(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd),
101 TP_ARGS(dd, rcd),
108 __entry->ctxt = rcd->ctxt;
109 if (rcd->do_interrupt ==
113 } else if (rcd->do_interrupt ==
117 } else if (rcd->do_interrupt ==
Dhfi.h215 int (*do_interrupt)(struct hfi1_ctxtdata *rcd, int threaded);
329 static inline u32 rcvhdrq_size(struct hfi1_ctxtdata *rcd) in rcvhdrq_size() argument
331 return PAGE_ALIGN(rcd->rcvhdrq_cnt * in rcvhdrq_size()
332 rcd->rcvhdrqentsize * sizeof(u32)); in rcvhdrq_size()
346 struct hfi1_ctxtdata *rcd; member
1301 struct hfi1_ctxtdata **rcd; member
1414 void handle_user_interrupt(struct hfi1_ctxtdata *rcd);
1416 int hfi1_create_rcvhdrq(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd);
1417 int hfi1_setup_eagerbufs(struct hfi1_ctxtdata *rcd);
1420 struct hfi1_ctxtdata **rcd);
[all …]
Ddebugfs.c139 struct hfi1_ctxtdata *rcd; in _opcode_stats_seq_show() local
142 rcd = hfi1_rcd_get_by_index(dd, j); in _opcode_stats_seq_show()
143 if (rcd) { in _opcode_stats_seq_show()
144 n_packets += rcd->opstats->stats[i].n_packets; in _opcode_stats_seq_show()
145 n_bytes += rcd->opstats->stats[i].n_bytes; in _opcode_stats_seq_show()
147 hfi1_rcd_put(rcd); in _opcode_stats_seq_show()
230 struct hfi1_ctxtdata *rcd; in _ctx_stats_seq_show() local
240 rcd = hfi1_rcd_get_by_index_safe(dd, i); in _ctx_stats_seq_show()
241 if (!rcd) in _ctx_stats_seq_show()
244 for (j = 0; j < ARRAY_SIZE(rcd->opstats->stats); j++) in _ctx_stats_seq_show()
[all …]
Drc.c740 ibp = rcd_to_iport(packet->rcd); in hfi1_queue_rc_ack()
761 struct hfi1_ibport *ibp = rcd_to_iport(packet->rcd); in hfi1_make_rc_ack_9B()
809 struct hfi1_ibport *ibp = rcd_to_iport(packet->rcd); in hfi1_make_rc_ack_16B()
877 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_send_rc_ack() local
879 struct hfi1_ibport *ibp = rcd_to_iport(rcd); in hfi1_send_rc_ack()
916 pbuf = sc_buffer_alloc(rcd->sc, plen, NULL, NULL); in hfi1_send_rc_ack()
1282 u64 val, struct hfi1_ctxtdata *rcd) in do_rc_ack() argument
1302 ibp = rcd_to_iport(rcd); in do_rc_ack()
1343 &rcd->qp_wait_list); in do_rc_ack()
1500 struct hfi1_ctxtdata *rcd) in rdma_seq_err() argument
[all …]
Dfault.c92 struct hfi1_ctxtdata *rcd; in _fault_stats_seq_show() local
95 rcd = hfi1_rcd_get_by_index(dd, j); in _fault_stats_seq_show()
96 if (rcd) { in _fault_stats_seq_show()
97 n_packets += rcd->opstats->stats[i].n_packets; in _fault_stats_seq_show()
98 n_bytes += rcd->opstats->stats[i].n_bytes; in _fault_stats_seq_show()
100 hfi1_rcd_put(rcd); in _fault_stats_seq_show()
367 struct hfi1_ibdev *ibd = &packet->rcd->dd->verbs_dev; in hfi1_dbg_should_fault_rx()
Dtrace_misc.h114 TP_STRUCT__entry(DD_DEV_ENTRY(packet->rcd->ppd->dd)
122 TP_fast_assign(DD_DEV_ASSIGN(packet->rcd->ppd->dd);
124 __entry->ctxt = packet->rcd->ctxt;
/Linux-v4.19/drivers/infiniband/hw/qib/
Dqib_file_ops.c101 struct qib_ctxtdata *rcd = ctxt_fp(fp); in qib_get_base_info() local
104 struct qib_devdata *dd = rcd->dd; in qib_get_base_info()
105 struct qib_pportdata *ppd = rcd->ppd; in qib_get_base_info()
110 subctxt_cnt = rcd->subctxt_cnt; in qib_get_base_info()
135 ret = dd->f_get_base_info(rcd, kinfo); in qib_get_base_info()
141 kinfo->spi_tidegrcnt = rcd->rcvegrcnt; in qib_get_base_info()
147 rcd->rcvegrbuf_chunks * rcd->rcvegrbuf_size; in qib_get_base_info()
148 kinfo->spi_rcv_egrperchunk = rcd->rcvegrbufs_perchunk; in qib_get_base_info()
150 rcd->rcvegrbuf_chunks; in qib_get_base_info()
184 kinfo->spi_rcvhdr_base = (u64) rcd->rcvhdrq_phys; in qib_get_base_info()
[all …]
Dqib_init.c135 dd->rcd = kcalloc(dd->ctxtcnt, sizeof(*dd->rcd), GFP_KERNEL); in qib_create_ctxts()
136 if (!dd->rcd) in qib_create_ctxts()
142 struct qib_ctxtdata *rcd; in qib_create_ctxts() local
149 rcd = qib_create_ctxtdata(ppd, i, dd->assigned_node_id); in qib_create_ctxts()
150 if (!rcd) { in qib_create_ctxts()
153 kfree(dd->rcd); in qib_create_ctxts()
154 dd->rcd = NULL; in qib_create_ctxts()
157 rcd->pkeys[0] = QIB_DEFAULT_P_KEY; in qib_create_ctxts()
158 rcd->seq_cnt = 1; in qib_create_ctxts()
170 struct qib_ctxtdata *rcd; in qib_create_ctxtdata() local
[all …]
Dqib_tx.c80 int qib_disarm_piobufs_ifneeded(struct qib_ctxtdata *rcd) in qib_disarm_piobufs_ifneeded() argument
82 struct qib_devdata *dd = rcd->dd; in qib_disarm_piobufs_ifneeded()
87 last = rcd->pio_base + rcd->piocnt; in qib_disarm_piobufs_ifneeded()
93 if (rcd->user_event_mask) { in qib_disarm_piobufs_ifneeded()
98 clear_bit(_QIB_EVENT_DISARM_BUFS_BIT, &rcd->user_event_mask[0]); in qib_disarm_piobufs_ifneeded()
99 for (i = 1; i < rcd->subctxt_cnt; i++) in qib_disarm_piobufs_ifneeded()
101 &rcd->user_event_mask[i]); in qib_disarm_piobufs_ifneeded()
104 for (i = rcd->pio_base; i < last; i++) { in qib_disarm_piobufs_ifneeded()
107 dd->f_sendctrl(rcd->ppd, QIB_SENDCTRL_DISARM_BUF(i)); in qib_disarm_piobufs_ifneeded()
134 struct qib_ctxtdata *rcd; in find_ctxt() local
[all …]
Dqib_intr.c191 struct qib_ctxtdata *rcd; in qib_handle_urcv() local
196 for (i = dd->first_user_ctxt; dd->rcd && i < dd->cfgctxts; i++) { in qib_handle_urcv()
199 rcd = dd->rcd[i]; in qib_handle_urcv()
200 if (!rcd || !rcd->cnt) in qib_handle_urcv()
203 if (test_and_clear_bit(QIB_CTXT_WAITING_RCV, &rcd->flag)) { in qib_handle_urcv()
204 wake_up_interruptible(&rcd->wait); in qib_handle_urcv()
205 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_INTRAVAIL_DIS, in qib_handle_urcv()
206 rcd->ctxt); in qib_handle_urcv()
208 &rcd->flag)) { in qib_handle_urcv()
209 rcd->urgent++; in qib_handle_urcv()
[all …]
Dqib_driver.c283 static inline void *qib_get_egrbuf(const struct qib_ctxtdata *rcd, u32 etail) in qib_get_egrbuf() argument
285 const u32 chunk = etail >> rcd->rcvegrbufs_perchunk_shift; in qib_get_egrbuf()
286 const u32 idx = etail & ((u32)rcd->rcvegrbufs_perchunk - 1); in qib_get_egrbuf()
288 return rcd->rcvegrbuf[chunk] + (idx << rcd->dd->rcvegrbufsize_shift); in qib_get_egrbuf()
295 static u32 qib_rcv_hdrerr(struct qib_ctxtdata *rcd, struct qib_pportdata *ppd, in qib_rcv_hdrerr() argument
409 &rcd->qp_wait_list); in qib_rcv_hdrerr()
444 u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts) in qib_kreceive() argument
446 struct qib_devdata *dd = rcd->dd; in qib_kreceive()
447 struct qib_pportdata *ppd = rcd->ppd; in qib_kreceive()
459 l = rcd->head; in qib_kreceive()
[all …]
Dqib_debugfs.c112 if (!dd->rcd[j]) in _opcode_stats_seq_show()
114 n_packets += dd->rcd[j]->opstats->stats[i].n_packets; in _opcode_stats_seq_show()
115 n_bytes += dd->rcd[j]->opstats->stats[i].n_bytes; in _opcode_stats_seq_show()
175 if (!dd->rcd[i]) in _ctx_stats_seq_show()
178 for (j = 0; j < ARRAY_SIZE(dd->rcd[i]->opstats->stats); j++) in _ctx_stats_seq_show()
179 n_packets += dd->rcd[i]->opstats->stats[j].n_packets; in _ctx_stats_seq_show()
Dqib.h729 struct qib_ctxtdata **rcd; /* Receive Context Data */ member
1101 struct qib_ctxtdata *rcd; member
1175 (((struct qib_filedata *)(fp)->private_data)->rcd)
1352 static inline void qib_clear_rcvhdrtail(const struct qib_ctxtdata *rcd) in qib_clear_rcvhdrtail() argument
1354 *((u64 *) rcd->rcvhdrtail_kvaddr) = 0ULL; in qib_clear_rcvhdrtail()
1357 static inline u32 qib_get_rcvhdrtail(const struct qib_ctxtdata *rcd) in qib_get_rcvhdrtail() argument
1364 *((volatile __le64 *)rcd->rcvhdrtail_kvaddr)); /* DMA'ed */ in qib_get_rcvhdrtail()
1367 static inline u32 qib_get_hdrqtail(const struct qib_ctxtdata *rcd) in qib_get_hdrqtail() argument
1369 const struct qib_devdata *dd = rcd->dd; in qib_get_hdrqtail()
1376 rhf_addr = (__le32 *) rcd->rcvhdrq + in qib_get_hdrqtail()
[all …]
Dqib_iba7322.c71 struct qib_ctxtdata *rcd);
2694 static void qib_update_rhdrq_dca(struct qib_ctxtdata *rcd, int cpu) in qib_update_rhdrq_dca() argument
2696 struct qib_devdata *dd = rcd->dd; in qib_update_rhdrq_dca()
2701 if (cspec->rhdr_cpu[rcd->ctxt] != cpu) { in qib_update_rhdrq_dca()
2704 cspec->rhdr_cpu[rcd->ctxt] = cpu; in qib_update_rhdrq_dca()
2705 rmp = &dca_rcvhdr_reg_map[rcd->ctxt]; in qib_update_rhdrq_dca()
2710 "Ctxt %d cpu %d dca %llx\n", rcd->ctxt, cpu, in qib_update_rhdrq_dca()
2796 struct qib_ctxtdata *rcd = (struct qib_ctxtdata *)n->arg; in qib_irq_notifier_notify() local
2798 qib_update_rhdrq_dca(rcd, cpu); in qib_irq_notifier_notify()
2813 struct qib_ctxtdata *rcd = (struct qib_ctxtdata *)n->arg; in qib_irq_notifier_release() local
[all …]
/Linux-v4.19/arch/x86/kernel/cpu/mcheck/
Dmce-apei.c85 struct cper_mce_record rcd; in apei_write_mce() local
87 memset(&rcd, 0, sizeof(rcd)); in apei_write_mce()
88 memcpy(rcd.hdr.signature, CPER_SIG_RECORD, CPER_SIG_SIZE); in apei_write_mce()
89 rcd.hdr.revision = CPER_RECORD_REV; in apei_write_mce()
90 rcd.hdr.signature_end = CPER_SIG_END; in apei_write_mce()
91 rcd.hdr.section_count = 1; in apei_write_mce()
92 rcd.hdr.error_severity = CPER_SEV_FATAL; in apei_write_mce()
94 rcd.hdr.validation_bits = 0; in apei_write_mce()
95 rcd.hdr.record_length = sizeof(rcd); in apei_write_mce()
96 rcd.hdr.creator_id = CPER_CREATOR_MCE; in apei_write_mce()
[all …]
/Linux-v4.19/drivers/media/rc/
Dzx-irdec.c37 struct rc_dev *rcd; member
64 rc_repeat(irdec->rcd); in zx_irdec_irq()
77 rc_keydown(irdec->rcd, rc_proto, scancode, 0); in zx_irdec_irq()
88 struct rc_dev *rcd; in zx_irdec_probe() local
105 rcd = devm_rc_allocate_device(dev, RC_DRIVER_SCANCODE); in zx_irdec_probe()
106 if (!rcd) { in zx_irdec_probe()
111 irdec->rcd = rcd; in zx_irdec_probe()
113 rcd->priv = irdec; in zx_irdec_probe()
114 rcd->input_phys = DRIVER_NAME "/input0"; in zx_irdec_probe()
115 rcd->input_id.bustype = BUS_HOST; in zx_irdec_probe()
[all …]
/Linux-v4.19/drivers/acpi/apei/
Derst.c985 struct cper_pstore_record *rcd; in erst_reader() local
986 size_t rcd_len = sizeof(*rcd) + erst_info.bufsize; in erst_reader()
991 rcd = kmalloc(rcd_len, GFP_KERNEL); in erst_reader()
992 if (!rcd) { in erst_reader()
1007 len = erst_read(record_id, &rcd->hdr, rcd_len); in erst_reader()
1011 else if (len < 0 || len < sizeof(*rcd)) { in erst_reader()
1015 if (uuid_le_cmp(rcd->hdr.creator_id, CPER_CREATOR_PSTORE) != 0) in erst_reader()
1023 memcpy(record->buf, rcd->data, len - sizeof(*rcd)); in erst_reader()
1027 if (uuid_le_cmp(rcd->sec_hdr.section_type, in erst_reader()
1031 } else if (uuid_le_cmp(rcd->sec_hdr.section_type, in erst_reader()
[all …]
/Linux-v4.19/drivers/net/vmxnet3/
Dvmxnet3_drv.c290 # define vmxnet3_getRxComp(dstrcd, rcd, tmp) do { \ argument
292 vmxnet3_RxCompToCPU((rcd), (tmp)); \
305 # define vmxnet3_getRxComp(dstrcd, rcd, tmp) (dstrcd) = (rcd) argument
657 vmxnet3_append_frag(struct sk_buff *skb, struct Vmxnet3_RxCompDesc *rcd, in vmxnet3_append_frag() argument
667 skb_frag_size_set(frag, rcd->len); in vmxnet3_append_frag()
668 skb->data_len += rcd->len; in vmxnet3_append_frag()
1170 if (!gdesc->rcd.cnc && adapter->netdev->features & NETIF_F_RXCSUM) { in vmxnet3_rx_csum()
1171 if (gdesc->rcd.v4 && in vmxnet3_rx_csum()
1175 BUG_ON(!(gdesc->rcd.tcp || gdesc->rcd.udp)); in vmxnet3_rx_csum()
1176 BUG_ON(gdesc->rcd.frg); in vmxnet3_rx_csum()
[all …]
/Linux-v4.19/drivers/cpufreq/
Dsa1110-cpufreq.c133 static inline void set_mdcas(u_int *mdcas, int delayed, u_int rcd) in set_mdcas() argument
137 rcd = 2 * rcd - 1; in set_mdcas()
138 shift = delayed + 1 + rcd; in set_mdcas()
140 mdcas[0] = (1 << rcd) - 1; in set_mdcas()

12