Lines Matching refs:rcd

129 	struct hfi1_ctxtdata *rcd;  in hfi1_create_kctxt()  local
135 ret = hfi1_create_ctxtdata(ppd, dd->node, &rcd); in hfi1_create_kctxt()
146 rcd->flags = HFI1_CAP_KGET(MULTI_PKT_EGR) | in hfi1_create_kctxt()
152 if (rcd->ctxt == HFI1_CTRL_CTXT) in hfi1_create_kctxt()
153 rcd->flags |= HFI1_CAP_DMA_RTAIL; in hfi1_create_kctxt()
154 rcd->fast_handler = get_dma_rtail_setting(rcd) ? in hfi1_create_kctxt()
157 rcd->slow_handler = handle_receive_interrupt; in hfi1_create_kctxt()
159 hfi1_set_seq_cnt(rcd, 1); in hfi1_create_kctxt()
161 rcd->sc = sc_alloc(dd, SC_ACK, rcd->rcvhdrqentsize, dd->node); in hfi1_create_kctxt()
162 if (!rcd->sc) { in hfi1_create_kctxt()
166 hfi1_init_ctxt(rcd->sc); in hfi1_create_kctxt()
179 dd->rcd = kcalloc_node(dd->num_rcv_contexts, sizeof(*dd->rcd), in hfi1_create_kctxts()
181 if (!dd->rcd) in hfi1_create_kctxts()
192 for (i = 0; dd->rcd && i < dd->first_dyn_alloc_ctxt; ++i) in hfi1_create_kctxts()
193 hfi1_free_ctxt(dd->rcd[i]); in hfi1_create_kctxts()
196 kfree(dd->rcd); in hfi1_create_kctxts()
197 dd->rcd = NULL; in hfi1_create_kctxts()
204 static void hfi1_rcd_init(struct hfi1_ctxtdata *rcd) in hfi1_rcd_init() argument
206 kref_init(&rcd->kref); in hfi1_rcd_init()
217 struct hfi1_ctxtdata *rcd = in hfi1_rcd_free() local
220 spin_lock_irqsave(&rcd->dd->uctxt_lock, flags); in hfi1_rcd_free()
221 rcd->dd->rcd[rcd->ctxt] = NULL; in hfi1_rcd_free()
222 spin_unlock_irqrestore(&rcd->dd->uctxt_lock, flags); in hfi1_rcd_free()
224 hfi1_free_ctxtdata(rcd->dd, rcd); in hfi1_rcd_free()
226 kfree(rcd); in hfi1_rcd_free()
235 int hfi1_rcd_put(struct hfi1_ctxtdata *rcd) in hfi1_rcd_put() argument
237 if (rcd) in hfi1_rcd_put()
238 return kref_put(&rcd->kref, hfi1_rcd_free); in hfi1_rcd_put()
252 int hfi1_rcd_get(struct hfi1_ctxtdata *rcd) in hfi1_rcd_get() argument
254 return kref_get_unless_zero(&rcd->kref); in hfi1_rcd_get()
268 struct hfi1_ctxtdata *rcd, u16 *index) in allocate_rcd_index() argument
275 if (!dd->rcd[ctxt]) in allocate_rcd_index()
279 rcd->ctxt = ctxt; in allocate_rcd_index()
280 dd->rcd[ctxt] = rcd; in allocate_rcd_index()
281 hfi1_rcd_init(rcd); in allocate_rcd_index()
329 struct hfi1_ctxtdata *rcd = NULL; in hfi1_rcd_get_by_index() local
332 if (dd->rcd[ctxt]) { in hfi1_rcd_get_by_index()
333 rcd = dd->rcd[ctxt]; in hfi1_rcd_get_by_index()
334 if (!hfi1_rcd_get(rcd)) in hfi1_rcd_get_by_index()
335 rcd = NULL; in hfi1_rcd_get_by_index()
339 return rcd; in hfi1_rcd_get_by_index()
350 struct hfi1_ctxtdata *rcd; in hfi1_create_ctxtdata() local
358 rcd = kzalloc_node(sizeof(*rcd), GFP_KERNEL, numa); in hfi1_create_ctxtdata()
359 if (rcd) { in hfi1_create_ctxtdata()
364 ret = allocate_rcd_index(dd, rcd, &ctxt); in hfi1_create_ctxtdata()
367 kfree(rcd); in hfi1_create_ctxtdata()
371 INIT_LIST_HEAD(&rcd->qp_wait_list); in hfi1_create_ctxtdata()
372 hfi1_exp_tid_group_init(rcd); in hfi1_create_ctxtdata()
373 rcd->ppd = ppd; in hfi1_create_ctxtdata()
374 rcd->dd = dd; in hfi1_create_ctxtdata()
375 rcd->numa_id = numa; in hfi1_create_ctxtdata()
376 rcd->rcv_array_groups = dd->rcv_entries.ngroups; in hfi1_create_ctxtdata()
377 rcd->rhf_rcv_function_map = normal_rhf_rcv_functions; in hfi1_create_ctxtdata()
378 rcd->msix_intr = CCE_NUM_MSIX_VECTORS; in hfi1_create_ctxtdata()
380 mutex_init(&rcd->exp_mutex); in hfi1_create_ctxtdata()
381 spin_lock_init(&rcd->exp_lock); in hfi1_create_ctxtdata()
382 INIT_LIST_HEAD(&rcd->flow_queue.queue_head); in hfi1_create_ctxtdata()
383 INIT_LIST_HEAD(&rcd->rarr_queue.queue_head); in hfi1_create_ctxtdata()
385 hfi1_cdbg(PROC, "setting up context %u\n", rcd->ctxt); in hfi1_create_ctxtdata()
397 rcd->rcv_array_groups++; in hfi1_create_ctxtdata()
409 rcd->rcv_array_groups++; in hfi1_create_ctxtdata()
415 rcd->eager_base = base * dd->rcv_entries.group_size; in hfi1_create_ctxtdata()
417 rcd->rcvhdrq_cnt = rcvhdrcnt; in hfi1_create_ctxtdata()
418 rcd->rcvhdrqentsize = hfi1_hdrq_entsize; in hfi1_create_ctxtdata()
419 rcd->rhf_offset = in hfi1_create_ctxtdata()
420 rcd->rcvhdrqentsize - sizeof(u64) / sizeof(u32); in hfi1_create_ctxtdata()
432 max_entries = rcd->rcv_array_groups * in hfi1_create_ctxtdata()
435 rcd->egrbufs.count = round_down(rcvtids, in hfi1_create_ctxtdata()
437 if (rcd->egrbufs.count > MAX_EAGER_ENTRIES) { in hfi1_create_ctxtdata()
439 rcd->ctxt); in hfi1_create_ctxtdata()
440 rcd->egrbufs.count = MAX_EAGER_ENTRIES; in hfi1_create_ctxtdata()
444 rcd->ctxt, rcd->egrbufs.count); in hfi1_create_ctxtdata()
454 rcd->egrbufs.buffers = in hfi1_create_ctxtdata()
455 kcalloc_node(rcd->egrbufs.count, in hfi1_create_ctxtdata()
456 sizeof(*rcd->egrbufs.buffers), in hfi1_create_ctxtdata()
458 if (!rcd->egrbufs.buffers) in hfi1_create_ctxtdata()
460 rcd->egrbufs.rcvtids = in hfi1_create_ctxtdata()
461 kcalloc_node(rcd->egrbufs.count, in hfi1_create_ctxtdata()
462 sizeof(*rcd->egrbufs.rcvtids), in hfi1_create_ctxtdata()
464 if (!rcd->egrbufs.rcvtids) in hfi1_create_ctxtdata()
466 rcd->egrbufs.size = eager_buffer_size; in hfi1_create_ctxtdata()
472 if (rcd->egrbufs.size < hfi1_max_mtu) { in hfi1_create_ctxtdata()
473 rcd->egrbufs.size = __roundup_pow_of_two(hfi1_max_mtu); in hfi1_create_ctxtdata()
476 rcd->ctxt, rcd->egrbufs.size); in hfi1_create_ctxtdata()
478 rcd->egrbufs.rcvtid_size = HFI1_MAX_EAGER_BUFFER_SIZE; in hfi1_create_ctxtdata()
482 rcd->opstats = kzalloc_node(sizeof(*rcd->opstats), in hfi1_create_ctxtdata()
484 if (!rcd->opstats) in hfi1_create_ctxtdata()
488 hfi1_kern_init_ctxt_generations(rcd); in hfi1_create_ctxtdata()
491 *context = rcd; in hfi1_create_ctxtdata()
497 hfi1_free_ctxt(rcd); in hfi1_create_ctxtdata()
511 void hfi1_free_ctxt(struct hfi1_ctxtdata *rcd) in hfi1_free_ctxt() argument
513 hfi1_rcd_put(rcd); in hfi1_free_ctxt()
725 struct hfi1_ctxtdata *rcd; in init_after_reset() local
732 rcd = hfi1_rcd_get_by_index(dd, i); in init_after_reset()
735 HFI1_RCVCTRL_TAILUPD_DIS, rcd); in init_after_reset()
736 hfi1_rcd_put(rcd); in init_after_reset()
747 struct hfi1_ctxtdata *rcd; in enable_chip() local
759 rcd = hfi1_rcd_get_by_index(dd, i); in enable_chip()
760 if (!rcd) in enable_chip()
763 rcvmask |= HFI1_CAP_KGET_MASK(rcd->flags, DMA_RTAIL) ? in enable_chip()
765 if (!HFI1_CAP_KGET_MASK(rcd->flags, MULTI_PKT_EGR)) in enable_chip()
767 if (HFI1_CAP_KGET_MASK(rcd->flags, NODROP_RHQ_FULL)) in enable_chip()
769 if (HFI1_CAP_KGET_MASK(rcd->flags, NODROP_EGR_FULL)) in enable_chip()
773 hfi1_rcvctrl(dd, rcvmask, rcd); in enable_chip()
774 sc_enable(rcd->sc); in enable_chip()
775 hfi1_rcd_put(rcd); in enable_chip()
893 struct hfi1_ctxtdata *rcd; in hfi1_init() local
936 for (i = 0; dd->rcd && i < dd->first_dyn_alloc_ctxt; ++i) { in hfi1_init()
943 rcd = hfi1_rcd_get_by_index(dd, i); in hfi1_init()
944 if (!rcd) in hfi1_init()
947 rcd->do_interrupt = &handle_receive_interrupt; in hfi1_init()
949 lastfail = hfi1_create_rcvhdrq(dd, rcd); in hfi1_init()
951 lastfail = hfi1_setup_eagerbufs(rcd); in hfi1_init()
953 lastfail = hfi1_kern_exp_rcv_init(rcd, reinit); in hfi1_init()
960 hfi1_rcd_put(rcd); in hfi1_init()
1066 struct hfi1_ctxtdata *rcd; in shutdown_device() local
1091 rcd = hfi1_rcd_get_by_index(dd, i); in shutdown_device()
1096 HFI1_RCVCTRL_ONE_PKT_EGR_DIS, rcd); in shutdown_device()
1097 hfi1_rcd_put(rcd); in shutdown_device()
1145 void hfi1_free_ctxtdata(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd) in hfi1_free_ctxtdata() argument
1149 if (!rcd) in hfi1_free_ctxtdata()
1152 if (rcd->rcvhdrq) { in hfi1_free_ctxtdata()
1153 dma_free_coherent(&dd->pcidev->dev, rcvhdrq_size(rcd), in hfi1_free_ctxtdata()
1154 rcd->rcvhdrq, rcd->rcvhdrq_dma); in hfi1_free_ctxtdata()
1155 rcd->rcvhdrq = NULL; in hfi1_free_ctxtdata()
1156 if (hfi1_rcvhdrtail_kvaddr(rcd)) { in hfi1_free_ctxtdata()
1158 (void *)hfi1_rcvhdrtail_kvaddr(rcd), in hfi1_free_ctxtdata()
1159 rcd->rcvhdrqtailaddr_dma); in hfi1_free_ctxtdata()
1160 rcd->rcvhdrtail_kvaddr = NULL; in hfi1_free_ctxtdata()
1165 kfree(rcd->egrbufs.rcvtids); in hfi1_free_ctxtdata()
1166 rcd->egrbufs.rcvtids = NULL; in hfi1_free_ctxtdata()
1168 for (e = 0; e < rcd->egrbufs.alloced; e++) { in hfi1_free_ctxtdata()
1169 if (rcd->egrbufs.buffers[e].dma) in hfi1_free_ctxtdata()
1171 rcd->egrbufs.buffers[e].len, in hfi1_free_ctxtdata()
1172 rcd->egrbufs.buffers[e].addr, in hfi1_free_ctxtdata()
1173 rcd->egrbufs.buffers[e].dma); in hfi1_free_ctxtdata()
1175 kfree(rcd->egrbufs.buffers); in hfi1_free_ctxtdata()
1176 rcd->egrbufs.alloced = 0; in hfi1_free_ctxtdata()
1177 rcd->egrbufs.buffers = NULL; in hfi1_free_ctxtdata()
1179 sc_free(rcd->sc); in hfi1_free_ctxtdata()
1180 rcd->sc = NULL; in hfi1_free_ctxtdata()
1182 vfree(rcd->subctxt_uregbase); in hfi1_free_ctxtdata()
1183 vfree(rcd->subctxt_rcvegrbuf); in hfi1_free_ctxtdata()
1184 vfree(rcd->subctxt_rcvhdr_base); in hfi1_free_ctxtdata()
1185 kfree(rcd->opstats); in hfi1_free_ctxtdata()
1187 rcd->subctxt_uregbase = NULL; in hfi1_free_ctxtdata()
1188 rcd->subctxt_rcvegrbuf = NULL; in hfi1_free_ctxtdata()
1189 rcd->subctxt_rcvhdr_base = NULL; in hfi1_free_ctxtdata()
1190 rcd->opstats = NULL; in hfi1_free_ctxtdata()
1557 for (ctxt = 0; dd->rcd && ctxt < dd->num_rcv_contexts; ctxt++) { in cleanup_device_data()
1558 struct hfi1_ctxtdata *rcd = dd->rcd[ctxt]; in cleanup_device_data() local
1560 if (rcd) { in cleanup_device_data()
1561 hfi1_free_ctxt_rcv_groups(rcd); in cleanup_device_data()
1562 hfi1_free_ctxt(rcd); in cleanup_device_data()
1566 kfree(dd->rcd); in cleanup_device_data()
1567 dd->rcd = NULL; in cleanup_device_data()
1803 int hfi1_create_rcvhdrq(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd) in hfi1_create_rcvhdrq() argument
1807 if (!rcd->rcvhdrq) { in hfi1_create_rcvhdrq()
1810 amt = rcvhdrq_size(rcd); in hfi1_create_rcvhdrq()
1812 if (rcd->ctxt < dd->first_dyn_alloc_ctxt || rcd->is_vnic) in hfi1_create_rcvhdrq()
1816 rcd->rcvhdrq = dma_alloc_coherent(&dd->pcidev->dev, amt, in hfi1_create_rcvhdrq()
1817 &rcd->rcvhdrq_dma, in hfi1_create_rcvhdrq()
1820 if (!rcd->rcvhdrq) { in hfi1_create_rcvhdrq()
1823 amt, rcd->ctxt); in hfi1_create_rcvhdrq()
1827 if (HFI1_CAP_KGET_MASK(rcd->flags, DMA_RTAIL) || in hfi1_create_rcvhdrq()
1828 HFI1_CAP_UGET_MASK(rcd->flags, DMA_RTAIL)) { in hfi1_create_rcvhdrq()
1829 rcd->rcvhdrtail_kvaddr = dma_alloc_coherent(&dd->pcidev->dev, in hfi1_create_rcvhdrq()
1831 &rcd->rcvhdrqtailaddr_dma, in hfi1_create_rcvhdrq()
1833 if (!rcd->rcvhdrtail_kvaddr) in hfi1_create_rcvhdrq()
1838 set_hdrq_regs(rcd->dd, rcd->ctxt, rcd->rcvhdrqentsize, in hfi1_create_rcvhdrq()
1839 rcd->rcvhdrq_cnt); in hfi1_create_rcvhdrq()
1846 rcd->ctxt); in hfi1_create_rcvhdrq()
1847 dma_free_coherent(&dd->pcidev->dev, amt, rcd->rcvhdrq, in hfi1_create_rcvhdrq()
1848 rcd->rcvhdrq_dma); in hfi1_create_rcvhdrq()
1849 rcd->rcvhdrq = NULL; in hfi1_create_rcvhdrq()
1863 int hfi1_setup_eagerbufs(struct hfi1_ctxtdata *rcd) in hfi1_setup_eagerbufs() argument
1865 struct hfi1_devdata *dd = rcd->dd; in hfi1_setup_eagerbufs()
1887 if (rcd->egrbufs.size < (round_mtu * dd->rcv_entries.group_size)) in hfi1_setup_eagerbufs()
1888 rcd->egrbufs.size = round_mtu * dd->rcv_entries.group_size; in hfi1_setup_eagerbufs()
1893 if (!HFI1_CAP_KGET_MASK(rcd->flags, MULTI_PKT_EGR)) in hfi1_setup_eagerbufs()
1894 rcd->egrbufs.rcvtid_size = round_mtu; in hfi1_setup_eagerbufs()
1900 if (rcd->egrbufs.size <= (1 << 20)) in hfi1_setup_eagerbufs()
1901 rcd->egrbufs.rcvtid_size = max((unsigned long)round_mtu, in hfi1_setup_eagerbufs()
1902 rounddown_pow_of_two(rcd->egrbufs.size / 8)); in hfi1_setup_eagerbufs()
1904 while (alloced_bytes < rcd->egrbufs.size && in hfi1_setup_eagerbufs()
1905 rcd->egrbufs.alloced < rcd->egrbufs.count) { in hfi1_setup_eagerbufs()
1906 rcd->egrbufs.buffers[idx].addr = in hfi1_setup_eagerbufs()
1908 rcd->egrbufs.rcvtid_size, in hfi1_setup_eagerbufs()
1909 &rcd->egrbufs.buffers[idx].dma, in hfi1_setup_eagerbufs()
1911 if (rcd->egrbufs.buffers[idx].addr) { in hfi1_setup_eagerbufs()
1912 rcd->egrbufs.buffers[idx].len = in hfi1_setup_eagerbufs()
1913 rcd->egrbufs.rcvtid_size; in hfi1_setup_eagerbufs()
1914 rcd->egrbufs.rcvtids[rcd->egrbufs.alloced].addr = in hfi1_setup_eagerbufs()
1915 rcd->egrbufs.buffers[idx].addr; in hfi1_setup_eagerbufs()
1916 rcd->egrbufs.rcvtids[rcd->egrbufs.alloced].dma = in hfi1_setup_eagerbufs()
1917 rcd->egrbufs.buffers[idx].dma; in hfi1_setup_eagerbufs()
1918 rcd->egrbufs.alloced++; in hfi1_setup_eagerbufs()
1919 alloced_bytes += rcd->egrbufs.rcvtid_size; in hfi1_setup_eagerbufs()
1931 if (rcd->egrbufs.rcvtid_size == round_mtu || in hfi1_setup_eagerbufs()
1932 !HFI1_CAP_KGET_MASK(rcd->flags, MULTI_PKT_EGR)) { in hfi1_setup_eagerbufs()
1934 rcd->ctxt); in hfi1_setup_eagerbufs()
1939 new_size = rcd->egrbufs.rcvtid_size / 2; in hfi1_setup_eagerbufs()
1947 rcd->egrbufs.rcvtid_size = new_size; in hfi1_setup_eagerbufs()
1955 rcd->egrbufs.alloced = 0; in hfi1_setup_eagerbufs()
1957 if (i >= rcd->egrbufs.count) in hfi1_setup_eagerbufs()
1959 rcd->egrbufs.rcvtids[i].dma = in hfi1_setup_eagerbufs()
1960 rcd->egrbufs.buffers[j].dma + offset; in hfi1_setup_eagerbufs()
1961 rcd->egrbufs.rcvtids[i].addr = in hfi1_setup_eagerbufs()
1962 rcd->egrbufs.buffers[j].addr + offset; in hfi1_setup_eagerbufs()
1963 rcd->egrbufs.alloced++; in hfi1_setup_eagerbufs()
1964 if ((rcd->egrbufs.buffers[j].dma + offset + in hfi1_setup_eagerbufs()
1966 (rcd->egrbufs.buffers[j].dma + in hfi1_setup_eagerbufs()
1967 rcd->egrbufs.buffers[j].len)) { in hfi1_setup_eagerbufs()
1974 rcd->egrbufs.rcvtid_size = new_size; in hfi1_setup_eagerbufs()
1977 rcd->egrbufs.numbufs = idx; in hfi1_setup_eagerbufs()
1978 rcd->egrbufs.size = alloced_bytes; in hfi1_setup_eagerbufs()
1982 rcd->ctxt, rcd->egrbufs.alloced, in hfi1_setup_eagerbufs()
1983 rcd->egrbufs.rcvtid_size / 1024, rcd->egrbufs.size / 1024); in hfi1_setup_eagerbufs()
1990 rcd->egrbufs.threshold = in hfi1_setup_eagerbufs()
1991 rounddown_pow_of_two(rcd->egrbufs.alloced / 2); in hfi1_setup_eagerbufs()
1997 max_entries = rcd->rcv_array_groups * dd->rcv_entries.group_size; in hfi1_setup_eagerbufs()
1998 egrtop = roundup(rcd->egrbufs.alloced, dd->rcv_entries.group_size); in hfi1_setup_eagerbufs()
1999 rcd->expected_count = max_entries - egrtop; in hfi1_setup_eagerbufs()
2000 if (rcd->expected_count > MAX_TID_PAIR_ENTRIES * 2) in hfi1_setup_eagerbufs()
2001 rcd->expected_count = MAX_TID_PAIR_ENTRIES * 2; in hfi1_setup_eagerbufs()
2003 rcd->expected_base = rcd->eager_base + egrtop; in hfi1_setup_eagerbufs()
2005 rcd->ctxt, rcd->egrbufs.alloced, rcd->expected_count, in hfi1_setup_eagerbufs()
2006 rcd->eager_base, rcd->expected_base); in hfi1_setup_eagerbufs()
2008 if (!hfi1_rcvbuf_validate(rcd->egrbufs.rcvtid_size, PT_EAGER, &order)) { in hfi1_setup_eagerbufs()
2011 rcd->ctxt, rcd->egrbufs.rcvtid_size); in hfi1_setup_eagerbufs()
2016 for (idx = 0; idx < rcd->egrbufs.alloced; idx++) { in hfi1_setup_eagerbufs()
2017 hfi1_put_tid(dd, rcd->eager_base + idx, PT_EAGER, in hfi1_setup_eagerbufs()
2018 rcd->egrbufs.rcvtids[idx].dma, order); in hfi1_setup_eagerbufs()
2025 for (idx = 0; idx < rcd->egrbufs.alloced && in hfi1_setup_eagerbufs()
2026 rcd->egrbufs.buffers[idx].addr; in hfi1_setup_eagerbufs()
2029 rcd->egrbufs.buffers[idx].len, in hfi1_setup_eagerbufs()
2030 rcd->egrbufs.buffers[idx].addr, in hfi1_setup_eagerbufs()
2031 rcd->egrbufs.buffers[idx].dma); in hfi1_setup_eagerbufs()
2032 rcd->egrbufs.buffers[idx].addr = NULL; in hfi1_setup_eagerbufs()
2033 rcd->egrbufs.buffers[idx].dma = 0; in hfi1_setup_eagerbufs()
2034 rcd->egrbufs.buffers[idx].len = 0; in hfi1_setup_eagerbufs()