/Linux-v5.4/lib/ |
D | test_xarray.c | 59 XA_BUG_ON(xa, xa_load(xa, index) != NULL); in xa_erase_index() 149 void *entry = xa_load(xa, j); in check_xa_load() 160 void *entry = xa_load(xa, j); in check_xa_load() 318 XA_BUG_ON(xa, xa_load(xa, 1) != NULL); in check_xa_shrink() 323 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0)); in check_xa_shrink() 330 XA_BUG_ON(xa, xa_load(xa, max) != xa_mk_value(0)); in check_xa_shrink() 331 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink() 340 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink() 353 XA_BUG_ON(xa, xa_load(xa, i - 1) != NULL); in check_insert() 354 XA_BUG_ON(xa, xa_load(xa, i + 1) != NULL); in check_insert() [all …]
|
D | xarray.c | 1289 void *xa_load(struct xarray *xa, unsigned long index) in xa_load() function 1304 EXPORT_SYMBOL(xa_load);
|
/Linux-v5.4/drivers/infiniband/hw/cxgb3/ |
D | iwch.h | 138 return xa_load(&rhp->cqs, cqid); in get_chp() 143 return xa_load(&rhp->qps, qpid); in get_qhp() 148 return xa_load(&rhp->mrs, mmid); in get_mhp()
|
D | iwch_ev.c | 52 qhp = xa_load(&rnicp->qps, CQE_QPID(rsp_msg->cqe)); in post_qp_event() 119 qhp = xa_load(&rnicp->qps, CQE_QPID(rsp_msg->cqe)); in iwch_ev_dispatch()
|
/Linux-v5.4/drivers/infiniband/hw/cxgb4/ |
D | ev.c | 127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch() 229 chp = xa_load(&dev->cqs, qid); in c4iw_ev_handler()
|
D | iw_cxgb4.h | 351 return xa_load(&rhp->cqs, cqid); in get_chp() 356 return xa_load(&rhp->qps, qpid); in get_qhp()
|
/Linux-v5.4/drivers/gpu/drm/lima/ |
D | lima_ctx.c | 69 ctx = xa_load(&mgr->handles, id); in lima_ctx_get()
|
/Linux-v5.4/drivers/infiniband/hw/mlx5/ |
D | devx.c | 1318 event = xa_load(&dev->devx_event_table.event_xa, in devx_cleanup_subscription() 1322 xa_val_level2 = xa_load(&event->object_ids, sub->xa_key_level2); in devx_cleanup_subscription() 1384 event = xa_load(&table->event_xa, MLX5_EVENT_TYPE_COMP); in devx_cq_comp() 1388 obj_event = xa_load(&event->object_ids, obj_id); in devx_cq_comp() 1785 event = xa_load(&devx_event_table->event_xa, key_level1); in subscribe_event_xa_dealloc() 1788 xa_val_level2 = xa_load(&event->object_ids, in subscribe_event_xa_dealloc() 1807 event = xa_load(&devx_event_table->event_xa, key_level1); in subscribe_event_xa_alloc() 1829 obj_event = xa_load(&event->object_ids, key_level2); in subscribe_event_xa_alloc() 2057 event = xa_load(&devx_event_table->event_xa, in UVERBS_HANDLER() 2067 obj_event = xa_load(&event->object_ids, obj_id); in UVERBS_HANDLER() [all …]
|
D | srq_cmd.c | 87 srq = xa_load(&table->array, srqn); in mlx5_cmd_get_srq() 676 srq = xa_load(&table->array, srqn); in srq_event_notifier()
|
D | odp.c | 796 mmkey = xa_load(&dev->mdev->priv.mkey_table, mlx5_base_mkey(key)); in pagefault_single_data_segment() 1646 mmkey = xa_load(&dev->mdev->priv.mkey_table, in num_pending_prefetch_dec() 1666 mmkey = xa_load(&dev->mdev->priv.mkey_table, in num_pending_prefetch_inc()
|
/Linux-v5.4/drivers/infiniband/core/ |
D | device.c | 315 device = xa_load(&devices, index); in ib_device_get_by_index() 434 struct ib_client *client = xa_load(&clients, index); in ib_device_rename() 628 if (xa_load(&devices, device->index) == device) in ib_dealloc_device() 711 client_data = xa_load(&device->client_data, client_id); in remove_client_context() 713 client = xa_load(&clients, client_id); in remove_client_context() 883 cdev = xa_load(&device->compat_devs, rnet->id); in add_one_compat_dev() 1706 if (xa_load(&clients, highest_client_id - 1)) in remove_client_id() 1831 struct ib_client *client = xa_load(&clients, index); in __ib_get_client_nl_info() 2515 struct ib_client *client = xa_load(&clients, index); in ib_get_net_dev_by_params()
|
D | restrack.c | 297 res = xa_load(&rt->xa, id); in rdma_restrack_get_byid()
|
D | rdma_core.c | 327 uobj = xa_load(&ufile->idr, id); in lookup_get_idr_uobject()
|
/Linux-v5.4/drivers/infiniband/hw/qedr/ |
D | qedr_iw_cm.c | 494 qp = xa_load(&dev->qps, conn_param->qpn); in qedr_iw_connect() 684 qp = xa_load(&dev->qps, conn_param->qpn); in qedr_iw_accept() 751 return xa_load(&dev->qps, qpn); in qedr_iw_get_qp()
|
/Linux-v5.4/drivers/infiniband/hw/hns/ |
D | hns_roce_cq.c | 526 cq = xa_load(&hr_dev->cq_table.array, cqn & (hr_dev->caps.num_cqs - 1)); in hns_roce_cq_completion() 542 cq = xa_load(&cq_table->array, cqn & (hr_dev->caps.num_cqs - 1)); in hns_roce_cq_event()
|
D | hns_roce_srq.c | 18 srq = xa_load(&srq_table->xa, srqn & (hr_dev->caps.num_srqs - 1)); in hns_roce_srq_event()
|
/Linux-v5.4/mm/ |
D | memremap.c | 403 pgmap = xa_load(&pgmap_array, PHYS_PFN(phys)); in get_dev_pagemap()
|
D | readahead.c | 182 page = xa_load(&mapping->i_pages, page_offset); in __do_page_cache_readahead()
|
/Linux-v5.4/drivers/infiniband/sw/siw/ |
D | siw_mem.c | 53 mem = xa_load(&sdev->mem_xa, stag_index); in siw_mem_id2obj()
|
D | siw.h | 615 qp = xa_load(&sdev->qp_xa, id); in siw_qp_id2obj()
|
/Linux-v5.4/drivers/infiniband/hw/mlx4/ |
D | cm.c | 283 ent = xa_load(&sriov->pv_id_table, *pv_cm_id); in id_map_get()
|
/Linux-v5.4/Documentation/core-api/ |
D | xarray.rst | 73 using xa_load(). xa_store will overwrite any entry with the 181 * xa_load() 223 before calling xa_load(), then take a reference count on the
|
/Linux-v5.4/drivers/pci/ |
D | p2pdma.c | 811 return xa_to_value(xa_load(&provider->p2pdma->map_types, in pci_p2pdma_map_type()
|
/Linux-v5.4/drivers/infiniband/ulp/opa_vnic/ |
D | opa_vnic_vema.c | 152 return xa_load(&port->vports, vport_num); in vema_get_vport_adapter()
|
/Linux-v5.4/include/linux/ |
D | xarray.h | 347 void *xa_load(struct xarray *, unsigned long index);
|