/Linux-v5.15/drivers/infiniband/sw/rxe/ |
D | rxe_mw.c | 10 struct rxe_mw *mw = to_rmw(ibmw); in rxe_alloc_mw() local 17 ret = rxe_add_to_pool(&rxe->mw_pool, mw); in rxe_alloc_mw() 23 rxe_add_index(mw); in rxe_alloc_mw() 24 ibmw->rkey = (mw->pelem.index << 8) | rxe_get_next_key(-1); in rxe_alloc_mw() 25 mw->state = (mw->ibmw.type == IB_MW_TYPE_2) ? in rxe_alloc_mw() 27 spin_lock_init(&mw->lock); in rxe_alloc_mw() 32 static void rxe_do_dealloc_mw(struct rxe_mw *mw) in rxe_do_dealloc_mw() argument 34 if (mw->mr) { in rxe_do_dealloc_mw() 35 struct rxe_mr *mr = mw->mr; in rxe_do_dealloc_mw() 37 mw->mr = NULL; in rxe_do_dealloc_mw() [all …]
|
D | rxe_verbs.h | 467 static inline struct rxe_mw *to_rmw(struct ib_mw *mw) in to_rmw() argument 469 return mw ? container_of(mw, struct rxe_mw, ibmw) : NULL; in to_rmw() 487 static inline struct rxe_pd *rxe_mw_pd(struct rxe_mw *mw) in rxe_mw_pd() argument 489 return to_rpd(mw->ibmw.pd); in rxe_mw_pd() 492 static inline u32 rxe_mw_rkey(struct rxe_mw *mw) in rxe_mw_rkey() argument 494 return mw->ibmw.rkey; in rxe_mw_rkey()
|
D | rxe_resp.c | 423 struct rxe_mw *mw = NULL; in check_rkey() local 465 mw = rxe_lookup_mw(qp, access, rkey); in check_rkey() 466 if (!mw) { in check_rkey() 472 mr = mw->mr; in check_rkey() 479 if (mw->access & IB_ZERO_BASED) in check_rkey() 480 qp->resp.offset = mw->addr; in check_rkey() 482 rxe_drop_ref(mw); in check_rkey() 527 if (mw) in check_rkey() 528 rxe_drop_ref(mw); in check_rkey()
|
/Linux-v5.15/net/netfilter/ipvs/ |
D | ip_vs_wrr.c | 63 int mw; /* maximum weight */ member 119 mark->mw = ip_vs_wrr_max_weight(svc) - (mark->di - 1); in ip_vs_wrr_init_svc() 120 mark->cw = mark->mw; in ip_vs_wrr_init_svc() 146 mark->mw = ip_vs_wrr_max_weight(svc) - (mark->di - 1); in ip_vs_wrr_dest_changed() 147 if (mark->cw > mark->mw || !mark->cw) in ip_vs_wrr_dest_changed() 148 mark->cw = mark->mw; in ip_vs_wrr_dest_changed() 172 if (mark->mw == 0) in ip_vs_wrr_schedule() 188 mark->cw = mark->mw; in ip_vs_wrr_schedule()
|
D | ip_vs_mh.c | 348 int mw, shift; in ip_vs_mh_shift_weight() local 366 mw = weight / gcd; in ip_vs_mh_shift_weight() 369 shift = fls(mw) - IP_VS_MH_TAB_BITS; in ip_vs_mh_shift_weight()
|
/Linux-v5.15/drivers/ntb/ |
D | ntb_transport.c | 617 struct ntb_transport_mw *mw; in ntb_transport_setup_qp_mw() local 629 mw = &nt->mw_vec[mw_num]; in ntb_transport_setup_qp_mw() 631 if (!mw->virt_addr) in ntb_transport_setup_qp_mw() 639 rx_size = (unsigned int)mw->xlat_size / num_qps_mw; in ntb_transport_setup_qp_mw() 640 qp->rx_buff = mw->virt_addr + rx_size * (qp_num / mw_count); in ntb_transport_setup_qp_mw() 793 struct ntb_transport_mw *mw = &nt->mw_vec[num_mw]; in ntb_free_mw() local 796 if (!mw->virt_addr) in ntb_free_mw() 800 dma_free_coherent(&pdev->dev, mw->alloc_size, in ntb_free_mw() 801 mw->alloc_addr, mw->dma_addr); in ntb_free_mw() 802 mw->xlat_size = 0; in ntb_free_mw() [all …]
|
/Linux-v5.15/arch/mips/txx9/generic/ |
D | mem_tx4927.c | 46 unsigned int mw = 0; in tx4927_process_sdccr() local 61 mw = 8 >> sdccr_mw; in tx4927_process_sdccr() 64 return rs * cs * mw * bs; in tx4927_process_sdccr()
|
/Linux-v5.15/fs/ocfs2/ |
D | dlmglue.c | 434 struct ocfs2_mask_waiter *mw, int ret) in ocfs2_update_lock_stats() argument 447 kt = ktime_sub(ktime_get(), mw->mw_lock_start); in ocfs2_update_lock_stats() 474 struct ocfs2_mask_waiter *mw; in ocfs2_track_lock_wait() local 481 mw = list_first_entry(&lockres->l_mask_waiters, in ocfs2_track_lock_wait() 484 ktime_to_us(ktime_mono_to_real(mw->mw_lock_start)); in ocfs2_track_lock_wait() 487 static inline void ocfs2_init_start_time(struct ocfs2_mask_waiter *mw) in ocfs2_init_start_time() argument 489 mw->mw_lock_start = ktime_get(); in ocfs2_init_start_time() 496 int level, struct ocfs2_mask_waiter *mw, int ret) in ocfs2_update_lock_stats() argument 505 static inline void ocfs2_init_start_time(struct ocfs2_mask_waiter *mw) in ocfs2_init_start_time() argument 891 struct ocfs2_mask_waiter *mw, *tmp; in lockres_set_flags() local [all …]
|
/Linux-v5.15/drivers/net/ethernet/mellanox/mlx4/ |
D | mr.c | 815 struct mlx4_mw *mw) in mlx4_mw_alloc() argument 829 mw->key = hw_index_to_key(index); in mlx4_mw_alloc() 830 mw->pd = pd; in mlx4_mw_alloc() 831 mw->type = type; in mlx4_mw_alloc() 832 mw->enabled = MLX4_MPT_DISABLED; in mlx4_mw_alloc() 838 int mlx4_mw_enable(struct mlx4_dev *dev, struct mlx4_mw *mw) in mlx4_mw_enable() argument 844 err = mlx4_mpt_alloc_icm(dev, key_to_hw_index(mw->key)); in mlx4_mw_enable() 858 mpt_entry->key = cpu_to_be32(key_to_hw_index(mw->key)); in mlx4_mw_enable() 859 mpt_entry->pd_flags = cpu_to_be32(mw->pd); in mlx4_mw_enable() 860 if (mw->type == MLX4_MW_TYPE_2) { in mlx4_mw_enable() [all …]
|
/Linux-v5.15/drivers/infiniband/hw/hns/ |
D | hns_roce_mr.c | 483 struct hns_roce_mw *mw) in hns_roce_mw_free() argument 488 if (mw->enabled) { in hns_roce_mw_free() 490 key_to_hw_index(mw->rkey) & in hns_roce_mw_free() 496 key_to_hw_index(mw->rkey)); in hns_roce_mw_free() 500 (int)key_to_hw_index(mw->rkey)); in hns_roce_mw_free() 504 struct hns_roce_mw *mw) in hns_roce_mw_enable() argument 509 unsigned long mtpt_idx = key_to_hw_index(mw->rkey); in hns_roce_mw_enable() 523 ret = hr_dev->hw->mw_write_mtpt(mailbox->buf, mw); in hns_roce_mw_enable() 536 mw->enabled = 1; in hns_roce_mw_enable() 556 struct hns_roce_mw *mw = to_hr_mw(ibmw); in hns_roce_alloc_mw() local [all …]
|
/Linux-v5.15/drivers/clk/rockchip/ |
D | clk.h | 462 #define COMPOSITE(_id, cname, pnames, f, mo, ms, mw, mf, ds, dw,\ argument 473 .mux_width = mw, \ 483 #define COMPOSITE_DIV_OFFSET(_id, cname, pnames, f, mo, ms, mw, \ argument 494 .mux_width = mw, \ 542 #define COMPOSITE_NODIV(_id, cname, pnames, f, mo, ms, mw, mf, \ argument 553 .mux_width = mw, \ 560 #define COMPOSITE_NOGATE(_id, cname, pnames, f, mo, ms, mw, mf, \ argument 571 .mux_width = mw, \ 580 mw, mf, ds, dw, df, dt) \ argument 590 .mux_width = mw, \ [all …]
|
/Linux-v5.15/drivers/net/ethernet/marvell/octeontx2/af/ |
D | rvu.c | 33 static int rvu_mbox_init(struct rvu *rvu, struct mbox_wq_info *mw, 2035 struct mbox_wq_info *mw; in __rvu_mbox_handler() local 2040 mw = &rvu->afpf_wq_info; in __rvu_mbox_handler() 2043 mw = &rvu->afvf_wq_info; in __rvu_mbox_handler() 2049 devid = mwork - mw->mbox_wrk; in __rvu_mbox_handler() 2050 mbox = &mw->mbox; in __rvu_mbox_handler() 2055 if (mw->mbox_wrk[devid].num_msgs == 0) in __rvu_mbox_handler() 2060 for (id = 0; id < mw->mbox_wrk[devid].num_msgs; id++) { in __rvu_mbox_handler() 2093 mw->mbox_wrk[devid].num_msgs = 0; in __rvu_mbox_handler() 2119 struct mbox_wq_info *mw; in __rvu_mbox_up_handler() local [all …]
|
/Linux-v5.15/include/linux/usb/ |
D | pd.h | 249 #define PDO_BATT_MAX_POWER(mw) ((((mw) / 250) & PDO_PWR_MASK) << PDO_BATT_MAX_PWR_SHIFT) argument 374 #define RDO_BATT_OP_PWR(mw) ((((mw) / 250) & RDO_PWR_MASK) << RDO_BATT_OP_PWR_SHIFT) argument 375 #define RDO_BATT_MAX_PWR(mw) ((((mw) / 250) & RDO_PWR_MASK) << RDO_BATT_MAX_PWR_SHIFT) argument
|
/Linux-v5.15/drivers/infiniband/hw/mlx4/ |
D | mr.c | 617 struct mlx4_ib_mw *mw = to_mmw(ibmw); in mlx4_ib_alloc_mw() local 621 to_mlx4_type(ibmw->type), &mw->mmw); in mlx4_ib_alloc_mw() 625 err = mlx4_mw_enable(dev->dev, &mw->mmw); in mlx4_ib_alloc_mw() 629 ibmw->rkey = mw->mmw.key; in mlx4_ib_alloc_mw() 633 mlx4_mw_free(dev->dev, &mw->mmw); in mlx4_ib_alloc_mw() 639 struct mlx4_ib_mw *mw = to_mmw(ibmw); in mlx4_ib_dealloc_mw() local 641 mlx4_mw_free(to_mdev(ibmw->device)->dev, &mw->mmw); in mlx4_ib_dealloc_mw()
|
/Linux-v5.15/drivers/mtd/maps/ |
D | physmap-core.c | 154 map_word mw; in physmap_addr_gpios_read() local 162 mw.x[0] = word; in physmap_addr_gpios_read() 163 return mw; in physmap_addr_gpios_read() 188 static void physmap_addr_gpios_write(struct map_info *map, map_word mw, in physmap_addr_gpios_write() argument 199 word = mw.x[0]; in physmap_addr_gpios_write()
|
/Linux-v5.15/drivers/pci/endpoint/functions/ |
D | pci-epf-ntb.c | 236 enum pci_epc_interface_type type, u32 mw) in epf_ntb_configure_mw() argument 252 peer_barno = peer_ntb_epc->epf_ntb_bar[mw + NTB_MW_OFFSET]; in epf_ntb_configure_mw() 259 if (mw + NTB_MW_OFFSET == BAR_DB_MW1) in epf_ntb_configure_mw() 262 if (size > ntb->mws_size[mw]) { in epf_ntb_configure_mw() 265 pci_epc_interface_string(type), mw, size, in epf_ntb_configure_mw() 266 ntb->mws_size[mw]); in epf_ntb_configure_mw() 278 pci_epc_interface_string(type), mw); in epf_ntb_configure_mw() 295 enum pci_epc_interface_type type, u32 mw) in epf_ntb_teardown_mw() argument 309 peer_barno = peer_ntb_epc->epf_ntb_bar[mw + NTB_MW_OFFSET]; in epf_ntb_teardown_mw() 314 if (mw + NTB_MW_OFFSET == BAR_DB_MW1) in epf_ntb_teardown_mw()
|
/Linux-v5.15/drivers/infiniband/hw/bnxt_re/ |
D | ib_verbs.h | 53 struct ib_mw *mw; member 207 int bnxt_re_dealloc_mw(struct ib_mw *mw);
|
D | ib_verbs.c | 409 fence->bind_rkey = ib_inc_rkey(fence->mw->rkey); in bnxt_re_create_fence_wqe() 447 if (fence->mw) { in bnxt_re_destroy_fence_mr() 448 bnxt_re_dealloc_mw(fence->mw); in bnxt_re_destroy_fence_mr() 449 fence->mw = NULL; in bnxt_re_destroy_fence_mr() 475 struct ib_mw *mw; in bnxt_re_create_fence_mr() local 519 mw = bnxt_re_alloc_mw(&pd->ib_pd, IB_MW_TYPE_1, NULL); in bnxt_re_create_fence_mr() 520 if (IS_ERR(mw)) { in bnxt_re_create_fence_mr() 523 rc = PTR_ERR(mw); in bnxt_re_create_fence_mr() 526 fence->mw = mw; in bnxt_re_create_fence_mr() 3712 struct bnxt_re_mw *mw; in bnxt_re_alloc_mw() local [all …]
|
/Linux-v5.15/include/dt-bindings/usb/ |
D | pd.h | 44 #define PDO_BATT_MAX_POWER(mw) ((((mw) / 250) & PDO_PWR_MASK) << PDO_BATT_MAX_PWR_SHIFT) argument
|
/Linux-v5.15/include/uapi/rdma/ |
D | rdma_user_rxe.h | 109 } mw; member
|
/Linux-v5.15/drivers/infiniband/core/ |
D | uverbs_main.c | 105 int uverbs_dealloc_mw(struct ib_mw *mw) in uverbs_dealloc_mw() argument 107 struct ib_pd *pd = mw->pd; in uverbs_dealloc_mw() 110 ret = mw->device->ops.dealloc_mw(mw); in uverbs_dealloc_mw() 115 kfree(mw); in uverbs_dealloc_mw()
|
D | uverbs.h | 245 int uverbs_dealloc_mw(struct ib_mw *mw);
|
D | uverbs_cmd.c | 909 struct ib_mw *mw; in ib_uverbs_alloc_mw() local 932 mw = rdma_zalloc_drv_obj(ib_dev, ib_mw); in ib_uverbs_alloc_mw() 933 if (!mw) { in ib_uverbs_alloc_mw() 938 mw->device = ib_dev; in ib_uverbs_alloc_mw() 939 mw->pd = pd; in ib_uverbs_alloc_mw() 940 mw->uobject = uobj; in ib_uverbs_alloc_mw() 941 mw->type = cmd.mw_type; in ib_uverbs_alloc_mw() 943 ret = pd->device->ops.alloc_mw(mw, &attrs->driver_udata); in ib_uverbs_alloc_mw() 949 uobj->object = mw; in ib_uverbs_alloc_mw() 953 resp.rkey = mw->rkey; in ib_uverbs_alloc_mw() [all …]
|
/Linux-v5.15/drivers/infiniband/hw/mlx5/ |
D | mr.c | 2265 struct mlx5_ib_mw *mw = to_mmw(ibmw); in mlx5_ib_alloc_mw() local 2307 err = mlx5_ib_create_mkey(dev, &mw->mmkey, in, inlen); in mlx5_ib_alloc_mw() 2311 mw->mmkey.type = MLX5_MKEY_MW; in mlx5_ib_alloc_mw() 2312 ibmw->rkey = mw->mmkey.key; in mlx5_ib_alloc_mw() 2313 mw->ndescs = ndescs; in mlx5_ib_alloc_mw() 2324 err = mlx5r_store_odp_mkey(dev, &mw->mmkey); in mlx5_ib_alloc_mw() 2333 mlx5_core_destroy_mkey(dev->mdev, &mw->mmkey); in mlx5_ib_alloc_mw() 2339 int mlx5_ib_dealloc_mw(struct ib_mw *mw) in mlx5_ib_dealloc_mw() argument 2341 struct mlx5_ib_dev *dev = to_mdev(mw->device); in mlx5_ib_dealloc_mw() 2342 struct mlx5_ib_mw *mmw = to_mmw(mw); in mlx5_ib_dealloc_mw()
|
/Linux-v5.15/drivers/net/ethernet/marvell/octeontx2/nic/ |
D | otx2_pf.c | 285 static void otx2_queue_work(struct mbox *mw, struct workqueue_struct *mbox_wq, in otx2_queue_work() argument 298 mbox = &mw->mbox; in otx2_queue_work() 311 mw[i].num_msgs = hdr->num_msgs; in otx2_queue_work() 318 queue_work(mbox_wq, &mw[i].mbox_wrk); in otx2_queue_work() 321 mbox = &mw->mbox_up; in otx2_queue_work() 327 mw[i].up_num_msgs = hdr->num_msgs; in otx2_queue_work() 334 queue_work(mbox_wq, &mw[i].mbox_up_wrk); in otx2_queue_work()
|