/Linux-v4.19/drivers/md/ |
D | dm-writecache.c | 98 #define WC_MODE_PMEM(wc) ((wc)->pmem_mode) argument 99 #define WC_MODE_FUA(wc) ((wc)->writeback_fua) argument 101 #define WC_MODE_PMEM(wc) false argument 102 #define WC_MODE_FUA(wc) false argument 104 #define WC_MODE_SORT_FREELIST(wc) (!WC_MODE_PMEM(wc)) argument 190 struct dm_writecache *wc; member 201 struct dm_writecache *wc; member 210 static void wc_lock(struct dm_writecache *wc) in wc_lock() argument 212 mutex_lock(&wc->lock); in wc_lock() 215 static void wc_unlock(struct dm_writecache *wc) in wc_unlock() argument [all …]
|
/Linux-v4.19/include/math-emu/ |
D | op-common.h | 27 #define _FP_DECL(wc, X) \ argument 29 _FP_FRAC_DECL_##wc(X) 36 #define _FP_UNPACK_CANONICAL(fs, wc, X) \ argument 42 _FP_FRAC_SLL_##wc(X, _FP_WORKBITS); \ 48 if (_FP_FRAC_ZEROP_##wc(X)) \ 54 _FP_FRAC_CLZ_##wc(_shift, X); \ 56 _FP_FRAC_SLL_##wc(X, (_shift+_FP_WORKBITS)); \ 69 if (_FP_FRAC_ZEROP_##wc(X)) \ 89 #define _FP_PACK_CANONICAL(fs, wc, X) \ argument 97 _FP_ROUND(wc, X); \ [all …]
|
D | soft-fp.h | 135 #define _FP_ROUND_NEAREST(wc, X) \ argument 137 if ((_FP_FRAC_LOW_##wc(X) & 15) != _FP_WORK_ROUND) \ 138 _FP_FRAC_ADDI_##wc(X, _FP_WORK_ROUND); \ 141 #define _FP_ROUND_ZERO(wc, X) 0 argument 143 #define _FP_ROUND_PINF(wc, X) \ argument 145 if (!X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \ 146 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \ 149 #define _FP_ROUND_MINF(wc, X) \ argument 151 if (X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \ 152 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \ [all …]
|
/Linux-v4.19/drivers/infiniband/hw/cxgb3/ |
D | iwch_cq.c | 36 struct iwch_qp *qhp, struct ib_wc *wc) in __iwch_poll_cq_one() argument 59 wc->wr_id = cookie; in __iwch_poll_cq_one() 60 wc->qp = qhp ? &qhp->ibqp : NULL; in __iwch_poll_cq_one() 61 wc->vendor_err = CQE_STATUS(cqe); in __iwch_poll_cq_one() 62 wc->wc_flags = 0; in __iwch_poll_cq_one() 72 wc->byte_len = CQE_LEN(cqe); in __iwch_poll_cq_one() 74 wc->byte_len = 0; in __iwch_poll_cq_one() 75 wc->opcode = IB_WC_RECV; in __iwch_poll_cq_one() 78 wc->ex.invalidate_rkey = CQE_WRID_STAG(cqe); in __iwch_poll_cq_one() 79 wc->wc_flags |= IB_WC_WITH_INVALIDATE; in __iwch_poll_cq_one() [all …]
|
/Linux-v4.19/drivers/infiniband/sw/rdmavt/ |
D | cq.c | 66 struct rvt_cq_wc *wc; in rvt_cq_enter() local 77 wc = cq->queue; in rvt_cq_enter() 78 head = wc->head; in rvt_cq_enter() 86 if (unlikely(next == wc->tail)) { in rvt_cq_enter() 100 wc->uqueue[head].wr_id = entry->wr_id; in rvt_cq_enter() 101 wc->uqueue[head].status = entry->status; in rvt_cq_enter() 102 wc->uqueue[head].opcode = entry->opcode; in rvt_cq_enter() 103 wc->uqueue[head].vendor_err = entry->vendor_err; in rvt_cq_enter() 104 wc->uqueue[head].byte_len = entry->byte_len; in rvt_cq_enter() 105 wc->uqueue[head].ex.imm_data = entry->ex.imm_data; in rvt_cq_enter() [all …]
|
D | trace_cq.h | 112 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), 113 TP_ARGS(cq, wc, idx), 125 __entry->wr_id = wc->wr_id; 126 __entry->status = wc->status; 127 __entry->opcode = wc->opcode; 128 __entry->length = wc->byte_len; 129 __entry->qpn = wc->qp->qp_num; 146 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), 147 TP_ARGS(cq, wc, idx)); 151 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), [all …]
|
/Linux-v4.19/drivers/infiniband/hw/mlx4/ |
D | cq.c | 514 struct ib_wc *wc) in mlx4_ib_handle_error_cqe() argument 528 wc->status = IB_WC_LOC_LEN_ERR; in mlx4_ib_handle_error_cqe() 531 wc->status = IB_WC_LOC_QP_OP_ERR; in mlx4_ib_handle_error_cqe() 534 wc->status = IB_WC_LOC_PROT_ERR; in mlx4_ib_handle_error_cqe() 537 wc->status = IB_WC_WR_FLUSH_ERR; in mlx4_ib_handle_error_cqe() 540 wc->status = IB_WC_MW_BIND_ERR; in mlx4_ib_handle_error_cqe() 543 wc->status = IB_WC_BAD_RESP_ERR; in mlx4_ib_handle_error_cqe() 546 wc->status = IB_WC_LOC_ACCESS_ERR; in mlx4_ib_handle_error_cqe() 549 wc->status = IB_WC_REM_INV_REQ_ERR; in mlx4_ib_handle_error_cqe() 552 wc->status = IB_WC_REM_ACCESS_ERR; in mlx4_ib_handle_error_cqe() [all …]
|
/Linux-v4.19/fs/ocfs2/ |
D | aops.c | 816 static void ocfs2_unlock_pages(struct ocfs2_write_ctxt *wc) in ocfs2_unlock_pages() argument 825 if (wc->w_target_locked) { in ocfs2_unlock_pages() 826 BUG_ON(!wc->w_target_page); in ocfs2_unlock_pages() 827 for (i = 0; i < wc->w_num_pages; i++) { in ocfs2_unlock_pages() 828 if (wc->w_target_page == wc->w_pages[i]) { in ocfs2_unlock_pages() 829 wc->w_pages[i] = NULL; in ocfs2_unlock_pages() 833 mark_page_accessed(wc->w_target_page); in ocfs2_unlock_pages() 834 put_page(wc->w_target_page); in ocfs2_unlock_pages() 836 ocfs2_unlock_and_free_pages(wc->w_pages, wc->w_num_pages); in ocfs2_unlock_pages() 855 struct ocfs2_write_ctxt *wc) in ocfs2_free_write_ctxt() argument [all …]
|
/Linux-v4.19/drivers/net/ethernet/brocade/bna/ |
D | bfa_cs.h | 81 bfa_wc_up(struct bfa_wc *wc) in bfa_wc_up() argument 83 wc->wc_count++; in bfa_wc_up() 87 bfa_wc_down(struct bfa_wc *wc) in bfa_wc_down() argument 89 wc->wc_count--; in bfa_wc_down() 90 if (wc->wc_count == 0) in bfa_wc_down() 91 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down() 96 bfa_wc_init(struct bfa_wc *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument 98 wc->wc_resume = wc_resume; in bfa_wc_init() 99 wc->wc_cbarg = wc_cbarg; in bfa_wc_init() 100 wc->wc_count = 0; in bfa_wc_init() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/mlx5/ |
D | cq.c | 115 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() argument 118 wc->wc_flags = 0; in handle_good_req() 121 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 124 wc->opcode = IB_WC_RDMA_WRITE; in handle_good_req() 127 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 131 wc->opcode = IB_WC_SEND; in handle_good_req() 134 wc->opcode = IB_WC_RDMA_READ; in handle_good_req() 135 wc->byte_len = be32_to_cpu(cqe->byte_cnt); in handle_good_req() 138 wc->opcode = IB_WC_COMP_SWAP; in handle_good_req() 139 wc->byte_len = 8; in handle_good_req() [all …]
|
D | gsi.c | 37 struct ib_wc wc; member 88 WARN_ON_ONCE(mlx5_ib_generate_wc(gsi_cq, &wr->wc)); in generate_completions() 96 static void handle_single_completion(struct ib_cq *cq, struct ib_wc *wc) in handle_single_completion() argument 100 container_of(wc->wr_cqe, struct mlx5_ib_gsi_wr, cqe); in handle_single_completion() 106 wr_id = wr->wc.wr_id; in handle_single_completion() 107 wr->wc = *wc; in handle_single_completion() 108 wr->wc.wr_id = wr_id; in handle_single_completion() 109 wr->wc.qp = &gsi->ibqp; in handle_single_completion() 415 struct ib_ud_wr *wr, struct ib_wc *wc) in mlx5_ib_add_outstanding_wr() argument 429 if (!wc) { in mlx5_ib_add_outstanding_wr() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/qib/ |
D | qib_ud.c | 61 struct ib_wc wc; in qib_ud_loopback() local 124 memset(&wc, 0, sizeof(wc)); in qib_ud_loopback() 125 wc.byte_len = length + sizeof(struct ib_grh); in qib_ud_loopback() 128 wc.wc_flags = IB_WC_WITH_IMM; in qib_ud_loopback() 129 wc.ex.imm_data = swqe->wr.ex.imm_data; in qib_ud_loopback() 154 if (unlikely(wc.byte_len > qp->r_len)) { in qib_ud_loopback() 167 wc.wc_flags |= IB_WC_GRH; in qib_ud_loopback() 205 wc.wr_id = qp->r_wr_id; in qib_ud_loopback() 206 wc.status = IB_WC_SUCCESS; in qib_ud_loopback() 207 wc.opcode = IB_WC_RECV; in qib_ud_loopback() [all …]
|
D | qib_uc.c | 245 struct ib_wc wc; in qib_uc_rcv() local 367 wc.ex.imm_data = ohdr->u.imm_data; in qib_uc_rcv() 369 wc.wc_flags = IB_WC_WITH_IMM; in qib_uc_rcv() 373 wc.ex.imm_data = 0; in qib_uc_rcv() 374 wc.wc_flags = 0; in qib_uc_rcv() 384 wc.byte_len = tlen + qp->r_rcv_len; in qib_uc_rcv() 385 if (unlikely(wc.byte_len > qp->r_len)) in qib_uc_rcv() 387 wc.opcode = IB_WC_RECV; in qib_uc_rcv() 391 wc.wr_id = qp->r_wr_id; in qib_uc_rcv() 392 wc.status = IB_WC_SUCCESS; in qib_uc_rcv() [all …]
|
/Linux-v4.19/drivers/infiniband/core/ |
D | cq.c | 42 struct ib_wc *wc = &wcs[i]; in __ib_process_cq() local 44 if (wc->wr_cqe) in __ib_process_cq() 45 wc->wr_cqe->done(cq, wc); in __ib_process_cq() 47 WARN_ON_ONCE(wc->status == IB_WC_SUCCESS); in __ib_process_cq() 91 completed = __ib_process_cq(cq, budget, cq->wc, IB_POLL_BATCH); in ib_poll_handler() 111 completed = __ib_process_cq(cq, IB_POLL_BUDGET_WORKQUEUE, cq->wc, in ib_cq_poll_work() 159 cq->wc = kmalloc_array(IB_POLL_BATCH, sizeof(*cq->wc), GFP_KERNEL); in __ib_alloc_cq() 160 if (!cq->wc) in __ib_alloc_cq() 190 kfree(cq->wc); in __ib_alloc_cq() 222 kfree(cq->wc); in ib_free_cq()
|
/Linux-v4.19/fs/ntfs/ |
D | unistr.c | 264 wchar_t wc; in ntfs_nlstoucs() local 273 &wc); in ntfs_nlstoucs() 276 if (likely(wc)) { in ntfs_nlstoucs() 277 ucs[o++] = cpu_to_le16(wc); in ntfs_nlstoucs() 341 int i, o, ns_len, wc; in ntfs_ucstonls() local 348 wc = -ENAMETOOLONG; in ntfs_ucstonls() 358 retry: wc = nls->uni2char(le16_to_cpu(ins[i]), ns + o, in ntfs_ucstonls() 360 if (wc > 0) { in ntfs_ucstonls() 361 o += wc; in ntfs_ucstonls() 363 } else if (!wc) in ntfs_ucstonls() [all …]
|
/Linux-v4.19/net/smc/ |
D | smc_wr.c | 64 static inline void smc_wr_tx_process_cqe(struct ib_wc *wc) in smc_wr_tx_process_cqe() argument 71 link = wc->qp->qp_context; in smc_wr_tx_process_cqe() 73 if (wc->opcode == IB_WC_REG_MR) { in smc_wr_tx_process_cqe() 74 if (wc->status) in smc_wr_tx_process_cqe() 82 pnd_snd_idx = smc_wr_tx_find_pending_index(link, wc->wr_id); in smc_wr_tx_process_cqe() 85 link->wr_tx_pends[pnd_snd_idx].wc_status = wc->status; in smc_wr_tx_process_cqe() 94 if (wc->status) { in smc_wr_tx_process_cqe() 107 pnd_snd.handler(&pnd_snd.priv, link, wc->status); in smc_wr_tx_process_cqe() 114 struct ib_wc wc[SMC_WR_MAX_POLL_CQE]; in smc_wr_tx_tasklet_fn() local 121 memset(&wc, 0, sizeof(wc)); in smc_wr_tx_tasklet_fn() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_cq.c | 334 struct ib_wc *wc) in pvrdma_poll_one() argument 367 wc->opcode = pvrdma_wc_opcode_to_ib(cqe->opcode); in pvrdma_poll_one() 368 wc->status = pvrdma_wc_status_to_ib(cqe->status); in pvrdma_poll_one() 369 wc->wr_id = cqe->wr_id; in pvrdma_poll_one() 370 wc->qp = &(*cur_qp)->ibqp; in pvrdma_poll_one() 371 wc->byte_len = cqe->byte_len; in pvrdma_poll_one() 372 wc->ex.imm_data = cqe->imm_data; in pvrdma_poll_one() 373 wc->src_qp = cqe->src_qp; in pvrdma_poll_one() 374 wc->wc_flags = pvrdma_wc_flags_to_ib(cqe->wc_flags); in pvrdma_poll_one() 375 wc->pkey_index = cqe->pkey_index; in pvrdma_poll_one() [all …]
|
/Linux-v4.19/drivers/scsi/bfa/ |
D | bfa_cs.h | 261 bfa_wc_up(struct bfa_wc_s *wc) in bfa_wc_up() argument 263 wc->wc_count++; in bfa_wc_up() 267 bfa_wc_down(struct bfa_wc_s *wc) in bfa_wc_down() argument 269 wc->wc_count--; in bfa_wc_down() 270 if (wc->wc_count == 0) in bfa_wc_down() 271 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down() 278 bfa_wc_init(struct bfa_wc_s *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument 280 wc->wc_resume = wc_resume; in bfa_wc_init() 281 wc->wc_cbarg = wc_cbarg; in bfa_wc_init() 282 wc->wc_count = 0; in bfa_wc_init() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/hfi1/ |
D | ud.c | 82 struct ib_wc wc; in ud_loopback() local 148 memset(&wc, 0, sizeof(wc)); in ud_loopback() 149 wc.byte_len = length + sizeof(struct ib_grh); in ud_loopback() 152 wc.wc_flags = IB_WC_WITH_IMM; in ud_loopback() 153 wc.ex.imm_data = swqe->wr.ex.imm_data; in ud_loopback() 178 if (unlikely(wc.byte_len > qp->r_len)) { in ud_loopback() 215 wc.wc_flags |= IB_WC_GRH; in ud_loopback() 254 wc.wr_id = qp->r_wr_id; in ud_loopback() 255 wc.status = IB_WC_SUCCESS; in ud_loopback() 256 wc.opcode = IB_WC_RECV; in ud_loopback() [all …]
|
D | uc.c | 315 struct ib_wc wc; in hfi1_uc_rcv() local 434 wc.ex.imm_data = ohdr->u.imm_data; in hfi1_uc_rcv() 435 wc.wc_flags = IB_WC_WITH_IMM; in hfi1_uc_rcv() 439 wc.ex.imm_data = 0; in hfi1_uc_rcv() 440 wc.wc_flags = 0; in hfi1_uc_rcv() 448 wc.byte_len = tlen + qp->r_rcv_len; in hfi1_uc_rcv() 449 if (unlikely(wc.byte_len > qp->r_len)) in hfi1_uc_rcv() 451 wc.opcode = IB_WC_RECV; in hfi1_uc_rcv() 455 wc.wr_id = qp->r_wr_id; in hfi1_uc_rcv() 456 wc.status = IB_WC_SUCCESS; in hfi1_uc_rcv() [all …]
|
/Linux-v4.19/arch/c6x/platforms/ |
D | cache.c | 141 unsigned int wc = 0; in cache_block_operation() local 143 for (; wcnt; wcnt -= wc, start += wc) { in cache_block_operation() 163 wc = 0xffff; in cache_block_operation() 165 wc = wcnt; in cache_block_operation() 168 imcr_set(wc_reg, wc & 0xffff); in cache_block_operation() 186 unsigned int wc = 0; in cache_block_operation_nowait() local 188 for (; wcnt; wcnt -= wc, start += wc) { in cache_block_operation_nowait() 195 wc = 0xffff; in cache_block_operation_nowait() 197 wc = wcnt; in cache_block_operation_nowait() 200 imcr_set(wc_reg, wc & 0xffff); in cache_block_operation_nowait()
|
/Linux-v4.19/net/sunrpc/xprtrdma/ |
D | frwr_ops.c | 295 __frwr_sendcompletion_flush(struct ib_wc *wc, const char *wr) in __frwr_sendcompletion_flush() argument 297 if (wc->status != IB_WC_WR_FLUSH_ERR) in __frwr_sendcompletion_flush() 299 wr, ib_wc_status_msg(wc->status), in __frwr_sendcompletion_flush() 300 wc->status, wc->vendor_err); in __frwr_sendcompletion_flush() 310 frwr_wc_fastreg(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_fastreg() argument 312 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_fastreg() 317 if (wc->status != IB_WC_SUCCESS) { in frwr_wc_fastreg() 319 __frwr_sendcompletion_flush(wc, "fastreg"); in frwr_wc_fastreg() 321 trace_xprtrdma_wc_fastreg(wc, frwr); in frwr_wc_fastreg() 331 frwr_wc_localinv(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_localinv() argument [all …]
|
/Linux-v4.19/drivers/infiniband/ulp/iser/ |
D | iser_initiator.c | 558 void iser_login_rsp(struct ib_cq *cq, struct ib_wc *wc) in iser_login_rsp() argument 560 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_login_rsp() 562 struct iser_login_desc *desc = iser_login(wc->wr_cqe); in iser_login_rsp() 567 if (unlikely(wc->status != IB_WC_SUCCESS)) { in iser_login_rsp() 568 iser_err_comp(wc, "login_rsp"); in iser_login_rsp() 578 length = wc->byte_len - ISER_HEADERS_LEN; in iser_login_rsp() 603 struct ib_wc *wc, in iser_check_remote_inv() argument 606 if (wc->wc_flags & IB_WC_WITH_INVALIDATE) { in iser_check_remote_inv() 608 u32 rkey = wc->ex.invalidate_rkey; in iser_check_remote_inv() 643 void iser_task_rsp(struct ib_cq *cq, struct ib_wc *wc) in iser_task_rsp() argument [all …]
|
/Linux-v4.19/sound/isa/wavefront/ |
D | wavefront_synth.c | 1505 wavefront_control *wc) in wavefront_synth_control() argument 1513 "cmd 0x%x\n", wc->cmd); in wavefront_synth_control() 1517 switch (wc->cmd) { in wavefront_synth_control() 1532 wc->rbuf[0] = dev->interrupts_are_midi; in wavefront_synth_control() 1536 dev->rom_samples_rdonly = wc->wbuf[0]; in wavefront_synth_control() 1537 wc->status = 0; in wavefront_synth_control() 1541 i = wc->wbuf[0] | (wc->wbuf[1] << 7); in wavefront_synth_control() 1545 wc->status = EINVAL; in wavefront_synth_control() 1548 wc->rbuf[0] = dev->sample_status[i]; in wavefront_synth_control() 1549 wc->status = 0; in wavefront_synth_control() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/cxgb4/ |
D | cq.c | 756 struct ib_wc *wc, struct c4iw_srq *srq) in __c4iw_poll_cq_one() argument 770 wc->wr_id = cookie; in __c4iw_poll_cq_one() 771 wc->qp = qhp ? &qhp->ibqp : NULL; in __c4iw_poll_cq_one() 772 wc->vendor_err = CQE_STATUS(&cqe); in __c4iw_poll_cq_one() 773 wc->wc_flags = 0; in __c4iw_poll_cq_one() 791 wc->byte_len = CQE_LEN(&cqe); in __c4iw_poll_cq_one() 793 wc->byte_len = 0; in __c4iw_poll_cq_one() 797 wc->opcode = IB_WC_RECV; in __c4iw_poll_cq_one() 801 wc->opcode = IB_WC_RECV; in __c4iw_poll_cq_one() 802 wc->ex.invalidate_rkey = CQE_WRID_STAG(&cqe); in __c4iw_poll_cq_one() [all …]
|