/Linux-v5.4/net/sunrpc/xprtrdma/ |
D | backchannel.c | 60 static int rpcrdma_bc_marshal_reply(struct rpc_rqst *rqst) in rpcrdma_bc_marshal_reply() argument 62 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(rqst->rq_xprt); in rpcrdma_bc_marshal_reply() 63 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in rpcrdma_bc_marshal_reply() 68 rdmab_data(req->rl_rdmabuf), rqst); in rpcrdma_bc_marshal_reply() 73 *p++ = rqst->rq_xid; in rpcrdma_bc_marshal_reply() 82 &rqst->rq_snd_buf, rpcrdma_noch)) in rpcrdma_bc_marshal_reply() 85 trace_xprtrdma_cb_reply(rqst); in rpcrdma_bc_marshal_reply() 101 int xprt_rdma_bc_send_reply(struct rpc_rqst *rqst) in xprt_rdma_bc_send_reply() argument 103 struct rpc_xprt *xprt = rqst->rq_xprt; in xprt_rdma_bc_send_reply() 105 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in xprt_rdma_bc_send_reply() [all …]
|
D | svc_rdma_backchannel.c | 115 struct rpc_rqst *rqst, in svc_rdma_bc_sendto() argument 120 ret = svc_rdma_map_reply_msg(rdma, ctxt, &rqst->rq_snd_buf, NULL); in svc_rdma_bc_sendto() 127 get_page(virt_to_page(rqst->rq_buffer)); in svc_rdma_bc_sendto() 139 struct rpc_rqst *rqst = task->tk_rqstp; in xprt_rdma_bc_allocate() local 140 size_t size = rqst->rq_callsize; in xprt_rdma_bc_allocate() 152 rqst->rq_buffer = page_address(page); in xprt_rdma_bc_allocate() 154 rqst->rq_rbuffer = kmalloc(rqst->rq_rcvsize, RPCRDMA_DEF_GFP); in xprt_rdma_bc_allocate() 155 if (!rqst->rq_rbuffer) { in xprt_rdma_bc_allocate() 165 struct rpc_rqst *rqst = task->tk_rqstp; in xprt_rdma_bc_free() local 167 put_page(virt_to_page(rqst->rq_buffer)); in xprt_rdma_bc_free() [all …]
|
D | rpc_rdma.c | 136 struct rpc_rqst *rqst) in rpcrdma_args_inline() argument 138 struct xdr_buf *xdr = &rqst->rq_snd_buf; in rpcrdma_args_inline() 167 struct rpc_rqst *rqst) in rpcrdma_results_inline() argument 169 return rqst->rq_rcv_buf.buflen <= r_xprt->rx_ep.rep_max_inline_recv; in rpcrdma_results_inline() 178 const struct rpc_rqst *rqst) in rpcrdma_nonpayload_inline() argument 180 const struct xdr_buf *buf = &rqst->rq_rcv_buf; in rpcrdma_nonpayload_inline() 387 struct rpc_rqst *rqst, in rpcrdma_encode_read_list() argument 399 pos = rqst->rq_snd_buf.head[0].iov_len; in rpcrdma_encode_read_list() 403 nsegs = rpcrdma_convert_iovs(r_xprt, &rqst->rq_snd_buf, pos, in rpcrdma_encode_read_list() 416 trace_xprtrdma_chunk_read(rqst->rq_task, pos, mr, nsegs); in rpcrdma_encode_read_list() [all …]
|
D | transport.c | 585 xprt_rdma_free_slot(struct rpc_xprt *xprt, struct rpc_rqst *rqst) in xprt_rdma_free_slot() argument 590 memset(rqst, 0, sizeof(*rqst)); in xprt_rdma_free_slot() 591 rpcrdma_buffer_put(&r_xprt->rx_buf, rpcr_to_rdmar(rqst)); in xprt_rdma_free_slot() 620 struct rpc_rqst *rqst = task->tk_rqstp; in xprt_rdma_allocate() local 621 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(rqst->rq_xprt); in xprt_rdma_allocate() 622 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in xprt_rdma_allocate() 629 if (!rpcrdma_check_regbuf(r_xprt, req->rl_sendbuf, rqst->rq_callsize, in xprt_rdma_allocate() 632 if (!rpcrdma_check_regbuf(r_xprt, req->rl_recvbuf, rqst->rq_rcvsize, in xprt_rdma_allocate() 636 rqst->rq_buffer = rdmab_data(req->rl_sendbuf); in xprt_rdma_allocate() 637 rqst->rq_rbuffer = rdmab_data(req->rl_recvbuf); in xprt_rdma_allocate() [all …]
|
D | xprt_rdma.h | 335 rpcr_to_rdmar(const struct rpc_rqst *rqst) in rpcr_to_rdmar() argument 337 return container_of(rqst, struct rpcrdma_req, rl_slot); in rpcr_to_rdmar() 577 int rpcrdma_marshal_req(struct rpcrdma_xprt *r_xprt, struct rpc_rqst *rqst); 608 int xprt_rdma_bc_send_reply(struct rpc_rqst *rqst);
|
/Linux-v5.4/fs/cifs/ |
D | smb2inode.c | 41 free_set_inf_compound(struct smb_rqst *rqst) in free_set_inf_compound() argument 43 if (rqst[1].rq_iov) in free_set_inf_compound() 44 SMB2_set_info_free(&rqst[1]); in free_set_inf_compound() 45 if (rqst[2].rq_iov) in free_set_inf_compound() 46 SMB2_close_free(&rqst[2]); in free_set_inf_compound() 64 struct smb_rqst rqst[3]; in smb2_compound_op() local 83 memset(rqst, 0, sizeof(rqst)); in smb2_compound_op() 109 rqst[num_rqst].rq_iov = open_iov; in smb2_compound_op() 110 rqst[num_rqst].rq_nvec = SMB2_CREATE_IOV_SIZE; in smb2_compound_op() 111 rc = SMB2_open_init(tcon, &rqst[num_rqst], &oplock, &oparms, in smb2_compound_op() [all …]
|
D | transport.c | 263 smb_rqst_len(struct TCP_Server_Info *server, struct smb_rqst *rqst) in smb_rqst_len() argument 271 rqst->rq_nvec >= 2 && rqst->rq_iov[0].iov_len == 4) { in smb_rqst_len() 272 iov = &rqst->rq_iov[1]; in smb_rqst_len() 273 nvec = rqst->rq_nvec - 1; in smb_rqst_len() 275 iov = rqst->rq_iov; in smb_rqst_len() 276 nvec = rqst->rq_nvec; in smb_rqst_len() 289 if (rqst->rq_npages) { in smb_rqst_len() 290 if (rqst->rq_npages == 1) in smb_rqst_len() 291 buflen += rqst->rq_tailsz; in smb_rqst_len() 297 buflen += rqst->rq_pagesz * (rqst->rq_npages - 1) - in smb_rqst_len() [all …]
|
D | smb2pdu.c | 787 struct smb_rqst rqst; in SMB2_negotiate() local 861 memset(&rqst, 0, sizeof(struct smb_rqst)); in SMB2_negotiate() 862 rqst.rq_iov = iov; in SMB2_negotiate() 863 rqst.rq_nvec = 1; in SMB2_negotiate() 865 rc = cifs_send_recv(xid, ses, &rqst, &resp_buftype, flags, &rsp_iov); in SMB2_negotiate() 1232 struct smb_rqst rqst; in SMB2_sess_sendreceive() local 1241 memset(&rqst, 0, sizeof(struct smb_rqst)); in SMB2_sess_sendreceive() 1242 rqst.rq_iov = sess_data->iov; in SMB2_sess_sendreceive() 1243 rqst.rq_nvec = 2; in SMB2_sess_sendreceive() 1247 &rqst, in SMB2_sess_sendreceive() [all …]
|
D | smb2proto.h | 49 struct smb_rqst *rqst); 51 struct TCP_Server_Info *server, struct smb_rqst *rqst); 56 extern int smb2_calc_signature(struct smb_rqst *rqst, 58 extern int smb3_calc_signature(struct smb_rqst *rqst, 119 struct smb_rqst *rqst); 121 struct smb_rqst *rqst); 122 extern void smb2_set_related(struct smb_rqst *rqst); 140 extern int SMB2_open_init(struct cifs_tcon *tcon, struct smb_rqst *rqst, 143 extern void SMB2_open_free(struct smb_rqst *rqst); 148 extern int SMB2_ioctl_init(struct cifs_tcon *tcon, struct smb_rqst *rqst, [all …]
|
D | smb2ops.c | 644 struct smb_rqst rqst[2]; in open_shroot() local 673 memset(rqst, 0, sizeof(rqst)); in open_shroot() 679 rqst[0].rq_iov = open_iov; in open_shroot() 680 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE; in open_shroot() 689 rc = SMB2_open_init(tcon, &rqst[0], &oplock, &oparms, &utf16_path); in open_shroot() 692 smb2_set_next_command(tcon, &rqst[0]); in open_shroot() 695 rqst[1].rq_iov = qi_iov; in open_shroot() 696 rqst[1].rq_nvec = 1; in open_shroot() 698 rc = SMB2_query_info_init(tcon, &rqst[1], COMPOUND_FID, in open_shroot() 706 smb2_set_related(&rqst[1]); in open_shroot() [all …]
|
D | cifsencrypt.c | 40 int __cifs_calc_signature(struct smb_rqst *rqst, in __cifs_calc_signature() argument 46 struct kvec *iov = rqst->rq_iov; in __cifs_calc_signature() 47 int n_vec = rqst->rq_nvec; in __cifs_calc_signature() 79 for (i = 0; i < rqst->rq_npages; i++) { in __cifs_calc_signature() 83 rqst_page_get_length(rqst, i, &len, &offset); in __cifs_calc_signature() 85 kaddr = (char *) kmap(rqst->rq_pages[i]) + offset; in __cifs_calc_signature() 91 kunmap(rqst->rq_pages[i]); in __cifs_calc_signature() 95 kunmap(rqst->rq_pages[i]); in __cifs_calc_signature() 112 static int cifs_calc_signature(struct smb_rqst *rqst, in cifs_calc_signature() argument 117 if (!rqst->rq_iov || !signature || !server) in cifs_calc_signature() [all …]
|
D | smb2transport.c | 166 smb2_calc_signature(struct smb_rqst *rqst, struct TCP_Server_Info *server) in smb2_calc_signature() argument 171 struct kvec *iov = rqst->rq_iov; in smb2_calc_signature() 213 drqst = *rqst; in smb2_calc_signature() 427 smb3_calc_signature(struct smb_rqst *rqst, struct TCP_Server_Info *server) in smb3_calc_signature() argument 432 struct kvec *iov = rqst->rq_iov; in smb3_calc_signature() 472 drqst = *rqst; in smb3_calc_signature() 494 smb2_sign_rqst(struct smb_rqst *rqst, struct TCP_Server_Info *server) in smb2_sign_rqst() argument 498 (struct smb2_sync_hdr *)rqst->rq_iov[0].iov_base; in smb2_sign_rqst() 509 rc = server->ops->calc_signature(rqst, server); in smb2_sign_rqst() 515 smb2_verify_signature(struct smb_rqst *rqst, struct TCP_Server_Info *server) in smb2_verify_signature() argument [all …]
|
D | misc.c | 986 void rqst_page_get_length(struct smb_rqst *rqst, unsigned int page, in rqst_page_get_length() argument 989 *len = rqst->rq_pagesz; in rqst_page_get_length() 990 *offset = (page == 0) ? rqst->rq_offset : 0; in rqst_page_get_length() 992 if (rqst->rq_npages == 1 || page == rqst->rq_npages-1) in rqst_page_get_length() 993 *len = rqst->rq_tailsz; in rqst_page_get_length() 995 *len = rqst->rq_pagesz - rqst->rq_offset; in rqst_page_get_length()
|
D | cifsproto.h | 94 struct smb_rqst *rqst, 99 struct smb_rqst *rqst, int *resp_buf_type, 103 struct smb_rqst *rqst, int *resp_buf_type, 468 extern int cifs_sign_rqst(struct smb_rqst *rqst, struct TCP_Server_Info *server, 473 extern int cifs_verify_signature(struct smb_rqst *rqst, 571 int __cifs_calc_signature(struct smb_rqst *rqst, 585 extern void rqst_page_get_length(struct smb_rqst *rqst, unsigned int page,
|
D | smbdirect.h | 279 int num_rqst, struct smb_rqst *rqst); 319 static inline int smbd_send(struct TCP_Server_Info *server, int num_rqst, struct smb_rqst *rqst) {r… in smbd_send() argument
|
D | smbdirect.c | 2081 struct smb_rqst *rqst; in smbd_send() local 2111 rqst = &rqst_array[rqst_idx]; in smbd_send() 2112 iov = rqst->rq_iov; in smbd_send() 2115 rqst_idx, smb_rqst_len(server, rqst)); in smbd_send() 2116 for (i = 0; i < rqst->rq_nvec; i++) in smbd_send() 2122 rqst_idx, rqst->rq_nvec, rqst->rq_npages, rqst->rq_pagesz, in smbd_send() 2123 rqst->rq_tailsz, smb_rqst_len(server, rqst)); in smbd_send() 2173 if (i == rqst->rq_nvec) in smbd_send() 2180 if (i == rqst->rq_nvec) { in smbd_send() 2199 for (i = 0; i < rqst->rq_npages; i++) { in smbd_send() [all …]
|
/Linux-v5.4/include/trace/events/ |
D | sunrpc.h | 360 xdr->rqst->rq_task->tk_client->cl_program->name) 362 xdr->rqst->rq_task->tk_msg.rpc_proc->p_name) 366 if (xdr->rqst) { 367 const struct rpc_task *task = xdr->rqst->rq_task; 427 xdr->rqst->rq_task->tk_client->cl_program->name) 429 xdr->rqst->rq_task->tk_msg.rpc_proc->p_name) 433 const struct rpc_task *task = xdr->rqst->rq_task; 699 const struct rpc_rqst *rqst, 703 TP_ARGS(rqst, status), 714 __entry->task_id = rqst->rq_task->tk_pid; [all …]
|
D | rpcrdma.h | 270 const struct rpc_rqst *rqst 273 TP_ARGS(rqst), 276 __field(const void *, rqst) 283 __entry->rqst = rqst; 284 __entry->req = rpcr_to_rdmar(rqst); 285 __entry->rep = rpcr_to_rdmar(rqst)->rl_reply; 286 __entry->xid = be32_to_cpu(rqst->rq_xid); 290 __entry->xid, __entry->rqst, __entry->req, __entry->rep 297 const struct rpc_rqst *rqst \ 299 TP_ARGS(rqst)) [all …]
|
D | rpcgss.h | 197 const struct rpc_rqst *rqst = task->tk_rqstp; 201 __entry->xid = be32_to_cpu(rqst->rq_xid); 202 __entry->seqno = rqst->rq_seqno;
|
/Linux-v5.4/include/linux/sunrpc/ |
D | svc.h | 304 #define SVC_NET(rqst) (rqst->rq_xprt ? rqst->rq_xprt->xpt_net : rqst->rq_bc_net) argument 309 static inline struct sockaddr_in *svc_addr_in(const struct svc_rqst *rqst) in svc_addr_in() argument 311 return (struct sockaddr_in *) &rqst->rq_addr; in svc_addr_in() 314 static inline struct sockaddr_in6 *svc_addr_in6(const struct svc_rqst *rqst) in svc_addr_in6() argument 316 return (struct sockaddr_in6 *) &rqst->rq_addr; in svc_addr_in6() 319 static inline struct sockaddr *svc_addr(const struct svc_rqst *rqst) in svc_addr() argument 321 return (struct sockaddr *) &rqst->rq_addr; in svc_addr() 324 static inline struct sockaddr_in *svc_daddr_in(const struct svc_rqst *rqst) in svc_daddr_in() argument 326 return (struct sockaddr_in *) &rqst->rq_daddr; in svc_daddr_in() 329 static inline struct sockaddr_in6 *svc_daddr_in6(const struct svc_rqst *rqst) in svc_daddr_in6() argument [all …]
|
/Linux-v5.4/drivers/nvme/target/ |
D | fc.c | 1231 acc->rqst.desc_tag = cpu_to_be32(FCNVME_LSDESC_RQST); in nvmet_fc_format_rsp_hdr() 1232 acc->rqst.desc_len = in nvmet_fc_format_rsp_hdr() 1234 acc->rqst.w0.ls_cmd = rqst_ls_cmd; in nvmet_fc_format_rsp_hdr() 1324 struct fcnvme_ls_cr_assoc_rqst *rqst = in nvmet_fc_ls_create_association() local 1343 else if (be32_to_cpu(rqst->desc_list_len) < in nvmet_fc_ls_create_association() 1346 else if (rqst->assoc_cmd.desc_tag != in nvmet_fc_ls_create_association() 1349 else if (be32_to_cpu(rqst->assoc_cmd.desc_len) < in nvmet_fc_ls_create_association() 1352 else if (!rqst->assoc_cmd.ersp_ratio || in nvmet_fc_ls_create_association() 1353 (be16_to_cpu(rqst->assoc_cmd.ersp_ratio) >= in nvmet_fc_ls_create_association() 1354 be16_to_cpu(rqst->assoc_cmd.sqsize))) in nvmet_fc_ls_create_association() [all …]
|
/Linux-v5.4/include/linux/ |
D | nvme-fc.h | 258 struct fcnvme_lsdesc_rqst rqst; member 266 struct fcnvme_lsdesc_rqst rqst; member
|
/Linux-v5.4/drivers/gpu/drm/msm/dsi/phy/ |
D | dsi_phy_14nm.c | 22 u32 rqst = clk_ln ? timing->hs_rqst_ckln : timing->hs_rqst; in dsi_14nm_dphy_set_timing() local 36 DSI_14nm_PHY_LN_TIMING_CTRL_8_HS_RQST(rqst)); in dsi_14nm_dphy_set_timing()
|
/Linux-v5.4/fs/nfsd/ |
D | trace.h | 15 TP_PROTO(const struct svc_rqst *rqst, 17 TP_ARGS(rqst, args_opcnt), 23 __entry->xid = be32_to_cpu(rqst->rq_xid);
|
/Linux-v5.4/net/sunrpc/ |
D | sched.c | 1011 struct rpc_rqst *rqst = task->tk_rqstp; in rpc_malloc() local 1012 size_t size = rqst->rq_callsize + rqst->rq_rcvsize; in rpc_malloc() 1031 rqst->rq_buffer = buf->data; in rpc_malloc() 1032 rqst->rq_rbuffer = (char *)rqst->rq_buffer + rqst->rq_callsize; in rpc_malloc()
|