/Linux-v5.15/drivers/net/ethernet/chelsio/cxgb/ |
D | sge.c | 242 struct sge *sge; member 256 struct sge { struct 270 u32 sge_control; /* shadow value of sge control reg */ argument 271 struct sge_intr_counts stats; argument 272 struct sge_port_stats __percpu *port_stats[MAX_NPORTS]; argument 284 static void tx_sched_stop(struct sge *sge) in tx_sched_stop() 299 unsigned int t1_sched_update_parms(struct sge *sge, unsigned int port, in t1_sched_update_parms() 373 static int tx_sched_init(struct sge *sge) in tx_sched_init() 400 static inline int sched_update_avail(struct sge *sge) in sched_update_avail() 434 static struct sk_buff *sched_skb(struct sge *sge, struct sk_buff *skb, in sched_skb() [all …]
|
/Linux-v5.15/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | adapter.h | 270 struct sge { struct 274 struct sge_eth_txq ethtxq[MAX_ETH_QSETS]; argument 275 struct sge_eth_rxq ethrxq[MAX_ETH_QSETS]; argument 281 struct sge_rspq fw_evtq ____cacheline_aligned_in_smp; argument 283 struct sge_rspq intrq ____cacheline_aligned_in_smp; argument 284 spinlock_t intrq_lock; 294 struct timer_list rx_timer; 299 struct timer_list tx_timer; 306 u16 max_ethqsets; /* # of available Ethernet queue sets */ 307 u16 ethqsets; /* # of active Ethernet queue sets */ [all …]
|
/Linux-v5.15/drivers/net/ethernet/huawei/hinic/ |
D | hinic_common.c | 55 void hinic_set_sge(struct hinic_sge *sge, dma_addr_t addr, int len) in hinic_set_sge() 68 dma_addr_t hinic_sge_to_dma(struct hinic_sge *sge) in hinic_sge_to_dma()
|
D | hinic_rx.c | 200 struct hinic_sge sge; in rx_alloc_pkts() local 249 struct hinic_sge sge; in free_all_rx_skbs() local 279 struct hinic_sge sge; in rx_recv_jumbo_pkt() local 367 struct hinic_sge sge; in rxq_recv() local
|
D | hinic_hw_wqe.h | 334 struct hinic_sge sge; member 353 struct hinic_sge sge; member 403 struct hinic_sge sge; member 430 struct hinic_sge sge; member
|
/Linux-v5.15/include/rdma/ |
D | rdmavt_mr.h | 78 struct rvt_sge sge; /* progress state for the current SGE */ member 102 static inline u32 rvt_get_sge_length(struct rvt_sge *sge, u32 length) in rvt_get_sge_length() 117 struct rvt_sge *sge = &ss->sge; in rvt_update_sge() local 141 struct rvt_sge *sge = &ss->sge; in rvt_skip_sge() local
|
D | rw.h | 24 struct ib_sge sge; member 36 struct ib_sge sge; member
|
/Linux-v5.15/net/core/ |
D | skmsg.c | 34 struct scatterlist *sge; in sk_msg_alloc() local 81 struct scatterlist *sge = sk_msg_elem(src, i); in sk_msg_clone() local 133 struct scatterlist *sge = sk_msg_elem(msg, i); in sk_msg_return_zero() local 157 struct scatterlist *sge = &msg->sg.data[i]; in sk_msg_return() local 170 struct scatterlist *sge = sk_msg_elem(msg, i); in sk_msg_free_elem() local 186 struct scatterlist *sge = sk_msg_elem(msg, i); in __sk_msg_free() local 216 struct scatterlist *sge; in __sk_msg_free_partial() local 363 struct scatterlist *sge; in sk_msg_memcopy_from_iter() local 413 struct scatterlist *sge; in sk_msg_recvmsg() local
|
/Linux-v5.15/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4.h | 437 struct sge_params sge; member 964 struct sge { struct 965 struct sge_eth_txq ethtxq[MAX_ETH_QSETS]; argument 966 struct sge_eth_txq ptptxq; argument 967 struct sge_ctrl_txq ctrlq[MAX_CTRL_QUEUES]; argument 969 struct sge_eth_rxq ethrxq[MAX_ETH_QSETS]; argument 970 struct sge_rspq fw_evtq ____cacheline_aligned_in_smp; argument 971 struct sge_uld_rxq_info **uld_rxq_info; argument 972 struct sge_uld_txq_info **uld_txq_info; argument 974 struct sge_rspq intrq ____cacheline_aligned_in_smp; argument [all …]
|
/Linux-v5.15/drivers/scsi/csiostor/ |
D | csio_wr.c | 58 csio_get_flbuf_size(struct csio_hw *hw, struct csio_sge *sge, uint32_t reg) in csio_get_flbuf_size() 66 csio_wr_fl_bufsz(struct csio_sge *sge, struct csio_dma_buf *buf) in csio_wr_fl_bufsz() 118 struct csio_sge *sge = &wrm->sge; in csio_wr_fill_fl() local 1050 struct csio_sge *sge = &wrm->sge; in csio_wr_process_fl() local 1313 struct csio_sge *sge = &wrm->sge; in csio_wr_fixup_host_params() local 1436 struct csio_sge *sge = &wrm->sge; in csio_init_intr_coalesce_parms() local 1458 struct csio_sge *sge = &wrm->sge; in csio_wr_get_sge() local 1527 struct csio_sge *sge = &wrm->sge; in csio_wr_set_sge() local
|
/Linux-v5.15/drivers/infiniband/sw/siw/ |
D | siw_qp_tx.c | 43 struct siw_sge *sge = &wqe->sqe.sge[0]; in siw_try_1seg() local 370 struct siw_sge *sge, unsigned int offset, in siw_0copy_tx() 434 struct siw_sge *sge = &wqe->sqe.sge[c_tx->sge_idx]; in siw_tx_hdt() local 764 struct siw_sge *sge = &wqe->sqe.sge[0]; in siw_check_sgl_tx() local
|
D | siw_verbs.h | 24 static inline void siw_copy_sgl(struct ib_sge *sge, struct siw_sge *siw_sge, in siw_copy_sgl()
|
D | siw_qp_rx.c | 457 struct siw_sge *sge; in siw_proc_send() local 792 struct siw_sge *sge; in siw_proc_rresp() local
|
/Linux-v5.15/net/sunrpc/xprtrdma/ |
D | rpc_rdma.c | 536 struct ib_sge *sge; in rpcrdma_sendctx_unmap() local 560 struct ib_sge *sge = &sc->sc_sges[req->rl_wr.num_sge++]; in rpcrdma_prepare_hdr_sge() local 577 struct ib_sge *sge = &sc->sc_sges[req->rl_wr.num_sge++]; in rpcrdma_prepare_head_iov() local 602 struct ib_sge *sge; in rpcrdma_prepare_pagelist() local 640 struct ib_sge *sge = &sc->sc_sges[req->rl_wr.num_sge++]; in rpcrdma_prepare_tail_iov() local
|
/Linux-v5.15/drivers/infiniband/sw/rdmavt/ |
D | mr.c | 682 struct ib_sge *sge) in rvt_sge_adjacent() 718 struct ib_sge *sge, int acc) in rvt_lkey_ok() 826 int rvt_rkey_ok(struct rvt_qp *qp, struct rvt_sge *sge, in rvt_rkey_ok()
|
D | qp.c | 628 struct rvt_sge *sge = &wqe->sg_list[i]; in rvt_swqe_has_lkey() local 2128 struct rvt_sge *sge = &wqe->sg_list[--j]; in rvt_post_one_wr() local 2254 static struct ib_sge *rvt_cast_sge(struct rvt_wqe_sge *sge) in rvt_cast_sge() 2301 struct rvt_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; in init_sge() local 2818 struct rvt_sge *sge = &ss->sge; in rvt_copy_sge() local 2910 struct rvt_sge *sge; in rvt_ruc_loopback() local
|
/Linux-v5.15/net/rds/ |
D | ib_recv.c | 53 struct ib_sge *sge; in rds_ib_recv_init_ring() local 311 struct ib_sge *sge; in rds_ib_recv_refill_one() local 576 struct ib_sge *sge = &ic->i_ack_sge; in rds_ib_recv_init_ack() local
|
/Linux-v5.15/drivers/infiniband/hw/irdma/ |
D | uk.c | 16 static void irdma_set_fragment(__le64 *wqe, u32 offset, struct irdma_sge *sge, in irdma_set_fragment() 41 struct irdma_sge *sge, u8 valid) in irdma_set_fragment_gen_1() 829 struct irdma_sge sge = {}; in irdma_uk_stag_local_invalidate() local 1337 void irdma_get_wqe_shift(struct irdma_uk_attrs *uk_attrs, u32 sge, in irdma_get_wqe_shift()
|
/Linux-v5.15/drivers/scsi/be2iscsi/ |
D | be_mgmt.c | 128 struct be_sge *sge; in mgmt_open_connection() local 230 struct be_sge *sge; in beiscsi_exec_nemb_cmd() local 1011 struct be_sge *sge; in beiscsi_boot_get_sinfo() local 1509 struct be_sge *sge; in beiscsi_mgmt_invalidate_icds() local
|
/Linux-v5.15/drivers/net/ethernet/microsoft/mana/ |
D | hw_channel.c | 84 struct gdma_sge *sge; in mana_hwc_post_rx_wqe() local 184 struct gdma_sge *sge; in mana_hwc_rx_event_handler() local 514 struct gdma_sge *sge; in mana_hwc_post_tx_wqe() local
|
/Linux-v5.15/drivers/infiniband/sw/rxe/ |
D | rxe_mr.c | 376 struct rxe_sge *sge = &dma->sge[dma->cur_sge]; in copy_data() local 462 struct rxe_sge *sge = &dma->sge[dma->cur_sge]; in advance_dma_data() local
|
/Linux-v5.15/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_qp.c | 678 struct pvrdma_sge *sge; in pvrdma_post_send() local 888 struct pvrdma_sge *sge; in pvrdma_post_recv() local
|
/Linux-v5.15/net/ipv4/ |
D | tcp_bpf.c | 17 struct scatterlist *sge; in bpf_tcp_ingress() local 70 struct scatterlist *sge; in tcp_bpf_push() local
|
/Linux-v5.15/drivers/net/ethernet/chelsio/cxgb3/ |
D | adapter.h | 215 struct sge { struct 216 struct sge_qset qs[SGE_QSETS]; argument 217 spinlock_t reg_lock; /* guards non-atomic SGE registers (eg context) */ 244 struct sge sge; member
|
/Linux-v5.15/include/uapi/rdma/ |
D | siw-abi.h | 119 struct siw_sge sge[SIW_MAX_SGE]; member 135 struct siw_sge sge[SIW_MAX_SGE]; member
|