/Linux-v5.4/drivers/infiniband/hw/mlx4/ |
D | alias_GUID.c | 86 guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid. in mlx4_ib_update_cache_on_guid_change() 102 memcpy(&dev->sriov.demux[port_index].guid_cache[slave_id], in mlx4_ib_update_cache_on_guid_change() 117 return *(__be64 *)&dev->sriov.demux[port - 1].guid_cache[index]; in get_cached_alias_guid() 136 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags); in mlx4_ib_slave_alias_guid_event() 137 if (dev->sriov.alias_guid.ports_guid[port_index].state_flags & in mlx4_ib_slave_alias_guid_event() 141 curr_guid = *(__be64 *)&dev->sriov. in mlx4_ib_slave_alias_guid_event() 154 *(__be64 *)&dev->sriov.alias_guid.ports_guid[port_index]. in mlx4_ib_slave_alias_guid_event() 157 dev->sriov.alias_guid.ports_guid[port_index]. in mlx4_ib_slave_alias_guid_event() 160 dev->sriov.alias_guid.ports_guid[port_index]. in mlx4_ib_slave_alias_guid_event() 164 dev->sriov.alias_guid.ports_guid[port_index]. in mlx4_ib_slave_alias_guid_event() [all …]
|
D | cm.c | 146 struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; in id_map_find_by_sl_id() 173 struct mlx4_ib_sriov *sriov = &dev->sriov; in id_map_ent_timeout() local 174 struct rb_root *sl_id_map = &sriov->sl_id_map; in id_map_ent_timeout() 176 spin_lock(&sriov->id_map_lock); in id_map_ent_timeout() 177 if (!xa_erase(&sriov->pv_id_table, ent->pv_cm_id)) in id_map_ent_timeout() 185 spin_unlock(&sriov->id_map_lock); in id_map_ent_timeout() 191 struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; in id_map_find_del() local 192 struct rb_root *sl_id_map = &sriov->sl_id_map; in id_map_find_del() 195 spin_lock(&sriov->id_map_lock); in id_map_find_del() 196 ent = xa_erase(&sriov->pv_id_table, pv_cm_id); in id_map_find_del() [all …]
|
D | mad.c | 284 if (!dev->sriov.is_going_down) in smp_snoop() 299 !dev->sriov.is_going_down) { in smp_snoop() 438 if (dev->sriov.demux[port - 1].guid_cache[i] == guid) in mlx4_ib_find_real_gid() 526 tun_ctx = dev->sriov.demux[port-1].tun[slave]; in mlx4_ib_send_to_slave() 739 atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix))) { in mlx4_ib_demux_mad() 1090 if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) in handle_lid_change_event() 1101 if (!dev->sriov.is_going_down) { in handle_client_rereg_event() 1102 mlx4_ib_mcg_port_cleanup(&dev->sriov.demux[port_num - 1], 0); in handle_client_rereg_event() 1228 (u64)atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix), in handle_port_mgmt_change_event() 1230 atomic64_set(&dev->sriov.demux[port - 1].subnet_prefix, in handle_port_mgmt_change_event() [all …]
|
D | sysfs.c | 87 spin_lock_irqsave(&mdev->sriov.alias_guid.ag_work_lock, flags); in store_admin_alias_guid() 89 *(__be64 *)&mdev->sriov.alias_guid.ports_guid[port->num - 1]. in store_admin_alias_guid() 95 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].status in store_admin_alias_guid() 102 mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].guid_indexes in store_admin_alias_guid() 105 spin_unlock_irqrestore(&mdev->sriov.alias_guid.ag_work_lock, flags); in store_admin_alias_guid()
|
D | mcg.c | 887 struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1]; in mlx4_ib_mcg_demux_handler() 936 struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1]; in mlx4_ib_mcg_multiplex_handler()
|
D | main.c | 3074 spin_lock_irqsave(&ibdev->sriov.going_down_lock, flags); in do_slave_init() 3075 if (!ibdev->sriov.is_going_down) { in do_slave_init() 3077 queue_work(ibdev->sriov.demux[i].ud_wq, &dm[i]->work); in do_slave_init() 3078 spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); in do_slave_init() 3080 spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); in do_slave_init()
|
D | mlx4_ib.h | 596 struct mlx4_ib_sriov sriov; member
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/ |
D | sriov.c | 41 struct mlx5_core_sriov *sriov = &dev->priv.sriov; in sriov_restore_guids() local 46 if (sriov->vfs_ctx[vf].node_guid || in sriov_restore_guids() 47 sriov->vfs_ctx[vf].port_guid || in sriov_restore_guids() 48 sriov->vfs_ctx[vf].policy != MLX5_POLICY_INVALID) { in sriov_restore_guids() 53 in->node_guid = sriov->vfs_ctx[vf].node_guid; in sriov_restore_guids() 54 in->port_guid = sriov->vfs_ctx[vf].port_guid; in sriov_restore_guids() 55 in->policy = sriov->vfs_ctx[vf].policy; in sriov_restore_guids() 73 struct mlx5_core_sriov *sriov = &dev->priv.sriov; in mlx5_device_enable_sriov() local 95 sriov->vfs_ctx[vf].enabled = 1; in mlx5_device_enable_sriov() 113 struct mlx5_core_sriov *sriov = &dev->priv.sriov; in mlx5_device_disable_sriov() local [all …]
|
D | Makefile | 16 transobj.o vport.o sriov.o fs_cmd.o fs_core.o pci_irq.o \
|
D | eswitch_offloads.c | 1373 mlx5_eswitch_update_num_of_vfs(esw, esw->dev->priv.sriov.num_vfs); in esw_offloads_start() 2594 vport_num <= esw->dev->priv.sriov.max_vfs; in mlx5_eswitch_is_vf_vport()
|
/Linux-v5.4/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_sriov_pf.c | 71 struct qlcnic_sriov *sriov = adapter->ahw->sriov; in qlcnic_sriov_pf_cal_res_limit() local 72 struct qlcnic_resources *res = &sriov->ff_max; in qlcnic_sriov_pf_cal_res_limit() 73 u16 num_macs = sriov->num_allowed_vlans + 1; in qlcnic_sriov_pf_cal_res_limit() 82 num_vfs = sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit() 104 info->max_tx_ques = res->num_tx_queues - sriov->num_vfs; in qlcnic_sriov_pf_cal_res_limit() 109 vp = sriov->vf_info[id].vp; in qlcnic_sriov_pf_cal_res_limit() 140 struct qlcnic_resources *ff_max = &adapter->ahw->sriov->ff_max; in qlcnic_sriov_pf_set_ff_max_res() 159 struct qlcnic_sriov *sriov = adapter->ahw->sriov; in qlcnic_sriov_set_vf_max_vlan() local 163 total_fn = sriov->num_vfs + 1; in qlcnic_sriov_set_vf_max_vlan() 166 sriov->num_allowed_vlans = temp - 1; in qlcnic_sriov_set_vf_max_vlan() [all …]
|
D | qlcnic_sriov_common.c | 143 struct qlcnic_sriov *sriov; in qlcnic_sriov_init() local 153 sriov = kzalloc(sizeof(struct qlcnic_sriov), GFP_KERNEL); in qlcnic_sriov_init() 154 if (!sriov) in qlcnic_sriov_init() 157 adapter->ahw->sriov = sriov; in qlcnic_sriov_init() 158 sriov->num_vfs = num_vfs; in qlcnic_sriov_init() 159 bc = &sriov->bc; in qlcnic_sriov_init() 160 sriov->vf_info = kcalloc(num_vfs, sizeof(struct qlcnic_vf_info), in qlcnic_sriov_init() 162 if (!sriov->vf_info) { in qlcnic_sriov_init() 191 vf = &sriov->vf_info[i]; in qlcnic_sriov_init() 210 sriov->vf_info[i].vp = vp; in qlcnic_sriov_init() [all …]
|
D | qlcnic_sriov.h | 267 static inline void qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov, in qlcnic_sriov_pf_handle_flr() argument
|
/Linux-v5.4/drivers/pci/ |
D | iov.c | 24 return dev->bus->number + ((dev->devfn + dev->sriov->offset + in pci_iov_virtfn_bus() 25 dev->sriov->stride * vf_id) >> 8); in pci_iov_virtfn_bus() 32 return (dev->devfn + dev->sriov->offset + in pci_iov_virtfn_devfn() 33 dev->sriov->stride * vf_id) & 0xff; in pci_iov_virtfn_devfn() 44 struct pci_sriov *iov = dev->sriov; in pci_iov_set_numvfs() 60 struct pci_sriov *iov = dev->sriov; in compute_max_vf_buses() 111 return dev->sriov->barsz[resno - PCI_IOV_RESOURCES]; in pci_iov_resource_size() 128 &physfn->sriov->class); in pci_read_vf_config_common() 130 &physfn->sriov->hdr_type); in pci_read_vf_config_common() 132 &physfn->sriov->subsystem_vendor); in pci_read_vf_config_common() [all …]
|
D | probe.c | 1613 return dev->physfn->sriov->class; in pci_class() 1623 *vendor = dev->physfn->sriov->subsystem_vendor; in pci_subsystem_ids() 1624 *device = dev->physfn->sriov->subsystem_device; in pci_subsystem_ids() 1638 return dev->physfn->sriov->hdr_type; in pci_hdr_type()
|
D | pci-driver.c | 402 return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || in pci_device_can_probe()
|
/Linux-v5.4/drivers/infiniband/hw/mlx5/ |
D | ib_virt.c | 99 struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; in mlx5_ib_set_vf_link_state() 157 struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; in set_vf_node_guid() 178 struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; in set_vf_port_guid()
|
/Linux-v5.4/drivers/net/ethernet/sfc/ |
D | Makefile | 6 sfc-$(CONFIG_SFC_SRIOV) += sriov.o siena_sriov.o ef10_sriov.o
|
D | siena_sriov.c | 1243 unsigned index, devfn, sriov, buftbl_base; in efx_siena_sriov_vfs_init() local 1248 sriov = pci_find_ext_capability(pci_dev, PCI_EXT_CAP_ID_SRIOV); in efx_siena_sriov_vfs_init() 1249 if (!sriov) in efx_siena_sriov_vfs_init() 1252 pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_OFFSET, &offset); in efx_siena_sriov_vfs_init() 1253 pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_STRIDE, &stride); in efx_siena_sriov_vfs_init()
|
/Linux-v5.4/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.h | 218 #define BNX2X_NR_VIRTFN(bp) ((bp)->vfdb->sriov.nr_virtfn) 240 #define GET_NUM_VFS_PER_PF(bp) ((bp)->vfdb ? (bp)->vfdb->sriov.total \ 332 struct bnx2x_sriov sriov; member 540 return bp->vfdb->sriov.nr_virtfn * BNX2X_CIDS_PER_VF; in bnx2x_vf_headroom()
|
D | bnx2x_sriov.c | 1055 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_bus() 1064 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_devfn() 1073 struct bnx2x_sriov *iov = &bp->vfdb->sriov; in bnx2x_vf_set_bars() 1242 iov = &(bp->vfdb->sriov); in bnx2x_iov_init_one() 1321 for (vf_idx = 0; vf_idx < bp->vfdb->sriov.total; vf_idx++) { in bnx2x_iov_remove_one() 1324 bp->vfdb->sriov.first_vf_in_pf + in bnx2x_iov_remove_one() 1327 bp->vfdb->sriov.first_vf_in_pf + vf_idx); in bnx2x_iov_remove_one() 1371 tot_size = (BP_VFDB(bp)->sriov.first_vf_in_pf + BNX2X_NR_VIRTFN(bp)) * in bnx2x_iov_alloc_mem() 1557 DP(BNX2X_MSG_IOV, "num of vfs: %d\n", (bp)->vfdb->sriov.nr_virtfn); in bnx2x_iov_nic_init() 1566 int base_vf_cid = (BP_VFDB(bp)->sriov.first_vf_in_pf + vfid) * in bnx2x_iov_nic_init() [all …]
|
/Linux-v5.4/include/linux/mlx5/ |
D | driver.h | 583 struct mlx5_core_sriov sriov; member 1142 return dev->priv.sriov.max_vfs; in mlx5_core_max_vfs()
|
/Linux-v5.4/drivers/scsi/bfa/ |
D | bfa_defs.h | 769 u8 sriov; member
|
/Linux-v5.4/drivers/net/hyperv/ |
D | hyperv_net.h | 517 u64 sriov:1; member
|
D | netvsc.c | 529 init_packet->msg.v2_msg.send_ndis_config.capability.sriov = 1; in negotiate_nvsp_ver()
|