Lines Matching refs:vpath
25 status = __vxge_hw_vpath_stats_access(vpath, \
50 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_wait_receive_idle() local
54 vpath = &hldev->virtual_paths[vp_id]; in vxge_hw_vpath_wait_receive_idle()
55 vp_reg = vpath->vp_reg; in vxge_hw_vpath_wait_receive_idle()
157 vxge_hw_vpath_fw_api(struct __vxge_hw_virtualpath *vpath, u32 action, in vxge_hw_vpath_fw_api() argument
161 struct vxge_hw_vpath_reg __iomem *vp_reg = vpath->vp_reg; in vxge_hw_vpath_fw_api()
166 spin_lock(&vpath->lock); in vxge_hw_vpath_fw_api()
167 if (!vpath->vp_open) { in vxge_hw_vpath_fw_api()
168 spin_unlock(&vpath->lock); in vxge_hw_vpath_fw_api()
196 if (!vpath->vp_open) in vxge_hw_vpath_fw_api()
216 if (vpath->vp_open) in vxge_hw_vpath_fw_api()
217 spin_unlock(&vpath->lock); in vxge_hw_vpath_fw_api()
226 struct __vxge_hw_virtualpath *vpath; in vxge_hw_upgrade_read_version() local
229 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_upgrade_read_version()
231 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_upgrade_read_version()
249 struct __vxge_hw_virtualpath *vpath; in vxge_hw_flash_fw() local
253 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_flash_fw()
255 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_flash_fw()
280 struct __vxge_hw_virtualpath *vpath; in vxge_update_fw_image() local
284 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_update_fw_image()
287 status = vxge_hw_vpath_fw_api(vpath, in vxge_update_fw_image()
306 status = vxge_hw_vpath_fw_api(vpath, in vxge_update_fw_image()
377 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_eprom_img_ver_get() local
381 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_vpath_eprom_img_ver_get()
387 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_vpath_eprom_img_ver_get()
426 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_channel_initialize() local
428 vpath = channel->vph->vpath; in __vxge_hw_channel_initialize()
437 vpath->fifoh = (struct __vxge_hw_fifo *)channel; in __vxge_hw_channel_initialize()
442 vpath->ringh = (struct __vxge_hw_ring *)channel; in __vxge_hw_channel_initialize()
804 __vxge_hw_vpath_fw_ver_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_fw_ver_get() argument
814 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_fw_ver_get()
871 __vxge_hw_vpath_card_info_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_card_info_get() argument
883 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
896 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
911 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
930 __vxge_hw_vpath_pci_func_mode_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_pci_func_mode_get() argument
938 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_pci_func_mode_get()
954 __vxge_hw_vpath_addr_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_addr_get() argument
963 status = vxge_hw_vpath_fw_api(vpath, action, in __vxge_hw_vpath_addr_get()
1009 struct __vxge_hw_virtualpath vpath; in vxge_hw_device_hw_info_get() local
1057 spin_lock_init(&vpath.lock); in vxge_hw_device_hw_info_get()
1058 vpath.vp_reg = bar0 + val64; in vxge_hw_device_hw_info_get()
1059 vpath.vp_open = VXGE_HW_VP_NOT_OPEN; in vxge_hw_device_hw_info_get()
1061 status = __vxge_hw_vpath_pci_func_mode_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1065 status = __vxge_hw_vpath_fw_ver_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1069 status = __vxge_hw_vpath_card_info_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1081 vpath.vp_reg = bar0 + val64; in vxge_hw_device_hw_info_get()
1082 vpath.vp_open = VXGE_HW_VP_NOT_OPEN; in vxge_hw_device_hw_info_get()
1084 status = __vxge_hw_vpath_addr_get(&vpath, in vxge_hw_device_hw_info_get()
1412 __vxge_hw_vpath_stats_access(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_stats_access() argument
1419 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_stats_access()
1424 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_stats_access()
1433 vpath->hldev->config.device_poll_millis); in __vxge_hw_vpath_stats_access()
1447 __vxge_hw_vpath_xmac_tx_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_xmac_tx_stats_get() argument
1457 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_xmac_tx_stats_get()
1463 status = __vxge_hw_vpath_stats_access(vpath, in __vxge_hw_vpath_xmac_tx_stats_get()
1479 __vxge_hw_vpath_xmac_rx_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_xmac_rx_stats_get() argument
1488 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_xmac_rx_stats_get()
1493 status = __vxge_hw_vpath_stats_access(vpath, in __vxge_hw_vpath_xmac_rx_stats_get()
1510 __vxge_hw_vpath_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_stats_get() argument
1517 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_stats_get()
1521 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_stats_get()
1581 status = __vxge_hw_vpath_xmac_tx_stats_get(vpath, &hw_stats->tx_stats); in __vxge_hw_vpath_stats_get()
1585 status = __vxge_hw_vpath_xmac_rx_stats_get(vpath, &hw_stats->rx_stats); in __vxge_hw_vpath_stats_get()
2193 hldev = vph->vpath->hldev; in __vxge_hw_channel_allocate()
2194 vp_id = vph->vpath->vp_id; in __vxge_hw_channel_allocate()
2769 struct __vxge_hw_ring *ring = vp->vpath->ringh; in __vxge_hw_ring_delete()
2776 vp->vpath->ringh = NULL; in __vxge_hw_ring_delete()
2805 hldev = vp->vpath->hldev; in __vxge_hw_ring_create()
2806 vp_id = vp->vpath->vp_id; in __vxge_hw_ring_create()
2823 vp->vpath->ringh = ring; in __vxge_hw_ring_create()
2825 ring->vp_reg = vp->vpath->vp_reg; in __vxge_hw_ring_create()
2827 ring->stats = &vp->vpath->sw_stats->ring_stats; in __vxge_hw_ring_create()
2833 ring->tim_rti_cfg1_saved = vp->vpath->tim_rti_cfg1_saved; in __vxge_hw_ring_create()
2834 ring->tim_rti_cfg3_saved = vp->vpath->tim_rti_cfg3_saved; in __vxge_hw_ring_create()
3373 struct __vxge_hw_fifo *fifo = vp->vpath->fifoh; in __vxge_hw_fifo_delete()
3380 vp->vpath->fifoh = NULL; in __vxge_hw_fifo_delete()
3442 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_fifo_create() local
3448 vpath = vp->vpath; in __vxge_hw_fifo_create()
3449 config = &vpath->hldev->config.vp_config[vpath->vp_id].fifo; in __vxge_hw_fifo_create()
3465 vpath->fifoh = fifo; in __vxge_hw_fifo_create()
3466 fifo->nofl_db = vpath->nofl_db; in __vxge_hw_fifo_create()
3468 fifo->vp_id = vpath->vp_id; in __vxge_hw_fifo_create()
3469 fifo->vp_reg = vpath->vp_reg; in __vxge_hw_fifo_create()
3470 fifo->stats = &vpath->sw_stats->fifo_stats; in __vxge_hw_fifo_create()
3476 fifo->tim_tti_cfg1_saved = vpath->tim_tti_cfg1_saved; in __vxge_hw_fifo_create()
3477 fifo->tim_tti_cfg3_saved = vpath->tim_tti_cfg3_saved; in __vxge_hw_fifo_create()
3527 __vxge_hw_mempool_create(vpath->hldev, in __vxge_hw_fifo_create()
3559 __vxge_hw_vpath_pci_read(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_pci_read() argument
3564 struct vxge_hw_vpath_reg __iomem *vp_reg = vpath->vp_reg; in __vxge_hw_vpath_pci_read()
3605 struct __vxge_hw_virtualpath *vpath; in vxge_hw_device_flick_link_led() local
3614 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_device_flick_link_led()
3617 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_device_flick_link_led()
3652 status = vxge_hw_vpath_fw_api(vp->vpath, action, rts_table, offset, in __vxge_hw_vpath_rts_table_get()
3688 status = vxge_hw_vpath_fw_api(vp->vpath, action, rts_table, offset, in __vxge_hw_vpath_rts_table_set()
3815 if (vp->vpath->hldev->config.rth_it_type in vxge_hw_vpath_rts_rth_itable_set()
3958 struct __vxge_hw_virtualpath *vpath) in __vxge_hw_vpath_mgmt_read() argument
3965 val64 = readq(&vpath->vpmgmt_reg-> in __vxge_hw_vpath_mgmt_read()
3975 vpath->max_mtu = mtu + VXGE_HW_MAC_HEADER_MAX_SIZE; in __vxge_hw_vpath_mgmt_read()
3977 val64 = readq(&vpath->vpmgmt_reg->xmac_vsport_choices_vp); in __vxge_hw_vpath_mgmt_read()
3981 vpath->vsport_number = i; in __vxge_hw_vpath_mgmt_read()
3984 val64 = readq(&vpath->vpmgmt_reg->xgmac_gen_status_vpmgmt_clone); in __vxge_hw_vpath_mgmt_read()
3987 VXGE_HW_DEVICE_LINK_STATE_SET(vpath->hldev, VXGE_HW_LINK_UP); in __vxge_hw_vpath_mgmt_read()
3989 VXGE_HW_DEVICE_LINK_STATE_SET(vpath->hldev, VXGE_HW_LINK_DOWN); in __vxge_hw_vpath_mgmt_read()
4000 __vxge_hw_vpath_reset_check(struct __vxge_hw_virtualpath *vpath) in __vxge_hw_vpath_reset_check() argument
4005 &vpath->hldev->common_reg->vpath_rst_in_prog, in __vxge_hw_vpath_reset_check()
4007 1 << (16 - vpath->vp_id)), in __vxge_hw_vpath_reset_check()
4008 vpath->hldev->config.device_poll_millis); in __vxge_hw_vpath_reset_check()
4038 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_sw_reset() local
4040 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_sw_reset()
4042 if (vpath->ringh) { in __vxge_hw_vpath_sw_reset()
4043 status = __vxge_hw_ring_reset(vpath->ringh); in __vxge_hw_vpath_sw_reset()
4048 if (vpath->fifoh) in __vxge_hw_vpath_sw_reset()
4049 status = __vxge_hw_fifo_reset(vpath->fifoh); in __vxge_hw_vpath_sw_reset()
4063 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_prc_configure() local
4067 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_prc_configure()
4068 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_prc_configure()
4069 vp_config = vpath->vp_config; in __vxge_hw_vpath_prc_configure()
4078 val64 = readq(&vpath->vp_reg->prc_cfg6); in __vxge_hw_vpath_prc_configure()
4080 writeq(val64, &vpath->vp_reg->prc_cfg6); in __vxge_hw_vpath_prc_configure()
4084 if (vpath->vp_config->ring.scatter_mode != in __vxge_hw_vpath_prc_configure()
4089 switch (vpath->vp_config->ring.scatter_mode) { in __vxge_hw_vpath_prc_configure()
4109 vpath->ringh) >> 3), &vp_reg->prc_cfg5); in __vxge_hw_vpath_prc_configure()
4137 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_kdfc_configure() local
4140 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_kdfc_configure()
4141 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_kdfc_configure()
4149 vpath->max_kdfc_db = in __vxge_hw_vpath_kdfc_configure()
4153 if (vpath->vp_config->fifo.enable == VXGE_HW_FIFO_ENABLE) { in __vxge_hw_vpath_kdfc_configure()
4155 vpath->max_nofl_db = vpath->max_kdfc_db; in __vxge_hw_vpath_kdfc_configure()
4157 if (vpath->max_nofl_db < in __vxge_hw_vpath_kdfc_configure()
4158 ((vpath->vp_config->fifo.memblock_size / in __vxge_hw_vpath_kdfc_configure()
4159 (vpath->vp_config->fifo.max_frags * in __vxge_hw_vpath_kdfc_configure()
4161 vpath->vp_config->fifo.fifo_blocks)) { in __vxge_hw_vpath_kdfc_configure()
4166 (vpath->max_nofl_db*2)-1); in __vxge_hw_vpath_kdfc_configure()
4191 vpath->nofl_db = in __vxge_hw_vpath_kdfc_configure()
4208 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_mac_configure() local
4212 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_mac_configure()
4213 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_mac_configure()
4214 vp_config = vpath->vp_config; in __vxge_hw_vpath_mac_configure()
4217 vpath->vsport_number), &vp_reg->xmac_vsport_choice); in __vxge_hw_vpath_mac_configure()
4238 VXGE_HW_MAC_HEADER_MAX_SIZE) < vpath->max_mtu) in __vxge_hw_vpath_mac_configure()
4244 vpath->max_mtu); in __vxge_hw_vpath_mac_configure()
4275 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_tim_configure() local
4279 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_tim_configure()
4280 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_tim_configure()
4281 config = vpath->vp_config; in __vxge_hw_vpath_tim_configure()
4343 vpath->tim_tti_cfg1_saved = val64; in __vxge_hw_vpath_tim_configure()
4401 vpath->tim_tti_cfg3_saved = val64; in __vxge_hw_vpath_tim_configure()
4450 vpath->tim_rti_cfg1_saved = val64; in __vxge_hw_vpath_tim_configure()
4508 vpath->tim_rti_cfg3_saved = val64; in __vxge_hw_vpath_tim_configure()
4538 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_initialize() local
4541 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_initialize()
4547 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_initialize()
4549 status = __vxge_hw_vpath_swapper_set(vpath->vp_reg); in __vxge_hw_vpath_initialize()
4568 status = __vxge_hw_vpath_pci_read(vpath, 1, 0x78, &val32); in __vxge_hw_vpath_initialize()
4597 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vp_terminate() local
4599 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vp_terminate()
4601 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) in __vxge_hw_vp_terminate()
4604 VXGE_HW_DEVICE_TIM_INT_MASK_RESET(vpath->hldev->tim_int_mask0, in __vxge_hw_vp_terminate()
4605 vpath->hldev->tim_int_mask1, vpath->vp_id); in __vxge_hw_vp_terminate()
4606 hldev->stats.hw_dev_info_stats.vpath_info[vpath->vp_id] = NULL; in __vxge_hw_vp_terminate()
4611 spin_lock(&vpath->lock); in __vxge_hw_vp_terminate()
4612 vpath->vp_open = VXGE_HW_VP_NOT_OPEN; in __vxge_hw_vp_terminate()
4613 spin_unlock(&vpath->lock); in __vxge_hw_vp_terminate()
4615 vpath->vpmgmt_reg = NULL; in __vxge_hw_vp_terminate()
4616 vpath->nofl_db = NULL; in __vxge_hw_vp_terminate()
4617 vpath->max_mtu = 0; in __vxge_hw_vp_terminate()
4618 vpath->vsport_number = 0; in __vxge_hw_vp_terminate()
4619 vpath->max_kdfc_db = 0; in __vxge_hw_vp_terminate()
4620 vpath->max_nofl_db = 0; in __vxge_hw_vp_terminate()
4621 vpath->ringh = NULL; in __vxge_hw_vp_terminate()
4622 vpath->fifoh = NULL; in __vxge_hw_vp_terminate()
4623 memset(&vpath->vpath_handles, 0, sizeof(struct list_head)); in __vxge_hw_vp_terminate()
4624 vpath->stats_block = NULL; in __vxge_hw_vp_terminate()
4625 vpath->hw_stats = NULL; in __vxge_hw_vp_terminate()
4626 vpath->hw_stats_sav = NULL; in __vxge_hw_vp_terminate()
4627 vpath->sw_stats = NULL; in __vxge_hw_vp_terminate()
4642 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vp_initialize() local
4650 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vp_initialize()
4652 spin_lock_init(&vpath->lock); in __vxge_hw_vp_initialize()
4653 vpath->vp_id = vp_id; in __vxge_hw_vp_initialize()
4654 vpath->vp_open = VXGE_HW_VP_OPEN; in __vxge_hw_vp_initialize()
4655 vpath->hldev = hldev; in __vxge_hw_vp_initialize()
4656 vpath->vp_config = config; in __vxge_hw_vp_initialize()
4657 vpath->vp_reg = hldev->vpath_reg[vp_id]; in __vxge_hw_vp_initialize()
4658 vpath->vpmgmt_reg = hldev->vpmgmt_reg[vp_id]; in __vxge_hw_vp_initialize()
4662 status = __vxge_hw_vpath_reset_check(vpath); in __vxge_hw_vp_initialize()
4664 memset(vpath, 0, sizeof(struct __vxge_hw_virtualpath)); in __vxge_hw_vp_initialize()
4668 status = __vxge_hw_vpath_mgmt_read(hldev, vpath); in __vxge_hw_vp_initialize()
4670 memset(vpath, 0, sizeof(struct __vxge_hw_virtualpath)); in __vxge_hw_vp_initialize()
4674 INIT_LIST_HEAD(&vpath->vpath_handles); in __vxge_hw_vp_initialize()
4676 vpath->sw_stats = &hldev->stats.sw_dev_info_stats.vpath_info[vp_id]; in __vxge_hw_vp_initialize()
4698 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_mtu_set() local
4704 vpath = vp->vpath; in vxge_hw_vpath_mtu_set()
4708 if ((new_mtu < VXGE_HW_MIN_MTU) || (new_mtu > vpath->max_mtu)) in vxge_hw_vpath_mtu_set()
4711 val64 = readq(&vpath->vp_reg->rxmac_vcfg0); in vxge_hw_vpath_mtu_set()
4716 writeq(val64, &vpath->vp_reg->rxmac_vcfg0); in vxge_hw_vpath_mtu_set()
4718 vpath->vp_config->mtu = new_mtu - VXGE_HW_MAC_HEADER_MAX_SIZE; in vxge_hw_vpath_mtu_set()
4733 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_stats_enable() local
4735 vpath = vp->vpath; in vxge_hw_vpath_stats_enable()
4737 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_stats_enable()
4742 memcpy(vpath->hw_stats_sav, vpath->hw_stats, in vxge_hw_vpath_stats_enable()
4745 status = __vxge_hw_vpath_stats_get(vpath, vpath->hw_stats); in vxge_hw_vpath_stats_enable()
4793 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_open() local
4797 vpath = &hldev->virtual_paths[attr->vp_id]; in vxge_hw_vpath_open()
4799 if (vpath->vp_open == VXGE_HW_VP_OPEN) { in vxge_hw_vpath_open()
4815 vp->vpath = vpath; in vxge_hw_vpath_open()
4817 if (vpath->vp_config->fifo.enable == VXGE_HW_FIFO_ENABLE) { in vxge_hw_vpath_open()
4823 if (vpath->vp_config->ring.enable == VXGE_HW_RING_ENABLE) { in vxge_hw_vpath_open()
4831 vpath->fifoh->tx_intr_num = in vxge_hw_vpath_open()
4835 vpath->stats_block = __vxge_hw_blockpool_block_allocate(hldev, in vxge_hw_vpath_open()
4837 if (vpath->stats_block == NULL) { in vxge_hw_vpath_open()
4842 vpath->hw_stats = vpath->stats_block->memblock; in vxge_hw_vpath_open()
4843 memset(vpath->hw_stats, 0, in vxge_hw_vpath_open()
4847 vpath->hw_stats; in vxge_hw_vpath_open()
4849 vpath->hw_stats_sav = in vxge_hw_vpath_open()
4851 memset(vpath->hw_stats_sav, 0, in vxge_hw_vpath_open()
4854 writeq(vpath->stats_block->dma_addr, &vpath->vp_reg->stats_cfg); in vxge_hw_vpath_open()
4860 list_add(&vp->item, &vpath->vpath_handles); in vxge_hw_vpath_open()
4862 hldev->vpaths_deployed |= vxge_mBIT(vpath->vp_id); in vxge_hw_vpath_open()
4866 attr->fifo_attr.userdata = vpath->fifoh; in vxge_hw_vpath_open()
4867 attr->ring_attr.userdata = vpath->ringh; in vxge_hw_vpath_open()
4872 if (vpath->ringh != NULL) in vxge_hw_vpath_open()
4875 if (vpath->fifoh != NULL) in vxge_hw_vpath_open()
4896 struct __vxge_hw_virtualpath *vpath = vp->vpath; in vxge_hw_vpath_rx_doorbell_init() local
4897 struct __vxge_hw_ring *ring = vpath->ringh; in vxge_hw_vpath_rx_doorbell_init()
4898 struct vxgedev *vdev = netdev_priv(vpath->hldev->ndev); in vxge_hw_vpath_rx_doorbell_init()
4902 new_count = readq(&vpath->vp_reg->rxdmem_size); in vxge_hw_vpath_rx_doorbell_init()
4910 &vpath->vp_reg->prc_rxd_doorbell); in vxge_hw_vpath_rx_doorbell_init()
4911 readl(&vpath->vp_reg->prc_rxd_doorbell); in vxge_hw_vpath_rx_doorbell_init()
4914 val64 = readq(&vpath->vp_reg->prc_cfg6); in vxge_hw_vpath_rx_doorbell_init()
4959 struct __vxge_hw_virtualpath *vpath = NULL; in vxge_hw_vpath_close() local
4961 u32 vp_id = vp->vpath->vp_id; in vxge_hw_vpath_close()
4965 vpath = vp->vpath; in vxge_hw_vpath_close()
4966 devh = vpath->hldev; in vxge_hw_vpath_close()
4968 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_close()
4975 if (!list_empty(&vpath->vpath_handles)) { in vxge_hw_vpath_close()
4976 list_add(&vp->item, &vpath->vpath_handles); in vxge_hw_vpath_close()
4987 if (vpath->ringh != NULL) in vxge_hw_vpath_close()
4990 if (vpath->fifoh != NULL) in vxge_hw_vpath_close()
4993 if (vpath->stats_block != NULL) in vxge_hw_vpath_close()
4994 __vxge_hw_blockpool_block_free(devh, vpath->stats_block); in vxge_hw_vpath_close()
5012 struct __vxge_hw_virtualpath *vpath = vp->vpath; in vxge_hw_vpath_reset() local
5014 vp_id = vpath->vp_id; in vxge_hw_vpath_reset()
5016 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_reset()
5021 status = __vxge_hw_vpath_reset(vpath->hldev, vp_id); in vxge_hw_vpath_reset()
5023 vpath->sw_stats->soft_reset_cnt++; in vxge_hw_vpath_reset()
5036 struct __vxge_hw_virtualpath *vpath = NULL; in vxge_hw_vpath_recover_from_reset() local
5041 vp_id = vp->vpath->vp_id; in vxge_hw_vpath_recover_from_reset()
5042 vpath = vp->vpath; in vxge_hw_vpath_recover_from_reset()
5043 hldev = vpath->hldev; in vxge_hw_vpath_recover_from_reset()
5045 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_recover_from_reset()
5050 status = __vxge_hw_vpath_reset_check(vpath); in vxge_hw_vpath_recover_from_reset()
5062 if (vpath->ringh != NULL) in vxge_hw_vpath_recover_from_reset()
5065 memset(vpath->hw_stats, 0, in vxge_hw_vpath_recover_from_reset()
5068 memset(vpath->hw_stats_sav, 0, in vxge_hw_vpath_recover_from_reset()
5071 writeq(vpath->stats_block->dma_addr, in vxge_hw_vpath_recover_from_reset()
5072 &vpath->vp_reg->stats_cfg); in vxge_hw_vpath_recover_from_reset()
5091 hldev = vp->vpath->hldev; in vxge_hw_vpath_enable()
5094 1 << (16 - vp->vpath->vp_id)); in vxge_hw_vpath_enable()