Lines Matching refs:vpath
25 status = __vxge_hw_vpath_stats_access(vpath, \
50 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_wait_receive_idle() local
54 vpath = &hldev->virtual_paths[vp_id]; in vxge_hw_vpath_wait_receive_idle()
55 vp_reg = vpath->vp_reg; in vxge_hw_vpath_wait_receive_idle()
157 vxge_hw_vpath_fw_api(struct __vxge_hw_virtualpath *vpath, u32 action, in vxge_hw_vpath_fw_api() argument
161 struct vxge_hw_vpath_reg __iomem *vp_reg = vpath->vp_reg; in vxge_hw_vpath_fw_api()
166 spin_lock(&vpath->lock); in vxge_hw_vpath_fw_api()
167 if (!vpath->vp_open) { in vxge_hw_vpath_fw_api()
168 spin_unlock(&vpath->lock); in vxge_hw_vpath_fw_api()
196 if (!vpath->vp_open) in vxge_hw_vpath_fw_api()
216 if (vpath->vp_open) in vxge_hw_vpath_fw_api()
217 spin_unlock(&vpath->lock); in vxge_hw_vpath_fw_api()
226 struct __vxge_hw_virtualpath *vpath; in vxge_hw_upgrade_read_version() local
229 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_upgrade_read_version()
231 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_upgrade_read_version()
249 struct __vxge_hw_virtualpath *vpath; in vxge_hw_flash_fw() local
253 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_flash_fw()
255 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_flash_fw()
280 struct __vxge_hw_virtualpath *vpath; in vxge_update_fw_image() local
284 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_update_fw_image()
287 status = vxge_hw_vpath_fw_api(vpath, in vxge_update_fw_image()
306 status = vxge_hw_vpath_fw_api(vpath, in vxge_update_fw_image()
377 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_eprom_img_ver_get() local
381 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_vpath_eprom_img_ver_get()
387 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_vpath_eprom_img_ver_get()
426 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_channel_initialize() local
428 vpath = channel->vph->vpath; in __vxge_hw_channel_initialize()
437 vpath->fifoh = (struct __vxge_hw_fifo *)channel; in __vxge_hw_channel_initialize()
442 vpath->ringh = (struct __vxge_hw_ring *)channel; in __vxge_hw_channel_initialize()
804 __vxge_hw_vpath_fw_ver_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_fw_ver_get() argument
814 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_fw_ver_get()
871 __vxge_hw_vpath_card_info_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_card_info_get() argument
883 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
896 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
911 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_card_info_get()
930 __vxge_hw_vpath_pci_func_mode_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_pci_func_mode_get() argument
938 status = vxge_hw_vpath_fw_api(vpath, in __vxge_hw_vpath_pci_func_mode_get()
954 __vxge_hw_vpath_addr_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_addr_get() argument
963 status = vxge_hw_vpath_fw_api(vpath, action, in __vxge_hw_vpath_addr_get()
1006 struct __vxge_hw_virtualpath vpath; in vxge_hw_device_hw_info_get() local
1054 spin_lock_init(&vpath.lock); in vxge_hw_device_hw_info_get()
1055 vpath.vp_reg = bar0 + val64; in vxge_hw_device_hw_info_get()
1056 vpath.vp_open = VXGE_HW_VP_NOT_OPEN; in vxge_hw_device_hw_info_get()
1058 status = __vxge_hw_vpath_pci_func_mode_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1062 status = __vxge_hw_vpath_fw_ver_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1066 status = __vxge_hw_vpath_card_info_get(&vpath, hw_info); in vxge_hw_device_hw_info_get()
1078 vpath.vp_reg = bar0 + val64; in vxge_hw_device_hw_info_get()
1079 vpath.vp_open = VXGE_HW_VP_NOT_OPEN; in vxge_hw_device_hw_info_get()
1081 status = __vxge_hw_vpath_addr_get(&vpath, in vxge_hw_device_hw_info_get()
1409 __vxge_hw_vpath_stats_access(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_stats_access() argument
1416 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_stats_access()
1421 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_stats_access()
1430 vpath->hldev->config.device_poll_millis); in __vxge_hw_vpath_stats_access()
1444 __vxge_hw_vpath_xmac_tx_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_xmac_tx_stats_get() argument
1454 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_xmac_tx_stats_get()
1460 status = __vxge_hw_vpath_stats_access(vpath, in __vxge_hw_vpath_xmac_tx_stats_get()
1476 __vxge_hw_vpath_xmac_rx_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_xmac_rx_stats_get() argument
1485 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_xmac_rx_stats_get()
1490 status = __vxge_hw_vpath_stats_access(vpath, in __vxge_hw_vpath_xmac_rx_stats_get()
1507 __vxge_hw_vpath_stats_get(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_stats_get() argument
1514 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in __vxge_hw_vpath_stats_get()
1518 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_stats_get()
1578 status = __vxge_hw_vpath_xmac_tx_stats_get(vpath, &hw_stats->tx_stats); in __vxge_hw_vpath_stats_get()
1582 status = __vxge_hw_vpath_xmac_rx_stats_get(vpath, &hw_stats->rx_stats); in __vxge_hw_vpath_stats_get()
2190 hldev = vph->vpath->hldev; in __vxge_hw_channel_allocate()
2191 vp_id = vph->vpath->vp_id; in __vxge_hw_channel_allocate()
2775 struct __vxge_hw_ring *ring = vp->vpath->ringh; in __vxge_hw_ring_delete()
2782 vp->vpath->ringh = NULL; in __vxge_hw_ring_delete()
2811 hldev = vp->vpath->hldev; in __vxge_hw_ring_create()
2812 vp_id = vp->vpath->vp_id; in __vxge_hw_ring_create()
2829 vp->vpath->ringh = ring; in __vxge_hw_ring_create()
2831 ring->vp_reg = vp->vpath->vp_reg; in __vxge_hw_ring_create()
2833 ring->stats = &vp->vpath->sw_stats->ring_stats; in __vxge_hw_ring_create()
2839 ring->tim_rti_cfg1_saved = vp->vpath->tim_rti_cfg1_saved; in __vxge_hw_ring_create()
2840 ring->tim_rti_cfg3_saved = vp->vpath->tim_rti_cfg3_saved; in __vxge_hw_ring_create()
3379 struct __vxge_hw_fifo *fifo = vp->vpath->fifoh; in __vxge_hw_fifo_delete()
3386 vp->vpath->fifoh = NULL; in __vxge_hw_fifo_delete()
3448 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_fifo_create() local
3454 vpath = vp->vpath; in __vxge_hw_fifo_create()
3455 config = &vpath->hldev->config.vp_config[vpath->vp_id].fifo; in __vxge_hw_fifo_create()
3471 vpath->fifoh = fifo; in __vxge_hw_fifo_create()
3472 fifo->nofl_db = vpath->nofl_db; in __vxge_hw_fifo_create()
3474 fifo->vp_id = vpath->vp_id; in __vxge_hw_fifo_create()
3475 fifo->vp_reg = vpath->vp_reg; in __vxge_hw_fifo_create()
3476 fifo->stats = &vpath->sw_stats->fifo_stats; in __vxge_hw_fifo_create()
3482 fifo->tim_tti_cfg1_saved = vpath->tim_tti_cfg1_saved; in __vxge_hw_fifo_create()
3483 fifo->tim_tti_cfg3_saved = vpath->tim_tti_cfg3_saved; in __vxge_hw_fifo_create()
3533 __vxge_hw_mempool_create(vpath->hldev, in __vxge_hw_fifo_create()
3565 __vxge_hw_vpath_pci_read(struct __vxge_hw_virtualpath *vpath, in __vxge_hw_vpath_pci_read() argument
3570 struct vxge_hw_vpath_reg __iomem *vp_reg = vpath->vp_reg; in __vxge_hw_vpath_pci_read()
3611 struct __vxge_hw_virtualpath *vpath; in vxge_hw_device_flick_link_led() local
3620 vpath = &hldev->virtual_paths[hldev->first_vp_id]; in vxge_hw_device_flick_link_led()
3623 status = vxge_hw_vpath_fw_api(vpath, in vxge_hw_device_flick_link_led()
3658 status = vxge_hw_vpath_fw_api(vp->vpath, action, rts_table, offset, in __vxge_hw_vpath_rts_table_get()
3694 status = vxge_hw_vpath_fw_api(vp->vpath, action, rts_table, offset, in __vxge_hw_vpath_rts_table_set()
3821 if (vp->vpath->hldev->config.rth_it_type in vxge_hw_vpath_rts_rth_itable_set()
3964 struct __vxge_hw_virtualpath *vpath) in __vxge_hw_vpath_mgmt_read() argument
3971 val64 = readq(&vpath->vpmgmt_reg-> in __vxge_hw_vpath_mgmt_read()
3981 vpath->max_mtu = mtu + VXGE_HW_MAC_HEADER_MAX_SIZE; in __vxge_hw_vpath_mgmt_read()
3983 val64 = readq(&vpath->vpmgmt_reg->xmac_vsport_choices_vp); in __vxge_hw_vpath_mgmt_read()
3987 vpath->vsport_number = i; in __vxge_hw_vpath_mgmt_read()
3990 val64 = readq(&vpath->vpmgmt_reg->xgmac_gen_status_vpmgmt_clone); in __vxge_hw_vpath_mgmt_read()
3993 VXGE_HW_DEVICE_LINK_STATE_SET(vpath->hldev, VXGE_HW_LINK_UP); in __vxge_hw_vpath_mgmt_read()
3995 VXGE_HW_DEVICE_LINK_STATE_SET(vpath->hldev, VXGE_HW_LINK_DOWN); in __vxge_hw_vpath_mgmt_read()
4006 __vxge_hw_vpath_reset_check(struct __vxge_hw_virtualpath *vpath) in __vxge_hw_vpath_reset_check() argument
4011 &vpath->hldev->common_reg->vpath_rst_in_prog, in __vxge_hw_vpath_reset_check()
4013 1 << (16 - vpath->vp_id)), in __vxge_hw_vpath_reset_check()
4014 vpath->hldev->config.device_poll_millis); in __vxge_hw_vpath_reset_check()
4044 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_sw_reset() local
4046 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_sw_reset()
4048 if (vpath->ringh) { in __vxge_hw_vpath_sw_reset()
4049 status = __vxge_hw_ring_reset(vpath->ringh); in __vxge_hw_vpath_sw_reset()
4054 if (vpath->fifoh) in __vxge_hw_vpath_sw_reset()
4055 status = __vxge_hw_fifo_reset(vpath->fifoh); in __vxge_hw_vpath_sw_reset()
4069 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_prc_configure() local
4073 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_prc_configure()
4074 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_prc_configure()
4075 vp_config = vpath->vp_config; in __vxge_hw_vpath_prc_configure()
4084 val64 = readq(&vpath->vp_reg->prc_cfg6); in __vxge_hw_vpath_prc_configure()
4086 writeq(val64, &vpath->vp_reg->prc_cfg6); in __vxge_hw_vpath_prc_configure()
4090 if (vpath->vp_config->ring.scatter_mode != in __vxge_hw_vpath_prc_configure()
4095 switch (vpath->vp_config->ring.scatter_mode) { in __vxge_hw_vpath_prc_configure()
4115 vpath->ringh) >> 3), &vp_reg->prc_cfg5); in __vxge_hw_vpath_prc_configure()
4143 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_kdfc_configure() local
4146 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_kdfc_configure()
4147 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_kdfc_configure()
4155 vpath->max_kdfc_db = in __vxge_hw_vpath_kdfc_configure()
4159 if (vpath->vp_config->fifo.enable == VXGE_HW_FIFO_ENABLE) { in __vxge_hw_vpath_kdfc_configure()
4161 vpath->max_nofl_db = vpath->max_kdfc_db; in __vxge_hw_vpath_kdfc_configure()
4163 if (vpath->max_nofl_db < in __vxge_hw_vpath_kdfc_configure()
4164 ((vpath->vp_config->fifo.memblock_size / in __vxge_hw_vpath_kdfc_configure()
4165 (vpath->vp_config->fifo.max_frags * in __vxge_hw_vpath_kdfc_configure()
4167 vpath->vp_config->fifo.fifo_blocks)) { in __vxge_hw_vpath_kdfc_configure()
4172 (vpath->max_nofl_db*2)-1); in __vxge_hw_vpath_kdfc_configure()
4197 vpath->nofl_db = in __vxge_hw_vpath_kdfc_configure()
4214 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_mac_configure() local
4218 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_mac_configure()
4219 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_mac_configure()
4220 vp_config = vpath->vp_config; in __vxge_hw_vpath_mac_configure()
4223 vpath->vsport_number), &vp_reg->xmac_vsport_choice); in __vxge_hw_vpath_mac_configure()
4244 VXGE_HW_MAC_HEADER_MAX_SIZE) < vpath->max_mtu) in __vxge_hw_vpath_mac_configure()
4250 vpath->max_mtu); in __vxge_hw_vpath_mac_configure()
4281 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_tim_configure() local
4285 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_tim_configure()
4286 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_tim_configure()
4287 config = vpath->vp_config; in __vxge_hw_vpath_tim_configure()
4349 vpath->tim_tti_cfg1_saved = val64; in __vxge_hw_vpath_tim_configure()
4407 vpath->tim_tti_cfg3_saved = val64; in __vxge_hw_vpath_tim_configure()
4456 vpath->tim_rti_cfg1_saved = val64; in __vxge_hw_vpath_tim_configure()
4514 vpath->tim_rti_cfg3_saved = val64; in __vxge_hw_vpath_tim_configure()
4544 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vpath_initialize() local
4547 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vpath_initialize()
4553 vp_reg = vpath->vp_reg; in __vxge_hw_vpath_initialize()
4555 status = __vxge_hw_vpath_swapper_set(vpath->vp_reg); in __vxge_hw_vpath_initialize()
4574 status = __vxge_hw_vpath_pci_read(vpath, 1, 0x78, &val32); in __vxge_hw_vpath_initialize()
4603 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vp_terminate() local
4605 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vp_terminate()
4607 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) in __vxge_hw_vp_terminate()
4610 VXGE_HW_DEVICE_TIM_INT_MASK_RESET(vpath->hldev->tim_int_mask0, in __vxge_hw_vp_terminate()
4611 vpath->hldev->tim_int_mask1, vpath->vp_id); in __vxge_hw_vp_terminate()
4612 hldev->stats.hw_dev_info_stats.vpath_info[vpath->vp_id] = NULL; in __vxge_hw_vp_terminate()
4617 spin_lock(&vpath->lock); in __vxge_hw_vp_terminate()
4618 vpath->vp_open = VXGE_HW_VP_NOT_OPEN; in __vxge_hw_vp_terminate()
4619 spin_unlock(&vpath->lock); in __vxge_hw_vp_terminate()
4621 vpath->vpmgmt_reg = NULL; in __vxge_hw_vp_terminate()
4622 vpath->nofl_db = NULL; in __vxge_hw_vp_terminate()
4623 vpath->max_mtu = 0; in __vxge_hw_vp_terminate()
4624 vpath->vsport_number = 0; in __vxge_hw_vp_terminate()
4625 vpath->max_kdfc_db = 0; in __vxge_hw_vp_terminate()
4626 vpath->max_nofl_db = 0; in __vxge_hw_vp_terminate()
4627 vpath->ringh = NULL; in __vxge_hw_vp_terminate()
4628 vpath->fifoh = NULL; in __vxge_hw_vp_terminate()
4629 memset(&vpath->vpath_handles, 0, sizeof(struct list_head)); in __vxge_hw_vp_terminate()
4630 vpath->stats_block = NULL; in __vxge_hw_vp_terminate()
4631 vpath->hw_stats = NULL; in __vxge_hw_vp_terminate()
4632 vpath->hw_stats_sav = NULL; in __vxge_hw_vp_terminate()
4633 vpath->sw_stats = NULL; in __vxge_hw_vp_terminate()
4648 struct __vxge_hw_virtualpath *vpath; in __vxge_hw_vp_initialize() local
4656 vpath = &hldev->virtual_paths[vp_id]; in __vxge_hw_vp_initialize()
4658 spin_lock_init(&vpath->lock); in __vxge_hw_vp_initialize()
4659 vpath->vp_id = vp_id; in __vxge_hw_vp_initialize()
4660 vpath->vp_open = VXGE_HW_VP_OPEN; in __vxge_hw_vp_initialize()
4661 vpath->hldev = hldev; in __vxge_hw_vp_initialize()
4662 vpath->vp_config = config; in __vxge_hw_vp_initialize()
4663 vpath->vp_reg = hldev->vpath_reg[vp_id]; in __vxge_hw_vp_initialize()
4664 vpath->vpmgmt_reg = hldev->vpmgmt_reg[vp_id]; in __vxge_hw_vp_initialize()
4668 status = __vxge_hw_vpath_reset_check(vpath); in __vxge_hw_vp_initialize()
4670 memset(vpath, 0, sizeof(struct __vxge_hw_virtualpath)); in __vxge_hw_vp_initialize()
4674 status = __vxge_hw_vpath_mgmt_read(hldev, vpath); in __vxge_hw_vp_initialize()
4676 memset(vpath, 0, sizeof(struct __vxge_hw_virtualpath)); in __vxge_hw_vp_initialize()
4680 INIT_LIST_HEAD(&vpath->vpath_handles); in __vxge_hw_vp_initialize()
4682 vpath->sw_stats = &hldev->stats.sw_dev_info_stats.vpath_info[vp_id]; in __vxge_hw_vp_initialize()
4704 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_mtu_set() local
4710 vpath = vp->vpath; in vxge_hw_vpath_mtu_set()
4714 if ((new_mtu < VXGE_HW_MIN_MTU) || (new_mtu > vpath->max_mtu)) in vxge_hw_vpath_mtu_set()
4717 val64 = readq(&vpath->vp_reg->rxmac_vcfg0); in vxge_hw_vpath_mtu_set()
4722 writeq(val64, &vpath->vp_reg->rxmac_vcfg0); in vxge_hw_vpath_mtu_set()
4724 vpath->vp_config->mtu = new_mtu - VXGE_HW_MAC_HEADER_MAX_SIZE; in vxge_hw_vpath_mtu_set()
4739 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_stats_enable() local
4741 vpath = vp->vpath; in vxge_hw_vpath_stats_enable()
4743 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_stats_enable()
4748 memcpy(vpath->hw_stats_sav, vpath->hw_stats, in vxge_hw_vpath_stats_enable()
4751 status = __vxge_hw_vpath_stats_get(vpath, vpath->hw_stats); in vxge_hw_vpath_stats_enable()
4799 struct __vxge_hw_virtualpath *vpath; in vxge_hw_vpath_open() local
4803 vpath = &hldev->virtual_paths[attr->vp_id]; in vxge_hw_vpath_open()
4805 if (vpath->vp_open == VXGE_HW_VP_OPEN) { in vxge_hw_vpath_open()
4821 vp->vpath = vpath; in vxge_hw_vpath_open()
4823 if (vpath->vp_config->fifo.enable == VXGE_HW_FIFO_ENABLE) { in vxge_hw_vpath_open()
4829 if (vpath->vp_config->ring.enable == VXGE_HW_RING_ENABLE) { in vxge_hw_vpath_open()
4837 vpath->fifoh->tx_intr_num = in vxge_hw_vpath_open()
4841 vpath->stats_block = __vxge_hw_blockpool_block_allocate(hldev, in vxge_hw_vpath_open()
4843 if (vpath->stats_block == NULL) { in vxge_hw_vpath_open()
4848 vpath->hw_stats = vpath->stats_block->memblock; in vxge_hw_vpath_open()
4849 memset(vpath->hw_stats, 0, in vxge_hw_vpath_open()
4853 vpath->hw_stats; in vxge_hw_vpath_open()
4855 vpath->hw_stats_sav = in vxge_hw_vpath_open()
4857 memset(vpath->hw_stats_sav, 0, in vxge_hw_vpath_open()
4860 writeq(vpath->stats_block->dma_addr, &vpath->vp_reg->stats_cfg); in vxge_hw_vpath_open()
4866 list_add(&vp->item, &vpath->vpath_handles); in vxge_hw_vpath_open()
4868 hldev->vpaths_deployed |= vxge_mBIT(vpath->vp_id); in vxge_hw_vpath_open()
4872 attr->fifo_attr.userdata = vpath->fifoh; in vxge_hw_vpath_open()
4873 attr->ring_attr.userdata = vpath->ringh; in vxge_hw_vpath_open()
4878 if (vpath->ringh != NULL) in vxge_hw_vpath_open()
4881 if (vpath->fifoh != NULL) in vxge_hw_vpath_open()
4902 struct __vxge_hw_virtualpath *vpath = vp->vpath; in vxge_hw_vpath_rx_doorbell_init() local
4903 struct __vxge_hw_ring *ring = vpath->ringh; in vxge_hw_vpath_rx_doorbell_init()
4904 struct vxgedev *vdev = netdev_priv(vpath->hldev->ndev); in vxge_hw_vpath_rx_doorbell_init()
4908 new_count = readq(&vpath->vp_reg->rxdmem_size); in vxge_hw_vpath_rx_doorbell_init()
4916 &vpath->vp_reg->prc_rxd_doorbell); in vxge_hw_vpath_rx_doorbell_init()
4917 readl(&vpath->vp_reg->prc_rxd_doorbell); in vxge_hw_vpath_rx_doorbell_init()
4920 val64 = readq(&vpath->vp_reg->prc_cfg6); in vxge_hw_vpath_rx_doorbell_init()
4965 struct __vxge_hw_virtualpath *vpath = NULL; in vxge_hw_vpath_close() local
4967 u32 vp_id = vp->vpath->vp_id; in vxge_hw_vpath_close()
4971 vpath = vp->vpath; in vxge_hw_vpath_close()
4972 devh = vpath->hldev; in vxge_hw_vpath_close()
4974 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_close()
4981 if (!list_empty(&vpath->vpath_handles)) { in vxge_hw_vpath_close()
4982 list_add(&vp->item, &vpath->vpath_handles); in vxge_hw_vpath_close()
4993 if (vpath->ringh != NULL) in vxge_hw_vpath_close()
4996 if (vpath->fifoh != NULL) in vxge_hw_vpath_close()
4999 if (vpath->stats_block != NULL) in vxge_hw_vpath_close()
5000 __vxge_hw_blockpool_block_free(devh, vpath->stats_block); in vxge_hw_vpath_close()
5018 struct __vxge_hw_virtualpath *vpath = vp->vpath; in vxge_hw_vpath_reset() local
5020 vp_id = vpath->vp_id; in vxge_hw_vpath_reset()
5022 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_reset()
5027 status = __vxge_hw_vpath_reset(vpath->hldev, vp_id); in vxge_hw_vpath_reset()
5029 vpath->sw_stats->soft_reset_cnt++; in vxge_hw_vpath_reset()
5042 struct __vxge_hw_virtualpath *vpath = NULL; in vxge_hw_vpath_recover_from_reset() local
5047 vp_id = vp->vpath->vp_id; in vxge_hw_vpath_recover_from_reset()
5048 vpath = vp->vpath; in vxge_hw_vpath_recover_from_reset()
5049 hldev = vpath->hldev; in vxge_hw_vpath_recover_from_reset()
5051 if (vpath->vp_open == VXGE_HW_VP_NOT_OPEN) { in vxge_hw_vpath_recover_from_reset()
5056 status = __vxge_hw_vpath_reset_check(vpath); in vxge_hw_vpath_recover_from_reset()
5068 if (vpath->ringh != NULL) in vxge_hw_vpath_recover_from_reset()
5071 memset(vpath->hw_stats, 0, in vxge_hw_vpath_recover_from_reset()
5074 memset(vpath->hw_stats_sav, 0, in vxge_hw_vpath_recover_from_reset()
5077 writeq(vpath->stats_block->dma_addr, in vxge_hw_vpath_recover_from_reset()
5078 &vpath->vp_reg->stats_cfg); in vxge_hw_vpath_recover_from_reset()
5097 hldev = vp->vpath->hldev; in vxge_hw_vpath_enable()
5100 1 << (16 - vp->vpath->vp_id)); in vxge_hw_vpath_enable()