Lines Matching refs:q_vector

167 				i40evf_force_wb(vsi, tx_ring->q_vector);  in i40evf_detect_recover_hung()
286 tx_ring->q_vector->tx.total_bytes += total_bytes; in i40e_clean_tx_irq()
287 tx_ring->q_vector->tx.total_packets += total_packets; in i40e_clean_tx_irq()
334 struct i40e_q_vector *q_vector) in i40e_enable_wb_on_itr() argument
336 u16 flags = q_vector->tx.ring[0].flags; in i40e_enable_wb_on_itr()
342 if (q_vector->arm_wb_state) in i40e_enable_wb_on_itr()
349 I40E_VFINT_DYN_CTLN1(q_vector->reg_idx), val); in i40e_enable_wb_on_itr()
350 q_vector->arm_wb_state = true; in i40e_enable_wb_on_itr()
359 void i40evf_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector) in i40evf_force_wb() argument
368 I40E_VFINT_DYN_CTLN1(q_vector->reg_idx), in i40evf_force_wb()
372 static inline bool i40e_container_is_rx(struct i40e_q_vector *q_vector, in i40e_container_is_rx() argument
375 return &q_vector->rx == rc; in i40e_container_is_rx()
378 static inline unsigned int i40e_itr_divisor(struct i40e_q_vector *q_vector) in i40e_itr_divisor() argument
382 switch (q_vector->adapter->link_speed) { in i40e_itr_divisor()
416 static void i40e_update_itr(struct i40e_q_vector *q_vector, in i40e_update_itr() argument
431 itr = i40e_container_is_rx(q_vector, rc) ? in i40e_update_itr()
449 if (q_vector->itr_countdown) { in i40e_update_itr()
457 if (i40e_container_is_rx(q_vector, rc)) { in i40e_update_itr()
464 (q_vector->tx.target_itr & I40E_ITR_ADAPTIVE_LATENCY)) { in i40e_update_itr()
475 (q_vector->rx.target_itr & I40E_ITR_MASK) == in i40e_update_itr()
503 itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); in i40e_update_itr()
590 itr += DIV_ROUND_UP(avg_wire_size, i40e_itr_divisor(q_vector)) * in i40e_update_itr()
864 struct i40e_q_vector *q_vector = rx_ring->q_vector; in i40e_receive_skb() local
870 napi_gro_receive(&q_vector->napi, skb); in i40e_receive_skb()
1310 skb = __napi_alloc_skb(&rx_ring->q_vector->napi, in i40e_construct_skb()
1574 rx_ring->q_vector->rx.total_packets += total_rx_packets; in i40e_clean_rx_irq()
1575 rx_ring->q_vector->rx.total_bytes += total_rx_bytes; in i40e_clean_rx_irq()
1628 struct i40e_q_vector *q_vector) in i40e_update_enable_itr() argument
1634 i40e_update_itr(q_vector, &q_vector->tx); in i40e_update_enable_itr()
1635 i40e_update_itr(q_vector, &q_vector->rx); in i40e_update_enable_itr()
1645 if (q_vector->rx.target_itr < q_vector->rx.current_itr) { in i40e_update_enable_itr()
1648 q_vector->rx.target_itr); in i40e_update_enable_itr()
1649 q_vector->rx.current_itr = q_vector->rx.target_itr; in i40e_update_enable_itr()
1650 q_vector->itr_countdown = ITR_COUNTDOWN_START; in i40e_update_enable_itr()
1651 } else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) || in i40e_update_enable_itr()
1652 ((q_vector->rx.target_itr - q_vector->rx.current_itr) < in i40e_update_enable_itr()
1653 (q_vector->tx.target_itr - q_vector->tx.current_itr))) { in i40e_update_enable_itr()
1658 q_vector->tx.target_itr); in i40e_update_enable_itr()
1659 q_vector->tx.current_itr = q_vector->tx.target_itr; in i40e_update_enable_itr()
1660 q_vector->itr_countdown = ITR_COUNTDOWN_START; in i40e_update_enable_itr()
1661 } else if (q_vector->rx.current_itr != q_vector->rx.target_itr) { in i40e_update_enable_itr()
1664 q_vector->rx.target_itr); in i40e_update_enable_itr()
1665 q_vector->rx.current_itr = q_vector->rx.target_itr; in i40e_update_enable_itr()
1666 q_vector->itr_countdown = ITR_COUNTDOWN_START; in i40e_update_enable_itr()
1670 if (q_vector->itr_countdown) in i40e_update_enable_itr()
1671 q_vector->itr_countdown--; in i40e_update_enable_itr()
1675 wr32(hw, INTREG(q_vector->reg_idx), intval); in i40e_update_enable_itr()
1689 struct i40e_q_vector *q_vector = in i40evf_napi_poll() local
1691 struct i40e_vsi *vsi = q_vector->vsi; in i40evf_napi_poll()
1706 i40e_for_each_ring(ring, q_vector->tx) { in i40evf_napi_poll()
1722 budget_per_ring = max(budget/q_vector->num_ringpairs, 1); in i40evf_napi_poll()
1724 i40e_for_each_ring(ring, q_vector->rx) { in i40evf_napi_poll()
1744 if (!cpumask_test_cpu(cpu_id, &q_vector->affinity_mask)) { in i40evf_napi_poll()
1749 i40evf_force_wb(vsi, q_vector); in i40evf_napi_poll()
1756 q_vector->tx.ring[0].tx_stats.tx_force_wb++; in i40evf_napi_poll()
1757 i40e_enable_wb_on_itr(vsi, q_vector); in i40evf_napi_poll()
1763 q_vector->arm_wb_state = false; in i40evf_napi_poll()
1768 i40e_update_enable_itr(vsi, q_vector); in i40evf_napi_poll()