Lines Matching refs:buff_

505 	struct aq_ring_buff_s *buff_ = buff;  in aq_add_rx_fragment()  local
515 buff_ = &ring->buff_ring[buff_->next]; in aq_add_rx_fragment()
517 buff_->rxdata.daddr, in aq_add_rx_fragment()
518 buff_->rxdata.pg_off, in aq_add_rx_fragment()
519 buff_->len, in aq_add_rx_fragment()
521 skb_frag_off_set(frag, buff_->rxdata.pg_off); in aq_add_rx_fragment()
522 skb_frag_size_set(frag, buff_->len); in aq_add_rx_fragment()
523 sinfo->xdp_frags_size += buff_->len; in aq_add_rx_fragment()
524 __skb_frag_set_page(frag, buff_->rxdata.page); in aq_add_rx_fragment()
526 buff_->is_cleaned = 1; in aq_add_rx_fragment()
528 buff->is_ip_cso &= buff_->is_ip_cso; in aq_add_rx_fragment()
529 buff->is_udp_cso &= buff_->is_udp_cso; in aq_add_rx_fragment()
530 buff->is_tcp_cso &= buff_->is_tcp_cso; in aq_add_rx_fragment()
531 buff->is_cso_err |= buff_->is_cso_err; in aq_add_rx_fragment()
533 if (page_is_pfmemalloc(buff_->rxdata.page)) in aq_add_rx_fragment()
536 } while (!buff_->is_eop); in aq_add_rx_fragment()
554 struct aq_ring_buff_s *buff_ = NULL; in __aq_ring_rx_clean() local
565 buff_ = buff; in __aq_ring_rx_clean()
569 if (buff_->next >= self->size) { in __aq_ring_rx_clean()
575 next_ = buff_->next, in __aq_ring_rx_clean()
576 buff_ = &self->buff_ring[next_]; in __aq_ring_rx_clean()
588 buff->is_error |= buff_->is_error; in __aq_ring_rx_clean()
589 buff->is_cso_err |= buff_->is_cso_err; in __aq_ring_rx_clean()
591 } while (!buff_->is_eop); in __aq_ring_rx_clean()
595 buff_ = buff; in __aq_ring_rx_clean()
597 if (buff_->next >= self->size) { in __aq_ring_rx_clean()
601 next_ = buff_->next, in __aq_ring_rx_clean()
602 buff_ = &self->buff_ring[next_]; in __aq_ring_rx_clean()
604 buff_->is_cleaned = true; in __aq_ring_rx_clean()
605 } while (!buff_->is_eop); in __aq_ring_rx_clean()
658 buff_ = buff; in __aq_ring_rx_clean()
660 next_ = buff_->next; in __aq_ring_rx_clean()
661 buff_ = &self->buff_ring[next_]; in __aq_ring_rx_clean()
664 buff_->rxdata.daddr, in __aq_ring_rx_clean()
665 buff_->rxdata.pg_off, in __aq_ring_rx_clean()
666 buff_->len, in __aq_ring_rx_clean()
669 buff_->rxdata.page, in __aq_ring_rx_clean()
670 buff_->rxdata.pg_off, in __aq_ring_rx_clean()
671 buff_->len, in __aq_ring_rx_clean()
673 page_ref_inc(buff_->rxdata.page); in __aq_ring_rx_clean()
674 buff_->is_cleaned = 1; in __aq_ring_rx_clean()
676 buff->is_ip_cso &= buff_->is_ip_cso; in __aq_ring_rx_clean()
677 buff->is_udp_cso &= buff_->is_udp_cso; in __aq_ring_rx_clean()
678 buff->is_tcp_cso &= buff_->is_tcp_cso; in __aq_ring_rx_clean()
679 buff->is_cso_err |= buff_->is_cso_err; in __aq_ring_rx_clean()
681 } while (!buff_->is_eop); in __aq_ring_rx_clean()
730 struct aq_ring_buff_s *buff_ = NULL; in __aq_ring_xdp_clean() local
740 buff_ = buff; in __aq_ring_xdp_clean()
742 if (buff_->next >= rx_ring->size) { in __aq_ring_xdp_clean()
746 next_ = buff_->next; in __aq_ring_xdp_clean()
747 buff_ = &rx_ring->buff_ring[next_]; in __aq_ring_xdp_clean()
756 buff->is_error |= buff_->is_error; in __aq_ring_xdp_clean()
757 buff->is_cso_err |= buff_->is_cso_err; in __aq_ring_xdp_clean()
758 } while (!buff_->is_eop); in __aq_ring_xdp_clean()
766 buff_ = buff; in __aq_ring_xdp_clean()
768 if (buff_->next >= rx_ring->size) { in __aq_ring_xdp_clean()
772 next_ = buff_->next; in __aq_ring_xdp_clean()
773 buff_ = &rx_ring->buff_ring[next_]; in __aq_ring_xdp_clean()
775 buff_->is_cleaned = true; in __aq_ring_xdp_clean()
776 } while (!buff_->is_eop); in __aq_ring_xdp_clean()