Lines Matching refs:buff_size
366 skb = netdev_alloc_skb(adapter->netdev, pool->buff_size); in replenish_rx_pool()
379 offset = index * pool->buff_size; in replenish_rx_pool()
381 memset(dst, 0, pool->buff_size); in replenish_rx_pool()
389 pool->rx_buff[index].size = pool->buff_size; in replenish_rx_pool()
407 sub_crq->rx_add.len = cpu_to_be32(pool->buff_size << shift); in replenish_rx_pool()
529 u64 buff_size; in reset_rx_pools() local
536 buff_size = adapter->cur_rx_buf_sz; in reset_rx_pools()
543 if (rx_pool->buff_size != buff_size) { in reset_rx_pools()
545 rx_pool->buff_size = ALIGN(buff_size, L1_CACHE_BYTES); in reset_rx_pools()
549 rx_pool->buff_size); in reset_rx_pools()
613 u64 buff_size; in init_rx_pools() local
617 buff_size = adapter->cur_rx_buf_sz; in init_rx_pools()
635 buff_size); in init_rx_pools()
639 rx_pool->buff_size = ALIGN(buff_size, L1_CACHE_BYTES); in init_rx_pools()
659 rx_pool->size * rx_pool->buff_size)) { in init_rx_pools()
793 u64 buff_size; in init_tx_pools() local
813 buff_size = adapter->req_mtu + VLAN_HLEN; in init_tx_pools()
814 buff_size = ALIGN(buff_size, L1_CACHE_BYTES); in init_tx_pools()
817 buff_size); in init_tx_pools()
5727 IOMMU_PAGE_ALIGN(adapter->rx_pool[i].buff_size, tbl); in ibmvnic_get_desired_dma()