Lines Matching refs:napi

130 	struct napi_struct napi;  member
138 struct napi_struct napi; member
314 static void virtqueue_napi_schedule(struct napi_struct *napi, in virtqueue_napi_schedule() argument
317 if (napi_schedule_prep(napi)) { in virtqueue_napi_schedule()
319 __napi_schedule(napi); in virtqueue_napi_schedule()
323 static void virtqueue_napi_complete(struct napi_struct *napi, in virtqueue_napi_complete() argument
329 if (napi_complete_done(napi, processed)) { in virtqueue_napi_complete()
331 virtqueue_napi_schedule(napi, vq); in virtqueue_napi_complete()
340 struct napi_struct *napi = &vi->sq[vq2txq(vq)].napi; in skb_xmit_done() local
345 if (napi->weight) in skb_xmit_done()
346 virtqueue_napi_schedule(napi, vq); in skb_xmit_done()
384 skb = napi_alloc_skb(&rq->napi, GOOD_COPY_LEN); in page_to_skb()
1061 napi_gro_receive(&rq->napi, skb); in receive_buf()
1247 virtqueue_napi_schedule(&rq->napi, rvq); in skb_recv_done()
1250 static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi) in virtnet_napi_enable() argument
1252 napi_enable(napi); in virtnet_napi_enable()
1259 virtqueue_napi_schedule(napi, vq); in virtnet_napi_enable()
1265 struct napi_struct *napi) in virtnet_napi_tx_enable() argument
1267 if (!napi->weight) in virtnet_napi_tx_enable()
1274 napi->weight = 0; in virtnet_napi_tx_enable()
1278 return virtnet_napi_enable(vq, napi); in virtnet_napi_tx_enable()
1281 static void virtnet_napi_tx_disable(struct napi_struct *napi) in virtnet_napi_tx_disable() argument
1283 if (napi->weight) in virtnet_napi_tx_disable()
1284 napi_disable(napi); in virtnet_napi_tx_disable()
1297 napi_disable(&rq->napi); in refill_work()
1299 virtnet_napi_enable(rq->vq, &rq->napi); in refill_work()
1405 if (!sq->napi.weight || is_xdp_raw_buffer_queue(vi, index)) in virtnet_poll_cleantx()
1417 static int virtnet_poll(struct napi_struct *napi, int budget) in virtnet_poll() argument
1420 container_of(napi, struct receive_queue, napi); in virtnet_poll()
1432 virtqueue_napi_complete(napi, rq->vq, received); in virtnet_poll()
1471 virtnet_napi_enable(vi->rq[i].vq, &vi->rq[i].napi); in virtnet_open()
1472 virtnet_napi_tx_enable(vi, vi->sq[i].vq, &vi->sq[i].napi); in virtnet_open()
1478 static int virtnet_poll_tx(struct napi_struct *napi, int budget) in virtnet_poll_tx() argument
1480 struct send_queue *sq = container_of(napi, struct send_queue, napi); in virtnet_poll_tx()
1487 napi_complete_done(napi, 0); in virtnet_poll_tx()
1496 virtqueue_napi_complete(napi, sq->vq, 0); in virtnet_poll_tx()
1559 bool use_napi = sq->napi.weight; in start_xmit()
1809 napi_disable(&vi->rq[i].napi); in virtnet_close()
1810 virtnet_napi_tx_disable(&vi->sq[i].napi); in virtnet_close()
2245 if (napi_weight ^ vi->sq[0].napi.weight) { in virtnet_set_coalesce()
2249 vi->sq[i].napi.weight = napi_weight; in virtnet_set_coalesce()
2266 if (vi->sq[0].napi.weight) in virtnet_get_coalesce()
2329 napi_disable(&vi->rq[i].napi); in virtnet_freeze_down()
2330 virtnet_napi_tx_disable(&vi->sq[i].napi); in virtnet_freeze_down()
2354 virtnet_napi_enable(vi->rq[i].vq, &vi->rq[i].napi); in virtnet_restore_up()
2356 &vi->sq[i].napi); in virtnet_restore_up()
2457 napi_disable(&vi->rq[i].napi); in virtnet_xdp_set()
2458 virtnet_napi_tx_disable(&vi->sq[i].napi); in virtnet_xdp_set()
2489 virtnet_napi_enable(vi->rq[i].vq, &vi->rq[i].napi); in virtnet_xdp_set()
2491 &vi->sq[i].napi); in virtnet_xdp_set()
2506 virtnet_napi_enable(vi->rq[i].vq, &vi->rq[i].napi); in virtnet_xdp_set()
2508 &vi->sq[i].napi); in virtnet_xdp_set()
2646 napi_hash_del(&vi->rq[i].napi); in virtnet_free_queues()
2647 netif_napi_del(&vi->rq[i].napi); in virtnet_free_queues()
2648 netif_napi_del(&vi->sq[i].napi); in virtnet_free_queues()
2850 netif_napi_add(vi->dev, &vi->rq[i].napi, virtnet_poll, in virtnet_alloc_queues()
2852 netif_tx_napi_add(vi->dev, &vi->sq[i].napi, virtnet_poll_tx, in virtnet_alloc_queues()