Home
last modified time | relevance | path

Searched refs:page_pool (Results 1 – 25 of 28) sorted by relevance

12

/Linux-v5.4/include/net/
Dpage_pool.h70 struct page_pool { struct
112 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp); argument
114 static inline struct page *page_pool_dev_alloc_pages(struct page_pool *pool) in page_pool_dev_alloc_pages()
125 inline enum dma_data_direction page_pool_get_dma_dir(struct page_pool *pool) in page_pool_get_dma_dir()
130 struct page_pool *page_pool_create(const struct page_pool_params *params);
132 void __page_pool_free(struct page_pool *pool);
133 static inline void page_pool_free(struct page_pool *pool) in page_pool_free()
144 static inline void page_pool_destroy(struct page_pool *pool) in page_pool_destroy()
153 void __page_pool_put_page(struct page_pool *pool,
156 static inline void page_pool_put_page(struct page_pool *pool, in page_pool_put_page()
[all …]
Dxdp_priv.h12 struct page_pool *page_pool; member
Dxdp.h53 struct page_pool;
/Linux-v5.4/net/core/
Dpage_pool.c21 static int page_pool_init(struct page_pool *pool, in page_pool_init()
61 struct page_pool *page_pool_create(const struct page_pool_params *params) in page_pool_create()
63 struct page_pool *pool; in page_pool_create()
82 static struct page *__page_pool_get_cached(struct page_pool *pool) in __page_pool_get_cached()
117 static struct page *__page_pool_alloc_pages_slow(struct page_pool *pool, in __page_pool_alloc_pages_slow()
172 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp) in page_pool_alloc_pages()
192 static s32 page_pool_inflight(struct page_pool *pool) in page_pool_inflight()
204 static bool __page_pool_safe_to_destroy(struct page_pool *pool) in __page_pool_safe_to_destroy()
215 static void __page_pool_clean_page(struct page_pool *pool, in __page_pool_clean_page()
236 void page_pool_unmap_page(struct page_pool *pool, struct page *page) in page_pool_unmap_page()
[all …]
Dxdp.c75 page_pool_free(xa->page_pool); in __xdp_mem_allocator_rcu_free()
105 safe_to_remove = page_pool_request_shutdown(xa->page_pool); in __mem_id_disconnect()
374 page_pool_get(xdp_alloc->page_pool); in xdp_rxq_info_reg_mem_model()
407 page_pool_put_page(xa->page_pool, page, napi_direct); in __xdp_return()
464 page_pool_release_page(xa->page_pool, page); in __xdp_release_frame()
DMakefile17 obj-$(CONFIG_PAGE_POOL) += page_pool.o
/Linux-v5.4/include/trace/events/
Dpage_pool.h3 #define TRACE_SYSTEM page_pool
15 TP_PROTO(const struct page_pool *pool,
21 __field(const struct page_pool *, pool)
40 TP_PROTO(const struct page_pool *pool,
46 __field(const struct page_pool *, pool)
63 TP_PROTO(const struct page_pool *pool,
69 __field(const struct page_pool *, pool)
/Linux-v5.4/drivers/net/ethernet/apm/xgene/
Dxgene_enet_main.c672 struct xgene_enet_desc_ring *buf_pool, *page_pool; in xgene_enet_rx_frame() local
690 page_pool = rx_ring->page_pool; in xgene_enet_rx_frame()
713 xgene_enet_free_pagepool(page_pool, raw_desc, exp_desc); in xgene_enet_rx_frame()
727 slots = page_pool->slots - 1; in xgene_enet_rx_frame()
728 head = page_pool->head; in xgene_enet_rx_frame()
739 page = page_pool->frag_page[head]; in xgene_enet_rx_frame()
745 page_pool->frag_page[head] = NULL; in xgene_enet_rx_frame()
749 page_pool->head = head; in xgene_enet_rx_frame()
762 ret = xgene_enet_refill_pagepool(page_pool, NUM_NXTBUFPOOL); in xgene_enet_rx_frame()
1058 struct xgene_enet_desc_ring *buf_pool, *page_pool; in xgene_enet_delete_desc_rings() local
[all …]
Dxgene_enet_cle.c709 if (pdata->rx_ring[idx]->page_pool) { in xgene_cle_set_rss_idt()
710 pool_id = pdata->rx_ring[idx]->page_pool->id; in xgene_cle_set_rss_idt()
786 if (pdata->rx_ring[0]->page_pool) { in xgene_enet_cle_init()
787 pool_id = pdata->rx_ring[0]->page_pool->id; in xgene_enet_cle_init()
Dxgene_enet_main.h120 struct xgene_enet_desc_ring *page_pool; member
/Linux-v5.4/block/
Dbounce.c32 static mempool_t page_pool, isa_page_pool; variable
61 ret = mempool_init_page_pool(&page_pool, POOL_SIZE, 0); in init_emergency_pool()
188 bounce_end_io(bio, &page_pool); in bounce_end_io_write()
209 __bounce_end_io_read(bio, &page_pool); in bounce_end_io_read()
346 if (pool == &page_pool) { in __blk_queue_bounce()
378 pool = &page_pool; in blk_queue_bounce()
/Linux-v5.4/mm/
Dreadahead.c162 LIST_HEAD(page_pool); in __do_page_cache_readahead()
190 read_pages(mapping, filp, &page_pool, nr_pages, in __do_page_cache_readahead()
200 list_add(&page->lru, &page_pool); in __do_page_cache_readahead()
212 read_pages(mapping, filp, &page_pool, nr_pages, gfp_mask); in __do_page_cache_readahead()
213 BUG_ON(!list_empty(&page_pool)); in __do_page_cache_readahead()
/Linux-v5.4/drivers/net/ethernet/socionext/
Dnetsec.c285 struct page_pool *page_pool; member
725 page = page_pool_dev_alloc_pages(dring->page_pool); in netsec_alloc_rx_data()
738 dma_dir = page_pool_get_dma_dir(dring->page_pool); in netsec_alloc_rx_data()
848 page_pool_get_dma_dir(rx_ring->page_pool); in netsec_xdp_queue_one()
938 dma_dir = page_pool_get_dma_dir(dring->page_pool); in netsec_process_rx()
1016 page_pool_recycle_direct(dring->page_pool, page); in netsec_process_rx()
1021 page_pool_release_page(dring->page_pool, page); in netsec_process_rx()
1194 page_pool_put_page(dring->page_pool, page, false); in netsec_uninit_pkt_dring()
1206 page_pool_destroy(dring->page_pool); in netsec_uninit_pkt_dring()
1285 dring->page_pool = page_pool_create(&pp_params); in netsec_setup_rx_dring()
[all …]
/Linux-v5.4/drivers/net/ethernet/stmicro/stmmac/
Dstmmac.h69 struct page_pool *page_pool; member
Dstmmac_main.c1208 buf->page = page_pool_dev_alloc_pages(rx_q->page_pool); in stmmac_init_rx_buffers()
1213 buf->sec_page = page_pool_dev_alloc_pages(rx_q->page_pool); in stmmac_init_rx_buffers()
1243 page_pool_put_page(rx_q->page_pool, buf->page, false); in stmmac_free_rx_buffer()
1247 page_pool_put_page(rx_q->page_pool, buf->sec_page, false); in stmmac_free_rx_buffer()
1505 if (rx_q->page_pool) { in free_dma_rx_desc_resources()
1506 page_pool_request_shutdown(rx_q->page_pool); in free_dma_rx_desc_resources()
1507 page_pool_destroy(rx_q->page_pool); in free_dma_rx_desc_resources()
1574 rx_q->page_pool = page_pool_create(&pp_params); in alloc_dma_rx_desc_resources()
1575 if (IS_ERR(rx_q->page_pool)) { in alloc_dma_rx_desc_resources()
1576 ret = PTR_ERR(rx_q->page_pool); in alloc_dma_rx_desc_resources()
[all …]
/Linux-v5.4/drivers/net/ethernet/ti/
Dcpsw.c543 page_pool_recycle_direct(cpsw->page_pool[ch], page); in cpsw_run_xdp()
555 static struct page_pool *cpsw_create_page_pool(struct cpsw_common *cpsw, in cpsw_create_page_pool()
559 struct page_pool *pool; in cpsw_create_page_pool()
579 struct page_pool *pool; in cpsw_ndev_create_xdp_rxq()
582 pool = cpsw->page_pool[ch]; in cpsw_ndev_create_xdp_rxq()
608 struct page_pool *pool; in cpsw_create_rx_pool()
616 cpsw->page_pool[ch] = pool; in cpsw_create_rx_pool()
635 page_pool_destroy(cpsw->page_pool[ch]); in cpsw_destroy_xdp_rxqs()
636 cpsw->page_pool[ch] = NULL; in cpsw_destroy_xdp_rxqs()
683 struct page_pool *pool; in cpsw_rx_handler()
[all …]
Dcpsw_priv.h350 struct page_pool *page_pool[CPSW_MAX_QUEUES]; member
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/
Den.h60 struct page_pool;
680 struct page_pool *page_pool; member
Den_rx.c246 dma_info->page = page_pool_dev_alloc_pages(rq->page_pool); in mlx5e_page_alloc_pool()
253 page_pool_recycle_direct(rq->page_pool, dma_info->page); in mlx5e_page_alloc_pool()
284 page_pool_recycle_direct(rq->page_pool, dma_info->page); in mlx5e_page_release_dynamic()
287 page_pool_release_page(rq->page_pool, dma_info->page); in mlx5e_page_release_dynamic()
Den_main.c560 rq->page_pool = page_pool_create(&pp_params); in mlx5e_alloc_rq()
561 if (IS_ERR(rq->page_pool)) { in mlx5e_alloc_rq()
562 err = PTR_ERR(rq->page_pool); in mlx5e_alloc_rq()
563 rq->page_pool = NULL; in mlx5e_alloc_rq()
567 MEM_TYPE_PAGE_POOL, rq->page_pool); in mlx5e_alloc_rq()
635 page_pool_destroy(rq->page_pool); in mlx5e_alloc_rq()
670 page_pool_destroy(rq->page_pool); in mlx5e_free_rq()
/Linux-v5.4/drivers/net/ethernet/broadcom/bnxt/
Dbnxt_xdp.c195 page_pool_recycle_direct(rxr->page_pool, page); in bnxt_rx_xdp()
Dbnxt.h29 struct page_pool;
899 struct page_pool *page_pool; member
Dbnxt.c692 page = page_pool_dev_alloc_pages(rxr->page_pool); in __bnxt_alloc_rx_page()
699 page_pool_recycle_direct(rxr->page_pool, page); in __bnxt_alloc_rx_page()
2584 page_pool_recycle_direct(rxr->page_pool, data); in bnxt_free_rx_skbs()
2780 page_pool_destroy(rxr->page_pool); in bnxt_free_rx_rings()
2781 rxr->page_pool = NULL; in bnxt_free_rx_rings()
2804 rxr->page_pool = page_pool_create(&pp); in bnxt_alloc_rx_page_pool()
2805 if (IS_ERR(rxr->page_pool)) { in bnxt_alloc_rx_page_pool()
2806 int err = PTR_ERR(rxr->page_pool); in bnxt_alloc_rx_page_pool()
2808 rxr->page_pool = NULL; in bnxt_alloc_rx_page_pool()
2840 rxr->page_pool); in bnxt_alloc_rx_rings()
/Linux-v5.4/drivers/md/
Ddm-crypt.c202 mempool_t page_pool; member
1312 page = mempool_alloc(&cc->page_pool, gfp_mask); in crypt_alloc_buffer()
1347 mempool_free(bv->bv_page, &cc->page_pool); in crypt_free_buffer_pages()
2129 mempool_exit(&cc->page_pool); in crypt_dtr()
2648 ret = mempool_init(&cc->page_pool, BIO_MAX_PAGES, crypt_page_alloc, crypt_page_free, cc); in crypt_ctr()
/Linux-v5.4/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_cmn.h1026 bnx2x_free_rx_mem_pool(bp, &fp->page_pool); in bnx2x_free_rx_sge_range()

12