/Linux-v4.19/include/trace/events/ |
D | kmem.h | 280 int alloc_order, int fallback_order, 284 alloc_order, fallback_order, 289 __field( int, alloc_order ) 298 __entry->alloc_order = alloc_order; 309 __entry->alloc_order,
|
/Linux-v4.19/Documentation/trace/postprocess/ |
D | trace-pagealloc-postprocess.pl | 242 my ($alloc_order, $fallback_order, $pageblock_order); 254 $alloc_order = $3;
|
/Linux-v4.19/arch/powerpc/sysdev/xive/ |
D | native.c | 221 unsigned int alloc_order; in xive_native_cleanup_queue() local 228 alloc_order = xive_alloc_order(xive_queue_shift); in xive_native_cleanup_queue() 229 free_pages((unsigned long)q->qpage, alloc_order); in xive_native_cleanup_queue()
|
D | spapr.c | 468 unsigned int alloc_order; in xive_spapr_cleanup_queue() local 477 alloc_order = xive_alloc_order(xive_queue_shift); in xive_spapr_cleanup_queue() 478 free_pages((unsigned long)q->qpage, alloc_order); in xive_spapr_cleanup_queue()
|
D | common.c | 1447 unsigned int alloc_order; in xive_queue_page_alloc() local 1451 alloc_order = xive_alloc_order(queue_shift); in xive_queue_page_alloc() 1452 pages = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL, alloc_order); in xive_queue_page_alloc()
|
/Linux-v4.19/drivers/lightnvm/ |
D | pblk-rb.c | 48 unsigned int alloc_order = power_size; in pblk_rb_init() local 64 if (alloc_order >= max_order) { in pblk_rb_init() 66 iter = (1 << (alloc_order - max_order)); in pblk_rb_init() 68 order = alloc_order; in pblk_rb_init()
|
/Linux-v4.19/drivers/dma/ioat/ |
D | dma.h | 139 u16 alloc_order; member 322 return 1 << ioat_chan->alloc_order; in ioat_ring_size()
|
D | init.c | 626 const int total_descs = 1 << ioat_chan->alloc_order; in ioat_free_chan_resources() 669 ioat_chan->alloc_order = 0; in ioat_free_chan_resources() 694 return 1 << ioat_chan->alloc_order; in ioat_alloc_chan_resources() 723 ioat_chan->alloc_order = order; in ioat_alloc_chan_resources() 737 return 1 << ioat_chan->alloc_order; in ioat_alloc_chan_resources()
|
D | sysfs.c | 127 return sprintf(page, "%d\n", (1 << ioat_chan->alloc_order) & ~1); in ring_size_show()
|
/Linux-v4.19/mm/ |
D | vmscan.c | 3636 static void kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_order, in kswapd_try_to_sleep() argument 3667 wakeup_kcompactd(pgdat, alloc_order, classzone_idx); in kswapd_try_to_sleep() 3731 unsigned int alloc_order, reclaim_order; in kswapd() local 3765 alloc_order = reclaim_order = pgdat->kswapd_order; in kswapd() 3769 kswapd_try_to_sleep(pgdat, alloc_order, reclaim_order, in kswapd() 3773 alloc_order = reclaim_order = pgdat->kswapd_order; in kswapd() 3798 alloc_order); in kswapd() 3799 reclaim_order = balance_pgdat(pgdat, alloc_order, classzone_idx); in kswapd() 3800 if (reclaim_order < alloc_order) in kswapd()
|
D | page_alloc.c | 2241 unsigned int alloc_order) in reserve_highatomic_pageblock() argument
|
/Linux-v4.19/drivers/net/ethernet/amd/xgbe/ |
D | xgbe-desc.c | 286 struct xgbe_page_alloc *pa, int alloc_order, in xgbe_alloc_pages() argument 295 order = alloc_order; in xgbe_alloc_pages()
|
/Linux-v4.19/drivers/net/ethernet/brocade/bna/ |
D | bnad.h | 261 int alloc_order; member
|
D | bnad.c | 260 unmap_q->alloc_order = -1; in bnad_rxq_alloc_uninit() 279 unmap_q->alloc_order = 0; in bnad_rxq_alloc_init() 283 unmap_q->alloc_order = 0; in bnad_rxq_alloc_init() 287 unmap_q->alloc_order = order; in bnad_rxq_alloc_init() 360 alloc_size = PAGE_SIZE << unmap_q->alloc_order; in bnad_rxq_refill_page() 368 unmap_q->alloc_order); in bnad_rxq_refill_page()
|
/Linux-v4.19/Documentation/trace/ |
D | events-kmem.rst | 107 …mm_page_alloc_extfrag page=%p pfn=%lu alloc_order=%d fallback_order=%d pageblock_order=%d alloc_m…
|
/Linux-v4.19/arch/powerpc/include/asm/ |
D | opal.h | 268 int64_t opal_xive_alloc_vp_block(uint32_t alloc_order);
|