/Linux-v4.19/include/trace/events/ |
D | compaction.h | 174 int order, 178 TP_ARGS(order, gfp_mask, prio), 181 __field(int, order) 187 __entry->order = order; 193 __entry->order, 202 int order, 205 TP_ARGS(zone, order, ret), 210 __field(int, order) 217 __entry->order = order; 224 __entry->order, [all …]
|
D | kmem.h | 153 TP_PROTO(struct page *page, unsigned int order), 155 TP_ARGS(page, order), 159 __field( unsigned int, order ) 164 __entry->order = order; 170 __entry->order) 194 TP_PROTO(struct page *page, unsigned int order, 197 TP_ARGS(page, order, gfp_flags, migratetype), 201 __field( unsigned int, order ) 208 __entry->order = order; 216 __entry->order, [all …]
|
D | vmscan.h | 60 TP_PROTO(int nid, int zid, int order), 62 TP_ARGS(nid, zid, order), 67 __field( int, order ) 73 __entry->order = order; 76 TP_printk("nid=%d zid=%d order=%d", __entry->nid, __entry->zid, __entry->order) 81 TP_PROTO(int nid, int zid, int order, gfp_t gfp_flags), 83 TP_ARGS(nid, zid, order, gfp_flags), 88 __field( int, order ) 95 __entry->order = order; 102 __entry->order, [all …]
|
D | oom.h | 35 int order, 42 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 47 __field( int, order) 58 __entry->order = order; 68 __entry->order, 157 TP_PROTO(int order, 164 TP_ARGS(order, priority, result, retries, max_retries, ret), 167 __field( int, order) 176 __entry->order = order; 185 __entry->order,
|
/Linux-v4.19/drivers/media/platform/soc_camera/ |
D | soc_mediabus.c | 26 .order = SOC_MBUS_ORDER_LE, 36 .order = SOC_MBUS_ORDER_LE, 46 .order = SOC_MBUS_ORDER_LE, 56 .order = SOC_MBUS_ORDER_LE, 66 .order = SOC_MBUS_ORDER_LE, 76 .order = SOC_MBUS_ORDER_BE, 86 .order = SOC_MBUS_ORDER_LE, 96 .order = SOC_MBUS_ORDER_BE, 106 .order = SOC_MBUS_ORDER_LE, 115 .order = SOC_MBUS_ORDER_LE, [all …]
|
/Linux-v4.19/drivers/gpu/drm/lib/ |
D | drm_random.c | 15 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument 23 swap(order[i], order[j]); in drm_random_reorder() 30 unsigned int *order, i; in drm_random_order() local 32 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order() 33 if (!order) in drm_random_order() 34 return order; in drm_random_order() 37 order[i] = i; in drm_random_order() 39 drm_random_reorder(order, count, state); in drm_random_order() 40 return order; in drm_random_order()
|
/Linux-v4.19/drivers/gpu/drm/i915/selftests/ |
D | i915_random.c | 44 void i915_random_reorder(unsigned int *order, unsigned int count, in i915_random_reorder() argument 52 swap(order[i], order[j]); in i915_random_reorder() 58 unsigned int *order, i; in i915_random_order() local 60 order = kmalloc_array(count, sizeof(*order), in i915_random_order() 62 if (!order) in i915_random_order() 63 return order; in i915_random_order() 66 order[i] = i; in i915_random_order() 68 i915_random_reorder(order, count, state); in i915_random_order() 69 return order; in i915_random_order()
|
D | i915_timeline.c | 63 int order, offset; in igt_sync() local 68 for (order = 1; order < 64; order++) { in igt_sync() 69 for (offset = -1; offset <= (order > 1); offset++) { in igt_sync() 70 u64 ctx = BIT_ULL(order) + offset; in igt_sync() 81 for (order = 1; order < 64; order++) { in igt_sync() 82 for (offset = -1; offset <= (order > 1); offset++) { in igt_sync() 83 u64 ctx = BIT_ULL(order) + offset; in igt_sync() 110 int order, last_order; in bench_sync() local 227 for (last_order = 1, order = 1; order < 32; in bench_sync() 228 ({ int tmp = last_order; last_order = order; order += tmp; })) { in bench_sync() [all …]
|
D | i915_syncmap.c | 274 unsigned int pass, order; in igt_syncmap_join_above() local 296 for (order = 0; order < 64; order += SHIFT) { in igt_syncmap_join_above() 297 u64 context = BIT_ULL(order); in igt_syncmap_join_above() 335 unsigned int step, order, idx; in igt_syncmap_join_below() local 345 for (order = 64 - SHIFT; order > 0; order -= SHIFT) { in igt_syncmap_join_below() 346 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below() 354 context, order, step, sync->height, sync->prefix); in igt_syncmap_join_below() 362 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_join_below() 363 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below() 367 context, order, step); in igt_syncmap_join_below() [all …]
|
/Linux-v4.19/include/linux/ |
D | compaction.h | 68 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument 83 return 2UL << order; in compact_gap() 95 extern int fragmentation_index(struct zone *zone, unsigned int order); 97 unsigned int order, unsigned int alloc_flags, 100 extern enum compact_result compaction_suitable(struct zone *zone, int order, 103 extern void defer_compaction(struct zone *zone, int order); 104 extern bool compaction_deferred(struct zone *zone, int order); 105 extern void compaction_defer_reset(struct zone *zone, int order, 107 extern bool compaction_restarting(struct zone *zone, int order); 174 bool compaction_zonelist_suitable(struct alloc_context *ac, int order, [all …]
|
D | page_owner.h | 11 extern void __reset_page_owner(struct page *page, unsigned int order); 13 unsigned int order, gfp_t gfp_mask); 14 extern void __split_page_owner(struct page *page, unsigned int order); 21 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument 24 __reset_page_owner(page, order); in reset_page_owner() 28 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument 31 __set_page_owner(page, order, gfp_mask); in set_page_owner() 34 static inline void split_page_owner(struct page *page, unsigned int order) in split_page_owner() argument 37 __split_page_owner(page, order); in split_page_owner() 55 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument [all …]
|
D | gfp.h | 460 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument 463 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument 467 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, 471 __alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid) in __alloc_pages() argument 473 return __alloc_pages_nodemask(gfp_mask, order, preferred_nid, NULL); in __alloc_pages() 481 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument 486 return __alloc_pages(gfp_mask, order, nid); in __alloc_pages_node() 495 unsigned int order) in alloc_pages_node() argument 500 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node() 504 extern struct page *alloc_pages_current(gfp_t gfp_mask, unsigned order); [all …]
|
/Linux-v4.19/mm/ |
D | compaction.c | 47 #define block_start_pfn(pfn, order) round_down(pfn, 1UL << (order)) argument 48 #define block_end_pfn(pfn, order) ALIGN((pfn) + 1, 1UL << (order)) argument 70 unsigned int i, order, nr_pages; in map_pages() local 77 order = page_private(page); in map_pages() 78 nr_pages = 1 << order; in map_pages() 80 post_alloc_hook(page, order, __GFP_MOVABLE); in map_pages() 81 if (order) in map_pages() 82 split_page(page, order); in map_pages() 141 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 146 if (order < zone->compact_order_failed) in defer_compaction() [all …]
|
D | page_alloc.c | 194 static void __free_pages_ok(struct page *page, unsigned int order); 569 void prep_compound_page(struct page *page, unsigned int order) in prep_compound_page() argument 572 int nr_pages = 1 << order; in prep_compound_page() 575 set_compound_order(page, order); in prep_compound_page() 644 unsigned int order, int migratetype) in set_page_guard() argument 651 if (order >= debug_guardpage_minorder()) in set_page_guard() 661 set_page_private(page, order); in set_page_guard() 663 __mod_zone_freepage_state(zone, -(1 << order), migratetype); in set_page_guard() 669 unsigned int order, int migratetype) in clear_page_guard() argument 684 __mod_zone_freepage_state(zone, (1 << order), migratetype); in clear_page_guard() [all …]
|
/Linux-v4.19/drivers/media/pci/cx18/ |
D | cx18-mailbox.c | 239 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument 248 mb = &order->mb; in epu_dma_done() 255 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done() 261 mdl_ack = order->mdl_ack; in epu_dma_done() 285 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done() 332 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument 335 char *str = order->str; in epu_debug() 337 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug() 343 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument 345 switch (order->rpu) { in epu_cmd() [all …]
|
/Linux-v4.19/Documentation/trace/postprocess/ |
D | trace-vmscan-postprocess.pl | 316 my $order = $1; 317 $perprocesspid{$process_pid}->{MM_VMSCAN_DIRECT_RECLAIM_BEGIN_PERORDER}[$order]++; 318 $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER} = $order; 327 my $order = $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER}; 329 $perprocesspid{$process_pid}->{HIGH_DIRECT_RECLAIM_LATENCY}[$index] = "$order-$latency"; 340 my $order = $2; 341 $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER} = $order; 346 $perprocesspid{$process_pid}->{MM_VMSCAN_KSWAPD_WAKE_PERORDER}[$order]++; 349 $perprocesspid{$process_pid}->{HIGH_KSWAPD_REWAKEUP_PERORDER}[$order]++; 359 my $order = $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER}; [all …]
|
/Linux-v4.19/arch/s390/mm/ |
D | page-states.c | 77 static inline void set_page_unused(struct page *page, int order) in set_page_unused() argument 81 for (i = 0; i < (1 << order); i++) in set_page_unused() 88 static inline void set_page_stable_dat(struct page *page, int order) in set_page_stable_dat() argument 92 for (i = 0; i < (1 << order); i++) in set_page_stable_dat() 99 static inline void set_page_stable_nodat(struct page *page, int order) in set_page_stable_nodat() argument 103 for (i = 0; i < (1 << order); i++) in set_page_stable_nodat() 216 void arch_free_page(struct page *page, int order) in arch_free_page() argument 220 set_page_unused(page, order); in arch_free_page() 223 void arch_alloc_page(struct page *page, int order) in arch_alloc_page() argument 228 set_page_stable_dat(page, order); in arch_alloc_page() [all …]
|
/Linux-v4.19/tools/testing/radix-tree/ |
D | multiorder.c | 23 #define for_each_index(i, base, order) \ argument 24 for (i = base; i < base + (1 << order); i++) 26 static void __multiorder_tag_test(int index, int order) in __multiorder_tag_test() argument 32 base = index & ~((1 << order) - 1); in __multiorder_tag_test() 37 err = item_insert_order(&tree, index, order); in __multiorder_tag_test() 45 for_each_index(i, base, order) { in __multiorder_tag_test() 46 err = __radix_tree_insert(&tree, i, order, in __multiorder_tag_test() 51 for_each_index(i, base, order) { in __multiorder_tag_test() 58 for_each_index(i, base, order) { in __multiorder_tag_test() 66 for_each_index(i, base, order) { in __multiorder_tag_test() [all …]
|
D | benchmark.c | 20 #define for_each_index(i, base, order) \ argument 21 for (i = base; i < base + (1 << order); i++) 64 unsigned long size, unsigned long step, int order) in benchmark_insert() argument 73 item_insert_order(root, index, order); in benchmark_insert() 81 size, step, order, nsec); in benchmark_insert() 85 unsigned long size, unsigned long step, int order) in benchmark_tagging() argument 102 size, step, order, nsec); in benchmark_tagging() 106 unsigned long size, unsigned long step, int order) in benchmark_delete() argument 115 for_each_index(i, index, order) in benchmark_delete() 124 size, step, order, nsec); in benchmark_delete() [all …]
|
/Linux-v4.19/drivers/gpu/drm/nouveau/nvkm/subdev/therm/ |
D | gk104.c | 34 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_enable() local 38 for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable() 39 if (!nvkm_device_subdev(dev, order[i].engine)) in gk104_clkgate_enable() 42 nvkm_mask(dev, 0x20200 + order[i].offset, 0xff00, 0x4500); in gk104_clkgate_enable() 50 for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable() 51 if (!nvkm_device_subdev(dev, order[i].engine)) in gk104_clkgate_enable() 54 nvkm_mask(dev, 0x20200 + order[i].offset, 0x00ff, 0x0045); in gk104_clkgate_enable() 63 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_fini() local 67 for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_fini() 68 if (!nvkm_device_subdev(dev, order[i].engine)) in gk104_clkgate_fini() [all …]
|
/Linux-v4.19/arch/arm/lib/ |
D | lib1funcs.S | 106 .macro ARM_DIV2_ORDER divisor, order 110 clz \order, \divisor 111 rsb \order, \order, #31 117 movhs \order, #16 118 movlo \order, #0 122 addhs \order, \order, #8 126 addhs \order, \order, #4 129 addhi \order, \order, #3 130 addls \order, \order, \divisor, lsr #1 137 .macro ARM_MOD_BODY dividend, divisor, order, spare [all …]
|
/Linux-v4.19/arch/c6x/mm/ |
D | dma-coherent.c | 47 static inline u32 __alloc_dma_pages(int order) in __alloc_dma_pages() argument 53 pos = bitmap_find_free_region(dma_bitmap, dma_pages, order); in __alloc_dma_pages() 59 static void __free_dma_pages(u32 addr, int order) in __free_dma_pages() argument 64 if (addr < dma_base || (pos + (1 << order)) >= dma_pages) { in __free_dma_pages() 70 bitmap_release_region(dma_bitmap, pos, order); in __free_dma_pages() 82 int order; in arch_dma_alloc() local 87 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); in arch_dma_alloc() 89 paddr = __alloc_dma_pages(order); in arch_dma_alloc() 106 int order; in arch_dma_free() local 111 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); in arch_dma_free() [all …]
|
/Linux-v4.19/drivers/pci/endpoint/ |
D | pci-epc-mem.c | 25 int order; in pci_epc_mem_get_order() local 31 order = fls(size); in pci_epc_mem_get_order() 33 order = fls64(size); in pci_epc_mem_get_order() 35 return order; in pci_epc_mem_get_order() 128 int order; in pci_epc_mem_alloc_addr() local 131 order = pci_epc_mem_get_order(mem, size); in pci_epc_mem_alloc_addr() 133 pageno = bitmap_find_free_region(mem->bitmap, mem->pages, order); in pci_epc_mem_alloc_addr() 140 bitmap_release_region(mem->bitmap, pageno, order); in pci_epc_mem_alloc_addr() 161 int order; in pci_epc_mem_free_addr() local 166 order = pci_epc_mem_get_order(mem, size); in pci_epc_mem_free_addr() [all …]
|
/Linux-v4.19/drivers/staging/android/ion/ |
D | ion_page_pool.c | 16 return alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages() 22 __free_pages(page, pool->order); in ion_page_pool_free_pages() 37 (1 << (PAGE_SHIFT + pool->order))); in ion_page_pool_add() 57 -(1 << (PAGE_SHIFT + pool->order))); in ion_page_pool_remove() 82 BUG_ON(pool->order != compound_order(page)); in ion_page_pool_free() 94 return count << pool->order; in ion_page_pool_total() 125 freed += (1 << pool->order); in ion_page_pool_shrink() 131 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order) in ion_page_pool_create() argument 142 pool->order = order; in ion_page_pool_create() 144 plist_node_init(&pool->list, order); in ion_page_pool_create()
|
/Linux-v4.19/arch/m68k/kernel/ |
D | dma.c | 28 int i, order; in arch_dma_alloc() local 33 order = get_order(size); in arch_dma_alloc() 35 page = alloc_pages(flag, order); in arch_dma_alloc() 40 map = kmalloc(sizeof(struct page *) << order, flag & ~__GFP_DMA); in arch_dma_alloc() 42 __free_pages(page, order); in arch_dma_alloc() 45 split_page(page, order); in arch_dma_alloc() 47 order = 1 << order; in arch_dma_alloc() 52 for (; i < order; i++) in arch_dma_alloc()
|