Home
last modified time | relevance | path

Searched refs:order (Results 1 – 25 of 1420) sorted by relevance

12345678910>>...57

/Linux-v4.19/include/trace/events/
Dcompaction.h174 int order,
178 TP_ARGS(order, gfp_mask, prio),
181 __field(int, order)
187 __entry->order = order;
193 __entry->order,
202 int order,
205 TP_ARGS(zone, order, ret),
210 __field(int, order)
217 __entry->order = order;
224 __entry->order,
[all …]
Dkmem.h153 TP_PROTO(struct page *page, unsigned int order),
155 TP_ARGS(page, order),
159 __field( unsigned int, order )
164 __entry->order = order;
170 __entry->order)
194 TP_PROTO(struct page *page, unsigned int order,
197 TP_ARGS(page, order, gfp_flags, migratetype),
201 __field( unsigned int, order )
208 __entry->order = order;
216 __entry->order,
[all …]
Dvmscan.h60 TP_PROTO(int nid, int zid, int order),
62 TP_ARGS(nid, zid, order),
67 __field( int, order )
73 __entry->order = order;
76 TP_printk("nid=%d zid=%d order=%d", __entry->nid, __entry->zid, __entry->order)
81 TP_PROTO(int nid, int zid, int order, gfp_t gfp_flags),
83 TP_ARGS(nid, zid, order, gfp_flags),
88 __field( int, order )
95 __entry->order = order;
102 __entry->order,
[all …]
Doom.h35 int order,
42 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check),
47 __field( int, order)
58 __entry->order = order;
68 __entry->order,
157 TP_PROTO(int order,
164 TP_ARGS(order, priority, result, retries, max_retries, ret),
167 __field( int, order)
176 __entry->order = order;
185 __entry->order,
/Linux-v4.19/drivers/media/platform/soc_camera/
Dsoc_mediabus.c26 .order = SOC_MBUS_ORDER_LE,
36 .order = SOC_MBUS_ORDER_LE,
46 .order = SOC_MBUS_ORDER_LE,
56 .order = SOC_MBUS_ORDER_LE,
66 .order = SOC_MBUS_ORDER_LE,
76 .order = SOC_MBUS_ORDER_BE,
86 .order = SOC_MBUS_ORDER_LE,
96 .order = SOC_MBUS_ORDER_BE,
106 .order = SOC_MBUS_ORDER_LE,
115 .order = SOC_MBUS_ORDER_LE,
[all …]
/Linux-v4.19/drivers/gpu/drm/lib/
Ddrm_random.c15 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument
23 swap(order[i], order[j]); in drm_random_reorder()
30 unsigned int *order, i; in drm_random_order() local
32 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order()
33 if (!order) in drm_random_order()
34 return order; in drm_random_order()
37 order[i] = i; in drm_random_order()
39 drm_random_reorder(order, count, state); in drm_random_order()
40 return order; in drm_random_order()
/Linux-v4.19/drivers/gpu/drm/i915/selftests/
Di915_random.c44 void i915_random_reorder(unsigned int *order, unsigned int count, in i915_random_reorder() argument
52 swap(order[i], order[j]); in i915_random_reorder()
58 unsigned int *order, i; in i915_random_order() local
60 order = kmalloc_array(count, sizeof(*order), in i915_random_order()
62 if (!order) in i915_random_order()
63 return order; in i915_random_order()
66 order[i] = i; in i915_random_order()
68 i915_random_reorder(order, count, state); in i915_random_order()
69 return order; in i915_random_order()
Di915_timeline.c63 int order, offset; in igt_sync() local
68 for (order = 1; order < 64; order++) { in igt_sync()
69 for (offset = -1; offset <= (order > 1); offset++) { in igt_sync()
70 u64 ctx = BIT_ULL(order) + offset; in igt_sync()
81 for (order = 1; order < 64; order++) { in igt_sync()
82 for (offset = -1; offset <= (order > 1); offset++) { in igt_sync()
83 u64 ctx = BIT_ULL(order) + offset; in igt_sync()
110 int order, last_order; in bench_sync() local
227 for (last_order = 1, order = 1; order < 32; in bench_sync()
228 ({ int tmp = last_order; last_order = order; order += tmp; })) { in bench_sync()
[all …]
Di915_syncmap.c274 unsigned int pass, order; in igt_syncmap_join_above() local
296 for (order = 0; order < 64; order += SHIFT) { in igt_syncmap_join_above()
297 u64 context = BIT_ULL(order); in igt_syncmap_join_above()
335 unsigned int step, order, idx; in igt_syncmap_join_below() local
345 for (order = 64 - SHIFT; order > 0; order -= SHIFT) { in igt_syncmap_join_below()
346 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below()
354 context, order, step, sync->height, sync->prefix); in igt_syncmap_join_below()
362 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_join_below()
363 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below()
367 context, order, step); in igt_syncmap_join_below()
[all …]
/Linux-v4.19/include/linux/
Dcompaction.h68 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument
83 return 2UL << order; in compact_gap()
95 extern int fragmentation_index(struct zone *zone, unsigned int order);
97 unsigned int order, unsigned int alloc_flags,
100 extern enum compact_result compaction_suitable(struct zone *zone, int order,
103 extern void defer_compaction(struct zone *zone, int order);
104 extern bool compaction_deferred(struct zone *zone, int order);
105 extern void compaction_defer_reset(struct zone *zone, int order,
107 extern bool compaction_restarting(struct zone *zone, int order);
174 bool compaction_zonelist_suitable(struct alloc_context *ac, int order,
[all …]
Dpage_owner.h11 extern void __reset_page_owner(struct page *page, unsigned int order);
13 unsigned int order, gfp_t gfp_mask);
14 extern void __split_page_owner(struct page *page, unsigned int order);
21 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument
24 __reset_page_owner(page, order); in reset_page_owner()
28 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument
31 __set_page_owner(page, order, gfp_mask); in set_page_owner()
34 static inline void split_page_owner(struct page *page, unsigned int order) in split_page_owner() argument
37 __split_page_owner(page, order); in split_page_owner()
55 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument
[all …]
Dgfp.h460 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument
463 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument
467 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid,
471 __alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid) in __alloc_pages() argument
473 return __alloc_pages_nodemask(gfp_mask, order, preferred_nid, NULL); in __alloc_pages()
481 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument
486 return __alloc_pages(gfp_mask, order, nid); in __alloc_pages_node()
495 unsigned int order) in alloc_pages_node() argument
500 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node()
504 extern struct page *alloc_pages_current(gfp_t gfp_mask, unsigned order);
[all …]
/Linux-v4.19/mm/
Dcompaction.c47 #define block_start_pfn(pfn, order) round_down(pfn, 1UL << (order)) argument
48 #define block_end_pfn(pfn, order) ALIGN((pfn) + 1, 1UL << (order)) argument
70 unsigned int i, order, nr_pages; in map_pages() local
77 order = page_private(page); in map_pages()
78 nr_pages = 1 << order; in map_pages()
80 post_alloc_hook(page, order, __GFP_MOVABLE); in map_pages()
81 if (order) in map_pages()
82 split_page(page, order); in map_pages()
141 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
146 if (order < zone->compact_order_failed) in defer_compaction()
[all …]
Dpage_alloc.c194 static void __free_pages_ok(struct page *page, unsigned int order);
569 void prep_compound_page(struct page *page, unsigned int order) in prep_compound_page() argument
572 int nr_pages = 1 << order; in prep_compound_page()
575 set_compound_order(page, order); in prep_compound_page()
644 unsigned int order, int migratetype) in set_page_guard() argument
651 if (order >= debug_guardpage_minorder()) in set_page_guard()
661 set_page_private(page, order); in set_page_guard()
663 __mod_zone_freepage_state(zone, -(1 << order), migratetype); in set_page_guard()
669 unsigned int order, int migratetype) in clear_page_guard() argument
684 __mod_zone_freepage_state(zone, (1 << order), migratetype); in clear_page_guard()
[all …]
/Linux-v4.19/drivers/media/pci/cx18/
Dcx18-mailbox.c239 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument
248 mb = &order->mb; in epu_dma_done()
255 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done()
261 mdl_ack = order->mdl_ack; in epu_dma_done()
285 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done()
332 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument
335 char *str = order->str; in epu_debug()
337 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug()
343 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument
345 switch (order->rpu) { in epu_cmd()
[all …]
/Linux-v4.19/Documentation/trace/postprocess/
Dtrace-vmscan-postprocess.pl316 my $order = $1;
317 $perprocesspid{$process_pid}->{MM_VMSCAN_DIRECT_RECLAIM_BEGIN_PERORDER}[$order]++;
318 $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER} = $order;
327 my $order = $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER};
329 $perprocesspid{$process_pid}->{HIGH_DIRECT_RECLAIM_LATENCY}[$index] = "$order-$latency";
340 my $order = $2;
341 $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER} = $order;
346 $perprocesspid{$process_pid}->{MM_VMSCAN_KSWAPD_WAKE_PERORDER}[$order]++;
349 $perprocesspid{$process_pid}->{HIGH_KSWAPD_REWAKEUP_PERORDER}[$order]++;
359 my $order = $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER};
[all …]
/Linux-v4.19/arch/s390/mm/
Dpage-states.c77 static inline void set_page_unused(struct page *page, int order) in set_page_unused() argument
81 for (i = 0; i < (1 << order); i++) in set_page_unused()
88 static inline void set_page_stable_dat(struct page *page, int order) in set_page_stable_dat() argument
92 for (i = 0; i < (1 << order); i++) in set_page_stable_dat()
99 static inline void set_page_stable_nodat(struct page *page, int order) in set_page_stable_nodat() argument
103 for (i = 0; i < (1 << order); i++) in set_page_stable_nodat()
216 void arch_free_page(struct page *page, int order) in arch_free_page() argument
220 set_page_unused(page, order); in arch_free_page()
223 void arch_alloc_page(struct page *page, int order) in arch_alloc_page() argument
228 set_page_stable_dat(page, order); in arch_alloc_page()
[all …]
/Linux-v4.19/tools/testing/radix-tree/
Dmultiorder.c23 #define for_each_index(i, base, order) \ argument
24 for (i = base; i < base + (1 << order); i++)
26 static void __multiorder_tag_test(int index, int order) in __multiorder_tag_test() argument
32 base = index & ~((1 << order) - 1); in __multiorder_tag_test()
37 err = item_insert_order(&tree, index, order); in __multiorder_tag_test()
45 for_each_index(i, base, order) { in __multiorder_tag_test()
46 err = __radix_tree_insert(&tree, i, order, in __multiorder_tag_test()
51 for_each_index(i, base, order) { in __multiorder_tag_test()
58 for_each_index(i, base, order) { in __multiorder_tag_test()
66 for_each_index(i, base, order) { in __multiorder_tag_test()
[all …]
Dbenchmark.c20 #define for_each_index(i, base, order) \ argument
21 for (i = base; i < base + (1 << order); i++)
64 unsigned long size, unsigned long step, int order) in benchmark_insert() argument
73 item_insert_order(root, index, order); in benchmark_insert()
81 size, step, order, nsec); in benchmark_insert()
85 unsigned long size, unsigned long step, int order) in benchmark_tagging() argument
102 size, step, order, nsec); in benchmark_tagging()
106 unsigned long size, unsigned long step, int order) in benchmark_delete() argument
115 for_each_index(i, index, order) in benchmark_delete()
124 size, step, order, nsec); in benchmark_delete()
[all …]
/Linux-v4.19/drivers/gpu/drm/nouveau/nvkm/subdev/therm/
Dgk104.c34 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_enable() local
38 for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable()
39 if (!nvkm_device_subdev(dev, order[i].engine)) in gk104_clkgate_enable()
42 nvkm_mask(dev, 0x20200 + order[i].offset, 0xff00, 0x4500); in gk104_clkgate_enable()
50 for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable()
51 if (!nvkm_device_subdev(dev, order[i].engine)) in gk104_clkgate_enable()
54 nvkm_mask(dev, 0x20200 + order[i].offset, 0x00ff, 0x0045); in gk104_clkgate_enable()
63 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_fini() local
67 for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_fini()
68 if (!nvkm_device_subdev(dev, order[i].engine)) in gk104_clkgate_fini()
[all …]
/Linux-v4.19/arch/arm/lib/
Dlib1funcs.S106 .macro ARM_DIV2_ORDER divisor, order
110 clz \order, \divisor
111 rsb \order, \order, #31
117 movhs \order, #16
118 movlo \order, #0
122 addhs \order, \order, #8
126 addhs \order, \order, #4
129 addhi \order, \order, #3
130 addls \order, \order, \divisor, lsr #1
137 .macro ARM_MOD_BODY dividend, divisor, order, spare
[all …]
/Linux-v4.19/arch/c6x/mm/
Ddma-coherent.c47 static inline u32 __alloc_dma_pages(int order) in __alloc_dma_pages() argument
53 pos = bitmap_find_free_region(dma_bitmap, dma_pages, order); in __alloc_dma_pages()
59 static void __free_dma_pages(u32 addr, int order) in __free_dma_pages() argument
64 if (addr < dma_base || (pos + (1 << order)) >= dma_pages) { in __free_dma_pages()
70 bitmap_release_region(dma_bitmap, pos, order); in __free_dma_pages()
82 int order; in arch_dma_alloc() local
87 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); in arch_dma_alloc()
89 paddr = __alloc_dma_pages(order); in arch_dma_alloc()
106 int order; in arch_dma_free() local
111 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); in arch_dma_free()
[all …]
/Linux-v4.19/drivers/pci/endpoint/
Dpci-epc-mem.c25 int order; in pci_epc_mem_get_order() local
31 order = fls(size); in pci_epc_mem_get_order()
33 order = fls64(size); in pci_epc_mem_get_order()
35 return order; in pci_epc_mem_get_order()
128 int order; in pci_epc_mem_alloc_addr() local
131 order = pci_epc_mem_get_order(mem, size); in pci_epc_mem_alloc_addr()
133 pageno = bitmap_find_free_region(mem->bitmap, mem->pages, order); in pci_epc_mem_alloc_addr()
140 bitmap_release_region(mem->bitmap, pageno, order); in pci_epc_mem_alloc_addr()
161 int order; in pci_epc_mem_free_addr() local
166 order = pci_epc_mem_get_order(mem, size); in pci_epc_mem_free_addr()
[all …]
/Linux-v4.19/drivers/staging/android/ion/
Dion_page_pool.c16 return alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages()
22 __free_pages(page, pool->order); in ion_page_pool_free_pages()
37 (1 << (PAGE_SHIFT + pool->order))); in ion_page_pool_add()
57 -(1 << (PAGE_SHIFT + pool->order))); in ion_page_pool_remove()
82 BUG_ON(pool->order != compound_order(page)); in ion_page_pool_free()
94 return count << pool->order; in ion_page_pool_total()
125 freed += (1 << pool->order); in ion_page_pool_shrink()
131 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order) in ion_page_pool_create() argument
142 pool->order = order; in ion_page_pool_create()
144 plist_node_init(&pool->list, order); in ion_page_pool_create()
/Linux-v4.19/arch/m68k/kernel/
Ddma.c28 int i, order; in arch_dma_alloc() local
33 order = get_order(size); in arch_dma_alloc()
35 page = alloc_pages(flag, order); in arch_dma_alloc()
40 map = kmalloc(sizeof(struct page *) << order, flag & ~__GFP_DMA); in arch_dma_alloc()
42 __free_pages(page, order); in arch_dma_alloc()
45 split_page(page, order); in arch_dma_alloc()
47 order = 1 << order; in arch_dma_alloc()
52 for (; i < order; i++) in arch_dma_alloc()

12345678910>>...57