/Linux-v6.1/arch/arm64/kvm/hyp/nvhe/ |
D | page_alloc.c | 25 * Order 2 1 0 28 * __find_buddy_nocheck(pool, page 0, order 0) => page 1 29 * __find_buddy_nocheck(pool, page 0, order 1) => page 2 30 * __find_buddy_nocheck(pool, page 1, order 0) => page 0 31 * __find_buddy_nocheck(pool, page 2, order 0) => page 3 35 unsigned short order) in __find_buddy_nocheck() argument 39 addr ^= (PAGE_SIZE << order); in __find_buddy_nocheck() 54 unsigned short order) in __find_buddy_avail() argument 56 struct hyp_page *buddy = __find_buddy_nocheck(pool, p, order); in __find_buddy_avail() 58 if (!buddy || buddy->order != order || buddy->refcount) in __find_buddy_avail() [all …]
|
/Linux-v6.1/include/trace/events/ |
D | compaction.h | 157 int order, 161 TP_ARGS(order, gfp_mask, prio), 164 __field(int, order) 170 __entry->order = order; 175 TP_printk("order=%d gfp_mask=%s priority=%d", 176 __entry->order, 184 int order, 187 TP_ARGS(zone, order, ret), 192 __field(int, order) 199 __entry->order = order; [all …]
|
D | vmscan.h | 68 TP_PROTO(int nid, int zid, int order), 70 TP_ARGS(nid, zid, order), 75 __field( int, order ) 81 __entry->order = order; 84 TP_printk("nid=%d order=%d", 86 __entry->order) 91 TP_PROTO(int nid, int zid, int order, gfp_t gfp_flags), 93 TP_ARGS(nid, zid, order, gfp_flags), 98 __field( int, order ) 105 __entry->order = order; [all …]
|
D | kmem.h | 138 TP_PROTO(struct page *page, unsigned int order), 140 TP_ARGS(page, order), 144 __field( unsigned int, order ) 149 __entry->order = order; 152 TP_printk("page=%p pfn=0x%lx order=%d", 155 __entry->order) 172 TP_printk("page=%p pfn=0x%lx order=0", 179 TP_PROTO(struct page *page, unsigned int order, 182 TP_ARGS(page, order, gfp_flags, migratetype), 186 __field( unsigned int, order ) [all …]
|
D | oom.h | 35 int order, 42 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 47 __field( int, order) 58 __entry->order = order; 66 …TP_printk("node=%d zone=%-8s order=%d reclaimable=%lu available=%lu min_wmark=%lu no_progress_loop… 68 __entry->order, 157 TP_PROTO(int order, 164 TP_ARGS(order, priority, result, retries, max_retries, ret), 167 __field( int, order) 176 __entry->order = order; [all …]
|
/Linux-v6.1/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_buddy.c | 31 /* Allocating max_order bitmaps, one for each order */ in mlx5dr_buddy_init() 41 /* In the beginning, we have only one order that is available for in mlx5dr_buddy_init() 77 unsigned int *order) in dr_buddy_find_free_seg() argument 90 "ICM Buddy: failed finding free mem for order %d\n", in dr_buddy_find_free_seg() 101 *order = order_iter; in dr_buddy_find_free_seg() 108 * @order: Order of the buddy to update. 112 * It uses the data structures of the buddy system in order to find the first 113 * area of free place, starting from the current order till the maximum order 122 unsigned int order, in mlx5dr_buddy_alloc_mem() argument 128 err = dr_buddy_find_free_seg(buddy, order, &seg, &order_iter); in mlx5dr_buddy_alloc_mem() [all …]
|
/Linux-v6.1/drivers/gpu/drm/tests/ |
D | drm_buddy_test.c | 21 static inline u64 get_size(int order, u64 chunk_size) in get_size() argument 23 return (1 << order) * chunk_size; in get_size() 49 …kunit_err(test, "block info: header=%llx, state=%u, order=%d, offset=%llx size=%llx root=%d buddy=… in __dump_block() 227 unsigned int order; in check_mm() local 243 order = drm_buddy_block_order(root); in check_mm() 246 if (order != mm->max_order) { in check_mm() 247 kunit_err(test, "max order root missing\n"); in check_mm() 267 block = list_first_entry_or_null(&mm->free_list[order], in check_mm() 270 kunit_err(test, "root mismatch at order=%u\n", order); in check_mm() 330 int order, top; in drm_test_buddy_alloc_pathological() local [all …]
|
/Linux-v6.1/drivers/gpu/drm/ttm/ |
D | ttm_pool.c | 53 * @vaddr: original vaddr return for the mapping and order in the lower bits 77 /* Allocate pages of size 1 << order with the given gfp_flags */ 79 unsigned int order) in ttm_pool_alloc_page() argument 86 /* Don't set the __GFP_COMP flag for higher order allocations. in ttm_pool_alloc_page() 90 if (order) in ttm_pool_alloc_page() 95 p = alloc_pages(gfp_flags, order); in ttm_pool_alloc_page() 97 p->private = order; in ttm_pool_alloc_page() 105 if (order) in ttm_pool_alloc_page() 108 vaddr = dma_alloc_attrs(pool->dev, (1ULL << order) * PAGE_SIZE, in ttm_pool_alloc_page() 121 dma->vaddr = (unsigned long)vaddr | order; in ttm_pool_alloc_page() [all …]
|
/Linux-v6.1/kernel/bpf/ |
D | cgroup_iter.c | 13 * 1. Walk the descendants of a cgroup in pre-order. 14 * 2. Walk the descendants of a cgroup in post-order. 18 * For walking descendants, cgroup_iter can walk in either pre-order or 19 * post-order. For walking ancestors, the iter walks up from a cgroup to 40 * EOPNOTSUPP. In order to work around, the user may have to update their 54 int order; member 77 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_start() 79 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_start() 110 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_next() 112 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_next() [all …]
|
/Linux-v6.1/arch/arm/lib/ |
D | lib1funcs.S | 106 .macro ARM_DIV2_ORDER divisor, order 110 clz \order, \divisor 111 rsb \order, \order, #31 117 movhs \order, #16 118 movlo \order, #0 122 addhs \order, \order, #8 126 addhs \order, \order, #4 129 addhi \order, \order, #3 130 addls \order, \order, \divisor, lsr #1 137 .macro ARM_MOD_BODY dividend, divisor, order, spare [all …]
|
/Linux-v6.1/mm/kmsan/ |
D | init.c | 105 * by their order: when kmsan_memblock_free_pages() is called for the first 106 * time with a certain order, it is reserved as a shadow block, for the second 109 * after which held_back[order] can be used again. 114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) in kmsan_memblock_free_pages() argument 118 if (!held_back[order].shadow) { in kmsan_memblock_free_pages() 119 held_back[order].shadow = page; in kmsan_memblock_free_pages() 122 if (!held_back[order].origin) { in kmsan_memblock_free_pages() 123 held_back[order].origin = page; in kmsan_memblock_free_pages() 126 shadow = held_back[order].shadow; in kmsan_memblock_free_pages() 127 origin = held_back[order].origin; in kmsan_memblock_free_pages() [all …]
|
/Linux-v6.1/lib/ |
D | test_xarray.c | 72 unsigned order, void *entry, gfp_t gfp) in xa_store_order() argument 74 XA_STATE_ORDER(xas, xa, index, order); in xa_store_order() 177 unsigned int order; in check_xa_mark_1() local 207 for (order = 2; order < max_order; order++) { in check_xa_mark_1() 208 unsigned long base = round_down(index, 1UL << order); in check_xa_mark_1() 209 unsigned long next = base + (1UL << order); in check_xa_mark_1() 217 xa_store_order(xa, index, order, xa_mk_index(index), in check_xa_mark_1() 328 unsigned int order; in check_xa_shrink() local 353 for (order = 0; order < max_order; order++) { in check_xa_shrink() 354 unsigned long max = (1UL << order) - 1; in check_xa_shrink() [all …]
|
/Linux-v6.1/arch/riscv/kvm/ |
D | tlb.c | 23 unsigned long order) in kvm_riscv_local_hfence_gvma_vmid_gpa() argument 27 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_vmid_gpa() 34 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa() 39 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa() 51 unsigned long order) in kvm_riscv_local_hfence_gvma_gpa() argument 55 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_gpa() 62 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa() 67 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa() 82 unsigned long order) in kvm_riscv_local_hfence_vvma_asid_gva() argument 86 if (PTRS_PER_PTE < (gvsz >> order)) { in kvm_riscv_local_hfence_vvma_asid_gva() [all …]
|
/Linux-v6.1/tools/testing/radix-tree/ |
D | multiorder.c | 3 * multiorder.c: Multi-order radix tree entry testing 16 unsigned order) in item_insert_order() argument 18 XA_STATE_ORDER(xas, xa, index, order); in item_insert_order() 19 struct item *item = item_create(index, order); in item_insert_order() 42 int order[NUM_ENTRIES] = {1, 1, 2, 3, 4, 1, 0, 1, 3, 0, 7}; in multiorder_iteration() local 47 err = item_insert_order(xa, index[i], order[i]); in multiorder_iteration() 53 if (j <= (index[i] | ((1 << order[i]) - 1))) in multiorder_iteration() 58 int height = order[i] / XA_CHUNK_SHIFT; in multiorder_iteration() 60 unsigned long mask = (1UL << order[i]) - 1; in multiorder_iteration() 66 assert(item->order == order[i]); in multiorder_iteration() [all …]
|
/Linux-v6.1/drivers/media/pci/cx18/ |
D | cx18-mailbox.c | 230 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument 239 mb = &order->mb; in epu_dma_done() 246 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done() 252 mdl_ack = order->mdl_ack; in epu_dma_done() 276 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done() 323 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument 326 char *str = order->str; in epu_debug() 328 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug() 334 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument 336 switch (order->rpu) { in epu_cmd() [all …]
|
/Linux-v6.1/drivers/gpu/drm/lib/ |
D | drm_random.c | 16 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument 24 swap(order[i], order[j]); in drm_random_reorder() 31 unsigned int *order, i; in drm_random_order() local 33 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order() 34 if (!order) in drm_random_order() 35 return order; in drm_random_order() 38 order[i] = i; in drm_random_order() 40 drm_random_reorder(order, count, state); in drm_random_order() 41 return order; in drm_random_order()
|
/Linux-v6.1/mm/ |
D | compaction.c | 53 #define block_start_pfn(pfn, order) round_down(pfn, 1UL << (order)) argument 54 #define block_end_pfn(pfn, order) ALIGN((pfn) + 1, 1UL << (order)) argument 57 * Page order with-respect-to which proactive compaction 87 unsigned int i, order, nr_pages; in split_map_pages() local 94 order = page_private(page); in split_map_pages() 95 nr_pages = 1 << order; in split_map_pages() 97 post_alloc_hook(page, order, __GFP_MOVABLE); in split_map_pages() 98 if (order) in split_map_pages() 99 split_page(page, order); in split_map_pages() 154 static void defer_compaction(struct zone *zone, int order) in defer_compaction() argument [all …]
|
D | page_alloc.c | 314 * change gfp_allowed_mask in order to avoid using I/O during memory allocations 353 static void __free_pages_ok(struct page *page, unsigned int order, 726 static inline unsigned int order_to_pindex(int migratetype, int order) in order_to_pindex() argument 728 int base = order; in order_to_pindex() 731 if (order > PAGE_ALLOC_COSTLY_ORDER) { in order_to_pindex() 732 VM_BUG_ON(order != pageblock_order); in order_to_pindex() 736 VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDER); in order_to_pindex() 744 int order = pindex / MIGRATE_PCPTYPES; in pindex_to_order() local 748 order = pageblock_order; in pindex_to_order() 750 VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDER); in pindex_to_order() [all …]
|
/Linux-v6.1/Documentation/userspace-api/media/v4l/ |
D | field-order.rst | 3 .. _field-order: 6 Field Order 25 which field of a frame is older, the *temporal order*. 30 even) fields, the *spatial order*: The first line of the top field is 39 creating a natural order. 44 and spatial order of fields. The diagrams below should make this 48 bus in the same order they were captured, so if the top field was 53 order. Some drivers may permit the selection of a different order, to 76 - Applications request this field order when any field format 78 e.g. the requested image size, and return the actual field order. [all …]
|
/Linux-v6.1/tools/testing/selftests/tc-testing/tc-tests/actions/ |
D | mpls.json | 20 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*pipe.*index 8 ref", 44 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*pass.*index 8 ref", 68 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*drop.*index 8 ref", 92 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*reclassify.*index 8 ref", 116 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*continue.*index 8 ref", 140 "matchPattern": "action order [0-9]+: mpls.*jump 10.*index 8 ref", 164 "matchPattern": "action order [0-9]+: mpls.*dec_ttl trap.*index 8 ref", 188 "matchPattern": "action order [0-9]+: mpls.*dec_ttl pipe.*index 8 ref.*cookie aabbccddeeff", 212 …"matchPattern": "action order [0-9]+: mpls.*dec_ttl continue.*index 8 ref.*cookie aa11bb22cc33dd44… 236 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*foo.*index 8 ref", [all …]
|
/Linux-v6.1/include/linux/ |
D | gfp.h | 63 * The zone fallback order is MOVABLE=>HIGHMEM=>NORMAL=>DMA32=>DMA. 171 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument 174 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument 177 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid, 179 struct folio *__folio_alloc(gfp_t gfp, unsigned int order, int preferred_nid, 191 /* Bulk allocate order-0 pages */ 232 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument 237 return __alloc_pages(gfp_mask, order, nid, NULL); in __alloc_pages_node() 241 struct folio *__folio_alloc_node(gfp_t gfp, unsigned int order, int nid) in __folio_alloc_node() argument 246 return __folio_alloc(gfp, order, nid, NULL); in __folio_alloc_node() [all …]
|
D | compaction.h | 61 * Number of free order-0 pages that should be available above given watermark 65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument 69 * free scanner may have up to 1 << order pages on its list and then in compact_gap() 70 * try to split an (order - 1) free page. At that point, a gap of in compact_gap() 71 * 1 << order might not be enough, so it's safer to require twice that in compact_gap() 80 return 2UL << order; in compact_gap() 92 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order); 93 extern int fragmentation_index(struct zone *zone, unsigned int order); 95 unsigned int order, unsigned int alloc_flags, 99 extern enum compact_result compaction_suitable(struct zone *zone, int order, [all …]
|
/Linux-v6.1/scripts/atomic/ |
D | gen-atomic-fallback.sh | 8 #gen_template_fallback(template, meta, pfx, name, sfx, order, atomic, int, args...) 16 local order="$1"; shift 20 local atomicname="arch_${atomic}_${pfx}${name}${sfx}${order}" 35 #gen_proto_fallback(meta, pfx, name, sfx, order, atomic, int, args...) 42 local order="$1"; shift 44 local tmpl="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")" 45 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@" 65 local order="$1"; shift 70 printf "#define ${basename}${order} ${basename}${order}\n" 84 local template="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")" [all …]
|
D | gen-atomic-instrumented.sh | 37 local order="$1"; shift 39 if [ "${order}" = "_release" ]; then 41 elif [ -z "${order}" ] && ! meta_in "$meta" "slv"; then 52 #gen_proto_order_variant(meta, pfx, name, sfx, order, atomic, int, arg...) 59 local order="$1"; shift 63 local atomicname="${atomic}_${pfx}${name}${sfx}${order}" 67 local checks="$(gen_params_checks "${meta}" "${order}" "$@")" 86 local order="$1"; shift 91 case "$order" in 100 #define ${xchg}${order}(ptr, oldp, ...) \\ [all …]
|
/Linux-v6.1/Documentation/trace/postprocess/ |
D | trace-vmscan-postprocess.pl | 31 # Per-order events 110 my $regex_direct_begin_default = 'order=([0-9]*) may_writepage=([0-9]*) gfp_flags=([A-Z_|]*)'; 112 my $regex_kswapd_wake_default = 'nid=([0-9]*) order=([0-9]*)'; 114 my $regex_wakeup_kswapd_default = 'nid=([0-9]*) zid=([0-9]*) order=([0-9]*) gfp_flags=([A-Z_|]*)'; 115 my $regex_lru_isolate_default = 'isolate_mode=([0-9]*) classzone_idx=([0-9]*) order=([0-9]*) nr_req… 166 # Verify fields are in the right order 187 "order", "may_writepage", 196 "nid", "order"); 204 "nid", "zid", "order", "gfp_flags"); 208 "isolate_mode", "classzone_idx", "order", [all …]
|