/Linux-v5.10/tools/testing/selftests/vm/ |
D | charge_reserved_hugetlb.sh | 152 local reserve="$9" 178 "$cgroup" "$path" "$method" "$private" "-l" "$reserve" 2>&1 | tee $output & 200 elif [[ "$reserve" != "-n" ]]; then 211 "$cgroup" "$path" "$method" "$private" "$reserve" 214 if [[ "$reserve" != "-n" ]]; then 270 local reserve="${10}" 288 "$reserve" 318 local reserve="${15}" 342 "$expect_failure" "$reserve" 359 "$expect_failure" "$reserve" [all …]
|
D | write_to_hugetlbfs.c | 75 int reserve = 1; in main() local 114 reserve = 0; in main() 159 if (!reserve) in main() 162 printf("RESERVE mapping.\n"); in main() 174 (reserve ? 0 : MAP_NORESERVE), in main() 188 (reserve ? 0 : MAP_NORESERVE), in main()
|
D | write_hugetlb_memory.sh | 14 reserve=$9 23 "$private" "$want_sleep" "$reserve"
|
/Linux-v5.10/arch/arm/mach-omap2/ |
D | board-generic.c | 56 .reserve = omap_reserve, 73 .reserve = omap_reserve, 119 .reserve = rx51_reserve, 137 .reserve = omap_reserve, 154 .reserve = omap_reserve, 171 .reserve = omap_reserve, 187 .reserve = omap_reserve, 206 .reserve = omap_reserve, 223 .reserve = omap_reserve, 241 .reserve = omap_reserve, [all …]
|
/Linux-v5.10/fs/btrfs/ |
D | block-rsv.c | 16 * currently reserved for this block reserve. 22 * -> Reserve 60 * to make our extent tree updates. This block reserve acts as an overflow 61 * buffer in case our delayed refs reserve does not reserve enough space to 77 * reservation. We use the transaction items and pre-reserve space for every 83 * the most dynamic block reserve in the system, we will try to refill this 84 * block reserve first with any excess returned by any other block reserve. 87 * This is the fallback block reserve to make us try to reserve space if we 90 * content to just reserve space from the space_info on demand. 94 * allocate a block reserve, set it to some size, and then truncate bytes [all …]
|
D | delalloc-space.c | 22 * -> Reserve 78 * necessary, either by attempting to reserve more space, or freeing up excess 88 * -> reserve 144 /* Use new btrfs_qgroup_reserve_data to reserve precious data space. */ in btrfs_check_data_free_space() 325 * over-reserve slightly, and clean up the mess when we are done. in btrfs_delalloc_reserve_metadata() 395 * When we reserve space we increase outstanding_extents for the extents we may 419 * btrfs_delalloc_reserve_space - reserve data and metadata space for 429 * - reserve space in data space info for num bytes 430 * and reserve precious corresponding qgroup space 433 * - reserve space for metadata space, based on the number of outstanding [all …]
|
/Linux-v5.10/drivers/scsi/ |
D | constants.c | 44 "Reserve(6)", 68 "Reserve track", "Send OPC info", "Mode Select(10)", 69 /* 56-5b */ "Reserve(10)", "Release(10)", "Repair track", "Read master cue", 71 /* 5c-5f */ "Read buffer capacity", "Send cue sheet", "Persistent reserve in", 72 "Persistent reserve out", 160 {0x0, "Persistent reserve in, read keys"}, 161 {0x1, "Persistent reserve in, read reservation"}, 162 {0x2, "Persistent reserve in, report capabilities"}, 163 {0x3, "Persistent reserve in, read full status"}, 168 {0x0, "Persistent reserve out, register"}, [all …]
|
/Linux-v5.10/arch/powerpc/platforms/pseries/ |
D | vio.c | 79 * @reserve: pool of memory from which devices reserve entitlement, incl. spare 92 struct vio_cmo_pool reserve; member 152 /* Determine the amount of free entitlement available in reserve */ in vio_cmo_alloc() 182 * the reserve pool is used to reduce device entitlement, the excess 183 * pool is used to increase the reserve pool toward the desired entitlement 209 /* Spare is a subset of the reserve pool, replenish it first. */ in vio_cmo_dealloc() 213 * Replenish the spare in the reserve pool from the excess pool. in vio_cmo_dealloc() 214 * This moves entitlement into the reserve pool. in vio_cmo_dealloc() 219 vio_cmo.reserve.size += tmp; in vio_cmo_dealloc() 227 * Replenish the spare in the reserve pool from the reserve pool. in vio_cmo_dealloc() [all …]
|
/Linux-v5.10/arch/x86/include/asm/trace/ |
D | irq_vectors.h | 271 bool reserve), 273 TP_ARGS(irq, is_managed, can_reserve, reserve), 279 __field( bool, reserve ) 286 __entry->reserve = reserve; 289 TP_printk("irq=%u is_managed=%d can_reserve=%d reserve=%d", 291 __entry->reserve) 297 bool can_reserve, bool reserve), \ 298 TP_ARGS(irq, is_managed, can_reserve, reserve), NULL, NULL); \
|
/Linux-v5.10/fs/xfs/ |
D | xfs_fsops.c | 157 * Reserve AG metadata blocks. ENOSPC here does not mean there was a in xfs_growfs_data_private() 310 * Reserve the requested number of blocks if available. Otherwise return 349 * retry if we end up trying to reserve more space than is available. in xfs_reserve_blocks() 376 * If the request is larger than the current reservation, reserve the in xfs_reserve_blocks() 377 * blocks before we update the reserve counters. Sample m_fdblocks and in xfs_reserve_blocks() 400 * reserve. in xfs_reserve_blocks() 402 * Don't set the reserved flag here - we don't want to reserve in xfs_reserve_blocks() 403 * the extra reserve blocks from the reserve..... in xfs_reserve_blocks() 411 * Update the reserve counters if blocks have been successfully in xfs_reserve_blocks() 517 * Reserve free space for per-AG metadata. [all …]
|
/Linux-v5.10/arch/x86/kernel/ |
D | ebda.c | 27 * to reserve the EBDA area in the 'RAM size' value ... 30 * chipset: reserve a page before VGA to prevent PCI prefetch 90 * then also reserve everything from the EBDA start address up to in reserve_bios_regions() 96 /* Reserve all memory between bios_start and the 1MB mark: */ in reserve_bios_regions()
|
/Linux-v5.10/arch/powerpc/include/asm/ |
D | kdump.h | 9 /* How many bytes to reserve at zero for kdump. The reserve limit should 11 * Reserve to the end of the FWNMI area, see head_64.S */
|
/Linux-v5.10/drivers/char/agp/ |
D | frontend.c | 799 struct agp_region reserve; in agpioc_reserve_wrap() local 804 if (copy_from_user(&reserve, arg, sizeof(struct agp_region))) in agpioc_reserve_wrap() 807 if ((unsigned) reserve.seg_count >= ~0U/sizeof(struct agp_segment)) in agpioc_reserve_wrap() 810 client = agp_find_client_by_pid(reserve.pid); in agpioc_reserve_wrap() 812 if (reserve.seg_count == 0) { in agpioc_reserve_wrap() 814 client_priv = agp_find_private(reserve.pid); in agpioc_reserve_wrap() 824 return agp_remove_client(reserve.pid); in agpioc_reserve_wrap() 828 if (reserve.seg_count >= 16384) in agpioc_reserve_wrap() 831 segment = kmalloc((sizeof(struct agp_segment) * reserve.seg_count), in agpioc_reserve_wrap() 837 if (copy_from_user(segment, (void __user *) reserve.seg_list, in agpioc_reserve_wrap() [all …]
|
/Linux-v5.10/tools/testing/selftests/bpf/benchs/ |
D | run_bench_ringbufs.sh | 58 header "Ringbuf back-to-back, reserve+commit vs output" 59 summarize "reserve" "$($RUN_BENCH --rb-b2b rb-custom)" 62 header "Ringbuf sampled, reserve+commit vs output" 63 summarize "reserve-sampled" "$($RUN_BENCH --rb-sampled rb-custom)"
|
/Linux-v5.10/Documentation/vm/ |
D | hugetlbfs_reserv.rst | 21 'reserve' huge pages at mmap() time to ensure that huge pages would be 23 describe how huge page reserve processing is done in the v4.10 kernel. 40 Reserve Map 41 A reserve map is described by the structure:: 52 There is one reserve map for each huge page mapping in the system. 92 Specifically, vma->vm_private_data. This reserve map is created at the 240 The routine vma_commit_reservation() is then called to adjust the reserve 244 in the reserve map already existed so no change is made. However, if there 248 It is possible that the reserve map could have been changed between the call 255 a race is detected, the subpool and global reserve counts are adjusted to [all …]
|
/Linux-v5.10/arch/um/kernel/ |
D | physmem.c | 82 unsigned long reserve = reserve_end - start; in setup_physmem() local 83 long map_size = len - reserve; in setup_physmem() 88 reserve, len); in setup_physmem() 94 err = os_map_memory((void *) reserve_end, physmem_fd, reserve, in setup_physmem() 112 memblock_reserve(__pa(start), reserve); in setup_physmem()
|
/Linux-v5.10/include/drm/ttm/ |
D | ttm_execbuf_util.h | 55 * @ticket: ww_acquire_ctx from reserve call 74 * Tries to reserve bos pointed to by the list entries for validation. 79 * reserve the same buffers in reverse order, all threads except one will 105 * @ticket: ww_acquire_ctx from reserve call
|
/Linux-v5.10/include/linux/ |
D | relay.h | 96 * NOTE: the client can reserve bytes at the beginning of the new 245 * relay_reserve - reserve slot in channel buffer 247 * @length: number of bytes to reserve 274 * subbuf_start_reserve - reserve bytes at the start of a sub-buffer 276 * @length: number of bytes to reserve 278 * Helper function used to reserve bytes at the beginning of
|
/Linux-v5.10/drivers/dax/pmem/ |
D | core.c | 39 /* reserve the metadata area, device-dax will reserve the data */ in __dax_pmem_probe() 45 dev_warn(dev, "could not reserve metadata\n"); in __dax_pmem_probe()
|
/Linux-v5.10/drivers/xen/ |
D | xen-stub.c | 3 * xen-stub.c - stub drivers to reserve space for Xen 43 /* just reserve space for Xen, block native driver loaded */ in xen_stub_memory_device_init() 78 /* just reserve space for Xen, block native driver loaded */ in xen_stub_processor_init()
|
/Linux-v5.10/drivers/md/bcache/ |
D | alloc.c | 306 /* Prios/gens are actually the most important reserve */ in bch_allocator_push() 392 long bch_bucket_alloc(struct cache *ca, unsigned int reserve, bool wait) in bch_bucket_alloc() argument 405 fifo_pop(&ca->free[reserve], r)) in bch_bucket_alloc() 409 trace_bcache_alloc_fail(ca, reserve); in bch_bucket_alloc() 421 !fifo_pop(&ca->free[reserve], r)); in bch_bucket_alloc() 428 trace_bcache_alloc(ca, reserve); in bch_bucket_alloc() 451 if (reserve <= RESERVE_PRIO) { in bch_bucket_alloc() 489 int __bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve, in __bch_bucket_alloc_set() argument 504 b = bch_bucket_alloc(ca, reserve, wait); in __bch_bucket_alloc_set() 521 int bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve, in bch_bucket_alloc_set() argument [all …]
|
/Linux-v5.10/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
D | nv04.c | 165 /* PRAMIN aperture maps over the end of VRAM, reserve it */ in nv04_instmem_oneinit() 172 /* 0x00000-0x10000: reserve for probable vbios image */ in nv04_instmem_oneinit() 178 /* 0x10000-0x18000: reserve for RAMHT */ in nv04_instmem_oneinit() 183 /* 0x18000-0x18800: reserve for RAMFC (enough for 32 nv30 channels) */ in nv04_instmem_oneinit() 189 /* 0x18800-0x18a00: reserve for RAMRO */ in nv04_instmem_oneinit()
|
/Linux-v5.10/drivers/input/joystick/ |
D | walkera0701.c | 13 #define RESERVE 20000 macro 149 if (abs(pulse_time - SYNC_PULSE) < RESERVE) /* new frame sync */ in walkera0701_irq_handler() 152 if ((pulse_time > (ANALOG_MIN_PULSE - RESERVE) in walkera0701_irq_handler() 153 && (pulse_time < (ANALOG_MAX_PULSE + RESERVE)))) { in walkera0701_irq_handler() 154 pulse_time -= (ANALOG_MIN_PULSE - RESERVE); in walkera0701_irq_handler() 161 RESERVE + BIN1_PULSE - BIN0_PULSE) /* frame sync .. */ in walkera0701_irq_handler()
|
/Linux-v5.10/drivers/target/ |
D | target_core_pr.c | 131 * From spc4r17 5.7.3 Exceptions to SPC-2 RESERVE and RELEASE in target_check_scsi2_reservation_conflict() 134 * A RESERVE(6) or RESERVE(10) command shall complete with GOOD in target_check_scsi2_reservation_conflict() 147 * In all other cases, a RESERVE(6) command, RESERVE(10) command, in target_check_scsi2_reservation_conflict() 168 * If a logical unit has executed a PERSISTENT RESERVE OUT in target_check_scsi2_reservation_conflict() 171 * initiator, all RESERVE commands and all RELEASE commands in target_check_scsi2_reservation_conflict() 181 pr_err("Received legacy SPC-2 RESERVE/RELEASE" in target_check_scsi2_reservation_conflict() 425 case RESERVE: in core_scsi3_pr_seq_non_holder() 907 pr_debug("SPC-3 PR [%s] Service Action: APTPL RESERVE created" in core_scsi3_aptpl_reserve() 912 pr_debug("SPC-3 PR [%s] RESERVE Node: %s%s\n", in core_scsi3_aptpl_reserve() 1246 * type, then a PERSISTENT RESERVE OUT command with REGISTER in core_scsi3_check_implicit_release() [all …]
|
/Linux-v5.10/drivers/mtd/ubi/ |
D | Kconfig | 47 To put it differently, if this value is 20, UBI will try to reserve 52 MTD partitions of the same size, UBI will reserve 40 eraseblocks when 77 into one PEB. UBI will reserve PEBs to hold two fastmaps.
|