| /Linux-v6.1/arch/powerpc/mm/nohash/ |
| D | kaslr_booke.c | 24 struct regions { struct 39 struct regions __initdata regions; argument 114 if (regions.reserved_mem < 0) in overlaps_reserved_region() 118 for (subnode = fdt_first_subnode(fdt, regions.reserved_mem); in overlaps_reserved_region() 126 while (len >= (regions.reserved_mem_addr_cells + in overlaps_reserved_region() 127 regions.reserved_mem_size_cells)) { in overlaps_reserved_region() 129 if (regions.reserved_mem_addr_cells == 2) in overlaps_reserved_region() 132 reg += regions.reserved_mem_addr_cells; in overlaps_reserved_region() 133 len -= 4 * regions.reserved_mem_addr_cells; in overlaps_reserved_region() 136 if (regions.reserved_mem_size_cells == 2) in overlaps_reserved_region() [all …]
|
| /Linux-v6.1/drivers/mtd/chips/ |
| D | jedec_probe.c | 274 const uint32_t regions[6]; member 306 .regions = { 318 .regions = { 333 .regions = { 348 .regions = { 363 .regions = { 378 .regions = { 394 .regions = { 411 .regions = { 428 .regions = { [all …]
|
| /Linux-v6.1/mm/damon/ |
| D | vaddr-test.h | 62 struct damon_addr_range regions[3] = {0,}; in damon_test_three_regions_in_vmas() local 76 __damon_va_three_regions(&mm, regions); in damon_test_three_regions_in_vmas() 78 KUNIT_EXPECT_EQ(test, 10ul, regions[0].start); in damon_test_three_regions_in_vmas() 79 KUNIT_EXPECT_EQ(test, 25ul, regions[0].end); in damon_test_three_regions_in_vmas() 80 KUNIT_EXPECT_EQ(test, 200ul, regions[1].start); in damon_test_three_regions_in_vmas() 81 KUNIT_EXPECT_EQ(test, 220ul, regions[1].end); in damon_test_three_regions_in_vmas() 82 KUNIT_EXPECT_EQ(test, 300ul, regions[2].start); in damon_test_three_regions_in_vmas() 83 KUNIT_EXPECT_EQ(test, 330ul, regions[2].end); in damon_test_three_regions_in_vmas() 122 unsigned long *regions, int nr_regions, in damon_do_test_apply_three_regions() argument 132 r = damon_new_region(regions[i * 2], regions[i * 2 + 1]); in damon_do_test_apply_three_regions() [all …]
|
| D | vaddr.c | 117 struct damon_addr_range regions[3]) in __damon_va_three_regions() 158 regions[0].start = ALIGN(start, DAMON_MIN_REGION); in __damon_va_three_regions() 159 regions[0].end = ALIGN(first_gap.start, DAMON_MIN_REGION); in __damon_va_three_regions() 160 regions[1].start = ALIGN(first_gap.end, DAMON_MIN_REGION); in __damon_va_three_regions() 161 regions[1].end = ALIGN(second_gap.start, DAMON_MIN_REGION); in __damon_va_three_regions() 162 regions[2].start = ALIGN(second_gap.end, DAMON_MIN_REGION); in __damon_va_three_regions() 163 regions[2].end = ALIGN(prev->vm_end, DAMON_MIN_REGION); in __damon_va_three_regions() 174 struct damon_addr_range regions[3]) in damon_va_three_regions() 184 rc = __damon_va_three_regions(mm, regions); in damon_va_three_regions() 238 struct damon_addr_range regions[3]; in __damon_va_init_regions() local [all …]
|
| /Linux-v6.1/drivers/gpu/drm/nouveau/nvkm/nvfw/ |
| D | acr.c | 89 hdr->regions.no_regions); in flcn_acr_desc_dump() 91 for (i = 0; i < ARRAY_SIZE(hdr->regions.region_props); i++) { in flcn_acr_desc_dump() 94 hdr->regions.region_props[i].start_addr); in flcn_acr_desc_dump() 96 hdr->regions.region_props[i].end_addr); in flcn_acr_desc_dump() 98 hdr->regions.region_props[i].region_id); in flcn_acr_desc_dump() 100 hdr->regions.region_props[i].read_mask); in flcn_acr_desc_dump() 102 hdr->regions.region_props[i].write_mask); in flcn_acr_desc_dump() 104 hdr->regions.region_props[i].client_mask); in flcn_acr_desc_dump() 132 hdr->regions.no_regions); in flcn_acr_desc_v1_dump() 134 for (i = 0; i < ARRAY_SIZE(hdr->regions.region_props); i++) { in flcn_acr_desc_v1_dump() [all …]
|
| /Linux-v6.1/drivers/vfio/platform/ |
| D | vfio_platform_common.c | 145 vdev->regions = kcalloc(cnt, sizeof(struct vfio_platform_region), in vfio_platform_regions_init() 147 if (!vdev->regions) in vfio_platform_regions_init() 157 vdev->regions[i].addr = res->start; in vfio_platform_regions_init() 158 vdev->regions[i].size = resource_size(res); in vfio_platform_regions_init() 159 vdev->regions[i].flags = 0; in vfio_platform_regions_init() 163 vdev->regions[i].type = VFIO_PLATFORM_REGION_TYPE_MMIO; in vfio_platform_regions_init() 164 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ; in vfio_platform_regions_init() 166 vdev->regions[i].flags |= in vfio_platform_regions_init() 173 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_platform_regions_init() 174 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_platform_regions_init() [all …]
|
| /Linux-v6.1/tools/testing/memblock/tests/ |
| D | alloc_nid_api.c | 55 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_try_nid_top_down_simple_check() 107 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_try_nid_top_down_end_misaligned_check() 158 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_try_nid_exact_address_generic_check() 210 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_try_nid_top_down_narrow_range_check() 300 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_try_nid_min_reserved_generic_check() 352 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_try_nid_max_reserved_generic_check() 405 struct memblock_region *rgn1 = &memblock.reserved.regions[1]; in alloc_try_nid_top_down_reserved_with_space_check() 406 struct memblock_region *rgn2 = &memblock.reserved.regions[0]; in alloc_try_nid_top_down_reserved_with_space_check() 470 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_try_nid_reserved_full_merge_generic_check() 532 struct memblock_region *rgn1 = &memblock.reserved.regions[1]; in alloc_try_nid_top_down_reserved_no_space_check() [all …]
|
| D | basic_api.c | 17 ASSERT_NE(memblock.memory.regions, NULL); in memblock_initialization_check() 22 ASSERT_NE(memblock.reserved.regions, NULL); in memblock_initialization_check() 45 rgn = &memblock.memory.regions[0]; in memblock_add_simple_check() 78 rgn = &memblock.memory.regions[0]; in memblock_add_node_simple_check() 121 rgn1 = &memblock.memory.regions[0]; in memblock_add_disjoint_check() 122 rgn2 = &memblock.memory.regions[1]; in memblock_add_disjoint_check() 175 rgn = &memblock.memory.regions[0]; in memblock_add_overlap_top_check() 227 rgn = &memblock.memory.regions[0]; in memblock_add_overlap_bottom_check() 276 rgn = &memblock.memory.regions[0]; in memblock_add_within_check() 347 rgn = &memblock.memory.regions[0]; in memblock_add_between_check() [all …]
|
| D | alloc_api.c | 26 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_simple_check() 73 struct memblock_region *rgn1 = &memblock.reserved.regions[1]; in alloc_top_down_disjoint_check() 74 struct memblock_region *rgn2 = &memblock.reserved.regions[0]; in alloc_top_down_disjoint_check() 125 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_before_check() 168 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_after_check() 217 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_second_fit_check() 266 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_in_between_generic_check() 416 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_limited_space_generic_check() 450 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_no_memory_generic_check() 484 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_too_large_generic_check() [all …]
|
| D | alloc_helpers_api.c | 20 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_simple_generic_check() 63 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_misaligned_generic_check() 110 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_high_addr_check() 153 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_no_space_above_check() 190 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_min_addr_cap_check() 236 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_high_addr_check() 278 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_no_space_above_check() 314 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_min_addr_cap_check()
|
| /Linux-v6.1/mm/ |
| D | memblock.c | 116 .memory.regions = memblock_memory_init_regions, 121 .reserved.regions = memblock_reserved_init_regions, 132 .regions = memblock_physmem_init_regions, 148 for (i = 0, rgn = &memblock_type->regions[0]; \ 150 i++, rgn = &memblock_type->regions[i]) 192 if (memblock_addrs_overlap(base, size, type->regions[i].base, in memblock_overlaps_region() 193 type->regions[i].size)) in memblock_overlaps_region() 345 type->total_size -= type->regions[r].size; in memblock_remove_region() 346 memmove(&type->regions[r], &type->regions[r + 1], in memblock_remove_region() 347 (type->cnt - (r + 1)) * sizeof(type->regions[r])); in memblock_remove_region() [all …]
|
| /Linux-v6.1/drivers/virt/nitro_enclaves/ |
| D | ne_misc_dev_test.c | 117 phys_contig_mem_regions.regions = kunit_kcalloc(test, MAX_PHYS_REGIONS, in ne_misc_dev_test_merge_phys_contig_memory_regions() 118 sizeof(*phys_contig_mem_regions.regions), in ne_misc_dev_test_merge_phys_contig_memory_regions() 120 KUNIT_ASSERT_TRUE(test, phys_contig_mem_regions.regions); in ne_misc_dev_test_merge_phys_contig_memory_regions() 135 KUNIT_EXPECT_EQ(test, phys_contig_mem_regions.regions[num - 1].start, in ne_misc_dev_test_merge_phys_contig_memory_regions() 137 KUNIT_EXPECT_EQ(test, range_len(&phys_contig_mem_regions.regions[num - 1]), in ne_misc_dev_test_merge_phys_contig_memory_regions() 141 kunit_kfree(test, phys_contig_mem_regions.regions); in ne_misc_dev_test_merge_phys_contig_memory_regions()
|
| /Linux-v6.1/drivers/vfio/fsl-mc/ |
| D | vfio_fsl_mc.c | 30 vdev->regions = kcalloc(count, sizeof(struct vfio_fsl_mc_region), in vfio_fsl_mc_open_device() 32 if (!vdev->regions) in vfio_fsl_mc_open_device() 36 struct resource *res = &mc_dev->regions[i]; in vfio_fsl_mc_open_device() 39 vdev->regions[i].addr = res->start; in vfio_fsl_mc_open_device() 40 vdev->regions[i].size = resource_size(res); in vfio_fsl_mc_open_device() 41 vdev->regions[i].type = mc_dev->regions[i].flags & IORESOURCE_BITS; in vfio_fsl_mc_open_device() 46 if (!no_mmap && !(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_fsl_mc_open_device() 47 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_fsl_mc_open_device() 48 vdev->regions[i].flags |= in vfio_fsl_mc_open_device() 50 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ; in vfio_fsl_mc_open_device() [all …]
|
| /Linux-v6.1/drivers/net/dsa/sja1105/ |
| D | sja1105_devlink.c | 85 priv->regions = kcalloc(num_regions, sizeof(struct devlink_region *), in sja1105_setup_devlink_regions() 87 if (!priv->regions) in sja1105_setup_devlink_regions() 97 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_setup_devlink_regions() 99 kfree(priv->regions); in sja1105_setup_devlink_regions() 103 priv->regions[i] = region; in sja1105_setup_devlink_regions() 115 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_teardown_devlink_regions() 117 kfree(priv->regions); in sja1105_teardown_devlink_regions()
|
| /Linux-v6.1/Documentation/admin-guide/device-mapper/ |
| D | dm-clone.rst | 58 3. A small metadata device - it records which regions are already valid in the 59 destination device, i.e., which regions have already been hydrated, or have 68 dm-clone divides the source and destination devices in fixed sized regions. 77 Reads and writes from/to hydrated regions are serviced from the destination 93 as a hint to skip hydration of the regions covered by the request, i.e., it 111 A message `hydration_threshold <#regions>` can be used to set the maximum number 112 of regions being copied, the default being 1 region. 116 region size. A message `hydration_batch_size <#regions>` can be used to tune the 118 dm-clone trying to batch together contiguous regions, so we copy the data in 119 batches of this many regions. [all …]
|
| /Linux-v6.1/drivers/soc/qcom/ |
| D | smem.c | 287 struct smem_region regions[]; member 430 header = smem->regions[0].virt_base; in qcom_smem_alloc_global() 517 header = smem->regions[0].virt_base; in qcom_smem_get_global() 525 region = &smem->regions[i]; in qcom_smem_get_global() 713 header = __smem->regions[0].virt_base; in qcom_smem_get_free_space() 716 if (ret > __smem->regions[0].size) in qcom_smem_get_free_space() 762 area = &__smem->regions[i]; in qcom_smem_virt_to_phys() 780 header = smem->regions[0].virt_base; in qcom_smem_get_sbl_version() 833 phys_addr = smem->regions[0].aux_base + le32_to_cpu(entry->offset); in qcom_smem_partition_header() 915 smem->global_partition.phys_base = smem->regions[0].aux_base + in qcom_smem_set_global_partition() [all …]
|
| /Linux-v6.1/drivers/perf/ |
| D | marvell_cn10k_tad_pmu.c | 32 struct tad_region *regions; member 53 new += readq(tad_pmu->regions[i].base + in tad_pmu_event_counter_read() 71 writeq_relaxed(0, tad_pmu->regions[i].base + in tad_pmu_event_counter_stop() 92 writeq_relaxed(0, tad_pmu->regions[i].base + in tad_pmu_event_counter_start() 100 writeq_relaxed(reg_val, tad_pmu->regions[i].base + in tad_pmu_event_counter_start() 258 struct tad_region *regions; in tad_pmu_probe() local 299 regions = devm_kcalloc(&pdev->dev, tad_cnt, in tad_pmu_probe() 300 sizeof(*regions), GFP_KERNEL); in tad_pmu_probe() 301 if (!regions) in tad_pmu_probe() 306 regions[i].base = devm_ioremap(&pdev->dev, in tad_pmu_probe() [all …]
|
| /Linux-v6.1/Documentation/networking/devlink/ |
| D | devlink-region.rst | 7 ``devlink`` regions enable access to driver defined address regions using 10 Each device can create and register its own supported address regions. The 22 address regions that are otherwise inaccessible to the user. 45 # Show all of the exposed regions with region sizes: 68 As regions are likely very device or driver specific, no generic regions are 70 specific regions a driver supports.
|
| /Linux-v6.1/drivers/gpu/drm/i915/gem/selftests/ |
| D | i915_gem_dmabuf.c | 91 struct intel_memory_region *lmem = i915->mm.regions[INTEL_REGION_LMEM_0]; in igt_dmabuf_import_same_driver_lmem() 144 struct intel_memory_region **regions, in igt_dmabuf_import_same_driver() argument 158 regions, num_regions); in igt_dmabuf_import_same_driver() 202 if (obj->mm.region != i915->mm.regions[INTEL_REGION_SMEM]) { in igt_dmabuf_import_same_driver() 246 struct intel_memory_region *smem = i915->mm.regions[INTEL_REGION_SMEM]; in igt_dmabuf_import_same_driver_smem() 254 struct intel_memory_region *regions[2]; in igt_dmabuf_import_same_driver_lmem_smem() local 256 if (!i915->mm.regions[INTEL_REGION_LMEM_0]) in igt_dmabuf_import_same_driver_lmem_smem() 259 regions[0] = i915->mm.regions[INTEL_REGION_LMEM_0]; in igt_dmabuf_import_same_driver_lmem_smem() 260 regions[1] = i915->mm.regions[INTEL_REGION_SMEM]; in igt_dmabuf_import_same_driver_lmem_smem() 261 return igt_dmabuf_import_same_driver(i915, regions, 2); in igt_dmabuf_import_same_driver_lmem_smem()
|
| /Linux-v6.1/Documentation/admin-guide/mm/damon/ |
| D | lru_sort.rst | 31 DAMON_LRU_SORT finds hot pages (pages of memory regions that showing access 33 memory regions that showing no access for a time that longer than a 85 Access frequency threshold for hot memory regions identification in permil. 94 Time threshold for cold memory regions identification in microseconds. 179 Minimum number of monitoring regions. 181 The minimal number of monitoring regions of DAMON for the cold memory 190 Maximum number of monitoring regions. 192 The maximum number of monitoring regions of DAMON for the cold memory 225 Number of hot memory regions that tried to be LRU-sorted. 230 Total bytes of hot memory regions that tried to be LRU-sorted. [all …]
|
| D | reclaim.rst | 33 DAMON_RECLAIM finds memory regions that didn't accessed for specific time 36 out memory regions that didn't accessed longer time first. System 77 Time threshold for cold memory regions identification in microseconds. 173 Minimum number of monitoring regions. 175 The minimal number of monitoring regions of DAMON for the cold memory 183 Maximum number of monitoring regions. 185 The maximum number of monitoring regions of DAMON for the cold memory 196 against. That is, DAMON_RECLAIM will find cold memory regions in this region 205 against. That is, DAMON_RECLAIM will find cold memory regions in this region 219 Number of memory regions that tried to be reclaimed by DAMON_RECLAIM. [all …]
|
| /Linux-v6.1/drivers/gpu/drm/nouveau/nvkm/subdev/acr/ |
| D | gp102.c | 200 desc->regions.no_regions = 2; in gp102_acr_load_load() 201 desc->regions.region_props[0].start_addr = acr->wpr_start >> 8; in gp102_acr_load_load() 202 desc->regions.region_props[0].end_addr = acr->wpr_end >> 8; in gp102_acr_load_load() 203 desc->regions.region_props[0].region_id = 1; in gp102_acr_load_load() 204 desc->regions.region_props[0].read_mask = 0xf; in gp102_acr_load_load() 205 desc->regions.region_props[0].write_mask = 0xc; in gp102_acr_load_load() 206 desc->regions.region_props[0].client_mask = 0x2; in gp102_acr_load_load() 207 desc->regions.region_props[0].shadow_mem_start_addr = in gp102_acr_load_load()
|
| /Linux-v6.1/drivers/virt/acrn/ |
| D | mm.c | 20 struct vm_memory_region_batch *regions; in modify_region() local 23 regions = kzalloc(sizeof(*regions), GFP_KERNEL); in modify_region() 24 if (!regions) in modify_region() 27 regions->vmid = vm->vmid; in modify_region() 28 regions->regions_num = 1; in modify_region() 29 regions->regions_gpa = virt_to_phys(region); in modify_region() 31 ret = hcall_set_memory_regions(virt_to_phys(regions)); in modify_region() 36 kfree(regions); in modify_region()
|
| /Linux-v6.1/Documentation/mm/damon/ |
| D | design.rst | 51 address regions is just wasteful. However, because DAMON can deal with some 52 level of noise using the adaptive regions adjustment mechanism, tracking every 58 distinct regions that cover every mapped area of the address space. The two 59 gaps between the three regions are the two biggest unmapped areas in the given 69 (small mmap()-ed regions and munmap()-ed regions) 98 ``update interval``, ``minimum number of regions``, and ``maximum number of 99 regions``. 139 overhead is controllable by setting the number of regions. DAMON allows users 140 to set the minimum and the maximum number of regions for the trade-off. 149 Even somehow the initial monitoring target regions are well constructed to [all …]
|
| /Linux-v6.1/drivers/net/wireless/ath/ath10k/ |
| D | coredump.c | 1297 .regions = qca6174_hw10_mem_regions, 1306 .regions = qca6174_hw10_mem_regions, 1315 .regions = qca6174_hw10_mem_regions, 1324 .regions = qca6174_hw21_mem_regions, 1333 .regions = qca6174_hw30_mem_regions, 1342 .regions = qca6174_hw30_mem_regions, 1351 .regions = qca6174_hw30_sdio_mem_regions, 1360 .regions = qca6174_hw30_mem_regions, 1369 .regions = qca988x_hw20_mem_regions, 1378 .regions = qca9984_hw10_mem_regions, [all …]
|