/Linux-v4.19/drivers/gpu/drm/ttm/ |
D | ttm_memory.c | 78 struct ttm_mem_zone *zone = in ttm_mem_zone_kobj_release() local 82 zone->name, (unsigned long long)zone->used_mem >> 10); in ttm_mem_zone_kobj_release() 83 kfree(zone); in ttm_mem_zone_kobj_release() 90 struct ttm_mem_zone *zone = in ttm_mem_zone_show() local 94 spin_lock(&zone->glob->lock); in ttm_mem_zone_show() 96 val = zone->zone_mem; in ttm_mem_zone_show() 98 val = zone->emer_mem; in ttm_mem_zone_show() 100 val = zone->max_mem; in ttm_mem_zone_show() 102 val = zone->swap_limit; in ttm_mem_zone_show() 104 val = zone->used_mem; in ttm_mem_zone_show() [all …]
|
/Linux-v4.19/mm/ |
D | compaction.c | 141 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 143 zone->compact_considered = 0; in defer_compaction() 144 zone->compact_defer_shift++; in defer_compaction() 146 if (order < zone->compact_order_failed) in defer_compaction() 147 zone->compact_order_failed = order; in defer_compaction() 149 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction() 150 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction() 152 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction() 156 bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument 158 unsigned long defer_limit = 1UL << zone->compact_defer_shift; in compaction_deferred() [all …]
|
D | page_alloc.c | 454 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument 462 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries() 463 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries() 464 sp = zone->spanned_pages; in page_outside_zone_boundaries() 465 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries() 467 } while (zone_span_seqretry(zone, seq)); in page_outside_zone_boundaries() 471 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries() 477 static int page_is_consistent(struct zone *zone, struct page *page) in page_is_consistent() argument 481 if (zone != page_zone(page)) in page_is_consistent() 489 static int __maybe_unused bad_range(struct zone *zone, struct page *page) in bad_range() argument [all …]
|
D | vmstat.c | 39 static void zero_zone_numa_counters(struct zone *zone) in zero_zone_numa_counters() argument 44 atomic_long_set(&zone->vm_numa_stat[item], 0); in zero_zone_numa_counters() 46 per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item] in zero_zone_numa_counters() 54 struct zone *zone; in zero_zones_numa_counters() local 56 for_each_populated_zone(zone) in zero_zones_numa_counters() 57 zero_zone_numa_counters(zone); in zero_zones_numa_counters() 171 int calculate_pressure_threshold(struct zone *zone) in calculate_pressure_threshold() argument 184 watermark_distance = low_wmark_pages(zone) - min_wmark_pages(zone); in calculate_pressure_threshold() 195 int calculate_normal_threshold(struct zone *zone) in calculate_normal_threshold() argument 230 mem = zone->managed_pages >> (27 - PAGE_SHIFT); in calculate_normal_threshold() [all …]
|
D | memory_hotplug.c | 319 static unsigned long find_smallest_section_pfn(int nid, struct zone *zone, in find_smallest_section_pfn() argument 334 if (zone && zone != page_zone(pfn_to_page(start_pfn))) in find_smallest_section_pfn() 344 static unsigned long find_biggest_section_pfn(int nid, struct zone *zone, in find_biggest_section_pfn() argument 362 if (zone && zone != page_zone(pfn_to_page(pfn))) in find_biggest_section_pfn() 371 static void shrink_zone_span(struct zone *zone, unsigned long start_pfn, in shrink_zone_span() argument 374 unsigned long zone_start_pfn = zone->zone_start_pfn; in shrink_zone_span() 375 unsigned long z = zone_end_pfn(zone); /* zone_end_pfn namespace clash */ in shrink_zone_span() 379 int nid = zone_to_nid(zone); in shrink_zone_span() 381 zone_span_writelock(zone); in shrink_zone_span() 389 pfn = find_smallest_section_pfn(nid, zone, end_pfn, in shrink_zone_span() [all …]
|
D | page_isolation.c | 21 struct zone *zone; in set_migratetype_isolate() local 27 zone = page_zone(page); in set_migratetype_isolate() 29 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 63 if (!has_unmovable_pages(zone, page, arg.pages_found, migratetype, in set_migratetype_isolate() 78 zone->nr_isolate_pageblock++; in set_migratetype_isolate() 79 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE, in set_migratetype_isolate() 82 __mod_zone_freepage_state(zone, -nr_pages, mt); in set_migratetype_isolate() 85 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 87 drain_all_pages(zone); in set_migratetype_isolate() 93 struct zone *zone; in unset_migratetype_isolate() local [all …]
|
D | mmzone.c | 30 struct zone *next_zone(struct zone *zone) in next_zone() argument 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 35 zone++; in next_zone() 39 zone = pgdat->node_zones; in next_zone() 41 zone = NULL; in next_zone() 43 return zone; in next_zone() 69 (z->zone && !zref_in_nodemask(z, nodes))) in __next_zones_zonelist() 77 struct page *page, struct zone *zone) in memmap_valid_within() argument 82 if (page_zone(page) != zone) in memmap_valid_within()
|
/Linux-v4.19/include/linux/ |
D | mmzone.h | 359 struct zone { struct 525 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() argument 527 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn() 530 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument 532 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 535 static inline bool zone_is_initialized(struct zone *zone) in zone_is_initialized() argument 537 return zone->initialized; in zone_is_initialized() 540 static inline bool zone_is_empty(struct zone *zone) in zone_is_empty() argument 542 return zone->spanned_pages == 0; in zone_is_empty() 549 static inline bool zone_intersects(struct zone *zone, in zone_intersects() argument [all …]
|
D | memory_hotplug.h | 11 struct zone; 61 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument 63 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin() 65 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument 67 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry() 69 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument 71 write_seqlock(&zone->span_seqlock); in zone_span_writelock() 73 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument 75 write_sequnlock(&zone->span_seqlock); in zone_span_writeunlock() 77 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument [all …]
|
D | vmstat.h | 136 static inline void zone_numa_state_add(long x, struct zone *zone, in zone_numa_state_add() argument 139 atomic_long_add(x, &zone->vm_numa_stat[item]); in zone_numa_state_add() 150 static inline unsigned long zone_numa_state_snapshot(struct zone *zone, in zone_numa_state_snapshot() argument 153 long x = atomic_long_read(&zone->vm_numa_stat[item]); in zone_numa_state_snapshot() 157 x += per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item]; in zone_numa_state_snapshot() 163 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument 166 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add() 197 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument 200 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state() 214 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument [all …]
|
D | compaction.h | 95 extern int fragmentation_index(struct zone *zone, unsigned int order); 100 extern enum compact_result compaction_suitable(struct zone *zone, int order, 103 extern void defer_compaction(struct zone *zone, int order); 104 extern bool compaction_deferred(struct zone *zone, int order); 105 extern void compaction_defer_reset(struct zone *zone, int order, 107 extern bool compaction_restarting(struct zone *zone, int order); 186 static inline enum compact_result compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument 192 static inline void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 196 static inline bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
|
D | page-isolation.h | 6 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument 8 return zone->nr_isolate_pageblock; in has_isolate_pageblock() 19 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument 33 bool has_unmovable_pages(struct zone *zone, struct page *page, int count, 36 int move_freepages_block(struct zone *zone, struct page *page,
|
/Linux-v4.19/fs/btrfs/ |
D | reada.c | 228 struct reada_zone *zone; in reada_find_zone() local 234 zone = NULL; in reada_find_zone() 236 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone() 238 if (ret == 1 && logical >= zone->start && logical <= zone->end) { in reada_find_zone() 239 kref_get(&zone->refcnt); in reada_find_zone() 241 return zone; in reada_find_zone() 254 zone = kzalloc(sizeof(*zone), GFP_KERNEL); in reada_find_zone() 255 if (!zone) in reada_find_zone() 260 kfree(zone); in reada_find_zone() 264 zone->start = start; in reada_find_zone() [all …]
|
/Linux-v4.19/include/net/netfilter/ |
D | nf_conntrack_zones.h | 14 return &ct->zone; in nf_ct_zone() 21 nf_ct_zone_init(struct nf_conntrack_zone *zone, u16 id, u8 dir, u8 flags) in nf_ct_zone_init() argument 23 zone->id = id; in nf_ct_zone_init() 24 zone->flags = flags; in nf_ct_zone_init() 25 zone->dir = dir; in nf_ct_zone_init() 27 return zone; in nf_ct_zone_init() 38 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 39 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl() 45 const struct nf_conntrack_zone *zone) in nf_ct_zone_add() argument 48 ct->zone = *zone; in nf_ct_zone_add() [all …]
|
/Linux-v4.19/drivers/md/ |
D | dm-zoned-metadata.c | 187 unsigned int dmz_id(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_id() argument 189 return ((unsigned int)(zone - zmd->zones)); in dmz_id() 192 sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_sect() argument 194 return (sector_t)dmz_id(zmd, zone) << zmd->dev->zone_nr_sectors_shift; in dmz_start_sect() 197 sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_block() argument 199 return (sector_t)dmz_id(zmd, zone) << zmd->dev->zone_nr_blocks_shift; in dmz_start_block() 1038 static int dmz_init_zone(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_init_zone() argument 1050 INIT_LIST_HEAD(&zone->link); in dmz_init_zone() 1051 atomic_set(&zone->refcount, 0); in dmz_init_zone() 1052 zone->chunk = DMZ_MAP_UNMAPPED; in dmz_init_zone() [all …]
|
D | dm-zoned-target.c | 20 struct dm_zone *zone; member 103 static int dmz_submit_read_bio(struct dmz_target *dmz, struct dm_zone *zone, in dmz_submit_read_bio() argument 112 sector = dmz_start_sect(dmz->metadata, zone) + dmz_blk2sect(chunk_block); in dmz_submit_read_bio() 162 static int dmz_handle_read(struct dmz_target *dmz, struct dm_zone *zone, in dmz_handle_read() argument 172 if (!zone) { in dmz_handle_read() 179 (dmz_is_rnd(zone) ? "RND" : "SEQ"), in dmz_handle_read() 180 dmz_id(dmz->metadata, zone), in dmz_handle_read() 184 bzone = zone->bzone; in dmz_handle_read() 187 if (dmz_is_rnd(zone) || chunk_block < zone->wp_block) { in dmz_handle_read() 189 ret = dmz_block_valid(dmz->metadata, zone, chunk_block); in dmz_handle_read() [all …]
|
/Linux-v4.19/drivers/block/ |
D | null_blk_zoned.c | 33 struct blk_zone *zone = &dev->zones[i]; in null_zone_init() local 35 zone->start = zone->wp = sector; in null_zone_init() 36 zone->len = dev->zone_size_sects; in null_zone_init() 37 zone->type = BLK_ZONE_TYPE_SEQWRITE_REQ; in null_zone_init() 38 zone->cond = BLK_ZONE_COND_EMPTY; in null_zone_init() 106 struct blk_zone *zone = &dev->zones[zno]; in null_zone_write() local 108 switch (zone->cond) { in null_zone_write() 116 if (sector != zone->wp) { in null_zone_write() 121 if (zone->cond == BLK_ZONE_COND_EMPTY) in null_zone_write() 122 zone->cond = BLK_ZONE_COND_IMP_OPEN; in null_zone_write() [all …]
|
/Linux-v4.19/kernel/power/ |
D | snapshot.c | 373 struct mem_zone_bm_rtree *zone; member 430 static int add_rtree_block(struct mem_zone_bm_rtree *zone, gfp_t gfp_mask, in add_rtree_block() argument 437 block_nr = zone->blocks; in add_rtree_block() 447 for (i = zone->levels; i < levels_needed; i++) { in add_rtree_block() 449 &zone->nodes); in add_rtree_block() 453 node->data[0] = (unsigned long)zone->rtree; in add_rtree_block() 454 zone->rtree = node; in add_rtree_block() 455 zone->levels += 1; in add_rtree_block() 459 block = alloc_rtree_node(gfp_mask, safe_needed, ca, &zone->leaves); in add_rtree_block() 464 node = zone->rtree; in add_rtree_block() [all …]
|
/Linux-v4.19/drivers/thermal/tegra/ |
D | tegra-bpmp-thermal.c | 44 struct tegra_bpmp_thermal_zone *zone = data; in tegra_bpmp_thermal_get_temp() local 52 req.get_temp.zone = zone->idx; in tegra_bpmp_thermal_get_temp() 61 err = tegra_bpmp_transfer(zone->tegra->bpmp, &msg); in tegra_bpmp_thermal_get_temp() 72 struct tegra_bpmp_thermal_zone *zone = data; in tegra_bpmp_thermal_set_trips() local 78 req.set_trip.zone = zone->idx; in tegra_bpmp_thermal_set_trips() 88 return tegra_bpmp_transfer(zone->tegra->bpmp, &msg); in tegra_bpmp_thermal_set_trips() 93 struct tegra_bpmp_thermal_zone *zone; in tz_device_update_work_fn() local 95 zone = container_of(work, struct tegra_bpmp_thermal_zone, in tz_device_update_work_fn() 98 thermal_zone_device_update(zone->tzd, THERMAL_TRIP_VIOLATED); in tz_device_update_work_fn() 118 if (tegra->zones[i]->idx != req->host_trip_reached.zone) in bpmp_mrq_thermal() [all …]
|
/Linux-v4.19/include/trace/events/ |
D | compaction.h | 201 TP_PROTO(struct zone *zone, 205 TP_ARGS(zone, order, ret), 215 __entry->nid = zone_to_nid(zone); 216 __entry->idx = zone_idx(zone); 230 TP_PROTO(struct zone *zone, 234 TP_ARGS(zone, order, ret) 239 TP_PROTO(struct zone *zone, 243 TP_ARGS(zone, order, ret) 248 TP_PROTO(struct zone *zone, int order), 250 TP_ARGS(zone, order), [all …]
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/ |
D | alloc.c | 251 struct mlx4_zone_entry *zone = kmalloc(sizeof(*zone), GFP_KERNEL); in mlx4_zone_add_one() local 253 if (NULL == zone) in mlx4_zone_add_one() 256 zone->flags = flags; in mlx4_zone_add_one() 257 zone->bitmap = bitmap; in mlx4_zone_add_one() 258 zone->use_rr = (flags & MLX4_ZONE_USE_RR) ? MLX4_USE_RR : 0; in mlx4_zone_add_one() 259 zone->priority = priority; in mlx4_zone_add_one() 260 zone->offset = offset; in mlx4_zone_add_one() 264 zone->uid = zone_alloc->last_uid++; in mlx4_zone_add_one() 265 zone->allocator = zone_alloc; in mlx4_zone_add_one() 275 list_add_tail(&zone->prio_list, &it->prio_list); in mlx4_zone_add_one() [all …]
|
/Linux-v4.19/drivers/mtd/ |
D | sm_ftl.c | 195 static loff_t sm_mkoffset(struct sm_ftl *ftl, int zone, int block, int boffset) in sm_mkoffset() argument 198 WARN_ON(zone < 0 || zone >= ftl->zone_count); in sm_mkoffset() 205 return (zone * SM_MAX_ZONE_SIZE + block) * ftl->block_size + boffset; in sm_mkoffset() 210 int *zone, int *block, int *boffset) in sm_break_offset() argument 215 *zone = offset >= ftl->zone_count ? -1 : offset; in sm_break_offset() 238 int zone, int block, int boffset, in sm_read_sector() argument 268 if (zone == 0 && block == ftl->cis_block && boffset == in sm_read_sector() 279 ret = mtd_read_oob(mtd, sm_mkoffset(ftl, zone, block, boffset), &ops); in sm_read_sector() 284 block, zone, ret); in sm_read_sector() 302 " as bad" , block, zone); in sm_read_sector() [all …]
|
/Linux-v4.19/drivers/thermal/int340x_thermal/ |
D | int340x_thermal_zone.c | 22 static int int340x_thermal_get_zone_temp(struct thermal_zone_device *zone, in int340x_thermal_get_zone_temp() argument 25 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_zone_temp() 30 return d->override_ops->get_temp(zone, temp); in int340x_thermal_get_zone_temp() 51 static int int340x_thermal_get_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_get_trip_temp() argument 54 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_temp() 58 return d->override_ops->get_trip_temp(zone, trip, temp); in int340x_thermal_get_trip_temp() 83 static int int340x_thermal_get_trip_type(struct thermal_zone_device *zone, in int340x_thermal_get_trip_type() argument 87 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_type() 91 return d->override_ops->get_trip_type(zone, trip, type); in int340x_thermal_get_trip_type() 116 static int int340x_thermal_set_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_set_trip_temp() argument [all …]
|
/Linux-v4.19/fs/adfs/ |
D | map.c | 192 scan_map(struct adfs_sb_info *asb, unsigned int zone, in scan_map() argument 199 dm = asb->s_map + zone; in scan_map() 200 zone = asb->s_map_size; in scan_map() 201 dm_end = asb->s_map + zone; in scan_map() 212 } while (--zone > 0); in scan_map() 235 unsigned int zone; in adfs_map_free() local 238 zone = asb->s_map_size; in adfs_map_free() 242 } while (--zone > 0); in adfs_map_free() 252 unsigned int zone, mapoff; in adfs_map_lookup() local 260 zone = asb->s_map_size >> 1; in adfs_map_lookup() [all …]
|
/Linux-v4.19/drivers/platform/x86/ |
D | alienware-wmi.c | 249 static int parse_rgb(const char *buf, struct platform_zone *zone) in parse_rgb() argument 269 zone->colors = repackager.cp; in parse_rgb() 275 u8 zone; in match_zone() local 277 for (zone = 0; zone < quirks->num_zones; zone++) { in match_zone() 278 if ((struct device_attribute *)zone_data[zone].attr == attr) { in match_zone() 280 zone_data[zone].location); in match_zone() 281 return &zone_data[zone]; in match_zone() 290 static int alienware_update_led(struct platform_zone *zone) in alienware_update_led() argument 299 wmax_basic_args.led_mask = 1 << zone->location; in alienware_update_led() 300 wmax_basic_args.colors = zone->colors; in alienware_update_led() [all …]
|