/Linux-v6.1/drivers/block/null_blk/ |
D | zoned.c | 35 struct nullb_zone *zone) in null_init_zone_lock() argument 38 spin_lock_init(&zone->spinlock); in null_init_zone_lock() 40 mutex_init(&zone->mutex); in null_init_zone_lock() 44 struct nullb_zone *zone) in null_lock_zone() argument 47 spin_lock_irq(&zone->spinlock); in null_lock_zone() 49 mutex_lock(&zone->mutex); in null_lock_zone() 53 struct nullb_zone *zone) in null_unlock_zone() argument 56 spin_unlock_irq(&zone->spinlock); in null_unlock_zone() 58 mutex_unlock(&zone->mutex); in null_unlock_zone() 64 struct nullb_zone *zone; in null_init_zoned_dev() local [all …]
|
/Linux-v6.1/fs/pstore/ |
D | zone.c | 160 static inline int buffer_datalen(struct pstore_zone *zone) in buffer_datalen() argument 162 return atomic_read(&zone->buffer->datalen); in buffer_datalen() 165 static inline int buffer_start(struct pstore_zone *zone) in buffer_start() argument 167 return atomic_read(&zone->buffer->start); in buffer_start() 175 static ssize_t psz_zone_read_buffer(struct pstore_zone *zone, char *buf, in psz_zone_read_buffer() argument 178 if (!buf || !zone || !zone->buffer) in psz_zone_read_buffer() 180 if (off > zone->buffer_size) in psz_zone_read_buffer() 182 len = min_t(size_t, len, zone->buffer_size - off); in psz_zone_read_buffer() 183 memcpy(buf, zone->buffer->data + off, len); in psz_zone_read_buffer() 187 static int psz_zone_read_oldbuf(struct pstore_zone *zone, char *buf, in psz_zone_read_oldbuf() argument [all …]
|
/Linux-v6.1/mm/ |
D | page_alloc.c | 639 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument 647 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries() 648 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries() 649 sp = zone->spanned_pages; in page_outside_zone_boundaries() 650 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries() 652 } while (zone_span_seqretry(zone, seq)); in page_outside_zone_boundaries() 656 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries() 662 static int page_is_consistent(struct zone *zone, struct page *page) in page_is_consistent() argument 664 if (zone != page_zone(page)) in page_is_consistent() 672 static int __maybe_unused bad_range(struct zone *zone, struct page *page) in bad_range() argument [all …]
|
D | vmstat.c | 38 static void zero_zone_numa_counters(struct zone *zone) in zero_zone_numa_counters() argument 43 atomic_long_set(&zone->vm_numa_event[item], 0); in zero_zone_numa_counters() 45 per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_event[item] in zero_zone_numa_counters() 54 struct zone *zone; in zero_zones_numa_counters() local 56 for_each_populated_zone(zone) in zero_zones_numa_counters() 57 zero_zone_numa_counters(zone); in zero_zones_numa_counters() 169 static void fold_vm_zone_numa_events(struct zone *zone) in fold_vm_zone_numa_events() argument 178 pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu); in fold_vm_zone_numa_events() 184 zone_numa_event_add(zone_numa_events[item], zone, item); in fold_vm_zone_numa_events() 189 struct zone *zone; in fold_vm_numa_events() local [all …]
|
D | compaction.c | 154 static void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 156 zone->compact_considered = 0; in defer_compaction() 157 zone->compact_defer_shift++; in defer_compaction() 159 if (order < zone->compact_order_failed) in defer_compaction() 160 zone->compact_order_failed = order; in defer_compaction() 162 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction() 163 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction() 165 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction() 169 static bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument 171 unsigned long defer_limit = 1UL << zone->compact_defer_shift; in compaction_deferred() [all …]
|
D | memory_hotplug.c | 347 static unsigned long find_smallest_section_pfn(int nid, struct zone *zone, in find_smallest_section_pfn() argument 358 if (zone != page_zone(pfn_to_page(start_pfn))) in find_smallest_section_pfn() 368 static unsigned long find_biggest_section_pfn(int nid, struct zone *zone, in find_biggest_section_pfn() argument 383 if (zone != page_zone(pfn_to_page(pfn))) in find_biggest_section_pfn() 392 static void shrink_zone_span(struct zone *zone, unsigned long start_pfn, in shrink_zone_span() argument 396 int nid = zone_to_nid(zone); in shrink_zone_span() 398 if (zone->zone_start_pfn == start_pfn) { in shrink_zone_span() 405 pfn = find_smallest_section_pfn(nid, zone, end_pfn, in shrink_zone_span() 406 zone_end_pfn(zone)); in shrink_zone_span() 408 zone->spanned_pages = zone_end_pfn(zone) - pfn; in shrink_zone_span() [all …]
|
D | page_isolation.c | 37 struct zone *zone = page_zone(page); in has_unmovable_pages() local 72 if (zone_idx(zone) == ZONE_MOVABLE) in has_unmovable_pages() 150 struct zone *zone = page_zone(page); in set_migratetype_isolate() local 155 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 163 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 185 zone->nr_isolate_pageblock++; in set_migratetype_isolate() 186 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE, in set_migratetype_isolate() 189 __mod_zone_freepage_state(zone, -nr_pages, mt); in set_migratetype_isolate() 190 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 194 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() [all …]
|
D | page_reporting.c | 114 page_reporting_cycle(struct page_reporting_dev_info *prdev, struct zone *zone, in page_reporting_cycle() argument 118 struct free_area *area = &zone->free_area[order]; in page_reporting_cycle() 132 spin_lock_irq(&zone->lock); in page_reporting_cycle() 190 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 202 spin_lock_irq(&zone->lock); in page_reporting_cycle() 222 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 229 struct scatterlist *sgl, struct zone *zone) in page_reporting_process_zone() argument 236 watermark = low_wmark_pages(zone) + in page_reporting_process_zone() 243 if (!zone_watermark_ok(zone, 0, watermark, 0, ALLOC_CMA)) in page_reporting_process_zone() 253 err = page_reporting_cycle(prdev, zone, order, mt, in page_reporting_process_zone() [all …]
|
D | mmzone.c | 30 struct zone *next_zone(struct zone *zone) in next_zone() argument 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 35 zone++; in next_zone() 39 zone = pgdat->node_zones; in next_zone() 41 zone = NULL; in next_zone() 43 return zone; in next_zone() 69 (z->zone && !zref_in_nodemask(z, nodes))) in __next_zones_zonelist()
|
/Linux-v6.1/include/linux/ |
D | memory_hotplug.h | 11 struct zone; 130 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument 132 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin() 134 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument 136 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry() 138 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument 140 write_seqlock(&zone->span_seqlock); in zone_span_writelock() 142 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument 144 write_sequnlock(&zone->span_seqlock); in zone_span_writeunlock() 146 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument [all …]
|
D | mmzone.h | 705 struct zone { struct 893 static inline unsigned long zone_managed_pages(struct zone *zone) in zone_managed_pages() argument 895 return (unsigned long)atomic_long_read(&zone->managed_pages); in zone_managed_pages() 898 static inline unsigned long zone_cma_pages(struct zone *zone) in zone_cma_pages() argument 901 return zone->cma_pages; in zone_cma_pages() 907 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() argument 909 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn() 912 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument 914 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 917 static inline bool zone_is_initialized(struct zone *zone) in zone_is_initialized() argument [all …]
|
D | vmstat.h | 139 static inline void zone_numa_event_add(long x, struct zone *zone, in zone_numa_event_add() argument 142 atomic_long_add(x, &zone->vm_numa_event[item]); in zone_numa_event_add() 146 static inline unsigned long zone_numa_event_state(struct zone *zone, in zone_numa_event_state() argument 149 return atomic_long_read(&zone->vm_numa_event[item]); in zone_numa_event_state() 159 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument 162 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add() 201 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument 204 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state() 218 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument 221 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state_snapshot() [all …]
|
/Linux-v6.1/drivers/md/ |
D | dm-zoned-metadata.c | 137 struct dm_zone *zone; member 221 static unsigned int dmz_dev_zone_id(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_dev_zone_id() argument 223 if (WARN_ON(!zone)) in dmz_dev_zone_id() 226 return zone->id - zone->dev->zone_offset; in dmz_dev_zone_id() 229 sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_sect() argument 231 unsigned int zone_id = dmz_dev_zone_id(zmd, zone); in dmz_start_sect() 236 sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_block() argument 238 unsigned int zone_id = dmz_dev_zone_id(zmd, zone); in dmz_start_block() 311 struct dm_zone *zone = kzalloc(sizeof(struct dm_zone), GFP_KERNEL); in dmz_insert() local 313 if (!zone) in dmz_insert() [all …]
|
D | dm-zoned-target.c | 21 struct dm_zone *zone; member 86 struct dm_zone *zone = bioctx->zone; in dmz_bio_endio() local 88 if (zone) { in dmz_bio_endio() 91 dmz_is_seq(zone)) in dmz_bio_endio() 92 set_bit(DMZ_SEQ_WRITE_ERR, &zone->flags); in dmz_bio_endio() 93 dmz_deactivate_zone(zone); in dmz_bio_endio() 116 static int dmz_submit_bio(struct dmz_target *dmz, struct dm_zone *zone, in dmz_submit_bio() argument 122 struct dmz_dev *dev = zone->dev; in dmz_submit_bio() 134 dmz_start_sect(dmz->metadata, zone) + dmz_blk2sect(chunk_block); in dmz_submit_bio() 144 if (bio_op(bio) == REQ_OP_WRITE && dmz_is_seq(zone)) in dmz_submit_bio() [all …]
|
/Linux-v6.1/include/net/netfilter/ |
D | nf_conntrack_zones.h | 12 return &ct->zone; in nf_ct_zone() 19 nf_ct_zone_init(struct nf_conntrack_zone *zone, u16 id, u8 dir, u8 flags) in nf_ct_zone_init() argument 21 zone->id = id; in nf_ct_zone_init() 22 zone->flags = flags; in nf_ct_zone_init() 23 zone->dir = dir; in nf_ct_zone_init() 25 return zone; in nf_ct_zone_init() 36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 37 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl() 43 const struct nf_conntrack_zone *zone) in nf_ct_zone_add() argument 46 ct->zone = *zone; in nf_ct_zone_add() [all …]
|
/Linux-v6.1/kernel/power/ |
D | snapshot.c | 404 struct mem_zone_bm_rtree *zone; member 465 static int add_rtree_block(struct mem_zone_bm_rtree *zone, gfp_t gfp_mask, in add_rtree_block() argument 472 block_nr = zone->blocks; in add_rtree_block() 482 for (i = zone->levels; i < levels_needed; i++) { in add_rtree_block() 484 &zone->nodes); in add_rtree_block() 488 node->data[0] = (unsigned long)zone->rtree; in add_rtree_block() 489 zone->rtree = node; in add_rtree_block() 490 zone->levels += 1; in add_rtree_block() 494 block = alloc_rtree_node(gfp_mask, safe_needed, ca, &zone->leaves); in add_rtree_block() 499 node = zone->rtree; in add_rtree_block() [all …]
|
/Linux-v6.1/fs/adfs/ |
D | map.c | 159 static int scan_map(struct adfs_sb_info *asb, unsigned int zone, in scan_map() argument 166 dm = asb->s_map + zone; in scan_map() 167 zone = asb->s_map_size; in scan_map() 168 dm_end = asb->s_map + zone; in scan_map() 179 } while (--zone > 0); in scan_map() 202 unsigned int zone; in adfs_map_statfs() local 205 zone = asb->s_map_size; in adfs_map_statfs() 209 } while (--zone > 0); in adfs_map_statfs() 220 unsigned int zone, mapoff; in adfs_map_lookup() local 228 zone = asb->s_map_size >> 1; in adfs_map_lookup() [all …]
|
/Linux-v6.1/drivers/thermal/tegra/ |
D | tegra-bpmp-thermal.c | 33 static int __tegra_bpmp_thermal_get_temp(struct tegra_bpmp_thermal_zone *zone, in __tegra_bpmp_thermal_get_temp() argument 43 req.get_temp.zone = zone->idx; in __tegra_bpmp_thermal_get_temp() 52 err = tegra_bpmp_transfer(zone->tegra->bpmp, &msg); in __tegra_bpmp_thermal_get_temp() 70 struct tegra_bpmp_thermal_zone *zone = tz->devdata; in tegra_bpmp_thermal_set_trips() local 77 req.set_trip.zone = zone->idx; in tegra_bpmp_thermal_set_trips() 87 err = tegra_bpmp_transfer(zone->tegra->bpmp, &msg); in tegra_bpmp_thermal_set_trips() 98 struct tegra_bpmp_thermal_zone *zone; in tz_device_update_work_fn() local 100 zone = container_of(work, struct tegra_bpmp_thermal_zone, in tz_device_update_work_fn() 103 thermal_zone_device_update(zone->tzd, THERMAL_TRIP_VIOLATED); in tz_device_update_work_fn() 123 if (tegra->zones[i]->idx != req->host_trip_reached.zone) in bpmp_mrq_thermal() [all …]
|
/Linux-v6.1/include/trace/events/ |
D | compaction.h | 183 TP_PROTO(struct zone *zone, 187 TP_ARGS(zone, order, ret), 197 __entry->nid = zone_to_nid(zone); 198 __entry->idx = zone_idx(zone); 212 TP_PROTO(struct zone *zone, 216 TP_ARGS(zone, order, ret) 221 TP_PROTO(struct zone *zone, 225 TP_ARGS(zone, order, ret) 230 TP_PROTO(struct zone *zone, int order), 232 TP_ARGS(zone, order), [all …]
|
/Linux-v6.1/virt/kvm/ |
D | coalesced_mmio.c | 36 if (addr < dev->zone.addr) in coalesced_mmio_in_range() 38 if (addr + len > dev->zone.addr + dev->zone.size) in coalesced_mmio_in_range() 89 ring->coalesced_mmio[insert].pio = dev->zone.pio; in coalesced_mmio_write() 138 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_register_coalesced_mmio() argument 143 if (zone->pio != 1 && zone->pio != 0) in kvm_vm_ioctl_register_coalesced_mmio() 153 dev->zone = *zone; in kvm_vm_ioctl_register_coalesced_mmio() 157 zone->pio ? KVM_PIO_BUS : KVM_MMIO_BUS, in kvm_vm_ioctl_register_coalesced_mmio() 158 zone->addr, zone->size, &dev->dev); in kvm_vm_ioctl_register_coalesced_mmio() 174 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_unregister_coalesced_mmio() argument 179 if (zone->pio != 1 && zone->pio != 0) in kvm_vm_ioctl_unregister_coalesced_mmio() [all …]
|
/Linux-v6.1/drivers/net/ethernet/mellanox/mlx4/ |
D | alloc.c | 250 struct mlx4_zone_entry *zone = kmalloc(sizeof(*zone), GFP_KERNEL); in mlx4_zone_add_one() local 252 if (NULL == zone) in mlx4_zone_add_one() 255 zone->flags = flags; in mlx4_zone_add_one() 256 zone->bitmap = bitmap; in mlx4_zone_add_one() 257 zone->use_rr = (flags & MLX4_ZONE_USE_RR) ? MLX4_USE_RR : 0; in mlx4_zone_add_one() 258 zone->priority = priority; in mlx4_zone_add_one() 259 zone->offset = offset; in mlx4_zone_add_one() 263 zone->uid = zone_alloc->last_uid++; in mlx4_zone_add_one() 264 zone->allocator = zone_alloc; in mlx4_zone_add_one() 274 list_add_tail(&zone->prio_list, &it->prio_list); in mlx4_zone_add_one() [all …]
|
/Linux-v6.1/drivers/mtd/ |
D | sm_ftl.c | 192 static loff_t sm_mkoffset(struct sm_ftl *ftl, int zone, int block, int boffset) in sm_mkoffset() argument 195 WARN_ON(zone < 0 || zone >= ftl->zone_count); in sm_mkoffset() 202 return (zone * SM_MAX_ZONE_SIZE + block) * ftl->block_size + boffset; in sm_mkoffset() 207 int *zone, int *block, int *boffset) in sm_break_offset() argument 212 *zone = offset >= ftl->zone_count ? -1 : offset; in sm_break_offset() 238 int zone, int block, int boffset, in sm_read_sector() argument 270 if (zone == 0 && block == ftl->cis_block && boffset == in sm_read_sector() 282 ret = mtd_read_oob(mtd, sm_mkoffset(ftl, zone, block, boffset), &ops); in sm_read_sector() 287 block, zone, ret); in sm_read_sector() 305 " as bad" , block, zone); in sm_read_sector() [all …]
|
/Linux-v6.1/drivers/thermal/intel/int340x_thermal/ |
D | int340x_thermal_zone.c | 14 static int int340x_thermal_get_zone_temp(struct thermal_zone_device *zone, in int340x_thermal_get_zone_temp() argument 17 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_zone_temp() 22 return d->override_ops->get_temp(zone, temp); in int340x_thermal_get_zone_temp() 43 static int int340x_thermal_get_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_get_trip_temp() argument 46 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_temp() 50 return d->override_ops->get_trip_temp(zone, trip, temp); in int340x_thermal_get_trip_temp() 75 static int int340x_thermal_get_trip_type(struct thermal_zone_device *zone, in int340x_thermal_get_trip_type() argument 79 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_type() 83 return d->override_ops->get_trip_type(zone, trip, type); in int340x_thermal_get_trip_type() 108 static int int340x_thermal_set_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_set_trip_temp() argument [all …]
|
/Linux-v6.1/arch/x86/mm/ |
D | highmem_32.c | 9 struct zone *zone; in set_highmem_pages_init() local 17 for_each_zone(zone) { in set_highmem_pages_init() 20 if (!is_highmem(zone)) in set_highmem_pages_init() 23 zone_start_pfn = zone->zone_start_pfn; in set_highmem_pages_init() 24 zone_end_pfn = zone_start_pfn + zone->spanned_pages; in set_highmem_pages_init() 26 nid = zone_to_nid(zone); in set_highmem_pages_init() 28 zone->name, nid, zone_start_pfn, zone_end_pfn); in set_highmem_pages_init()
|
/Linux-v6.1/Documentation/filesystems/ |
D | zonefs.rst | 10 zonefs is a very simple file system exposing each zone of a zoned block device 24 by allowing SSTables to be stored in a zone file similarly to a regular file 26 of the higher level construct "one file is one zone" can help reducing the 34 space that is divided into zones. A zone is a group of consecutive LBAs and all 41 sequentially. Each sequential zone has a write pointer maintained by the 43 to the device. As a result of this write constraint, LBAs in a sequential zone 45 command (zone reset) before rewriting. 61 representing zones are grouped by zone type, which are themselves represented 62 by sub-directories. This file structure is built entirely using zone information 71 mount, zonefs uses blkdev_report_zones() to obtain the device zone configuration [all …]
|