Lines Matching full:zone

33  * blocks indicating zone block validity.
39 * the first conventional zone found on disk.
87 * and give the zone ID (dzone_id) mapping the chunk on disk.
88 * This zone may be sequential or random. If it is a sequential
89 * zone, a second zone (bzone_id) used as a write buffer may
90 * also be specified. This second zone will always be a randomly
91 * writeable zone.
137 struct dm_zone *zone; member
174 /* Zone information array */
192 /* Zone allocation management */
221 static unsigned int dmz_dev_zone_id(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_dev_zone_id() argument
223 if (WARN_ON(!zone)) in dmz_dev_zone_id()
226 return zone->id - zone->dev->zone_offset; in dmz_dev_zone_id()
229 sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_sect() argument
231 unsigned int zone_id = dmz_dev_zone_id(zmd, zone); in dmz_start_sect()
236 sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_block() argument
238 unsigned int zone_id = dmz_dev_zone_id(zmd, zone); in dmz_start_block()
311 struct dm_zone *zone = kzalloc(sizeof(struct dm_zone), GFP_KERNEL); in dmz_insert() local
313 if (!zone) in dmz_insert()
316 if (xa_insert(&zmd->zones, zone_id, zone, GFP_KERNEL)) { in dmz_insert()
317 kfree(zone); in dmz_insert()
321 INIT_LIST_HEAD(&zone->link); in dmz_insert()
322 atomic_set(&zone->refcount, 0); in dmz_insert()
323 zone->id = zone_id; in dmz_insert()
324 zone->chunk = DMZ_MAP_UNMAPPED; in dmz_insert()
325 zone->dev = dev; in dmz_insert()
327 return zone; in dmz_insert()
359 * The map lock also protects all the zone lists.
375 * the map lock and zone state management (active and reclaim state are
807 sb_block = zmd->sb[set].zone->id << zmd->zone_nr_blocks_shift; in dmz_write_sb()
1030 if (sb_block != (u64)dsb->zone->id << zmd->zone_nr_blocks_shift ) { in dmz_check_sb()
1034 (u64)dsb->zone->id << zmd->zone_nr_blocks_shift); in dmz_check_sb()
1134 unsigned int zone_id = zmd->sb[0].zone->id; in dmz_lookup_secondary_sb()
1147 zmd->sb[1].zone = dmz_get(zmd, zone_id + 1); in dmz_lookup_secondary_sb()
1155 zmd->sb[1].zone = dmz_get(zmd, zone_id + i); in dmz_lookup_secondary_sb()
1160 zmd->sb[1].zone = NULL; in dmz_lookup_secondary_sb()
1206 zmd->sb[0].block = dmz_start_block(zmd, zmd->sb[0].zone); in dmz_recover_mblocks()
1208 zmd->sb[1].block = dmz_start_block(zmd, zmd->sb[1].zone); in dmz_recover_mblocks()
1252 if (!zmd->sb[0].zone) { in dmz_load_sb()
1253 dmz_zmd_err(zmd, "Primary super block zone not set"); in dmz_load_sb()
1258 zmd->sb[0].block = dmz_start_block(zmd, zmd->sb[0].zone); in dmz_load_sb()
1259 zmd->sb[0].dev = zmd->sb[0].zone->dev; in dmz_load_sb()
1271 if (!zmd->sb[1].zone) { in dmz_load_sb()
1273 zmd->sb[0].zone->id + zmd->nr_meta_zones; in dmz_load_sb()
1275 zmd->sb[1].zone = dmz_get(zmd, zone_id); in dmz_load_sb()
1277 zmd->sb[1].block = dmz_start_block(zmd, zmd->sb[1].zone); in dmz_load_sb()
1342 sb->zone = dmz_get(zmd, zmd->dev[i].zone_offset); in dmz_load_sb()
1344 if (!dmz_is_meta(sb->zone)) { in dmz_load_sb()
1346 "Tertiary super block zone %u not marked as metadata zone", in dmz_load_sb()
1347 sb->zone->id); in dmz_load_sb()
1370 * Initialize a zone descriptor.
1377 struct dm_zone *zone; in dmz_init_zone() local
1379 zone = dmz_insert(zmd, idx, dev); in dmz_init_zone()
1380 if (IS_ERR(zone)) in dmz_init_zone()
1381 return PTR_ERR(zone); in dmz_init_zone()
1385 /* Ignore the eventual runt (smaller) zone */ in dmz_init_zone()
1386 set_bit(DMZ_OFFLINE, &zone->flags); in dmz_init_zone()
1394 * Devices that have zones with a capacity smaller than the zone size in dmz_init_zone()
1402 set_bit(DMZ_RND, &zone->flags); in dmz_init_zone()
1406 set_bit(DMZ_SEQ, &zone->flags); in dmz_init_zone()
1412 if (dmz_is_rnd(zone)) in dmz_init_zone()
1413 zone->wp_block = 0; in dmz_init_zone()
1415 zone->wp_block = dmz_sect2blk(blkz->wp - blkz->start); in dmz_init_zone()
1418 set_bit(DMZ_OFFLINE, &zone->flags); in dmz_init_zone()
1420 set_bit(DMZ_READ_ONLY, &zone->flags); in dmz_init_zone()
1423 if (dmz_is_rnd(zone)) { in dmz_init_zone()
1425 if (zmd->nr_devs == 1 && !zmd->sb[0].zone) { in dmz_init_zone()
1426 /* Primary super block zone */ in dmz_init_zone()
1427 zmd->sb[0].zone = zone; in dmz_init_zone()
1434 * as metadata zone. in dmz_init_zone()
1436 set_bit(DMZ_META, &zone->flags); in dmz_init_zone()
1448 struct dm_zone *zone; in dmz_emulate_zones() local
1450 zone = dmz_insert(zmd, idx, dev); in dmz_emulate_zones()
1451 if (IS_ERR(zone)) in dmz_emulate_zones()
1452 return PTR_ERR(zone); in dmz_emulate_zones()
1453 set_bit(DMZ_CACHE, &zone->flags); in dmz_emulate_zones()
1454 zone->wp_block = 0; in dmz_emulate_zones()
1458 /* Disable runt zone */ in dmz_emulate_zones()
1459 set_bit(DMZ_OFFLINE, &zone->flags); in dmz_emulate_zones()
1475 struct dm_zone *zone = xa_load(&zmd->zones, idx); in dmz_drop_zones() local
1477 kfree(zone); in dmz_drop_zones()
1484 * Allocate and initialize zone descriptors using the zone
1503 /* Allocate zone array */ in dmz_init_zones()
1526 DMDEBUG("(%s): Using %zu B for zone information", in dmz_init_zones()
1539 * Primary superblock zone is always at zone 0 when multiple in dmz_init_zones()
1542 zmd->sb[0].zone = dmz_get(zmd, 0); in dmz_init_zones()
1561 * Get zone information and initialize zone descriptors. At the same in dmz_init_zones()
1563 * first randomly writable zone. in dmz_init_zones()
1580 struct dm_zone *zone = data; in dmz_update_zone_cb() local
1582 clear_bit(DMZ_OFFLINE, &zone->flags); in dmz_update_zone_cb()
1583 clear_bit(DMZ_READ_ONLY, &zone->flags); in dmz_update_zone_cb()
1585 set_bit(DMZ_OFFLINE, &zone->flags); in dmz_update_zone_cb()
1587 set_bit(DMZ_READ_ONLY, &zone->flags); in dmz_update_zone_cb()
1589 if (dmz_is_seq(zone)) in dmz_update_zone_cb()
1590 zone->wp_block = dmz_sect2blk(blkz->wp - blkz->start); in dmz_update_zone_cb()
1592 zone->wp_block = 0; in dmz_update_zone_cb()
1597 * Update a zone information.
1599 static int dmz_update_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_update_zone() argument
1601 struct dmz_dev *dev = zone->dev; in dmz_update_zone()
1609 * Get zone information from disk. Since blkdev_report_zones() uses in dmz_update_zone()
1615 ret = blkdev_report_zones(dev->bdev, dmz_start_sect(zmd, zone), 1, in dmz_update_zone()
1616 dmz_update_zone_cb, zone); in dmz_update_zone()
1622 dmz_dev_err(dev, "Get zone %u report failed", in dmz_update_zone()
1623 zone->id); in dmz_update_zone()
1632 * Check a zone write pointer position when the zone is marked
1636 struct dm_zone *zone) in dmz_handle_seq_write_err() argument
1638 struct dmz_dev *dev = zone->dev; in dmz_handle_seq_write_err()
1642 wp = zone->wp_block; in dmz_handle_seq_write_err()
1643 ret = dmz_update_zone(zmd, zone); in dmz_handle_seq_write_err()
1647 dmz_dev_warn(dev, "Processing zone %u write error (zone wp %u/%u)", in dmz_handle_seq_write_err()
1648 zone->id, zone->wp_block, wp); in dmz_handle_seq_write_err()
1650 if (zone->wp_block < wp) { in dmz_handle_seq_write_err()
1651 dmz_invalidate_blocks(zmd, zone, zone->wp_block, in dmz_handle_seq_write_err()
1652 wp - zone->wp_block); in dmz_handle_seq_write_err()
1659 * Reset a zone write pointer.
1661 static int dmz_reset_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_reset_zone() argument
1669 if (dmz_is_offline(zone) || in dmz_reset_zone()
1670 dmz_is_readonly(zone) || in dmz_reset_zone()
1671 dmz_is_rnd(zone)) in dmz_reset_zone()
1674 if (!dmz_is_empty(zone) || dmz_seq_write_err(zone)) { in dmz_reset_zone()
1675 struct dmz_dev *dev = zone->dev; in dmz_reset_zone()
1678 dmz_start_sect(zmd, zone), in dmz_reset_zone()
1681 dmz_dev_err(dev, "Reset zone %u failed %d", in dmz_reset_zone()
1682 zone->id, ret); in dmz_reset_zone()
1688 clear_bit(DMZ_SEQ_WRITE_ERR, &zone->flags); in dmz_reset_zone()
1689 zone->wp_block = 0; in dmz_reset_zone()
1694 static void dmz_get_zone_weight(struct dmz_metadata *zmd, struct dm_zone *zone);
1714 /* Get chunk mapping table blocks and initialize zone mapping */ in dmz_load_mapping()
1727 /* Check data zone */ in dmz_load_mapping()
1733 dmz_zmd_err(zmd, "Chunk %u mapping: invalid data zone ID %u", in dmz_load_mapping()
1740 dmz_zmd_err(zmd, "Chunk %u mapping: data zone %u not present", in dmz_load_mapping()
1755 /* Check buffer zone */ in dmz_load_mapping()
1761 dmz_zmd_err(zmd, "Chunk %u mapping: invalid buffer zone ID %u", in dmz_load_mapping()
1768 dmz_zmd_err(zmd, "Chunk %u mapping: buffer zone %u not present", in dmz_load_mapping()
1773 dmz_zmd_err(zmd, "Chunk %u mapping: invalid buffer zone %u", in dmz_load_mapping()
1821 /* Unmapped data zone */ in dmz_load_mapping()
1863 * This rotates a zone at the end of its map list.
1865 static void __dmz_lru_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in __dmz_lru_zone() argument
1867 if (list_empty(&zone->link)) in __dmz_lru_zone()
1870 list_del_init(&zone->link); in __dmz_lru_zone()
1871 if (dmz_is_seq(zone)) { in __dmz_lru_zone()
1872 /* LRU rotate sequential zone */ in __dmz_lru_zone()
1873 list_add_tail(&zone->link, &zone->dev->map_seq_list); in __dmz_lru_zone()
1874 } else if (dmz_is_cache(zone)) { in __dmz_lru_zone()
1875 /* LRU rotate cache zone */ in __dmz_lru_zone()
1876 list_add_tail(&zone->link, &zmd->map_cache_list); in __dmz_lru_zone()
1878 /* LRU rotate random zone */ in __dmz_lru_zone()
1879 list_add_tail(&zone->link, &zone->dev->map_rnd_list); in __dmz_lru_zone()
1885 * in LRU order. This rotates a zone at the end of the list.
1887 static void dmz_lru_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_lru_zone() argument
1889 __dmz_lru_zone(zmd, zone); in dmz_lru_zone()
1890 if (zone->bzone) in dmz_lru_zone()
1891 __dmz_lru_zone(zmd, zone->bzone); in dmz_lru_zone()
1895 * Wait for any zone to be freed.
1913 * Lock a zone for reclaim (set the zone RECLAIM bit).
1914 * Returns false if the zone cannot be locked or if it is already locked
1917 int dmz_lock_zone_reclaim(struct dm_zone *zone) in dmz_lock_zone_reclaim() argument
1920 if (dmz_is_active(zone)) in dmz_lock_zone_reclaim()
1923 return !test_and_set_bit(DMZ_RECLAIM, &zone->flags); in dmz_lock_zone_reclaim()
1927 * Clear a zone reclaim flag.
1929 void dmz_unlock_zone_reclaim(struct dm_zone *zone) in dmz_unlock_zone_reclaim() argument
1931 WARN_ON(dmz_is_active(zone)); in dmz_unlock_zone_reclaim()
1932 WARN_ON(!dmz_in_reclaim(zone)); in dmz_unlock_zone_reclaim()
1934 clear_bit_unlock(DMZ_RECLAIM, &zone->flags); in dmz_unlock_zone_reclaim()
1936 wake_up_bit(&zone->flags, DMZ_RECLAIM); in dmz_unlock_zone_reclaim()
1940 * Wait for a zone reclaim to complete.
1942 static void dmz_wait_for_reclaim(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_wait_for_reclaim() argument
1946 set_bit(DMZ_RECLAIM_TERMINATE, &zone->flags); in dmz_wait_for_reclaim()
1947 wait_on_bit_timeout(&zone->flags, DMZ_RECLAIM, TASK_UNINTERRUPTIBLE, HZ); in dmz_wait_for_reclaim()
1948 clear_bit(DMZ_RECLAIM_TERMINATE, &zone->flags); in dmz_wait_for_reclaim()
1954 * Select a cache or random write zone for reclaim.
1960 struct dm_zone *zone, *maxw_z = NULL; in dmz_get_rnd_zone_for_reclaim() local
1963 /* If we have cache zones select from the cache zone list */ in dmz_get_rnd_zone_for_reclaim()
1973 * Find the buffer zone with the heaviest weight or the first (oldest) in dmz_get_rnd_zone_for_reclaim()
1974 * data zone that can be reclaimed. in dmz_get_rnd_zone_for_reclaim()
1976 list_for_each_entry(zone, zone_list, link) { in dmz_get_rnd_zone_for_reclaim()
1977 if (dmz_is_buf(zone)) { in dmz_get_rnd_zone_for_reclaim()
1978 dzone = zone->bzone; in dmz_get_rnd_zone_for_reclaim()
1984 dzone = zone; in dmz_get_rnd_zone_for_reclaim()
1996 * first zone that can be reclaimed regardless of its weitght. in dmz_get_rnd_zone_for_reclaim()
1998 list_for_each_entry(zone, zone_list, link) { in dmz_get_rnd_zone_for_reclaim()
1999 if (dmz_is_buf(zone)) { in dmz_get_rnd_zone_for_reclaim()
2000 dzone = zone->bzone; in dmz_get_rnd_zone_for_reclaim()
2004 dzone = zone; in dmz_get_rnd_zone_for_reclaim()
2013 * Select a buffered sequential zone for reclaim.
2018 struct dm_zone *zone; in dmz_get_seq_zone_for_reclaim() local
2020 list_for_each_entry(zone, &zmd->dev[idx].map_seq_list, link) { in dmz_get_seq_zone_for_reclaim()
2021 if (!zone->bzone) in dmz_get_seq_zone_for_reclaim()
2023 if (dmz_lock_zone_reclaim(zone)) in dmz_get_seq_zone_for_reclaim()
2024 return zone; in dmz_get_seq_zone_for_reclaim()
2031 * Select a zone for reclaim.
2036 struct dm_zone *zone = NULL; in dmz_get_zone_for_reclaim() local
2039 * Search for a zone candidate to reclaim: 2 cases are possible. in dmz_get_zone_for_reclaim()
2040 * (1) There is no free sequential zones. Then a random data zone in dmz_get_zone_for_reclaim()
2041 * cannot be reclaimed. So choose a sequential zone to reclaim so in dmz_get_zone_for_reclaim()
2042 * that afterward a random zone can be reclaimed. in dmz_get_zone_for_reclaim()
2043 * (2) At least one free sequential zone is available, then choose in dmz_get_zone_for_reclaim()
2044 * the oldest random zone (data or buffer) that can be locked. in dmz_get_zone_for_reclaim()
2048 zone = dmz_get_seq_zone_for_reclaim(zmd, dev_idx); in dmz_get_zone_for_reclaim()
2049 if (!zone) in dmz_get_zone_for_reclaim()
2050 zone = dmz_get_rnd_zone_for_reclaim(zmd, dev_idx, idle); in dmz_get_zone_for_reclaim()
2053 return zone; in dmz_get_zone_for_reclaim()
2057 * Get the zone mapping a chunk, if the chunk is mapped already.
2058 * If no mapping exist and the operation is WRITE, a zone is
2060 * The zone returned will be set to the active state.
2084 /* Allocate a random zone */ in dmz_get_chunk_mapping()
2098 /* The chunk is already mapped: get the mapping zone */ in dmz_get_chunk_mapping()
2121 * If the zone is being reclaimed, the chunk mapping may change in dmz_get_chunk_mapping()
2122 * to a different zone. So wait for reclaim and retry. Otherwise, in dmz_get_chunk_mapping()
2123 * activate the zone (this will prevent reclaim from touching it). in dmz_get_chunk_mapping()
2154 /* Empty buffer zone: reclaim it */ in dmz_put_chunk_mapping()
2161 /* Deactivate the data zone */ in dmz_put_chunk_mapping()
2166 /* Unbuffered inactive empty data zone: reclaim it */ in dmz_put_chunk_mapping()
2175 * Allocate and map a random zone to buffer a chunk
2176 * already mapped to a sequential zone.
2190 /* Allocate a random zone */ in dmz_get_chunk_buffer()
2219 * Get an unmapped (free) zone.
2226 struct dm_zone *zone; in dmz_alloc_zone() local
2246 * No free zone: return NULL if this is for not reclaim. in dmz_alloc_zone()
2262 zone = list_first_entry_or_null(&zmd->reserved_seq_zones_list, in dmz_alloc_zone()
2264 if (zone) { in dmz_alloc_zone()
2265 list_del_init(&zone->link); in dmz_alloc_zone()
2268 return zone; in dmz_alloc_zone()
2271 zone = list_first_entry(list, struct dm_zone, link); in dmz_alloc_zone()
2272 list_del_init(&zone->link); in dmz_alloc_zone()
2274 if (dmz_is_cache(zone)) in dmz_alloc_zone()
2276 else if (dmz_is_rnd(zone)) in dmz_alloc_zone()
2277 atomic_dec(&zone->dev->unmap_nr_rnd); in dmz_alloc_zone()
2279 atomic_dec(&zone->dev->unmap_nr_seq); in dmz_alloc_zone()
2281 if (dmz_is_offline(zone)) { in dmz_alloc_zone()
2282 dmz_zmd_warn(zmd, "Zone %u is offline", zone->id); in dmz_alloc_zone()
2283 zone = NULL; in dmz_alloc_zone()
2286 if (dmz_is_meta(zone)) { in dmz_alloc_zone()
2287 dmz_zmd_warn(zmd, "Zone %u has metadata", zone->id); in dmz_alloc_zone()
2288 zone = NULL; in dmz_alloc_zone()
2291 return zone; in dmz_alloc_zone()
2295 * Free a zone.
2298 void dmz_free_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_free_zone() argument
2300 /* If this is a sequential zone, reset it */ in dmz_free_zone()
2301 if (dmz_is_seq(zone)) in dmz_free_zone()
2302 dmz_reset_zone(zmd, zone); in dmz_free_zone()
2304 /* Return the zone to its type unmap list */ in dmz_free_zone()
2305 if (dmz_is_cache(zone)) { in dmz_free_zone()
2306 list_add_tail(&zone->link, &zmd->unmap_cache_list); in dmz_free_zone()
2308 } else if (dmz_is_rnd(zone)) { in dmz_free_zone()
2309 list_add_tail(&zone->link, &zone->dev->unmap_rnd_list); in dmz_free_zone()
2310 atomic_inc(&zone->dev->unmap_nr_rnd); in dmz_free_zone()
2311 } else if (dmz_is_reserved(zone)) { in dmz_free_zone()
2312 list_add_tail(&zone->link, &zmd->reserved_seq_zones_list); in dmz_free_zone()
2315 list_add_tail(&zone->link, &zone->dev->unmap_seq_list); in dmz_free_zone()
2316 atomic_inc(&zone->dev->unmap_nr_seq); in dmz_free_zone()
2323 * Map a chunk to a zone.
2342 * Unmap a zone.
2345 void dmz_unmap_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_unmap_zone() argument
2347 unsigned int chunk = zone->chunk; in dmz_unmap_zone()
2355 if (test_and_clear_bit(DMZ_BUF, &zone->flags)) { in dmz_unmap_zone()
2357 * Unmapping the chunk buffer zone: clear only in dmz_unmap_zone()
2360 dzone_id = zone->bzone->id; in dmz_unmap_zone()
2361 zone->bzone->bzone = NULL; in dmz_unmap_zone()
2362 zone->bzone = NULL; in dmz_unmap_zone()
2366 * Unmapping the chunk data zone: the zone must in dmz_unmap_zone()
2369 if (WARN_ON(zone->bzone)) { in dmz_unmap_zone()
2370 zone->bzone->bzone = NULL; in dmz_unmap_zone()
2371 zone->bzone = NULL; in dmz_unmap_zone()
2378 zone->chunk = DMZ_MAP_UNMAPPED; in dmz_unmap_zone()
2379 list_del_init(&zone->link); in dmz_unmap_zone()
2415 * Get the bitmap block storing the bit for chunk_block in zone.
2418 struct dm_zone *zone, in dmz_get_bitmap() argument
2422 (sector_t)(zone->id * zmd->zone_nr_bitmap_blocks) + in dmz_get_bitmap()
2474 /* Get a valid region from the source zone */ in dmz_merge_valid_blocks()
2493 int dmz_validate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_validate_blocks() argument
2501 dmz_zmd_debug(zmd, "=> VALIDATE zone %u, block %llu, %u blocks", in dmz_validate_blocks()
2502 zone->id, (unsigned long long)chunk_block, in dmz_validate_blocks()
2509 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_validate_blocks()
2528 if (likely(zone->weight + n <= zone_nr_blocks)) in dmz_validate_blocks()
2529 zone->weight += n; in dmz_validate_blocks()
2531 dmz_zmd_warn(zmd, "Zone %u: weight %u should be <= %u", in dmz_validate_blocks()
2532 zone->id, zone->weight, in dmz_validate_blocks()
2534 zone->weight = zone_nr_blocks; in dmz_validate_blocks()
2574 int dmz_invalidate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_invalidate_blocks() argument
2581 dmz_zmd_debug(zmd, "=> INVALIDATE zone %u, block %llu, %u blocks", in dmz_invalidate_blocks()
2582 zone->id, (u64)chunk_block, nr_blocks); in dmz_invalidate_blocks()
2588 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_invalidate_blocks()
2608 if (zone->weight >= n) in dmz_invalidate_blocks()
2609 zone->weight -= n; in dmz_invalidate_blocks()
2611 dmz_zmd_warn(zmd, "Zone %u: weight %u should be >= %u", in dmz_invalidate_blocks()
2612 zone->id, zone->weight, n); in dmz_invalidate_blocks()
2613 zone->weight = 0; in dmz_invalidate_blocks()
2622 static int dmz_test_block(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_test_block() argument
2631 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_test_block()
2648 static int dmz_to_next_set_block(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_to_next_set_block() argument
2662 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_to_next_set_block()
2691 int dmz_block_valid(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_block_valid() argument
2696 valid = dmz_test_block(zmd, zone, chunk_block); in dmz_block_valid()
2701 return dmz_to_next_set_block(zmd, zone, chunk_block, in dmz_block_valid()
2706 * Find the first valid block from @chunk_block in @zone.
2711 int dmz_first_valid_block(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_first_valid_block() argument
2717 ret = dmz_to_next_set_block(zmd, zone, start_block, in dmz_first_valid_block()
2725 return dmz_to_next_set_block(zmd, zone, start_block, in dmz_first_valid_block()
2758 * Get a zone weight.
2760 static void dmz_get_zone_weight(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_get_zone_weight() argument
2771 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_get_zone_weight()
2789 zone->weight = n; in dmz_get_zone_weight()
2801 /* Release zone mapping resources */ in dmz_cleanup_metadata()
2845 /* Free the zone descriptors */ in dmz_cleanup_metadata()
2888 struct dm_zone *zone; in dmz_ctr_metadata() local
2916 /* Initialize zone descriptors */ in dmz_ctr_metadata()
2928 zone = dmz_get(zmd, zmd->sb[0].zone->id + i); in dmz_ctr_metadata()
2929 if (!zone) { in dmz_ctr_metadata()
2931 "metadata zone %u not present", i); in dmz_ctr_metadata()
2935 if (!dmz_is_rnd(zone) && !dmz_is_cache(zone)) { in dmz_ctr_metadata()
2937 "metadata zone %d is not random", i); in dmz_ctr_metadata()
2941 set_bit(DMZ_META, &zone->flags); in dmz_ctr_metadata()
2992 dmz_zmd_debug(zmd, " %u data zone mapping blocks", in dmz_ctr_metadata()
3019 * Check zone information on resume.
3023 struct dm_zone *zone; in dmz_resume_metadata() local
3030 zone = dmz_get(zmd, i); in dmz_resume_metadata()
3031 if (!zone) { in dmz_resume_metadata()
3032 dmz_zmd_err(zmd, "Unable to get zone %u", i); in dmz_resume_metadata()
3035 wp_block = zone->wp_block; in dmz_resume_metadata()
3037 ret = dmz_update_zone(zmd, zone); in dmz_resume_metadata()
3039 dmz_zmd_err(zmd, "Broken zone %u", i); in dmz_resume_metadata()
3043 if (dmz_is_offline(zone)) { in dmz_resume_metadata()
3044 dmz_zmd_warn(zmd, "Zone %u is offline", i); in dmz_resume_metadata()
3049 if (!dmz_is_seq(zone)) in dmz_resume_metadata()
3050 zone->wp_block = 0; in dmz_resume_metadata()
3051 else if (zone->wp_block != wp_block) { in dmz_resume_metadata()
3052 dmz_zmd_err(zmd, "Zone %u: Invalid wp (%llu / %llu)", in dmz_resume_metadata()
3053 i, (u64)zone->wp_block, (u64)wp_block); in dmz_resume_metadata()
3054 zone->wp_block = wp_block; in dmz_resume_metadata()
3055 dmz_invalidate_blocks(zmd, zone, zone->wp_block, in dmz_resume_metadata()
3056 zmd->zone_nr_blocks - zone->wp_block); in dmz_resume_metadata()