Lines Matching full:zone

33  * blocks indicating zone block validity.
39 * the first conventional zone found on disk.
87 * and give the zone ID (dzone_id) mapping the chunk on disk.
88 * This zone may be sequential or random. If it is a sequential
89 * zone, a second zone (bzone_id) used as a write buffer may
90 * also be specified. This second zone will always be a randomly
91 * writeable zone.
137 struct dm_zone *zone; member
174 /* Zone information array */
192 /* Zone allocation management */
221 static unsigned int dmz_dev_zone_id(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_dev_zone_id() argument
223 if (WARN_ON(!zone)) in dmz_dev_zone_id()
226 return zone->id - zone->dev->zone_offset; in dmz_dev_zone_id()
229 sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_sect() argument
231 unsigned int zone_id = dmz_dev_zone_id(zmd, zone); in dmz_start_sect()
236 sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_block() argument
238 unsigned int zone_id = dmz_dev_zone_id(zmd, zone); in dmz_start_block()
311 struct dm_zone *zone = kzalloc(sizeof(struct dm_zone), GFP_KERNEL); in dmz_insert() local
313 if (!zone) in dmz_insert()
316 if (xa_insert(&zmd->zones, zone_id, zone, GFP_KERNEL)) { in dmz_insert()
317 kfree(zone); in dmz_insert()
321 INIT_LIST_HEAD(&zone->link); in dmz_insert()
322 atomic_set(&zone->refcount, 0); in dmz_insert()
323 zone->id = zone_id; in dmz_insert()
324 zone->chunk = DMZ_MAP_UNMAPPED; in dmz_insert()
325 zone->dev = dev; in dmz_insert()
327 return zone; in dmz_insert()
359 * The map lock also protects all the zone lists.
375 * the map lock and zone state management (active and reclaim state are
793 sb_block = zmd->sb[set].zone->id << zmd->zone_nr_blocks_shift; in dmz_write_sb()
1016 if (sb_block != (u64)dsb->zone->id << zmd->zone_nr_blocks_shift) { in dmz_check_sb()
1018 sb_block, (u64)dsb->zone->id << zmd->zone_nr_blocks_shift); in dmz_check_sb()
1116 unsigned int zone_id = zmd->sb[0].zone->id; in dmz_lookup_secondary_sb()
1129 zmd->sb[1].zone = dmz_get(zmd, zone_id + 1); in dmz_lookup_secondary_sb()
1137 zmd->sb[1].zone = dmz_get(zmd, zone_id + i); in dmz_lookup_secondary_sb()
1142 zmd->sb[1].zone = NULL; in dmz_lookup_secondary_sb()
1188 zmd->sb[0].block = dmz_start_block(zmd, zmd->sb[0].zone); in dmz_recover_mblocks()
1190 zmd->sb[1].block = dmz_start_block(zmd, zmd->sb[1].zone); in dmz_recover_mblocks()
1234 if (!zmd->sb[0].zone) { in dmz_load_sb()
1235 dmz_zmd_err(zmd, "Primary super block zone not set"); in dmz_load_sb()
1240 zmd->sb[0].block = dmz_start_block(zmd, zmd->sb[0].zone); in dmz_load_sb()
1241 zmd->sb[0].dev = zmd->sb[0].zone->dev; in dmz_load_sb()
1253 if (!zmd->sb[1].zone) { in dmz_load_sb()
1255 zmd->sb[0].zone->id + zmd->nr_meta_zones; in dmz_load_sb()
1257 zmd->sb[1].zone = dmz_get(zmd, zone_id); in dmz_load_sb()
1259 zmd->sb[1].block = dmz_start_block(zmd, zmd->sb[1].zone); in dmz_load_sb()
1324 sb->zone = dmz_get(zmd, zmd->dev[i].zone_offset); in dmz_load_sb()
1326 if (!dmz_is_meta(sb->zone)) { in dmz_load_sb()
1328 "Tertiary super block zone %u not marked as metadata zone", in dmz_load_sb()
1329 sb->zone->id); in dmz_load_sb()
1352 * Initialize a zone descriptor.
1359 struct dm_zone *zone; in dmz_init_zone() local
1361 zone = dmz_insert(zmd, idx, dev); in dmz_init_zone()
1362 if (IS_ERR(zone)) in dmz_init_zone()
1363 return PTR_ERR(zone); in dmz_init_zone()
1367 /* Ignore the eventual runt (smaller) zone */ in dmz_init_zone()
1368 set_bit(DMZ_OFFLINE, &zone->flags); in dmz_init_zone()
1376 * Devices that have zones with a capacity smaller than the zone size in dmz_init_zone()
1384 set_bit(DMZ_RND, &zone->flags); in dmz_init_zone()
1388 set_bit(DMZ_SEQ, &zone->flags); in dmz_init_zone()
1394 if (dmz_is_rnd(zone)) in dmz_init_zone()
1395 zone->wp_block = 0; in dmz_init_zone()
1397 zone->wp_block = dmz_sect2blk(blkz->wp - blkz->start); in dmz_init_zone()
1400 set_bit(DMZ_OFFLINE, &zone->flags); in dmz_init_zone()
1402 set_bit(DMZ_READ_ONLY, &zone->flags); in dmz_init_zone()
1405 if (dmz_is_rnd(zone)) { in dmz_init_zone()
1407 if (zmd->nr_devs == 1 && !zmd->sb[0].zone) { in dmz_init_zone()
1408 /* Primary super block zone */ in dmz_init_zone()
1409 zmd->sb[0].zone = zone; in dmz_init_zone()
1416 * as metadata zone. in dmz_init_zone()
1418 set_bit(DMZ_META, &zone->flags); in dmz_init_zone()
1430 struct dm_zone *zone; in dmz_emulate_zones() local
1432 zone = dmz_insert(zmd, idx, dev); in dmz_emulate_zones()
1433 if (IS_ERR(zone)) in dmz_emulate_zones()
1434 return PTR_ERR(zone); in dmz_emulate_zones()
1435 set_bit(DMZ_CACHE, &zone->flags); in dmz_emulate_zones()
1436 zone->wp_block = 0; in dmz_emulate_zones()
1440 /* Disable runt zone */ in dmz_emulate_zones()
1441 set_bit(DMZ_OFFLINE, &zone->flags); in dmz_emulate_zones()
1457 struct dm_zone *zone = xa_load(&zmd->zones, idx); in dmz_drop_zones() local
1459 kfree(zone); in dmz_drop_zones()
1466 * Allocate and initialize zone descriptors using the zone
1485 /* Allocate zone array */ in dmz_init_zones()
1508 DMDEBUG("(%s): Using %zu B for zone information", in dmz_init_zones()
1521 * Primary superblock zone is always at zone 0 when multiple in dmz_init_zones()
1524 zmd->sb[0].zone = dmz_get(zmd, 0); in dmz_init_zones()
1543 * Get zone information and initialize zone descriptors. At the same in dmz_init_zones()
1545 * first randomly writable zone. in dmz_init_zones()
1562 struct dm_zone *zone = data; in dmz_update_zone_cb() local
1564 clear_bit(DMZ_OFFLINE, &zone->flags); in dmz_update_zone_cb()
1565 clear_bit(DMZ_READ_ONLY, &zone->flags); in dmz_update_zone_cb()
1567 set_bit(DMZ_OFFLINE, &zone->flags); in dmz_update_zone_cb()
1569 set_bit(DMZ_READ_ONLY, &zone->flags); in dmz_update_zone_cb()
1571 if (dmz_is_seq(zone)) in dmz_update_zone_cb()
1572 zone->wp_block = dmz_sect2blk(blkz->wp - blkz->start); in dmz_update_zone_cb()
1574 zone->wp_block = 0; in dmz_update_zone_cb()
1579 * Update a zone information.
1581 static int dmz_update_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_update_zone() argument
1583 struct dmz_dev *dev = zone->dev; in dmz_update_zone()
1591 * Get zone information from disk. Since blkdev_report_zones() uses in dmz_update_zone()
1597 ret = blkdev_report_zones(dev->bdev, dmz_start_sect(zmd, zone), 1, in dmz_update_zone()
1598 dmz_update_zone_cb, zone); in dmz_update_zone()
1604 dmz_dev_err(dev, "Get zone %u report failed", in dmz_update_zone()
1605 zone->id); in dmz_update_zone()
1614 * Check a zone write pointer position when the zone is marked
1618 struct dm_zone *zone) in dmz_handle_seq_write_err() argument
1620 struct dmz_dev *dev = zone->dev; in dmz_handle_seq_write_err()
1624 wp = zone->wp_block; in dmz_handle_seq_write_err()
1625 ret = dmz_update_zone(zmd, zone); in dmz_handle_seq_write_err()
1629 dmz_dev_warn(dev, "Processing zone %u write error (zone wp %u/%u)", in dmz_handle_seq_write_err()
1630 zone->id, zone->wp_block, wp); in dmz_handle_seq_write_err()
1632 if (zone->wp_block < wp) { in dmz_handle_seq_write_err()
1633 dmz_invalidate_blocks(zmd, zone, zone->wp_block, in dmz_handle_seq_write_err()
1634 wp - zone->wp_block); in dmz_handle_seq_write_err()
1641 * Reset a zone write pointer.
1643 static int dmz_reset_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_reset_zone() argument
1651 if (dmz_is_offline(zone) || in dmz_reset_zone()
1652 dmz_is_readonly(zone) || in dmz_reset_zone()
1653 dmz_is_rnd(zone)) in dmz_reset_zone()
1656 if (!dmz_is_empty(zone) || dmz_seq_write_err(zone)) { in dmz_reset_zone()
1657 struct dmz_dev *dev = zone->dev; in dmz_reset_zone()
1660 dmz_start_sect(zmd, zone), in dmz_reset_zone()
1663 dmz_dev_err(dev, "Reset zone %u failed %d", in dmz_reset_zone()
1664 zone->id, ret); in dmz_reset_zone()
1670 clear_bit(DMZ_SEQ_WRITE_ERR, &zone->flags); in dmz_reset_zone()
1671 zone->wp_block = 0; in dmz_reset_zone()
1676 static void dmz_get_zone_weight(struct dmz_metadata *zmd, struct dm_zone *zone);
1696 /* Get chunk mapping table blocks and initialize zone mapping */ in dmz_load_mapping()
1709 /* Check data zone */ in dmz_load_mapping()
1715 dmz_zmd_err(zmd, "Chunk %u mapping: invalid data zone ID %u", in dmz_load_mapping()
1722 dmz_zmd_err(zmd, "Chunk %u mapping: data zone %u not present", in dmz_load_mapping()
1737 /* Check buffer zone */ in dmz_load_mapping()
1743 dmz_zmd_err(zmd, "Chunk %u mapping: invalid buffer zone ID %u", in dmz_load_mapping()
1750 dmz_zmd_err(zmd, "Chunk %u mapping: buffer zone %u not present", in dmz_load_mapping()
1755 dmz_zmd_err(zmd, "Chunk %u mapping: invalid buffer zone %u", in dmz_load_mapping()
1803 /* Unmapped data zone */ in dmz_load_mapping()
1845 * This rotates a zone at the end of its map list.
1847 static void __dmz_lru_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in __dmz_lru_zone() argument
1849 if (list_empty(&zone->link)) in __dmz_lru_zone()
1852 list_del_init(&zone->link); in __dmz_lru_zone()
1853 if (dmz_is_seq(zone)) { in __dmz_lru_zone()
1854 /* LRU rotate sequential zone */ in __dmz_lru_zone()
1855 list_add_tail(&zone->link, &zone->dev->map_seq_list); in __dmz_lru_zone()
1856 } else if (dmz_is_cache(zone)) { in __dmz_lru_zone()
1857 /* LRU rotate cache zone */ in __dmz_lru_zone()
1858 list_add_tail(&zone->link, &zmd->map_cache_list); in __dmz_lru_zone()
1860 /* LRU rotate random zone */ in __dmz_lru_zone()
1861 list_add_tail(&zone->link, &zone->dev->map_rnd_list); in __dmz_lru_zone()
1867 * in LRU order. This rotates a zone at the end of the list.
1869 static void dmz_lru_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_lru_zone() argument
1871 __dmz_lru_zone(zmd, zone); in dmz_lru_zone()
1872 if (zone->bzone) in dmz_lru_zone()
1873 __dmz_lru_zone(zmd, zone->bzone); in dmz_lru_zone()
1877 * Wait for any zone to be freed.
1895 * Lock a zone for reclaim (set the zone RECLAIM bit).
1896 * Returns false if the zone cannot be locked or if it is already locked
1899 int dmz_lock_zone_reclaim(struct dm_zone *zone) in dmz_lock_zone_reclaim() argument
1902 if (dmz_is_active(zone)) in dmz_lock_zone_reclaim()
1905 return !test_and_set_bit(DMZ_RECLAIM, &zone->flags); in dmz_lock_zone_reclaim()
1909 * Clear a zone reclaim flag.
1911 void dmz_unlock_zone_reclaim(struct dm_zone *zone) in dmz_unlock_zone_reclaim() argument
1913 WARN_ON(dmz_is_active(zone)); in dmz_unlock_zone_reclaim()
1914 WARN_ON(!dmz_in_reclaim(zone)); in dmz_unlock_zone_reclaim()
1916 clear_bit_unlock(DMZ_RECLAIM, &zone->flags); in dmz_unlock_zone_reclaim()
1918 wake_up_bit(&zone->flags, DMZ_RECLAIM); in dmz_unlock_zone_reclaim()
1922 * Wait for a zone reclaim to complete.
1924 static void dmz_wait_for_reclaim(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_wait_for_reclaim() argument
1928 set_bit(DMZ_RECLAIM_TERMINATE, &zone->flags); in dmz_wait_for_reclaim()
1929 wait_on_bit_timeout(&zone->flags, DMZ_RECLAIM, TASK_UNINTERRUPTIBLE, HZ); in dmz_wait_for_reclaim()
1930 clear_bit(DMZ_RECLAIM_TERMINATE, &zone->flags); in dmz_wait_for_reclaim()
1936 * Select a cache or random write zone for reclaim.
1942 struct dm_zone *zone, *maxw_z = NULL; in dmz_get_rnd_zone_for_reclaim() local
1945 /* If we have cache zones select from the cache zone list */ in dmz_get_rnd_zone_for_reclaim()
1955 * Find the buffer zone with the heaviest weight or the first (oldest) in dmz_get_rnd_zone_for_reclaim()
1956 * data zone that can be reclaimed. in dmz_get_rnd_zone_for_reclaim()
1958 list_for_each_entry(zone, zone_list, link) { in dmz_get_rnd_zone_for_reclaim()
1959 if (dmz_is_buf(zone)) { in dmz_get_rnd_zone_for_reclaim()
1960 dzone = zone->bzone; in dmz_get_rnd_zone_for_reclaim()
1966 dzone = zone; in dmz_get_rnd_zone_for_reclaim()
1978 * first zone that can be reclaimed regardless of its weitght. in dmz_get_rnd_zone_for_reclaim()
1980 list_for_each_entry(zone, zone_list, link) { in dmz_get_rnd_zone_for_reclaim()
1981 if (dmz_is_buf(zone)) { in dmz_get_rnd_zone_for_reclaim()
1982 dzone = zone->bzone; in dmz_get_rnd_zone_for_reclaim()
1986 dzone = zone; in dmz_get_rnd_zone_for_reclaim()
1995 * Select a buffered sequential zone for reclaim.
2000 struct dm_zone *zone; in dmz_get_seq_zone_for_reclaim() local
2002 list_for_each_entry(zone, &zmd->dev[idx].map_seq_list, link) { in dmz_get_seq_zone_for_reclaim()
2003 if (!zone->bzone) in dmz_get_seq_zone_for_reclaim()
2005 if (dmz_lock_zone_reclaim(zone)) in dmz_get_seq_zone_for_reclaim()
2006 return zone; in dmz_get_seq_zone_for_reclaim()
2013 * Select a zone for reclaim.
2018 struct dm_zone *zone = NULL; in dmz_get_zone_for_reclaim() local
2021 * Search for a zone candidate to reclaim: 2 cases are possible. in dmz_get_zone_for_reclaim()
2022 * (1) There is no free sequential zones. Then a random data zone in dmz_get_zone_for_reclaim()
2023 * cannot be reclaimed. So choose a sequential zone to reclaim so in dmz_get_zone_for_reclaim()
2024 * that afterward a random zone can be reclaimed. in dmz_get_zone_for_reclaim()
2025 * (2) At least one free sequential zone is available, then choose in dmz_get_zone_for_reclaim()
2026 * the oldest random zone (data or buffer) that can be locked. in dmz_get_zone_for_reclaim()
2030 zone = dmz_get_seq_zone_for_reclaim(zmd, dev_idx); in dmz_get_zone_for_reclaim()
2031 if (!zone) in dmz_get_zone_for_reclaim()
2032 zone = dmz_get_rnd_zone_for_reclaim(zmd, dev_idx, idle); in dmz_get_zone_for_reclaim()
2035 return zone; in dmz_get_zone_for_reclaim()
2039 * Get the zone mapping a chunk, if the chunk is mapped already.
2040 * If no mapping exist and the operation is WRITE, a zone is
2042 * The zone returned will be set to the active state.
2067 /* Allocate a random zone */ in dmz_get_chunk_mapping()
2081 /* The chunk is already mapped: get the mapping zone */ in dmz_get_chunk_mapping()
2104 * If the zone is being reclaimed, the chunk mapping may change in dmz_get_chunk_mapping()
2105 * to a different zone. So wait for reclaim and retry. Otherwise, in dmz_get_chunk_mapping()
2106 * activate the zone (this will prevent reclaim from touching it). in dmz_get_chunk_mapping()
2137 /* Empty buffer zone: reclaim it */ in dmz_put_chunk_mapping()
2144 /* Deactivate the data zone */ in dmz_put_chunk_mapping()
2149 /* Unbuffered inactive empty data zone: reclaim it */ in dmz_put_chunk_mapping()
2158 * Allocate and map a random zone to buffer a chunk
2159 * already mapped to a sequential zone.
2173 /* Allocate a random zone */ in dmz_get_chunk_buffer()
2202 * Get an unmapped (free) zone.
2209 struct dm_zone *zone; in dmz_alloc_zone() local
2229 * No free zone: return NULL if this is for not reclaim. in dmz_alloc_zone()
2245 zone = list_first_entry_or_null(&zmd->reserved_seq_zones_list, in dmz_alloc_zone()
2247 if (zone) { in dmz_alloc_zone()
2248 list_del_init(&zone->link); in dmz_alloc_zone()
2251 return zone; in dmz_alloc_zone()
2254 zone = list_first_entry(list, struct dm_zone, link); in dmz_alloc_zone()
2255 list_del_init(&zone->link); in dmz_alloc_zone()
2257 if (dmz_is_cache(zone)) in dmz_alloc_zone()
2259 else if (dmz_is_rnd(zone)) in dmz_alloc_zone()
2260 atomic_dec(&zone->dev->unmap_nr_rnd); in dmz_alloc_zone()
2262 atomic_dec(&zone->dev->unmap_nr_seq); in dmz_alloc_zone()
2264 if (dmz_is_offline(zone)) { in dmz_alloc_zone()
2265 dmz_zmd_warn(zmd, "Zone %u is offline", zone->id); in dmz_alloc_zone()
2266 zone = NULL; in dmz_alloc_zone()
2269 if (dmz_is_meta(zone)) { in dmz_alloc_zone()
2270 dmz_zmd_warn(zmd, "Zone %u has metadata", zone->id); in dmz_alloc_zone()
2271 zone = NULL; in dmz_alloc_zone()
2274 return zone; in dmz_alloc_zone()
2278 * Free a zone.
2281 void dmz_free_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_free_zone() argument
2283 /* If this is a sequential zone, reset it */ in dmz_free_zone()
2284 if (dmz_is_seq(zone)) in dmz_free_zone()
2285 dmz_reset_zone(zmd, zone); in dmz_free_zone()
2287 /* Return the zone to its type unmap list */ in dmz_free_zone()
2288 if (dmz_is_cache(zone)) { in dmz_free_zone()
2289 list_add_tail(&zone->link, &zmd->unmap_cache_list); in dmz_free_zone()
2291 } else if (dmz_is_rnd(zone)) { in dmz_free_zone()
2292 list_add_tail(&zone->link, &zone->dev->unmap_rnd_list); in dmz_free_zone()
2293 atomic_inc(&zone->dev->unmap_nr_rnd); in dmz_free_zone()
2294 } else if (dmz_is_reserved(zone)) { in dmz_free_zone()
2295 list_add_tail(&zone->link, &zmd->reserved_seq_zones_list); in dmz_free_zone()
2298 list_add_tail(&zone->link, &zone->dev->unmap_seq_list); in dmz_free_zone()
2299 atomic_inc(&zone->dev->unmap_nr_seq); in dmz_free_zone()
2306 * Map a chunk to a zone.
2325 * Unmap a zone.
2328 void dmz_unmap_zone(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_unmap_zone() argument
2330 unsigned int chunk = zone->chunk; in dmz_unmap_zone()
2338 if (test_and_clear_bit(DMZ_BUF, &zone->flags)) { in dmz_unmap_zone()
2340 * Unmapping the chunk buffer zone: clear only in dmz_unmap_zone()
2343 dzone_id = zone->bzone->id; in dmz_unmap_zone()
2344 zone->bzone->bzone = NULL; in dmz_unmap_zone()
2345 zone->bzone = NULL; in dmz_unmap_zone()
2349 * Unmapping the chunk data zone: the zone must in dmz_unmap_zone()
2352 if (WARN_ON(zone->bzone)) { in dmz_unmap_zone()
2353 zone->bzone->bzone = NULL; in dmz_unmap_zone()
2354 zone->bzone = NULL; in dmz_unmap_zone()
2361 zone->chunk = DMZ_MAP_UNMAPPED; in dmz_unmap_zone()
2362 list_del_init(&zone->link); in dmz_unmap_zone()
2398 * Get the bitmap block storing the bit for chunk_block in zone.
2401 struct dm_zone *zone, in dmz_get_bitmap() argument
2405 (sector_t)(zone->id * zmd->zone_nr_bitmap_blocks) + in dmz_get_bitmap()
2457 /* Get a valid region from the source zone */ in dmz_merge_valid_blocks()
2476 int dmz_validate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_validate_blocks() argument
2484 dmz_zmd_debug(zmd, "=> VALIDATE zone %u, block %llu, %u blocks", in dmz_validate_blocks()
2485 zone->id, (unsigned long long)chunk_block, in dmz_validate_blocks()
2492 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_validate_blocks()
2511 if (likely(zone->weight + n <= zone_nr_blocks)) in dmz_validate_blocks()
2512 zone->weight += n; in dmz_validate_blocks()
2514 dmz_zmd_warn(zmd, "Zone %u: weight %u should be <= %u", in dmz_validate_blocks()
2515 zone->id, zone->weight, in dmz_validate_blocks()
2517 zone->weight = zone_nr_blocks; in dmz_validate_blocks()
2557 int dmz_invalidate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_invalidate_blocks() argument
2564 dmz_zmd_debug(zmd, "=> INVALIDATE zone %u, block %llu, %u blocks", in dmz_invalidate_blocks()
2565 zone->id, (u64)chunk_block, nr_blocks); in dmz_invalidate_blocks()
2571 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_invalidate_blocks()
2591 if (zone->weight >= n) in dmz_invalidate_blocks()
2592 zone->weight -= n; in dmz_invalidate_blocks()
2594 dmz_zmd_warn(zmd, "Zone %u: weight %u should be >= %u", in dmz_invalidate_blocks()
2595 zone->id, zone->weight, n); in dmz_invalidate_blocks()
2596 zone->weight = 0; in dmz_invalidate_blocks()
2605 static int dmz_test_block(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_test_block() argument
2614 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_test_block()
2631 static int dmz_to_next_set_block(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_to_next_set_block() argument
2645 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_to_next_set_block()
2674 int dmz_block_valid(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_block_valid() argument
2679 valid = dmz_test_block(zmd, zone, chunk_block); in dmz_block_valid()
2684 return dmz_to_next_set_block(zmd, zone, chunk_block, in dmz_block_valid()
2689 * Find the first valid block from @chunk_block in @zone.
2694 int dmz_first_valid_block(struct dmz_metadata *zmd, struct dm_zone *zone, in dmz_first_valid_block() argument
2700 ret = dmz_to_next_set_block(zmd, zone, start_block, in dmz_first_valid_block()
2708 return dmz_to_next_set_block(zmd, zone, start_block, in dmz_first_valid_block()
2741 * Get a zone weight.
2743 static void dmz_get_zone_weight(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_get_zone_weight() argument
2754 mblk = dmz_get_bitmap(zmd, zone, chunk_block); in dmz_get_zone_weight()
2772 zone->weight = n; in dmz_get_zone_weight()
2784 /* Release zone mapping resources */ in dmz_cleanup_metadata()
2828 /* Free the zone descriptors */ in dmz_cleanup_metadata()
2870 struct dm_zone *zone; in dmz_ctr_metadata() local
2898 /* Initialize zone descriptors */ in dmz_ctr_metadata()
2910 zone = dmz_get(zmd, zmd->sb[0].zone->id + i); in dmz_ctr_metadata()
2911 if (!zone) { in dmz_ctr_metadata()
2913 "metadata zone %u not present", i); in dmz_ctr_metadata()
2917 if (!dmz_is_rnd(zone) && !dmz_is_cache(zone)) { in dmz_ctr_metadata()
2919 "metadata zone %d is not random", i); in dmz_ctr_metadata()
2923 set_bit(DMZ_META, &zone->flags); in dmz_ctr_metadata()
2980 dmz_zmd_debug(zmd, " %u data zone mapping blocks", in dmz_ctr_metadata()
3007 * Check zone information on resume.
3011 struct dm_zone *zone; in dmz_resume_metadata() local
3018 zone = dmz_get(zmd, i); in dmz_resume_metadata()
3019 if (!zone) { in dmz_resume_metadata()
3020 dmz_zmd_err(zmd, "Unable to get zone %u", i); in dmz_resume_metadata()
3023 wp_block = zone->wp_block; in dmz_resume_metadata()
3025 ret = dmz_update_zone(zmd, zone); in dmz_resume_metadata()
3027 dmz_zmd_err(zmd, "Broken zone %u", i); in dmz_resume_metadata()
3031 if (dmz_is_offline(zone)) { in dmz_resume_metadata()
3032 dmz_zmd_warn(zmd, "Zone %u is offline", i); in dmz_resume_metadata()
3037 if (!dmz_is_seq(zone)) in dmz_resume_metadata()
3038 zone->wp_block = 0; in dmz_resume_metadata()
3039 else if (zone->wp_block != wp_block) { in dmz_resume_metadata()
3040 dmz_zmd_err(zmd, "Zone %u: Invalid wp (%llu / %llu)", in dmz_resume_metadata()
3041 i, (u64)zone->wp_block, (u64)wp_block); in dmz_resume_metadata()
3042 zone->wp_block = wp_block; in dmz_resume_metadata()
3043 dmz_invalidate_blocks(zmd, zone, zone->wp_block, in dmz_resume_metadata()
3044 zmd->zone_nr_blocks - zone->wp_block); in dmz_resume_metadata()