Home
last modified time | relevance | path

Searched full:zone (Results 1 – 25 of 661) sorted by relevance

12345678910>>...27

/linux-6.8/drivers/block/null_blk/
Dzoned.c35 struct nullb_zone *zone) in null_init_zone_lock() argument
38 spin_lock_init(&zone->spinlock); in null_init_zone_lock()
40 mutex_init(&zone->mutex); in null_init_zone_lock()
44 struct nullb_zone *zone) in null_lock_zone() argument
47 spin_lock_irq(&zone->spinlock); in null_lock_zone()
49 mutex_lock(&zone->mutex); in null_lock_zone()
53 struct nullb_zone *zone) in null_unlock_zone() argument
56 spin_unlock_irq(&zone->spinlock); in null_unlock_zone()
58 mutex_unlock(&zone->mutex); in null_unlock_zone()
64 struct nullb_zone *zone; in null_init_zoned_dev() local
[all …]
/linux-6.8/fs/pstore/
Dzone.c26 * struct psz_buffer - header of zone to flush to storage
31 * @data: zone data.
66 * @off: zone offset of storage
67 * @type: front-end type for this zone
68 * @name: front-end name for this zone
69 * @buffer: pointer to data buffer managed by this zone
72 * @should_recover: whether this zone should recover from storage
75 * zone structure in memory.
90 * struct psz_context - all about running state of pstore/zone
93 * @ppsz: pmsg storage zone
[all …]
/linux-6.8/include/uapi/linux/
Dblkzoned.h25 * @BLK_ZONE_TYPE_CONVENTIONAL: The zone has no write pointer and can be writen
26 * randomly. Zone reset has no effect on the zone.
27 * @BLK_ZONE_TYPE_SEQWRITE_REQ: The zone must be written sequentially
28 * @BLK_ZONE_TYPE_SEQWRITE_PREF: The zone can be written non-sequentially
39 * enum blk_zone_cond - Condition [state] of a zone in a zoned device.
41 * @BLK_ZONE_COND_NOT_WP: The zone has no write pointer, it is conventional.
42 * @BLK_ZONE_COND_EMPTY: The zone is empty.
43 * @BLK_ZONE_COND_IMP_OPEN: The zone is open, but not explicitly opened.
45 * OPEN ZONE command.
46 * @BLK_ZONE_COND_CLOSED: The zone was [explicitly] closed after writing.
[all …]
/linux-6.8/Documentation/filesystems/
Dzonefs.rst4 ZoneFS - Zone filesystem for Zoned block devices
10 zonefs is a very simple file system exposing each zone of a zoned block device
24 by allowing SSTables to be stored in a zone file similarly to a regular file
26 of the higher level construct "one file is one zone" can help reducing the
34 space that is divided into zones. A zone is a group of consecutive LBAs and all
41 sequentially. Each sequential zone has a write pointer maintained by the
43 to the device. As a result of this write constraint, LBAs in a sequential zone
45 command (zone reset) before rewriting.
61 representing zones are grouped by zone type, which are themselves represented
62 by sub-directories. This file structure is built entirely using zone information
[all …]
/linux-6.8/drivers/md/
Ddm-zoned-metadata.c33 * blocks indicating zone block validity.
39 * the first conventional zone found on disk.
87 * and give the zone ID (dzone_id) mapping the chunk on disk.
88 * This zone may be sequential or random. If it is a sequential
89 * zone, a second zone (bzone_id) used as a write buffer may
90 * also be specified. This second zone will always be a randomly
91 * writeable zone.
137 struct dm_zone *zone; member
174 /* Zone information array */
192 /* Zone allocation management */
[all …]
Ddm-zone.c14 #define DM_MSG_PREFIX "zone"
19 * For internal zone reports bypassing the top BIO submission path.
52 * User facing dm device block device report zone operation. This calls the
77 static int dm_report_zones_cb(struct blk_zone *zone, unsigned int idx, in dm_report_zones_cb() argument
86 if (zone->start >= args->start + args->tgt->len) in dm_report_zones_cb()
90 * Remap the start sector and write pointer position of the zone in dm_report_zones_cb()
93 zone->start += sector_diff; in dm_report_zones_cb()
94 if (zone->type != BLK_ZONE_TYPE_CONVENTIONAL) { in dm_report_zones_cb()
95 if (zone->cond == BLK_ZONE_COND_FULL) in dm_report_zones_cb()
96 zone->wp = zone->start + zone->len; in dm_report_zones_cb()
[all …]
Ddm-zoned.h93 * Zone descriptor.
96 /* For listing the zone depending on its state */
99 /* Device containing this zone */
102 /* Zone type and state */
105 /* Zone activation reference count */
108 /* Zone id */
111 /* Zone write pointer block (relative to the zone start block) */
114 /* Zone weight (number of valid blocks in the zone) */
117 /* The chunk that the zone maps */
121 * For a sequential data zone, pointer to the random zone
[all …]
Ddm-zoned-target.c17 * Zone BIO context.
21 struct dm_zone *zone; member
86 struct dm_zone *zone = bioctx->zone; in dmz_bio_endio() local
88 if (zone) { in dmz_bio_endio()
91 dmz_is_seq(zone)) in dmz_bio_endio()
92 set_bit(DMZ_SEQ_WRITE_ERR, &zone->flags); in dmz_bio_endio()
93 dmz_deactivate_zone(zone); in dmz_bio_endio()
116 static int dmz_submit_bio(struct dmz_target *dmz, struct dm_zone *zone, in dmz_submit_bio() argument
122 struct dmz_dev *dev = zone->dev; in dmz_submit_bio()
134 dmz_start_sect(dmz->metadata, zone) + dmz_blk2sect(chunk_block); in dmz_submit_bio()
[all …]
/linux-6.8/mm/
Dshow_mem.c26 static inline void show_node(struct zone *zone) in show_node() argument
29 printk("Node %d ", zone_to_nid(zone)); in show_node()
38 struct zone *zone; in si_mem_available() local
40 for_each_zone(zone) in si_mem_available()
41 wmark_low += low_wmark_pages(zone); in si_mem_available()
104 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node() local
106 if (is_highmem(zone)) { in si_meminfo_node()
107 managed_highpages += zone_managed_pages(zone); in si_meminfo_node()
108 free_highpages += zone_page_state(zone, NR_FREE_PAGES); in si_meminfo_node()
190 struct zone *zone; in show_free_areas() local
[all …]
Dpage_alloc.c13 * Zone balancing, Kanoj Sarcar, SGI, Jan 2000
80 * shuffle the whole zone).
89 /* prevent >1 _updater_ of zone percpu pageset ->high and ->batch fields */
293 /* movable_zone is the "real" zone pages in ZONE_MOVABLE are taken from */
306 static bool try_to_accept_memory(struct zone *zone, unsigned int order);
332 _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
334 return deferred_grow_zone(zone, order); in _deferred_grow_zone()
442 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument
450 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries()
451 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries()
[all …]
Dvmstat.c37 /* zero numa counters within a zone */
38 static void zero_zone_numa_counters(struct zone *zone) in zero_zone_numa_counters() argument
43 atomic_long_set(&zone->vm_numa_event[item], 0); in zero_zone_numa_counters()
45 per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_event[item] in zero_zone_numa_counters()
54 struct zone *zone; in zero_zones_numa_counters() local
56 for_each_populated_zone(zone) in zero_zones_numa_counters()
57 zero_zone_numa_counters(zone); in zero_zones_numa_counters()
158 * Manage combined zone based / global counters
169 static void fold_vm_zone_numa_events(struct zone *zone) in fold_vm_zone_numa_events() argument
178 pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu); in fold_vm_zone_numa_events()
[all …]
Dmm_init.c49 struct zone *zone; in mminit_verify_zonelist() local
57 /* Identify the zone and nodelist */ in mminit_verify_zonelist()
61 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
62 if (!populated_zone(zone)) in mminit_verify_zonelist()
68 zone->name); in mminit_verify_zonelist()
71 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
72 pr_cont("%d:%s ", zone_to_nid(zone), zone->name); in mminit_verify_zonelist()
87 "Section %d Node %d Zone %d Lastcpupid %d Kasantag %d Gen %d Tier %d Flags %d\n", in mminit_verify_pageflags_layout()
97 "Section %d Node %d Zone %d Lastcpupid %d Kasantag %d\n", in mminit_verify_pageflags_layout()
104 "Section %lu Node %lu Zone %lu Lastcpupid %lu Kasantag %lu\n", in mminit_verify_pageflags_layout()
[all …]
Dcompaction.c59 * the "fragmentation score" of a node/zone.
153 static void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
155 zone->compact_considered = 0; in defer_compaction()
156 zone->compact_defer_shift++; in defer_compaction()
158 if (order < zone->compact_order_failed) in defer_compaction()
159 zone->compact_order_failed = order; in defer_compaction()
161 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction()
162 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction()
164 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction()
168 static bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
[all …]
/linux-6.8/tools/power/cpupower/lib/
Dpowercap.c116 static int sysfs_powercap_get64_val(struct powercap_zone *zone, in sysfs_powercap_get64_val() argument
124 strcat(file, zone->sys_name); in sysfs_powercap_get64_val()
138 int powercap_get_max_energy_range_uj(struct powercap_zone *zone, uint64_t *val) in powercap_get_max_energy_range_uj() argument
140 return sysfs_powercap_get64_val(zone, GET_MAX_ENERGY_RANGE_UJ, val); in powercap_get_max_energy_range_uj()
143 int powercap_get_energy_uj(struct powercap_zone *zone, uint64_t *val) in powercap_get_energy_uj() argument
145 return sysfs_powercap_get64_val(zone, GET_ENERGY_UJ, val); in powercap_get_energy_uj()
148 int powercap_get_max_power_range_uw(struct powercap_zone *zone, uint64_t *val) in powercap_get_max_power_range_uw() argument
150 return sysfs_powercap_get64_val(zone, GET_MAX_POWER_RANGE_UW, val); in powercap_get_max_power_range_uw()
153 int powercap_get_power_uw(struct powercap_zone *zone, uint64_t *val) in powercap_get_power_uw() argument
155 return sysfs_powercap_get64_val(zone, GET_POWER_UW, val); in powercap_get_power_uw()
[all …]
/linux-6.8/include/linux/
Dmemory_hotplug.h11 struct zone;
70 /* Types for control the zone type of onlined and offlined memory */
74 /* Online the memory. Zone depends, see default_zone_for_pfn(). */
126 * Zone resizing functions
128 * Note: any attempt to resize a zone should has pgdat_resize_lock()
129 * zone_span_writelock() both held. This ensure the size of a zone
132 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument
134 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin()
136 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument
138 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry()
[all …]
Dmmzone.h126 NUMA_INTERLEAVE_HIT, /* interleaver preferred this zone */
679 * zone lock contention and keep cache-hot pages reusing.
768 * faulted, they come from the right zone right away. However, it is
772 * to a different zone. When migration fails - pinning fails.
793 * on different platforms may end up in a movable zone. ZERO_PAGE(0)
796 * memory to the MOVABLE zone, the vmemmap pages are also placed in
797 * such zone. Such pages cannot be really moved around as they are
819 struct zone { struct
822 /* zone watermarks, access with *_wmark_pages(zone) macros */
831 * wasting several GB of ram we must reserve some of the lower zone argument
[all …]
Dvmstat.h138 * Zone and node-based page accounting with per cpu differentials.
145 static inline void zone_numa_event_add(long x, struct zone *zone, in zone_numa_event_add() argument
148 atomic_long_add(x, &zone->vm_numa_event[item]); in zone_numa_event_add()
152 static inline unsigned long zone_numa_event_state(struct zone *zone, in zone_numa_event_state() argument
155 return atomic_long_read(&zone->vm_numa_event[item]); in zone_numa_event_state()
165 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument
168 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add()
207 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument
210 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state()
224 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument
[all …]
Dpowercap.h16 * controlled. Each power zone can have one or more constraints.
28 * limits. If disabled power zone can only be monitored
75 * struct powercap_zone_ops - Define power zone callbacks
83 * @set_enable: Enable/Disable power zone controls.
92 * This structure defines zone callbacks to be implemented by client drivers.
112 * struct powercap_zone- Defines instance of a power cap zone
114 * @name: Power zone name.
115 * @control_type_inst: Control type instance for this zone.
116 * @ops: Pointer to the zone operation structure.
121 * @private_data: Private data pointer if any for this zone.
[all …]
/linux-6.8/Documentation/driver-api/thermal/
Dsysfs-api.rst15 The generic thermal sysfs provides a set of interfaces for thermal zone
19 This how-to focuses on enabling new thermal zone and cooling devices to
21 This solution is platform independent and any type of thermal zone devices
24 The main task of the thermal sysfs driver is to expose thermal zone attributes
27 inputs from thermal zone attributes (the current temperature and trip point
36 1.1 thermal zone device interface
48 This interface function adds a new thermal zone device (sensor) to
53 the thermal zone type.
55 the total number of trip points this thermal zone supports.
61 thermal zone device call-backs.
[all …]
/linux-6.8/fs/zonefs/
Dsuper.c31 * Get the name of a zone group directory.
47 * Manage the active zone count.
65 * If the zone is active, that is, if it is explicitly open or in zonefs_account_active()
78 /* The zone is not active. If it was, update the active count */ in zonefs_account_active()
86 * Manage the active zone count. Called with zi->i_truncate_mutex held.
96 * Execute a zone management operation.
104 * With ZNS drives, closing an explicitly open zone that has not been in zonefs_zone_mgmt()
105 * written will change the zone state to "closed", that is, the zone in zonefs_zone_mgmt()
107 * open operation on other zones if the drive active zone resources in zonefs_zone_mgmt()
108 * are exceeded, make sure that the zone does not remain active by in zonefs_zone_mgmt()
[all …]
/linux-6.8/fs/adfs/
Dmap.c14 * zone which contains a bitstream made up of variable sized fragments.
30 * large or fragmented files. The first map zone a fragment starts in
32 * from any zone on the disk.
63 * return the map bit offset of the fragment frag_id in the zone dm.
109 * Scan the free space map, for this zone, calculating the total
133 * exist in this zone. in scan_free_map()
159 static int scan_map(struct adfs_sb_info *asb, unsigned int zone, in scan_map() argument
166 dm = asb->s_map + zone; in scan_map()
167 zone = asb->s_map_size; in scan_map()
168 dm_end = asb->s_map + zone; in scan_map()
[all …]
/linux-6.8/Documentation/admin-guide/device-mapper/
Ddm-zoned.rst54 Data in these zones may be directly mapped to the conventional zone, but
55 later moved to a sequential zone so that the conventional zone can be
65 1) The first block of the first conventional zone found contains the
73 indicates the zone number of the device storing the chunk of data. Each
74 mapping entry may also indicate if the zone number of a conventional
75 zone used to buffer random modification to the data zone.
80 data chunk, a block is always valid only in the data zone mapping the
81 chunk or in the buffer zone of the chunk.
83 For a logical chunk mapped to a conventional zone, all write operations
84 are processed by directly writing to the zone. If the mapping zone is a
[all …]
/linux-6.8/drivers/scsi/
Dsd_zbc.c27 * sd_zbc_get_zone_wp_offset - Get zone write pointer offset.
28 * @zone: Zone for which to return the write pointer offset.
30 * Return: offset of the write pointer from the start of the zone.
32 static unsigned int sd_zbc_get_zone_wp_offset(struct blk_zone *zone) in sd_zbc_get_zone_wp_offset() argument
34 if (zone->type == ZBC_ZONE_TYPE_CONV) in sd_zbc_get_zone_wp_offset()
37 switch (zone->cond) { in sd_zbc_get_zone_wp_offset()
41 return zone->wp - zone->start; in sd_zbc_get_zone_wp_offset()
43 return zone->len; in sd_zbc_get_zone_wp_offset()
50 * write pointer. Use 0 as for an empty zone. in sd_zbc_get_zone_wp_offset()
56 /* Whether or not a SCSI zone descriptor describes a gap zone. */
[all …]
/linux-6.8/include/net/netfilter/
Dnf_conntrack_zones.h12 return &ct->zone; in nf_ct_zone()
19 nf_ct_zone_init(struct nf_conntrack_zone *zone, u16 id, u8 dir, u8 flags) in nf_ct_zone_init() argument
21 zone->id = id; in nf_ct_zone_init()
22 zone->flags = flags; in nf_ct_zone_init()
23 zone->dir = dir; in nf_ct_zone_init()
25 return zone; in nf_ct_zone_init()
36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl()
37 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl()
43 const struct nf_conntrack_zone *zone) in nf_ct_zone_add() argument
46 ct->zone = *zone; in nf_ct_zone_add()
[all …]
/linux-6.8/include/trace/events/
Dcompaction.h194 TP_PROTO(struct zone *zone,
198 TP_ARGS(zone, order, ret),
208 __entry->nid = zone_to_nid(zone);
209 __entry->idx = zone_idx(zone);
214 TP_printk("node=%d zone=%-8s order=%d ret=%s",
223 TP_PROTO(struct zone *zone,
227 TP_ARGS(zone, order, ret)
232 TP_PROTO(struct zone *zone,
236 TP_ARGS(zone, order, ret)
241 TP_PROTO(struct zone *zone, int order),
[all …]

12345678910>>...27