Home
last modified time | relevance | path

Searched full:zone (Results 1 – 25 of 583) sorted by relevance

12345678910>>...24

/linux/drivers/block/null_blk/
H A Dzoned.c25 struct nullb_zone *zone) in null_init_zone_lock() argument
28 spin_lock_init(&zone->spinlock); in null_init_zone_lock()
30 mutex_init(&zone->mutex); in null_init_zone_lock()
34 struct nullb_zone *zone) in null_lock_zone() argument
37 spin_lock_irq(&zone->spinlock); in null_lock_zone()
39 mutex_lock(&zone->mutex); in null_lock_zone()
43 struct nullb_zone *zone) in null_unlock_zone() argument
46 spin_unlock_irq(&zone->spinlock); in null_unlock_zone()
48 mutex_unlock(&zone->mutex); in null_unlock_zone()
55 struct nullb_zone *zone; in null_init_zoned_dev() local
[all …]
/linux/drivers/md/dm-vdo/
H A Dlogical-zone.c6 #include "logical-zone.h"
21 #include "physical-zone.h"
47 * initialize_zone() - Initialize a logical zone.
48 * @zones: The logical_zones to which this zone belongs.
55 struct logical_zone *zone = &zones->zones[zone_number]; in initialize_zone() local
58 result = vdo_int_map_create(VDO_LOCK_MAP_CAPACITY, &zone->lbn_operations); in initialize_zone()
63 zone->next = &zones->zones[zone_number + 1]; in initialize_zone()
65 vdo_initialize_completion(&zone->completion, vdo, in initialize_zone()
67 zone->zones = zones; in initialize_zone()
68 zone->zone_number = zone_number; in initialize_zone()
[all …]
H A Dblock-map.c23 #include "physical-zone.h"
60 struct block_map_zone *zone; member
98 struct block_map_zone *zone; member
121 * For adjusting VDO page cache statistic fields which are only mutated on the logical zone thread.
200 info->vio->completion.callback_thread_id = cache->zone->thread_id; in initialize_info()
251 VDO_ASSERT_LOG_ONLY((thread_id == cache->zone->thread_id), in assert_on_cache_thread()
253 function_name, cache->zone->thread_id, thread_id); in assert_on_cache_thread()
259 VDO_ASSERT_LOG_ONLY(!vdo_is_state_quiescent(&cache->zone->state), in assert_io_allowed()
637 static void check_for_drain_complete(struct block_map_zone *zone) in check_for_drain_complete() argument
639 if (vdo_is_state_draining(&zone->state) && in check_for_drain_complete()
[all …]
/linux/fs/pstore/
H A Dzone.c26 * struct psz_buffer - header of zone to flush to storage
31 * @data: zone data.
66 * @off: zone offset of storage
67 * @type: front-end type for this zone
68 * @name: front-end name for this zone
69 * @buffer: pointer to data buffer managed by this zone
72 * @should_recover: whether this zone should recover from storage
75 * zone structure in memory.
90 * struct psz_context - all about running state of pstore/zone
93 * @ppsz: pmsg storage zone
[all …]
/linux/include/uapi/linux/
H A Dblkzoned.h25 * @BLK_ZONE_TYPE_CONVENTIONAL: The zone has no write pointer and can be writen
26 * randomly. Zone reset has no effect on the zone.
27 * @BLK_ZONE_TYPE_SEQWRITE_REQ: The zone must be written sequentially
28 * @BLK_ZONE_TYPE_SEQWRITE_PREF: The zone can be written non-sequentially
39 * enum blk_zone_cond - Condition [state] of a zone in a zoned device.
41 * @BLK_ZONE_COND_NOT_WP: The zone has no write pointer, it is conventional.
42 * @BLK_ZONE_COND_EMPTY: The zone is empty.
43 * @BLK_ZONE_COND_IMP_OPEN: The zone is open, but not explicitly opened.
45 * OPEN ZONE command.
46 * @BLK_ZONE_COND_CLOSED: The zone was [explicitly] closed after writing.
[all …]
/linux/block/
H A Dblk-zoned.c41 * Per-zone write plug.
45 * @rcu_head: RCU head to free zone write plugs with an RCU grace period.
48 * @ref: Zone write plug reference counter. A zone write plug reference is
54 * reference is dropped whenever the zone of the zone write plug is reset,
55 * finished and when the zone becomes full (last write BIO to the zone
58 * @zone_no: The number of the zone the plug is managing.
59 * @wp_offset: The zone write pointer location relative to the start of the zone
61 * @cond: Condition of the zone
80 * All request-based zoned devices need zone resources so that the in disk_need_zone_resources()
83 * handling zone write ordering and do not need zone resources, unless in disk_need_zone_resources()
[all …]
/linux/drivers/block/
H A Dzloop.c7 * zone as backing storage.
66 /* Arbitrary limit on the zone size (16GB). */
158 struct zloop_zone *zone = &zlo->zones[zone_no]; in zloop_update_seq_zone() local
164 lockdep_assert_held(&zone->lock); in zloop_update_seq_zone()
166 ret = vfs_getattr(&zone->file->f_path, &stat, STATX_SIZE, 0); in zloop_update_seq_zone()
168 pr_err("Failed to get zone %u file stat (err=%d)\n", in zloop_update_seq_zone()
170 set_bit(ZLOOP_ZONE_SEQ_ERROR, &zone->flags); in zloop_update_seq_zone()
176 pr_err("Zone %u file too large (%llu sectors > %llu)\n", in zloop_update_seq_zone()
182 pr_err("Zone %u file size not aligned to block size %u\n", in zloop_update_seq_zone()
187 spin_lock_irqsave(&zone->wp_lock, flags); in zloop_update_seq_zone()
[all …]
/linux/drivers/md/
H A Ddm-zoned-metadata.c33 * blocks indicating zone block validity.
39 * the first conventional zone found on disk.
87 * and give the zone ID (dzone_id) mapping the chunk on disk.
88 * This zone may be sequential or random. If it is a sequential
89 * zone, a second zone (bzone_id) used as a write buffer may
90 * also be specified. This second zone will always be a randomly
91 * writeable zone.
137 struct dm_zone *zone; member
174 /* Zone information array */
192 /* Zone allocation management */
[all …]
H A Ddm-zoned-target.c17 * Zone BIO context.
21 struct dm_zone *zone; member
86 struct dm_zone *zone = bioctx->zone; in dmz_bio_endio() local
88 if (zone) { in dmz_bio_endio()
91 dmz_is_seq(zone)) in dmz_bio_endio()
92 set_bit(DMZ_SEQ_WRITE_ERR, &zone->flags); in dmz_bio_endio()
93 dmz_deactivate_zone(zone); in dmz_bio_endio()
116 static int dmz_submit_bio(struct dmz_target *dmz, struct dm_zone *zone, in dmz_submit_bio() argument
122 struct dmz_dev *dev = zone->dev; in dmz_submit_bio()
134 dmz_start_sect(dmz->metadata, zone) + dmz_blk2sect(chunk_block); in dmz_submit_bio()
[all …]
/linux/Documentation/filesystems/
H A Dzonefs.rst4 ZoneFS - Zone filesystem for Zoned block devices
10 zonefs is a very simple file system exposing each zone of a zoned block device
24 by allowing SSTables to be stored in a zone file similarly to a regular file
26 of the higher level construct "one file is one zone" can help reducing the
34 space that is divided into zones. A zone is a group of consecutive LBAs and all
41 sequentially. Each sequential zone has a write pointer maintained by the
43 to the device. As a result of this write constraint, LBAs in a sequential zone
45 command (zone reset) before rewriting.
61 representing zones are grouped by zone type, which are themselves represented
62 by sub-directories. This file structure is built entirely using zone information
[all …]
/linux/mm/
H A Dshow_mem.c26 static inline void show_node(struct zone *zone) in show_node() argument
29 printk("Node %d ", zone_to_nid(zone)); in show_node()
38 struct zone *zone; in si_mem_available() local
40 for_each_zone(zone) in si_mem_available()
41 wmark_low += low_wmark_pages(zone); in si_mem_available()
98 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node() local
99 managed_pages += zone_managed_pages(zone); in si_meminfo_node()
184 struct zone *zone; show_free_areas() local
405 struct zone *zone; __show_mem() local
[all...]
H A Dpage_alloc.c12 * Zone balancing, Kanoj Sarcar, SGI, Jan 2000
81 * shuffle the whole zone).
93 /* prevent >1 _updater_ of zone percpu pageset ->high and ->batch fields */
308 /* movable_zone is the "real" zone pages in ZONE_MOVABLE are taken from */
320 static bool cond_accept_memory(struct zone *zone, unsigned int order,
346 _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
348 return deferred_grow_zone(zone, order); in _deferred_grow_zone()
356 static inline bool _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
604 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument
612 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries()
[all …]
H A Dvmstat.c39 /* zero numa counters within a zone */
40 static void zero_zone_numa_counters(struct zone *zone) in zero_zone_numa_counters() argument
45 atomic_long_set(&zone->vm_numa_event[item], 0); in zero_zone_numa_counters()
47 per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_event[item] in zero_zone_numa_counters()
56 struct zone *zone; in zero_zones_numa_counters() local
58 for_each_populated_zone(zone) in zero_zones_numa_counters()
59 zero_zone_numa_counters(zone); in zero_zones_numa_counters()
161 * Manage combined zone base
172 fold_vm_zone_numa_events(struct zone * zone) fold_vm_zone_numa_events() argument
192 struct zone *zone; fold_vm_numa_events() local
201 calculate_pressure_threshold(struct zone * zone) calculate_pressure_threshold() argument
225 calculate_normal_threshold(struct zone * zone) calculate_normal_threshold() argument
278 struct zone *zone; refresh_zone_stat_thresholds() local
321 set_pgdat_percpu_threshold(pg_data_t * pgdat,int (* calculate_pressure)(struct zone *)) set_pgdat_percpu_threshold() argument
323 struct zone *zone; set_pgdat_percpu_threshold() local
345 __mod_zone_page_state(struct zone * zone,enum zone_stat_item item,long delta) __mod_zone_page_state() argument
435 __inc_zone_state(struct zone * zone,enum zone_stat_item item) __inc_zone_state() argument
491 __dec_zone_state(struct zone * zone,enum zone_stat_item item) __dec_zone_state() argument
560 mod_zone_state(struct zone * zone,enum zone_stat_item item,long delta,int overstep_mode) mod_zone_state() argument
599 mod_zone_page_state(struct zone * zone,enum zone_stat_item item,long delta) mod_zone_page_state() argument
695 mod_zone_page_state(struct zone * zone,enum zone_stat_item item,long delta) mod_zone_page_state() argument
709 struct zone *zone; inc_zone_page_state() local
817 struct zone *zone; refresh_cpu_vm_stats() local
907 struct zone *zone; cpu_vm_stats_fold() local
963 drain_zonestat(struct zone * zone,struct per_cpu_zonestat * pzstats) drain_zonestat() argument
1078 fill_contig_page_info(struct zone * zone,unsigned int suitable_order,struct contig_page_info * info) fill_contig_page_info() argument
1145 extfrag_for_order(struct zone * zone,unsigned int order) extfrag_for_order() argument
1159 fragmentation_index(struct zone * zone,unsigned int order) fragmentation_index() argument
1542 walk_zones_in_node(struct seq_file * m,pg_data_t * pgdat,bool assert_populated,bool nolock,void (* print)(struct seq_file * m,pg_data_t *,struct zone *)) walk_zones_in_node() argument
1544 struct zone *zone; walk_zones_in_node() local
1563 frag_show_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) frag_show_print() argument
1588 pagetypeinfo_showfree_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) pagetypeinfo_showfree_print() argument
1645 pagetypeinfo_showblockcount_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) pagetypeinfo_showblockcount_print() argument
1752 is_zone_first_populated(pg_data_t * pgdat,struct zone * zone) is_zone_first_populated() argument
1767 zoneinfo_show_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) zoneinfo_show_print() argument
2070 struct zone *zone; need_update() local
2331 unusable_show_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) unusable_show_print() argument
2381 extfrag_show_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) extfrag_show_print() argument
[all...]
H A Dcompaction.c73 * the "fragmentation score" of a node/zone.
126 static void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
128 zone->compact_considered = 0; in defer_compaction()
129 zone->compact_defer_shift++; in defer_compaction()
131 if (order < zone->compact_order_failed) in defer_compaction()
132 zone->compact_order_failed = order; in defer_compaction()
134 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction()
135 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction()
137 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction()
141 static bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
[all …]
H A Dmm_init.c69 struct zone *zone; in mminit_verify_zonelist() local
76 /* Identify the zone and nodelist */ in mminit_verify_zonelist()
80 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
81 if (!populated_zone(zone)) in mminit_verify_zonelist()
87 zone->name); in mminit_verify_zonelist()
90 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
91 pr_cont("%d:%s ", zone_to_nid(zone), zone->name); in mminit_verify_zonelist()
105 "Section %d Node %d Zone %d Lastcpupid %d Kasantag %d Gen %d Tier %d Flags %d\n", in mminit_verify_pageflags_layout()
115 "Section %d Node %d Zone %d Lastcpupid %d Kasantag %d\n", in mminit_verify_pageflags_layout()
122 "Section %lu Node %lu Zone %lu Lastcpupid %lu Kasantag %lu\n", in mminit_verify_pageflags_layout()
[all …]
/linux/tools/power/cpupower/lib/
H A Dpowercap.c124 static int sysfs_powercap_get64_val(struct powercap_zone *zone, in sysfs_powercap_get64_val() argument
132 strcat(file, zone->sys_name); in sysfs_powercap_get64_val()
146 int powercap_get_max_energy_range_uj(struct powercap_zone *zone, uint64_t *val) in powercap_get_max_energy_range_uj() argument
148 return sysfs_powercap_get64_val(zone, GET_MAX_ENERGY_RANGE_UJ, val); in powercap_get_max_energy_range_uj()
151 int powercap_get_energy_uj(struct powercap_zone *zone, uint64_t *val) in powercap_get_energy_uj() argument
153 return sysfs_powercap_get64_val(zone, GET_ENERGY_UJ, val); in powercap_get_energy_uj()
156 int powercap_get_max_power_range_uw(struct powercap_zone *zone, uint64_t *val) in powercap_get_max_power_range_uw() argument
158 return sysfs_powercap_get64_val(zone, GET_MAX_POWER_RANGE_UW, val); in powercap_get_max_power_range_uw()
161 int powercap_get_power_uw(struct powercap_zone *zone, uint64_t *val) in powercap_get_power_uw() argument
163 return sysfs_powercap_get64_val(zone, GET_POWER_UW, val); in powercap_get_power_uw()
[all …]
/linux/drivers/md/dm-vdo/indexer/
H A Dindex.c30 * get different numbers of records, some zones may fall behind others. Each time a zone fills up
32 * and also informs all other zones that it has closed the chapter. Each other zone will then close
34 * Once every zone has closed the chapter, the chapter writer will commit that chapter to storage.
36 * The last zone to close the chapter also removes the oldest chapter from the volume index.
38 * means that those zones will never ask the volume index about it. No zone is allowed to get more
39 * than one chapter ahead of any other. If a zone is so far ahead that it tries to close another
43 * request wants to add a chapter to the sparse cache, it sends a barrier message to each zone
44 * during the triage stage that acts as a rendezvous. Once every zone has reached the barrier and
45 * paused its operations, the cache membership is changed and each zone is then informed that it
48 * If a sparse cache has only one zone, it will not create a triage queue, but it still needs the
[all …]
/linux/fs/zonefs/
H A Dsuper.c32 * Get the name of a zone group directory.
48 * Manage the active zone count.
66 * If the zone is active, that is, if it is explicitly open or in zonefs_account_active()
79 /* The zone is not active. If it was, update the active count */ in zonefs_account_active()
87 * Manage the active zone count. Called with zi->i_truncate_mutex held.
97 * Execute a zone management operation.
105 * With ZNS drives, closing an explicitly open zone that has not been in zonefs_zone_mgmt()
106 * written will change the zone state to "closed", that is, the zone in zonefs_zone_mgmt()
108 * open operation on other zones if the drive active zone resources in zonefs_zone_mgmt()
109 * are exceeded, make sure that the zone does not remain active by in zonefs_zone_mgmt()
[all …]
/linux/fs/adfs/
H A Dmap.c14 * zone which contains a bitstream made up of variable sized fragments.
30 * large or fragmented files. The first map zone a fragment starts in
32 * from any zone on the disk.
63 * return the map bit offset of the fragment frag_id in the zone dm.
109 * Scan the free space map, for this zone, calculating the total
133 * exist in this zone. in scan_free_map()
159 static int scan_map(struct adfs_sb_info *asb, unsigned int zone, in scan_map() argument
166 dm = asb->s_map + zone; in scan_map()
167 zone = asb->s_map_size; in scan_map()
168 dm_end = asb->s_map + zone; in scan_map()
[all …]
/linux/include/linux/
H A Dmmzone.h150 NUMA_INTERLEAVE_HIT, /* interleaver preferred this zone */
739 * zone lock contention and keep cache-hot pages reusing.
828 * faulted, they come from the right zone right away. However, it is
832 * to a different zone. When migration fails - pinning fails.
853 * on different platforms may end up in a movable zone. ZERO_PAGE(0)
856 * memory to the MOVABLE zone, the vmemmap pages are also placed in
857 * such zone. Such pages cannot be really moved around as they are
879 struct zone { struct
882 /* zone watermarks, access with *_wmark_pages(zone) macros */
892 * wasting several GB of ram we must reserve some of the lower zone argument
[all …]
H A Dvmstat.h135 * Zone and node-based page accounting with per cpu differentials.
142 static inline void zone_numa_event_add(long x, struct zone *zone, in zone_numa_event_add() argument
145 atomic_long_add(x, &zone->vm_numa_event[item]); in zone_numa_event_add()
149 static inline unsigned long zone_numa_event_state(struct zone *zone, in zone_numa_event_state() argument
152 return atomic_long_read(&zone->vm_numa_event[item]); in zone_numa_event_state()
162 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument
165 atomic_long_add(x, &zone in zone_page_state_add()
204 zone_page_state(struct zone * zone,enum zone_stat_item item) zone_page_state() argument
221 zone_page_state_snapshot(struct zone * zone,enum zone_stat_item item) zone_page_state_snapshot() argument
240 __count_numa_event(struct zone * zone,enum numa_stat_item item) __count_numa_event() argument
248 __count_numa_events(struct zone * zone,enum numa_stat_item item,long delta) __count_numa_events() argument
312 __mod_zone_page_state(struct zone * zone,enum zone_stat_item item,long delta) __mod_zone_page_state() argument
335 __inc_zone_state(struct zone * zone,enum zone_stat_item item) __inc_zone_state() argument
347 __dec_zone_state(struct zone * zone,enum zone_stat_item item) __dec_zone_state() argument
407 drain_zonestat(struct zone * zone,struct per_cpu_zonestat * pzstats) drain_zonestat() argument
[all...]
H A Dpowercap.h16 * controlled. Each power zone can have one or more constraints.
28 * limits. If disabled power zone can only be monitored
75 * struct powercap_zone_ops - Define power zone callbacks
83 * @set_enable: Enable/Disable power zone controls.
92 * This structure defines zone callbacks to be implemented by client drivers.
112 * struct powercap_zone- Defines instance of a power cap zone
114 * @name: Power zone name.
115 * @control_type_inst: Control type instance for this zone.
116 * @ops: Pointer to the zone operation structure.
121 * @private_data: Private data pointer if any for this zone.
[all …]
/linux/Documentation/admin-guide/device-mapper/
H A Ddm-zoned.rst54 Data in these zones may be directly mapped to the conventional zone, but
55 later moved to a sequential zone so that the conventional zone can be
65 1) The first block of the first conventional zone found contains the
73 indicates the zone number of the device storing the chunk of data. Each
74 mapping entry may also indicate if the zone number of a conventional
75 zone used to buffer random modification to the data zone.
80 data chunk, a block is always valid only in the data zone mapping the
81 chunk or in the buffer zone of the chunk.
83 For a logical chunk mapped to a conventional zone, all write operations
84 are processed by directly writing to the zone. If the mapping zone is a
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dalloc.c250 struct mlx4_zone_entry *zone = kmalloc_obj(*zone); in mlx4_zone_add_one() local
252 if (NULL == zone) in mlx4_zone_add_one()
255 zone->flags = flags; in mlx4_zone_add_one()
256 zone->bitmap = bitmap; in mlx4_zone_add_one()
257 zone->use_rr = (flags & MLX4_ZONE_USE_RR) ? MLX4_USE_RR : 0; in mlx4_zone_add_one()
258 zone->priority = priority; in mlx4_zone_add_one()
259 zone->offset = offset; in mlx4_zone_add_one()
263 zone->uid = zone_alloc->last_uid++; in mlx4_zone_add_one()
264 zone->allocator = zone_alloc; in mlx4_zone_add_one()
274 list_add_tail(&zone->prio_list, &it->prio_list); in mlx4_zone_add_one()
[all …]
/linux/include/trace/events/
H A Dcompaction.h194 TP_PROTO(struct zone *zone,
198 TP_ARGS(zone, order, ret),
208 __entry->nid = zone_to_nid(zone);
209 __entry->idx = zone_idx(zone);
214 TP_printk("node=%d zone=%-8s order=%d ret=%s",
223 TP_PROTO(struct zone *zone,
227 TP_ARGS(zone, order, ret)
232 TP_PROTO(struct zone *zone,
236 TP_ARGS(zone, order, ret)
241 TP_PROTO(struct zone *zone, int order),
[all …]

12345678910>>...24