/linux-6.3-rc2/drivers/block/null_blk/ |
A D | zoned.c | 129 zone->capacity = zone->len; in null_init_zoned_dev() 130 zone->wp = zone->start + zone->len; in null_init_zoned_dev() 141 zone->start = zone->wp = sector; in null_init_zoned_dev() 143 zone->len = dev_capacity_sects - zone->start; in null_init_zoned_dev() 273 if (zone->wp == zone->start) { in __null_close_zone() 412 if (zone->wp + nr_sectors > zone->start + zone->capacity) { in null_zone_write() 444 if (zone->wp == zone->start + zone->capacity) { in null_zone_write() 557 zone->wp = zone->start + zone->len; in null_finish_zone() 595 zone->wp = zone->start; in null_reset_zone() 719 zone->wp = zone->start; in null_set_zone_cond() [all …]
|
/linux-6.3-rc2/fs/pstore/ |
A D | zone.c | 413 zone->name, i, zone->off, in psz_kmsg_recover_meta() 421 zone->name, i, zone->off, in psz_kmsg_recover_meta() 444 zone->name, i, zone->off, in psz_kmsg_recover_meta() 453 zone->name, i, zone->off, in psz_kmsg_recover_meta() 489 if (!zone || zone->oldbuf) in psz_recover_zone() 516 zone->name, zone->off, zone->buffer_size); in psz_recover_zone() 523 zone->name, zone->off, zone->buffer_size, in psz_recover_zone() 529 zone->name, zone->off, zone->buffer_size, in psz_recover_zone() 652 if (zone && zone->buffer && buffer_datalen(zone)) in psz_ok() 763 zone->oldbuf = zone->buffer; in psz_kmsg_write_record() [all …]
|
/linux-6.3-rc2/mm/ |
A D | page_alloc.c | 2114 struct zone *zone; in deferred_init_memmap() local 2275 struct zone *zone; in page_alloc_init_late() local 2938 struct zone *zone; in unreserve_highatomic_pageblock() local 3219 struct zone *zone; in drain_pages() local 3478 struct zone *zone; in free_unref_page() local 4212 struct zone *zone; in get_page_from_freelist() local 4672 struct zone *zone; in should_compact_retry() local 4842 struct zone *zone; in wake_all_kswapds() local 4963 struct zone *zone; in should_reclaim_retry() local 5410 struct zone *zone; in __alloc_pages_bulk() local [all …]
|
A D | vmstat.c | 54 struct zone *zone; in zero_zones_numa_counters() local 189 struct zone *zone; in fold_vm_numa_events() local 275 struct zone *zone; in refresh_zone_stat_thresholds() local 320 struct zone *zone; in set_pgdat_percpu_threshold() local 704 struct zone *zone; in inc_zone_page_state() local 809 struct zone *zone; in refresh_cpu_vm_stats() local 895 struct zone *zone; in cpu_vm_stats_fold() local 1441 struct zone *zone; in walk_zones_in_node() local 1460 struct zone *zone) in frag_show_print() argument 1664 struct zone *zone) in zoneinfo_show_print() argument [all …]
|
A D | compaction.c | 416 struct zone *zone = cc->zone; in update_cached_migrate() local 438 struct zone *zone = cc->zone; in update_pageblock_skip() local 1560 struct zone *zone = cc->zone; in isolate_freepages() local 2268 struct zone *zone; in compaction_zonelist_suitable() local 2531 .zone = zone, in compact_zone_order() 2594 struct zone *zone; in try_to_compact_pages() local 2668 struct zone *zone; in proactive_compact_node() local 2683 cc.zone = zone; in proactive_compact_node() 2699 struct zone *zone; in compact_node() local 2715 cc.zone = zone; in compact_node() [all …]
|
A D | memory_hotplug.c | 408 zone->spanned_pages = zone_end_pfn(zone) - pfn; in shrink_zone_span() 424 zone->spanned_pages = pfn - zone->zone_start_pfn + 1; in shrink_zone_span() 435 struct zone *zone; in update_pgdat_span() local 620 struct zone *zone, struct memory_notify *arg) in node_states_check_changes_online() argument 732 struct zone *zone) in auto_movable_stats_account_zone() argument 787 struct zone *zone; in auto_movable_can_online_movable() local 841 struct zone *zone = &pgdat->node_zones[zid]; in default_kernel_zone_for_pfn() local 1008 struct zone *zone = page_zone(page); in adjust_present_page_count() local 1027 struct zone *zone) in mhp_init_memmap_on_memory() argument 1733 struct zone *zone, struct memory_notify *arg) in node_states_check_changes_offline() argument [all …]
|
A D | page_isolation.c | 37 struct zone *zone = page_zone(page); in has_unmovable_pages() local 72 if (zone_idx(zone) == ZONE_MOVABLE) in has_unmovable_pages() 150 struct zone *zone = page_zone(page); in set_migratetype_isolate() local 185 zone->nr_isolate_pageblock++; in set_migratetype_isolate() 208 struct zone *zone; in unset_migratetype_isolate() local 214 zone = page_zone(page); in unset_migratetype_isolate() 261 zone->nr_isolate_pageblock--; in unset_migratetype_isolate() 312 struct zone *zone; in isolate_single_pageblock() local 330 zone->zone_start_pfn); in isolate_single_pageblock() 640 struct zone *zone; in test_pages_isolated() local [all …]
|
A D | page_reporting.c | 146 page_reporting_cycle(struct page_reporting_dev_info *prdev, struct zone *zone, in page_reporting_cycle() argument 164 spin_lock_irq(&zone->lock); in page_reporting_cycle() 222 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 234 spin_lock_irq(&zone->lock); in page_reporting_cycle() 254 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 261 struct scatterlist *sgl, struct zone *zone) in page_reporting_process_zone() argument 268 watermark = low_wmark_pages(zone) + in page_reporting_process_zone() 299 spin_lock_irq(&zone->lock); in page_reporting_process_zone() 301 spin_unlock_irq(&zone->lock); in page_reporting_process_zone() 314 struct zone *zone; in page_reporting_process() local [all …]
|
A D | mmzone.c | 30 struct zone *next_zone(struct zone *zone) in next_zone() argument 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 35 zone++; in next_zone() 39 zone = pgdat->node_zones; in next_zone() 41 zone = NULL; in next_zone() 43 return zone; in next_zone() 69 (z->zone && !zref_in_nodemask(z, nodes))) in __next_zones_zonelist()
|
/linux-6.3-rc2/include/linux/ |
A D | memory_hotplug.h | 11 struct zone; 130 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument 138 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument 142 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument 146 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument 155 struct zone *zone); 158 struct zone *zone, struct memory_group *group); 307 struct zone *zone, struct memory_group *group); 316 struct zone *zone, struct memory_group *group) in offline_pages() argument 329 extern void set_zone_contiguous(struct zone *zone); [all …]
|
A D | mmzone.h | 1011 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn() 1178 struct zone *zone; /* Pointer to actual zone */ member 1446 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) argument 1466 static inline bool managed_zone(struct zone *zone) in managed_zone() argument 1478 static inline int zone_to_nid(struct zone *zone) in zone_to_nid() argument 1488 static inline int zone_to_nid(struct zone *zone) in zone_to_nid() argument 1515 static inline int is_highmem(struct zone *zone) in is_highmem() argument 1567 extern struct zone *next_zone(struct zone *zone); 1587 zone = next_zone(zone)) 1592 zone = next_zone(zone)) \ [all …]
|
A D | vmstat.h | 139 static inline void zone_numa_event_add(long x, struct zone *zone, in zone_numa_event_add() argument 146 static inline unsigned long zone_numa_event_state(struct zone *zone, in zone_numa_event_state() argument 159 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument 201 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument 218 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument 237 __count_numa_event(struct zone *zone, enum numa_stat_item item) in __count_numa_event() argument 301 void drain_zonestat(struct zone *zone, struct per_cpu_zonestat *); 303 int calculate_pressure_threshold(struct zone *zone); 304 int calculate_normal_threshold(struct zone *zone); 313 static inline void __mod_zone_page_state(struct zone *zone, in __mod_zone_page_state() argument [all …]
|
/linux-6.3-rc2/tools/power/cpupower/lib/ |
A D | powercap.c | 115 strcat(file, zone->sys_name); in sysfs_powercap_get64_val() 158 strcat(path, zone->sys_name); in powercap_zone_get_enabled() 189 strcat(file, zone->sys_name); in powercap_read_zone() 192 if (zone->parent) in powercap_read_zone() 193 zone->tree_depth = zone->parent->tree_depth + 1; in powercap_read_zone() 196 zone->has_energy_uj = 1; in powercap_read_zone() 199 zone->has_power_uw = 1; in powercap_read_zone() 232 child_zone->parent = zone; in powercap_read_zone() 278 if (!zone) in powercap_walk_zones() 281 ret = f(zone); in powercap_walk_zones() [all …]
|
/linux-6.3-rc2/drivers/md/ |
A D | dm-zoned-metadata.c | 226 return zone->id - zone->dev->zone_offset; in dmz_dev_zone_id() 313 if (!zone) in dmz_insert() 327 return zone; in dmz_insert() 1409 zmd->sb[0].zone = zone; in dmz_init_zone() 1630 zone->id, zone->wp_block, wp); in dmz_handle_seq_write_err() 1633 dmz_invalidate_blocks(zmd, zone, zone->wp_block, in dmz_handle_seq_write_err() 2515 zone->id, zone->weight, in dmz_validate_blocks() 2595 zone->id, zone->weight, n); in dmz_invalidate_blocks() 2911 zone = dmz_get(zmd, zmd->sb[0].zone->id + i); in dmz_ctr_metadata() 2918 if (!dmz_is_rnd(zone) && !dmz_is_cache(zone)) { in dmz_ctr_metadata() [all …]
|
A D | dm-zoned-target.c | 86 struct dm_zone *zone = bioctx->zone; in dmz_bio_endio() local 88 if (zone) { in dmz_bio_endio() 190 zone->id, in dmz_handle_read() 197 if (dmz_is_rnd(zone) || dmz_is_cache(zone) || in dmz_handle_read() 325 if (!zone) in dmz_handle_write() 333 zone->id, in dmz_handle_write() 336 if (dmz_is_rnd(zone) || dmz_is_cache(zone) || in dmz_handle_write() 367 if (!zone) in dmz_handle_discard() 383 if (dmz_is_rnd(zone) || dmz_is_cache(zone) || in dmz_handle_discard() 421 bioctx->zone = zone; in dmz_handle_bio() [all …]
|
/linux-6.3-rc2/include/net/netfilter/ |
A D | nf_conntrack_zones.h | 12 return &ct->zone; in nf_ct_zone() 21 zone->id = id; in nf_ct_zone_init() 22 zone->flags = flags; in nf_ct_zone_init() 23 zone->dir = dir; in nf_ct_zone_init() 25 return zone; in nf_ct_zone_init() 36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 43 const struct nf_conntrack_zone *zone) in nf_ct_zone_add() argument 46 ct->zone = *zone; in nf_ct_zone_add() 53 return zone->dir & (1 << dir); in nf_ct_zone_matches_dir() 60 return nf_ct_zone_matches_dir(zone, dir) ? in nf_ct_zone_id() [all …]
|
/linux-6.3-rc2/kernel/power/ |
A D | snapshot.c | 626 struct zone *zone; in create_mem_extents() local 751 zone = bm->cur.zone; in memory_bm_find_bit() 753 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit() 781 if (zone == bm->cur.zone && in memory_bm_find_bit() 799 bm->cur.zone = zone; in memory_bm_find_bit() 1239 struct zone *zone; in count_free_highmem_pages() local 1287 struct zone *zone; in count_highmem_pages() local 1355 struct zone *zone; in count_data_pages() local 1451 struct zone *zone; in copy_data_pages() local 1738 struct zone *zone; in hibernate_preallocate_memory() local [all …]
|
/linux-6.3-rc2/fs/adfs/ |
A D | map.c | 179 } while (--zone > 0); in scan_map() 202 unsigned int zone; in adfs_map_statfs() local 209 } while (--zone > 0); in adfs_map_statfs() 322 for (zone = 1; zone < nzones; zone++) { in adfs_map_layout() 324 dm[zone].dm_startblk = zone * zone_size - ADFS_DR_SIZE_BITS; in adfs_map_layout() 337 unsigned int zone; in adfs_map_read() local 339 for (zone = 0; zone < nzones; zone++) { in adfs_map_read() 340 dm[zone].dm_bh = sb_bread(sb, map_addr + zone); in adfs_map_read() 341 if (!dm[zone].dm_bh) in adfs_map_read() 350 unsigned int zone; in adfs_map_relse() local [all …]
|
/linux-6.3-rc2/include/trace/events/ |
A D | compaction.h | 183 TP_PROTO(struct zone *zone, 212 TP_PROTO(struct zone *zone, 221 TP_PROTO(struct zone *zone, 230 TP_PROTO(struct zone *zone, int order), 232 TP_ARGS(zone, order), 263 TP_PROTO(struct zone *zone, int order), 265 TP_ARGS(zone, order) 270 TP_PROTO(struct zone *zone, int order), 272 TP_ARGS(zone, order) 277 TP_PROTO(struct zone *zone, int order), [all …]
|
/linux-6.3-rc2/virt/kvm/ |
A D | coalesced_mmio.c | 36 if (addr < dev->zone.addr) in coalesced_mmio_in_range() 38 if (addr + len > dev->zone.addr + dev->zone.size) in coalesced_mmio_in_range() 138 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_register_coalesced_mmio() argument 143 if (zone->pio != 1 && zone->pio != 0) in kvm_vm_ioctl_register_coalesced_mmio() 153 dev->zone = *zone; in kvm_vm_ioctl_register_coalesced_mmio() 157 zone->pio ? KVM_PIO_BUS : KVM_MMIO_BUS, in kvm_vm_ioctl_register_coalesced_mmio() 158 zone->addr, zone->size, &dev->dev); in kvm_vm_ioctl_register_coalesced_mmio() 174 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_unregister_coalesced_mmio() argument 179 if (zone->pio != 1 && zone->pio != 0) in kvm_vm_ioctl_unregister_coalesced_mmio() 185 if (zone->pio == dev->zone.pio && in kvm_vm_ioctl_unregister_coalesced_mmio() [all …]
|
/linux-6.3-rc2/drivers/thermal/tegra/ |
A D | tegra-bpmp-thermal.c | 43 req.get_temp.zone = zone->idx; in __tegra_bpmp_thermal_get_temp() 77 req.set_trip.zone = zone->idx; in tegra_bpmp_thermal_set_trips() 98 struct tegra_bpmp_thermal_zone *zone; in tz_device_update_work_fn() local 133 req.host_trip_reached.zone); in bpmp_mrq_thermal() 202 zone = devm_kzalloc(&pdev->dev, sizeof(*zone), GFP_KERNEL); in tegra_bpmp_thermal_probe() 203 if (!zone) in tegra_bpmp_thermal_probe() 206 zone->idx = i; in tegra_bpmp_thermal_probe() 207 zone->tegra = tegra; in tegra_bpmp_thermal_probe() 211 devm_kfree(&pdev->dev, zone); in tegra_bpmp_thermal_probe() 220 devm_kfree(&pdev->dev, zone); in tegra_bpmp_thermal_probe() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/mellanox/mlx4/ |
A D | alloc.c | 250 struct mlx4_zone_entry *zone = kmalloc(sizeof(*zone), GFP_KERNEL); in mlx4_zone_add_one() local 252 if (NULL == zone) in mlx4_zone_add_one() 279 *puid = zone->uid; in mlx4_zone_add_one() 328 kfree(zone); in mlx4_zone_allocator_destroy() 349 uid = zone->uid; in __mlx4_alloc_from_zone() 421 mlx4_bitmap_free_range(zone->bitmap, obj - zone->offset, count, zone->use_rr); in __mlx4_free_from_zone() 432 return zone; in __mlx4_find_zone_by_uid() 447 bitmap = zone == NULL ? NULL : zone->bitmap; in mlx4_zone_get_bitmap() 463 if (NULL == zone) { in mlx4_zone_remove_one() 472 kfree(zone); in mlx4_zone_remove_one() [all …]
|
/linux-6.3-rc2/drivers/mtd/ |
A D | sm_ftl.c | 195 WARN_ON(zone < 0 || zone >= ftl->zone_count); in sm_mkoffset() 287 block, zone, ret); in sm_read_sector() 314 block, zone); in sm_read_sector() 353 block, zone, ret); in sm_write_sector() 856 zone->initialized = 1; in sm_init_zone() 882 struct ftl_zone *zone; in sm_get_zone() local 894 return zone; in sm_get_zone() 932 struct ftl_zone *zone; in sm_cache_flush() local 1017 struct ftl_zone *zone; in sm_read() local 1026 if (IS_ERR(zone)) { in sm_read() [all …]
|
/linux-6.3-rc2/arch/x86/mm/ |
A D | highmem_32.c | 9 struct zone *zone; in set_highmem_pages_init() local 17 for_each_zone(zone) { in set_highmem_pages_init() 20 if (!is_highmem(zone)) in set_highmem_pages_init() 23 zone_start_pfn = zone->zone_start_pfn; in set_highmem_pages_init() 24 zone_end_pfn = zone_start_pfn + zone->spanned_pages; in set_highmem_pages_init() 26 nid = zone_to_nid(zone); in set_highmem_pages_init() 28 zone->name, nid, zone_start_pfn, zone_end_pfn); in set_highmem_pages_init()
|
/linux-6.3-rc2/Documentation/filesystems/ |
A D | zonefs.rst | 78 the zone containing the super block is a sequential zone, the mkzonefs format 79 tool always "finishes" the zone, that is, it transitions the zone to a full 90 the device only has a single conventional zone at sector 0, the zone will not 122 Conventional zone files 125 The size of conventional zone files is fixed to the size of the zone they 132 Sequential zone files 136 the file's zone write pointer position relative to the zone start sector. 159 the zone, or up to the zone capacity, in which case the file's zone is 321 * zone-ro 322 * zone-offline [all …]
|