| /linux/drivers/block/null_blk/ |
| A D | zoned.c | 126 zone->capacity = zone->len; in null_init_zoned_dev() 127 zone->wp = zone->start + zone->len; in null_init_zoned_dev() 138 zone->start = zone->wp = sector; in null_init_zoned_dev() 140 zone->len = dev_capacity_sects - zone->start; in null_init_zoned_dev() 143 zone->capacity = in null_init_zoned_dev() 270 if (zone->wp == zone->start) { in __null_close_zone() 407 if (zone->wp + nr_sectors > zone->start + zone->capacity) { in null_zone_write() 439 if (zone->wp == zone->start + zone->capacity) { in null_zone_write() 552 zone->wp = zone->start + zone->len; in null_finish_zone() 590 zone->wp = zone->start; in null_reset_zone() [all …]
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| A D | ttm_memory.c | 88 zone->name, (unsigned long long)zone->used_mem >> 10); in ttm_mem_zone_kobj_release() 89 kfree(zone); in ttm_mem_zone_kobj_release() 319 zone->swap_limit = zone->max_mem - (mem >> 3); in ttm_mem_init_kernel_zone() 344 zone = kzalloc(sizeof(*zone), GFP_KERNEL); in ttm_mem_init_highmem_zone() 355 zone->swap_limit = zone->max_mem - (mem >> 3); in ttm_mem_init_highmem_zone() 361 zone->name); in ttm_mem_init_highmem_zone() 388 kfree(zone); in ttm_mem_init_dma32_zone() 403 zone->swap_limit = zone->max_mem - (mem >> 3); in ttm_mem_init_dma32_zone() 458 zone->name, (unsigned long long)zone->max_mem >> 10); in ttm_mem_global_init() 492 if (zone->used_mem > zone->swap_limit) { in ttm_check_swapping() [all …]
|
| /linux/fs/pstore/ |
| A D | zone.c | 413 zone->name, i, zone->off, in psz_kmsg_recover_meta() 421 zone->name, i, zone->off, in psz_kmsg_recover_meta() 444 zone->name, i, zone->off, in psz_kmsg_recover_meta() 453 zone->name, i, zone->off, in psz_kmsg_recover_meta() 489 if (!zone || zone->oldbuf) in psz_recover_zone() 516 zone->name, zone->off, zone->buffer_size); in psz_recover_zone() 523 zone->name, zone->off, zone->buffer_size, in psz_recover_zone() 529 zone->name, zone->off, zone->buffer_size, in psz_recover_zone() 652 if (zone && zone->buffer && buffer_datalen(zone)) in psz_ok() 763 zone->oldbuf = zone->buffer; in psz_kmsg_write_record() [all …]
|
| /linux/mm/ |
| A D | page_alloc.c | 152 struct zone *zone; member 2038 struct zone *zone; in deferred_init_memmap() local 2199 struct zone *zone; in page_alloc_init_late() local 2808 struct zone *zone; in unreserve_highatomic_pageblock() local 3108 struct zone *zone; in drain_pages() local 3498 struct zone *zone; in __isolate_free_page() local 4037 struct zone *zone; in get_page_from_freelist() local 4491 struct zone *zone; in should_compact_retry() local 4635 struct zone *zone; in wake_all_kswapds() local 4746 struct zone *zone; in should_reclaim_retry() local [all …]
|
| A D | vmstat.c | 54 struct zone *zone; in zero_zones_numa_counters() local 189 struct zone *zone; in fold_vm_numa_events() local 275 struct zone *zone; in refresh_zone_stat_thresholds() local 320 struct zone *zone; in set_pgdat_percpu_threshold() local 716 struct zone *zone; in inc_zone_page_state() local 821 struct zone *zone; in refresh_cpu_vm_stats() local 907 struct zone *zone; in cpu_vm_stats_fold() local 1429 struct zone *zone; in walk_zones_in_node() local 1448 struct zone *zone) in frag_show_print() argument 1652 struct zone *zone) in zoneinfo_show_print() argument [all …]
|
| A D | compaction.c | 423 struct zone *zone = cc->zone; in update_cached_migrate() local 445 struct zone *zone = cc->zone; in update_pageblock_skip() local 1543 struct zone *zone = cc->zone; in isolate_freepages() local 2245 struct zone *zone; in compaction_zonelist_suitable() local 2496 .zone = zone, in compact_zone_order() 2562 struct zone *zone; in try_to_compact_pages() local 2636 struct zone *zone; in proactive_compact_node() local 2651 cc.zone = zone; in proactive_compact_node() 2665 struct zone *zone; in compact_node() local 2681 cc.zone = zone; in compact_node() [all …]
|
| A D | memory_hotplug.c | 435 struct zone *zone; in update_pgdat_span() local 618 struct zone *zone, struct memory_notify *arg) in node_states_check_changes_online() argument 724 struct zone *zone) in auto_movable_stats_account_zone() argument 779 struct zone *zone; in auto_movable_can_online_movable() local 833 struct zone *zone = &pgdat->node_zones[zid]; in default_kernel_zone_for_pfn() local 1000 struct zone *zone = page_zone(page); in adjust_present_page_count() local 1019 struct zone *zone) in mhp_init_memmap_on_memory() argument 1600 struct zone *zone = NULL; in test_pages_in_a_zone() local 1612 if (zone && !zone_spans_pfn(zone, pfn)) in test_pages_in_a_zone() 1615 if (zone && page_zone(page) != zone) in test_pages_in_a_zone() [all …]
|
| A D | page_isolation.c | 20 struct zone *zone = page_zone(page); in set_migratetype_isolate() local 24 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 32 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 46 zone->nr_isolate_pageblock++; in set_migratetype_isolate() 69 struct zone *zone; in unset_migratetype_isolate() local 76 zone = page_zone(page); in unset_migratetype_isolate() 77 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate() 125 zone->nr_isolate_pageblock--; in unset_migratetype_isolate() 275 struct zone *zone; in test_pages_isolated() local 295 zone = page_zone(page); in test_pages_isolated() [all …]
|
| A D | page_reporting.c | 114 page_reporting_cycle(struct page_reporting_dev_info *prdev, struct zone *zone, in page_reporting_cycle() argument 132 spin_lock_irq(&zone->lock); in page_reporting_cycle() 190 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 202 spin_lock_irq(&zone->lock); in page_reporting_cycle() 222 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 229 struct scatterlist *sgl, struct zone *zone) in page_reporting_process_zone() argument 236 watermark = low_wmark_pages(zone) + in page_reporting_process_zone() 267 spin_lock_irq(&zone->lock); in page_reporting_process_zone() 269 spin_unlock_irq(&zone->lock); in page_reporting_process_zone() 282 struct zone *zone; in page_reporting_process() local [all …]
|
| /linux/include/linux/ |
| A D | mmzone.h | 703 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn() 764 struct zone *zone; /* Pointer to actual zone */ member 988 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) argument 1008 static inline bool managed_zone(struct zone *zone) in managed_zone() argument 1020 static inline int zone_to_nid(struct zone *zone) in zone_to_nid() argument 1030 static inline int zone_to_nid(struct zone *zone) in zone_to_nid() argument 1057 static inline int is_highmem(struct zone *zone) in is_highmem() argument 1105 extern struct zone *next_zone(struct zone *zone); 1125 zone = next_zone(zone)) 1130 zone = next_zone(zone)) \ [all …]
|
| A D | memory_hotplug.h | 11 struct zone; 81 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument 89 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument 93 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument 97 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument 106 struct zone *zone); 109 struct zone *zone, struct memory_group *group); 242 static inline void zone_seqlock_init(struct zone *zone) {} in zone_seqlock_init() argument 322 extern void set_zone_contiguous(struct zone *zone); 323 extern void clear_zone_contiguous(struct zone *zone); [all …]
|
| A D | vmstat.h | 145 static inline void zone_numa_event_add(long x, struct zone *zone, in zone_numa_event_add() argument 152 static inline unsigned long zone_numa_event_state(struct zone *zone, in zone_numa_event_state() argument 165 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument 207 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument 224 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument 243 __count_numa_event(struct zone *zone, enum numa_stat_item item) in __count_numa_event() argument 307 void drain_zonestat(struct zone *zone, struct per_cpu_zonestat *); 309 int calculate_pressure_threshold(struct zone *zone); 310 int calculate_normal_threshold(struct zone *zone); 319 static inline void __mod_zone_page_state(struct zone *zone, in __mod_zone_page_state() argument [all …]
|
| /linux/drivers/md/ |
| A D | dm-zoned-metadata.c | 226 return zone->id - zone->dev->zone_offset; in dmz_dev_zone_id() 313 if (!zone) in dmz_insert() 327 return zone; in dmz_insert() 1427 zmd->sb[0].zone = zone; in dmz_init_zone() 1648 zone->id, zone->wp_block, wp); in dmz_handle_seq_write_err() 1651 dmz_invalidate_blocks(zmd, zone, zone->wp_block, in dmz_handle_seq_write_err() 2532 zone->id, zone->weight, in dmz_validate_blocks() 2612 zone->id, zone->weight, n); in dmz_invalidate_blocks() 2928 zone = dmz_get(zmd, zmd->sb[0].zone->id + i); in dmz_ctr_metadata() 2935 if (!dmz_is_rnd(zone) && !dmz_is_cache(zone)) { in dmz_ctr_metadata() [all …]
|
| A D | dm-zoned-target.c | 86 struct dm_zone *zone = bioctx->zone; in dmz_bio_endio() local 88 if (zone) { in dmz_bio_endio() 191 zone->id, in dmz_handle_read() 198 if (dmz_is_rnd(zone) || dmz_is_cache(zone) || in dmz_handle_read() 326 if (!zone) in dmz_handle_write() 334 zone->id, in dmz_handle_write() 337 if (dmz_is_rnd(zone) || dmz_is_cache(zone) || in dmz_handle_write() 368 if (!zone) in dmz_handle_discard() 384 if (dmz_is_rnd(zone) || dmz_is_cache(zone) || in dmz_handle_discard() 422 bioctx->zone = zone; in dmz_handle_bio() [all …]
|
| /linux/fs/btrfs/ |
| A D | reada.c | 240 zone = NULL; in reada_find_zone() 247 return zone; in reada_find_zone() 260 zone = kzalloc(sizeof(*zone), GFP_KERNEL); in reada_find_zone() 261 if (!zone) in reada_find_zone() 266 kfree(zone); in reada_find_zone() 287 zone); in reada_find_zone() 290 kfree(zone); in reada_find_zone() 301 return zone; in reada_find_zone() 370 if (!zone) in reada_find_extent() 548 kfree(zone); in reada_zone_release() [all …]
|
| /linux/kernel/power/ |
| A D | snapshot.c | 622 struct zone *zone; in create_mem_extents() local 747 zone = bm->cur.zone; in memory_bm_find_bit() 749 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit() 777 if (zone == bm->cur.zone && in memory_bm_find_bit() 795 bm->cur.zone = zone; in memory_bm_find_bit() 1242 struct zone *zone; in count_free_highmem_pages() local 1290 struct zone *zone; in count_highmem_pages() local 1358 struct zone *zone; in count_data_pages() local 1454 struct zone *zone; in copy_data_pages() local 1741 struct zone *zone; in hibernate_preallocate_memory() local [all …]
|
| /linux/include/net/netfilter/ |
| A D | nf_conntrack_zones.h | 12 return &ct->zone; in nf_ct_zone() 21 zone->id = id; in nf_ct_zone_init() 22 zone->flags = flags; in nf_ct_zone_init() 23 zone->dir = dir; in nf_ct_zone_init() 25 return zone; in nf_ct_zone_init() 36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 43 const struct nf_conntrack_zone *zone) in nf_ct_zone_add() argument 46 ct->zone = *zone; in nf_ct_zone_add() 53 return zone->dir & (1 << dir); in nf_ct_zone_matches_dir() 60 return nf_ct_zone_matches_dir(zone, dir) ? in nf_ct_zone_id() [all …]
|
| /linux/fs/adfs/ |
| A D | map.c | 179 } while (--zone > 0); in scan_map() 202 unsigned int zone; in adfs_map_statfs() local 209 } while (--zone > 0); in adfs_map_statfs() 322 for (zone = 1; zone < nzones; zone++) { in adfs_map_layout() 324 dm[zone].dm_startblk = zone * zone_size - ADFS_DR_SIZE_BITS; in adfs_map_layout() 337 unsigned int zone; in adfs_map_read() local 339 for (zone = 0; zone < nzones; zone++) { in adfs_map_read() 340 dm[zone].dm_bh = sb_bread(sb, map_addr + zone); in adfs_map_read() 341 if (!dm[zone].dm_bh) in adfs_map_read() 350 unsigned int zone; in adfs_map_relse() local [all …]
|
| /linux/drivers/thermal/tegra/ |
| A D | tegra-bpmp-thermal.c | 43 req.get_temp.zone = zone->idx; in tegra_bpmp_thermal_get_temp() 69 req.set_trip.zone = zone->idx; in tegra_bpmp_thermal_set_trips() 84 struct tegra_bpmp_thermal_zone *zone; in tz_device_update_work_fn() local 118 req->host_trip_reached.zone); in bpmp_mrq_thermal() 185 zone = devm_kzalloc(&pdev->dev, sizeof(*zone), GFP_KERNEL); in tegra_bpmp_thermal_probe() 186 if (!zone) in tegra_bpmp_thermal_probe() 189 zone->idx = i; in tegra_bpmp_thermal_probe() 190 zone->tegra = tegra; in tegra_bpmp_thermal_probe() 194 devm_kfree(&pdev->dev, zone); in tegra_bpmp_thermal_probe() 203 devm_kfree(&pdev->dev, zone); in tegra_bpmp_thermal_probe() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| A D | alloc.c | 250 struct mlx4_zone_entry *zone = kmalloc(sizeof(*zone), GFP_KERNEL); in mlx4_zone_add_one() local 252 if (NULL == zone) in mlx4_zone_add_one() 279 *puid = zone->uid; in mlx4_zone_add_one() 328 kfree(zone); in mlx4_zone_allocator_destroy() 349 uid = zone->uid; in __mlx4_alloc_from_zone() 421 mlx4_bitmap_free_range(zone->bitmap, obj - zone->offset, count, zone->use_rr); in __mlx4_free_from_zone() 432 return zone; in __mlx4_find_zone_by_uid() 447 bitmap = zone == NULL ? NULL : zone->bitmap; in mlx4_zone_get_bitmap() 463 if (NULL == zone) { in mlx4_zone_remove_one() 472 kfree(zone); in mlx4_zone_remove_one() [all …]
|
| /linux/include/trace/events/ |
| A D | compaction.h | 199 TP_PROTO(struct zone *zone, 228 TP_PROTO(struct zone *zone, 237 TP_PROTO(struct zone *zone, 246 TP_PROTO(struct zone *zone, int order), 248 TP_ARGS(zone, order), 279 TP_PROTO(struct zone *zone, int order), 281 TP_ARGS(zone, order) 286 TP_PROTO(struct zone *zone, int order), 288 TP_ARGS(zone, order) 293 TP_PROTO(struct zone *zone, int order), [all …]
|
| /linux/virt/kvm/ |
| A D | coalesced_mmio.c | 36 if (addr < dev->zone.addr) in coalesced_mmio_in_range() 38 if (addr + len > dev->zone.addr + dev->zone.size) in coalesced_mmio_in_range() 138 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_register_coalesced_mmio() argument 143 if (zone->pio != 1 && zone->pio != 0) in kvm_vm_ioctl_register_coalesced_mmio() 153 dev->zone = *zone; in kvm_vm_ioctl_register_coalesced_mmio() 157 zone->pio ? KVM_PIO_BUS : KVM_MMIO_BUS, in kvm_vm_ioctl_register_coalesced_mmio() 158 zone->addr, zone->size, &dev->dev); in kvm_vm_ioctl_register_coalesced_mmio() 174 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_unregister_coalesced_mmio() argument 179 if (zone->pio != 1 && zone->pio != 0) in kvm_vm_ioctl_unregister_coalesced_mmio() 185 if (zone->pio == dev->zone.pio && in kvm_vm_ioctl_unregister_coalesced_mmio() [all …]
|
| /linux/drivers/mtd/ |
| A D | sm_ftl.c | 195 WARN_ON(zone < 0 || zone >= ftl->zone_count); in sm_mkoffset() 287 block, zone, ret); in sm_read_sector() 314 block, zone); in sm_read_sector() 353 block, zone, ret); in sm_write_sector() 856 zone->initialized = 1; in sm_init_zone() 882 struct ftl_zone *zone; in sm_get_zone() local 894 return zone; in sm_get_zone() 932 struct ftl_zone *zone; in sm_cache_flush() local 1017 struct ftl_zone *zone; in sm_read() local 1026 if (IS_ERR(zone)) { in sm_read() [all …]
|
| /linux/drivers/thermal/intel/int340x_thermal/ |
| A D | int340x_thermal_zone.c | 17 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_zone_temp() 22 return d->override_ops->get_temp(zone, temp); in int340x_thermal_get_zone_temp() 46 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_temp() 50 return d->override_ops->get_trip_temp(zone, trip, temp); in int340x_thermal_get_trip_temp() 79 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_type() 83 return d->override_ops->get_trip_type(zone, trip, type); in int340x_thermal_get_trip_type() 111 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_set_trip_temp() 133 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_hyst() 151 dev_dbg(&zone->device, "%s: critical temperature reached\n", zone->type); in int340x_thermal_critical() 268 if (IS_ERR(int34x_thermal_zone->zone)) { in int340x_thermal_zone_add() [all …]
|
| /linux/Documentation/filesystems/ |
| A D | zonefs.rst | 78 the zone containing the super block is a sequential zone, the mkzonefs format 79 tool always "finishes" the zone, that is, it transitions the zone to a full 90 the device only has a single conventional zone at sector 0, the zone will not 122 Conventional zone files 125 The size of conventional zone files is fixed to the size of the zone they 132 Sequential zone files 136 the file's zone write pointer position relative to the zone start sector. 159 the zone, or up to the zone capacity, in which case the file's zone is 314 * zone-ro 315 * zone-offline [all …]
|