/linux-6.3-rc2/include/linux/ |
A D | oom.h | 13 struct zonelist; 31 struct zonelist *zonelist; member
|
A D | mmzone.h | 1196 struct zonelist { struct 1261 struct zonelist node_zonelists[MAX_ZONELISTS]; 1657 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1661 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist() 1704 struct zonelist *zonelist; in movable_only_nodes() local 1717 zonelist = &NODE_DATA(nid)->node_zonelists[ZONELIST_FALLBACK]; in movable_only_nodes() 1718 z = first_zones_zonelist(zonelist, ZONE_NORMAL, nodes); in movable_only_nodes()
|
A D | swap.h | 413 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
A D | gfp.h | 165 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
/linux-6.3-rc2/mm/ |
A D | mm_init.c | 34 struct zonelist *zonelist; in mminit_verify_zonelist() local 43 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 54 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
A D | page_alloc.c | 2935 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 4427 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 4684 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_compact_retry() 4992 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_reclaim_retry() 5111 struct zoneref *z = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 5346 ac->zonelist = node_zonelist(preferred_nid, gfp_mask); in prepare_alloc_pages() 5936 struct zonelist *zonelist = node_zonelist(numa_node_id(), GFP_KERNEL); in nr_free_zone_pages() local 5938 for_each_zone_zonelist(zone, z, zonelist, offset) { in nr_free_zone_pages() 9489 struct zonelist *zonelist; in alloc_contig_pages() local 9493 zonelist = node_zonelist(nid, gfp_mask); in alloc_contig_pages() [all …]
|
A D | vmscan.c | 6632 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 6653 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 6755 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 6773 shrink_zones(zonelist, sc); in do_try_to_free_pages() 6893 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 6931 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 6973 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 7008 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() 7082 struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask); in try_to_free_mem_cgroup_pages() local 7088 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_mem_cgroup_pages() [all …]
|
A D | internal.h | 219 struct zonelist *zonelist; member
|
A D | oom_kill.c | 270 if (!oc->zonelist) in constrained_alloc() 294 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
|
A D | mempolicy.c | 1931 struct zonelist *zonelist; in mempolicy_slab_node() local 1933 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 1934 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
A D | slab.c | 3072 struct zonelist *zonelist; in fallback_alloc() local 3086 zonelist = node_zonelist(mempolicy_slab_node(), flags); in fallback_alloc() 3093 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in fallback_alloc()
|
A D | compaction.c | 2275 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in compaction_zonelist_suitable() 2607 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in try_to_compact_pages()
|
A D | slub.c | 2318 struct zonelist *zonelist; in get_any_partial() local 2349 zonelist = node_zonelist(mempolicy_slab_node(), pc->flags); in get_any_partial() 2350 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in get_any_partial()
|
A D | hugetlb.c | 1314 struct zonelist *zonelist; in dequeue_hugetlb_folio_nodemask() local 1319 zonelist = node_zonelist(nid, gfp_mask); in dequeue_hugetlb_folio_nodemask() 1323 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_hugetlb_folio_nodemask()
|
A D | memcontrol.c | 1702 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
/linux-6.3-rc2/Documentation/translations/zh_CN/mm/ |
A D | numa.rst | 57 中的一个或多个]构建了一个有序的“区列表”。zonelist指定了当一个选定的区/节点不能满足分配请求 63 代表了相对稀缺的资源。Linux选择了一个默认的Node ordered zonelist。这意味着在使用按NUMA距
|
/linux-6.3-rc2/Documentation/mm/ |
A D | numa.rst | 74 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 84 a default Node ordered zonelist. This means it tries to fallback to other zones 89 Linux will attempt to allocate from the first node in the appropriate zonelist 92 nodes' zones in the selected zonelist looking for the first zone in the list 120 zonelist--will not be the node itself. Rather, it will be the node that the
|
/linux-6.3-rc2/drivers/tty/ |
A D | sysrq.c | 389 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
/linux-6.3-rc2/Documentation/admin-guide/sysctl/ |
A D | vm.rst | 648 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 654 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
/linux-6.3-rc2/Documentation/admin-guide/mm/ |
A D | numa_memory_policy.rst | 234 node zonelist.
|
/linux-6.3-rc2/Documentation/admin-guide/ |
A D | kernel-parameters.txt | 3909 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|