Lines Matching refs:gfp_mask

138 	gfp_t gfp_mask;  member
805 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, in shrink_slab_memcg() argument
824 .gfp_mask = gfp_mask, in shrink_slab_memcg()
879 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, in shrink_slab_memcg() argument
906 static unsigned long shrink_slab(gfp_t gfp_mask, int nid, in shrink_slab() argument
921 return shrink_slab_memcg(gfp_mask, nid, memcg, priority); in shrink_slab()
928 .gfp_mask = gfp_mask, in shrink_slab()
1469 .gfp_mask = (GFP_HIGHUSER_MOVABLE & ~__GFP_RECLAIM) | in alloc_demote_page()
1557 may_enter_fs = (sc->gfp_mask & __GFP_FS) || in shrink_page_list()
1558 (PageSwapCache(page) && (sc->gfp_mask & __GFP_IO)); in shrink_page_list()
1694 if (!(sc->gfp_mask & __GFP_IO)) in shrink_page_list()
1858 if (!try_to_release_page(page, sc->gfp_mask)) in shrink_page_list()
1969 .gfp_mask = GFP_KERNEL, in reclaim_clean_pages_from_list()
2290 if ((sc->gfp_mask & (__GFP_IO | __GFP_FS)) == (__GFP_IO | __GFP_FS)) in too_many_isolated()
2601 .gfp_mask = GFP_KERNEL, in reclaim_pages()
3172 shrink_slab(sc->gfp_mask, pgdat->node_id, memcg, in shrink_node_memcgs()
3176 vmpressure(sc->gfp_mask, memcg, false, in shrink_node_memcgs()
3301 vmpressure(sc->gfp_mask, sc->target_mem_cgroup, true, in shrink_node()
3466 orig_mask = sc->gfp_mask; in shrink_zones()
3468 sc->gfp_mask |= __GFP_HIGHMEM; in shrink_zones()
3469 sc->reclaim_idx = gfp_zone(sc->gfp_mask); in shrink_zones()
3516 sc->order, sc->gfp_mask, in shrink_zones()
3540 sc->gfp_mask = orig_mask; in shrink_zones()
3585 vmpressure_prio(sc->gfp_mask, sc->target_mem_cgroup, in do_try_to_free_pages()
3708 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument
3747 gfp_zone(gfp_mask), nodemask) { in throttle_direct_reclaim()
3773 if (!(gfp_mask & __GFP_FS)) in throttle_direct_reclaim()
3789 gfp_t gfp_mask, nodemask_t *nodemask) in try_to_free_pages() argument
3794 .gfp_mask = current_gfp_context(gfp_mask), in try_to_free_pages()
3795 .reclaim_idx = gfp_zone(gfp_mask), in try_to_free_pages()
3817 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages()
3821 trace_mm_vmscan_direct_reclaim_begin(order, sc.gfp_mask); in try_to_free_pages()
3835 gfp_t gfp_mask, bool noswap, in mem_cgroup_shrink_node() argument
3851 sc.gfp_mask = (gfp_mask & GFP_RECLAIM_MASK) | in mem_cgroup_shrink_node()
3855 sc.gfp_mask); in mem_cgroup_shrink_node()
3875 gfp_t gfp_mask, in try_to_free_mem_cgroup_pages() argument
3882 .gfp_mask = (current_gfp_context(gfp_mask) & GFP_RECLAIM_MASK) | in try_to_free_mem_cgroup_pages()
3896 struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask); in try_to_free_mem_cgroup_pages()
3899 trace_mm_vmscan_memcg_reclaim_begin(0, sc.gfp_mask); in try_to_free_mem_cgroup_pages()
4140 .gfp_mask = GFP_KERNEL, in balance_pgdat()
4252 sc.gfp_mask, &nr_soft_scanned); in balance_pgdat()
4596 .gfp_mask = GFP_HIGHUSER_MOVABLE, in shrink_all_memory()
4604 struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask); in shrink_all_memory()
4608 fs_reclaim_acquire(sc.gfp_mask); in shrink_all_memory()
4616 fs_reclaim_release(sc.gfp_mask); in shrink_all_memory()
4741 static int __node_reclaim(struct pglist_data *pgdat, gfp_t gfp_mask, unsigned int order) in __node_reclaim() argument
4749 .gfp_mask = current_gfp_context(gfp_mask), in __node_reclaim()
4755 .reclaim_idx = gfp_zone(gfp_mask), in __node_reclaim()
4760 sc.gfp_mask); in __node_reclaim()
4764 fs_reclaim_acquire(sc.gfp_mask); in __node_reclaim()
4787 fs_reclaim_release(sc.gfp_mask); in __node_reclaim()
4795 int node_reclaim(struct pglist_data *pgdat, gfp_t gfp_mask, unsigned int order) in node_reclaim() argument
4817 if (!gfpflags_allow_blocking(gfp_mask) || (current->flags & PF_MEMALLOC)) in node_reclaim()
4832 ret = __node_reclaim(pgdat, gfp_mask, order); in node_reclaim()