| /mm/ |
| A D | vmscan.c | 247 if (sc->proactive && sc->proactive_swappiness) in sc_swappiness() 351 if (sc && sc->no_demotion) in can_demote() 4872 if (sc->nr_reclaimed >= max(sc->nr_to_reclaim, compact_gap(sc->order))) in should_abort_scan() 4925 if (sc->nr.unqueued_dirty && sc->nr.unqueued_dirty == sc->nr.file_taken) in try_to_shrink_lruvec() 6077 memset(&sc->nr, 0, sizeof(sc->nr)); in shrink_node() 6085 memset(&sc->nr, 0, sizeof(sc->nr)); in shrink_node() 6124 if (sc->nr.writeback && sc->nr.writeback == sc->nr.taken) in shrink_node() 6150 if (sc->nr.dirty && sc->nr.dirty == sc->nr.congested) { in shrink_node() 6279 sc->reclaim_idx, sc->nodemask) { in shrink_zones() 6322 sc->order, sc->gfp_mask, in shrink_zones() [all …]
|
| A D | shrinker_debug.c | 27 struct shrink_control sc = { in shrinker_count_objects() local 33 nr = shrinker->count_objects(shrinker, &sc); in shrinker_count_objects() 110 struct shrink_control sc = { in shrinker_debugfs_scan_write() local 144 sc.nid = nid; in shrinker_debugfs_scan_write() 145 sc.memcg = memcg; in shrinker_debugfs_scan_write() 146 sc.nr_to_scan = nr_to_scan; in shrinker_debugfs_scan_write() 147 sc.nr_scanned = nr_to_scan; in shrinker_debugfs_scan_write() 149 shrinker->scan_objects(shrinker, &sc); in shrinker_debugfs_scan_write()
|
| A D | shrinker.c | 337 struct shrink_control *sc) in xchg_nr_deferred() argument 339 int nid = sc->nid; in xchg_nr_deferred() 344 if (sc->memcg && in xchg_nr_deferred() 347 sc->memcg); in xchg_nr_deferred() 354 struct shrink_control *sc) in add_nr_deferred() argument 356 int nid = sc->nid; in add_nr_deferred() 361 if (sc->memcg && in add_nr_deferred() 364 sc->memcg); in add_nr_deferred() 528 struct shrink_control sc = { in shrink_slab_memcg() local 653 struct shrink_control sc = { in shrink_slab() local [all …]
|
| A D | workingset.c | 641 struct shrink_control *sc) in count_shadow_nodes() argument 647 nodes = list_lru_shrink_count(&shadow_nodes, sc); in count_shadow_nodes() 674 if (sc->memcg) { in count_shadow_nodes() 678 mem_cgroup_flush_stats_ratelimited(sc->memcg); in count_shadow_nodes() 679 lruvec = mem_cgroup_lruvec(sc->memcg, NODE_DATA(sc->nid)); in count_shadow_nodes() 689 pages = node_present_pages(sc->nid); in count_shadow_nodes() 768 struct shrink_control *sc) in scan_shadow_nodes() argument 771 return list_lru_shrink_walk_irq(&shadow_nodes, sc, shadow_lru_isolate, in scan_shadow_nodes()
|
| A D | swap_cgroup.c | 41 struct swap_cgroup *sc = &map[offset / ID_PER_SC]; in __swap_cgroup_id_xchg() local 43 unsigned int new_ids, old_ids = atomic_read(&sc->ids); in __swap_cgroup_id_xchg() 49 } while (!atomic_try_cmpxchg(&sc->ids, &old_ids, new_ids)); in __swap_cgroup_id_xchg()
|
| A D | zswap.c | 1252 struct shrink_control *sc) in zswap_shrinker_scan() argument 1258 !mem_cgroup_zswap_writeback_enabled(sc->memcg)) { in zswap_shrinker_scan() 1259 sc->nr_scanned = 0; in zswap_shrinker_scan() 1263 shrink_ret = list_lru_shrink_walk(&zswap_list_lru, sc, &shrink_memcg_cb, in zswap_shrinker_scan() 1273 struct shrink_control *sc) in zswap_shrinker_count() argument 1275 struct mem_cgroup *memcg = sc->memcg; in zswap_shrinker_count() 1276 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(sc->nid)); in zswap_shrinker_count() 1290 if (!gfp_has_io_fs(sc->gfp_mask)) in zswap_shrinker_count() 1313 nr_freeable = list_lru_shrink_count(&zswap_list_lru, sc); in zswap_shrinker_count()
|
| A D | huge_memory.c | 72 struct shrink_control *sc); 74 struct shrink_control *sc); 270 struct shrink_control *sc) in shrink_huge_zero_page_count() argument 277 struct shrink_control *sc) in shrink_huge_zero_page_scan() argument 4103 struct shrink_control *sc) in deferred_split_count() argument 4109 if (sc->memcg) in deferred_split_count() 4110 ds_queue = &sc->memcg->deferred_split_queue; in deferred_split_count() 4149 struct shrink_control *sc) in deferred_split_scan() argument 4159 if (sc->memcg) in deferred_split_scan() 4160 ds_queue = &sc->memcg->deferred_split_queue; in deferred_split_scan() [all …]
|
| A D | memory-failure.c | 1260 #define sc ((1UL << PG_swapcache) | (1UL << PG_swapbacked)) macro 1276 { sc|dirty, sc|dirty, MF_MSG_DIRTY_SWAPCACHE, me_swapcache_dirty }, 1277 { sc|dirty, sc, MF_MSG_CLEAN_SWAPCACHE, me_swapcache_clean }, 1295 #undef sc
|
| A D | slab_common.c | 2090 kfree_rcu_shrink_count(struct shrinker *shrink, struct shrink_control *sc) in kfree_rcu_shrink_count() argument 2108 kfree_rcu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) in kfree_rcu_shrink_scan() argument 2120 sc->nr_to_scan -= count; in kfree_rcu_shrink_scan() 2123 if (sc->nr_to_scan <= 0) in kfree_rcu_shrink_scan()
|
| A D | shmem.c | 748 struct shrink_control *sc, unsigned long nr_to_free) in shmem_unused_huge_shrink() argument 754 unsigned long batch = sc ? sc->nr_to_scan : 128; in shmem_unused_huge_shrink() 855 struct shrink_control *sc) in shmem_unused_huge_scan() argument 862 return shmem_unused_huge_shrink(sbinfo, sc, 0); in shmem_unused_huge_scan() 866 struct shrink_control *sc) in shmem_unused_huge_count() argument 876 struct shrink_control *sc, unsigned long nr_to_free) in shmem_unused_huge_shrink() argument
|
| A D | zsmalloc.c | 2013 struct shrink_control *sc) in zs_shrinker_scan() argument 2029 struct shrink_control *sc) in zs_shrinker_count() argument
|
| A D | vmalloc.c | 5211 vmap_node_shrink_count(struct shrinker *shrink, struct shrink_control *sc) in vmap_node_shrink_count() argument 5226 vmap_node_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) in vmap_node_shrink_scan() argument
|