Home
last modified time | relevance | path

Searched refs:first (Results 1 – 13 of 13) sorted by relevance

/mm/
A Dinterval_tree.c93 unsigned long first, unsigned long last) in anon_vma_interval_tree_iter_first() argument
95 return __anon_vma_interval_tree_iter_first(root, first, last); in anon_vma_interval_tree_iter_first()
100 unsigned long first, unsigned long last) in anon_vma_interval_tree_iter_next() argument
102 return __anon_vma_interval_tree_iter_next(node, first, last); in anon_vma_interval_tree_iter_next()
A Ddmapool.c306 struct dma_block *block, *first = NULL, *last = NULL; in pool_initialise_page() local
323 first = block; in pool_initialise_page()
331 pool->next_block = first; in pool_initialise_page()
A Drmap.c1250 int first = 0, nr = 0; in __folio_add_rmap() local
1272 first += atomic_inc_and_test(&page->_mapcount); in __folio_add_rmap()
1275 if (first && in __folio_add_rmap()
1276 atomic_add_return_relaxed(first, mapped) < ENTIRELY_MAPPED) in __folio_add_rmap()
1277 nr = first; in __folio_add_rmap()
1283 first = atomic_inc_and_test(&folio->_entire_mapcount); in __folio_add_rmap()
1285 if (level == RMAP_LEVEL_PMD && first) in __folio_add_rmap()
1296 if (first) { in __folio_add_rmap()
A Dvmscan.c3073 bool first = false; in iterate_mm_list() local
3102 first = true; in iterate_mm_list()
3124 if (mm && first) in iterate_mm_list()
3617 if (*first == -1) { in walk_pmd_range_locked()
3618 *first = addr; in walk_pmd_range_locked()
3623 i = addr == -1 ? 0 : pmd_index(addr) - pmd_index(*first); in walk_pmd_range_locked()
3629 pmd = pmd_offset(pud, *first); in walk_pmd_range_locked()
3642 addr = i ? (*first & PMD_MASK) + i * PMD_SIZE : *first; in walk_pmd_range_locked()
3685 *first = -1; in walk_pmd_range_locked()
3697 unsigned long first = -1; in walk_pmd_range() local
[all …]
A Dcma_debug.c82 mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); in cma_get_entry_from_list()
A Dmempolicy.c623 struct vm_area_struct *first; member
845 if (!qp->first) { in queue_pages_test_walk()
846 qp->first = vma; in queue_pages_test_walk()
916 .first = NULL, in queue_pages_range()
923 if (!qp.first) in queue_pages_range()
A Dmmu_notifier.c325 subscription = hlist_entry(subscriptions->list.first, in mn_hlist_release()
A Dmemory-failure.c1909 llist_for_each_entry(p, raw_hwp_head->first, node) { in is_raw_hwpoison_page_in_hugepage()
1954 llist_for_each_entry(p, head->first, node) { in folio_set_hugetlb_hwpoison()
A Dfilemap.c4377 pgoff_t first = start >> PAGE_SHIFT; in filemap_invalidate_inode() local
4379 pgoff_t nr = end == LLONG_MAX ? ULONG_MAX : last - first + 1; in filemap_invalidate_inode()
4390 unmap_mapping_pages(mapping, first, nr, false); in filemap_invalidate_inode()
A Dksm.c1697 BUG_ON(stable_node->hlist.first->next); in stable_node_dup()
1721 } else if (stable_node->hlist.first != &found->hlist_dup && in stable_node_dup()
A DKconfig684 two situations. The first is on NUMA systems to put pages nearer
/mm/damon/
A Dcore.c178 static int damon_fill_regions_holes(struct damon_region *first, in damon_fill_regions_holes() argument
181 struct damon_region *r = first; in damon_fill_regions_holes()
230 struct damon_region *first = NULL, *last, *newr; in damon_set_regions() local
237 if (!first) in damon_set_regions()
238 first = r; in damon_set_regions()
244 if (!first) { in damon_set_regions()
255 first->ar.start = ALIGN_DOWN(range->start, in damon_set_regions()
260 err = damon_fill_regions_holes(first, last, t); in damon_set_regions()
A DKconfig95 pages reclaimed first under memory pressure.

Completed in 68 milliseconds