Home
last modified time | relevance | path

Searched refs:HPAGE_PMD_NR (Results 1 – 17 of 17) sorted by relevance

/linux-6.3-rc2/mm/
A Dkhugepaged.c271 if (err || max_ptes_none > HPAGE_PMD_NR - 1) in max_ptes_none_store()
296 if (err || max_ptes_swap > HPAGE_PMD_NR - 1) in max_ptes_swap_store()
322 if (err || max_ptes_shared > HPAGE_PMD_NR - 1) in max_ptes_shared_store()
398 khugepaged_pages_to_scan = HPAGE_PMD_NR * 8; in khugepaged_init()
399 khugepaged_max_ptes_none = HPAGE_PMD_NR - 1; in khugepaged_init()
400 khugepaged_max_ptes_swap = HPAGE_PMD_NR / 8; in khugepaged_init()
555 for (_pte = pte; _pte < pte + HPAGE_PMD_NR; in __collapse_huge_page_isolate()
692 for (_pte = pte; _pte < pte + HPAGE_PMD_NR; in __collapse_huge_page_copy()
1784 VM_BUG_ON(start & (HPAGE_PMD_NR - 1)); in collapse_file()
2075 folio_ref_add(folio, HPAGE_PMD_NR - 1); in collapse_file()
[all …]
A Dpage_vma_mapped.c113 if ((pfn + HPAGE_PMD_NR - 1) < pvmw->pfn) in check_pmd()
249 (pvmw->nr_pages >= HPAGE_PMD_NR)) { in page_vma_mapped_walk()
A Dmemfd.c45 cache_count = HPAGE_PMD_NR; in memfd_tag_pins()
104 cache_count = HPAGE_PMD_NR; in memfd_wait_for_pins()
A Dhuge_memory.c226 return HPAGE_PMD_NR; in shrink_huge_zero_page_scan()
679 clear_huge_page(page, vmf->address, HPAGE_PMD_NR); in __do_huge_pmd_anonymous_page()
714 add_mm_counter(vma->vm_mm, MM_ANONPAGES, HPAGE_PMD_NR); in __do_huge_pmd_anonymous_page()
1115 add_mm_counter(dst_mm, MM_ANONPAGES, HPAGE_PMD_NR); in copy_huge_pmd()
1158 add_mm_counter(dst_mm, MM_ANONPAGES, HPAGE_PMD_NR); in copy_huge_pmd()
1573 task_numa_fault(last_cpupid, page_nid, HPAGE_PMD_NR, in do_huge_pmd_numa_page()
1721 add_mm_counter(tlb->mm, MM_ANONPAGES, -HPAGE_PMD_NR); in zap_huge_pmd()
1931 ret = HPAGE_PMD_NR; in change_huge_pmd()
2056 for (i = 0; i < HPAGE_PMD_NR; i++, haddr += PAGE_SIZE) { in __split_huge_zero_page_pmd()
2113 add_mm_counter(mm, mm_counter_file(page), -HPAGE_PMD_NR); in __split_huge_pmd_locked()
[all …]
A Dmprotect.c428 if (nr_ptes == HPAGE_PMD_NR) { in change_pmd_range()
429 pages += HPAGE_PMD_NR; in change_pmd_range()
A Dvmstat.c1674 pages /= HPAGE_PMD_NR; in zoneinfo_show_print()
1806 v[i] /= HPAGE_PMD_NR; in vmstat_start()
A Dmemcontrol.c5994 mc.precharge += HPAGE_PMD_NR; in mem_cgroup_count_precharge_pte_range()
6188 if (mc.precharge < HPAGE_PMD_NR) { in mem_cgroup_move_charge_pte_range()
6198 mc.precharge -= HPAGE_PMD_NR; in mem_cgroup_move_charge_pte_range()
6199 mc.moved_charge += HPAGE_PMD_NR; in mem_cgroup_move_charge_pte_range()
6209 mc.precharge -= HPAGE_PMD_NR; in mem_cgroup_move_charge_pte_range()
6210 mc.moved_charge += HPAGE_PMD_NR; in mem_cgroup_move_charge_pte_range()
A Dshmem.c489 index = round_up(index + 1, HPAGE_PMD_NR); in shmem_is_huge()
1554 hindex = round_down(index, HPAGE_PMD_NR); in shmem_alloc_hugefolio()
1555 if (xa_find(&mapping->i_pages, &hindex, hindex + HPAGE_PMD_NR - 1, in shmem_alloc_hugefolio()
1590 nr = huge ? HPAGE_PMD_NR : 1; in shmem_alloc_and_acct_folio()
A Dfilemap.c3144 ractl._index &= ~((unsigned long)HPAGE_PMD_NR - 1); in do_sync_mmap_readahead()
3145 ra->size = HPAGE_PMD_NR; in do_sync_mmap_readahead()
3152 ra->async_size = HPAGE_PMD_NR; in do_sync_mmap_readahead()
A Dmemory-failure.c707 if (pfn <= hwp->pfn && hwp->pfn < pfn + HPAGE_PMD_NR) { in check_hwpoisoned_pmd_entry()
A Dmigrate.c1506 #define NR_MAX_BATCHED_MIGRATION HPAGE_PMD_NR
A Dgup.c707 ctx->page_mask = HPAGE_PMD_NR - 1; in follow_pmd_mask()
A Dmemory.c4224 for (i = 0; i < HPAGE_PMD_NR; i++) in do_set_pmd()
4231 add_mm_counter(vma->vm_mm, mm_counter_file(page), HPAGE_PMD_NR); in do_set_pmd()
A Dswapfile.c271 #define SWAPFILE_CLUSTER HPAGE_PMD_NR
/linux-6.3-rc2/include/linux/
A Dhuge_mm.h106 #define HPAGE_PMD_NR (1<<HPAGE_PMD_ORDER) macro
146 HPAGE_PMD_NR)) in transhuge_vma_suitable()
/linux-6.3-rc2/drivers/gpu/drm/amd/amdgpu/
A Damdgpu_vram_mgr.c412 pages_per_block = HPAGE_PMD_NR; in amdgpu_vram_mgr_new()
/linux-6.3-rc2/drivers/base/
A Dnode.c534 pages /= HPAGE_PMD_NR; in node_read_vmstat()

Completed in 77 milliseconds