Searched refs:vmemmap_start (Results 1 – 3 of 3) sorted by relevance
465 vmemmap_end = vmemmap_start + hugetlb_vmemmap_size(h); in __hugetlb_vmemmap_restore_folio()466 vmemmap_reuse = vmemmap_start; in __hugetlb_vmemmap_restore_folio()467 vmemmap_start += HUGETLB_VMEMMAP_RESERVE_SIZE; in __hugetlb_vmemmap_restore_folio()591 vmemmap_end = vmemmap_start + hugetlb_vmemmap_size(h); in __hugetlb_vmemmap_optimize_folio()592 vmemmap_reuse = vmemmap_start; in __hugetlb_vmemmap_optimize_folio()593 vmemmap_start += HUGETLB_VMEMMAP_RESERVE_SIZE; in __hugetlb_vmemmap_optimize_folio()601 ret = vmemmap_remap_free(vmemmap_start, vmemmap_end, vmemmap_reuse, in __hugetlb_vmemmap_optimize_folio()637 vmemmap_end = vmemmap_start + hugetlb_vmemmap_size(h); in hugetlb_vmemmap_split_folio()638 vmemmap_reuse = vmemmap_start; in hugetlb_vmemmap_split_folio()639 vmemmap_start += HUGETLB_VMEMMAP_RESERVE_SIZE; in hugetlb_vmemmap_split_folio()[all …]
239 unsigned long vmemmap_start; in setup_kernel_memory_layout() local327 vmemmap_start = round_down(__abs_lowcore - pages * sizeof(struct page), rte_size); in setup_kernel_memory_layout()329 ident_map_size = min(ident_map_size, vmemmap_start); in setup_kernel_memory_layout()332 if (vmemmap_start + vmemmap_size > __abs_lowcore) { in setup_kernel_memory_layout()336 vmemmap = (struct page *)vmemmap_start; in setup_kernel_memory_layout()340 max_mappable = min(max_mappable, vmemmap_start); in setup_kernel_memory_layout()342 __identity_base = round_down(vmemmap_start - max_mappable, rte_size); in setup_kernel_memory_layout()
331 u64 vmemmap_start = (u64)virt_to_page((void *)page_offset); in ptdump_init() local343 { vmemmap_start, "vmemmap start" }, in ptdump_init()
Completed in 9 milliseconds