Lines Matching refs:va

58 	phys_addr_t (*alloc_pte)(uintptr_t va);
61 phys_addr_t (*alloc_pmd)(uintptr_t va);
288 static inline phys_addr_t __init alloc_pte_early(uintptr_t va) in alloc_pte_early() argument
297 static inline phys_addr_t __init alloc_pte_fixmap(uintptr_t va) in alloc_pte_fixmap() argument
302 static phys_addr_t __init alloc_pte_late(uintptr_t va) in alloc_pte_late() argument
313 uintptr_t va, phys_addr_t pa, in create_pte_mapping() argument
316 uintptr_t pte_idx = pte_index(va); in create_pte_mapping()
353 static phys_addr_t __init alloc_pmd_early(uintptr_t va) in alloc_pmd_early() argument
355 BUG_ON((va - kernel_map.virt_addr) >> PGDIR_SHIFT); in alloc_pmd_early()
360 static phys_addr_t __init alloc_pmd_fixmap(uintptr_t va) in alloc_pmd_fixmap() argument
365 static phys_addr_t __init alloc_pmd_late(uintptr_t va) in alloc_pmd_late() argument
375 uintptr_t va, phys_addr_t pa, in create_pmd_mapping() argument
380 uintptr_t pmd_idx = pmd_index(va); in create_pmd_mapping()
389 pte_phys = pt_ops.alloc_pte(va); in create_pmd_mapping()
398 create_pte_mapping(ptep, va, pa, sz, prot); in create_pmd_mapping()
418 uintptr_t va, phys_addr_t pa, in create_pgd_mapping() argument
423 uintptr_t pgd_idx = pgd_index(va); in create_pgd_mapping()
432 next_phys = alloc_pgd_next(va); in create_pgd_mapping()
441 create_pgd_next_mapping(nextp, va, pa, sz, prot); in create_pgd_mapping()
466 static __init pgprot_t pgprot_from_va(uintptr_t va) in pgprot_from_va() argument
468 if (is_va_kernel_text(va)) in pgprot_from_va()
477 if (IS_ENABLED(CONFIG_64BIT) && is_va_kernel_lm_alias_text(va)) in pgprot_from_va()
493 static __init pgprot_t pgprot_from_va(uintptr_t va) in pgprot_from_va() argument
495 if (IS_ENABLED(CONFIG_64BIT) && !is_kernel_mapping(va)) in pgprot_from_va()
524 uintptr_t va, end_va; in create_kernel_page_table() local
528 for (va = kernel_map.virt_addr; va < end_va; va += PMD_SIZE) in create_kernel_page_table()
529 create_pgd_mapping(pgdir, va, in create_kernel_page_table()
530 kernel_map.xiprom + (va - kernel_map.virt_addr), in create_kernel_page_table()
535 for (va = kernel_map.virt_addr + XIP_OFFSET; va < end_va; va += PMD_SIZE) in create_kernel_page_table()
536 create_pgd_mapping(pgdir, va, in create_kernel_page_table()
537 kernel_map.phys_addr + (va - (kernel_map.virt_addr + XIP_OFFSET)), in create_kernel_page_table()
543 uintptr_t va, end_va; in create_kernel_page_table() local
546 for (va = kernel_map.virt_addr; va < end_va; va += PMD_SIZE) in create_kernel_page_table()
547 create_pgd_mapping(pgdir, va, in create_kernel_page_table()
548 kernel_map.phys_addr + (va - kernel_map.virt_addr), in create_kernel_page_table()
551 PAGE_KERNEL_EXEC : pgprot_from_va(va)); in create_kernel_page_table()
700 uintptr_t va, map_size; in setup_vm_final() local
731 va = (uintptr_t)__va(pa); in setup_vm_final()
733 create_pgd_mapping(swapper_pg_dir, va, pa, map_size, in setup_vm_final()
734 pgprot_from_va(va)); in setup_vm_final()