| /arch/x86/mm/ |
| A D | ioremap.c | 217 &phys_addr, &last_addr); in __ioremap_caller() 225 phys_addr &= PAGE_MASK; in __ioremap_caller() 234 retval = memtype_reserve(phys_addr, (u64)phys_addr + size, in __ioremap_caller() 295 area->phys_addr = phys_addr; in __ioremap_caller() 318 memtype_free(phys_addr, phys_addr + size); in __ioremap_caller() 496 memtype_free(p->phys_addr, p->phys_addr + get_vm_area_size(p)); in iounmap() 580 switch (e820__get_entry_type(phys_addr, phys_addr + size - 1)) { in memremap_should_map_decrypted() 614 if (phys_addr == paddr) in memremap_is_efi_data() 620 if (phys_addr == paddr) in memremap_is_efi_data() 652 if (phys_addr == paddr) in __memremap_is_setup_data() [all …]
|
| A D | physaddr.c | 74 unsigned long phys_addr = x - PAGE_OFFSET; in __phys_addr() local 80 VIRTUAL_BUG_ON((phys_addr >> PAGE_SHIFT) > max_low_pfn); in __phys_addr() 81 BUG_ON(slow_virt_to_phys((void *)x) != phys_addr); in __phys_addr() 83 return phys_addr; in __phys_addr()
|
| /arch/nios2/mm/ |
| A D | ioremap.c | 37 pfn = PFN_DOWN(phys_addr); in remap_area_pte() 60 phys_addr -= address; in remap_area_pmd() 83 phys_addr -= address; in remap_area_pages() 104 phys_addr + address, flags)) in remap_area_pages() 127 last_addr = phys_addr + size - 1; in ioremap() 129 if (!size || last_addr < phys_addr) in ioremap() 133 if (phys_addr > PHYS_OFFSET && phys_addr < virt_to_phys(high_memory)) { in ioremap() 137 t_addr = __va(phys_addr); in ioremap() 149 if (IS_MAPPABLE_UNCACHEABLE(phys_addr) && in ioremap() 154 offset = phys_addr & ~PAGE_MASK; in ioremap() [all …]
|
| /arch/mips/mm/ |
| A D | ioremap.c | 56 cpu_addr = plat_ioremap(phys_addr, size, flags); in ioremap_prot() 60 phys_addr = fixup_bigphys_addr(phys_addr, size); in ioremap_prot() 63 last_addr = phys_addr + size - 1; in ioremap_prot() 64 if (!size || last_addr < phys_addr) in ioremap_prot() 71 if (IS_LOW512(phys_addr) && IS_LOW512(last_addr) && in ioremap_prot() 73 return (void __iomem *) CKSEG1ADDR(phys_addr); in ioremap_prot() 83 pfn = PFN_DOWN(phys_addr); in ioremap_prot() 88 &phys_addr, &last_addr); in ioremap_prot() 95 offset = phys_addr & ~PAGE_MASK; in ioremap_prot() 96 phys_addr &= PAGE_MASK; in ioremap_prot() [all …]
|
| /arch/parisc/mm/ |
| A D | ioremap.c | 16 void __iomem *ioremap_prot(phys_addr_t phys_addr, size_t size, in ioremap_prot() argument 20 unsigned long end = phys_addr + size - 1; in ioremap_prot() 22 if ((phys_addr >= 0x00080000 && end < 0x000fffff) || in ioremap_prot() 23 (phys_addr >= 0x00500000 && end < 0x03bfffff)) in ioremap_prot() 24 phys_addr |= F_EXTEND(0xfc000000); in ioremap_prot() 30 if (phys_addr < virt_to_phys(high_memory)) { in ioremap_prot() 34 t_addr = __va(phys_addr); in ioremap_prot() 44 return generic_ioremap_prot(phys_addr, size, prot); in ioremap_prot()
|
| /arch/mips/alchemy/common/ |
| A D | setup.c | 80 phys_addr_t fixup_bigphys_addr(phys_addr_t phys_addr, phys_addr_t size) in fixup_bigphys_addr() argument 86 if ((phys_addr >> 32) != 0) in fixup_bigphys_addr() 87 return phys_addr; in fixup_bigphys_addr() 90 if (phys_addr >= start && (phys_addr + size - 1) <= end) in fixup_bigphys_addr() 91 return (phys_addr_t)(AU1500_PCI_MEM_PHYS_ADDR + phys_addr); in fixup_bigphys_addr() 94 return phys_addr; in fixup_bigphys_addr() 100 phys_addr_t phys_addr = fixup_bigphys_addr(pfn << PAGE_SHIFT, size); in io_remap_pfn_range() local 102 return remap_pfn_range(vma, vaddr, phys_addr >> PAGE_SHIFT, size, prot); in io_remap_pfn_range()
|
| /arch/loongarch/mm/ |
| A D | ioremap.c | 9 void __init __iomem *early_ioremap(u64 phys_addr, unsigned long size) in early_ioremap() argument 11 return ((void __iomem *)TO_CACHE(phys_addr)); in early_ioremap() 19 void * __init early_memremap_ro(resource_size_t phys_addr, unsigned long size) in early_memremap_ro() argument 21 return early_memremap(phys_addr, size); in early_memremap_ro() 24 void * __init early_memremap_prot(resource_size_t phys_addr, unsigned long size, in early_memremap_prot() argument 27 return early_memremap(phys_addr, size); in early_memremap_prot()
|
| /arch/riscv/mm/ |
| A D | kasan_init.c | 31 phys_addr_t phys_addr; in kasan_populate_pte() local 52 phys_addr_t phys_addr; in kasan_populate_pmd() local 69 if (phys_addr) { in kasan_populate_pmd() 83 phys_addr_t phys_addr; in kasan_populate_pud() local 100 if (phys_addr) { in kasan_populate_pud() 114 phys_addr_t phys_addr; in kasan_populate_p4d() local 131 if (phys_addr) { in kasan_populate_p4d() 145 phys_addr_t phys_addr; in kasan_populate_pgd() local 154 if (phys_addr) { in kasan_populate_pgd() 239 phys_addr_t phys_addr; in kasan_early_populate_pud() local [all …]
|
| /arch/arm64/mm/ |
| A D | ioremap.c | 17 void __iomem *ioremap_prot(phys_addr_t phys_addr, size_t size, in ioremap_prot() argument 20 unsigned long last_addr = phys_addr + size - 1; in ioremap_prot() 27 if (WARN_ON(pfn_is_map_memory(__phys_to_pfn(phys_addr)))) in ioremap_prot() 35 WARN_ON(ioremap_prot_hook(phys_addr, size, &pgprot))) { in ioremap_prot() 39 return generic_ioremap_prot(phys_addr, size, pgprot); in ioremap_prot()
|
| /arch/sh/mm/ |
| A D | ioremap_fixed.c | 48 ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) in ioremap_fixed() argument 59 offset = phys_addr & ~PAGE_MASK; in ioremap_fixed() 60 phys_addr &= PAGE_MASK; in ioremap_fixed() 61 size = PAGE_ALIGN(phys_addr + size) - phys_addr; in ioremap_fixed() 90 __set_fixmap(idx, phys_addr, prot); in ioremap_fixed() 91 phys_addr += PAGE_SIZE; in ioremap_fixed()
|
| A D | ioremap.c | 75 void __iomem __ref *ioremap_prot(phys_addr_t phys_addr, size_t size, in ioremap_prot() argument 80 mapped = __ioremap_trapped(phys_addr, size); in ioremap_prot() 84 mapped = __ioremap_29bit(phys_addr, size, pgprot); in ioremap_prot() 92 return ioremap_fixed(phys_addr, size, pgprot); in ioremap_prot() 98 mapped = pmb_remap_caller(phys_addr, size, pgprot, in ioremap_prot() 103 return generic_ioremap_prot(phys_addr, size, pgprot); in ioremap_prot()
|
| /arch/x86/platform/efi/ |
| A D | memmap.c | 128 start = md->phys_addr; in efi_memmap_split_count() 193 start = md->phys_addr; in efi_memmap_insert() 203 md->num_pages = (m_end - md->phys_addr + 1) >> in efi_memmap_insert() 209 md->phys_addr = m_end + 1; in efi_memmap_insert() 210 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert() 216 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 223 md->phys_addr = m_start; in efi_memmap_insert() 230 md->phys_addr = m_end + 1; in efi_memmap_insert() 238 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 244 md->phys_addr = m_start; in efi_memmap_insert() [all …]
|
| A D | efi_32.c | 42 start_pfn = PFN_DOWN(md->phys_addr); in efi_map_region() 44 end = md->phys_addr + size; in efi_map_region() 48 va = __va(md->phys_addr); in efi_map_region() 53 va = ioremap_cache(md->phys_addr, size); in efi_map_region() 58 pr_err("ioremap of 0x%llX failed!\n", md->phys_addr); in efi_map_region() 88 void __init parse_efi_setup(u64 phys_addr, u32 data_len) {} in parse_efi_setup() argument
|
| A D | efi.c | 123 unsigned long long start = md->phys_addr; in do_add_efi_memmap() 247 (md->phys_addr >> EFI_PAGE_SHIFT)) { in efi_memmap_entry_valid() 262 md->phys_addr, end_hi, end); in efi_memmap_entry_valid() 266 md->phys_addr, end); in efi_memmap_entry_valid() 331 start = md->phys_addr; in efi_remove_e820_mmio() 357 md->phys_addr, in efi_print_memmap() 358 md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1, in efi_print_memmap() 529 if (md->phys_addr == (prev_md->phys_addr + prev_size)) { in efi_merge_regions() 877 bool efi_is_table_address(unsigned long phys_addr) in efi_is_table_address() argument 881 if (phys_addr == EFI_INVALID_TABLE_ADDR) in efi_is_table_address() [all …]
|
| /arch/powerpc/boot/ |
| A D | cuboot-pq2.c | 33 u32 phys_addr; member 178 if (mem->phys_addr + mem->size[1] == mmio->phys_addr) in fixup_pci() 180 else if (mmio->phys_addr + mmio->size[1] == mem->phys_addr) in fixup_pci() 185 out_be32(&pci_regs[1][0], mem_base->phys_addr | 1); in fixup_pci() 188 out_be32(&pci_regs[1][1], io->phys_addr | 1); in fixup_pci() 192 out_le32(&pci_regs[0][2], mem->phys_addr >> 12); in fixup_pci() 196 out_le32(&pci_regs[0][8], mmio->phys_addr >> 12); in fixup_pci() 200 out_le32(&pci_regs[0][14], io->phys_addr >> 12); in fixup_pci()
|
| /arch/arm/mm/ |
| A D | ioremap.c | 63 if (vm->phys_addr > paddr || in find_static_vm_paddr() 64 paddr + size - 1 > vm->phys_addr + vm->size - 1) in find_static_vm_paddr() 309 addr += paddr - svm->vm.phys_addr; in __arm_ioremap_pfn_caller() 326 area->phys_addr = paddr; in __arm_ioremap_pfn_caller() 356 unsigned long offset = phys_addr & ~PAGE_MASK; in __arm_ioremap_caller() 357 unsigned long pfn = __phys_to_pfn(phys_addr); in __arm_ioremap_caller() 362 last_addr = phys_addr + size - 1; in __arm_ioremap_caller() 363 if (!size || last_addr < phys_addr) in __arm_ioremap_caller() 421 __arm_ioremap_exec(phys_addr_t phys_addr, size_t size, bool cached) in __arm_ioremap_exec() argument 430 return __arm_ioremap_caller(phys_addr, size, mtype, in __arm_ioremap_exec() [all …]
|
| /arch/sparc/kernel/ |
| A D | auxio_32.c | 67 r.start = auxregs[0].phys_addr; in auxio_probe() 68 r.end = auxregs[0].phys_addr + auxregs[0].reg_size - 1; in auxio_probe() 71 if ((((unsigned long) auxregs[0].phys_addr) & 3) == 3) in auxio_probe() 131 r.start = regs.phys_addr; in auxio_power_probe() 132 r.end = regs.phys_addr + regs.reg_size - 1; in auxio_power_probe()
|
| A D | prom_64.c | 81 (unsigned int) (regs->phys_addr >> 32UL), in sun4v_path_component() 82 (unsigned int) (regs->phys_addr & 0xffffffffUL)); in sun4v_path_component() 86 type = regs->phys_addr >> 60UL; in sun4v_path_component() 88 low_bits = (regs->phys_addr & 0xffffffffUL); in sun4v_path_component() 122 (unsigned int) (regs->phys_addr >> 32UL), in sun4u_path_component() 123 (unsigned int) (regs->phys_addr & 0xffffffffUL)); in sun4u_path_component() 139 (unsigned int) (regs->phys_addr & mask)); in sun4u_path_component() 158 regs->phys_addr); in sbus_path_component() 207 (unsigned int) (regs->phys_addr & 0xffffffffUL)); in upa_path_component() 241 (unsigned int) (regs->phys_addr >> 32UL), in ebus_path_component() [all …]
|
| A D | chmc.c | 284 if (phys_addr < dp->base_addr || in jbusmc_find_dimm_group() 285 (dp->base_addr + dp->size) <= phys_addr) in jbusmc_find_dimm_group() 295 unsigned long phys_addr, in jbusmc_print_dimm() argument 303 dp = jbusmc_find_dimm_group(phys_addr); in jbusmc_print_dimm() 322 get_pin_and_dimm_str(syndrome_code, phys_addr, &pin, in jbusmc_print_dimm() 355 this_base = ent->phys_addr; in jbusmc_dimm_group_size() 520 static struct chmc_bank_info *chmc_find_bank(unsigned long phys_addr) in chmc_find_bank() argument 531 if (chmc_bank_match(bp, phys_addr)) in chmc_find_bank() 541 unsigned long phys_addr, in chmc_print_dimm() argument 548 bp = chmc_find_bank(phys_addr); in chmc_print_dimm() [all …]
|
| /arch/xtensa/mm/ |
| A D | ioremap.c | 13 void __iomem *ioremap_prot(phys_addr_t phys_addr, size_t size, in ioremap_prot() argument 16 unsigned long pfn = __phys_to_pfn((phys_addr)); in ioremap_prot() 19 return generic_ioremap_prot(phys_addr, size, prot); in ioremap_prot()
|
| /arch/x86/kvm/vmx/ |
| A D | vmx_ops.h | 15 void vmclear_error(struct vmcs *vmcs, u64 phys_addr); 16 void vmptrld_error(struct vmcs *vmcs, u64 phys_addr); 291 u64 phys_addr = __pa(vmcs); in vmcs_clear() local 293 vmx_asm1(vmclear, "m"(phys_addr), vmcs, phys_addr); in vmcs_clear() 298 u64 phys_addr = __pa(vmcs); in vmcs_load() local 301 return evmcs_load(phys_addr); in vmcs_load() 303 vmx_asm1(vmptrld, "m"(phys_addr), vmcs, phys_addr); in vmcs_load()
|
| /arch/arm/mach-imx/ |
| A D | mm-imx3.c | 55 static void __iomem *imx3_ioremap_caller(phys_addr_t phys_addr, size_t size, in imx3_ioremap_caller() argument 64 if (phys_addr < 0x80000000 && in imx3_ioremap_caller() 65 !addr_in_module(phys_addr, MX3x_L2CC)) in imx3_ioremap_caller() 69 return __arm_ioremap_caller(phys_addr, size, mtype, caller); in imx3_ioremap_caller()
|
| /arch/x86/include/asm/ |
| A D | fixmap.h | 182 void __init *early_memremap_encrypted(resource_size_t phys_addr, 184 void __init *early_memremap_encrypted_wp(resource_size_t phys_addr, 186 void __init *early_memremap_decrypted(resource_size_t phys_addr, 188 void __init *early_memremap_decrypted_wp(resource_size_t phys_addr,
|
| /arch/powerpc/platforms/ps3/ |
| A D | mm.c | 127 BUG_ON(is_kernel_addr(phys_addr)); in ps3_mm_phys_to_lpar() 128 return (phys_addr < map.rm.size || phys_addr >= map.total) in ps3_mm_phys_to_lpar() 129 ? phys_addr : phys_addr + map.r1.offset; in ps3_mm_phys_to_lpar() 526 c->lpar_addr = ps3_mm_phys_to_lpar(phys_addr); in dma_sb_map_pages() 563 phys_addr, ps3_mm_phys_to_lpar(phys_addr), len); in dma_ioc0_map_pages() 572 c->lpar_addr = ps3_mm_phys_to_lpar(phys_addr); in dma_ioc0_map_pages() 793 phys_addr); in dma_sb_map_area() 844 phys_addr, aligned_phys, aligned_len); in dma_ioc0_map_area() 852 *bus_addr = c->bus_addr + phys_addr - aligned_phys; in dma_ioc0_map_area() 868 *bus_addr = c->bus_addr + phys_addr - aligned_phys; in dma_ioc0_map_area() [all …]
|
| /arch/loongarch/include/asm/ |
| A D | dmi.h | 15 static inline void *dmi_remap(u64 phys_addr, unsigned long size) in dmi_remap() argument 17 return ((void *)TO_CACHE(phys_addr)); in dmi_remap()
|