| /arch/m68k/mm/ |
| A D | sun3kmap.c | 37 sun3_put_pte(virt, pte); in do_page_mapin() 41 print_pte_vaddr(virt); in do_page_mapin() 50 if(sun3_get_segmap(virt & ~SUN3_PMEG_MASK) == SUN3_INVALID_PMEG) in do_pmeg_mapin() 51 mmu_emu_map_pmeg(sun3_get_context(), virt); in do_pmeg_mapin() 54 do_page_mapin(phys, virt, type); in do_pmeg_mapin() 56 virt += PAGE_SIZE; in do_pmeg_mapin() 65 unsigned long offset, virt, ret; in sun3_ioremap() local 86 virt = (unsigned long)area->addr; in sun3_ioremap() 87 ret = virt + offset; in sun3_ioremap() 96 do_pmeg_mapin(phys, virt, type, seg_pages); in sun3_ioremap() [all …]
|
| /arch/parisc/kernel/ |
| A D | kexec.c | 81 void *virt = (void *)__fix_to_virt(FIX_TEXT_KEXEC); in machine_kexec() local 90 desc.addr = (long long)virt; in machine_kexec() 92 reloc = (void *)virt; in machine_kexec() 95 memcpy(virt, dereference_function_descriptor(relocate_new_kernel), in machine_kexec() 98 *(unsigned long *)(virt + kexec_cmdline_offset) = arch->cmdline; in machine_kexec() 99 *(unsigned long *)(virt + kexec_initrd_start_offset) = arch->initrd_start; in machine_kexec() 100 *(unsigned long *)(virt + kexec_initrd_end_offset) = arch->initrd_end; in machine_kexec() 101 *(unsigned long *)(virt + kexec_free_mem_offset) = PAGE0->mem_free; in machine_kexec()
|
| /arch/arm/mach-hisi/ |
| A D | platsmp.c | 108 void __iomem *virt; in hix5hd2_set_scu_boot_addr() local 110 virt = ioremap(start_addr, PAGE_SIZE); in hix5hd2_set_scu_boot_addr() 112 writel_relaxed(0xe51ff004, virt); /* ldr pc, [pc, #-4] */ in hix5hd2_set_scu_boot_addr() 113 writel_relaxed(jump_addr, virt + 4); /* pc jump phy address */ in hix5hd2_set_scu_boot_addr() 114 iounmap(virt); in hix5hd2_set_scu_boot_addr() 144 void __iomem *virt; in hip01_set_boot_addr() local 146 virt = phys_to_virt(start_addr); in hip01_set_boot_addr() 148 writel_relaxed(0xe51ff004, virt); in hip01_set_boot_addr() 149 writel_relaxed(jump_addr, virt + 4); in hip01_set_boot_addr()
|
| /arch/powerpc/mm/nohash/ |
| A D | e500.c | 111 TLBCAM[index].MAS2 = virt & PAGE_MASK; in settlbcam() 125 if (!is_kernel_addr(virt)) { in settlbcam() 133 tlbcam_addrs[index].start = virt; in settlbcam() 134 tlbcam_addrs[index].limit = virt + size - 1; in settlbcam() 142 unsigned int align = __ffs(virt | phys); in calc_cam_sz() 181 cam_sz = calc_cam_sz(boundary, virt, phys); in map_mem_in_cams_addr() 183 settlbcam(i, virt, phys, cam_sz, pgprot_val(prot), 0); in map_mem_in_cams_addr() 187 virt += cam_sz; in map_mem_in_cams_addr() 194 cam_sz = calc_cam_sz(ram, virt, phys); in map_mem_in_cams_addr() 200 virt += cam_sz; in map_mem_in_cams_addr() [all …]
|
| A D | 44x.c | 55 static void __init ppc44x_pin_tlb(unsigned int virt, unsigned int phys) in ppc44x_pin_tlb() argument 70 "r" (virt | PPC44x_TLB_VALID | PPC44x_TLB_256M), in ppc44x_pin_tlb() 125 static void __init ppc47x_pin_tlb(unsigned int virt, unsigned int phys) in ppc47x_pin_tlb() argument 141 virt, phys, bolted); in ppc47x_pin_tlb() 157 "r" (virt | PPC47x_TLB0_VALID | PPC47x_TLB0_256M), in ppc47x_pin_tlb()
|
| /arch/x86/pci/ |
| A D | mmconfig_64.c | 23 if (cfg && cfg->virt) in pci_dev_base() 24 return cfg->virt + (PCI_MMCFG_BUS_OFFSET(bus) | (devfn << 12)); in pci_dev_base() 116 cfg->virt = mcfg_ioremap(cfg); in pci_mmcfg_arch_map() 117 if (!cfg->virt) { in pci_mmcfg_arch_map() 127 if (cfg && cfg->virt) { in pci_mmcfg_arch_unmap() 128 iounmap(cfg->virt + PCI_MMCFG_BUS_OFFSET(cfg->start_bus)); in pci_mmcfg_arch_unmap() 129 cfg->virt = NULL; in pci_mmcfg_arch_unmap()
|
| /arch/riscv/crypto/ |
| A D | aes-riscv64-glue.c | 148 walk.dst.virt.addr, in riscv64_aes_ecb_crypt() 152 walk.dst.virt.addr, in riscv64_aes_ecb_crypt() 186 walk.dst.virt.addr, in riscv64_aes_cbc_crypt() 191 walk.dst.virt.addr, in riscv64_aes_cbc_crypt() 259 aes_cbc_cts_crypt_zvkned(ctx, walk.src.virt.addr, walk.dst.virt.addr, in riscv64_aes_cbc_cts_crypt() 317 walk.dst.virt.addr, in riscv64_aes_ctr_crypt() 324 walk.src.virt.addr + p1_nbytes, in riscv64_aes_ctr_crypt() 407 &ctx->ctx1, walk.src.virt.addr, in riscv64_aes_xts_crypt() 411 &ctx->ctx1, walk.src.virt.addr, in riscv64_aes_xts_crypt() 436 &ctx->ctx1, walk.src.virt.addr, in riscv64_aes_xts_crypt() [all …]
|
| /arch/arm/mm/ |
| A D | ioremap.c | 111 int ioremap_page(unsigned long virt, unsigned long phys, in ioremap_page() argument 114 return vmap_page_range(virt, virt + PAGE_SIZE, phys, in ioremap_page() 173 static void unmap_area_sections(unsigned long virt, unsigned long size) in unmap_area_sections() argument 175 unsigned long addr = virt, end = virt + (size & ~(SZ_1M - 1)); in unmap_area_sections() 209 flush_tlb_kernel_range(virt, end); in unmap_area_sections() 213 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 216 unsigned long addr = virt, end = virt + size; in remap_area_sections() 223 unmap_area_sections(virt, size); in remap_area_sections() 240 remap_area_supersections(unsigned long virt, unsigned long pfn, in remap_area_supersections() argument 243 unsigned long addr = virt, end = virt + size; in remap_area_supersections() [all …]
|
| /arch/xtensa/mm/ |
| A D | cache.c | 146 unsigned long virt; in flush_dcache_folio() local 160 virt = TLBTEMP_BASE_1 + (phys & DCACHE_ALIAS_MASK); in flush_dcache_folio() 161 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_folio() 163 virt = TLBTEMP_BASE_1 + (temp & DCACHE_ALIAS_MASK); in flush_dcache_folio() 166 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_folio() 169 __invalidate_icache_page_alias(virt, phys); in flush_dcache_folio() 205 unsigned long virt = TLBTEMP_BASE_1 + (address & DCACHE_ALIAS_MASK); in local_flush_cache_page() local 208 __flush_invalidate_dcache_page_alias(virt, phys); in local_flush_cache_page() 209 __invalidate_icache_page_alias(virt, phys); in local_flush_cache_page()
|
| /arch/um/kernel/ |
| A D | physmem.c | 25 void map_memory(unsigned long virt, unsigned long phys, unsigned long len, in map_memory() argument 32 err = os_map_memory((void *) virt, fd, offset, len, r, w, x); in map_memory() 39 "err = %d\n", virt, fd, offset, len, r, w, x, err); in map_memory() 166 return region->virt; in find_iomem() 189 region->virt = iomem_start; in setup_iomem() 190 region->phys = __pa(region->virt); in setup_iomem()
|
| A D | tlb.c | 22 unsigned long virt, unsigned long len, int prot, 25 unsigned long virt, unsigned long len); 29 unsigned long virt, unsigned long len, int prot, in kern_map() argument 33 return os_map_memory((void *)virt, phys_fd, offset, len, in kern_map() 39 unsigned long virt, unsigned long len) in kern_unmap() argument 41 return os_unmap_memory((void *)virt, len); in kern_unmap()
|
| /arch/arm/crypto/ |
| A D | aes-ce-glue.c | 177 ce_aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_encrypt() 197 ce_aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_decrypt() 215 ce_aes_cbc_encrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_encrypt_walk() 245 ce_aes_cbc_decrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_decrypt_walk() 315 ce_aes_cbc_cts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_encrypt() 373 ce_aes_cbc_cts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_decrypt() 392 ce_aes_ctr_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ctr_encrypt() 461 ce_aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt() 483 ce_aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt() 533 ce_aes_xts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_decrypt() [all …]
|
| /arch/powerpc/boot/ |
| A D | oflib.c | 152 unsigned int of_claim(unsigned long virt, unsigned long size, in of_claim() argument 161 return of_call_prom("claim", 3, 1, virt, size, align); in of_claim() 164 align, size, virt); in of_claim() 168 align, size, virt); in of_claim() 171 0x12, size, virt, virt); in of_claim() 172 return virt; in of_claim()
|
| /arch/powerpc/kernel/ |
| A D | exceptions-64s.S | 265 .if \virt 306 .if !\virt 373 GEN_BRANCH_TO_COMMON \name \virt 1430 GEN_INT_ENTRY data_access, virt=0 1433 GEN_INT_ENTRY data_access, virt=1 1674 GEN_INT_ENTRY alignment, virt=0 1677 GEN_INT_ENTRY alignment, virt=1 1834 GEN_INT_ENTRY decrementer, virt=0 1837 GEN_INT_ENTRY decrementer, virt=1 1970 .macro SYSTEM_CALL virt [all …]
|
| /arch/arm64/kvm/hyp/nvhe/ |
| A D | mem_protect.c | 776 void *virt = __hyp_va(phys); in __pkvm_host_donate_hyp() local 790 WARN_ON(pkvm_create_mappings_locked(virt, virt + size, PAGE_HYP)); in __pkvm_host_donate_hyp() 1257 WARN_ON(!virt); in pkvm_ownership_selftest() 1276 assert_transition_res(-EPERM, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest() 1287 assert_transition_res(-EPERM, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest() 1299 assert_transition_res(0, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest() 1300 assert_transition_res(0, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest() 1301 hyp_unpin_shared_mem(virt, virt + size); in pkvm_ownership_selftest() 1311 hyp_unpin_shared_mem(virt, virt + size); in pkvm_ownership_selftest() 1329 assert_transition_res(-EPERM, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest() [all …]
|
| /arch/arm64/crypto/ |
| A D | sm4-neon-glue.c | 45 const u8 *src = walk.src.virt.addr; in sm4_ecb_do_crypt() 46 u8 *dst = walk.dst.virt.addr; in sm4_ecb_do_crypt() 92 const u8 *src = walk.src.virt.addr; in sm4_cbc_encrypt() 93 u8 *dst = walk.dst.virt.addr; in sm4_cbc_encrypt() 123 const u8 *src = walk.src.virt.addr; in sm4_cbc_decrypt() 124 u8 *dst = walk.dst.virt.addr; in sm4_cbc_decrypt() 154 const u8 *src = walk.src.virt.addr; in sm4_ctr_crypt() 155 u8 *dst = walk.dst.virt.addr; in sm4_ctr_crypt()
|
| A D | aes-glue.c | 191 aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_encrypt() 211 aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_decrypt() 229 aes_cbc_encrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_encrypt_walk() 258 aes_cbc_decrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_decrypt_walk() 327 aes_cbc_cts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_encrypt() 384 aes_cbc_cts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_decrypt() 420 aes_essiv_cbc_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in essiv_cbc_encrypt() 583 aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt() 605 aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt() 655 aes_xts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_decrypt() [all …]
|
| A D | sm4-ce-glue.c | 116 const u8 *src = walk.src.virt.addr; in sm4_ecb_do_crypt() 117 u8 *dst = walk.dst.virt.addr; in sm4_ecb_do_crypt() 164 const u8 *src = walk.src.virt.addr; in sm4_cbc_crypt() 165 u8 *dst = walk.dst.virt.addr; in sm4_cbc_crypt() 256 walk.src.virt.addr, walk.iv, walk.nbytes); in sm4_cbc_cts_crypt() 259 walk.src.virt.addr, walk.iv, walk.nbytes); in sm4_cbc_cts_crypt() 287 const u8 *src = walk.src.virt.addr; in sm4_ctr_crypt() 288 u8 *dst = walk.dst.virt.addr; in sm4_ctr_crypt() 366 walk.src.virt.addr, walk.iv, nbytes, in sm4_xts_crypt() 370 walk.src.virt.addr, walk.iv, nbytes, in sm4_xts_crypt() [all …]
|
| /arch/x86/crypto/ |
| A D | aegis128-aesni-glue.c | 115 aegis128_aesni_enc(state, walk->src.virt.addr, in crypto_aegis128_aesni_process_crypt() 116 walk->dst.virt.addr, in crypto_aegis128_aesni_process_crypt() 120 aegis128_aesni_dec(state, walk->src.virt.addr, in crypto_aegis128_aesni_process_crypt() 121 walk->dst.virt.addr, in crypto_aegis128_aesni_process_crypt() 132 aegis128_aesni_enc_tail(state, walk->src.virt.addr, in crypto_aegis128_aesni_process_crypt() 133 walk->dst.virt.addr, in crypto_aegis128_aesni_process_crypt() 136 aegis128_aesni_dec_tail(state, walk->src.virt.addr, in crypto_aegis128_aesni_process_crypt() 137 walk->dst.virt.addr, in crypto_aegis128_aesni_process_crypt()
|
| /arch/arm/mach-axxia/ |
| A D | platsmp.c | 24 u32 *virt = (u32 *) phys_to_virt(release_phys); in write_release_addr() local 25 writel_relaxed(__pa_symbol(secondary_startup), virt); in write_release_addr() 28 __cpuc_flush_dcache_area(virt, sizeof(u32)); in write_release_addr()
|
| /arch/sparc/crypto/ |
| A D | des_glue.c | 111 des_sparc64_ecb_crypt(walk.src.virt.addr, walk.dst.virt.addr, in __ecb_crypt() 153 des_sparc64_cbc_encrypt(walk.src.virt.addr, in __cbc_crypt() 154 walk.dst.virt.addr, in __cbc_crypt() 159 des_sparc64_cbc_decrypt(walk.src.virt.addr, in __cbc_crypt() 160 walk.dst.virt.addr, in __cbc_crypt() 262 des3_ede_sparc64_ecb_crypt(K, walk.src.virt.addr, in __ecb3_crypt() 263 walk.dst.virt.addr, in __ecb3_crypt() 309 des3_ede_sparc64_cbc_encrypt(K, walk.src.virt.addr, in __cbc3_crypt() 310 walk.dst.virt.addr, in __cbc3_crypt() 315 des3_ede_sparc64_cbc_decrypt(K, walk.src.virt.addr, in __cbc3_crypt() [all …]
|
| /arch/um/include/shared/ |
| A D | mem.h | 12 static inline unsigned long uml_to_phys(void *virt) in uml_to_phys() argument 14 return(((unsigned long) virt) - uml_physmem); in uml_to_phys()
|
| /arch/powerpc/crypto/ |
| A D | aes-spe-glue.c | 196 ppc_encrypt_ecb(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ecb_crypt() 199 ppc_decrypt_ecb(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ecb_crypt() 235 ppc_encrypt_cbc(walk.dst.virt.addr, walk.src.virt.addr, in ppc_cbc_crypt() 239 ppc_decrypt_cbc(walk.dst.virt.addr, walk.src.virt.addr, in ppc_cbc_crypt() 276 ppc_crypt_ctr(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ctr_crypt() 304 ppc_encrypt_xts(walk.dst.virt.addr, walk.src.virt.addr, in ppc_xts_crypt() 308 ppc_decrypt_xts(walk.dst.virt.addr, walk.src.virt.addr, in ppc_xts_crypt()
|
| /arch/arm64/kvm/hyp/include/nvhe/ |
| A D | memory.h | 91 #define hyp_virt_to_page(virt) hyp_phys_to_page(__hyp_pa(virt)) argument 92 #define hyp_virt_to_pfn(virt) hyp_phys_to_pfn(__hyp_pa(virt)) argument
|
| /arch/powerpc/mm/book3s32/ |
| A D | mmu.c | 116 static void setibat(int index, unsigned long virt, phys_addr_t phys, in setibat() argument 128 bat[0].batu = virt | (bl << 2) | 2; /* Vs=1, Vp=0 */ in setibat() 130 if (!is_kernel_addr(virt)) in setibat() 260 void __init setbat(int index, unsigned long virt, phys_addr_t phys, in setbat() argument 286 bat[1].batu = virt | (bl << 2) | 2; /* Vs=1, Vp=0 */ in setbat() 288 if (!is_kernel_addr(virt)) in setbat() 295 bat_addrs[index].start = virt; in setbat() 296 bat_addrs[index].limit = virt + ((bl + 1) << 17) - 1; in setbat()
|