Home
last modified time | relevance | path

Searched refs:virt (Results 1 – 25 of 211) sorted by relevance

123456789

/arch/m68k/mm/
A Dsun3kmap.c37 sun3_put_pte(virt, pte); in do_page_mapin()
41 print_pte_vaddr(virt); in do_page_mapin()
50 if(sun3_get_segmap(virt & ~SUN3_PMEG_MASK) == SUN3_INVALID_PMEG) in do_pmeg_mapin()
51 mmu_emu_map_pmeg(sun3_get_context(), virt); in do_pmeg_mapin()
54 do_page_mapin(phys, virt, type); in do_pmeg_mapin()
56 virt += PAGE_SIZE; in do_pmeg_mapin()
65 unsigned long offset, virt, ret; in sun3_ioremap() local
86 virt = (unsigned long)area->addr; in sun3_ioremap()
87 ret = virt + offset; in sun3_ioremap()
96 do_pmeg_mapin(phys, virt, type, seg_pages); in sun3_ioremap()
[all …]
/arch/parisc/kernel/
A Dkexec.c81 void *virt = (void *)__fix_to_virt(FIX_TEXT_KEXEC); in machine_kexec() local
90 desc.addr = (long long)virt; in machine_kexec()
92 reloc = (void *)virt; in machine_kexec()
95 memcpy(virt, dereference_function_descriptor(relocate_new_kernel), in machine_kexec()
98 *(unsigned long *)(virt + kexec_cmdline_offset) = arch->cmdline; in machine_kexec()
99 *(unsigned long *)(virt + kexec_initrd_start_offset) = arch->initrd_start; in machine_kexec()
100 *(unsigned long *)(virt + kexec_initrd_end_offset) = arch->initrd_end; in machine_kexec()
101 *(unsigned long *)(virt + kexec_free_mem_offset) = PAGE0->mem_free; in machine_kexec()
/arch/arm/mach-hisi/
A Dplatsmp.c108 void __iomem *virt; in hix5hd2_set_scu_boot_addr() local
110 virt = ioremap(start_addr, PAGE_SIZE); in hix5hd2_set_scu_boot_addr()
112 writel_relaxed(0xe51ff004, virt); /* ldr pc, [pc, #-4] */ in hix5hd2_set_scu_boot_addr()
113 writel_relaxed(jump_addr, virt + 4); /* pc jump phy address */ in hix5hd2_set_scu_boot_addr()
114 iounmap(virt); in hix5hd2_set_scu_boot_addr()
144 void __iomem *virt; in hip01_set_boot_addr() local
146 virt = phys_to_virt(start_addr); in hip01_set_boot_addr()
148 writel_relaxed(0xe51ff004, virt); in hip01_set_boot_addr()
149 writel_relaxed(jump_addr, virt + 4); in hip01_set_boot_addr()
/arch/powerpc/mm/nohash/
A De500.c111 TLBCAM[index].MAS2 = virt & PAGE_MASK; in settlbcam()
125 if (!is_kernel_addr(virt)) { in settlbcam()
133 tlbcam_addrs[index].start = virt; in settlbcam()
134 tlbcam_addrs[index].limit = virt + size - 1; in settlbcam()
142 unsigned int align = __ffs(virt | phys); in calc_cam_sz()
181 cam_sz = calc_cam_sz(boundary, virt, phys); in map_mem_in_cams_addr()
183 settlbcam(i, virt, phys, cam_sz, pgprot_val(prot), 0); in map_mem_in_cams_addr()
187 virt += cam_sz; in map_mem_in_cams_addr()
194 cam_sz = calc_cam_sz(ram, virt, phys); in map_mem_in_cams_addr()
200 virt += cam_sz; in map_mem_in_cams_addr()
[all …]
A D44x.c55 static void __init ppc44x_pin_tlb(unsigned int virt, unsigned int phys) in ppc44x_pin_tlb() argument
70 "r" (virt | PPC44x_TLB_VALID | PPC44x_TLB_256M), in ppc44x_pin_tlb()
125 static void __init ppc47x_pin_tlb(unsigned int virt, unsigned int phys) in ppc47x_pin_tlb() argument
141 virt, phys, bolted); in ppc47x_pin_tlb()
157 "r" (virt | PPC47x_TLB0_VALID | PPC47x_TLB0_256M), in ppc47x_pin_tlb()
/arch/x86/pci/
A Dmmconfig_64.c23 if (cfg && cfg->virt) in pci_dev_base()
24 return cfg->virt + (PCI_MMCFG_BUS_OFFSET(bus) | (devfn << 12)); in pci_dev_base()
116 cfg->virt = mcfg_ioremap(cfg); in pci_mmcfg_arch_map()
117 if (!cfg->virt) { in pci_mmcfg_arch_map()
127 if (cfg && cfg->virt) { in pci_mmcfg_arch_unmap()
128 iounmap(cfg->virt + PCI_MMCFG_BUS_OFFSET(cfg->start_bus)); in pci_mmcfg_arch_unmap()
129 cfg->virt = NULL; in pci_mmcfg_arch_unmap()
/arch/riscv/crypto/
A Daes-riscv64-glue.c148 walk.dst.virt.addr, in riscv64_aes_ecb_crypt()
152 walk.dst.virt.addr, in riscv64_aes_ecb_crypt()
186 walk.dst.virt.addr, in riscv64_aes_cbc_crypt()
191 walk.dst.virt.addr, in riscv64_aes_cbc_crypt()
259 aes_cbc_cts_crypt_zvkned(ctx, walk.src.virt.addr, walk.dst.virt.addr, in riscv64_aes_cbc_cts_crypt()
317 walk.dst.virt.addr, in riscv64_aes_ctr_crypt()
324 walk.src.virt.addr + p1_nbytes, in riscv64_aes_ctr_crypt()
407 &ctx->ctx1, walk.src.virt.addr, in riscv64_aes_xts_crypt()
411 &ctx->ctx1, walk.src.virt.addr, in riscv64_aes_xts_crypt()
436 &ctx->ctx1, walk.src.virt.addr, in riscv64_aes_xts_crypt()
[all …]
/arch/arm/mm/
A Dioremap.c111 int ioremap_page(unsigned long virt, unsigned long phys, in ioremap_page() argument
114 return vmap_page_range(virt, virt + PAGE_SIZE, phys, in ioremap_page()
173 static void unmap_area_sections(unsigned long virt, unsigned long size) in unmap_area_sections() argument
175 unsigned long addr = virt, end = virt + (size & ~(SZ_1M - 1)); in unmap_area_sections()
209 flush_tlb_kernel_range(virt, end); in unmap_area_sections()
213 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument
216 unsigned long addr = virt, end = virt + size; in remap_area_sections()
223 unmap_area_sections(virt, size); in remap_area_sections()
240 remap_area_supersections(unsigned long virt, unsigned long pfn, in remap_area_supersections() argument
243 unsigned long addr = virt, end = virt + size; in remap_area_supersections()
[all …]
/arch/xtensa/mm/
A Dcache.c146 unsigned long virt; in flush_dcache_folio() local
160 virt = TLBTEMP_BASE_1 + (phys & DCACHE_ALIAS_MASK); in flush_dcache_folio()
161 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_folio()
163 virt = TLBTEMP_BASE_1 + (temp & DCACHE_ALIAS_MASK); in flush_dcache_folio()
166 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_folio()
169 __invalidate_icache_page_alias(virt, phys); in flush_dcache_folio()
205 unsigned long virt = TLBTEMP_BASE_1 + (address & DCACHE_ALIAS_MASK); in local_flush_cache_page() local
208 __flush_invalidate_dcache_page_alias(virt, phys); in local_flush_cache_page()
209 __invalidate_icache_page_alias(virt, phys); in local_flush_cache_page()
/arch/um/kernel/
A Dphysmem.c25 void map_memory(unsigned long virt, unsigned long phys, unsigned long len, in map_memory() argument
32 err = os_map_memory((void *) virt, fd, offset, len, r, w, x); in map_memory()
39 "err = %d\n", virt, fd, offset, len, r, w, x, err); in map_memory()
166 return region->virt; in find_iomem()
189 region->virt = iomem_start; in setup_iomem()
190 region->phys = __pa(region->virt); in setup_iomem()
A Dtlb.c22 unsigned long virt, unsigned long len, int prot,
25 unsigned long virt, unsigned long len);
29 unsigned long virt, unsigned long len, int prot, in kern_map() argument
33 return os_map_memory((void *)virt, phys_fd, offset, len, in kern_map()
39 unsigned long virt, unsigned long len) in kern_unmap() argument
41 return os_unmap_memory((void *)virt, len); in kern_unmap()
/arch/arm/crypto/
A Daes-ce-glue.c177 ce_aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_encrypt()
197 ce_aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_decrypt()
215 ce_aes_cbc_encrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_encrypt_walk()
245 ce_aes_cbc_decrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_decrypt_walk()
315 ce_aes_cbc_cts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_encrypt()
373 ce_aes_cbc_cts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_decrypt()
392 ce_aes_ctr_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ctr_encrypt()
461 ce_aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt()
483 ce_aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt()
533 ce_aes_xts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_decrypt()
[all …]
/arch/powerpc/boot/
A Doflib.c152 unsigned int of_claim(unsigned long virt, unsigned long size, in of_claim() argument
161 return of_call_prom("claim", 3, 1, virt, size, align); in of_claim()
164 align, size, virt); in of_claim()
168 align, size, virt); in of_claim()
171 0x12, size, virt, virt); in of_claim()
172 return virt; in of_claim()
/arch/powerpc/kernel/
A Dexceptions-64s.S265 .if \virt
306 .if !\virt
373 GEN_BRANCH_TO_COMMON \name \virt
1430 GEN_INT_ENTRY data_access, virt=0
1433 GEN_INT_ENTRY data_access, virt=1
1674 GEN_INT_ENTRY alignment, virt=0
1677 GEN_INT_ENTRY alignment, virt=1
1834 GEN_INT_ENTRY decrementer, virt=0
1837 GEN_INT_ENTRY decrementer, virt=1
1970 .macro SYSTEM_CALL virt
[all …]
/arch/arm64/kvm/hyp/nvhe/
A Dmem_protect.c776 void *virt = __hyp_va(phys); in __pkvm_host_donate_hyp() local
790 WARN_ON(pkvm_create_mappings_locked(virt, virt + size, PAGE_HYP)); in __pkvm_host_donate_hyp()
1257 WARN_ON(!virt); in pkvm_ownership_selftest()
1276 assert_transition_res(-EPERM, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest()
1287 assert_transition_res(-EPERM, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest()
1299 assert_transition_res(0, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest()
1300 assert_transition_res(0, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest()
1301 hyp_unpin_shared_mem(virt, virt + size); in pkvm_ownership_selftest()
1311 hyp_unpin_shared_mem(virt, virt + size); in pkvm_ownership_selftest()
1329 assert_transition_res(-EPERM, hyp_pin_shared_mem, virt, virt + size); in pkvm_ownership_selftest()
[all …]
/arch/arm64/crypto/
A Dsm4-neon-glue.c45 const u8 *src = walk.src.virt.addr; in sm4_ecb_do_crypt()
46 u8 *dst = walk.dst.virt.addr; in sm4_ecb_do_crypt()
92 const u8 *src = walk.src.virt.addr; in sm4_cbc_encrypt()
93 u8 *dst = walk.dst.virt.addr; in sm4_cbc_encrypt()
123 const u8 *src = walk.src.virt.addr; in sm4_cbc_decrypt()
124 u8 *dst = walk.dst.virt.addr; in sm4_cbc_decrypt()
154 const u8 *src = walk.src.virt.addr; in sm4_ctr_crypt()
155 u8 *dst = walk.dst.virt.addr; in sm4_ctr_crypt()
A Daes-glue.c191 aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_encrypt()
211 aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in ecb_decrypt()
229 aes_cbc_encrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_encrypt_walk()
258 aes_cbc_decrypt(walk->dst.virt.addr, walk->src.virt.addr, in cbc_decrypt_walk()
327 aes_cbc_cts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_encrypt()
384 aes_cbc_cts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in cts_cbc_decrypt()
420 aes_essiv_cbc_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in essiv_cbc_encrypt()
583 aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt()
605 aes_xts_encrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_encrypt()
655 aes_xts_decrypt(walk.dst.virt.addr, walk.src.virt.addr, in xts_decrypt()
[all …]
A Dsm4-ce-glue.c116 const u8 *src = walk.src.virt.addr; in sm4_ecb_do_crypt()
117 u8 *dst = walk.dst.virt.addr; in sm4_ecb_do_crypt()
164 const u8 *src = walk.src.virt.addr; in sm4_cbc_crypt()
165 u8 *dst = walk.dst.virt.addr; in sm4_cbc_crypt()
256 walk.src.virt.addr, walk.iv, walk.nbytes); in sm4_cbc_cts_crypt()
259 walk.src.virt.addr, walk.iv, walk.nbytes); in sm4_cbc_cts_crypt()
287 const u8 *src = walk.src.virt.addr; in sm4_ctr_crypt()
288 u8 *dst = walk.dst.virt.addr; in sm4_ctr_crypt()
366 walk.src.virt.addr, walk.iv, nbytes, in sm4_xts_crypt()
370 walk.src.virt.addr, walk.iv, nbytes, in sm4_xts_crypt()
[all …]
/arch/x86/crypto/
A Daegis128-aesni-glue.c115 aegis128_aesni_enc(state, walk->src.virt.addr, in crypto_aegis128_aesni_process_crypt()
116 walk->dst.virt.addr, in crypto_aegis128_aesni_process_crypt()
120 aegis128_aesni_dec(state, walk->src.virt.addr, in crypto_aegis128_aesni_process_crypt()
121 walk->dst.virt.addr, in crypto_aegis128_aesni_process_crypt()
132 aegis128_aesni_enc_tail(state, walk->src.virt.addr, in crypto_aegis128_aesni_process_crypt()
133 walk->dst.virt.addr, in crypto_aegis128_aesni_process_crypt()
136 aegis128_aesni_dec_tail(state, walk->src.virt.addr, in crypto_aegis128_aesni_process_crypt()
137 walk->dst.virt.addr, in crypto_aegis128_aesni_process_crypt()
/arch/arm/mach-axxia/
A Dplatsmp.c24 u32 *virt = (u32 *) phys_to_virt(release_phys); in write_release_addr() local
25 writel_relaxed(__pa_symbol(secondary_startup), virt); in write_release_addr()
28 __cpuc_flush_dcache_area(virt, sizeof(u32)); in write_release_addr()
/arch/sparc/crypto/
A Ddes_glue.c111 des_sparc64_ecb_crypt(walk.src.virt.addr, walk.dst.virt.addr, in __ecb_crypt()
153 des_sparc64_cbc_encrypt(walk.src.virt.addr, in __cbc_crypt()
154 walk.dst.virt.addr, in __cbc_crypt()
159 des_sparc64_cbc_decrypt(walk.src.virt.addr, in __cbc_crypt()
160 walk.dst.virt.addr, in __cbc_crypt()
262 des3_ede_sparc64_ecb_crypt(K, walk.src.virt.addr, in __ecb3_crypt()
263 walk.dst.virt.addr, in __ecb3_crypt()
309 des3_ede_sparc64_cbc_encrypt(K, walk.src.virt.addr, in __cbc3_crypt()
310 walk.dst.virt.addr, in __cbc3_crypt()
315 des3_ede_sparc64_cbc_decrypt(K, walk.src.virt.addr, in __cbc3_crypt()
[all …]
/arch/um/include/shared/
A Dmem.h12 static inline unsigned long uml_to_phys(void *virt) in uml_to_phys() argument
14 return(((unsigned long) virt) - uml_physmem); in uml_to_phys()
/arch/powerpc/crypto/
A Daes-spe-glue.c196 ppc_encrypt_ecb(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ecb_crypt()
199 ppc_decrypt_ecb(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ecb_crypt()
235 ppc_encrypt_cbc(walk.dst.virt.addr, walk.src.virt.addr, in ppc_cbc_crypt()
239 ppc_decrypt_cbc(walk.dst.virt.addr, walk.src.virt.addr, in ppc_cbc_crypt()
276 ppc_crypt_ctr(walk.dst.virt.addr, walk.src.virt.addr, in ppc_ctr_crypt()
304 ppc_encrypt_xts(walk.dst.virt.addr, walk.src.virt.addr, in ppc_xts_crypt()
308 ppc_decrypt_xts(walk.dst.virt.addr, walk.src.virt.addr, in ppc_xts_crypt()
/arch/arm64/kvm/hyp/include/nvhe/
A Dmemory.h91 #define hyp_virt_to_page(virt) hyp_phys_to_page(__hyp_pa(virt)) argument
92 #define hyp_virt_to_pfn(virt) hyp_phys_to_pfn(__hyp_pa(virt)) argument
/arch/powerpc/mm/book3s32/
A Dmmu.c116 static void setibat(int index, unsigned long virt, phys_addr_t phys, in setibat() argument
128 bat[0].batu = virt | (bl << 2) | 2; /* Vs=1, Vp=0 */ in setibat()
130 if (!is_kernel_addr(virt)) in setibat()
260 void __init setbat(int index, unsigned long virt, phys_addr_t phys, in setbat() argument
286 bat[1].batu = virt | (bl << 2) | 2; /* Vs=1, Vp=0 */ in setbat()
288 if (!is_kernel_addr(virt)) in setbat()
295 bat_addrs[index].start = virt; in setbat()
296 bat_addrs[index].limit = virt + ((bl + 1) << 17) - 1; in setbat()

Completed in 44 milliseconds

123456789