Home
last modified time | relevance | path

Searched refs:vaddr (Results 1 – 25 of 290) sorted by relevance

12345678910>>...12

/arch/m68k/include/asm/
A Dbitops.h59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) argument
61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) argument
102 #define clear_bit(nr, vaddr) bclr_reg_clear_bit(nr, vaddr) argument
104 #define clear_bit(nr, vaddr) bclr_mem_clear_bit(nr, vaddr) argument
145 #define change_bit(nr, vaddr) bchg_reg_change_bit(nr, vaddr) argument
147 #define change_bit(nr, vaddr) bchg_mem_change_bit(nr, vaddr) argument
201 #define test_and_set_bit(nr, vaddr) bset_reg_test_and_set_bit(nr, vaddr) argument
203 #define test_and_set_bit(nr, vaddr) bset_mem_test_and_set_bit(nr, vaddr) argument
254 #define test_and_clear_bit(nr, vaddr) bclr_reg_test_and_clear_bit(nr, vaddr) argument
256 #define test_and_clear_bit(nr, vaddr) bclr_mem_test_and_clear_bit(nr, vaddr) argument
[all …]
/arch/riscv/mm/
A Dkasan_init.c47 } while (ptep++, vaddr += PAGE_SIZE, vaddr != end); in kasan_populate_pte()
77 } while (pmdp++, vaddr = next, vaddr != end); in kasan_populate_pmd()
108 } while (pudp++, vaddr = next, vaddr != end); in kasan_populate_pud()
139 } while (p4dp++, vaddr = next, vaddr != end); in kasan_populate_p4d()
162 } while (pgdp++, vaddr = next, vaddr != end); in kasan_populate_pgd()
187 } while (pudp++, vaddr = next, vaddr != end); in kasan_early_clear_pud()
213 } while (p4dp++, vaddr = next, vaddr != end); in kasan_early_clear_p4d()
231 } while (pgdp++, vaddr = next, vaddr != end); in kasan_early_clear_pgd()
260 } while (pudp++, vaddr = next, vaddr != end); in kasan_early_populate_pud()
298 } while (p4dp++, vaddr = next, vaddr != end); in kasan_early_populate_p4d()
[all …]
A Dpageattr.c102 pmdp = pmd_offset(pudp, vaddr); in __split_linear_mapping_pmd()
107 if (next - vaddr >= PMD_SIZE && in __split_linear_mapping_pmd()
108 vaddr <= (vaddr & PMD_MASK) && end >= next) in __split_linear_mapping_pmd()
130 } while (pmdp++, vaddr = next, vaddr != end); in __split_linear_mapping_pmd()
142 pudp = pud_offset(p4dp, vaddr); in __split_linear_mapping_pud()
148 vaddr <= (vaddr & PUD_MASK) && end >= next) in __split_linear_mapping_pud()
175 } while (pudp++, vaddr = next, vaddr != end); in __split_linear_mapping_pud()
197 vaddr <= (vaddr & P4D_MASK) && end >= next) in __split_linear_mapping_p4d()
233 } while (p4dp++, vaddr = next, vaddr != end); in __split_linear_mapping_p4d()
239 unsigned long vaddr, in __split_linear_mapping_pgd() argument
[all …]
/arch/parisc/kernel/
A Dpci-dma.c83 vaddr &= ~PMD_MASK; in map_pte_uncached()
84 end = vaddr + size; in map_pte_uncached()
120 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in map_pmd_uncached()
145 if (map_pmd_uncached(pmd, vaddr, end - vaddr, &paddr)) in map_uncached_pages()
147 vaddr = vaddr + PGDIR_SIZE; in map_uncached_pages()
149 } while (vaddr && (vaddr < end)); in map_uncached_pages()
203 pmd = pmd_offset(pud_offset(p4d_offset(dir, vaddr), vaddr), vaddr); in unmap_uncached_pmd()
210 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in unmap_uncached_pmd()
223 unmap_uncached_pmd(dir, vaddr, end - vaddr); in unmap_uncached_pages()
224 vaddr = vaddr + PGDIR_SIZE; in unmap_uncached_pages()
[all …]
/arch/arm/mm/
A Dcache-xsc3l2.c88 unsigned long vaddr; in xsc3_l2_inv_range() local
101 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); in xsc3_l2_inv_range()
111 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range()
120 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range()
125 l2_unmap_va(vaddr); in xsc3_l2_inv_range()
132 unsigned long vaddr; in xsc3_l2_clean_range() local
138 vaddr = l2_map_va(start, vaddr); in xsc3_l2_clean_range()
143 l2_unmap_va(vaddr); in xsc3_l2_clean_range()
170 unsigned long vaddr; in xsc3_l2_flush_range() local
181 vaddr = l2_map_va(start, vaddr); in xsc3_l2_flush_range()
[all …]
/arch/s390/include/asm/
A Didals.h35 dma64_t paddr = virt_to_dma64(vaddr); in idal_is_needed()
71 dma64_t paddr = virt_to_dma64(vaddr); in idal_create_words()
102 vaddr = idal; in set_normalized_cda()
104 ccw->cda = virt_to_dma32(vaddr); in set_normalized_cda()
136 void *vaddr; in idal_buffer_alloc() local
151 if (!vaddr) in idal_buffer_alloc()
172 void *vaddr; in idal_buffer_free() local
198 void *vaddr; in idal_buffer_set_cda() local
211 ccw->cda = virt_to_dma32(vaddr); in idal_buffer_set_cda()
222 void *vaddr; in idal_buffer_to_user() local
[all …]
/arch/parisc/mm/
A Dfixmap.c15 unsigned long vaddr = __fix_to_virt(idx); in set_fixmap() local
16 pgd_t *pgd = pgd_offset_k(vaddr); in set_fixmap()
17 p4d_t *p4d = p4d_offset(pgd, vaddr); in set_fixmap()
18 pud_t *pud = pud_offset(p4d, vaddr); in set_fixmap()
19 pmd_t *pmd = pmd_offset(pud, vaddr); in set_fixmap()
22 pte = pte_offset_kernel(pmd, vaddr); in set_fixmap()
24 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); in set_fixmap()
29 unsigned long vaddr = __fix_to_virt(idx); in clear_fixmap() local
30 pte_t *pte = virt_to_kpte(vaddr); in clear_fixmap()
35 pte_clear(&init_mm, vaddr, pte); in clear_fixmap()
[all …]
/arch/arm/include/asm/
A Dpage.h114 unsigned long vaddr, struct vm_area_struct *vma);
117 unsigned long vaddr, struct vm_area_struct *vma);
120 unsigned long vaddr, struct vm_area_struct *vma);
123 unsigned long vaddr, struct vm_area_struct *vma);
126 unsigned long vaddr, struct vm_area_struct *vma);
129 unsigned long vaddr, struct vm_area_struct *vma);
132 unsigned long vaddr, struct vm_area_struct *vma);
151 #define clear_user_highpage(page,vaddr) \ argument
152 __cpu_clear_user_highpage(page, vaddr)
155 #define copy_user_highpage(to,from,vaddr,vma) \ argument
[all …]
/arch/m68k/sun3/
A Ddvma.c23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument
35 if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { in dvma_page()
36 sun3_put_pte(vaddr, pte); in dvma_page()
37 ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; in dvma_page()
40 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page()
49 unsigned long vaddr; in dvma_map_iommu() local
51 vaddr = dvma_btov(baddr); in dvma_map_iommu()
53 end = vaddr + len; in dvma_map_iommu()
55 while(vaddr < end) { in dvma_map_iommu()
56 dvma_page(kaddr, vaddr); in dvma_map_iommu()
[all …]
A Dmmu_emu.c119 void print_pte_vaddr (unsigned long vaddr) in print_pte_vaddr() argument
121 pr_cont(" vaddr=%lx [%02lx]", vaddr, sun3_get_segmap (vaddr)); in print_pte_vaddr()
122 print_pte (__pte (sun3_get_pte (vaddr))); in print_pte_vaddr()
289 vaddr &= ~SUN3_PMEG_MASK; in mmu_emu_map_pmeg()
298 curr_pmeg, context, vaddr); in mmu_emu_map_pmeg()
310 if(vaddr >= PAGE_OFFSET) { in mmu_emu_map_pmeg()
316 sun3_put_segmap (vaddr, curr_pmeg); in mmu_emu_map_pmeg()
326 sun3_put_segmap (vaddr, curr_pmeg); in mmu_emu_map_pmeg()
329 pmeg_vaddr[curr_pmeg] = vaddr; in mmu_emu_map_pmeg()
401 mmu_emu_map_pmeg (context, vaddr); in mmu_emu_handle_fault()
[all …]
/arch/x86/mm/
A Dmem_encrypt_amd.c169 vaddr += PMD_SIZE; in __sme_early_map_unmap_mem()
261 while (vaddr < vaddr_end) { in enc_dec_hypercall()
266 kpte = lookup_address(vaddr, &level); in enc_dec_hypercall()
281 vaddr = (vaddr & pmask) + psize; in enc_dec_hypercall()
293 snp_set_memory_shared(vaddr, npages); in amd_enc_status_change_prepare()
401 start = vaddr; in early_set_memory_enc_dec()
402 vaddr_next = vaddr; in early_set_memory_enc_dec()
403 vaddr_end = vaddr + size; in early_set_memory_enc_dec()
405 for (; vaddr < vaddr_end; vaddr = vaddr_next) { in early_set_memory_enc_dec()
427 if (vaddr == (vaddr & pmask) && in early_set_memory_enc_dec()
[all …]
A Dpgtable_32.c27 void set_pte_vaddr(unsigned long vaddr, pte_t pteval) in set_pte_vaddr() argument
35 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pte_vaddr()
40 p4d = p4d_offset(pgd, vaddr); in set_pte_vaddr()
45 pud = pud_offset(p4d, vaddr); in set_pte_vaddr()
50 pmd = pmd_offset(pud, vaddr); in set_pte_vaddr()
55 pte = pte_offset_kernel(pmd, vaddr); in set_pte_vaddr()
57 set_pte_at(&init_mm, vaddr, pte, pteval); in set_pte_vaddr()
59 pte_clear(&init_mm, vaddr, pte); in set_pte_vaddr()
65 flush_tlb_one_kernel(vaddr); in set_pte_vaddr()
A Dkaslr.c82 unsigned long vaddr_start, vaddr; in kernel_randomize_memory() local
89 vaddr = vaddr_start; in kernel_randomize_memory()
151 vaddr += entropy; in kernel_randomize_memory()
152 *kaslr_regions[i].base = vaddr; in kernel_randomize_memory()
155 vaddr += get_padding(&kaslr_regions[i]); in kernel_randomize_memory()
166 vaddr = round_up(vaddr + 1, PUD_SIZE); in kernel_randomize_memory()
175 unsigned long paddr, vaddr; in init_trampoline_kaslr() local
188 vaddr = (unsigned long)__va(paddr); in init_trampoline_kaslr()
189 pgd = pgd_offset_k(vaddr); in init_trampoline_kaslr()
191 p4d = p4d_offset(pgd, vaddr); in init_trampoline_kaslr()
[all …]
A Dinit_32.c132 unsigned long vaddr; in page_table_range_init_count() local
137 vaddr = start; in page_table_range_init_count()
138 pgd_idx = pgd_index(vaddr); in page_table_range_init_count()
139 pmd_idx = pmd_index(vaddr); in page_table_range_init_count()
147 vaddr += PMD_SIZE; in page_table_range_init_count()
207 unsigned long vaddr; in page_table_range_init() local
217 vaddr = start; in page_table_range_init()
218 pgd_idx = pgd_index(vaddr); in page_table_range_init()
230 vaddr += PMD_SIZE; in page_table_range_init()
393 page_table_range_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in permanent_kmaps_init()
[all …]
/arch/sh/mm/
A Dkmap.c21 unsigned long vaddr; in kmap_coherent_init() local
24 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); in kmap_coherent_init()
25 kmap_coherent_pte = virt_to_kpte(vaddr); in kmap_coherent_init()
32 unsigned long vaddr; in kmap_coherent() local
43 vaddr = __fix_to_virt(idx); in kmap_coherent()
48 return (void *)vaddr; in kmap_coherent()
54 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; in kunmap_coherent() local
55 enum fixed_addresses idx = __virt_to_fix(vaddr); in kunmap_coherent()
58 __flush_purge_region((void *)vaddr, PAGE_SIZE); in kunmap_coherent()
60 pte_clear(&init_mm, vaddr, kmap_coherent_pte - idx); in kunmap_coherent()
[all …]
/arch/m68k/sun3x/
A Ddvma.c79 unsigned long vaddr, int len) in dvma_map_cpu() argument
88 vaddr &= PAGE_MASK; in dvma_map_cpu()
90 end = PAGE_ALIGN(vaddr + len); in dvma_map_cpu()
93 pgd = pgd_offset_k(vaddr); in dvma_map_cpu()
94 p4d = p4d_offset(pgd, vaddr); in dvma_map_cpu()
95 pud = pud_offset(p4d, vaddr); in dvma_map_cpu()
127 __pa(kaddr), vaddr); in dvma_map_cpu()
132 vaddr += PAGE_SIZE; in dvma_map_cpu()
133 } while(vaddr < end3); in dvma_map_cpu()
135 } while(vaddr < end2); in dvma_map_cpu()
[all …]
/arch/mips/mm/
A Dpgtable-32.c43 unsigned long vaddr; in pagetable_init() local
62 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1); in pagetable_init()
63 fixrange_init(vaddr & PMD_MASK, vaddr + FIXADDR_SIZE, pgd_base); in pagetable_init()
69 vaddr = PKMAP_BASE; in pagetable_init()
70 fixrange_init(vaddr & PMD_MASK, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in pagetable_init()
72 pgd = swapper_pg_dir + pgd_index(vaddr); in pagetable_init()
73 p4d = p4d_offset(pgd, vaddr); in pagetable_init()
74 pud = pud_offset(p4d, vaddr); in pagetable_init()
75 pmd = pmd_offset(pud, vaddr); in pagetable_init()
76 pte = pte_offset_kernel(pmd, vaddr); in pagetable_init()
/arch/sparc/include/asm/
A Dviking.h218 vaddr &= PAGE_MASK; in viking_hwprobe()
222 : "r" (vaddr | 0x400), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe()
229 : "r" (vaddr | 0x200), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe()
231 vaddr &= ~PGDIR_MASK; in viking_hwprobe()
232 vaddr >>= PAGE_SHIFT; in viking_hwprobe()
233 return val | (vaddr << 8); in viking_hwprobe()
239 : "r" (vaddr | 0x100), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe()
241 vaddr &= ~PMD_MASK; in viking_hwprobe()
242 vaddr >>= PAGE_SHIFT; in viking_hwprobe()
243 return val | (vaddr << 8); in viking_hwprobe()
[all …]
/arch/sparc/mm/
A Dio-unit.c136 vaddr = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + (vaddr & ~PAGE_MASK); in iounit_get_area()
141 IOD(("%08lx\n", vaddr)); in iounit_get_area()
142 return vaddr; in iounit_get_area()
189 vaddr = (vaddr - IOUNIT_DMA_BASE) >> PAGE_SHIFT; in iounit_unmap_page()
190 IOD(("iounit_release %08lx-%08lx\n", (long)vaddr, (long)len+vaddr)); in iounit_unmap_page()
191 for (len += vaddr; vaddr < len; vaddr++) in iounit_unmap_page()
192 clear_bit(vaddr, iounit->bmap); in iounit_unmap_page()
200 unsigned long flags, vaddr, len; in iounit_unmap_sg() local
208 IOD(("iounit_release %08lx-%08lx\n", (long)vaddr, (long)len+vaddr)); in iounit_unmap_sg()
209 for (len += vaddr; vaddr < len; vaddr++) in iounit_unmap_sg()
[all …]
/arch/nios2/mm/
A Ddma-mapping.c24 void *vaddr = phys_to_virt(paddr); in arch_sync_dma_for_device() local
28 invalidate_dcache_range((unsigned long)vaddr, in arch_sync_dma_for_device()
29 (unsigned long)(vaddr + size)); in arch_sync_dma_for_device()
37 flush_dcache_range((unsigned long)vaddr, in arch_sync_dma_for_device()
38 (unsigned long)(vaddr + size)); in arch_sync_dma_for_device()
48 void *vaddr = phys_to_virt(paddr); in arch_sync_dma_for_cpu() local
53 invalidate_dcache_range((unsigned long)vaddr, in arch_sync_dma_for_cpu()
54 (unsigned long)(vaddr + size)); in arch_sync_dma_for_cpu()
/arch/x86/include/asm/
A Dpage.h25 static inline void clear_user_page(void *page, unsigned long vaddr, in clear_user_page() argument
31 static inline void copy_user_page(void *to, void *from, unsigned long vaddr, in copy_user_page() argument
37 #define vma_alloc_zeroed_movable_folio(vma, vaddr) \ argument
38 vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
77 static __always_inline u64 __canonical_address(u64 vaddr, u8 vaddr_bits) in __canonical_address() argument
79 return ((s64)vaddr << (64 - vaddr_bits)) >> (64 - vaddr_bits); in __canonical_address()
82 static __always_inline u64 __is_canonical_address(u64 vaddr, u8 vaddr_bits) in __is_canonical_address() argument
84 return __canonical_address(vaddr, vaddr_bits) == vaddr; in __is_canonical_address()
/arch/csky/mm/
A Dhighmem.c22 unsigned long vaddr; in kmap_init() local
28 vaddr = PKMAP_BASE; in kmap_init()
29 fixrange_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, swapper_pg_dir); in kmap_init()
31 pgd = swapper_pg_dir + pgd_index(vaddr); in kmap_init()
33 pmd = pmd_offset(pud, vaddr); in kmap_init()
34 pte = pte_offset_kernel(pmd, vaddr); in kmap_init()
A Dtcm.c29 unsigned long vaddr, paddr; local
42 vaddr = __fix_to_virt(FIX_TCM - i);
45 pte_offset_kernel((pmd_t *)pgd_offset_k(vaddr), vaddr);
49 flush_tlb_one(vaddr);
61 vaddr = __fix_to_virt(FIX_TCM - CONFIG_ITCM_NR_PAGES - i);
64 pte_offset_kernel((pmd_t *) pgd_offset_k(vaddr), vaddr);
68 flush_tlb_one(vaddr);
111 unsigned long vaddr; local
116 vaddr = gen_pool_alloc(tcm_pool, len);
117 if (!vaddr)
[all …]
A Dinit.c95 unsigned long vaddr; in fixrange_init() local
97 vaddr = start; in fixrange_init()
98 i = pgd_index(vaddr); in fixrange_init()
99 j = pud_index(vaddr); in fixrange_init()
100 k = pmd_index(vaddr); in fixrange_init()
103 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { in fixrange_init()
105 for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) { in fixrange_init()
107 for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) { in fixrange_init()
118 vaddr += PMD_SIZE; in fixrange_init()
128 unsigned long vaddr; in fixaddr_init() local
[all …]
/arch/xtensa/mm/
A Dcache.c60 unsigned long vaddr) in kmap_invalidate_coherent() argument
62 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { in kmap_invalidate_coherent()
82 unsigned long vaddr, unsigned long *paddr) in coherent_kvaddr() argument
85 return (void *)(base + (vaddr & DCACHE_ALIAS_MASK)); in coherent_kvaddr()
95 kmap_invalidate_coherent(page, vaddr); in clear_user_highpage()
103 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument
113 kmap_invalidate_coherent(dst, vaddr); in copy_user_highpage()
274 unsigned long vaddr, void *dst, const void *src, in copy_to_user_page() argument
278 unsigned long alias = !(DCACHE_ALIAS_EQ(vaddr, phys)); in copy_to_user_page()
314 unsigned long vaddr, void *dst, const void *src, in copy_from_user_page() argument
[all …]

Completed in 49 milliseconds

12345678910>>...12