| /arch/hexagon/kernel/ |
| A D | vdso.c | 53 unsigned long vdso_base; in arch_setup_additional_pages() local 64 vdso_base = STACK_TOP; in arch_setup_additional_pages() 66 vdso_base = get_unmapped_area(NULL, vdso_base, PAGE_SIZE, 0, 0); in arch_setup_additional_pages() 67 if (IS_ERR_VALUE(vdso_base)) { in arch_setup_additional_pages() 68 ret = vdso_base; in arch_setup_additional_pages() 74 vma = _install_special_mapping(mm, vdso_base, PAGE_SIZE, in arch_setup_additional_pages() 83 mm->context.vdso = (void *)vdso_base; in arch_setup_additional_pages()
|
| /arch/csky/kernel/ |
| A D | vdso.c | 45 unsigned long vdso_base, vdso_len; in arch_setup_additional_pages() local 54 vdso_base = get_unmapped_area(NULL, 0, vdso_len, 0, 0); in arch_setup_additional_pages() 55 if (IS_ERR_VALUE(vdso_base)) { in arch_setup_additional_pages() 56 ret = vdso_base; in arch_setup_additional_pages() 65 mm->context.vdso = (void *)vdso_base; in arch_setup_additional_pages() 69 _install_special_mapping(mm, vdso_base, vdso_pages << PAGE_SHIFT, in arch_setup_additional_pages() 79 vdso_base += (vdso_pages << PAGE_SHIFT); in arch_setup_additional_pages()
|
| /arch/riscv/kernel/ |
| A D | vdso.c | 115 unsigned long vdso_base, vdso_text_len, vdso_mapping_len; in __setup_additional_pages() local 124 vdso_base = get_unmapped_area(NULL, 0, vdso_mapping_len, 0, 0); in __setup_additional_pages() 125 if (IS_ERR_VALUE(vdso_base)) { in __setup_additional_pages() 126 ret = ERR_PTR(vdso_base); in __setup_additional_pages() 130 ret = vdso_install_vvar_mapping(mm, vdso_base); in __setup_additional_pages() 134 vdso_base += VVAR_SIZE; in __setup_additional_pages() 135 mm->context.vdso = (void *)vdso_base; in __setup_additional_pages() 138 _install_special_mapping(mm, vdso_base, vdso_text_len, in __setup_additional_pages()
|
| /arch/powerpc/kernel/ |
| A D | vdso.c | 101 unsigned long vdso_size, vdso_base, mappings_size; in __arch_setup_additional_pages() local 122 vdso_base = get_unmapped_area(NULL, 0, mappings_size, 0, 0); in __arch_setup_additional_pages() 123 if (IS_ERR_VALUE(vdso_base)) in __arch_setup_additional_pages() 124 return vdso_base; in __arch_setup_additional_pages() 127 vdso_base = ALIGN(vdso_base, VDSO_ALIGNMENT); in __arch_setup_additional_pages() 129 vma = vdso_install_vvar_mapping(mm, vdso_base); in __arch_setup_additional_pages() 143 vma = _install_special_mapping(mm, vdso_base + vvar_size, vdso_size, in __arch_setup_additional_pages() 147 do_munmap(mm, vdso_base, vvar_size, NULL); in __arch_setup_additional_pages() 152 mm->context.vdso = (void __user *)vdso_base + vvar_size; in __arch_setup_additional_pages()
|
| /arch/arm64/kernel/ |
| A D | vdso.c | 106 unsigned long vdso_base, vdso_text_len, vdso_mapping_len; in __setup_additional_pages() local 116 vdso_base = get_unmapped_area(NULL, 0, vdso_mapping_len, 0, 0); in __setup_additional_pages() 117 if (IS_ERR_VALUE(vdso_base)) { in __setup_additional_pages() 118 ret = ERR_PTR(vdso_base); in __setup_additional_pages() 122 ret = vdso_install_vvar_mapping(mm, vdso_base); in __setup_additional_pages() 129 vdso_base += VDSO_NR_PAGES * PAGE_SIZE; in __setup_additional_pages() 130 mm->context.vdso = (void *)vdso_base; in __setup_additional_pages() 131 ret = _install_special_mapping(mm, vdso_base, vdso_text_len, in __setup_additional_pages()
|
| /arch/parisc/include/asm/ |
| A D | vdso.h | 12 #define VDSO64_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso64_offset_##name)) 13 #define VDSO32_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso32_offset_##name))
|
| A D | mmu.h | 7 unsigned long vdso_base; member
|
| A D | elf.h | 359 #define VDSO_CURRENT_BASE current->mm->context.vdso_base
|
| /arch/s390/include/asm/ |
| A D | vdso-symbols.h | 10 #define VDSO64_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso64_offset_##name)) 12 #define VDSO32_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso32_offset_##name))
|
| A D | mmu.h | 18 unsigned long vdso_base; member
|
| A D | elf.h | 270 (unsigned long)current->mm->context.vdso_base); \
|
| /arch/parisc/kernel/ |
| A D | vdso.c | 32 current->mm->context.vdso_base = vma->vm_start; in vdso_mremap() 91 current->mm->context.vdso_base = vdso_text_start; in arch_setup_additional_pages()
|
| /arch/loongarch/kernel/ |
| A D | vdso.c | 65 static unsigned long vdso_base(void) in vdso_base() function 94 data_addr = get_unmapped_area(NULL, vdso_base(), size, 0, 0); in arch_setup_additional_pages()
|
| /arch/mips/kernel/ |
| A D | vdso.c | 60 static unsigned long vdso_base(void) in vdso_base() function 119 base = get_unmapped_area(NULL, vdso_base(), size, 0, 0); in arch_setup_additional_pages()
|
| /arch/s390/kernel/ |
| A D | vdso.c | 32 current->mm->context.vdso_base = vma->vm_start; in vdso_mremap() 90 current->mm->context.vdso_base = vdso_text_start; in map_vdso()
|
| A D | stacktrace.c | 98 return in_range(ip, current->mm->context.vdso_base, vdso_text_size()); in ip_within_vdso()
|