| /arch/x86/include/asm/ |
| A D | pgtable_32_areas.h | 48 #define MODULES_END VMALLOC_END macro 49 #define MODULES_LEN (MODULES_VADDR - MODULES_END)
|
| A D | pgtable_64_types.h | 171 # define MODULES_END _AC(0xffffffffff000000, UL) macro 173 # define MODULES_END _AC(0xfffffffffe000000, UL) macro 175 #define MODULES_LEN (MODULES_END - MODULES_VADDR)
|
| /arch/arm/include/asm/ |
| A D | memory.h | 74 #define MODULES_END (PAGE_OFFSET - PMD_SIZE) macro 76 #define MODULES_END (PAGE_OFFSET) macro 126 #define MODULES_END (END_MEM) macro
|
| /arch/powerpc/mm/ |
| A D | mem.c | 383 for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE) in prealloc_execmem_pgtable() 407 if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) { in execmem_arch_setup() 410 fallback_end = MODULES_END; in execmem_arch_setup() 415 end = MODULES_END; in execmem_arch_setup()
|
| /arch/sparc/mm/ |
| A D | execmem.c | 13 .end = MODULES_END, in execmem_arch_setup()
|
| /arch/powerpc/include/asm/nohash/32/ |
| A D | mmu-8xx.h | 173 #define MODULES_END PAGE_OFFSET macro 175 #define MODULES_VADDR (MODULES_END - MODULES_SIZE)
|
| /arch/riscv/kernel/ |
| A D | vmcore_info.c | 18 vmcoreinfo_append_str("NUMBER(MODULES_END)=0x%lx\n", MODULES_END); in arch_crash_save_vmcoreinfo()
|
| /arch/arm64/kernel/ |
| A D | vmcore_info.c | 25 vmcoreinfo_append_str("NUMBER(MODULES_END)=0x%lx\n", MODULES_END); in arch_crash_save_vmcoreinfo()
|
| /arch/x86/kernel/ |
| A D | head64.c | 231 MAYBE_BUILD_BUG_ON(!(((MODULES_END - 1) & PGDIR_MASK) == in x86_64_start_kernel() 233 BUILD_BUG_ON(__fix_to_virt(__end_of_fixed_addresses) <= MODULES_END); in x86_64_start_kernel()
|
| /arch/s390/boot/ |
| A D | startup.c | 36 unsigned long __bootdata_preserved(MODULES_END); 400 MODULES_END = round_down(kernel_start, _SEGMENT_SIZE); in setup_kernel_memory_layout() 401 MODULES_VADDR = MODULES_END - MODULES_LEN; in setup_kernel_memory_layout() 405 boot_debug("modules area: 0x%016lx-0x%016lx\n", MODULES_VADDR, MODULES_END); in setup_kernel_memory_layout()
|
| /arch/x86/mm/ |
| A D | init.c | 1097 .end = MODULES_END, in execmem_arch_setup() 1104 .end = MODULES_END, in execmem_arch_setup() 1111 .end = MODULES_END, in execmem_arch_setup() 1118 .end = MODULES_END, in execmem_arch_setup() 1125 .end = MODULES_END, in execmem_arch_setup()
|
| /arch/arm64/include/asm/ |
| A D | memory.h | 46 #define KIMAGE_VADDR (MODULES_END) 47 #define MODULES_END (MODULES_VADDR + MODULES_VSIZE) macro
|
| /arch/arm/mm/ |
| A D | pageattr.c | 66 if (!range_in_range(start, size, MODULES_VADDR, MODULES_END) && in change_memory_common()
|
| A D | kasan_init.c | 285 create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END)); in kasan_init()
|
| /arch/nios2/mm/ |
| A D | init.c | 147 .end = MODULES_END, in execmem_arch_setup()
|
| /arch/powerpc/include/asm/book3s/32/ |
| A D | pgtable.h | 198 #define MODULES_END ALIGN_DOWN(PAGE_OFFSET, SZ_256M) macro 200 #define MODULES_VADDR (MODULES_END - MODULES_SIZE)
|
| /arch/powerpc/mm/kasan/ |
| A D | init_32.c | 121 k_end = (unsigned long)kasan_mem_to_shadow((void *)MODULES_END); in kasan_unmap_early_shadow_vmalloc()
|
| /arch/loongarch/mm/ |
| A D | init.c | 238 .end = MODULES_END, in execmem_arch_setup()
|
| /arch/s390/mm/ |
| A D | init.c | 310 .end = MODULES_END, in execmem_arch_setup()
|
| /arch/mips/mm/ |
| A D | fault.c | 87 if (unlikely(address >= MODULES_VADDR && address < MODULES_END)) in __do_page_fault()
|
| /arch/mips/include/asm/ |
| A D | pgtable-64.h | 151 #define MODULES_END (FIXADDR_START-2*PAGE_SIZE) macro
|
| /arch/loongarch/include/asm/ |
| A D | pgtable.h | 83 #define MODULES_END (MODULES_VADDR + SZ_256M) macro 91 #define VMALLOC_START MODULES_END
|
| /arch/riscv/mm/ |
| A D | kasan_init.c | 504 (void *)kasan_mem_to_shadow((void *)MODULES_END)); in kasan_init() 522 kasan_populate(kasan_mem_to_shadow((const void *)MODULES_END), in kasan_init()
|
| /arch/nios2/include/asm/ |
| A D | pgtable.h | 31 #define MODULES_END (CONFIG_NIOS2_KERNEL_REGION_BASE - 1) macro
|
| /arch/riscv/include/asm/ |
| A D | pgtable.h | 50 #define BPF_JIT_REGION_END (MODULES_END) 61 #define MODULES_END (PFN_ALIGN((unsigned long)&_start)) macro 64 #define MODULES_END VMALLOC_END macro
|