| /arch/powerpc/include/asm/ |
| A D | video.h | 8 unsigned long vm_start, unsigned long vm_end, in pgprot_framebuffer() argument 11 return __phys_mem_access_prot(PHYS_PFN(offset), vm_end - vm_start, prot); in pgprot_framebuffer()
|
| /arch/x86/um/ |
| A D | mem_32.c | 18 gate_vma.vm_end = FIXADDR_USER_END; in gate_vma_init() 49 return (addr >= vma->vm_start) && (addr < vma->vm_end); in in_gate_area()
|
| /arch/s390/mm/ |
| A D | gmap_helpers.c | 90 zap_page_range_single(vma, vmaddr, min(end, vma->vm_end) - vmaddr, NULL); in gmap_helper_discard() 91 vmaddr = vma->vm_end; in gmap_helper_discard() 158 rc = walk_page_range_vma(vma, addr, vma->vm_end, in __gmap_helper_unshare_zeropages()
|
| /arch/x86/include/asm/ |
| A D | video.h | 12 unsigned long vm_start, unsigned long vm_end,
|
| /arch/m68k/include/asm/ |
| A D | video.h | 9 unsigned long vm_start, unsigned long vm_end, in pgprot_framebuffer() argument
|
| /arch/mips/include/asm/ |
| A D | video.h | 7 unsigned long vm_start, unsigned long vm_end, in pgprot_framebuffer() argument
|
| /arch/x86/video/ |
| A D | video-common.c | 17 unsigned long vm_start, unsigned long vm_end, in pgprot_framebuffer() argument
|
| /arch/nios2/kernel/ |
| A D | sys_nios2.c | 46 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) { in sys_cacheflush()
|
| /arch/sparc/include/asm/ |
| A D | video.h | 14 unsigned long vm_start, unsigned long vm_end, in pgprot_framebuffer() argument
|
| /arch/riscv/kvm/ |
| A D | mmu.c | 209 hva_t vm_start, vm_end; in kvm_arch_prepare_memory_region() local 226 vm_end = min(reg_end, vma->vm_end); in kvm_arch_prepare_memory_region() 241 ret = kvm_riscv_mmu_ioremap(kvm, gpa, pa, vm_end - vm_start, in kvm_arch_prepare_memory_region() 246 hva = vm_end; in kvm_arch_prepare_memory_region()
|
| /arch/arc/kernel/ |
| A D | arc_hostlink.c | 23 vma->vm_end - vma->vm_start, in arc_hl_mmap()
|
| /arch/powerpc/mm/book3s32/ |
| A D | tlb.c | 93 hash__flush_range(mp->vm_mm, mp->vm_start, mp->vm_end); in hash__flush_tlb_mm()
|
| /arch/sh/kernel/ |
| A D | sys_sh.c | 74 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) { in sys_cacheflush()
|
| /arch/powerpc/kvm/ |
| A D | book3s_hv_uvmem.c | 416 ret = ksm_madvise(vma, vma->vm_start, vma->vm_end, in kvmppc_memslot_page_merge() 423 start = vma->vm_end; in kvmppc_memslot_page_merge() 424 } while (end > vma->vm_end); in kvmppc_memslot_page_merge() 628 if (!vma || addr >= vma->vm_end) { in kvmppc_uvmem_drop_pages() 812 if (!vma || vma->vm_start > start || vma->vm_end < end) in kvmppc_uv_migrate_mem_slot() 972 if (!vma || vma->vm_start > start || vma->vm_end < end) in kvmppc_h_svm_page_in() 1072 if (!vma || vma->vm_start > start || vma->vm_end < end) in kvmppc_h_svm_page_out()
|
| /arch/xtensa/kernel/ |
| A D | syscall.c | 90 addr = vmm->vm_end; in arch_get_unmapped_area()
|
| /arch/x86/entry/vsyscall/ |
| A D | vsyscall_64.c | 287 .vm_end = VSYSCALL_ADDR + PAGE_SIZE, 311 return (addr >= vma->vm_start) && (addr < vma->vm_end); in in_gate_area()
|
| /arch/um/drivers/ |
| A D | mmapper_kern.c | 57 size = vma->vm_end - vma->vm_start; in mmapper_mmap()
|
| /arch/arm/mm/ |
| A D | flush.c | 276 if (start + nr * PAGE_SIZE > vma->vm_end) in __flush_dcache_aliases() 277 nr = (vma->vm_end - start) / PAGE_SIZE; in __flush_dcache_aliases()
|
| /arch/powerpc/kernel/ |
| A D | proc_powerpc.c | 36 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) in page_map_mmap()
|
| /arch/parisc/mm/ |
| A D | fault.c | 262 vma->vm_start, vma->vm_end); in show_signal_msg() 377 address < vma->vm_start || address >= vma->vm_end) { in do_page_fault()
|
| /arch/powerpc/platforms/book3s/ |
| A D | vas-api.c | 516 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) { in coproc_mmap() 518 (vma->vm_end - vma->vm_start), PAGE_SIZE); in coproc_mmap() 574 vma->vm_end - vma->vm_start, prot); in coproc_mmap()
|
| /arch/arm/kernel/ |
| A D | process.c | 316 gate_vma.vm_end = 0xffff0000 + PAGE_SIZE; in gate_vma_init() 329 return (addr >= gate_vma.vm_start) && (addr < gate_vma.vm_end); in in_gate_area()
|
| /arch/parisc/kernel/ |
| A D | cache.c | 518 if (addr + nr * PAGE_SIZE > vma->vm_end) in flush_dcache_folio() 519 nr = (vma->vm_end - addr) / PAGE_SIZE; in flush_dcache_folio() 722 usize += vma->vm_end - vma->vm_start; in mm_total_size() 751 flush_cache_pages(vma, vma->vm_start, vma->vm_end); in flush_cache_mm()
|
| /arch/powerpc/platforms/powernv/ |
| A D | opal-prd.c | 119 vma->vm_start, vma->vm_end, vma->vm_pgoff, in opal_prd_mmap() 123 size = vma->vm_end - vma->vm_start; in opal_prd_mmap()
|
| /arch/arm64/kvm/ |
| A D | mmu.c | 1016 hva_t vm_start, vm_end; in stage2_unmap_memslot() local 1026 vm_end = min(reg_end, vma->vm_end); in stage2_unmap_memslot() 1030 kvm_stage2_unmap_range(&kvm->arch.mmu, gpa, vm_end - vm_start, true); in stage2_unmap_memslot() 1032 hva = vm_end; in stage2_unmap_memslot() 1412 ALIGN(hva, PUD_SIZE) <= vma->vm_end) in get_vma_page_shift() 1418 ALIGN(hva, PMD_SIZE) <= vma->vm_end) in get_vma_page_shift() 2260 hva = min(reg_end, vma->vm_end); in kvm_arch_prepare_memory_region()
|