Lines Matching refs:is_user
28 static inline bool get_mmap_lock_carefully(struct mm_struct *mm, bool is_user) in get_mmap_lock_carefully() argument
33 if (!is_user) in get_mmap_lock_carefully()
52 static inline bool upgrade_mmap_lock_carefully(struct mm_struct *mm, bool is_user) in upgrade_mmap_lock_carefully() argument
55 if (!is_user) in upgrade_mmap_lock_carefully()
82 unsigned long addr, bool is_user) in um_lock_mm_and_find_vma() argument
86 if (!get_mmap_lock_carefully(mm, is_user)) in um_lock_mm_and_find_vma()
112 if (!upgrade_mmap_lock_carefully(mm, is_user)) in um_lock_mm_and_find_vma()
141 int is_write, int is_user, int *code_out) in handle_page_fault() argument
159 if (is_user) in handle_page_fault()
162 vma = um_lock_mm_and_find_vma(mm, address, is_user); in handle_page_fault()
233 if (!is_user) in handle_page_fault()
308 unsigned long segv(struct faultinfo fi, unsigned long ip, int is_user, in segv() argument
316 if (!is_user && regs) in segv()
319 if (!is_user && init_mm.context.sync_tlb_range_to) { in segv()
349 else if (!is_user && address > PAGE_SIZE && address < TASK_SIZE) { in segv()
356 err = handle_page_fault(address, ip, is_write, is_user, in segv()
370 else if (!is_user && arch_fixup(ip, regs)) in segv()
373 if (!is_user) { in segv()