| /arch/x86/include/asm/ |
| A D | futex.h | 15 #define unsafe_atomic_op1(insn, oval, uaddr, oparg, label) \ argument 21 : "=r" (oldval), "=r" (ret), "+m" (*uaddr) \ 29 #define unsafe_atomic_op2(insn, oval, uaddr, oparg, label) \ argument 41 "+m" (*uaddr), "=&r" (tem) \ 49 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument 52 uaddr = masked_user_access_begin(uaddr); in arch_futex_atomic_op_inuser() 53 else if (!user_access_begin(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 62 uaddr, oparg, Efault); in arch_futex_atomic_op_inuser() 90 uaddr = masked_user_access_begin(uaddr); in futex_atomic_cmpxchg_inatomic() 91 else if (!user_access_begin(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/mips/include/asm/ |
| A D | futex.h | 52 "=" GCC_OFF_SMALL_ASM() (*uaddr) \ 53 : "0" (0), GCC_OFF_SMALL_ASM() (*uaddr), "Jr" (oparg), \ 82 "=" GCC_OFF_SMALL_ASM() (*uaddr) \ 83 : "0" (0), GCC_OFF_SMALL_ASM() (*uaddr), "Jr" (oparg), \ 97 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 107 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 111 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 115 ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser() 119 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 132 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument [all …]
|
| /arch/hexagon/include/asm/ |
| A D | futex.h | 13 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 29 : "=&r" (oldval), "=&r" (ret), "+m" (*uaddr) \ 30 : "r" (uaddr), "r" (oparg), "i" (-EFAULT) \ 39 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 44 __futex_atomic_op("%1 = %4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 47 __futex_atomic_op("%1 = add(%0,%4)\n", ret, oldval, uaddr, in arch_futex_atomic_op_inuser() 51 __futex_atomic_op("%1 = or(%0,%4)\n", ret, oldval, uaddr, in arch_futex_atomic_op_inuser() 56 oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 79 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() 98 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/arm/include/asm/ |
| A D | futex.h | 29 prefetchw(uaddr); \ 40 : "r" (uaddr), "r" (oparg), "Ir" (-EFAULT) \ 46 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 53 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() 58 prefetchw(uaddr); in futex_atomic_cmpxchg_inatomic() 70 : "r" (oldval), "r" (newval), "r" (uaddr), "Ir" (-EFAULT) in futex_atomic_cmpxchg_inatomic() 94 : "r" (uaddr), "r" (oparg), "Ir" (-EFAULT) \ 100 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 107 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() 120 : "r" (oldval), "r" (newval), "r" (uaddr), "Ir" (-EFAULT) in futex_atomic_cmpxchg_inatomic() [all …]
|
| A D | tlbflush.h | 418 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in __local_flush_tlb_page() 422 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", uaddr); in __local_flush_tlb_page() 423 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", uaddr); in __local_flush_tlb_page() 424 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", uaddr); in __local_flush_tlb_page() 429 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", uaddr); in __local_flush_tlb_page() 430 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", uaddr); in __local_flush_tlb_page() 431 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", uaddr); in __local_flush_tlb_page() 439 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in local_flush_tlb_page() 444 __local_flush_tlb_page(vma, uaddr); in local_flush_tlb_page() 456 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in __flush_tlb_page() [all …]
|
| /arch/xtensa/include/asm/ |
| A D | futex.h | 24 #define __futex_atomic_op(insn, ret, old, uaddr, arg) \ argument 44 : [addr] "r" (uaddr), [oparg] "r" (arg), \ 68 [mem] "+m" (*(uaddr)) \ 74 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument 79 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 85 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 89 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 93 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 97 ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser() 101 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() [all …]
|
| /arch/s390/include/asm/ |
| A D | futex.h | 14 __futex_atomic_##name(int oparg, int *old, u32 __user *uaddr) \ 19 instrument_copy_from_user_before(old, uaddr, sizeof(*old)); \ 34 [new] "=&d" (new), [uaddr] "+Q" (*uaddr) \ 39 instrument_copy_from_user_after(old, uaddr, sizeof(*old), 0); \ 56 rc = __futex_atomic_set(oparg, &old, uaddr); in arch_futex_atomic_op_inuser() 59 rc = __futex_atomic_add(oparg, &old, uaddr); in arch_futex_atomic_op_inuser() 62 rc = __futex_atomic_or(oparg, &old, uaddr); in arch_futex_atomic_op_inuser() 65 rc = __futex_atomic_and(~oparg, &old, uaddr); in arch_futex_atomic_op_inuser() 68 rc = __futex_atomic_xor(oparg, &old, uaddr); in arch_futex_atomic_op_inuser() 84 instrument_copy_from_user_before(uval, uaddr, sizeof(*uval)); in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/arc/include/asm/ |
| A D | futex.h | 18 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg)\ argument 40 : "r" (uaddr), "r" (oparg), "ir" (-EFAULT) \ 46 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg)\ argument 67 : "r" (uaddr), "r" (oparg), "ir" (-EFAULT) \ 74 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument 78 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 87 __futex_atomic_op("mov %0, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 91 __futex_atomic_op("add %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 94 __futex_atomic_op("or %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 127 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/loongarch/include/asm/ |
| A D | futex.h | 14 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 25 "=ZC" (*uaddr) \ 26 : "0" (0), "ZC" (*uaddr), "Jr" (oparg) \ 39 __futex_atomic_op("move $t0, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 42 __futex_atomic_op("add.w $t0, %1, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 45 __futex_atomic_op("or $t0, %1, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 48 __futex_atomic_op("and $t0, %1, %z5", ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser() 51 __futex_atomic_op("xor $t0, %1, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 71 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() 85 : "+r" (ret), "=&r" (val), "=ZC" (*uaddr) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/openrisc/include/asm/ |
| A D | futex.h | 11 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 27 : "=&r" (oldval), "=&r" (ret), "+m" (*uaddr) \ 38 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 43 __futex_atomic_op("l.or %1,%4,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 46 __futex_atomic_op("l.add %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 49 __futex_atomic_op("l.or %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 52 __futex_atomic_op("l.and %1,%0,%4", ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser() 55 __futex_atomic_op("l.xor %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 68 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 74 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/powerpc/include/asm/ |
| A D | futex.h | 12 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 28 : "b" (uaddr), "i" (-EFAULT), "r" (oparg) \ 32 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument 36 if (!user_access_begin(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 41 __futex_atomic_op("mr %1,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 44 __futex_atomic_op("add %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 47 __futex_atomic_op("or %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 53 __futex_atomic_op("xor %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 66 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 72 if (!user_access_begin(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/microblaze/include/asm/ |
| A D | futex.h | 11 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 28 : "r" (uaddr), "i" (-EFAULT), "r" (oparg) \ 37 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 42 __futex_atomic_op("or %1,%4,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 45 __futex_atomic_op("add %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 48 __futex_atomic_op("or %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 51 __futex_atomic_op("andn %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 54 __futex_atomic_op("xor %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 67 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 73 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/alpha/include/asm/ |
| A D | futex.h | 12 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 26 : "r" (uaddr), "r"(oparg) \ 30 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument 34 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 39 __futex_atomic_op("mov %3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 42 __futex_atomic_op("addl %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 45 __futex_atomic_op("or %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 51 __futex_atomic_op("xor %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 64 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 70 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/sparc/include/asm/ |
| A D | futex_64.h | 9 #define __futex_cas_op(insn, ret, oldval, uaddr, oparg) \ argument 30 : "r" (uaddr), "r" (oparg), "i" (-EFAULT) \ 34 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument 38 if (unlikely((((unsigned long) uaddr) & 0x3UL))) in arch_futex_atomic_op_inuser() 43 __futex_cas_op("mov\t%4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 46 __futex_cas_op("add\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 49 __futex_cas_op("or\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 52 __futex_cas_op("andn\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 55 __futex_cas_op("xor\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 68 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument [all …]
|
| /arch/csky/include/asm/ |
| A D | futex.h | 14 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 34 [u] "+m" (*uaddr), [t] "=&r" (tmp) \ 46 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 52 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 56 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 60 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 64 ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser() 68 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 83 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 89 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/riscv/include/asm/ |
| A D | futex.h | 22 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 30 [u] "+m" (*uaddr) \ 41 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser() 47 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 51 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 55 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 59 ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser() 63 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser() 76 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 83 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/arm64/include/asm/ |
| A D | futex.h | 15 #define __futex_atomic_op(insn, ret, oldval, uaddr, tmp, oparg) \ argument 33 : "=&r" (ret), "=&r" (oldval), "+Q" (*uaddr), "=&r" (tmp), \ 44 u32 __user *uaddr = __uaccess_mask_ptr(_uaddr); in arch_futex_atomic_op_inuser() local 52 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser() 56 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser() 60 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser() 64 ret, oldval, uaddr, tmp, ~oparg); in arch_futex_atomic_op_inuser() 68 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser() 87 u32 __user *uaddr; in futex_atomic_cmpxchg_inatomic() local 92 uaddr = __uaccess_mask_ptr(_uaddr); in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/um/kernel/skas/ |
| A D | uaccess.c | 259 if (!access_ok(uaddr, sizeof(*uaddr))) in arch_futex_atomic_op_inuser() 276 oldval = *uaddr; in arch_futex_atomic_op_inuser() 282 *uaddr = oparg; in arch_futex_atomic_op_inuser() 285 *uaddr += oparg; in arch_futex_atomic_op_inuser() 288 *uaddr |= oparg; in arch_futex_atomic_op_inuser() 291 *uaddr &= ~oparg; in arch_futex_atomic_op_inuser() 294 *uaddr ^= oparg; in arch_futex_atomic_op_inuser() 337 if (!access_ok(uaddr, sizeof(*uaddr))) in futex_atomic_cmpxchg_inatomic() 348 uaddr = page_address(page) + (((unsigned long) uaddr) & ~PAGE_MASK); in futex_atomic_cmpxchg_inatomic() 350 uaddr = kmap_atomic(page) + ((unsigned long) uaddr & ~PAGE_MASK); in futex_atomic_cmpxchg_inatomic() [all …]
|
| /arch/parisc/include/asm/ |
| A D | futex.h | 35 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument 38 unsigned long ua = (unsigned long)uaddr; in arch_futex_atomic_op_inuser() 48 if (unlikely(get_user(oldval, uaddr) != 0)) { in arch_futex_atomic_op_inuser() 77 if (unlikely(put_user(tmp, uaddr) != 0)) in arch_futex_atomic_op_inuser() 90 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 94 unsigned long ua = (unsigned long)uaddr; in futex_atomic_cmpxchg_inatomic() 99 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() 111 if (unlikely(get_user(val, uaddr) != 0)) { in futex_atomic_cmpxchg_inatomic() 116 if (val == oldval && unlikely(put_user(newval, uaddr) != 0)) { in futex_atomic_cmpxchg_inatomic()
|
| /arch/sh/include/asm/ |
| A D | futex.h | 20 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument 23 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic() 26 return atomic_futex_op_cmpxchg_inatomic(uval, uaddr, oldval, newval); in futex_atomic_cmpxchg_inatomic() 30 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument 36 ret = get_user(oldval, uaddr); in arch_futex_atomic_op_inuser() 63 ret = futex_atomic_cmpxchg_inatomic(&prev, uaddr, oldval, newval); in arch_futex_atomic_op_inuser()
|
| A D | futex-irq.h | 6 u32 __user *uaddr, in atomic_futex_op_cmpxchg_inatomic() argument 15 ret = get_user(prev, uaddr); in atomic_futex_op_cmpxchg_inatomic() 17 ret = put_user(newval, uaddr); in atomic_futex_op_cmpxchg_inatomic()
|
| /arch/arc/kernel/ |
| A D | process.c | 46 SYSCALL_DEFINE3(arc_usr_cmpxchg, int __user *, uaddr, int, expected, int, new) in SYSCALL_DEFINE3() argument 63 ret = access_ok(uaddr, sizeof(*uaddr)); in SYSCALL_DEFINE3() 70 ret = __get_user(uval, uaddr); in SYSCALL_DEFINE3() 77 ret = __put_user(new, uaddr); in SYSCALL_DEFINE3() 94 ret = fixup_user_fault(current->mm, (unsigned long) uaddr, in SYSCALL_DEFINE3()
|
| /arch/arm64/kvm/vgic/ |
| A D | vgic-kvm-device.c | 95 u64 __user *uaddr = (u64 __user *)attr->addr; in kvm_vgic_addr() local 104 if (get_user(addr, uaddr)) in kvm_vgic_addr() 205 r = put_user(addr, uaddr); in kvm_vgic_addr() 224 if (get_user(val, uaddr)) in vgic_set_common_attr() 304 VGIC_NR_PRIVATE_IRQS, uaddr); in vgic_get_common_attr() 383 if (get_user(val, uaddr)) in vgic_v2_attr_regs_access() 417 ret = put_user(val, uaddr); in vgic_v2_attr_regs_access() 561 if (get_user(val, uaddr)) in vgic_v3_attr_regs_access() 616 ret = put_user(val, uaddr); in vgic_v3_attr_regs_access() 632 u32 __user *uaddr = (u32 __user *)attr->addr; in vgic_v3_set_attr() local [all …]
|
| /arch/arm/mm/ |
| A D | flush.c | 128 void __flush_ptrace_access(struct page *page, unsigned long uaddr, void *kaddr, in __flush_ptrace_access() argument 140 flush_pfn_alias(page_to_pfn(page), uaddr); in __flush_ptrace_access() 149 flush_icache_alias(page_to_pfn(page), uaddr, len); in __flush_ptrace_access() 160 unsigned long uaddr, void *kaddr, unsigned long len) in flush_ptrace_access() argument 167 __flush_ptrace_access(page, uaddr, kaddr, len, flags); in flush_ptrace_access() 170 void flush_uprobe_xol_access(struct page *page, unsigned long uaddr, in flush_uprobe_xol_access() argument 175 __flush_ptrace_access(page, uaddr, kaddr, len, flags); in flush_uprobe_xol_access() 186 unsigned long uaddr, void *dst, const void *src, in copy_to_user_page() argument 193 flush_ptrace_access(vma, page, uaddr, dst, len); in copy_to_user_page()
|
| /arch/riscv/kvm/ |
| A D | aia_device.c | 306 if (copy_from_user(&nr, uaddr, sizeof(nr))) in aia_set_attr() 316 if (copy_from_user(&addr, uaddr, sizeof(addr))) in aia_set_attr() 341 if (copy_from_user(&nr, uaddr, sizeof(nr))) in aia_set_attr() 350 if (copy_from_user(&v, uaddr, sizeof(v))) in aia_set_attr() 373 if (copy_from_user(&nr, uaddr, sizeof(nr))) in aia_get_attr() 382 if (copy_to_user(uaddr, &nr, sizeof(nr))) in aia_get_attr() 401 if (copy_to_user(uaddr, &addr, sizeof(addr))) in aia_get_attr() 406 if (copy_from_user(&nr, uaddr, sizeof(nr))) in aia_get_attr() 415 if (copy_to_user(uaddr, &nr, sizeof(nr))) in aia_get_attr() 420 if (copy_from_user(&v, uaddr, sizeof(v))) in aia_get_attr() [all …]
|