| /arch/sh/mm/ |
| A D | flush-sh4.c | 25 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 26 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 27 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 28 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 29 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 30 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 31 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 32 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 37 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 59 __ocbp(v); v += L1_CACHE_BYTES; in sh4__flush_purge_region() [all …]
|
| A D | cache-sh2a.c | 52 unsigned long v; in sh2a__flush_wback_region() local 70 for (v = begin; v < end; v += L1_CACHE_BYTES) { in sh2a__flush_wback_region() 78 for (v = begin; v < end; v += L1_CACHE_BYTES) in sh2a__flush_wback_region() 79 sh2a_flush_oc_line(v, way); in sh2a__flush_wback_region() 93 unsigned long v; in sh2a__flush_purge_region() local 104 for (v = begin; v < end; v+=L1_CACHE_BYTES) { in sh2a__flush_purge_region() 109 sh2a_flush_oc_line(v, way); in sh2a__flush_purge_region() 123 unsigned long v; in sh2a__flush_invalidate_region() local 139 for (v = begin; v < end; v += L1_CACHE_BYTES) in sh2a__flush_invalidate_region() 154 unsigned long v; in sh2a_flush_icache_range() local [all …]
|
| A D | cache-sh2.c | 20 unsigned long v; in sh2__flush_wback_region() local 26 for (v = begin; v < end; v+=L1_CACHE_BYTES) { in sh2__flush_wback_region() 27 unsigned long addr = CACHE_OC_ADDRESS_ARRAY | (v & 0x00000ff0); in sh2__flush_wback_region() 41 unsigned long v; in sh2__flush_purge_region() local 48 for (v = begin; v < end; v+=L1_CACHE_BYTES) in sh2__flush_purge_region() 49 __raw_writel((v & CACHE_PHYSADDR_MASK), in sh2__flush_purge_region() 50 CACHE_OC_ADDRESS_ARRAY | (v & 0x00000ff0) | 0x00000008); in sh2__flush_purge_region() 72 unsigned long v; in sh2__flush_invalidate_region() 79 for (v = begin; v < end; v+=L1_CACHE_BYTES) in sh2__flush_invalidate_region() 80 __raw_writel((v & CACHE_PHYSADDR_MASK), in sh2__flush_invalidate_region() [all …]
|
| /arch/x86/lib/ |
| A D | atomic64_386_32.S | 41 #undef v 49 #undef v 59 #undef v 67 #undef v 77 #undef v 85 #undef v 98 #undef v 106 #undef v 118 #undef v 126 #undef v [all …]
|
| /arch/x86/include/asm/ |
| A D | atomic64_64.h | 15 return __READ_ONCE((v)->counter); in arch_atomic64_read() 20 __WRITE_ONCE(v->counter, i); in arch_atomic64_set() 26 : "=m" (v->counter) in arch_atomic64_add() 33 : "=m" (v->counter) in arch_atomic64_sub() 46 : "=m" (v->counter) in arch_atomic64_inc() 54 : "=m" (v->counter) in arch_atomic64_dec() 83 #define arch_atomic64_sub_return(i, v) arch_atomic64_add_return(-(i), v) argument 91 #define arch_atomic64_fetch_sub(i, v) arch_atomic64_fetch_add(-(i), v) argument 114 : "+m" (v->counter) in arch_atomic64_and() 132 : "+m" (v->counter) in arch_atomic64_or() [all …]
|
| A D | atomic.h | 23 return __READ_ONCE((v)->counter); in arch_atomic_read() 28 __WRITE_ONCE(v->counter, i); in arch_atomic_set() 34 : "+m" (v->counter) in arch_atomic_add() 41 : "+m" (v->counter) in arch_atomic_sub() 89 #define arch_atomic_sub_return(i, v) arch_atomic_add_return(-(i), v) argument 93 return xadd(&v->counter, i); in arch_atomic_fetch_add() 97 #define arch_atomic_fetch_sub(i, v) arch_atomic_fetch_add(-(i), v) argument 120 : "+m" (v->counter) in arch_atomic_and() 127 int val = arch_atomic_read(v); in arch_atomic_fetch_and() 138 : "+m" (v->counter) in arch_atomic_or() [all …]
|
| A D | atomic64_32.h | 40 return __READ_ONCE(v->counter); in arch_atomic64_read_nonatomic() 139 ASM_OUTPUT("+A" (i), "+c" (v)), in arch_atomic64_add_return() 149 ASM_OUTPUT("+A" (i), "+c" (v)), in arch_atomic64_sub_return() 161 "S" (v), in arch_atomic64_inc_return() 172 "S" (v), in arch_atomic64_dec_return() 198 "S" (v), in arch_atomic64_inc() 207 "S" (v), in arch_atomic64_dec() 218 "S" (v), in arch_atomic64_add_unless() 229 "S" (v), in arch_atomic64_inc_not_zero() 240 "S" (v), in arch_atomic64_dec_if_positive() [all …]
|
| /arch/s390/include/asm/ |
| A D | atomic.h | 20 return __atomic_read(&v->counter); in arch_atomic_read() 26 __atomic_set(&v->counter, i); in arch_atomic_set() 38 return __atomic_add_barrier(i, &v->counter); in arch_atomic_fetch_add() 44 __atomic_add(i, &v->counter); in arch_atomic_add() 50 __atomic_add_const(1, &v->counter); in arch_atomic_inc() 56 __atomic_add_const(-1, &v->counter); in arch_atomic_dec() 85 __atomic_##op(i, &v->counter); \ 107 return arch_xchg(&v->counter, new); in ATOMIC_OPS() 133 __atomic64_set((long *)&v->counter, i); in arch_atomic64_set() 151 __atomic64_add(i, (long *)&v->counter); in arch_atomic64_add() [all …]
|
| /arch/alpha/kernel/ |
| A D | termios.c | 7 struct termio v; in user_termio_to_kernel_termios() local 19 canon = v.c_lflag & ICANON; in user_termio_to_kernel_termios() 20 termios->c_cc[VINTR] = v.c_cc[_VINTR]; in user_termio_to_kernel_termios() 35 struct termio v; in kernel_termios_to_user_termio() local 38 memset(&v, 0, sizeof(struct termio)); in kernel_termios_to_user_termio() 39 v.c_iflag = termios->c_iflag; in kernel_termios_to_user_termio() 40 v.c_oflag = termios->c_oflag; in kernel_termios_to_user_termio() 41 v.c_cflag = termios->c_cflag; in kernel_termios_to_user_termio() 42 v.c_lflag = termios->c_lflag; in kernel_termios_to_user_termio() 43 v.c_line = termios->c_line; in kernel_termios_to_user_termio() [all …]
|
| /arch/mips/include/asm/mach-bcm63xx/ |
| A D | bcm63xx_io.h | 61 #define bcm_writeq(v, a) (*(volatile u64 *) BCM_REGS_VA((a)) = (v)) argument 69 #define bcm_rset_writeb(s, v, o) bcm_writeb((v), \ argument 71 #define bcm_rset_writew(s, v, o) bcm_writew((v), \ argument 73 #define bcm_rset_writel(s, v, o) bcm_writel((v), \ argument 80 #define bcm_perf_writel(v, o) bcm_rset_writel(RSET_PERF, (v), (o)) argument 84 #define bcm_wdt_writel(v, o) bcm_rset_writel(RSET_WDT, (v), (o)) argument 86 #define bcm_gpio_writel(v, o) bcm_rset_writel(RSET_GPIO, (v), (o)) argument 90 #define bcm_mpi_writel(v, o) bcm_rset_writel(RSET_MPI, (v), (o)) argument 94 #define bcm_pcie_writel(v, o) bcm_rset_writel(RSET_PCIE, (v), (o)) argument 98 #define bcm_memc_writel(v, o) bcm_rset_writel(RSET_MEMC, (v), (o)) argument [all …]
|
| /arch/m68k/include/asm/ |
| A D | atomic.h | 19 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 20 #define arch_atomic_set(v, i) WRITE_ONCE(((v)->counter), (i)) argument 92 t = v->counter; \ 93 v->counter c_op i; \ 157 : "=d" (c), "=m" (*v) in arch_atomic_dec_and_test_lt() 158 : "m" (*v)); in arch_atomic_dec_and_test_lt() 178 prev = arch_atomic_read(v); in arch_atomic_cmpxchg() 180 arch_atomic_set(v, new); in arch_atomic_cmpxchg() 192 prev = arch_atomic_read(v); in arch_atomic_xchg() 193 arch_atomic_set(v, new); in arch_atomic_xchg() [all …]
|
| /arch/x86/kernel/ |
| A D | bootflag.c | 29 if (!parity8(v)) in sbf_write() 30 v ^= SBF_PARITY; in sbf_write() 33 sbf_port, v); in sbf_write() 44 u8 v; in sbf_read() local 53 return v; in sbf_read() 60 if (!parity8(v)) in sbf_value_valid() 68 u8 v; in sbf_init() local 73 v = sbf_read(); in sbf_init() 81 v &= ~SBF_DIAG; in sbf_init() 83 v |= SBF_PNPOS; in sbf_init() [all …]
|
| /arch/parisc/include/asm/ |
| A D | atomic.h | 64 v->counter = i; in arch_atomic_set() 69 #define arch_atomic_set_release(v, i) arch_atomic_set((v), (i)) argument 73 return READ_ONCE((v)->counter); in arch_atomic_read() 82 v->counter c_op i; \ 106 ret = v->counter; \ 107 v->counter c_op i; \ 154 v->counter c_op i; \ 178 ret = v->counter; \ 179 v->counter c_op i; \ 222 v->counter = i; in arch_atomic64_set() [all …]
|
| /arch/powerpc/include/asm/ |
| A D | atomic.h | 44 __asm__ __volatile__("stw %1,0(%2)" : "=m"(v->counter) : "r"(i), "b"(&v->counter)); in arch_atomic_set() 182 : "r" (&v->counter) in arch_atomic_dec_if_positive() 210 __asm__ __volatile__("std %1,0(%2)" : "=m"(v->counter) : "r"(i), "b"(&v->counter)); in arch_atomic64_set() 307 : "=&r" (t), "+m" (v->counter) in ATOMIC64_OPS() 308 : "r" (&v->counter) in ATOMIC64_OPS() 323 : "r" (&v->counter) in arch_atomic64_inc_return_relaxed() 339 : "r" (&v->counter) in arch_atomic64_dec() 354 : "r" (&v->counter) in arch_atomic64_dec_return_relaxed() 381 : "r" (&v->counter) in arch_atomic64_dec_if_positive() 443 : "r" (&v->counter) in arch_atomic64_inc_not_zero() [all …]
|
| /arch/arm/include/asm/ |
| A D | atomic.h | 25 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 26 #define arch_atomic_set(v,i) WRITE_ONCE(((v)->counter), (i)) argument 134 prefetchw(&v->counter); in arch_atomic_fetch_add_unless() 180 val = v->counter; \ 216 ret = v->counter; in arch_atomic_cmpxchg() 218 v->counter = new; in arch_atomic_cmpxchg() 267 : "r" (&v->counter), "Qo" (v->counter) in arch_atomic64_read() 277 : "=Qo" (v->counter) in arch_atomic64_set() 289 : "r" (&v->counter), "Qo" (v->counter) in arch_atomic64_read() 299 prefetchw(&v->counter); in arch_atomic64_set() [all …]
|
| /arch/sparc/include/asm/ |
| A D | atomic_32.h | 47 #define arch_atomic_set_release(v, i) arch_atomic_set((v), (i)) argument 49 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 51 #define arch_atomic_add(i, v) ((void)arch_atomic_add_return( (int)(i), (v))) argument 52 #define arch_atomic_sub(i, v) ((void)arch_atomic_add_return(-(int)(i), (v))) argument 54 #define arch_atomic_and(i, v) ((void)arch_atomic_fetch_and((i), (v))) argument 55 #define arch_atomic_or(i, v) ((void)arch_atomic_fetch_or((i), (v))) argument 56 #define arch_atomic_xor(i, v) ((void)arch_atomic_fetch_xor((i), (v))) argument 58 #define arch_atomic_sub_return(i, v) (arch_atomic_add_return(-(int)(i), (v))) argument 59 #define arch_atomic_fetch_sub(i, v) (arch_atomic_fetch_add (-(int)(i), (v))) argument
|
| /arch/mips/kernel/ |
| A D | vpe.c | 65 res = v; in get_vpe() 95 struct vpe *v; in alloc_vpe() local 98 if (v == NULL) in alloc_vpe() 110 return v; in alloc_vpe() 139 kfree(v); in release_vpe() 289 if (v % 4) { in apply_r_mips_26() 365 val += v; in apply_r_mips_lo16() 565 if ((v->__start == 0) || (v->shared_ptr == NULL)) in find_vpe_symbols() 732 (unsigned long)v->load_addr + v->len); in vpe_elfload() 789 v->len = 0; in vpe_open() [all …]
|
| A D | module.c | 35 *location = base + v; in apply_r_mips_32() 39 Elf_Addr v) in apply_r_mips_26() argument 41 if (v % 4) { in apply_r_mips_26() 80 n->value = v; in apply_r_mips_hi16() 133 val += v; in apply_r_mips_lo16() 155 val = v + vallo; in apply_r_mips_lo16() 177 if (v % 4) { in apply_r_mips_pc() 202 Elf_Addr v) in apply_r_mips_pc16() argument 208 Elf_Addr v) in apply_r_mips_pc21() argument 214 Elf_Addr v) in apply_r_mips_pc26() argument [all …]
|
| /arch/powerpc/perf/ |
| A D | isa207-common.h | 95 #define p10_SDAR_MODE(v) (((v) >> p10_SDAR_MODE_SHIFT) & \ argument 153 #define CNST_THRESH_VAL(v) (((v) & EVENT_THRESH_MASK) << 32) argument 156 #define CNST_THRESH_CTL_SEL_VAL(v) (((v) & 0x7ffull) << 32) argument 159 #define p10_CNST_THRESH_CMP_VAL(v) (((v) & 0x7ffull) << 43) argument 162 #define CNST_EBB_VAL(v) (((v) & EVENT_EBB_MASK) << 24) argument 165 #define CNST_IFM_VAL(v) (((v) & EVENT_IFM_MASK) << 25) argument 168 #define CNST_L1_QUAL_VAL(v) (((v) & 3) << 22) argument 171 #define CNST_SAMPLE_VAL(v) (((v) & EVENT_SAMPLE_MASK) << 16) argument 174 #define CNST_CACHE_GROUP_VAL(v) (((v) & 0xffull) << 55) argument 179 #define CNST_L2L3_GROUP_VAL(v) (((v) & 0x1full) << 55) argument [all …]
|
| /arch/arm/mach-omap2/ |
| A D | cm33xx.c | 60 u32 v; in am33xx_cm_rmw_reg_bits() local 64 v |= bits; in am33xx_cm_rmw_reg_bits() 67 return v; in am33xx_cm_rmw_reg_bits() 72 u32 v; in am33xx_cm_read_reg_bits() local 78 return v; in am33xx_cm_read_reg_bits() 94 return v; in _clkctrl_idlest() 107 u32 v; in _is_module_ready() local 126 u32 v; in _clktrctrl_write() local 146 u32 v; in am33xx_cm_is_clkdm_in_hwsup() local 271 u32 v; in am33xx_cm_module_enable() local [all …]
|
| A D | prm33xx.c | 34 u32 v; in am33xx_prm_rmw_reg_bits() local 37 v &= ~mask; in am33xx_prm_rmw_reg_bits() 38 v |= bits; in am33xx_prm_rmw_reg_bits() 41 return v; in am33xx_prm_rmw_reg_bits() 59 u32 v; in am33xx_prm_is_hardreset_asserted() local 65 return v; in am33xx_prm_is_hardreset_asserted() 148 u32 v; in am33xx_pwrdm_read_next_pwrst() local 159 u32 v; in am33xx_pwrdm_read_pwrst() local 200 u32 v; in am33xx_pwrdm_read_logic_pwrst() local 218 v &= m; in am33xx_pwrdm_read_logic_retst() [all …]
|
| /arch/alpha/include/asm/ |
| A D | atomic.h | 29 #define arch_atomic_read(v) READ_ONCE((v)->counter) argument 30 #define arch_atomic64_read(v) READ_ONCE((v)->counter) argument 32 #define arch_atomic_set(v,i) WRITE_ONCE((v)->counter, (i)) argument 33 #define arch_atomic64_set(v,i) WRITE_ONCE((v)->counter, (i)) argument 53 :"=&r" (temp), "=m" (v->counter) \ 54 :"Ir" (i), "m" (v->counter)); \ 71 :"Ir" (i), "m" (v->counter) : "memory"); \ 106 :"=&r" (temp), "=m" (v->counter) \ 107 :"Ir" (i), "m" (v->counter)); \ 242 : [mem] "m"(*v), [a] "rI"(a), [u] "rI"(u) in arch_atomic64_fetch_add_unless() [all …]
|
| /arch/riscv/kernel/ |
| A D | module.c | 82 if (v != (u32)v) { in apply_r_riscv_32_rela() 98 Elf_Addr v) in apply_r_riscv_branch_rela() argument 110 Elf_Addr v) in apply_r_riscv_jal_rela() argument 122 Elf_Addr v) in apply_r_riscv_rvc_branch_rela() argument 153 Elf_Addr v) in apply_r_riscv_pcrel_hi20_rela() argument 191 Elf_Addr v) in apply_r_riscv_hi20_rela() argument 356 Elf_Addr v) in dynamic_linking_not_supported() argument 373 u8 value = v; in apply_r_riscv_sub6_rela() 382 u8 value = v; in apply_r_riscv_set6_rela() 777 Elf_Addr v; in apply_relocate_add() local [all …]
|
| /arch/arm64/include/asm/ |
| A D | atomic_lse.h | 20 : [v] "+Q" (v->counter) \ 31 __lse_atomic_add(-i, v); in ATOMIC_OP() 45 : [v] "+Q" (v->counter), \ 103 return __lse_atomic_andnot(~i, v); 127 : [v] "+Q" (v->counter) \ 138 __lse_atomic64_add(-i, v); in ATOMIC64_OP() 152 : [v] "+Q" (v->counter), \ 241 : [ret] "+&r" (v), [v] "+Q" (v->counter), [tmp] "=&r" (tmp) in __lse_atomic64_dec_if_positive() 245 return (long)v; in __lse_atomic64_dec_if_positive() 257 : [v] "+Q" (*(u##sz *)ptr), \ [all …]
|
| /arch/sparc/lib/ |
| A D | atomic32.c | 38 ret = v->counter; \ 39 v->counter c_op i; \ 53 ret = (v->counter c_op i); \ 76 ret = v->counter; in arch_atomic_xchg() 77 v->counter = new; in arch_atomic_xchg() 89 ret = v->counter; in arch_atomic_cmpxchg() 91 v->counter = new; in arch_atomic_cmpxchg() 104 ret = v->counter; in arch_atomic_fetch_add_unless() 106 v->counter += a; in arch_atomic_fetch_add_unless() 113 void arch_atomic_set(atomic_t *v, int i) in arch_atomic_set() argument [all …]
|