Home
last modified time | relevance | path

Searched refs:tmp1 (Results 1 – 25 of 62) sorted by relevance

123

/arch/arm/mach-at91/
A Dpm_suspend.S138 bic tmp1, tmp1, #0x1
142 bic tmp1, tmp1, #0x1
146 bic tmp1, tmp1, #0x1
150 bic tmp1, tmp1, #0x1
154 bic tmp1, tmp1, #0x1
310 orr tmp1, tmp1, #0x1
314 orr tmp1, tmp1, #0x1
318 orr tmp1, tmp1, #0x1
322 orr tmp1, tmp1, #0x1
326 orr tmp1, tmp1, #0x1
[all …]
/arch/arm64/include/asm/
A Dasm_pointer_auth.h13 mov \tmp1, #THREAD_KEYS_KERNEL
14 add \tmp1, \tsk, \tmp1
15 ldp \tmp2, \tmp3, [\tmp1, #PTRAUTH_KERNEL_KEY_APIA]
53 mov \tmp1, #THREAD_KEYS_USER
54 add \tmp1, \tsk, \tmp1
55 ldp \tmp2, \tmp3, [\tmp1, #PTRAUTH_USER_KEY_APIA]
61 mrs \tmp1, id_aa64isar1_el1
62 ubfx \tmp1, \tmp1, #ID_AA64ISAR1_EL1_APA_SHIFT, #8
65 orr \tmp1, \tmp1, \tmp2
66 cbz \tmp1, .Lno_addr_auth\@
[all …]
A Dasm-uaccess.h16 .macro __uaccess_ttbr0_disable, tmp1
17 mrs \tmp1, ttbr1_el1 // swapper_pg_dir
18 bic \tmp1, \tmp1, #TTBR_ASID_MASK
19 sub \tmp1, \tmp1, #RESERVED_SWAPPER_OFFSET // reserved_pg_dir
21 add \tmp1, \tmp1, #RESERVED_SWAPPER_OFFSET
26 .macro __uaccess_ttbr0_enable, tmp1, tmp2
27 get_current_task \tmp1
28 ldr \tmp1, [\tmp1, #TSK_TI_TTBR0] // load saved TTBR0_EL1
30 extr \tmp2, \tmp2, \tmp1, #48
40 __uaccess_ttbr0_disable \tmp1
[all …]
A Dassembler.h356 cmp \tmp0, \tmp1
357 csel \tmp0, \tmp1, \tmp0, hi
422 dcache_line_size \tmp1, \tmp2
436 sub \tmp2, \tmp1, #1
440 add \tmp2, \tmp2, \tmp1
455 phys_to_ttbr \tmp1, \pgtbl
456 offset_ttbr1 \tmp1, \tmp2
457 msr ttbr1_el1, \tmp1
629 mrs \tmp1, midr_el1
632 and \tmp1, \tmp1, \tmp2
[all …]
/arch/sparc/include/asm/
A Dhead_64.h37 lduw [%tmp1 + %lo(is_sun4v)], %tmp1; \
42 rdpr %ver, %tmp1; \
44 srlx %tmp1, 32, %tmp1; \
46 cmp %tmp1, %tmp2; \
51 rdpr %ver, %tmp1; \
53 srlx %tmp1, 32, %tmp1; \
55 cmp %tmp1, %tmp2; \
60 rdpr %ver, %tmp1; \
64 sllx %tmp1, 16, %tmp1; \
71 rdpr %ver, %tmp1; \
[all …]
A Dpgtsrmmu.h89 #define WINDOW_FLUSH(tmp1, tmp2) \ argument
90 mov 0, tmp1; \
93 add tmp1, 1, tmp1; \
96 99: subcc tmp1, 1, tmp1; \
/arch/arm/mach-tegra/
A Dsleep.h82 mrc p15, 0, \tmp1, c0, c0, 0
83 ubfx \tmp1, \tmp1, #4, #12
85 cmp \tmp1, \tmp2
89 .macro exit_smp, tmp1, tmp2
91 bic \tmp1, \tmp1, #(1<<6) | (1<<0) @ clear ACTLR.SMP | ACTLR.FW
97 andeq \tmp1, \tmp1, #0xF
98 moveq \tmp1, \tmp1, lsl #2
110 mov32 \tmp1, \base
111 ldr \tmp1, [\tmp1, #APB_MISC_GP_HIDREV]
112 and \tmp1, \tmp1, #0xff00
[all …]
/arch/loongarch/lib/
A Dcsum.c61 __uint128_t tmp1, tmp2, tmp3, tmp4; in do_csum() local
63 tmp1 = *(__uint128_t *)ptr; in do_csum()
72 tmp1 += (tmp1 >> 64) | (tmp1 << 64); in do_csum()
76 tmp1 = ((tmp1 >> 64) << 64) | (tmp2 >> 64); in do_csum()
77 tmp1 += (tmp1 >> 64) | (tmp1 << 64); in do_csum()
80 tmp1 = ((tmp1 >> 64) << 64) | (tmp3 >> 64); in do_csum()
81 tmp1 += (tmp1 >> 64) | (tmp1 << 64); in do_csum()
82 tmp1 = ((tmp1 >> 64) << 64) | sum64; in do_csum()
83 tmp1 += (tmp1 >> 64) | (tmp1 << 64); in do_csum()
84 sum64 = tmp1 >> 64; in do_csum()
/arch/arm64/lib/
A Dcsum.c63 __uint128_t tmp1, tmp2, tmp3, tmp4; in do_csum() local
65 tmp1 = *(__uint128_t *)ptr; in do_csum()
74 tmp1 += (tmp1 >> 64) | (tmp1 << 64); in do_csum()
78 tmp1 = ((tmp1 >> 64) << 64) | (tmp2 >> 64); in do_csum()
79 tmp1 += (tmp1 >> 64) | (tmp1 << 64); in do_csum()
82 tmp1 = ((tmp1 >> 64) << 64) | (tmp3 >> 64); in do_csum()
83 tmp1 += (tmp1 >> 64) | (tmp1 << 64); in do_csum()
84 tmp1 = ((tmp1 >> 64) << 64) | sum64; in do_csum()
85 tmp1 += (tmp1 >> 64) | (tmp1 << 64); in do_csum()
86 sum64 = tmp1 >> 64; in do_csum()
A Dstrlen.S30 #define tmp1 x4 macro
109 clz tmp1, has_nul1
126 orr tmp2, tmp1, tmp3
132 orr tmp2, tmp1, tmp3
163 clz tmp1, has_nul1
195 lsl tmp1, srcin, 3
199 lsr tmp1, tmp4, tmp1 /* Shift (tmp1 & 63). */
202 lsl tmp1, tmp4, tmp1 /* Shift (tmp1 & 63). */
204 orr tmp1, tmp1, REP8_80
205 orn data1, data1, tmp1
[all …]
A Dstrnlen.S38 tmp1 .req x8 label
54 ands tmp1, srcin, #15
74 sub tmp1, data1, zeroones
78 bic has_nul1, tmp1, tmp2
81 orr tmp1, has_nul1, has_nul2
82 ccmp tmp1, #0, #0, pl /* NZCV = 0000 */
106 CPU_BE( sub tmp1, data2, zeroones )
108 CPU_BE( bic has_nul2, tmp1, tmp2 )
135 add tmp3, tmp3, tmp1
138 neg tmp4, tmp1
[all …]
A Dstrncmp.S37 #define tmp1 x8 macro
63 eor tmp1, src1, src2
65 tst tmp1, #7
79 sub tmp1, data1, zeroones
111 add tmp1, limit, 8
138 sub tmp1, tmp3, zeroones
140 bic has_nul, tmp1, tmp2
238 ldp tmp1, tmp2, [src2], #16
246 LS_FW data2, tmp1, offset
247 LS_BK tmp1, tmp2, neg_offset
[all …]
A Dmemcpy.S45 #define tmp1 x14 macro
99 lsr tmp1, count, 1
102 ldrb B_lw, [src, tmp1]
104 strb B_lw, [dstin, tmp1]
148 sub tmp1, dstin, src
149 cbz tmp1, L(copy0)
150 cmp tmp1, count
156 and tmp1, dstin, 15
158 sub src, src, tmp1
202 and tmp1, dstend, 15
[all …]
/arch/x86/lib/
A Dmemmove_32.S25 .set tmp1, %ebx define
71 movl 1*4(src), tmp1
73 movl tmp1, 1*4(dest)
75 movl 3*4(src), tmp1
77 movl tmp1, 3*4(dest)
91 movl tmp0, (tmp1)
98 movl dest, tmp1
104 movl tmp0,(tmp1)
129 movl -2*4(src), tmp1
133 movl -4*4(src), tmp1
[all …]
/arch/powerpc/include/asm/book3s/32/
A Dmmu-hash.h99 .macro update_user_segments_by_4 tmp1 tmp2 tmp3 tmp4
100 uus_addi 1, \tmp2, \tmp1, 0x111
101 uus_addi 2, \tmp3, \tmp1, 0x222
102 uus_addi 3, \tmp4, \tmp1, 0x333
104 uus_mtsr 0, \tmp1
109 uus_addi 4, \tmp1, \tmp1, 0x444
114 uus_mtsr 4, \tmp1
119 uus_addi 8, \tmp1, \tmp1, 0x444
124 uus_mtsr 8, \tmp1
129 uus_addi 12, \tmp1, \tmp1, 0x444
[all …]
/arch/arc/include/asm/
A Duaccess.h143 unsigned long tmp1, tmp2, tmp3, tmp4; in raw_copy_from_user() local
243 "=r"(tmp1), "=r"(tmp2) in raw_copy_from_user()
263 : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1) in raw_copy_from_user()
283 : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1) in raw_copy_from_user()
301 : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1) in raw_copy_from_user()
358 "=r"(tmp1), "=r"(tmp2), "=r"(tmp3), "=r"(tmp4) in raw_copy_from_user()
371 unsigned long tmp1, tmp2, tmp3, tmp4; in raw_copy_to_user() local
466 "=r"(tmp1), "=r"(tmp2) in raw_copy_to_user()
486 : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1) in raw_copy_to_user()
506 : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1) in raw_copy_to_user()
[all …]
/arch/m68k/lib/
A Dchecksum.c40 unsigned long tmp1, tmp2; in csum_partial() local
118 "=&d" (tmp1), "=&d" (tmp2) in csum_partial()
139 unsigned long tmp1, tmp2; in csum_and_copy_from_user() local
261 "=&d" (tmp1), "=d" (tmp2) in csum_and_copy_from_user()
276 unsigned long tmp1, tmp2; in csum_partial_copy_nocheck() local
364 "=&d" (tmp1), "=&d" (tmp2) in csum_partial_copy_nocheck()
/arch/arm/include/asm/
A Dtls.h10 .macro switch_tls_none, base, tp, tpuser, tmp1, tmp2
13 .macro switch_tls_v6k, base, tp, tpuser, tmp1, tmp2
20 .macro switch_tls_v6, base, tp, tpuser, tmp1, tmp2
27 ldr_va \tmp1, elf_hwcap
29 tst \tmp1, #HWCAP_TLS @ hardware TLS available?
37 .L1_\@: switch_tls_v6k \base, \tp, \tpuser, \tmp1, \tmp2
41 .macro switch_tls_software, base, tp, tpuser, tmp1, tmp2
42 mov \tmp1, #0xffff0fff
43 str \tp, [\tmp1, #-15] @ set TLS value at 0xffff0ff0
/arch/x86/crypto/
A Dcurve25519-x86_64.c985 u64 *a = tmp1; in point_add_and_double()
987 u64 *ab = tmp1; in point_add_and_double()
1010 a1 = tmp1; in point_add_and_double()
1011 b1 = tmp1 + (u32)4U; in point_add_and_double()
1012 d = tmp1 + (u32)8U; in point_add_and_double()
1013 c = tmp1 + (u32)12U; in point_add_and_double()
1014 ab1 = tmp1; in point_add_and_double()
1033 u64 *a = tmp1; in point_double()
1037 u64 *ab = tmp1; in point_double()
1069 u64 *tmp1; in montgomery_ladder() local
[all …]
/arch/alpha/lib/
A Ddivide.S58 #define tmp1 $3 macro
110 stq tmp1,24($30)
142 subq modulus,divisor,tmp1
145 cmovne compare,tmp1,modulus
151 ldq tmp1,24($30)
184 stq tmp1,24($30)
191 subq $31,$27,tmp1
194 cmovlt $28,tmp1,$27
195 ldq tmp1,24($30)
A Dev6-divide.S68 #define tmp1 $3 macro
122 stq tmp1,24($30) # L :
173 subq modulus,divisor,tmp1 # E :
182 cmovne compare,tmp1,modulus # E : Latency 2, extra map slot
187 cmovne compare,tmp1,modulus # E : Latency 2, extra map slot
198 ldq tmp1,24($30) # L :
243 stq tmp1,24($30) # L :
251 subq $31,$27,tmp1 # E : U U L L
255 cmovlt $28,tmp1,$27 # E : Latency 2, extra map slot
258 ldq tmp1,24($30) # L :
/arch/xtensa/lib/
A Dumulsidi3.S199 .macro mul_mulsi3_body dst, src1, src2, tmp1, tmp2
201 1: add \tmp1, \src2, \dst
203 movnez \dst, \tmp1, \tmp2
205 do_addx2 \tmp1, \src2, \dst, \tmp1
207 movnez \dst, \tmp1, \tmp2
209 do_addx4 \tmp1, \src2, \dst, \tmp1
211 movnez \dst, \tmp1, \tmp2
213 do_addx8 \tmp1, \src2, \dst, \tmp1
215 movnez \dst, \tmp1, \tmp2
/arch/alpha/kernel/
A Dtraps.c380 long error, tmp1, tmp2, tmp3, tmp4; in do_entUna() local
403 : "=r"(error), "=&r"(tmp1), "=&r"(tmp2) in do_entUna()
407 una_reg(reg) = tmp1|tmp2; in do_entUna()
419 : "=r"(error), "=&r"(tmp1), "=&r"(tmp2) in do_entUna()
423 una_reg(reg) = (int)(tmp1|tmp2); in do_entUna()
439 una_reg(reg) = tmp1|tmp2; in do_entUna()
667 unsigned long tmp1, tmp2, tmp3, tmp4; in do_entUnaUser() local
730 *reg_addr = tmp1|tmp2; in do_entUnaUser()
762 alpha_write_fp_reg(reg, tmp1|tmp2); in do_entUnaUser()
778 *reg_addr = (int)(tmp1|tmp2); in do_entUnaUser()
[all …]
/arch/csky/abiv2/
A Dfpu.c152 unsigned long tmp1, tmp2; in save_to_user_fp() local
157 tmp1 = mfcr("cr<1, 2>"); in save_to_user_fp()
160 user_fp->fcr = tmp1; in save_to_user_fp()
202 : "=a"(tmp1), "=a"(tmp2), "=a"(tmp3), in save_to_user_fp()
214 unsigned long tmp1, tmp2; in restore_from_user_fp() local
219 tmp1 = user_fp->fcr; in restore_from_user_fp()
222 mtcr("cr<1, 2>", tmp1); in restore_from_user_fp()
264 : "=a"(tmp1), "=a"(tmp2), "=a"(tmp3), in restore_from_user_fp()
/arch/loongarch/include/asm/
A Dasmmacro.h68 la.pcrel \tmp1, 1f
69 alsl.d \tmp1, \tmp0, \tmp1, 3
70 jr \tmp1
93 move \tmp1, \tmp0
114 movgr2cf $fcc0, \tmp1
116 movgr2cf $fcc1, \tmp1
118 movgr2cf $fcc2, \tmp1
120 movgr2cf $fcc3, \tmp1
122 movgr2cf $fcc4, \tmp1
124 movgr2cf $fcc5, \tmp1
[all …]

Completed in 34 milliseconds

123