| /linux/arch/x86/lib/ |
| A D | memmove_64.S | 37 add %rdx, %r8 52 cmp $680, %rdx 61 sub $0x20, %rdx 66 sub $0x20, %rdx 86 movq %rdx, %rcx 119 cmp $680, %rdx 154 cmpq $16, %rdx 170 cmpq $8, %rdx 181 cmpq $4, %rdx 192 cmp $2, %rdx [all …]
|
| A D | memcpy_64.S | 37 movq %rdx, %rcx 49 cmpq $0x20, %rdx 58 subq $0x20, %rdx 60 subq $0x20, %rdx 84 addq %rdx, %rsi 85 addq %rdx, %rdi 86 subq $0x20, %rdx 93 subq $0x20, %rdx 110 subq %rdx, %rsi 111 subq %rdx, %rdi [all …]
|
| A D | hweight.S | 46 pushq %rdx 48 movq %rdi, %rdx # w -> t 50 shrq %rdx # t >>= 1 52 movabsq $0x3333333333333333, %rdx 59 addq %rdx, %rax # w = w_tmp + t 61 movq %rax, %rdx # w -> t 62 shrq $4, %rdx # t >>= 4 63 addq %rdx, %rax # w_tmp += t 64 movabsq $0x0f0f0f0f0f0f0f0f, %rdx 66 movabsq $0x0101010101010101, %rdx [all …]
|
| A D | getuser.S | 42 movq $0x0123456789abcdef,%rdx 47 cmp %rax, %rdx 48 sbb %rdx, %rdx 49 or %rdx, %rax 102 UACCESS movq (%_ASM_AX),%rdx 148 UACCESS movq (%_ASM_AX),%rdx
|
| A D | memset_64.S | 36 movq %rdx,%rcx 60 movq %rdx,%rcx 108 cmpq $7,%rdx 114 subq %r8,%rdx
|
| A D | copy_page_64.S | 35 movq 0x8*2(%rsi), %rdx 46 movq %rdx, 0x8*2(%rdi) 65 movq 0x8*2(%rsi), %rdx 74 movq %rdx, 0x8*2(%rdi)
|
| /linux/tools/arch/x86/lib/ |
| A D | memcpy_64.S | 36 movq %rdx, %rcx 48 cmpq $0x20, %rdx 57 subq $0x20, %rdx 59 subq $0x20, %rdx 83 addq %rdx, %rsi 84 addq %rdx, %rdi 85 subq $0x20, %rdx 92 subq $0x20, %rdx 109 subq %rdx, %rsi 110 subq %rdx, %rdi [all …]
|
| A D | memset_64.S | 36 movq %rdx,%rcx 60 movq %rdx,%rcx 108 cmpq $7,%rdx 114 subq %r8,%rdx
|
| /linux/arch/x86/crypto/ |
| A D | sm4-aesni-avx-asm_64.S | 157 vmovdqu 0*16(%rdx), RA0; 163 vmovdqu 1*16(%rdx), RA1; 165 vmovdqu 2*16(%rdx), RA2; 168 vmovdqu 3*16(%rdx), RA3; 376 vmovdqu (0 * 16)(%rdx), RA0; 377 vmovdqu (1 * 16)(%rdx), RA1; 378 vmovdqu (2 * 16)(%rdx), RA2; 379 vmovdqu (3 * 16)(%rdx), RA3; 380 vmovdqu (4 * 16)(%rdx), RB0; 385 vmovdqu (5 * 16)(%rdx), RB1; [all …]
|
| A D | sm4-aesni-avx2-asm_64.S | 367 vpxor (0 * 32)(%rdx), RA0, RA0; 368 vpxor (1 * 32)(%rdx), RA1, RA1; 369 vpxor (2 * 32)(%rdx), RA2, RA2; 405 vmovdqu (0 * 32)(%rdx), RA0; 406 vmovdqu (1 * 32)(%rdx), RA1; 407 vmovdqu (2 * 32)(%rdx), RA2; 408 vmovdqu (3 * 32)(%rdx), RA3; 409 vmovdqu (4 * 32)(%rdx), RB0; 410 vmovdqu (5 * 32)(%rdx), RB1; 411 vmovdqu (6 * 32)(%rdx), RB2; [all …]
|
| A D | cast5-avx-x86_64-asm_64.S | 63 #define RGI1 %rdx 451 movq %rdx, %r12; 453 vmovdqu (0*16)(%rdx), RL1; 454 vmovdqu (1*16)(%rdx), RR1; 455 vmovdqu (2*16)(%rdx), RL2; 456 vmovdqu (3*16)(%rdx), RR2; 457 vmovdqu (4*16)(%rdx), RL3; 458 vmovdqu (5*16)(%rdx), RR3; 459 vmovdqu (6*16)(%rdx), RL4; 460 vmovdqu (7*16)(%rdx), RR4; [all …]
|
| A D | chacha-avx512vl-x86_64.S | 30 # %rdx: up to 2 data blocks input, i 117 vpxord 0x00(%rdx),%xmm7,%xmm6 124 vpxord 0x10(%rdx),%xmm7,%xmm6 131 vpxord 0x20(%rdx),%xmm7,%xmm6 138 vpxord 0x30(%rdx),%xmm7,%xmm6 146 vpxord 0x40(%rdx),%xmm7,%xmm6 152 vpxord 0x50(%rdx),%xmm7,%xmm6 158 vpxord 0x60(%rdx),%xmm7,%xmm6 164 vpxord 0x70(%rdx),%xmm7,%xmm6 329 vpxord 0x00(%rdx),%xmm10,%xmm9 [all …]
|
| A D | chacha-avx2-x86_64.S | 144 vpxor 0x00(%rdx),%xmm7,%xmm6 151 vpxor 0x10(%rdx),%xmm7,%xmm6 158 vpxor 0x20(%rdx),%xmm7,%xmm6 165 vpxor 0x30(%rdx),%xmm7,%xmm6 173 vpxor 0x40(%rdx),%xmm7,%xmm6 179 vpxor 0x50(%rdx),%xmm7,%xmm6 185 vpxor 0x60(%rdx),%xmm7,%xmm6 191 vpxor 0x70(%rdx),%xmm7,%xmm6 211 lea (%rdx,%rax),%rsi 516 lea (%rdx,%rax),%rsi [all …]
|
| A D | chacha-ssse3-x86_64.S | 139 movdqu 0x00(%rdx),%xmm4 147 movdqu 0x10(%rdx),%xmm0 155 movdqu 0x20(%rdx),%xmm0 163 movdqu 0x30(%rdx),%xmm0 184 lea (%rdx,%rax),%rsi 654 movdqu 0x00(%rdx),%xmm1 661 movdqu 0x10(%rdx),%xmm1 668 movdqu 0x20(%rdx),%xmm1 675 movdqu 0x30(%rdx),%xmm1 682 movdqu 0x40(%rdx),%xmm1 [all …]
|
| /linux/arch/x86/mm/ |
| A D | mem_encrypt_boot.S | 44 movq %rdx, %r12 /* Area length */ 55 movq %r8, %rdx /* Pagetables used for encryption */ 99 mov %rdx, %cr3 102 mov %cr4, %rdx 103 andq $~X86_CR4_PGE, %rdx 104 mov %rdx, %cr4 105 orq $X86_CR4_PGE, %rdx 106 mov %rdx, %cr4 118 mov %rdx, %r15 /* Save original PAT value */ 151 mov %r15, %rdx /* Restore original PAT value */
|
| /linux/arch/x86/power/ |
| A D | hibernate_asm_64.S | 34 movq %rax, %rdx 35 andq $~(X86_CR4_PGE), %rdx 36 movq %rdx, %cr4; # turn off PGE 49 movq pt_regs_dx(%rax), %rdx 80 movq %rdx, pt_regs_dx(%rax) 112 movq restore_pblist(%rip), %rdx 132 testq %rdx, %rdx 136 movq pbe_address(%rdx), %rsi 137 movq pbe_orig_address(%rdx), %rdi 143 movq pbe_next(%rdx), %rdx
|
| /linux/arch/x86/entry/ |
| A D | calling.h | 68 .macro PUSH_REGS rdx=%rdx rcx=%rcx rax=%rax save_ret=0 unwind_hint=1 77 pushq \rdx /* pt_regs->dx */ 126 PUSH_REGS rdx=\rdx, rcx=\rcx, rax=\rax, save_ret=\save_ret unwind_hint=\unwind_hint 143 popq %rdx 311 shl $32, %rdx 312 or %rdx, %rax 323 shr $32, %rdx 339 mov \save_reg, %rdx 346 shr $32, %rdx 439 pushq %rdx [all …]
|
| A D | entry_64.S | 1150 pushq %rdx 1169 movq %rsp, %rdx 1179 PUSH_AND_CLEAR_REGS rdx=(%rdx) 1254 cmpq 8(%rsp), %rdx 1257 cmpq 8(%rsp), %rdx 1281 lea 6*8(%rsp), %rdx 1283 cmpq %rdx, 4*8(%rsp) 1288 cmpq %rdx, 4*8(%rsp) 1307 pushq %rdx 1316 popq %rdx [all …]
|
| /linux/arch/x86/kernel/ |
| A D | sev_verify_cbit.S | 38 movq %rsi, %rdx 39 andq $(~X86_CR4_PGE), %rdx 40 movq %rdx, %cr4 49 1: rdrand %rdx 53 movq %rdx, sev_check_data(%rip) 65 cmpq %rdx, sev_check_data(%rip)
|
| A D | ftrace_64.S | 85 movq %rdx, RDX(%rsp) 96 movq MCOUNT_REG_SIZE-8(%rsp), %rdx 98 movq %rbp, %rdx 100 movq %rdx, RBP(%rsp) 126 movq RDX(%rsp), %rdx 167 movq function_trace_op(%rip), %rdx 212 movq function_trace_op(%rip), %rdx 355 movq %rdx, 8(%rsp) 362 movq 8(%rsp), %rdx
|
| A D | relocate_kernel_64.S | 122 pushq %rdx 214 popq %rdx 217 call *%rdx 293 movq %rdi, %rdx /* Save destination page to %rdx */ 303 movq %rdx, %rsi 308 movq %rdx, %rdi
|
| A D | head_64.S | 66 leaq INIT_PER_CPU_VAR(fixed_percpu_data)(%rip), %rdx 68 shrq $32, %rdx 304 movq __per_cpu_offset(,%rcx,8), %rdx 315 movq pcpu_hot + X86_current_task(%rdx), %rax 336 leaq gdt_page(%rdx), %rax 364 leaq INIT_PER_CPU_VAR(fixed_percpu_data)(%rip), %rdx 367 shrq $32, %rdx 517 pushq %rdx /* pt_regs->dx */
|
| /linux/arch/x86/kernel/acpi/ |
| A D | madt_playdead.S | 18 movq %cr4, %rdx 19 andq $~(X86_CR4_PGE), %rdx 20 movq %rdx, %cr4
|
| /linux/arch/x86/kvm/svm/ |
| A D | vmenter.S | 321 mov %rbp, SEV_ES_RBP (%rdx) 322 mov %r15, SEV_ES_R15 (%rdx) 323 mov %r14, SEV_ES_R14 (%rdx) 324 mov %r13, SEV_ES_R13 (%rdx) 325 mov %r12, SEV_ES_R12 (%rdx) 326 mov %rbx, SEV_ES_RBX (%rdx) 332 mov %rdi, SEV_ES_RDI (%rdx) 333 mov %rsi, SEV_ES_RSI (%rdx)
|
| /linux/tools/testing/selftests/kvm/x86_64/ |
| A D | xen_vmcall_test.c | 34 unsigned long rdx = ARGVALUE(3); in guest_code() local 43 "a"(rax), "D"(rdi), "S"(rsi), "d"(rdx), in guest_code() 65 "a"(rax), "D"(rdi), "S"(rsi), "d"(rdx), in guest_code() 72 rdx = 0x5a5a5a5a; /* ingpa (badly aligned) */ in guest_code() 75 "a"(rax), "c"(rcx), "d"(rdx), in guest_code()
|