| /arch/x86/kernel/cpu/sgx/ |
| A D | encls.h | 80 #define __encls_ret_1(rax, rcx) \ argument 82 __encls_ret_N(rax, "c"(rcx)); \ 85 #define __encls_ret_2(rax, rbx, rcx) \ argument 87 __encls_ret_N(rax, "b"(rbx), "c"(rcx)); \ 90 #define __encls_ret_3(rax, rbx, rcx, rdx) \ argument 92 __encls_ret_N(rax, "b"(rbx), "c"(rcx), "d"(rdx)); \ 124 #define __encls_2(rax, rbx, rcx) \ argument 127 __encls_N(rax, ign_rbx_out, "b"(rbx), "c"(rcx)); \ 130 #define __encls_1_1(rax, data, rcx) \ argument 133 int ret = __encls_N(rax, rbx_out, "c"(rcx)); \
|
| /arch/x86/power/ |
| A D | hibernate_asm_64.S | 38 movq %cr3, %rcx; # flush TLB 39 movq %rcx, %cr3 49 movq pt_regs_cx(%rax), %rcx 80 movq %rcx, pt_regs_cx(%rax) 118 jmpq *%rcx 127 movq %rbx, %rcx 128 andq $~(X86_CR4_PGE), %rcx 129 movq %rcx, %cr4; # turn off PGE 130 movq %cr3, %rcx; # flush TLB 131 movq %rcx, %cr3; [all …]
|
| /arch/x86/lib/ |
| A D | copy_user_64.S | 36 cmpq $64,%rcx 50 dec %rcx 95 subq %rax,%rcx 98 movq %rcx,%rax 99 shrq $3,%rcx 107 1: leaq (%rax,%rcx,8),%rcx
|
| A D | memset_64.S | 37 movq %rdx,%rcx 54 imulq %rcx,%rax 62 movq %rdx,%rcx 63 shrq $6,%rcx 68 decq %rcx
|
| A D | csum-copy_64.S | 134 movq %rcx, %r10 219 subq $2, %rcx 229 subq $4, %rcx 242 leaq -1(%rcx, %rcx), %rcx 243 rorq $1, %rcx
|
| A D | clear_page_64.S | 69 cmpq $64,%rcx 81 dec %rcx 108 subq $64,%rcx 109 cmpq $64,%rcx
|
| A D | bhi.S | 82 cmovne %r10, %rcx 95 cmovne %r10, %rcx 116 cmovne %r10, %rcx 131 cmovne %r10, %rcx
|
| A D | memmove_64.S | 43 #define MEMMOVE_BYTES movq %rdx, %rcx; rep movsb; RET 87 movq %rdx, %rcx 90 shrq $3, %rcx 101 movq %rdx, %rcx 106 shrq $3, %rcx
|
| /arch/x86/kernel/ |
| A D | ftrace_64.S | 84 movq %rcx, RCX(%rsp) 127 movq RCX(%rsp), %rcx 164 movq %rcx, RSP(%rsp) 172 leaq (%rsp), %rcx 227 movq %rcx, EFLAGS(%rsp) 229 movq $__KERNEL_DS, %rcx 230 movq %rcx, SS(%rsp) 231 movq $__KERNEL_CS, %rcx 232 movq %rcx, CS(%rsp) 235 movq %rcx, RSP(%rsp) [all …]
|
| A D | head_64.S | 117 leaq early_top_pgt(%rip), %rcx 118 addq %rcx, %rax 222 movq %cr4, %rcx 227 movq %rcx, %cr4 233 movq %rcx, %cr4 278 movq apic_mmio_base(%rip), %rcx 279 addq $APIC_ID, %rcx 280 movl (%rcx), %eax 294 cmpl (%rbx,%rcx,4), %eax 437 movq TASK_threadsp(%rcx), %rsp [all …]
|
| A D | sev_verify_cbit.S | 56 movq %cr3, %rcx 68 movq %rcx, %cr3
|
| A D | relocate_kernel_64.S | 115 movq %rcx, %r11 326 movq %rdi, %rcx /* Put the indirection_page in %rcx */ 333 movq (%rbx), %rcx 338 movq %rcx, %rdi 344 movq %rcx, %rbx 354 movq %rcx, %rsi /* For ever source page do a copy */ 473 movq $16, %rcx 543 pushq %rcx 595 popq %rcx
|
| /arch/x86/xen/ |
| A D | xen-asm.S | 39 push %rcx 45 pop %rcx 66 push %rcx 82 pop %rcx 154 pop %rcx 198 pop %rcx 224 push %rcx 276 popq %rcx 296 popq %rcx 318 popq %rcx
|
| A D | xen-head.S | 93 push %rcx 115 lea xen_hypercall_amd(%rip), %rcx 116 cmp %rax, %rcx 124 pop %rcx
|
| /arch/x86/mm/ |
| A D | mem_encrypt_boot.S | 37 movq %rcx, %rax /* Workarea stack page */ 49 movq $(.L__enc_copy_end - __enc_copy), %rcx /* Encryption routine length */ 56 movq %r12, %rcx /* Area length */ 112 movq %rcx, %r9 /* Save area length */ 136 movq %r12, %rcx 141 movq %r12, %rcx
|
| /arch/x86/virt/vmx/tdx/ |
| A D | tdx.c | 126 args.rcx = 0; in try_init_module_global() 1029 .rcx = tdmr->base, in init_tdmr() 1289 switch (args.rcx) { in paddr_is_tdx_private() 1544 .rcx = gpa, in tdh_mem_page_add() 1554 *ext_err1 = args.rcx; in tdh_mem_page_add() 1564 .rcx = gpa | level, in tdh_mem_sept_add() 1595 .rcx = gpa | level, in tdh_mem_page_aug() 1614 .rcx = gpa | level, in tdh_mem_range_block() 1682 .rcx = gpa, in tdh_mr_extend() 1809 *tdx_pt = args.rcx; in tdh_phymem_page_reclaim() [all …]
|
| /arch/x86/entry/vdso/ |
| A D | vsgx.S | 45 mov SGX_ENCLAVE_OFFSET_OF_RUN(%rbp), %rcx 50 cmpq $0, (%rcx, %rbx) 57 mov SGX_ENCLAVE_RUN_TCS(%rcx), %rbx 58 lea .Lasync_exit_pointer(%rip), %rcx 104 mov %rsp, %rcx
|
| /arch/x86/crypto/ |
| A D | camellia-aesni-avx-asm_64.S | 195 %rcx, (%r9)); 727 leaq 8 * 16(%rax), %rcx; 731 %xmm15, %rax, %rcx); 735 %xmm15, %rax, %rcx, 0); 747 %xmm15, %rax, %rcx, 8); 767 vmovdqu 0 * 16(%rcx), %xmm8; 768 vmovdqu 1 * 16(%rcx), %xmm9; 814 leaq 8 * 16(%rax), %rcx; 818 %xmm15, %rax, %rcx); 853 vmovdqu 0 * 16(%rcx), %xmm8; [all …]
|
| A D | camellia-aesni-avx2-asm_64.S | 228 %rcx, (%r9)); 761 leaq 8 * 32(%rax), %rcx; 765 %ymm15, %rax, %rcx); 769 %ymm15, %rax, %rcx, 0); 781 %ymm15, %rax, %rcx, 8); 801 vmovdqu 0 * 32(%rcx), %ymm8; 802 vmovdqu 1 * 32(%rcx), %ymm9; 848 leaq 8 * 32(%rax), %rcx; 852 %ymm15, %rax, %rcx); 887 vmovdqu 0 * 32(%rcx), %ymm8; [all …]
|
| A D | sm4-aesni-avx-asm_64.S | 161 cmpq $2, %rcx; 166 cmpq $3, %rcx; 236 cmpq $2, %rcx; 241 cmpq $3, %rcx; 371 cmpq $5, %rcx; 386 cmpq $7, %rcx; 395 cmpq $6, %rcx; 405 cmpq $7, %rcx; 429 vmovdqu (%rcx), RA0; 462 vmovdqu RTMP1, (%rcx); [all …]
|
| /arch/x86/entry/ |
| A D | calling.h | 68 .macro PUSH_REGS rdx=%rdx rcx=%rcx rax=%rax save_ret=0 unwind_hint=1 80 pushq \rcx /* pt_regs->cx */ 127 .macro PUSH_AND_CLEAR_REGS rdx=%rdx rcx=%rcx rax=%rax save_ret=0 clear_bp=1 unwind_hint=1 128 PUSH_REGS rdx=\rdx, rcx=\rcx, rax=\rax, save_ret=\save_ret unwind_hint=\unwind_hint 144 popq %rcx 443 pushq %rcx 457 popq %rcx
|
| A D | entry_64_compat.S | 206 pushq %rcx /* pt_regs->ip */ 210 PUSH_AND_CLEAR_REGS rcx=%rbp rax=$-ENOSYS 238 movq RIP(%rsp), %rcx /* pt_regs->ip (in rcx) */
|
| /arch/x86/platform/efi/ |
| A D | efi_stub_64.S | 22 mov %rcx, %r8 23 mov %rsi, %rcx
|
| /arch/x86/net/ |
| A D | bpf_timed_may_goto.S | 29 pushq %rcx 47 popq %rcx
|
| /arch/x86/purgatory/ |
| A D | entry64.S | 40 movq rcx(%rip), %rcx 63 rcx: .quad 0x0 label
|