Lines Matching refs:esp

71 	testl	$USER_SEGMENT_RPL_MASK, PT_CS(%esp)
111 andl $0x0000ffff, 4*4(%esp)
114 testl $X86_EFLAGS_VM, 5*4(%esp)
117 testl $USER_SEGMENT_RPL_MASK, 4*4(%esp)
120 orl $CS_FROM_KERNEL, 4*4(%esp)
155 pushl %esp # sp (points at ss)
156 addl $7*4, (%esp) # point sp back at the previous context
157 pushl 7*4(%esp) # flags
158 pushl 7*4(%esp) # cs
159 pushl 7*4(%esp) # ip
160 pushl 7*4(%esp) # orig_eax
161 pushl 7*4(%esp) # gs / function
162 pushl 7*4(%esp) # fs
174 testl $CS_FROM_KERNEL, 1*4(%esp)
184 movl 5*4(%esp), %eax # (modified) regs->sp
186 movl 4*4(%esp), %ecx # flags
189 movl 3*4(%esp), %ecx # cs
193 movl 2*4(%esp), %ecx # ip
196 movl 1*4(%esp), %ecx # eax
200 lea -4*4(%eax), %esp
271 4: addl $(4 + \pop), %esp /* pop the unused "gs" slot */
312 movl PT_EFLAGS(%esp), %eax # mix EFLAGS, SS and CS
318 movb PT_OLDSS(%esp), %ah
319 movb PT_CS(%esp), %al
335 mov %esp, %edx /* load kernel esp */
336 mov PT_OLDESP(%esp), %eax /* load userspace esp */
350 lss (%esp), %esp /* switch to espfix segment */
380 subl %esp, %ecx /* ecx = (end of entry_stack) - esp */
385 movl %esp, %esi
397 movl PT_EFLAGS(%esp), %ecx # mix EFLAGS and CS
398 movb PT_CS(%esp), %cl
401 movl PT_CS(%esp), %ecx
427 movl %edi, %esp
480 orl $CS_FROM_ENTRY_STACK, PT_CS(%esp)
488 orl $CS_FROM_USER_CR3, PT_CS(%esp)
520 testl $(X86_EFLAGS_VM), PT_EFLAGS(%esp)
532 movl %esp, %esi
547 movl %ebx, %esp
566 testl $CS_FROM_ENTRY_STACK, PT_CS(%esp)
572 andl $(~CS_FROM_ENTRY_STACK), PT_CS(%esp)
575 movl %esp, %esi
598 movl %ebx, %esp
604 testl $CS_FROM_USER_CR3, PT_CS(%esp)
608 andl $(~CS_FROM_USER_CR3), PT_CS(%esp)
644 movl %esp, %eax
645 movl PT_ORIG_EAX(%esp), %edx /* get the vector from stack */
646 movl $-1, PT_ORIG_EAX(%esp) /* no syscall to restart */
689 movl %esp, TASK_threadsp(%eax)
690 movl TASK_threadsp(%edx), %esp
727 movl %esp, %edx /* regs */
737 addl $4, %esp
800 movl TSS_entry2task_stack(%esp), %esp
830 testl $X86_EFLAGS_NT|X86_EFLAGS_AC|X86_EFLAGS_TF, PT_EFLAGS(%esp)
834 movl %esp, %eax
853 movl PT_EFLAGS(%esp), %edi
854 movl PT_EAX(%esp), %esi
859 movl PT_EIP(%esp), %edx /* pt_regs->ip */
860 movl PT_OLDESP(%esp), %ecx /* pt_regs->sp */
861 1: mov PT_FS(%esp), %fs
864 addl $2*4, %esp /* skip pt_regs->cx and pt_regs->dx */
870 movl %eax, %esp
882 btrl $X86_EFLAGS_IF_BIT, (%esp)
894 2: movl $0, PT_FS(%esp)
939 movl %esp, %eax
1000 subl $2*4, %esp
1001 sgdt (%esp)
1002 movl 2(%esp), %ecx /* GDT address */
1010 addl $2*4, %esp
1012 addl %esp, %eax /* the adjusted stack pointer */
1015 lss (%esp), %esp /* switch to the normal stack segment */
1037 movl PT_GS(%esp), %edi # get the function address
1040 movl PT_ORIG_EAX(%esp), %edx # get the error code
1041 movl $-1, PT_ORIG_EAX(%esp) # no syscall to restart
1043 movl %esp, %eax # pt_regs pointer
1048 movl PT_EFLAGS(%esp), %eax # mix EFLAGS and CS
1049 movb PT_CS(%esp), %al
1055 movl PT_CS(%esp), %eax
1067 movl %esp, %eax
1137 movl %esp, %eax # pt_regs pointer
1155 movl %esp, %ebx
1156 movl PER_CPU_VAR(cpu_current_top_of_stack), %esp
1158 movl %ebx, %esp
1162 testl $CS_FROM_ESPFIX, PT_CS(%esp)
1177 pushl %esp
1178 addl $4, (%esp)
1181 pushl 4*4(%esp) # flags
1182 pushl 4*4(%esp) # cs
1183 pushl 4*4(%esp) # ip
1191 xorl $(CS_FROM_ESPFIX | CS_FROM_KERNEL), PT_CS(%esp)
1194 movl %esp, %eax # pt_regs pointer
1209 lss (1+5+6)*4(%esp), %esp # back to espfix stack
1221 leal -TOP_OF_KERNEL_STACK_PADDING-PTREGS_SIZE(%esi), %esp