| /arch/x86/purgatory/ |
| A D | setup-x86_64.S | 19 lgdt gdt(%rip) 39 SYM_DATA_START_LOCAL(gdt) 44 .word gdt_end - gdt - 1 45 .quad gdt 53 SYM_DATA_END_LABEL(gdt, SYM_L_LOCAL, gdt_end)
|
| A D | entry64.S | 19 lgdt gdt(%rip) 84 SYM_DATA_START_LOCAL(gdt) 90 .word gdt_end - gdt - 1 91 .quad gdt 99 SYM_DATA_END_LABEL(gdt, SYM_L_LOCAL, gdt_end)
|
| /arch/x86/boot/ |
| A D | pm.c | 83 static struct gdt_ptr gdt; in setup_gdt() local 85 gdt.len = sizeof(boot_gdt)-1; in setup_gdt() 86 gdt.ptr = (u32)&boot_gdt + (ds() << 4); in setup_gdt() 88 asm volatile("lgdtl %0" : : "m" (gdt)); in setup_gdt()
|
| /arch/x86/include/asm/ |
| A D | desc.h | 45 struct desc_struct gdt[GDT_ENTRIES]; member 53 return per_cpu(gdt_page, cpu).gdt; in get_cpu_gdt_rw() 59 return this_cpu_ptr(&gdt_page)->gdt; in get_current_gdt_rw() 65 return (struct desc_struct *)&get_cpu_entry_area(cpu)->gdt; in get_cpu_gdt_ro() 155 default: size = sizeof(*gdt); break; in native_write_gdt_entry() 158 memcpy(&gdt[entry], desc, size); in native_write_gdt_entry() 256 struct desc_ptr gdt; in native_load_tr_desc() local 261 native_store_gdt(&gdt); in native_load_tr_desc() 268 if (gdt.address == (unsigned long)fixmap_gdt) { in native_load_tr_desc() 294 struct desc_struct *gdt = get_cpu_gdt_rw(cpu); in native_load_tls() local [all …]
|
| A D | cpu_entry_area.h | 91 char gdt[PAGE_SIZE]; member
|
| /arch/x86/boot/compressed/ |
| A D | head_32.S | 64 leal gdt@GOTOFF(%edx), %eax 129 leal gdt@GOTOFF(%ebx), %eax 172 SYM_DATA_START_LOCAL(gdt) 173 .word gdt_end - gdt - 1 179 SYM_DATA_END_LABEL(gdt, SYM_L_LOCAL, gdt_end)
|
| A D | head_64.S | 106 leal rva(gdt)(%ebp), %eax 433 leaq rva(gdt)(%rbx), %rdx 491 .word gdt_end - gdt - 1 492 .quad gdt - gdt64 495 SYM_DATA_START_LOCAL(gdt) 496 .word gdt_end - gdt - 1 504 SYM_DATA_END_LABEL(gdt, SYM_L_LOCAL, gdt_end)
|
| /arch/x86/xen/ |
| A D | smp_pv.c | 228 struct desc_struct *gdt; in cpu_initialize_context() local 240 gdt = get_cpu_gdt_rw(cpu); in cpu_initialize_context() 258 BUG_ON((unsigned long)gdt & ~PAGE_MASK); in cpu_initialize_context() 260 gdt_mfn = arbitrary_virt_to_mfn(gdt); in cpu_initialize_context() 261 make_lowmem_page_readonly(gdt); in cpu_initialize_context()
|
| A D | enlighten_pv.c | 622 struct desc_struct *gdt; in load_TLS_descriptor() local 631 gdt = get_cpu_gdt_rw(cpu); in load_TLS_descriptor() 632 maddr = arbitrary_virt_to_machine(&gdt[GDT_ENTRY_TLS_MIN+i]); in load_TLS_descriptor()
|
| /arch/x86/hyperv/ |
| A D | hv_vtl.c | 118 struct desc_struct *gdt; in hv_vtl_bringup_vcpu() local 128 gdt = (struct desc_struct *)((void *)(gdt_ptr.address)); in hv_vtl_bringup_vcpu() 129 tss = (struct ldttss_desc *)(gdt + GDT_ENTRY_TSS); in hv_vtl_bringup_vcpu() 130 ldt = (struct ldttss_desc *)(gdt + GDT_ENTRY_LDT); in hv_vtl_bringup_vcpu()
|
| /arch/x86/platform/pvh/ |
| A D | head.S | 76 leal rva(gdt)(%ebp), %eax 228 SYM_DATA_START_LOCAL(gdt) 230 .long gdt_start - gdt 232 SYM_DATA_END(gdt)
|
| /arch/x86/kernel/ |
| A D | apm_32.c | 594 struct desc_struct *gdt; in __apm_bios_call() local 600 gdt = get_cpu_gdt_rw(cpu); in __apm_bios_call() 601 save_desc_40 = gdt[0x40 / 8]; in __apm_bios_call() 602 gdt[0x40 / 8] = bad_bios_desc; in __apm_bios_call() 615 gdt[0x40 / 8] = save_desc_40; in __apm_bios_call() 675 struct desc_struct *gdt; in __apm_bios_call_simple() local 681 gdt = get_cpu_gdt_rw(cpu); in __apm_bios_call_simple() 682 save_desc_40 = gdt[0x40 / 8]; in __apm_bios_call_simple() 695 gdt[0x40 / 8] = save_desc_40; in __apm_bios_call_simple() 2226 struct desc_struct *gdt; in apm_init() local [all …]
|
| A D | head_32.S | 274 movl %eax,%ss # after changing gdt.
|
| /arch/x86/kernel/acpi/ |
| A D | wakeup_32.S | 22 # reload the gdt, as we need the full 32 bit address 31 # and restore the stack ... but you need gdt for this to work
|
| /arch/x86/mm/ |
| A D | fault.c | 477 static void show_ldttss(const struct desc_ptr *gdt, const char *name, u16 index) in show_ldttss() argument 488 if (offset + sizeof(struct ldttss_desc) >= gdt->size) { in show_ldttss() 493 if (copy_from_kernel_nofault(&desc, (void *)(gdt->address + offset), in show_ldttss() 556 struct desc_ptr idt, gdt; in show_fault_oops() local 572 native_store_gdt(&gdt); in show_fault_oops() 575 idt.address, idt.size, gdt.address, gdt.size); in show_fault_oops() 578 show_ldttss(&gdt, "LDTR", ldtr); in show_fault_oops() 581 show_ldttss(&gdt, "TR", tr); in show_fault_oops()
|
| A D | cpu_entry_area.c | 199 cea_set_pte(&cea->gdt, get_cpu_gdt_paddr(cpu), gdt_prot); in setup_cpu_entry_area()
|
| /arch/x86/realmode/rm/ |
| A D | trampoline_64.S | 87 lgdtl tr_gdt # load gdt with whatever is appropriate 262 .short tr_gdt_end - tr_gdt - 1 # gdt limit 271 .short tr_gdt_end - tr_gdt - 1 # gdt limit
|
| A D | trampoline_32.S | 51 lgdtl tr_gdt # load gdt with whatever is appropriate
|
| /arch/x86/boot/startup/ |
| A D | gdt_idt.c | 55 .address = (unsigned long)gp->gdt, in startup_64_setup_gdt_idt()
|
| /arch/x86/include/uapi/asm/ |
| A D | kvm.h | 149 struct kvm_dtable gdt, idt; member 160 struct kvm_dtable gdt, idt; member
|
| /arch/x86/kernel/cpu/ |
| A D | common.c | 202 DEFINE_PER_CPU_PAGE_ALIGNED(struct gdt_page, gdt_page) = { .gdt = {
|
| /arch/x86/events/ |
| A D | core.c | 2840 desc = raw_cpu_ptr(gdt_page.gdt) + idx; in get_segment_base()
|
| /arch/x86/kvm/ |
| A D | x86.c | 11692 sregs->gdt.limit = dt.size; in __get_sregs_common() 11693 sregs->gdt.base = dt.address; in __get_sregs_common() 11894 dt.size = sregs->gdt.limit; in __set_sregs_common() 11895 dt.address = sregs->gdt.base; in __set_sregs_common()
|
| /arch/x86/kvm/vmx/ |
| A D | vmx.c | 1428 void *gdt = get_current_gdt_ro(); in vmx_vcpu_load_vmcs() local 1442 vmcs_writel(HOST_GDTR_BASE, (unsigned long)gdt); /* 22.2.4 */ in vmx_vcpu_load_vmcs()
|