| /arch/sparc/vdso/ |
| A D | vma.c | 75 shdrs = (void *)e->hdr + e->hdr->e_shoff; in one_section64() 76 snames = (void *)e->hdr + shdrs[e->hdr->e_shstrndx].sh_offset; in one_section64() 91 e->hdr = image->data; in find_sections64() 92 e->dynsym = one_section64(e, ".dynsym", &e->dynsymsize); in find_sections64() 93 e->dynstr = one_section64(e, ".dynstr", NULL); in find_sections64() 95 if (!e->dynsym || !e->dynstr) { in find_sections64() 143 shdrs = (void *)e->hdr + e->hdr->e_shoff; in one_section32() 144 snames = (void *)e->hdr + shdrs[e->hdr->e_shstrndx].sh_offset; in one_section32() 160 e->dynsym = one_section32(e, ".dynsym", &e->dynsymsize); in find_sections32() 161 e->dynstr = one_section32(e, ".dynstr", NULL); in find_sections32() [all …]
|
| /arch/arm64/kvm/vgic/ |
| A D | vgic-irqfd.c | 46 e->set = vgic_irqfd_set_irq; in kvm_set_routing_entry() 54 e->set = kvm_set_msi; in kvm_set_routing_entry() 57 e->msi.data = ue->u.msi.data; in kvm_set_routing_entry() 58 e->msi.flags = ue->flags; in kvm_set_routing_entry() 59 e->msi.devid = ue->u.msi.devid; in kvm_set_routing_entry() 74 msi->data = e->msi.data; in kvm_populate_msi() 75 msi->flags = e->msi.flags; in kvm_populate_msi() 76 msi->devid = e->msi.devid; in kvm_populate_msi() 98 kvm_populate_msi(e, &msi); in kvm_set_msi() 112 switch (e->type) { in kvm_arch_set_irq_inatomic() [all …]
|
| /arch/loongarch/kvm/ |
| A D | irqfd.c | 32 pch_msi_set_irq(kvm, e->msi.data, level); in kvm_set_msi() 47 struct kvm_kernel_irq_routing_entry *e, in kvm_set_routing_entry() argument 52 e->set = kvm_set_pic_irq; in kvm_set_routing_entry() 53 e->irqchip.irqchip = ue->u.irqchip.irqchip; in kvm_set_routing_entry() 54 e->irqchip.pin = ue->u.irqchip.pin; in kvm_set_routing_entry() 61 e->set = kvm_set_msi; in kvm_set_routing_entry() 62 e->msi.address_lo = ue->u.msi.address_lo; in kvm_set_routing_entry() 63 e->msi.address_hi = ue->u.msi.address_hi; in kvm_set_routing_entry() 64 e->msi.data = ue->u.msi.data; in kvm_set_routing_entry() 74 switch (e->type) { in kvm_arch_set_irq_inatomic() [all …]
|
| /arch/x86/xen/ |
| A D | enlighten_pvh.c | 78 if (e->type != E820_TYPE_RAM) in pvh_reserve_extra_memory() 80 ram_pages += PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr); in pvh_reserve_extra_memory() 94 if (e->type != E820_TYPE_UNUSABLE) in pvh_reserve_extra_memory() 98 PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr)); in pvh_reserve_extra_memory() 100 if (pages != (PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr))) { in pvh_reserve_extra_memory() 109 next = e + 1; in pvh_reserve_extra_memory() 110 memmove(next, e, in pvh_reserve_extra_memory() 111 (bootp->e820_entries - i) * sizeof(*e)); in pvh_reserve_extra_memory() 114 e->size = next->addr - e->addr; in pvh_reserve_extra_memory() 115 next->size -= e->size; in pvh_reserve_extra_memory() [all …]
|
| /arch/riscv/kvm/ |
| A D | vm.c | 77 msi.address_lo = e->msi.address_lo; in kvm_set_msi() 78 msi.address_hi = e->msi.address_hi; in kvm_set_msi() 79 msi.data = e->msi.data; in kvm_set_msi() 80 msi.flags = e->msi.flags; in kvm_set_msi() 81 msi.devid = e->msi.devid; in kvm_set_msi() 127 e->set = kvm_riscv_set_irq; in kvm_set_routing_entry() 135 e->set = kvm_set_msi; in kvm_set_routing_entry() 138 e->msi.data = ue->u.msi.data; in kvm_set_routing_entry() 139 e->msi.flags = ue->flags; in kvm_set_routing_entry() 140 e->msi.devid = ue->u.msi.devid; in kvm_set_routing_entry() [all …]
|
| /arch/x86/mm/ |
| A D | extable.c | 36 if (e->data & EX_FLAG_CLEAR_AX) in ex_handler_default() 38 if (e->data & EX_FLAG_CLEAR_DX) in ex_handler_default() 41 regs->ip = ex_fixup_addr(e); in ex_handler_default() 66 next_ip = ex_fixup_addr(e); in ex_handler_zeropad() 95 return ex_handler_default(e, regs); in ex_handler_zeropad() 296 return e ? FIELD_GET(EX_DATA_TYPE_MASK, e->data) : EX_TYPE_NONE; in ex_get_fixup_type() 302 const struct exception_table_entry *e; in fixup_exception() local 320 if (!e) in fixup_exception() 330 return ex_handler_default(e, regs); in fixup_exception() 337 return ex_handler_clear_fs(e, regs); in fixup_exception() [all …]
|
| /arch/um/kernel/ |
| A D | time.c | 406 if (e->pending) in __time_travel_add_event() 410 e->time = time; in __time_travel_add_event() 422 (tmp->time == e->time && tmp->onstack && e->onstack)) { in __time_travel_add_event() 485 e->fn(e); in deliver_time_travel_irqs() 498 e->fn(e); in time_travel_deliver_event() 512 e->fn(e); in time_travel_deliver_event() 522 if (!e->pending) in time_travel_del_event() 545 BUG_ON(!e); in time_travel_update_time() 553 if (e == &ne) { in time_travel_update_time() 558 time_travel_time, e->time, e); in time_travel_update_time() [all …]
|
| /arch/powerpc/crypto/ |
| A D | aes-tab-4k.S | 67 .long R(36, 1b, 1b, 2d), R(dc, 6e, 6e, b2) 102 .long R(fc, 7e, 7e, 82), R(7a, 3d, 3d, 47) 111 .long R(a7, de, de, 79), R(bc, 5e, 5e, e2) 114 .long R(74, 3a, 3a, 4e), R(14, 0a, 0a, 1e) 124 .long R(9c, 4e, 4e, d2), R(49, a9, a9, e0) 130 .long R(4a, 25, 25, 6f), R(5c, 2e, 2e, 72) 137 .long R(e0, 70, 70, 90), R(7c, 3e, 3e, 42) 140 .long R(f7, f6, f6, 01), R(1c, 0e, 0e, 12) 144 .long R(3a, 1d, 1d, 27), R(27, 9e, 9e, b9) 148 .long R(07, 8e, 8e, 89), R(33, 94, 94, a7) [all …]
|
| /arch/x86/crypto/ |
| A D | sm3-avx-asm_64.S | 112 #define e %r12d macro 185 leal K##round(t0, e, 1), t1; \ 200 GG##i(e, f, g, t2, t1); \ 360 movl state_h4(RSTATE), e; 374 R1(d, a, b, c, h, e, f, g, 1, 1, IW); 375 R1(c, d, a, b, g, h, e, f, 2, 2, IW); 379 R1(a, b, c, d, e, f, g, h, 4, 0, IW); 380 R1(d, a, b, c, h, e, f, g, 5, 1, IW); 467 R2(c, d, a, b, g, h, e, f, 58, 1, XW); 487 xorl state_h4(RSTATE), e; [all …]
|
| /arch/um/include/asm/ |
| A D | pgtable-4level.h | 45 #define pte_ERROR(e) \ argument 46 printk("%s:%d: bad pte %p(%016lx).\n", __FILE__, __LINE__, &(e), \ 47 pte_val(e)) 48 #define pmd_ERROR(e) \ argument 49 printk("%s:%d: bad pmd %p(%016lx).\n", __FILE__, __LINE__, &(e), \ 50 pmd_val(e)) 51 #define pud_ERROR(e) \ argument 52 printk("%s:%d: bad pud %p(%016lx).\n", __FILE__, __LINE__, &(e), \ 53 pud_val(e)) 54 #define pgd_ERROR(e) \ argument [all …]
|
| A D | pgtable-2level.h | 27 #define pte_ERROR(e) \ argument 28 printk("%s:%d: bad pte %p(%08lx).\n", __FILE__, __LINE__, &(e), \ 29 pte_val(e)) 30 #define pgd_ERROR(e) \ argument 31 printk("%s:%d: bad pgd %p(%08lx).\n", __FILE__, __LINE__, &(e), \ 32 pgd_val(e))
|
| /arch/parisc/kernel/ |
| A D | unwind.c | 61 e = &table->table[mid]; in find_unwind_entry_in_table() 62 if (addr < e->region_start) in find_unwind_entry_in_table() 64 else if (addr > e->region_end) in find_unwind_entry_in_table() 67 return e; in find_unwind_entry_in_table() 90 if (e) { in find_unwind_entry() 99 return e; in find_unwind_entry() 156 unwind_table_sort(s, e); in unwind_table_add() 283 if (e == NULL) { in unwind_frame_regs() 325 e->region_start, e->region_end, e->Save_SP, e->Save_RP, in unwind_frame_regs() 326 e->Millicode, e->Total_frame_size); in unwind_frame_regs() [all …]
|
| /arch/um/include/linux/ |
| A D | time-internal.h | 27 time_travel_set_event_fn(struct time_travel_event *e, in time_travel_set_event_fn() argument 30 e->fn = fn; in time_travel_set_event_fn() 49 void time_travel_add_irq_event(struct time_travel_event *e); 50 void time_travel_add_event_rel(struct time_travel_event *e, 52 bool time_travel_del_event(struct time_travel_event *e); 62 #define time_travel_set_event_fn(e, fn) do {} while (0) argument 72 static inline void time_travel_add_irq_event(struct time_travel_event *e) in time_travel_add_irq_event() argument
|
| /arch/xtensa/mm/ |
| A D | tlb.c | 31 int e = w + (i << PAGE_SHIFT); in __flush_itlb_all() local 32 invalidate_itlb_entry_no_isync(e); in __flush_itlb_all() 44 int e = w + (i << PAGE_SHIFT); in __flush_dtlb_all() local 45 invalidate_dtlb_entry_no_isync(e); in __flush_dtlb_all() 234 unsigned tlbidx = w | (e << PAGE_SHIFT); in check_tlb_entry() 248 dtlb ? 'D' : 'I', w, e, vpn, in check_tlb_entry() 278 unsigned w, e; in check_tlb_sanity() local 283 for (e = 0; e < (1 << XCHAL_DTLB_ARF_ENTRIES_LOG2); ++e) in check_tlb_sanity() 284 bug |= check_tlb_entry(w, e, true); in check_tlb_sanity() 286 for (e = 0; e < (1 << XCHAL_ITLB_ARF_ENTRIES_LOG2); ++e) in check_tlb_sanity() [all …]
|
| /arch/x86/include/asm/ |
| A D | pgtable_64.h | 34 #define pte_ERROR(e) \ argument 36 __FILE__, __LINE__, &(e), pte_val(e)) 37 #define pmd_ERROR(e) \ argument 39 __FILE__, __LINE__, &(e), pmd_val(e)) 40 #define pud_ERROR(e) \ argument 42 __FILE__, __LINE__, &(e), pud_val(e)) 44 #define p4d_ERROR(e) \ argument 46 __FILE__, __LINE__, &(e), p4d_val(e)) 48 #define pgd_ERROR(e) \ argument 50 __FILE__, __LINE__, &(e), pgd_val(e))
|
| A D | pgtable-3level.h | 12 #define pte_ERROR(e) \ argument 14 __FILE__, __LINE__, &(e), (e).pte_high, (e).pte_low) 15 #define pmd_ERROR(e) \ argument 17 __FILE__, __LINE__, &(e), pmd_val(e)) 18 #define pgd_ERROR(e) \ argument 20 __FILE__, __LINE__, &(e), pgd_val(e))
|
| /arch/x86/kvm/ |
| A D | ioapic.c | 116 e = &ioapic->redirtbl[RTC_GSI]; in __rtc_irq_eoi_tracking_restore_one() 118 e->fields.dest_id, in __rtc_irq_eoi_tracking_restore_one() 383 e->bits &= 0xffffffff; in ioapic_write_indirect() 384 e->bits |= (u64) val << 32; in ioapic_write_indirect() 386 e->bits &= ~0xffffffffULL; in ioapic_write_indirect() 387 e->bits |= (u32) val; in ioapic_write_indirect() 399 e->fields.remote_irr = 0; in ioapic_write_indirect() 401 mask_after = e->fields.mask; in ioapic_write_indirect() 405 ioapic->irr & (1 << index) && !e->fields.mask && !e->fields.remote_irr) { in ioapic_write_indirect() 462 !!e->fields.dest_mode); in ioapic_write_indirect() [all …]
|
| A D | irq.c | 261 .data = e->msi.data }; in kvm_msi_to_lapic_irq() 287 if (kvm_msi_route_invalid(kvm, e)) in kvm_set_msi() 293 kvm_msi_to_lapic_irq(kvm, e, &irq); in kvm_set_msi() 305 switch (e->type) { in kvm_arch_set_irq_inatomic() 313 if (kvm_msi_route_invalid(kvm, e)) in kvm_arch_set_irq_inatomic() 374 e->set = kvm_pic_set_irq; in kvm_set_routing_entry() 379 e->set = kvm_ioapic_set_irq; in kvm_set_routing_entry() 388 e->set = kvm_set_msi; in kvm_set_routing_entry() 391 e->msi.data = ue->u.msi.data; in kvm_set_routing_entry() 393 if (kvm_msi_route_invalid(kvm, e)) in kvm_set_routing_entry() [all …]
|
| /arch/powerpc/lib/ |
| A D | rheap.c | 164 e = s + size; in attach_free_block() 184 if (e == bs) in attach_free_block() 337 e = s + size; in rh_attach_region() 344 e = e & ~m; in rh_attach_region() 346 if (IS_ERR_VALUE(e) || (e < s)) in rh_attach_region() 351 size = e - s; in rh_attach_region() 382 e = s + size; in rh_detach_region() 389 e = e & ~m; in rh_detach_region() 428 newblk->start = e; in rh_detach_region() 539 e = s + size; in rh_alloc_fixed() [all …]
|
| /arch/arc/include/asm/ |
| A D | tlbflush.h | 23 #define flush_tlb_range(vma, s, e) local_flush_tlb_range(vma, s, e) argument 25 #define flush_tlb_kernel_range(s, e) local_flush_tlb_kernel_range(s, e) argument 29 #define flush_pmd_tlb_range(vma, s, e) local_flush_pmd_tlb_range(vma, s, e) argument
|
| A D | linkage.h | 17 .macro ST2 e, o, off 19 std \e, [sp, \off] 21 st \e, [sp, \off] 26 .macro LD2 e, o, off 28 ldd \e, [sp, \off] 30 ld \e, [sp, \off]
|
| /arch/mips/kernel/ |
| A D | jump_label.c | 38 void arch_jump_label_transform(struct jump_entry *e, in arch_jump_label_transform() argument 45 insn_p = (union mips_instruction *)msk_isa16_mode(e->code); in arch_jump_label_transform() 48 BUG_ON((e->target & J_ALIGN_MASK) != J_ISA_BIT); in arch_jump_label_transform() 52 offset = e->target - ((unsigned long)insn_p + 4); in arch_jump_label_transform() 69 WARN_ON((e->target & ~J_RANGE_MASK) != in arch_jump_label_transform() 70 ((e->code + 4) & ~J_RANGE_MASK)); in arch_jump_label_transform() 73 insn.j_format.target = e->target >> J_RANGE_SHIFT; in arch_jump_label_transform()
|
| /arch/x86/kernel/ |
| A D | early_printk.c | 161 char *e; in early_serial_init() local 175 port = simple_strtoul(s, &e, 10); in early_serial_init() 176 if (port > 1 || s == e) in early_serial_init() 186 baud = simple_strtoull(s, &e, 0); in early_serial_init() 188 if (baud == 0 || s == e) in early_serial_init() 223 char *e; in early_mmio_serial_init() local 245 if (baudrate == 0 || s == e) in early_mmio_serial_init() 269 char *e; in early_pci_serial_init() local 288 s = e; in early_pci_serial_init() 293 s = e; in early_pci_serial_init() [all …]
|
| /arch/powerpc/kvm/ |
| A D | e500.c | 368 sregs->u.e.impl_id = KVM_SREGS_E_IMPL_FSL; in kvmppc_core_get_sregs_e500() 370 sregs->u.e.impl.fsl.features = 0; in kvmppc_core_get_sregs_e500() 371 sregs->u.e.impl.fsl.svr = vcpu_e500->svr; in kvmppc_core_get_sregs_e500() 372 sregs->u.e.impl.fsl.hid0 = vcpu_e500->hid0; in kvmppc_core_get_sregs_e500() 373 sregs->u.e.impl.fsl.mcar = vcpu_e500->mcar; in kvmppc_core_get_sregs_e500() 378 sregs->u.e.ivor_high[3] = in kvmppc_core_get_sregs_e500() 393 vcpu_e500->svr = sregs->u.e.impl.fsl.svr; in kvmppc_core_set_sregs_e500() 407 sregs->u.e.ivor_high[0]; in kvmppc_core_set_sregs_e500() 409 sregs->u.e.ivor_high[1]; in kvmppc_core_set_sregs_e500() 411 sregs->u.e.ivor_high[2]; in kvmppc_core_set_sregs_e500() [all …]
|
| A D | e500mc.c | 226 sregs->u.e.impl_id = KVM_SREGS_E_IMPL_FSL; in kvmppc_core_get_sregs_e500mc() 228 sregs->u.e.impl.fsl.features = 0; in kvmppc_core_get_sregs_e500mc() 229 sregs->u.e.impl.fsl.svr = vcpu_e500->svr; in kvmppc_core_get_sregs_e500mc() 230 sregs->u.e.impl.fsl.hid0 = vcpu_e500->hid0; in kvmppc_core_get_sregs_e500mc() 231 sregs->u.e.impl.fsl.mcar = vcpu_e500->mcar; in kvmppc_core_get_sregs_e500mc() 235 sregs->u.e.ivor_high[3] = in kvmppc_core_get_sregs_e500mc() 250 vcpu_e500->svr = sregs->u.e.impl.fsl.svr; in kvmppc_core_set_sregs_e500mc() 262 if (sregs->u.e.features & KVM_SREGS_E_PM) { in kvmppc_core_set_sregs_e500mc() 264 sregs->u.e.ivor_high[3]; in kvmppc_core_set_sregs_e500mc() 269 sregs->u.e.ivor_high[4]; in kvmppc_core_set_sregs_e500mc() [all …]
|