| /arch/sparc/include/asm/ |
| A D | spitfire.h | 139 : "r" (entry << 3), "i" (ASI_DTLB_TAG_READ)); in spitfire_get_dtlb_tag() 148 : "r" (data), "r" (entry << 3), in spitfire_put_dtlb_data() 181 : "r" (data), "r" (entry << 3), in spitfire_put_itlb_data() 239 : "r" ((0 << 16) | (entry << 3)), in cheetah_get_ldtlb_data() 252 : "r" ((0 << 16) | (entry << 3)), in cheetah_get_litlb_data() 264 : "r" ((0 << 16) | (entry << 3)), in cheetah_get_ldtlb_tag() 276 : "r" ((0 << 16) | (entry << 3)), in cheetah_get_litlb_tag() 288 "r" ((0 << 16) | (entry << 3)), in cheetah_put_ldtlb_data() 298 "r" ((0 << 16) | (entry << 3)), in cheetah_put_litlb_data() 330 "r" ((tlb << 16) | (entry << 3)), in cheetah_put_dtlb_data() [all …]
|
| /arch/x86/kvm/ |
| A D | cpuid.c | 665 &entry.eax, &entry.ebx, &entry.ecx, &entry.edx); in raw_cpuid_get() 1286 &entry->eax, &entry->ebx, &entry->ecx, &entry->edx); in do_host_cpuid() 1434 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func() 1523 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func() 1553 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func() 1565 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func() 1576 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func() 1584 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func() 1720 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func() 1756 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func() [all …]
|
| A D | reverse_cpuid.h | 172 return &entry->eax; in __cpuid_entry_get_reg() 174 return &entry->ebx; in __cpuid_entry_get_reg() 176 return &entry->ecx; in __cpuid_entry_get_reg() 178 return &entry->edx; in __cpuid_entry_get_reg() 190 return __cpuid_entry_get_reg(entry, cpuid.reg); in cpuid_entry_get_reg() 193 static __always_inline u32 cpuid_entry_get(struct kvm_cpuid_entry2 *entry, in cpuid_entry_get() argument 196 u32 *reg = cpuid_entry_get_reg(entry, x86_feature); in cpuid_entry_get() 204 return cpuid_entry_get(entry, x86_feature); in cpuid_entry_has() 210 u32 *reg = cpuid_entry_get_reg(entry, x86_feature); in cpuid_entry_clear() 218 u32 *reg = cpuid_entry_get_reg(entry, x86_feature); in cpuid_entry_set() [all …]
|
| /arch/sparc/mm/ |
| A D | hugetlbpage.c | 25 return entry; in sun4u_hugepage_shift_to_tte() 32 pte_val(entry) = pte_val(entry) & ~_PAGE_SZALL_4V; in sun4v_hugepage_shift_to_tte() 57 pte_val(entry) = pte_val(entry) | hugepage_size; in sun4v_hugepage_shift_to_tte() 58 return entry; in sun4v_hugepage_shift_to_tte() 73 entry = pte_mkhuge(entry); in arch_make_huge_pte() 152 pte_val(entry)); in huge_tte_to_shift() 267 pte_t entry; in huge_ptep_get_and_clear() local 269 entry = *ptep; in huge_ptep_get_and_clear() 281 orig_shift = pte_none(entry) ? PAGE_SHIFT : huge_tte_to_shift(entry); in huge_ptep_get_and_clear() 283 if (pte_present(entry)) in huge_ptep_get_and_clear() [all …]
|
| /arch/s390/kernel/ |
| A D | jump_label.c | 23 insn->offset = (jump_entry_target(entry) - jump_entry_code(entry)) >> 1; in jump_label_make_nop() 30 insn->offset = (jump_entry_target(entry) - jump_entry_code(entry)) >> 1; in jump_label_make_branch() 50 void *code = (void *)jump_entry_code(entry); in jump_label_transform() 54 jump_label_make_nop(entry, &old); in jump_label_transform() 55 jump_label_make_branch(entry, &new); in jump_label_transform() 57 jump_label_make_branch(entry, &old); in jump_label_transform() 58 jump_label_make_nop(entry, &new); in jump_label_transform() 61 jump_label_bug(entry, &old, &new); in jump_label_transform() 65 void arch_jump_label_transform(struct jump_entry *entry, in arch_jump_label_transform() argument 68 jump_label_transform(entry, type); in arch_jump_label_transform() [all …]
|
| /arch/x86/platform/efi/ |
| A D | runtime-map.c | 92 entry = to_map_entry(kobj); in map_release() 93 kfree(entry); in map_release() 117 entry = kzalloc(sizeof(*entry), GFP_KERNEL); in add_sysfs_runtime_map_entry() 118 if (!entry) { in add_sysfs_runtime_map_entry() 127 entry->kobj.kset = map_kset; in add_sysfs_runtime_map_entry() 130 kobject_put(&entry->kobj); in add_sysfs_runtime_map_entry() 136 return entry; in add_sysfs_runtime_map_entry() 178 if (IS_ERR(entry)) { in efi_runtime_map_init() 179 ret = PTR_ERR(entry); in efi_runtime_map_init() 188 entry = *(map_entries + j); in efi_runtime_map_init() [all …]
|
| /arch/x86/kernel/ |
| A D | e820.c | 87 if (entry->addr >= end || entry->addr + entry->size <= start) in _e820__mapped_any() 124 if (entry->addr >= end || entry->addr + entry->size <= start) in __e820__mapped_all() 132 start = entry->addr + entry->size; in __e820__mapped_all() 160 return entry ? entry->type : -EINVAL; in e820__get_entry_type() 501 entry->size = start - entry->addr; in __e820__range_update() 569 memset(entry, 0, sizeof(*entry)); in e820__range_remove() 576 entry->size = start - entry->addr; in e820__range_remove() 1096 end = entry->addr + entry->size - 1; in e820__reserve_resources() 1123 firmware_map_add_early(entry->addr, entry->addr + entry->size, e820_type_to_string(entry)); in e820__reserve_resources() 1171 start = entry->addr + entry->size; in e820__reserve_resources_late() [all …]
|
| A D | jump_label.c | 20 int arch_jump_entry_size(struct jump_entry *entry) in arch_jump_entry_size() argument 24 insn_decode_kernel(&insn, (void *)jump_entry_code(entry)); in arch_jump_entry_size() 42 addr = (void *)jump_entry_code(entry); in __jump_label_patch() 43 dest = (void *)jump_entry_target(entry); in __jump_label_patch() 45 size = arch_jump_entry_size(entry); in __jump_label_patch() 83 __jump_label_transform(struct jump_entry *entry, in __jump_label_transform() argument 113 __jump_label_transform(entry, type, init); in jump_label_transform() 117 void arch_jump_label_transform(struct jump_entry *entry, in arch_jump_label_transform() argument 120 jump_label_transform(entry, type, 0); in arch_jump_label_transform() 132 arch_jump_label_transform(entry, type); in arch_jump_label_transform_queue() [all …]
|
| /arch/um/kernel/ |
| A D | irq.c | 124 struct irq_entry *entry; in irq_do_pending_events() local 281 os_add_epoll_fd(events, entry->fd, entry); in update_irq_entry() 294 return entry; in update_or_remove_irq_entry() 414 struct irq_entry *entry; in deactivate_fd() local 422 if (!entry) in deactivate_fd() 432 entry = update_or_remove_irq_entry(entry); in deactivate_fd() 435 kfree(entry); in deactivate_fd() 449 struct irq_entry *entry; in deactivate_all_fds() local 560 struct irq_entry *entry; in um_irqs_suspend() local 603 struct irq_entry *entry; in um_irqs_resume() local [all …]
|
| /arch/x86/boot/compressed/ |
| A D | idt_64.c | 10 gate_desc entry; in set_idt_entry() local 12 memset(&entry, 0, sizeof(entry)); in set_idt_entry() 14 entry.offset_low = (u16)(address & 0xffff); in set_idt_entry() 15 entry.segment = __KERNEL_CS; in set_idt_entry() 16 entry.bits.type = GATE_TRAP; in set_idt_entry() 17 entry.bits.p = 1; in set_idt_entry() 18 entry.offset_middle = (u16)((address >> 16) & 0xffff); in set_idt_entry() 19 entry.offset_high = (u32)(address >> 32); in set_idt_entry() 21 memcpy(&boot_idt[vector], &entry, sizeof(entry)); in set_idt_entry()
|
| /arch/mips/kernel/ |
| A D | perf_event.c | 26 static void save_raw_perf_callchain(struct perf_callchain_entry_ctx *entry, in save_raw_perf_callchain() argument 35 perf_callchain_store(entry, addr); in save_raw_perf_callchain() 36 if (entry->nr >= entry->max_stack) in save_raw_perf_callchain() 42 void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument 55 save_raw_perf_callchain(entry, sp); in perf_callchain_kernel() 59 perf_callchain_store(entry, pc); in perf_callchain_kernel() 60 if (entry->nr >= entry->max_stack) in perf_callchain_kernel() 65 save_raw_perf_callchain(entry, sp); in perf_callchain_kernel()
|
| /arch/arm/kernel/ |
| A D | perf_callchain.c | 35 struct perf_callchain_entry_ctx *entry) in user_backtrace() argument 50 perf_callchain_store(entry, buftail.lr); in user_backtrace() 63 perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) in perf_callchain_user() argument 67 perf_callchain_store(entry, regs->ARM_pc); in perf_callchain_user() 74 while ((entry->nr < entry->max_stack) && in perf_callchain_user() 76 tail = user_backtrace(tail, entry); in perf_callchain_user() 87 struct perf_callchain_entry_ctx *entry = data; in callchain_trace() local 88 return perf_callchain_store(entry, pc) == 0; in callchain_trace() 92 perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) in perf_callchain_kernel() argument 97 walk_stackframe(&fr, callchain_trace, entry); in perf_callchain_kernel()
|
| A D | jump_label.c | 7 static void __arch_jump_label_transform(struct jump_entry *entry, in __arch_jump_label_transform() argument 11 void *addr = (void *)entry->code; in __arch_jump_label_transform() 15 insn = arm_gen_branch(entry->code, entry->target); in __arch_jump_label_transform() 25 void arch_jump_label_transform(struct jump_entry *entry, in arch_jump_label_transform() argument 28 __arch_jump_label_transform(entry, type, false); in arch_jump_label_transform()
|
| /arch/csky/kernel/ |
| A D | perf_callchain.c | 36 struct perf_callchain_entry_ctx *entry) in walk_stackframe() argument 39 perf_callchain_store(entry, fr->lr); in walk_stackframe() 47 static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry, in user_backtrace() argument 67 perf_callchain_store(entry, lr); in user_backtrace() 86 void perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument 92 perf_callchain_store(entry, regs->pc); in perf_callchain_user() 100 fp = user_backtrace(entry, fp, regs->lr); in perf_callchain_user() 102 while (fp && !(fp & 0x3) && entry->nr < entry->max_stack) in perf_callchain_user() 103 fp = user_backtrace(entry, fp, 0); in perf_callchain_user() 106 void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument [all …]
|
| /arch/sh/mm/ |
| A D | tlbex_32.c | 31 pte_t entry; in handle_tlbmiss() local 57 entry = *pte; in handle_tlbmiss() 58 if (unlikely(pte_none(entry) || pte_not_present(entry))) in handle_tlbmiss() 60 if (unlikely(error_code && !pte_write(entry))) in handle_tlbmiss() 64 entry = pte_mkdirty(entry); in handle_tlbmiss() 65 entry = pte_mkyoung(entry); in handle_tlbmiss() 67 set_pte(pte, entry); in handle_tlbmiss()
|
| /arch/mips/mm/ |
| A D | pgtable-64.c | 19 unsigned long entry; in pgd_init() local 22 entry = (unsigned long)invalid_pud_table; in pgd_init() 24 entry = (unsigned long)invalid_pmd_table; in pgd_init() 33 p[0] = entry; in pgd_init() 34 p[1] = entry; in pgd_init() 35 p[2] = entry; in pgd_init() 36 p[3] = entry; in pgd_init() 37 p[4] = entry; in pgd_init() 39 p[-3] = entry; in pgd_init() 40 p[-2] = entry; in pgd_init() [all …]
|
| /arch/x86/mm/pat/ |
| A D | memtype_interval.c | 36 static inline u64 interval_start(struct memtype *entry) in interval_start() argument 38 return entry->start; in interval_start() 41 static inline u64 interval_end(struct memtype *entry) in interval_end() argument 43 return entry->end - 1; in interval_end() 107 struct memtype *entry = interval_iter_first(&memtype_rbroot, start, end - 1); in memtype_erase() local 109 while (entry && entry->start < end) { in memtype_erase() 110 if (entry->start == start && entry->end == end) { in memtype_erase() 111 interval_remove(entry, &memtype_rbroot); in memtype_erase() 112 return entry; in memtype_erase() 114 entry = interval_iter_next(entry, start, end - 1); in memtype_erase()
|
| /arch/arm64/mm/ |
| A D | mteswap.c | 50 void mte_restore_tags(swp_entry_t entry, struct page *page) in mte_restore_tags() argument 52 void *tags = xa_load(&mte_pages, entry.val); in mte_restore_tags() 65 swp_entry_t entry = swp_entry(type, offset); in mte_invalidate_tags() local 66 void *tags = xa_erase(&mte_pages, entry.val); in mte_invalidate_tags() 73 swp_entry_t entry = page_swap_entry(page); in __mte_invalidate_tags() local 75 mte_invalidate_tags(swp_type(entry), swp_offset(entry)); in __mte_invalidate_tags() 80 swp_entry_t entry = swp_entry(type, 0); in mte_invalidate_tags_area() local 84 XA_STATE(xa_state, &mte_pages, entry.val); in mte_invalidate_tags_area() 117 void arch_swap_restore(swp_entry_t entry, struct folio *folio) in arch_swap_restore() argument 127 mte_restore_tags(entry, folio_page(folio, i)); in arch_swap_restore() [all …]
|
| /arch/parisc/kernel/ |
| A D | alternative.c | 26 struct alt_instr *entry; in apply_alternatives() local 46 for (entry = start; entry < end; entry++, index++) { in apply_alternatives() 52 from = (u32 *)((ulong)&entry->orig_offset + entry->orig_offset); in apply_alternatives() 53 len = entry->len; in apply_alternatives() 54 cond = entry->cond; in apply_alternatives() 55 replacement = entry->replacement; in apply_alternatives() 90 source = (u32 *)((ulong)&entry->replacement + entry->replacement); in apply_alternatives()
|
| /arch/x86/xen/ |
| A D | setup.c | 173 e_pfn = PFN_DOWN(entry->addr + entry->size); in xen_find_pfn_range() 429 phys_addr_t end = entry->addr + entry->size; in xen_foreach_remap_area() 583 if (entry->type == E820_TYPE_RAM && entry->addr <= start && in xen_is_e820_reserved() 584 (entry->addr + entry->size) >= end) in xen_is_e820_reserved() 587 entry++; in xen_is_e820_reserved() 608 if (entry->type != E820_TYPE_RAM || entry->size < size) in xen_find_free_area() 615 if (start + size > entry->addr + entry->size) in xen_find_free_area() 644 entry_end = entry->addr + entry->size; in xen_e820_swap_entry_with_ram() 676 entry++; in xen_e820_swap_entry_with_ram() 704 if (entry->addr + entry->size > start && in xen_e820_resolve_conflicts() [all …]
|
| /arch/riscv/kernel/ |
| A D | perf_callchain.c | 9 static bool fill_callchain(void *entry, unsigned long pc) in fill_callchain() argument 11 return perf_callchain_store(entry, pc) == 0; in fill_callchain() 28 void perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument 36 arch_stack_walk_user(fill_callchain, entry, regs); in perf_callchain_user() 39 void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument 47 walk_stackframe(NULL, regs, fill_callchain, entry); in perf_callchain_kernel()
|
| /arch/loongarch/mm/ |
| A D | pgtable.c | 43 unsigned long entry; in pgd_init() local 46 entry = (unsigned long)invalid_pud_table; in pgd_init() 48 entry = (unsigned long)invalid_pmd_table; in pgd_init() 57 p[0] = entry; in pgd_init() 58 p[1] = entry; in pgd_init() 59 p[2] = entry; in pgd_init() 60 p[3] = entry; in pgd_init() 61 p[4] = entry; in pgd_init() 63 p[-3] = entry; in pgd_init() 64 p[-2] = entry; in pgd_init() [all …]
|
| /arch/arm64/kernel/ |
| A D | perf_callchain.c | 15 struct perf_callchain_entry_ctx *entry = data; in callchain_trace() local 17 return perf_callchain_store(entry, pc) == 0; in callchain_trace() 20 void perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument 28 arch_stack_walk_user(callchain_trace, entry, regs); in perf_callchain_user() 31 void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, in perf_callchain_kernel() argument 39 arch_stack_walk(callchain_trace, entry, current, regs); in perf_callchain_kernel()
|
| /arch/m68k/include/asm/ |
| A D | sun3mmu.h | 88 register unsigned long entry; in sun3_get_segmap() local 95 entry = c; in sun3_get_segmap() 97 return entry; in sun3_get_segmap() 101 static inline void sun3_put_segmap(unsigned long addr, unsigned long entry) in sun3_put_segmap() argument 107 SET_CONTROL_BYTE (AC_SEGMAP | (addr & SUN3_CONTROL_MASK), entry); in sun3_put_segmap() 116 register unsigned long entry; in sun3_get_pte() local 121 GET_CONTROL_WORD (AC_PAGEMAP | (addr & SUN3_CONTROL_MASK), entry); in sun3_get_pte() 124 return entry; in sun3_get_pte() 128 static inline void sun3_put_pte(unsigned long addr, unsigned long entry) in sun3_put_pte() argument 134 SET_CONTROL_WORD (AC_PAGEMAP | (addr & SUN3_CONTROL_MASK), entry); in sun3_put_pte()
|
| /arch/powerpc/kernel/ |
| A D | systbl.c | 19 #define __SYSCALL_WITH_COMPAT(nr, entry, compat) __SYSCALL(nr, entry) argument 23 #define __SYSCALL(nr, entry) [nr] = entry, argument 29 #define __SYSCALL(nr, entry) [nr] = (void *) entry, argument
|