| /arch/powerpc/mm/book3s64/ |
| A D | hash_tlb.c | 51 i = batch->index; in hpte_need_flush() 103 if (!batch->active) { in hpte_need_flush() 119 if (i != 0 && (mm != batch->mm || batch->psize != psize || in hpte_need_flush() 125 batch->mm = mm; in hpte_need_flush() 129 batch->pte[i] = rpte; in hpte_need_flush() 130 batch->vpn[i] = vpn; in hpte_need_flush() 131 batch->index = ++i; in hpte_need_flush() 148 i = batch->index; in __flush_tlb_pending() 151 flush_hash_page(batch->vpn[0], batch->pte[0], in __flush_tlb_pending() 152 batch->psize, batch->ssize, local); in __flush_tlb_pending() [all …]
|
| A D | hash_native.c | 784 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in native_flush_hash_range() local 785 unsigned long psize = batch->psize; in native_flush_hash_range() 786 int ssize = batch->ssize; in native_flush_hash_range() 792 vpn = batch->vpn[i]; in native_flush_hash_range() 793 pte = batch->pte[i]; in native_flush_hash_range() 826 vpn = batch->vpn[i]; in native_flush_hash_range() 827 pte = batch->pte[i]; in native_flush_hash_range() 843 vpn = batch->vpn[i]; in native_flush_hash_range() 844 pte = batch->pte[i]; in native_flush_hash_range()
|
| A D | hash_utils.c | 2266 struct ppc64_tlb_batch *batch = in flush_hash_range() local 2270 flush_hash_page(batch->vpn[i], batch->pte[i], in flush_hash_range() 2271 batch->psize, batch->ssize, local); in flush_hash_range()
|
| /arch/powerpc/include/asm/book3s/64/ |
| A D | tlbflush-hash.h | 25 extern void __flush_tlb_pending(struct ppc64_tlb_batch *batch); 31 struct ppc64_tlb_batch *batch; in arch_enter_lazy_mmu_mode() local 40 batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_enter_lazy_mmu_mode() 41 batch->active = 1; in arch_enter_lazy_mmu_mode() 46 struct ppc64_tlb_batch *batch; in arch_leave_lazy_mmu_mode() local 50 batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_leave_lazy_mmu_mode() 52 if (batch->index) in arch_leave_lazy_mmu_mode() 53 __flush_tlb_pending(batch); in arch_leave_lazy_mmu_mode() 54 batch->active = 0; in arch_leave_lazy_mmu_mode()
|
| /arch/riscv/mm/ |
| A D | tlbflush.c | 230 void arch_tlbbatch_add_pending(struct arch_tlbflush_unmap_batch *batch, in arch_tlbbatch_add_pending() argument 233 cpumask_or(&batch->cpumask, &batch->cpumask, mm_cpumask(mm)); in arch_tlbbatch_add_pending() 237 void arch_tlbbatch_flush(struct arch_tlbflush_unmap_batch *batch) in arch_tlbbatch_flush() argument 239 __flush_tlb_range(NULL, &batch->cpumask, in arch_tlbbatch_flush() 241 cpumask_clear(&batch->cpumask); in arch_tlbbatch_flush()
|
| /arch/x86/include/asm/ |
| A D | tlbflush.h | 350 static inline void arch_tlbbatch_add_pending(struct arch_tlbflush_unmap_batch *batch, in arch_tlbbatch_add_pending() argument 354 cpumask_or(&batch->cpumask, &batch->cpumask, mm_cpumask(mm)); in arch_tlbbatch_add_pending() 355 batch->unmapped_pages = true; in arch_tlbbatch_add_pending() 359 extern void arch_tlbbatch_flush(struct arch_tlbflush_unmap_batch *batch);
|
| /arch/riscv/include/asm/ |
| A D | tlbflush.h | 64 void arch_tlbbatch_add_pending(struct arch_tlbflush_unmap_batch *batch, 66 void arch_tlbbatch_flush(struct arch_tlbflush_unmap_batch *batch);
|
| /arch/powerpc/platforms/pseries/ |
| A D | lpar.c | 1359 psize = batch->psize; in do_block_remove() 1360 ssize = batch->ssize; in do_block_remove() 1363 vpn = batch->vpn[i]; in do_block_remove() 1364 pte = batch->pte[i]; in do_block_remove() 1531 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in pSeries_lpar_flush_hash_range() local 1541 if (is_supported_hlbkrm(batch->psize, batch->psize)) { in pSeries_lpar_flush_hash_range() 1542 do_block_remove(number, batch, param); in pSeries_lpar_flush_hash_range() 1546 psize = batch->psize; in pSeries_lpar_flush_hash_range() 1547 ssize = batch->ssize; in pSeries_lpar_flush_hash_range() 1550 vpn = batch->vpn[i]; in pSeries_lpar_flush_hash_range() [all …]
|
| /arch/x86/mm/ |
| A D | tlb.c | 1708 void arch_tlbbatch_flush(struct arch_tlbflush_unmap_batch *batch) in arch_tlbbatch_flush() argument 1721 if (cpu_feature_enabled(X86_FEATURE_INVLPGB) && batch->unmapped_pages) { in arch_tlbbatch_flush() 1723 batch->unmapped_pages = false; in arch_tlbbatch_flush() 1724 } else if (cpumask_any_but(&batch->cpumask, cpu) < nr_cpu_ids) { in arch_tlbbatch_flush() 1725 flush_tlb_multi(&batch->cpumask, info); in arch_tlbbatch_flush() 1726 } else if (cpumask_test_cpu(cpu, &batch->cpumask)) { in arch_tlbbatch_flush() 1733 cpumask_clear(&batch->cpumask); in arch_tlbbatch_flush()
|
| /arch/arm64/include/asm/ |
| A D | tlbflush.h | 335 static inline void arch_tlbbatch_flush(struct arch_tlbflush_unmap_batch *batch) in arch_tlbbatch_flush() argument 522 static inline void arch_tlbbatch_add_pending(struct arch_tlbflush_unmap_batch *batch, in arch_tlbbatch_add_pending() argument
|
| /arch/powerpc/kernel/ |
| A D | process.c | 1285 struct ppc64_tlb_batch *batch; in __switch_to() local 1294 batch = this_cpu_ptr(&ppc64_tlb_batch); in __switch_to() 1295 if (batch->active) { in __switch_to() 1297 if (batch->index) in __switch_to() 1298 __flush_tlb_pending(batch); in __switch_to() 1299 batch->active = 0; in __switch_to() 1381 batch = this_cpu_ptr(&ppc64_tlb_batch); in __switch_to() 1382 batch->active = 1; in __switch_to()
|
| /arch/x86/xen/ |
| A D | mmu_pv.c | 2517 int batch = min(REMAP_BATCH_SIZE, nr); in xen_remap_pfn() local 2518 int batch_left = batch; in xen_remap_pfn() 2520 range = (unsigned long)batch << PAGE_SHIFT; in xen_remap_pfn() 2558 nr -= batch; in xen_remap_pfn() 2561 err_ptr += batch; in xen_remap_pfn()
|
| /arch/arm/boot/dts/allwinner/ |
| A D | sun8i-h2-plus-bananapi-m2-zero.dts | 140 * On the production batch of this board the card detect GPIO is
|
| /arch/arm/lib/ |
| A D | lib1funcs.S | 171 @ Do comparisons in batch of 4 first.
|
| /arch/x86/kvm/mmu/ |
| A D | mmu.c | 6617 int nr_zapped, batch = 0; in kvm_zap_obsolete_pages() local 6647 if (batch >= BATCH_ZAP_PAGES && in kvm_zap_obsolete_pages() 6649 batch = 0; in kvm_zap_obsolete_pages() 6655 batch += nr_zapped; in kvm_zap_obsolete_pages()
|