| /arch/arm64/kvm/hyp/nvhe/ |
| A D | page_alloc.c | 54 u8 order) in __find_buddy_avail() argument 58 if (!buddy || buddy->order != order || buddy->refcount) in __find_buddy_avail() 97 u8 order = p->order; in __hyp_attach_page() local 113 for (; (order + 1) <= pool->max_order; order++) { in __hyp_attach_page() 126 p->order = order; in __hyp_attach_page() 137 while (p->order > order) { in __hyp_extract_page() 144 p->order--; in __hyp_extract_page() 146 buddy->order = p->order; in __hyp_extract_page() 186 u8 order = p->order; in hyp_split_page() local 189 p->order = 0; in hyp_split_page() [all …]
|
| /arch/riscv/kvm/ |
| A D | tlb.c | 25 unsigned long order) in kvm_riscv_local_hfence_gvma_vmid_gpa() argument 53 unsigned long order) in kvm_riscv_local_hfence_gvma_gpa() argument 84 unsigned long order) in kvm_riscv_local_hfence_vvma_asid_gva() argument 124 unsigned long order) in kvm_riscv_local_hfence_vvma_gva() argument 252 d.size, d.order); in kvm_riscv_hfence_process() 267 d.size, d.order); in kvm_riscv_hfence_process() 280 d.addr, d.size, d.order); in kvm_riscv_hfence_process() 283 d.size, d.order); in kvm_riscv_hfence_process() 353 data.order = order; in kvm_riscv_hfence_gvma_vmid_gpa() 383 data.order = order; in kvm_riscv_hfence_vvma_asid_gva() [all …]
|
| A D | gstage.c | 111 unsigned long order = PAGE_SHIFT; in gstage_tlb_flush() local 113 if (gstage_level_to_page_order(level, &order)) in gstage_tlb_flush() 115 addr &= ~(BIT(order) - 1); in gstage_tlb_flush() 118 kvm_riscv_local_hfence_gvma_vmid_gpa(gstage->vmid, addr, BIT(order), order); in gstage_tlb_flush() 120 kvm_riscv_hfence_gvma_vmid_gpa(gstage->kvm, -1UL, 0, addr, BIT(order), order, in gstage_tlb_flush()
|
| /arch/riscv/mm/ |
| A D | hugetlbpage.c | 35 unsigned long order; in huge_pte_alloc() local 68 for_each_napot_order(order) { in huge_pte_alloc() 88 unsigned long order; in huge_pte_offset() local 119 for_each_napot_order(order) { in huge_pte_offset() 189 unsigned long order; in arch_make_huge_pte() local 191 for_each_napot_order(order) { in arch_make_huge_pte() 197 if (order == NAPOT_ORDER_MAX) in arch_make_huge_pte() 281 unsigned long order; in huge_ptep_set_access_flags() local 326 unsigned long order; in huge_ptep_set_wrprotect() local 383 unsigned long order; in is_napot_size() local [all …]
|
| /arch/riscv/include/asm/ |
| A D | kvm_tlb.h | 25 unsigned long order; member 36 unsigned long order); 39 unsigned long order); 45 unsigned long order); 50 unsigned long order); 64 unsigned long order, unsigned long vmid); 71 unsigned long order, unsigned long asid, 79 unsigned long order, unsigned long vmid);
|
| A D | pgtable-64.h | 95 #define for_each_napot_order(order) \ argument 96 for (order = NAPOT_CONT_ORDER_BASE; order < NAPOT_ORDER_MAX; order++) 97 #define for_each_napot_order_rev(order) \ argument 98 for (order = NAPOT_ORDER_MAX - 1; \ 99 order >= NAPOT_CONT_ORDER_BASE; order--) 102 #define napot_cont_shift(order) ((order) + PAGE_SHIFT) argument 103 #define napot_cont_size(order) BIT(napot_cont_shift(order)) argument 104 #define napot_cont_mask(order) (~(napot_cont_size(order) - 1UL)) argument 105 #define napot_pte_num(order) BIT(order) argument
|
| /arch/arm/lib/ |
| A D | lib1funcs.S | 111 rsb \order, \order, #31 122 addhs \order, \order, #8 126 addhs \order, \order, #4 129 addhi \order, \order, #3 130 addls \order, \order, \divisor, lsr #1 143 sub \order, \order, \spare 148 mov \order, #0 157 addlo \order, \order, #4 165 addlo \order, \order, #1 172 subs \order, \order, #3 @ yes, 3 is intended here [all …]
|
| /arch/s390/mm/ |
| A D | page-states.c | 17 void arch_free_page(struct page *page, int order) in arch_free_page() argument 21 __set_page_unused(page_to_virt(page), 1UL << order); in arch_free_page() 24 void arch_alloc_page(struct page *page, int order) in arch_alloc_page() argument 29 __set_page_stable_dat(page_to_virt(page), 1UL << order); in arch_alloc_page() 31 __set_page_stable_nodat(page_to_virt(page), 1UL << order); in arch_alloc_page()
|
| A D | init.c | 75 unsigned int order; in setup_zero_pages() local 78 order = 7; in setup_zero_pages() 81 while (order > 2 && (total_pages >> 10) < (1UL << order)) in setup_zero_pages() 82 order--; in setup_zero_pages() 84 empty_zero_page = (unsigned long)memblock_alloc_or_panic(PAGE_SIZE << order, PAGE_SIZE); in setup_zero_pages() 86 zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; in setup_zero_pages()
|
| /arch/s390/include/asm/ |
| A D | sigp.h | 43 static inline int ____pcpu_sigp(u16 addr, u8 order, unsigned long parm, in ____pcpu_sigp() argument 53 : [addr] "d" (addr), [order] "a" (order) in ____pcpu_sigp() 59 static inline int __pcpu_sigp(u16 addr, u8 order, unsigned long parm, in __pcpu_sigp() argument 65 cc = ____pcpu_sigp(addr, order, parm, &_status); in __pcpu_sigp()
|
| /arch/x86/kernel/ |
| A D | aperture_64.c | 184 old_order = *order; in read_agp() 191 *order = 7 - nbits; in read_agp() 192 if ((int)*order < 0) /* < 32MB */ in read_agp() 193 *order = 0; in read_agp() 206 if (aper + (32ULL<<(20 + *order)) > 0x100000000ULL) { in read_agp() 208 bus, slot, func, 32 << *order, apsizereg); in read_agp() 209 *order = old_order; in read_agp() 213 bus, slot, func, aper, aper + (32ULL << (*order + 20)) - 1, in read_agp() 214 32 << *order, apsizereg); in read_agp() 216 if (!aperture_valid(aper, (32*1024*1024) << *order, 32<<20)) in read_agp() [all …]
|
| A D | kvmclock.c | 205 unsigned int order; in kvmclock_init_mem() local 213 order = get_order(ncpus * sizeof(*hvclock_mem)); in kvmclock_init_mem() 215 p = alloc_pages(GFP_KERNEL, order); in kvmclock_init_mem() 217 pr_warn("%s: failed to alloc %d pages", __func__, (1U << order)); in kvmclock_init_mem() 229 1UL << order); in kvmclock_init_mem() 231 __free_pages(p, order); in kvmclock_init_mem() 238 memset(hvclock_mem, 0, PAGE_SIZE << order); in kvmclock_init_mem()
|
| /arch/powerpc/mm/ |
| A D | hugetlbpage.c | 205 unsigned long order = 0; in gigantic_hugetlb_cma_reserve() local 208 order = PUD_SHIFT - PAGE_SHIFT; in gigantic_hugetlb_cma_reserve() 213 order = mmu_psize_to_shift(MMU_PAGE_16G) - PAGE_SHIFT; in gigantic_hugetlb_cma_reserve() 215 if (order) in gigantic_hugetlb_cma_reserve() 216 hugetlb_cma_reserve(order); in gigantic_hugetlb_cma_reserve()
|
| /arch/x86/kernel/cpu/mce/ |
| A D | genpool.c | 121 int mce_numrecords, mce_poolsz, order; in mce_gen_pool_create() local 125 order = order_base_2(sizeof(struct mce_evt_llist)); in mce_gen_pool_create() 126 gpool = gen_pool_create(order, -1); in mce_gen_pool_create() 131 mce_poolsz = mce_numrecords * (1 << order); in mce_gen_pool_create()
|
| /arch/x86/include/asm/xen/ |
| A D | swiotlb-xen.h | 6 int xen_create_contiguous_region(phys_addr_t pstart, unsigned int order, 9 void xen_destroy_contiguous_region(phys_addr_t pstart, unsigned int order);
|
| /arch/sparc/kernel/ |
| A D | pci_sun4v.c | 196 order = get_order(size); in dma_4v_alloc_coherent() 255 free_pages(first_page, order); in dma_4v_alloc_coherent() 350 order = get_order(size); in dma_4v_free_coherent() 351 if (order < 10) in dma_4v_free_coherent() 747 unsigned long order; in pci_sun4v_atu_alloc_iotsb() local 759 order = get_order(table_size); in pci_sun4v_atu_alloc_iotsb() 1062 order = get_order(alloc_size); in pci_sun4v_msiq_alloc() 1066 order); in pci_sun4v_msiq_alloc() 1105 free_pages(pages, order); in pci_sun4v_msiq_alloc() 1122 order = get_order(alloc_size); in pci_sun4v_msiq_free() [all …]
|
| A D | pci_fire.c | 234 unsigned long pages, order, i; in pci_fire_msiq_alloc() local 236 order = get_order(512 * 1024); in pci_fire_msiq_alloc() 237 pages = __get_free_pages(GFP_KERNEL | __GFP_COMP, order); in pci_fire_msiq_alloc() 240 order); in pci_fire_msiq_alloc() 243 memset((char *)pages, 0, PAGE_SIZE << order); in pci_fire_msiq_alloc() 266 unsigned long pages, order; in pci_fire_msiq_free() local 268 order = get_order(512 * 1024); in pci_fire_msiq_free() 271 free_pages(pages, order); in pci_fire_msiq_free()
|
| A D | iommu.c | 97 unsigned long i, order, sz, num_tsb_entries; in iommu_table_init() local 132 order = get_order(tsbsize); in iommu_table_init() 201 unsigned long order, first_page; in dma_4u_alloc_coherent() local 209 order = get_order(size); in dma_4u_alloc_coherent() 210 if (order >= 10) in dma_4u_alloc_coherent() 214 page = alloc_pages_node(nid, gfp, order); in dma_4u_alloc_coherent() 226 free_pages(first_page, order); in dma_4u_alloc_coherent() 251 unsigned long order, npages; in dma_4u_free_coherent() local 258 order = get_order(size); in dma_4u_free_coherent() 259 if (order < 10) in dma_4u_free_coherent() [all …]
|
| /arch/powerpc/sysdev/xive/ |
| A D | native.c | 132 __be32 *qpage, u32 order, bool can_escalate) in xive_native_configure_queue() argument 140 if (order) { in xive_native_configure_queue() 148 q->msk = order ? ((1u << (order - 2)) - 1) : 0; in xive_native_configure_queue() 174 rc = opal_xive_set_queue_info(vp_id, prio, qpage_phys, order, flags); in xive_native_configure_queue() 670 u32 order; in xive_native_alloc_vp_block() local 672 order = fls(max_vcpus) - 1; in xive_native_alloc_vp_block() 673 if (max_vcpus > (1 << order)) in xive_native_alloc_vp_block() 674 order++; in xive_native_alloc_vp_block() 677 max_vcpus, order); in xive_native_alloc_vp_block() 680 rc = opal_xive_alloc_vp_block(order); in xive_native_alloc_vp_block() [all …]
|
| /arch/powerpc/kvm/ |
| A D | book3s_64_mmu_hv.c | 55 u32 order; member 80 if ((order < PPC_MIN_HPT_ORDER) || (order > PPC_MAX_HPT_ORDER)) in kvmppc_allocate_hpt() 98 npte = 1ul << (order - 4); in kvmppc_allocate_hpt() 106 free_pages(hpt, order - PAGE_SHIFT); in kvmppc_allocate_hpt() 110 info->order = order; in kvmppc_allocate_hpt() 149 if (kvm->arch.hpt.order == order) { in kvmppc_alloc_reset_hpt() 191 info->order = 0; in kvmppc_free_hpt() 1323 BUG_ON(new->order >= old->order); in resize_hpt_rehash_hpte() 1433 resize->order); in resize_hpt_prepare_work() 1476 if (resize->order == shift) { in kvm_vm_ioctl_resize_hpt_prepare() [all …]
|
| /arch/parisc/kernel/ |
| A D | pci-dma.c | 406 int order; in arch_dma_alloc() local 411 order = get_order(size); in arch_dma_alloc() 412 size = 1 << (order + PAGE_SHIFT); in arch_dma_alloc() 414 paddr = __get_free_pages(gfp | __GFP_ZERO, order); in arch_dma_alloc() 426 int order = get_order(size); in arch_dma_free() local 431 size = 1 << (order + PAGE_SHIFT); in arch_dma_free() 435 free_pages((unsigned long)__va(dma_handle), order); in arch_dma_free()
|
| /arch/mips/kernel/ |
| A D | irq.c | 56 unsigned int order = get_order(IRQ_STACK_SIZE); in init_IRQ() local 67 void *s = (void *)__get_free_pages(GFP_KERNEL, order); in init_IRQ()
|
| /arch/m68k/ifpsp060/src/ |
| A D | README-SRC | 6 assembler, however it is being included in order to comply with the 9 You don't need to actually assemble these files in order to compile a
|
| /arch/um/include/shared/ |
| A D | kern_util.h | 21 extern unsigned long alloc_stack(int order, int atomic); 22 extern void free_stack(unsigned long stack, int order);
|
| /arch/um/kernel/ |
| A D | process.c | 49 void free_stack(unsigned long stack, int order) in free_stack() argument 51 free_pages(stack, order); in free_stack() 54 unsigned long alloc_stack(int order, int atomic) in alloc_stack() argument 61 page = __get_free_pages(flags, order); in alloc_stack()
|