Searched refs:PAGE_ALIGN (Results 1 – 25 of 135) sorted by relevance
123456
100 size_t size = PAGE_ALIGN(real_mode_blob_end - real_mode_blob); in setup_real_mode()191 size_t size = PAGE_ALIGN(real_mode_blob_end - real_mode_blob); in set_real_mode_permissions()194 PAGE_ALIGN(real_mode_header->ro_end) - in set_real_mode_permissions()198 PAGE_ALIGN(real_mode_header->ro_end) - in set_real_mode_permissions()
49 #define TASK_UNMAPPED_BASE_USER32 (PAGE_ALIGN(TASK_SIZE_USER32 / 4))50 #define TASK_UNMAPPED_BASE_USER64 (PAGE_ALIGN(DEFAULT_MAP_WINDOW_USER64 / 4))
124 offset = PAGE_ALIGN((unsigned long)(&_end)) - loadbase; in node_getfirstfree()128 return KDM_TO_PHYS(PAGE_ALIGN(SYMMON_STK_ADDR(nasid, 0))) >> PAGE_SHIFT; in node_getfirstfree()
71 pages = PAGE_ALIGN(size) / PAGE_SIZE; in omap_sram_push()124 pages = PAGE_ALIGN(omap_sram_size) / PAGE_SIZE; in omap_detect_and_map_sram()
53 vdso_info.size = PAGE_ALIGN(vdso_end - vdso_start); in init_vdso()71 base = PAGE_ALIGN(base); in vdso_base()
105 romfs_size = PAGE_ALIGN(get_romfs_len((unsigned *)romfs_base)); in machine_early_init()108 romfs_size = PAGE_ALIGN(get_romfs_len((unsigned *)romfs_base)); in machine_early_init()
33 return PAGE_ALIGN(size); in gcs_size()36 size = PAGE_ALIGN(min_t(unsigned long long, in gcs_size()96 alloc_size = PAGE_ALIGN(size); in SYSCALL_DEFINE3()
65 return PAGE_ALIGN(STACK_TOP - gap - rnd); in mmap_base()94 addr = PAGE_ALIGN(addr); in arch_get_unmapped_area()132 addr = PAGE_ALIGN(addr); in arch_get_unmapped_area_topdown()
21 unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave()
22 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave()
130 return PAGE_ALIGN(size); in adjust_shstk_size()132 return PAGE_ALIGN(min_t(unsigned long long, rlimit(RLIMIT_STACK), SZ_4G)); in adjust_shstk_size()311 need_to_check_vma = PAGE_ALIGN(*ssp) == *ssp; in shstk_pop_sigframe()528 aligned_size = PAGE_ALIGN(size); in SYSCALL_DEFINE3()
64 addr = PAGE_ALIGN(addr); in arch_get_unmapped_area()115 addr = PAGE_ALIGN(addr); in arch_get_unmapped_area_topdown()
544 .size = PAGE_ALIGN(size), in __dma_alloc()613 .size = PAGE_ALIGN(size), in __arm_dma_free()770 count = PAGE_ALIGN(size) >> PAGE_SHIFT; in __alloc_iova()1010 size = PAGE_ALIGN((iova & ~PAGE_MASK) + size); in __iommu_remove_mapping()1084 size = PAGE_ALIGN(size); in arm_iommu_alloc_attrs()1148 size = PAGE_ALIGN(size); in arm_iommu_free_attrs()1196 size = PAGE_ALIGN(size); in __map_sg_chunk()1374 int ret, prot, len = PAGE_ALIGN(size + offset); in arm_iommu_map_page()1412 int len = PAGE_ALIGN(size + offset); in arm_iommu_unmap_page()1442 size_t len = PAGE_ALIGN(size + offset); in arm_iommu_map_resource()[all …]
52 #define TASK_UNMAPPED_BASE PAGE_ALIGN((UL(1) << MMAP_MIN_VA_BITS) / 3)54 #define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3)
25 int num_pages = PAGE_ALIGN(end - start) >> PAGE_SHIFT; in set_kernel_memory()
57 cur = start + PAGE_ALIGN(size); in __pkvm_alloc_private_va_range()100 size = PAGE_ALIGN(size + offset_in_page(phys)); in __pkvm_create_private_mapping()123 end = PAGE_ALIGN(end); in pkvm_create_mappings_locked()165 end = PAGE_ALIGN((u64)hyp_phys_to_page(end)); in hyp_back_vmemmap()
28 #define PAGE_ALIGN(addr) _ALIGN(addr, PAGE_SIZE) macro
54 #define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE/3))
42 # define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE / 3))
84 addr = PAGE_ALIGN(addr); in arch_get_unmapped_area()133 addr = PAGE_ALIGN(addr); in arch_get_unmapped_area_topdown()
106 start = PAGE_ALIGN(start); in vdso_addr()131 return PAGE_ALIGN(size); in vdso_text_size()
26 size = PAGE_ALIGN(addr + size) - paligned; in __ioremap_caller()
93 pages = PAGE_ALIGN(size) / PAGE_SIZE; in omap_sram_push()217 pages = PAGE_ALIGN(omap_sram_size) / PAGE_SIZE; in omap2_map_sram()
62 addr = PAGE_ALIGN(addr); in arch_get_unmapped_area()
372 size = PAGE_ALIGN(image_data.size); in free_image_buf()409 image_data.data = vzalloc(PAGE_ALIGN(image_data.size)); in alloc_image_buf()417 size = PAGE_ALIGN(image_data.size); in alloc_image_buf()
Completed in 39 milliseconds