| /include/linux/ |
| A D | highmem.h | 395 VM_BUG_ON(dst_off + len > PAGE_SIZE || src_off + len > PAGE_SIZE); in memcpy_page() 414 chunk = PAGE_SIZE - offset_in_page(dst_off); in memcpy_folio() 417 chunk = PAGE_SIZE - offset_in_page(src_off); in memcpy_folio() 433 VM_BUG_ON(offset + len > PAGE_SIZE); in memset_page() 443 VM_BUG_ON(offset + len > PAGE_SIZE); in memcpy_from_page() 453 VM_BUG_ON(offset + len > PAGE_SIZE); in memcpy_to_page() 463 VM_BUG_ON(offset + len > PAGE_SIZE); in memzero_page() 487 chunk = PAGE_SIZE - offset_in_page(offset); in memcpy_from_folio() 515 chunk = PAGE_SIZE - offset_in_page(offset); in memcpy_to_folio() 554 max = PAGE_SIZE; in folio_zero_tail() [all …]
|
| A D | pfn.h | 9 #define PFN_ALIGN(x) (((unsigned long)(x) + (PAGE_SIZE - 1)) & PAGE_MASK) 10 #define PFN_UP(x) (((x) + PAGE_SIZE-1) >> PAGE_SHIFT)
|
| A D | mm_types_task.h | 38 #if (BITS_PER_LONG > 32) || (PAGE_SIZE >= 65536) 58 #if (PAGE_SIZE < PAGE_FRAG_CACHE_MAX_SIZE) && (BITS_PER_LONG <= 32)
|
| A D | bvec.h | 62 unsigned long nr = offset / PAGE_SIZE; in bvec_set_folio() 65 bvec_set_page(bv, folio_page(folio, nr), len, offset % PAGE_SIZE); in bvec_set_folio() 115 (mp_bvec_iter_offset((bvec), (iter)) / PAGE_SIZE) 126 (mp_bvec_iter_offset((bvec), (iter)) % PAGE_SIZE) 130 PAGE_SIZE - bvec_iter_offset((bvec), (iter))) 225 bv->bv_len = min_t(unsigned int, PAGE_SIZE - bv->bv_offset, in bvec_advance()
|
| A D | ndctl.h | 19 ND_MIN_NAMESPACE_SIZE = PAGE_SIZE,
|
| A D | execmem.h | 12 #define MODULE_ALIGN (PAGE_SIZE << KASAN_SHADOW_SCALE_SHIFT) 14 #define MODULE_ALIGN PAGE_SIZE
|
| A D | highmem-internal.h | 86 struct page *page = folio_page(folio, offset / PAGE_SIZE); in kmap_local_folio() 87 return __kmap_local_page_prot(page, kmap_prot) + offset % PAGE_SIZE; in kmap_local_folio() 214 kunmap_flush_on_unmap(PTR_ALIGN_DOWN(addr, PAGE_SIZE)); in __kunmap_local() 241 kunmap_flush_on_unmap(PTR_ALIGN_DOWN(addr, PAGE_SIZE)); in __kunmap_atomic()
|
| A D | threads.h | 34 #define PID_MAX_LIMIT (IS_ENABLED(CONFIG_BASE_SMALL) ? PAGE_SIZE * 8 : \
|
| A D | iov_iter.h | 118 void *kaddr = kmap_local_page(p->bv_page + offset / PAGE_SIZE); in iterate_bvec() 122 (size_t)(PAGE_SIZE - offset % PAGE_SIZE)); in iterate_bvec() 123 remain = step(kaddr + offset % PAGE_SIZE, progress, part, priv, priv2); in iterate_bvec() 172 part = umin(len, PAGE_SIZE - skip % PAGE_SIZE); in iterate_folioq() 208 pgoff_t index = start / PAGE_SIZE; in iterate_xarray() 229 PAGE_SIZE - offset_in_page(offset)); in iterate_xarray()
|
| A D | nfsacl.h | 19 #define NFSACL_MAXPAGES ((2*(8+12*NFS_ACL_MAX_ENTRIES) + PAGE_SIZE-1) \
|
| A D | page_frag_cache.h | 11 #if (PAGE_SIZE < PAGE_FRAG_CACHE_MAX_SIZE)
|
| A D | io-mapping.h | 176 return io_mapping_map_wc(mapping, offset, PAGE_SIZE); in io_mapping_map_atomic_wc() 193 return io_mapping_map_wc(mapping, offset, PAGE_SIZE); in io_mapping_map_local_wc()
|
| A D | lp.h | 31 #define LP_BUFFER_SIZE PAGE_SIZE
|
| /include/vdso/ |
| A D | page.h | 15 #define PAGE_SIZE (_AC(1,UL) << CONFIG_PAGE_SHIFT) macro 28 #define PAGE_MASK (~(PAGE_SIZE - 1))
|
| A D | datapage.h | 176 #define VDSO_ARCH_DATA_SIZE ALIGN(sizeof(struct vdso_arch_data), PAGE_SIZE) 208 #define __vdso_u_rng_data PROVIDE(vdso_u_rng_data = vdso_u_data + 2 * PAGE_SIZE); 214 #define __vdso_u_arch_data PROVIDE(vdso_u_arch_data = vdso_u_data + 3 * PAGE_SIZE); 220 PROVIDE(vdso_u_data = . - __VDSO_PAGES * PAGE_SIZE); \
|
| /include/asm-generic/ |
| A D | tlb.h | 213 ((PAGE_SIZE - sizeof(struct mmu_table_batch)) / sizeof(void *)) 279 ((PAGE_SIZE - sizeof(struct mmu_gather_batch)) / sizeof(void *)) 503 return tlb_remove_page_size(tlb, page, PAGE_SIZE); in tlb_remove_page() 642 tlb_flush_pte_range(tlb, address, PAGE_SIZE); \ 656 tlb_flush_pte_range(tlb, address, PAGE_SIZE * nr); in tlb_remove_tlb_entries() 662 address += PAGE_SIZE; in tlb_remove_tlb_entries() 729 tlb_flush_pmd_range(tlb, address, PAGE_SIZE); \ 738 tlb_flush_pud_range(tlb, address, PAGE_SIZE); \ 747 tlb_flush_p4d_range(tlb, address, PAGE_SIZE); \ 756 __tlb_adjust_range(tlb, address, PAGE_SIZE); \
|
| A D | shmparam.h | 5 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */
|
| A D | pgalloc.h | 159 BUG_ON((unsigned long)pmd & (PAGE_SIZE-1)); in pmd_free() 207 BUG_ON((unsigned long)pud & (PAGE_SIZE-1)); in __pud_free() 252 BUG_ON((unsigned long)p4d & (PAGE_SIZE-1)); in __p4d_free() 288 BUG_ON((unsigned long)pgd & (PAGE_SIZE-1)); in __pgd_free()
|
| A D | vmlinux.lds.h | 382 . = ALIGN(PAGE_SIZE); \ 385 . = ALIGN(PAGE_SIZE); \ 596 . = ALIGN(PAGE_SIZE); \ 670 . = ALIGN(PAGE_SIZE); \ 674 . = ALIGN(PAGE_SIZE); \ 769 . = ALIGN(PAGE_SIZE); \ 771 . = ALIGN(PAGE_SIZE); \ 990 . = ALIGN(PAGE_SIZE); \ 992 . = ALIGN(PAGE_SIZE); 1070 . = ALIGN(PAGE_SIZE); \ [all …]
|
| /include/linux/raid/ |
| A D | pq.h | 41 #ifndef PAGE_SIZE 42 # define PAGE_SIZE 4096 macro 47 extern const char raid6_empty_zero_page[PAGE_SIZE]; 179 # define __get_free_pages(x, y) ((unsigned long)mmap(NULL, PAGE_SIZE << (y), \ 183 # define free_pages(x, y) munmap((void *)(x), PAGE_SIZE << (y))
|
| /include/crypto/ |
| A D | if_alg.h | 190 return max_t(int, max_t(int, sk->sk_sndbuf & PAGE_MASK, PAGE_SIZE) - in af_alg_sndbuf() 202 return PAGE_SIZE <= af_alg_sndbuf(sk); in af_alg_writable() 216 return max_t(int, max_t(int, sk->sk_rcvbuf & PAGE_MASK, PAGE_SIZE) - in af_alg_rcvbuf() 228 return PAGE_SIZE <= af_alg_rcvbuf(sk); in af_alg_readable()
|
| A D | scatterwalk.h | 129 limit = PAGE_SIZE - offset_in_page(walk->offset); in scatterwalk_clamp() 131 limit = PAGE_SIZE; in scatterwalk_clamp() 260 PAGE_SIZE - 1) >> PAGE_SHIFT; in scatterwalk_done_dst()
|
| /include/uapi/linux/ |
| A D | binfmts.h | 15 #define MAX_ARG_STRLEN (PAGE_SIZE * 32)
|
| /include/xen/ |
| A D | page.h | 23 #define XEN_PFN_PER_PAGE (PAGE_SIZE / XEN_PAGE_SIZE)
|
| /include/crypto/internal/ |
| A D | rsa.h | 77 if (err > PAGE_SIZE) in rsa_set_key()
|