/linux-6.3-rc2/drivers/iommu/iommufd/ |
A D | pages.c | 620 for (; pages != end; pages++) in batch_from_pages() 878 pages->last_npinned = pages->npinned; in do_update_pinned() 890 if (pages->npinned == pages->last_npinned) in update_unpinned() 892 do_update_pinned(pages, pages->last_npinned - pages->npinned, false, in update_unpinned() 911 if (pages->npinned == pages->last_npinned) in pfn_reader_user_update_pinned() 961 struct iopt_pages *pages = pfns->pages; in pfn_reader_unpin() local 1069 pfns->pages = pages; in pfn_reader_init() 1093 struct iopt_pages *pages = pfns->pages; in pfn_reader_release_pins() local 1113 struct iopt_pages *pages = pfns->pages; in pfn_reader_destroy() local 1905 struct iopt_pages *pages = area->pages; in iopt_area_add_access() local [all …]
|
A D | io_pagetable.c | 356 elm->area->pages = elm->pages; in iopt_map_pages() 439 elm->pages = area->pages; in iopt_get_pages() 504 pages = area->pages; in iopt_unmap_iova_range() 695 struct iopt_pages *pages = area->pages; in iopt_unfill_domain() local 714 struct iopt_pages *pages = area->pages; in iopt_unfill_domain() local 716 if (!pages) in iopt_unfill_domain() 748 struct iopt_pages *pages = area->pages; in iopt_fill_domain() local 773 struct iopt_pages *pages = area->pages; in iopt_fill_domain() local 983 struct iopt_pages *pages = area->pages; in iopt_area_split() local 1044 lhs->pages = area->pages; in iopt_area_split() [all …]
|
/linux-6.3-rc2/net/ceph/ |
A D | pagevec.c | 19 set_page_dirty_lock(pages[i]); in ceph_put_page_vector() 20 put_page(pages[i]); in ceph_put_page_vector() 22 kvfree(pages); in ceph_put_page_vector() 31 __free_pages(pages[i], 0); in ceph_release_page_vector() 32 kfree(pages); in ceph_release_page_vector() 41 struct page **pages; in ceph_alloc_page_vector() local 44 pages = kmalloc_array(num_pages, sizeof(*pages), flags); in ceph_alloc_page_vector() 45 if (!pages) in ceph_alloc_page_vector() 49 if (pages[i] == NULL) { in ceph_alloc_page_vector() 50 ceph_release_page_vector(pages, i); in ceph_alloc_page_vector() [all …]
|
/linux-6.3-rc2/mm/ |
A D | percpu-vm.c | 34 static struct page **pages; in pcpu_get_pages() local 39 if (!pages) in pcpu_get_pages() 41 return pages; in pcpu_get_pages() 165 pages[pcpu_page_idx(cpu, i)] = page; in pcpu_unmap_pages() 197 PAGE_KERNEL, pages, PAGE_SHIFT); in __pcpu_map_pages() 279 struct page **pages; in pcpu_populate_chunk() local 281 pages = pcpu_get_pages(); in pcpu_populate_chunk() 282 if (!pages) in pcpu_populate_chunk() 315 struct page **pages; in pcpu_depopulate_chunk() local 322 pages = pcpu_get_pages(); in pcpu_depopulate_chunk() [all …]
|
A D | gup_test.c | 19 put_page(pages[i]); in put_back_pages() 32 put_page(pages[i]); in put_back_pages() 50 page = pages[i]; in verify_dma_pinned() 105 struct page **pages; in __gup_test_ioctl() local 115 if (!pages) in __gup_test_ioctl() 139 pages + i); in __gup_test_ioctl() 147 pages + i); in __gup_test_ioctl() 203 kvfree(pages); in __gup_test_ioctl() 228 struct page **pages; in pin_longterm_test_start() local 250 if (!pages) in pin_longterm_test_start() [all …]
|
A D | gup.c | 1103 pages ? &pages[i] : NULL); in __get_user_pages() 1169 if (pages) { in __get_user_pages() 1179 if (pages) { in __get_user_pages() 1415 pages += ret; in __get_user_pages_locked() 1463 pages++; in __get_user_pages_locked() 1710 if (pages) { in __get_user_pages_locked() 1712 if (pages[i]) in __get_user_pages_locked() 1995 pages[i] = NULL; in migrate_longterm_unpinnable_pages() 2018 if (pages[i]) in migrate_longterm_unpinnable_pages() 2053 pages); in check_and_migrate_movable_pages() [all …]
|
A D | mincore.c | 196 end = min(vma->vm_end, addr + (pages << PAGE_SHIFT)); in do_mincore() 198 unsigned long pages = DIV_ROUND_UP(end - addr, PAGE_SIZE); in do_mincore() local 199 memset(vec, 1, pages); in do_mincore() 200 return pages; in do_mincore() 236 unsigned long pages; in SYSCALL_DEFINE3() local 250 pages = len >> PAGE_SHIFT; in SYSCALL_DEFINE3() 251 pages += (offset_in_page(len)) != 0; in SYSCALL_DEFINE3() 253 if (!access_ok(vec, pages)) in SYSCALL_DEFINE3() 261 while (pages) { in SYSCALL_DEFINE3() 267 retval = do_mincore(start, min(pages, PAGE_SIZE), tmp); in SYSCALL_DEFINE3() [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/i915/gem/selftests/ |
A D | huge_gem_object.c | 12 struct sg_table *pages) in huge_free_pages() argument 24 sg_free_table(pages); in huge_free_pages() 25 kfree(pages); in huge_free_pages() 34 struct sg_table *pages; in huge_get_pages() local 41 pages = kmalloc(sizeof(*pages), GFP); in huge_get_pages() 42 if (!pages) in huge_get_pages() 46 kfree(pages); in huge_get_pages() 50 sg = pages->sgl; in huge_get_pages() 68 src = pages->sgl; in huge_get_pages() 80 huge_free_pages(obj, pages); in huge_get_pages() [all …]
|
/linux-6.3-rc2/fs/isofs/ |
A D | compress.c | 68 if (!pages[i]) in zisofs_uncompress_block() 121 if (pages[curpage]) { in zisofs_uncompress_block() 174 if (pages[curpage]) { in zisofs_uncompress_block() 290 if (poffset && *pages) { in zisofs_fill_pages() 313 struct page **pages; in zisofs_read_folio() local 340 if (!pages) { in zisofs_read_folio() 344 pages[full_page] = page; in zisofs_read_folio() 349 if (pages[i]) in zisofs_read_folio() 357 if (pages[i]) { in zisofs_read_folio() 363 put_page(pages[i]); in zisofs_read_folio() [all …]
|
/linux-6.3-rc2/fs/erofs/ |
A D | pcpubuf.c | 15 struct page **pages; member 67 pages = kmalloc_array(nrpages, sizeof(*pages), GFP_KERNEL); in erofs_pcpubuf_growsize() 68 if (!pages) { in erofs_pcpubuf_growsize() 75 if (!pages[i]) { in erofs_pcpubuf_growsize() 77 oldpages = pages; in erofs_pcpubuf_growsize() 84 oldpages = pages; in erofs_pcpubuf_growsize() 91 pcb->pages = pages; in erofs_pcpubuf_growsize() 139 if (!pcb->pages) in erofs_pcpubuf_exit() 143 if (pcb->pages[i]) in erofs_pcpubuf_exit() 145 kfree(pcb->pages); in erofs_pcpubuf_exit() [all …]
|
/linux-6.3-rc2/drivers/xen/ |
A D | xlate_mmu.c | 71 struct page **pages; member 148 struct page **pages) in xen_xlate_remap_gfn_array() argument 163 data.pages = pages; in xen_xlate_remap_gfn_array() 217 struct page **pages; in xen_xlate_map_ballooned_pages() local 226 pages = kcalloc(nr_pages, sizeof(pages[0]), GFP_KERNEL); in xen_xlate_map_ballooned_pages() 227 if (!pages) in xen_xlate_map_ballooned_pages() 232 kfree(pages); in xen_xlate_map_ballooned_pages() 239 kfree(pages); in xen_xlate_map_ballooned_pages() 253 kfree(pages); in xen_xlate_map_ballooned_pages() 257 kfree(pages); in xen_xlate_map_ballooned_pages() [all …]
|
/linux-6.3-rc2/kernel/dma/ |
A D | remap.c | 15 return area->pages; in dma_common_find_pages() 22 void *dma_common_pages_remap(struct page **pages, size_t size, in dma_common_pages_remap() argument 27 vaddr = vmap(pages, PAGE_ALIGN(size) >> PAGE_SHIFT, in dma_common_pages_remap() 30 find_vm_area(vaddr)->pages = pages; in dma_common_pages_remap() 42 struct page **pages; in dma_common_contiguous_remap() local 46 pages = kmalloc_array(count, sizeof(struct page *), GFP_KERNEL); in dma_common_contiguous_remap() 47 if (!pages) in dma_common_contiguous_remap() 50 pages[i] = nth_page(page, i); in dma_common_contiguous_remap() 51 vaddr = vmap(pages, count, VM_DMA_COHERENT, prot); in dma_common_contiguous_remap() 52 kfree(pages); in dma_common_contiguous_remap()
|
/linux-6.3-rc2/drivers/gpu/drm/xen/ |
A D | xen_drm_front_gem.c | 29 struct page **pages; member 55 kvfree(xen_obj->pages); in gem_free_pages_array() 56 xen_obj->pages = NULL; in gem_free_pages_array() 176 if (IS_ERR(xen_obj->pages)) { in gem_create() 178 xen_obj->pages = NULL; in gem_create() 209 if (xen_obj->pages) { in xen_drm_front_gem_free_object_unlocked() 212 xen_obj->pages); in xen_drm_front_gem_free_object_unlocked() 228 return xen_obj->pages; in xen_drm_front_gem_get_pages() 235 if (!xen_obj->pages) in xen_drm_front_gem_get_sg_table() 271 xen_obj->pages); in xen_drm_front_gem_import_sg_table() [all …]
|
/linux-6.3-rc2/drivers/staging/media/ipu3/ |
A D | ipu3-dmamap.c | 26 __free_page(pages[count]); in imgu_dmamap_free_buffer() 27 kvfree(pages); in imgu_dmamap_free_buffer() 36 struct page **pages; in imgu_dmamap_alloc_buffer() local 42 pages = kvmalloc_array(count, sizeof(*pages), GFP_KERNEL); in imgu_dmamap_alloc_buffer() 44 if (!pages) in imgu_dmamap_alloc_buffer() 77 pages[i++] = page++; in imgu_dmamap_alloc_buffer() 80 return pages; in imgu_dmamap_alloc_buffer() 100 struct page **pages; in imgu_dmamap_alloc() local 113 if (!pages) in imgu_dmamap_alloc() 131 map->pages = pages; in imgu_dmamap_alloc() [all …]
|
/linux-6.3-rc2/drivers/media/common/videobuf2/ |
A D | frame_vector.c | 80 struct page **pages; in put_vaddr_frames() local 84 pages = frame_vector_pages(vec); in put_vaddr_frames() 90 if (WARN_ON(IS_ERR(pages))) in put_vaddr_frames() 93 unpin_user_pages(pages, vec->nr_frames); in put_vaddr_frames() 112 struct page **pages; in frame_vector_to_pages() local 120 pages = (struct page **)nums; in frame_vector_to_pages() 122 pages[i] = pfn_to_page(nums[i]); in frame_vector_to_pages() 138 struct page **pages; in frame_vector_to_pfns() local 142 pages = (struct page **)(vec->ptrs); in frame_vector_to_pfns() 143 nums = (unsigned long *)pages; in frame_vector_to_pfns() [all …]
|
/linux-6.3-rc2/Documentation/admin-guide/mm/ |
A D | hugetlbpage.rst | 30 and surplus huge pages in the pool of huge pages of default size. 46 is the size of the pool of huge pages. 69 pages of all sizes. 80 pages in the kernel's huge page pool. "Persistent" huge pages will be 171 default sized persistent huge pages:: 205 huge pages can grow, if more huge pages than ``/proc/sys/vm/nr_hugepages`` are 213 surplus pages will first be promoted to persistent huge pages. Then, additional 226 of the in-use huge pages to surplus huge pages. This will occur even if 247 pages may exist:: 270 pages size are allowed. [all …]
|
A D | zswap.rst | 8 Zswap is a lightweight compressed cache for swap pages. It takes pages that are 40 When zswap is disabled at runtime it will stop storing pages that are 42 back into memory all of the pages stored in the compressed pool. The 46 fault back into memory all swapped out pages, including those in the 60 pages are freed. The pool is not preallocated. By default, a zpool 107 Some of the pages in zswap are same-value filled pages (i.e. contents of the 123 checking for the same-value filled pages during store operation. 125 However, the existing pages which are marked as same-value filled pages remain 141 pages by zswap. 157 of pages stored, same-value filled pages and various counters for the reasons [all …]
|
/linux-6.3-rc2/include/linux/ |
A D | balloon_compaction.h | 57 struct list_head pages; /* Pages enqueued & handled to Host */ member 67 struct list_head *pages); 69 struct list_head *pages, size_t n_req_pages); 75 INIT_LIST_HEAD(&balloon->pages); in balloon_devinfo_init() 97 list_add(&page->lru, &balloon->pages); in balloon_page_insert() 141 list_add(&page->lru, &balloon->pages); in balloon_page_insert() 164 static inline void balloon_page_push(struct list_head *pages, struct page *page) in balloon_page_push() argument 166 list_add(&page->lru, pages); in balloon_page_push() 176 static inline struct page *balloon_page_pop(struct list_head *pages) in balloon_page_pop() argument 178 struct page *page = list_first_entry_or_null(pages, struct page, lru); in balloon_page_pop()
|
/linux-6.3-rc2/net/rds/ |
A D | info.c | 65 struct page **pages; member 122 iter->addr = kmap_atomic(*iter->pages); in rds_info_copy() 127 "bytes %lu\n", *iter->pages, iter->addr, in rds_info_copy() 140 iter->pages++; in rds_info_copy() 166 struct page **pages = NULL; in rds_info_getsockopt() local 191 if (!pages) { in rds_info_getsockopt() 195 ret = pin_user_pages_fast(start, nr_pages, FOLL_WRITE, pages); in rds_info_getsockopt() 214 iter.pages = pages; in rds_info_getsockopt() 237 if (pages) in rds_info_getsockopt() 238 unpin_user_pages(pages, nr_pages); in rds_info_getsockopt() [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/ |
A D | drm_scatter.c | 51 for (i = 0; i < entry->pages; i++) { in drm_sg_cleanup() 83 unsigned long pages, i, j; in drm_legacy_sg_alloc() local 103 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_legacy_sg_alloc() 104 DRM_DEBUG("size=%ld pages=%ld\n", request->size, pages); in drm_legacy_sg_alloc() 106 entry->pages = pages; in drm_legacy_sg_alloc() 107 entry->pagelist = kcalloc(pages, sizeof(*entry->pagelist), GFP_KERNEL); in drm_legacy_sg_alloc() 113 entry->busaddr = kcalloc(pages, sizeof(*entry->busaddr), GFP_KERNEL); in drm_legacy_sg_alloc() 120 entry->virtual = vmalloc_32(pages << PAGE_SHIFT); in drm_legacy_sg_alloc() 131 memset(entry->virtual, 0, pages << PAGE_SHIFT); in drm_legacy_sg_alloc() 138 for (i = (unsigned long)entry->virtual, j = 0; j < pages; in drm_legacy_sg_alloc() [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_pages.c | 19 struct sg_table *pages) in __i915_gem_object_set_pages() argument 36 drm_clflush_sg(pages); in __i915_gem_object_set_pages() 45 obj->mm.pages = pages; in __i915_gem_object_set_pages() 212 pages = fetch_and_zero(&obj->mm.pages); in __i915_gem_object_unset_pages() 214 return pages; in __i915_gem_object_unset_pages() 232 return pages; in __i915_gem_object_unset_pages() 309 pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); in i915_gem_object_map_page() 310 if (!pages) in i915_gem_object_map_page() 316 pages[i++] = page; in i915_gem_object_map_page() 318 if (pages != stack) in i915_gem_object_map_page() [all …]
|
A D | i915_gem_phys.c | 101 struct sg_table *pages) in i915_gem_object_put_pages_phys() argument 103 dma_addr_t dma = sg_dma_address(pages->sgl); in i915_gem_object_put_pages_phys() 104 void *vaddr = sg_page(pages->sgl); in i915_gem_object_put_pages_phys() 136 sg_free_table(pages); in i915_gem_object_put_pages_phys() 137 kfree(pages); in i915_gem_object_put_pages_phys() 197 struct sg_table *pages; in i915_gem_object_shmem_to_phys() local 200 pages = __i915_gem_object_unset_pages(obj); in i915_gem_object_shmem_to_phys() 209 if (!IS_ERR_OR_NULL(pages)) in i915_gem_object_shmem_to_phys() 210 i915_gem_object_put_pages_shmem(obj, pages); in i915_gem_object_shmem_to_phys() 216 if (!IS_ERR_OR_NULL(pages)) in i915_gem_object_shmem_to_phys() [all …]
|
/linux-6.3-rc2/arch/m68k/mm/ |
A D | sun3kmap.c | 49 unsigned long type, int pages) in do_pmeg_mapin() argument 55 while(pages) { in do_pmeg_mapin() 59 pages--; in do_pmeg_mapin() 68 int pages; in sun3_ioremap() local 87 pages = size / PAGE_SIZE; in sun3_ioremap() 91 while(pages) { in sun3_ioremap() 95 if(seg_pages > pages) in sun3_ioremap() 96 seg_pages = pages; in sun3_ioremap() 100 pages -= seg_pages; in sun3_ioremap()
|
/linux-6.3-rc2/fs/squashfs/ |
A D | file_direct.c | 33 int i, n, pages, bytes, res = -ENOMEM; in squashfs_readpage_block() local 41 pages = end_index - start_index + 1; in squashfs_readpage_block() 43 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); in squashfs_readpage_block() 64 pages = i; in squashfs_readpage_block() 70 actor = squashfs_page_actor_init_special(msblk, page, pages, expected); in squashfs_readpage_block() 89 if (page[pages - 1]->index == end_index && bytes) { in squashfs_readpage_block() 90 pageaddr = kmap_local_page(page[pages - 1]); in squashfs_readpage_block() 96 for (i = 0; i < pages; i++) { in squashfs_readpage_block() 112 for (i = 0; i < pages; i++) { in squashfs_readpage_block()
|
A D | page_actor.c | 32 if (actor->next_page == actor->pages) in cache_next_page() 44 int pages, int length) in squashfs_page_actor_init() argument 51 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init() 53 actor->pages = pages; in squashfs_page_actor_init() 70 if ((actor->next_page == actor->pages) || in handle_next_page() 106 struct page **page, int pages, int length) in squashfs_page_actor_init_special() argument 123 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init_special() 125 actor->pages = pages; in squashfs_page_actor_init_special()
|