Home
last modified time | relevance | path

Searched refs:page_count (Results 1 – 25 of 136) sorted by relevance

123456

/linux/drivers/media/pci/ivtv/
A Divtv-udma.c102 if (dma->SG_length || dma->page_count) { in ivtv_udma_setup()
104 dma->SG_length, dma->page_count); in ivtv_udma_setup()
110 if (user_dma.page_count <= 0) { in ivtv_udma_setup()
120 if (user_dma.page_count != err) { in ivtv_udma_setup()
122 err, user_dma.page_count); in ivtv_udma_setup()
130 dma->page_count = user_dma.page_count; in ivtv_udma_setup()
135 dma->page_count = 0; in ivtv_udma_setup()
141 dma->page_count, DMA_TO_DEVICE); in ivtv_udma_setup()
150 return dma->page_count; in ivtv_udma_setup()
160 if (dma->page_count == 0) in ivtv_udma_unmap()
[all …]
A Divtv-yuv.c54 if (dma->SG_length || dma->page_count) { in ivtv_yuv_prep_user_dma()
57 dma->SG_length, dma->page_count); in ivtv_yuv_prep_user_dma()
68 if (y_pages == y_dma.page_count) { in ivtv_yuv_prep_user_dma()
70 uv_dma.page_count, &dma->map[y_pages], in ivtv_yuv_prep_user_dma()
74 if (y_pages != y_dma.page_count || uv_pages != uv_dma.page_count) { in ivtv_yuv_prep_user_dma()
77 if (y_pages == y_dma.page_count) { in ivtv_yuv_prep_user_dma()
80 uv_pages, uv_dma.page_count); in ivtv_yuv_prep_user_dma()
91 y_pages, y_dma.page_count); in ivtv_yuv_prep_user_dma()
107 dma->page_count = y_pages + uv_pages; in ivtv_yuv_prep_user_dma()
113 dma->page_count = 0; in ivtv_yuv_prep_user_dma()
[all …]
/linux/tools/testing/selftests/vm/
A Dksm_tests.c278 if (assert_ksm_pages_count(page_count)) { in check_ksm_merge()
280 munmap(map_ptr, page_size * page_count); in check_ksm_merge()
286 munmap(map_ptr, page_size * page_count); in check_ksm_merge()
294 int page_count = 2; in check_ksm_unmerge() local
326 munmap(map_ptr, page_size * page_count); in check_ksm_unmerge()
365 munmap(map_ptr, page_size * page_count); in check_ksm_zero_page_merge()
370 munmap(map_ptr, page_size * page_count); in check_ksm_zero_page_merge()
400 int page_count = 2; in check_ksm_numa_merge() local
602 size_t page_count = 4000; in ksm_cow_time() local
685 page_count = atol(optarg); in main()
[all …]
/linux/drivers/gpu/drm/nouveau/
A Dnouveau_bo90b5.c40 u32 page_count = new_reg->num_pages; in nvc0_bo_move_copy() local
43 page_count = new_reg->num_pages; in nvc0_bo_move_copy()
44 while (page_count) { in nvc0_bo_move_copy()
45 int line_count = (page_count > 8191) ? 8191 : page_count; in nvc0_bo_move_copy()
61 page_count -= line_count; in nvc0_bo_move_copy()
A Dnouveau_bo85b5.c47 u32 page_count = new_reg->num_pages; in nva3_bo_move_copy() local
50 page_count = new_reg->num_pages; in nva3_bo_move_copy()
51 while (page_count) { in nva3_bo_move_copy()
52 int line_count = (page_count > 8191) ? 8191 : page_count; in nva3_bo_move_copy()
68 page_count -= line_count; in nva3_bo_move_copy()
A Dnouveau_bo9039.c45 u32 page_count = new_reg->num_pages; in nvc0_bo_move_m2mf() local
48 page_count = new_reg->num_pages; in nvc0_bo_move_m2mf()
49 while (page_count) { in nvc0_bo_move_m2mf()
50 int line_count = (page_count > 2047) ? 2047 : page_count; in nvc0_bo_move_m2mf()
78 page_count -= line_count; in nvc0_bo_move_m2mf()
A Dnouveau_bo0039.c55 u32 page_count = new_reg->num_pages; in nv04_bo_move_m2mf() local
65 page_count = new_reg->num_pages; in nv04_bo_move_m2mf()
66 while (page_count) { in nv04_bo_move_m2mf()
67 int line_count = (page_count > 2047) ? 2047 : page_count; in nv04_bo_move_m2mf()
88 page_count -= line_count; in nv04_bo_move_m2mf()
/linux/drivers/firewire/
A Dcore-iso.c33 buffer->page_count = 0; in fw_iso_buffer_alloc()
40 for (i = 0; i < page_count; i++) { in fw_iso_buffer_alloc()
45 buffer->page_count = i; in fw_iso_buffer_alloc()
46 if (i < page_count) { in fw_iso_buffer_alloc()
62 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_map_dma()
71 if (i < buffer->page_count) in fw_iso_buffer_map_dma()
78 int page_count, enum dma_data_direction direction) in fw_iso_buffer_init() argument
82 ret = fw_iso_buffer_alloc(buffer, page_count); in fw_iso_buffer_init()
105 for (i = 0; i < buffer->page_count; i++) in fw_iso_buffer_destroy()
110 buffer->page_count = 0; in fw_iso_buffer_destroy()
[all …]
/linux/drivers/char/agp/
A Dgeneric.c181 if (curr->page_count != 0) { in agp_free_memory()
186 for (i = 0; i < curr->page_count; i++) { in agp_free_memory()
219 size_t page_count, u32 type) in agp_allocate_memory() argument
231 (cur_memory + page_count < page_count)) in agp_allocate_memory()
264 for (i = 0; i < page_count; i++) { in agp_allocate_memory()
272 new->page_count++; in agp_allocate_memory()
1038 if (mem->page_count == 0) in agp_generic_insert_memory()
1117 if (mem->page_count == 0) in agp_generic_remove_memory()
1170 for (i = 0; i < page_count; i++) in agp_generic_alloc_user()
1172 new->page_count = 0; in agp_generic_alloc_user()
[all …]
A Di460-agp.c311 if ((io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count) > num_entries) { in i460_insert_memory_small_io_page()
317 while (j < (io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count)) { in i460_insert_memory_small_io_page()
327 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { in i460_insert_memory_small_io_page()
346 for (i = pg_start; i < (pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count); i++) in i460_remove_memory_small_io_page()
415 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; in i460_insert_memory_large_io_page()
417 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; in i460_insert_memory_large_io_page()
473 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; in i460_remove_memory_large_io_page()
475 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; in i460_remove_memory_large_io_page()
A Dnvidia-agp.c210 if (mem->page_count == 0) in nvidia_insert_memory()
213 if ((pg_start + mem->page_count) > in nvidia_insert_memory()
217 for (j = pg_start; j < (pg_start + mem->page_count); j++) { in nvidia_insert_memory()
226 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in nvidia_insert_memory()
250 if (mem->page_count == 0) in nvidia_remove_memory()
253 for (i = pg_start; i < (mem->page_count + pg_start); i++) in nvidia_remove_memory()
A Dintel-gtt.c127 DBG("try unmapping %lu pages\n", (unsigned long)mem->page_count); in intel_gtt_unmap_memory()
217 if ((pg_start + mem->page_count) in i810_insert_dcache_entries()
224 for (i = pg_start; i < (pg_start + mem->page_count); i++) { in i810_insert_dcache_entries()
269 new->page_count = pg_count; in alloc_agpphysmem_i8xx()
280 if (curr->page_count == 4) in intel_i810_free_by_type()
914 if (mem->page_count == 0) in intel_fake_agp_insert_entries()
932 ret = intel_gtt_map_memory(mem->pages, mem->page_count, &st); in intel_fake_agp_insert_entries()
940 intel_gtt_insert_pages(pg_start, mem->page_count, mem->pages, in intel_fake_agp_insert_entries()
967 if (mem->page_count == 0) in intel_fake_agp_remove_entries()
970 intel_gtt_clear_range(pg_start, mem->page_count); in intel_fake_agp_remove_entries()
[all …]
A Dati-agp.c282 if (mem->page_count == 0) in ati_insert_memory()
285 if ((pg_start + mem->page_count) > num_entries) in ati_insert_memory()
289 while (j < (pg_start + mem->page_count)) { in ati_insert_memory()
303 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in ati_insert_memory()
328 if (mem->page_count == 0) in ati_remove_memory()
331 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in ati_remove_memory()
A Duninorth-agp.c165 if (mem->page_count == 0) in uninorth_insert_memory()
171 if ((pg_start + mem->page_count) > num_entries) in uninorth_insert_memory()
175 for (i = 0; i < mem->page_count; ++i) { in uninorth_insert_memory()
184 for (i = 0; i < mem->page_count; i++) { in uninorth_insert_memory()
214 if (mem->page_count == 0) in uninorth_remove_memory()
218 for (i = 0; i < mem->page_count; ++i) { in uninorth_remove_memory()
A Defficeon-agp.c238 int i, count = mem->page_count, num_entries; in efficeon_insert_memory()
246 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory()
287 int i, count = mem->page_count, num_entries; in efficeon_remove_memory()
293 if ((pg_start + mem->page_count) > num_entries) in efficeon_remove_memory()
A Dagp.h204 struct agp_memory *agp_generic_alloc_by_type(size_t page_count, int type);
208 struct agp_memory *memory, size_t page_count);
225 struct agp_memory *agp_generic_alloc_user(size_t page_count, int type);
/linux/arch/x86/hyperv/
A Dhv_proc.c29 int i, j, page_count; in hv_call_deposit_pages() local
87 for (i = 0, page_count = 0; i < num_allocations; ++i) { in hv_call_deposit_pages()
89 for (j = 0; j < counts[i]; ++j, ++page_count) in hv_call_deposit_pages()
90 input_page->gpa_page_list[page_count] = base_pfn + j; in hv_call_deposit_pages()
93 page_count, 0, input_page, NULL); in hv_call_deposit_pages()
/linux/drivers/target/
A Dtarget_core_rd.c68 u32 i, j, page_count = 0, sg_per_table; in rd_release_sgl_table() local
78 page_count++; in rd_release_sgl_table()
85 return page_count; in rd_release_sgl_table()
90 u32 page_count; in rd_release_device_space() local
95 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_table_array, in rd_release_device_space()
100 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, in rd_release_device_space()
101 rd_dev->sg_table_count, (unsigned long)page_count * PAGE_SIZE); in rd_release_device_space()
216 u32 page_count; in rd_release_prot_space() local
221 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_prot_array, in rd_release_prot_space()
226 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, in rd_release_prot_space()
[all …]
/linux/fs/btrfs/
A Dscrub.c96 int page_count; member
103 int page_count; member
595 sbio->page_count = 0; in scrub_setup_ctx()
1368 sblock->page_count++; in scrub_setup_recheck_block()
1663 if (sbio->page_count == 0) { in scrub_add_page_to_wr_bio()
1698 if (sbio->page_count < 1) { in scrub_add_page_to_wr_bio()
1710 sbio->page_count++; in scrub_add_page_to_wr_bio()
2096 if (sbio->page_count == 0) { in scrub_add_page_to_rd_bio()
2137 sbio->page_count++; in scrub_add_page_to_rd_bio()
2317 sblock->page_count++; in scrub_pages()
[all …]
/linux/fs/squashfs/
A Dblock.c85 const int page_count = DIV_ROUND_UP(total_len + offset, PAGE_SIZE); in squashfs_bio_read() local
89 if (page_count <= BIO_MAX_VECS) in squashfs_bio_read()
90 bio = bio_alloc(GFP_NOIO, page_count); in squashfs_bio_read()
92 bio = bio_kmalloc(GFP_NOIO, page_count); in squashfs_bio_read()
101 for (i = 0; i < page_count; ++i) { in squashfs_bio_read()
/linux/fs/orangefs/
A Dorangefs-bufmap.c152 int page_count; member
171 unpin_user_pages(bufmap->page_array, bufmap->page_count); in orangefs_bufmap_unmap()
240 bufmap->page_count = bufmap->total_size / PAGE_SIZE; in orangefs_bufmap_alloc()
244 kcalloc(bufmap->page_count, sizeof(struct page *), GFP_KERNEL); in orangefs_bufmap_alloc()
269 bufmap->page_count, FOLL_WRITE, bufmap->page_array); in orangefs_bufmap_map()
274 if (ret != bufmap->page_count) { in orangefs_bufmap_map()
276 bufmap->page_count, ret); in orangefs_bufmap_map()
291 for (i = 0; i < bufmap->page_count; i++) in orangefs_bufmap_map()
/linux/drivers/virt/vboxguest/
A Dvboxguest_utils.c199 u32 page_count; in hgcm_call_add_pagelist_size() local
201 page_count = hgcm_call_buf_size_in_pages(buf, len); in hgcm_call_add_pagelist_size()
202 *extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]); in hgcm_call_add_pagelist_size()
337 u32 i, page_count; in hgcm_call_init_linaddr() local
348 page_count = hgcm_call_buf_size_in_pages(buf, len); in hgcm_call_init_linaddr()
356 dst_pg_lst->page_count = page_count; in hgcm_call_init_linaddr()
358 for (i = 0; i < page_count; i++) { in hgcm_call_init_linaddr()
368 *off_extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]); in hgcm_call_init_linaddr()
/linux/tools/testing/radix-tree/
A Dregression2.c61 unsigned long page_count = 0; variable
71 p->index = page_count++; in page_alloc()
/linux/drivers/gpu/drm/
A Ddrm_bufs.c863 dma->page_count += byte_count >> PAGE_SHIFT; in drm_legacy_addbufs_agp()
898 int page_count; in drm_legacy_addbufs_pci() local
966 temp_pagelist = kmalloc_array(dma->page_count + (count << page_order), in drm_legacy_addbufs_pci()
977 dma->pagelist, dma->page_count * sizeof(*dma->pagelist)); in drm_legacy_addbufs_pci()
979 dma->page_count + (count << page_order)); in drm_legacy_addbufs_pci()
984 page_count = 0; in drm_legacy_addbufs_pci()
1019 dma->page_count + page_count, in drm_legacy_addbufs_pci()
1021 temp_pagelist[dma->page_count + page_count++] in drm_legacy_addbufs_pci()
1080 if (dma->page_count) { in drm_legacy_addbufs_pci()
1087 dma->page_count += entry->seg_count << page_order; in drm_legacy_addbufs_pci()
[all …]
/linux/fs/pstore/
A Dram_core.c408 unsigned int page_count; in persistent_ram_vmap() local
414 page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE); in persistent_ram_vmap()
431 pages = kmalloc_array(page_count, sizeof(struct page *), GFP_KERNEL); in persistent_ram_vmap()
434 __func__, page_count); in persistent_ram_vmap()
438 for (i = 0; i < page_count; i++) { in persistent_ram_vmap()
442 vaddr = vmap(pages, page_count, VM_MAP, prot); in persistent_ram_vmap()

Completed in 43 milliseconds

123456