| /mm/ |
| A D | swap_cgroup.c | 19 struct swap_cgroup *map; member 68 struct swap_cgroup *map; in swap_cgroup_record() local 74 map = swap_cgroup_ctrl[swp_type(ent)].map; in swap_cgroup_record() 96 struct swap_cgroup *map; in swap_cgroup_clear() local 101 map = swap_cgroup_ctrl[swp_type(ent)].map; in swap_cgroup_clear() 132 struct swap_cgroup *map; in swap_cgroup_swapon() local 142 if (!map) in swap_cgroup_swapon() 147 ctrl->map = map; in swap_cgroup_swapon() 167 map = ctrl->map; in swap_cgroup_swapoff() 168 ctrl->map = NULL; in swap_cgroup_swapoff() [all …]
|
| A D | vma.c | 2328 map->vm_flags = ksm_vma_flags(map->mm, map->file, map->vm_flags); in update_ksm_flags() 2369 map->next = vma_iter_next_rewind(vmi, &map->prev); in __mmap_prepare() 2373 if (!may_expand_vm(map->mm, map->vm_flags, map->pglen - vms->nr_pages)) in __mmap_prepare() 2377 if (accountable_mapping(map->file, map->vm_flags)) { in __mmap_prepare() 2378 map->charged = map->pglen; in __mmap_prepare() 2420 unmap_region(&vmi->mas, vma, map->prev, map->next); in __mmap_new_file_vma() 2465 vma_iter_config(vmi, map->addr, map->end); in __mmap_new_vma() 2466 vma_set_range(vma, map->addr, map->end, map->pgoff); in __mmap_new_vma() 2649 map.check_ksm_early = can_set_ksm_flags_early(&map); in __mmap_region() 2661 if (map.prev || map.next) { in __mmap_region() [all …]
|
| A D | hmm.c | 653 map->pfn_list = kvcalloc(nr_entries, sizeof(*map->pfn_list), in hmm_dma_map_alloc() 655 if (!map->pfn_list) in hmm_dma_map_alloc() 661 map->dma_list = kvcalloc(nr_entries, sizeof(*map->dma_list), in hmm_dma_map_alloc() 663 if (!map->dma_list) in hmm_dma_map_alloc() 669 kvfree(map->pfn_list); in hmm_dma_map_alloc() 683 if (dma_use_iova(&map->state)) in hmm_dma_map_free() 684 dma_iova_free(dev, &map->state); in hmm_dma_map_free() 685 kvfree(map->pfn_list); in hmm_dma_map_free() 686 kvfree(map->dma_list); in hmm_dma_map_free() 712 unsigned long *pfns = map->pfn_list; in hmm_dma_map_pfn() [all …]
|
| A D | sparse.c | 185 bitmap_set(map, idx, end - idx + 1); in subsection_mask_set() 422 struct page *map = sparse_buffer_alloc(size); in __populate_section_memmap() local 425 if (map) in __populate_section_memmap() 426 return map; in __populate_section_memmap() 429 if (!map) in __populate_section_memmap() 433 return map; in __populate_section_memmap() 540 struct page *map; in sparse_init_nid() local 562 if (!map) { in sparse_init_nid() 570 sparse_init_early_section(nid, map, pnum, 0); in sparse_init_nid() 702 subsection_mask_set(map, pfn, nr_pages); in clear_subsection_map() [all …]
|
| A D | swapfile.c | 199 if (*map != count) in swap_is_last_map() 1576 *map = 0; in swap_entries_free() 1661 unsigned char *map; in swp_swapcount() local 1686 kunmap_local(map); in swp_swapcount() 3763 count = *map; in add_swap_count_continuation() 3764 kunmap_local(map); in add_swap_count_continuation() 3801 unsigned char *map; in swap_count_continued() local 3838 *map += 1; in swap_count_continued() 3839 kunmap_local(map); in swap_count_continued() 3859 *map -= 1; in swap_count_continued() [all …]
|
| A D | memblock.c | 2426 map->start = start; in reserved_mem_add() 2427 map->size = size; in reserved_mem_add() 2438 if (!map->size) in reserve_mem_find_by_name_nolock() 2441 return map; in reserve_mem_find_by_name_nolock() 2462 if (!map) in reserve_mem_find_by_name() 2466 *size = map->size; in reserve_mem_find_by_name() 2488 if (!map) in reserve_mem_release_by_name() 2495 map->size = 0; in reserve_mem_release_by_name() 2513 err |= kho_preserve_phys(map->start, map->size); in reserve_mem_kho_finalize() 2560 err |= fdt_property(fdt, "start", &map->start, sizeof(map->start)); in prepare_kho_fdt() [all …]
|
| A D | hugetlb_vmemmap.c | 794 void *map; in hugetlb_vmemmap_init_early() local 817 map = pfn_to_page(pfn); in hugetlb_vmemmap_init_early() 818 start = (unsigned long)map; in hugetlb_vmemmap_init_early() 831 sparse_init_early_section(nid, map, pnum, in hugetlb_vmemmap_init_early() 833 map += section_map_size(); in hugetlb_vmemmap_init_early() 847 void *map; in hugetlb_vmemmap_init_late() local 871 map = pfn_to_page(pfn); in hugetlb_vmemmap_init_late() 873 start = (unsigned long)map; in hugetlb_vmemmap_init_late()
|
| A D | shrinker.c | 208 set_bit(shrinker_id_to_offset(shrinker_id), unit->map); in set_shrinker_bit() 527 for_each_set_bit(offset, unit->map, SHRINKER_UNIT_BITS) { in shrink_slab_memcg() 539 clear_bit(offset, unit->map); in shrink_slab_memcg() 552 clear_bit(offset, unit->map); in shrink_slab_memcg()
|
| A D | mm_init.c | 1651 struct page *map; in alloc_node_mem_map() local 1666 map = memmap_alloc(size, SMP_CACHE_BYTES, MEMBLOCK_LOW_LIMIT, in alloc_node_mem_map() 1668 if (!map) in alloc_node_mem_map() 1671 pgdat->node_mem_map = map + offset; in alloc_node_mem_map()
|
| A D | percpu.c | 1161 static unsigned long pcpu_find_zero_area(unsigned long *map, in pcpu_find_zero_area() argument 1171 index = find_next_zero_bit(map, size, start); in pcpu_find_zero_area() 1180 i = find_next_bit(map, end, index); in pcpu_find_zero_area()
|
| A D | hugetlb.c | 578 hugetlb_resv_map_add(struct resv_map *map, struct list_head *rg, long from, in hugetlb_resv_map_add() argument 585 nrg = get_file_region_entry_from_cache(map, from, to); in hugetlb_resv_map_add() 586 record_hugetlb_cgroup_uncharge_info(cg, h, map, nrg); in hugetlb_resv_map_add() 588 coalesce_file_region(map, nrg); in hugetlb_resv_map_add() 1193 static void set_vma_resv_map(struct vm_area_struct *vma, struct resv_map *map) in set_vma_resv_map() argument 1198 set_vma_private_data(vma, (unsigned long)map); in set_vma_resv_map()
|
| A D | Kconfig | 765 Programs which use vm86 functionality or have some need to map
|