/linux-6.3-rc2/mm/ |
A D | memremap.c | 101 return (range->start + range_len(range)) >> PAGE_SHIFT; in pfn_end() 121 PHYS_PFN(range_len(range))); in pageunmap_range() 124 PHYS_PFN(range_len(range)), NULL); in pageunmap_range() 126 arch_remove_memory(range->start, range_len(range), in pageunmap_range() 128 kasan_remove_zero_shadow(__va(range->start), range_len(range)); in pageunmap_range() 195 is_ram = region_intersects(range->start, range_len(range), in pagemap_range() 214 range_len(range)); in pagemap_range() 238 PHYS_PFN(range_len(range)), params); in pagemap_range() 246 error = arch_add_memory(nid, range->start, range_len(range), in pagemap_range() 255 PHYS_PFN(range_len(range)), params->altmap, in pagemap_range() [all …]
|
/linux-6.3-rc2/drivers/comedi/ |
A D | range.c | 110 int chan, range_len, i; in comedi_check_chanlist() local 116 range_len = s->range_table->length; in comedi_check_chanlist() 118 range_len = s->range_table_list[chan]->length; in comedi_check_chanlist() 120 range_len = 0; in comedi_check_chanlist() 122 CR_RANGE(chanspec) >= range_len) { in comedi_check_chanlist() 125 i, chanspec, chan, range_len); in comedi_check_chanlist()
|
/linux-6.3-rc2/drivers/dax/ |
A D | kmem.c | 83 total_len += range_len(&range); in dev_dax_kmem_probe() 116 res = request_mem_region(range.start, range_len(&range), data->res_name); in dev_dax_kmem_probe() 144 range_len(&range), kmem_name, MHP_NID_IS_MGID); in dev_dax_kmem_probe() 196 rc = remove_memory(range.start, range_len(&range)); in dev_dax_kmem_remove()
|
A D | bus.c | 186 size += range_len(&dev_dax->ranges[i].range); in dev_dax_size() 424 __release_region(&dax_region->res, range->start, range_len(range)); in trim_dev_dax_range() 603 || !IS_ALIGNED(range_len(range), align)) in alloc_dax_region() 830 pgoff += PHYS_PFN(range_len(&ranges[i].range)); in alloc_dev_dax_range() 923 shrink = min_t(u64, to_shrink, range_len(range)); in dev_dax_shrink() 924 if (shrink >= range_len(range)) { in dev_dax_shrink() 944 return adjust_dev_dax_range(dev_dax, adjust, range_len(range) in dev_dax_shrink() 1131 to_alloc = range_len(&r); in mapping_store() 1155 size_t len = range_len(&dev_dax->ranges[i].range); in dev_dax_validate_align() 1409 if (dev_dax->nr_range && range_len(&dev_dax->ranges[0].range)) { in devm_create_dev_dax()
|
A D | cxl.c | 29 .size = range_len(&cxlr_dax->hpa_range), in cxl_dax_region_probe()
|
A D | pmem.c | 66 .size = range_len(&range), in __dax_pmem_probe()
|
A D | device.c | 65 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys() 443 range_len(range), dev_name(dev))) { in dev_dax_probe()
|
/linux-6.3-rc2/fs/btrfs/ |
A D | defrag.c | 838 u64 range_len; in defrag_collect_targets() local 873 range_len = em->len - (cur - em->start); in defrag_collect_targets() 906 if (range_len >= extent_thresh) in defrag_collect_targets() 957 last->len += range_len; in defrag_collect_targets() 971 new->len = range_len; in defrag_collect_targets() 1155 u32 range_len = entry->len; in defrag_one_cluster() local 1164 range_len = min_t(u32, range_len, in defrag_one_cluster() 1173 if (entry->start + range_len <= *last_scanned_ret) in defrag_one_cluster() 1179 ((entry->start + range_len - 1) >> PAGE_SHIFT) - in defrag_one_cluster() 1187 ret = defrag_one_range(inode, entry->start, range_len, in defrag_one_cluster() [all …]
|
A D | ordered-data.h | 202 const u64 range_start, const u64 range_len); 204 const u64 range_start, const u64 range_len);
|
A D | ordered-data.c | 628 const u64 range_start, const u64 range_len) in btrfs_wait_ordered_extents() argument 636 const u64 range_end = range_start + range_len; in btrfs_wait_ordered_extents() 684 const u64 range_start, const u64 range_len) in btrfs_wait_ordered_roots() argument 705 range_start, range_len); in btrfs_wait_ordered_roots()
|
A D | file.c | 3140 const u64 range_len = last_byte - cur_offset; in btrfs_fallocate() local 3142 ret = add_falloc_range(&reserve_list, cur_offset, range_len); in btrfs_fallocate() 3148 &data_reserved, cur_offset, range_len); in btrfs_fallocate() 3153 qgroup_reserved += range_len; in btrfs_fallocate() 3154 data_space_needed += range_len; in btrfs_fallocate()
|
/linux-6.3-rc2/drivers/net/wireless/intel/iwlwifi/fw/ |
A D | dbg.c | 522 u32 range_len, void *ptr) in iwl_dump_prph() argument 538 for (i = 0; i < range_len; i++) { in iwl_dump_prph() 601 u32 range_len, void *ptr) in iwl_fw_get_prph_len() argument 609 for (i = 0; i < range_len; i++) { in iwl_fw_get_prph_len() 626 u32 range_len; in iwl_fw_prph_handler() local 1027 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_prph_mac_iter() argument 1102 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_csr_iter() argument 1307 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_txf_iter() argument 1442 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_rxf_iter() argument 1598 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_imr_iter() argument [all …]
|
/linux-6.3-rc2/tools/testing/nvdimm/ |
A D | dax-dev.c | 20 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys()
|
/linux-6.3-rc2/include/linux/ |
A D | range.h | 11 static inline u64 range_len(const struct range *range) in range_len() function
|
/linux-6.3-rc2/drivers/nvdimm/ |
A D | badrange.c | 237 len = range->start + range_len(range) in badblocks_populate() 252 len = range_len(range); in badblocks_populate()
|
A D | pmem.c | 514 range_len(&pmem->pgmap.range); in pmem_attach_disk()
|
/linux-6.3-rc2/drivers/virt/nitro_enclaves/ |
A D | ne_misc_dev_test.c | 137 KUNIT_EXPECT_EQ(test, range_len(&phys_contig_mem_regions.regions[num - 1]), in ne_misc_dev_test_merge_phys_contig_memory_regions()
|
A D | ne_misc_dev.c | 1006 u64 phys_region_size = range_len(&phys_contig_mem_regions.regions[i]); in ne_set_user_memory_region_ioctl() 1024 slot_add_mem_req.size = range_len(&phys_contig_mem_regions.regions[i]); in ne_set_user_memory_region_ioctl()
|
/linux-6.3-rc2/drivers/dax/hmem/ |
A D | hmem.c | 39 .size = region_idle ? 0 : range_len(&mri->range), in dax_hmem_probe()
|
/linux-6.3-rc2/drivers/gpu/drm/nouveau/ |
A D | nouveau_dmem.c | 299 release_mem_region(chunk->pagemap.range.start, range_len(&chunk->pagemap.range)); in nouveau_dmem_chunk_alloc() 376 unsigned long i, npages = range_len(&chunk->pagemap.range) >> PAGE_SHIFT; in nouveau_dmem_evict_chunk() 434 range_len(&chunk->pagemap.range)); in nouveau_dmem_fini()
|
/linux-6.3-rc2/Documentation/translations/zh_CN/mm/ |
A D | hmm.rst | 246 release_mem_region(pagemap.range.start, range_len(&pagemap.range));
|
/linux-6.3-rc2/drivers/cxl/core/ |
A D | hdm.c | 624 size = range_len(&cxld->hpa_range); in cxl_decoder_commit() 724 if (!range_len(&info->dvsec_range[which])) in cxl_setup_hdm_decoder_from_dvsec() 809 if (cxled && !committed && range_len(&info->dvsec_range[which])) in init_hdm_decoder()
|
/linux-6.3-rc2/lib/ |
A D | test_hmm.c | 568 pfn_last = pfn_first + (range_len(&devmem->pagemap.range) >> PAGE_SHIFT); in dmirror_allocate_chunk() 599 range_len(&devmem->pagemap.range)); in dmirror_allocate_chunk() 1293 range_len(&devmem->pagemap.range)); in dmirror_device_remove_chunks()
|
/linux-6.3-rc2/arch/powerpc/kvm/ |
A D | book3s_hv_uvmem.c | 703 (range_len(&kvmppc_uvmem_pgmap.range) >> PAGE_SHIFT); in kvmppc_uvmem_get_page() 1220 range_len(&kvmppc_uvmem_pgmap.range)); in kvmppc_uvmem_free()
|
/linux-6.3-rc2/drivers/pci/ |
A D | p2pdma.c | 341 range_len(&pgmap->range), dev_to_node(&pdev->dev), in pci_p2pdma_add_resource()
|