| /linux/mm/ |
| A D | memremap.c | 102 return (range->start + range_len(range)) >> PAGE_SHIFT; in pfn_end() 122 PHYS_PFN(range_len(range))); in pageunmap_range() 125 PHYS_PFN(range_len(range)), NULL); in pageunmap_range() 127 arch_remove_memory(range->start, range_len(range), in pageunmap_range() 129 kasan_remove_zero_shadow(__va(range->start), range_len(range)); in pageunmap_range() 196 is_ram = region_intersects(range->start, range_len(range), in pagemap_range() 215 range_len(range)); in pagemap_range() 239 PHYS_PFN(range_len(range)), params); in pagemap_range() 247 error = arch_add_memory(nid, range->start, range_len(range), in pagemap_range() 256 PHYS_PFN(range_len(range)), params->altmap, in pagemap_range() [all …]
|
| /linux/drivers/comedi/ |
| A D | range.c | 110 int chan, range_len, i; in comedi_check_chanlist() local 116 range_len = s->range_table->length; in comedi_check_chanlist() 118 range_len = s->range_table_list[chan]->length; in comedi_check_chanlist() 120 range_len = 0; in comedi_check_chanlist() 122 CR_RANGE(chanspec) >= range_len) { in comedi_check_chanlist() 125 i, chanspec, chan, range_len); in comedi_check_chanlist()
|
| /linux/drivers/dax/ |
| A D | kmem.c | 106 total_len += range_len(&range); in dev_dax_kmem_probe() 139 res = request_mem_region(range.start, range_len(&range), data->res_name); in dev_dax_kmem_probe() 171 range_len(&range), kmem_name, mhp_flags); in dev_dax_kmem_probe() 223 rc = remove_memory(range.start, range_len(&range)); in dev_dax_kmem_remove()
|
| A D | bus.c | 195 size += range_len(&dev_dax->ranges[i].range); in dev_dax_size() 452 __release_region(&dax_region->res, range->start, range_len(range)); in trim_dev_dax_range() 650 || !IS_ALIGNED(range_len(range), align)) in alloc_dax_region() 873 pgoff += PHYS_PFN(range_len(&ranges[i].range)); in alloc_dev_dax_range() 968 shrink = min_t(u64, to_shrink, range_len(range)); in dev_dax_shrink() 969 if (shrink >= range_len(range)) { in dev_dax_shrink() 989 return adjust_dev_dax_range(dev_dax, adjust, range_len(range) in dev_dax_shrink() 1191 to_alloc = range_len(&r); in mapping_store() 1215 size_t len = range_len(&dev_dax->ranges[i].range); in dev_dax_validate_align() 1516 if (dev_dax->nr_range && range_len(&dev_dax->ranges[0].range)) { in __devm_create_dev_dax()
|
| A D | cxl.c | 28 .size = range_len(&cxlr_dax->hpa_range), in cxl_dax_region_probe()
|
| A D | pmem.c | 65 .size = range_len(&range), in __dax_pmem_probe()
|
| A D | device.c | 65 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys() 437 range_len(range), dev_name(dev))) { in dev_dax_probe()
|
| /linux/fs/btrfs/ |
| A D | defrag.c | 975 u64 range_len; in defrag_collect_targets() local 1009 range_len = em->len - (cur - em->start); in defrag_collect_targets() 1080 range_len = min(extent_map_end(em), start + len) - cur; in defrag_collect_targets() 1093 last->len += range_len; in defrag_collect_targets() 1107 new->len = range_len; in defrag_collect_targets() 1290 u32 range_len = entry->len; in defrag_one_cluster() local 1299 range_len = min_t(u32, range_len, in defrag_one_cluster() 1308 if (entry->start + range_len <= *last_scanned_ret) in defrag_one_cluster() 1313 ((entry->start + range_len - 1) >> PAGE_SHIFT) - in defrag_one_cluster() 1321 ret = defrag_one_range(inode, entry->start, range_len, in defrag_one_cluster() [all …]
|
| A D | ordered-data.c | 749 u64 range_start, range_len; in btrfs_wait_ordered_extents() local 754 range_len = bg->length; in btrfs_wait_ordered_extents() 757 range_len = U64_MAX; in btrfs_wait_ordered_extents() 759 range_end = range_start + range_len; in btrfs_wait_ordered_extents()
|
| A D | file.c | 3166 const u64 range_len = last_byte - cur_offset; in btrfs_fallocate() local 3168 ret = add_falloc_range(&reserve_list, cur_offset, range_len); in btrfs_fallocate() 3174 &data_reserved, cur_offset, range_len); in btrfs_fallocate() 3179 qgroup_reserved += range_len; in btrfs_fallocate() 3180 data_space_needed += range_len; in btrfs_fallocate()
|
| /linux/drivers/net/wireless/intel/iwlwifi/fw/ |
| A D | dbg.c | 521 u32 range_len, void *ptr) in iwl_dump_prph() argument 537 for (i = 0; i < range_len; i++) { in iwl_dump_prph() 600 u32 range_len, void *ptr) in iwl_fw_get_prph_len() argument 608 for (i = 0; i < range_len; i++) { in iwl_fw_get_prph_len() 625 u32 range_len; in iwl_fw_prph_handler() local 1055 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_prph_mac_block_iter() argument 1129 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_prph_phy_block_iter() argument 1142 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_csr_iter() argument 1343 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_txf_iter() argument 1525 void *range_ptr, u32 range_len, int idx) in iwl_dump_ini_rxf_iter() argument [all …]
|
| /linux/include/linux/ |
| A D | range.h | 11 static inline u64 range_len(const struct range *range) in range_len() function
|
| /linux/tools/testing/nvdimm/ |
| A D | dax-dev.c | 20 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys()
|
| /linux/drivers/nvdimm/ |
| A D | badrange.c | 237 len = range->start + range_len(range) in badblocks_populate() 252 len = range_len(range); in badblocks_populate()
|
| A D | pmem.c | 524 range_len(&pmem->pgmap.range); in pmem_attach_disk()
|
| /linux/drivers/virt/nitro_enclaves/ |
| A D | ne_misc_dev_test.c | 137 KUNIT_EXPECT_EQ(test, range_len(&phys_contig_mem_regions.regions[num - 1]), in ne_misc_dev_test_merge_phys_contig_memory_regions()
|
| A D | ne_misc_dev.c | 1006 u64 phys_region_size = range_len(&phys_contig_mem_regions.regions[i]); in ne_set_user_memory_region_ioctl() 1024 slot_add_mem_req.size = range_len(&phys_contig_mem_regions.regions[i]); in ne_set_user_memory_region_ioctl()
|
| /linux/drivers/dax/hmem/ |
| A D | hmem.c | 38 .size = region_idle ? 0 : range_len(&mri->range), in dax_hmem_probe()
|
| /linux/drivers/gpu/drm/nouveau/ |
| A D | nouveau_dmem.c | 299 release_mem_region(chunk->pagemap.range.start, range_len(&chunk->pagemap.range)); in nouveau_dmem_chunk_alloc() 376 unsigned long i, npages = range_len(&chunk->pagemap.range) >> PAGE_SHIFT; in nouveau_dmem_evict_chunk() 434 range_len(&chunk->pagemap.range)); in nouveau_dmem_fini()
|
| /linux/Documentation/translations/zh_CN/mm/ |
| A D | hmm.rst | 246 release_mem_region(pagemap.range.start, range_len(&pagemap.range));
|
| /linux/tools/testing/cxl/test/ |
| A D | cxl.c | 396 range_len(&res->range)); in depopulate_all_mock_resources() 1017 if (range_len(&ram_range)) in mock_cxl_endpoint_parse_cdat() 1020 if (range_len(&pmem_range)) in mock_cxl_endpoint_parse_cdat()
|
| /linux/lib/ |
| A D | test_hmm.c | 562 pfn_last = pfn_first + (range_len(&devmem->pagemap.range) >> PAGE_SHIFT); in dmirror_allocate_chunk() 593 range_len(&devmem->pagemap.range)); in dmirror_allocate_chunk() 1284 range_len(&devmem->pagemap.range)); in dmirror_device_remove_chunks()
|
| /linux/arch/powerpc/kvm/ |
| A D | book3s_hv_uvmem.c | 704 (range_len(&kvmppc_uvmem_pgmap.range) >> PAGE_SHIFT); in kvmppc_uvmem_get_page() 1221 range_len(&kvmppc_uvmem_pgmap.range)); in kvmppc_uvmem_free()
|
| /linux/drivers/cxl/core/ |
| A D | hdm.c | 672 size = range_len(&cxld->hpa_range); in cxl_decoder_commit() 804 len = range_len(&info->dvsec_range[which]); in cxl_setup_hdm_decoder_from_dvsec()
|
| /linux/drivers/pci/ |
| A D | p2pdma.c | 341 range_len(&pgmap->range), dev_to_node(&pdev->dev), in pci_p2pdma_add_resource()
|