Home
last modified time | relevance | path

Searched refs:regions (Results 1 – 25 of 136) sorted by relevance

123456

/drivers/mtd/chips/
A Djedec_probe.c307 .regions = {
319 .regions = {
334 .regions = {
349 .regions = {
364 .regions = {
379 .regions = {
395 .regions = {
412 .regions = {
429 .regions = {
444 .regions = {
[all …]
A Dcfi_util.c365 struct mtd_erase_region_info *regions = mtd->eraseregions; in cfi_varsize_frob() local
379 while (i < mtd->numeraseregions && ofs >= regions[i].offset) in cfi_varsize_frob()
389 if (ofs & (regions[i].erasesize-1)) in cfi_varsize_frob()
399 while (i<mtd->numeraseregions && (ofs + len) >= regions[i].offset) in cfi_varsize_frob()
407 if ((ofs + len) & (regions[i].erasesize-1)) in cfi_varsize_frob()
416 int size = regions[i].erasesize; in cfi_varsize_frob()
427 if (ofs == regions[i].offset + size * regions[i].numblocks) in cfi_varsize_frob()
/drivers/mtd/devices/
A Dmtd_intel_dg.c205 if ((nvm->regions[i].offset + nvm->regions[i].size - 1) >= from && in idg_nvm_get_region()
438 u8 id = nvm->regions[i].id; in intel_dg_nvm_init()
467 nvm->regions[i].name, in intel_dg_nvm_init()
468 nvm->regions[i].id, in intel_dg_nvm_init()
469 nvm->regions[i].offset, in intel_dg_nvm_init()
470 nvm->regions[i].size, in intel_dg_nvm_init()
568 nvm->regions[idx].id, nvm->regions[idx].name, from, len); in intel_dg_mtd_read()
607 nvm->regions[idx].id, nvm->regions[idx].name, to, len); in intel_dg_mtd_write()
726 if (invm->regions[i].name) in intel_dg_mtd_probe()
743 if (!invm->regions[i].name) in intel_dg_mtd_probe()
[all …]
/drivers/vfio/platform/
A Dvfio_platform_common.c146 if (!vdev->regions) in vfio_platform_regions_init()
153 vdev->regions[i].addr = res->start; in vfio_platform_regions_init()
154 vdev->regions[i].size = resource_size(res); in vfio_platform_regions_init()
155 vdev->regions[i].flags = 0; in vfio_platform_regions_init()
162 vdev->regions[i].flags |= in vfio_platform_regions_init()
170 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_platform_regions_init()
171 vdev->regions[i].flags |= in vfio_platform_regions_init()
187 kfree(vdev->regions); in vfio_platform_regions_init()
196 iounmap(vdev->regions[i].ioaddr); in vfio_platform_regions_cleanup()
199 kfree(vdev->regions); in vfio_platform_regions_cleanup()
[all …]
/drivers/gpu/drm/nouveau/nvkm/nvfw/
A Dacr.c130 hdr->regions.no_regions); in flcn_acr_desc_dump()
135 hdr->regions.region_props[i].start_addr); in flcn_acr_desc_dump()
137 hdr->regions.region_props[i].end_addr); in flcn_acr_desc_dump()
139 hdr->regions.region_props[i].region_id); in flcn_acr_desc_dump()
141 hdr->regions.region_props[i].read_mask); in flcn_acr_desc_dump()
143 hdr->regions.region_props[i].write_mask); in flcn_acr_desc_dump()
173 hdr->regions.no_regions); in flcn_acr_desc_v1_dump()
178 hdr->regions.region_props[i].start_addr); in flcn_acr_desc_v1_dump()
180 hdr->regions.region_props[i].end_addr); in flcn_acr_desc_v1_dump()
182 hdr->regions.region_props[i].region_id); in flcn_acr_desc_v1_dump()
[all …]
/drivers/vfio/cdx/
A Dmain.c21 if (!vdev->regions) in vfio_cdx_open_device()
27 vdev->regions[i].addr = res->start; in vfio_cdx_open_device()
28 vdev->regions[i].size = resource_size(res); in vfio_cdx_open_device()
29 vdev->regions[i].type = res->flags; in vfio_cdx_open_device()
34 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_cdx_open_device()
35 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_cdx_open_device()
36 vdev->regions[i].flags |= in vfio_cdx_open_device()
44 kfree(vdev->regions); in vfio_cdx_open_device()
45 vdev->regions = NULL; in vfio_cdx_open_device()
62 kfree(vdev->regions); in vfio_cdx_close_device()
[all …]
/drivers/vfio/fsl-mc/
A Dvfio_fsl_mc.c32 if (!vdev->regions) in vfio_fsl_mc_open_device()
36 struct resource *res = &mc_dev->regions[i]; in vfio_fsl_mc_open_device()
39 vdev->regions[i].addr = res->start; in vfio_fsl_mc_open_device()
40 vdev->regions[i].size = resource_size(res); in vfio_fsl_mc_open_device()
41 vdev->regions[i].type = mc_dev->regions[i].flags & IORESOURCE_BITS; in vfio_fsl_mc_open_device()
47 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_fsl_mc_open_device()
48 vdev->regions[i].flags |= in vfio_fsl_mc_open_device()
64 iounmap(vdev->regions[i].ioaddr); in vfio_fsl_mc_regions_cleanup()
65 kfree(vdev->regions); in vfio_fsl_mc_regions_cleanup()
255 region = &vdev->regions[index]; in vfio_fsl_mc_read()
[all …]
/drivers/md/
A Ddm-bio-prison-v1.c29 struct prison_region regions[] __counted_by(num_locks);
53 spin_lock_init(&prison->regions[i].lock); in dm_bio_prison_create()
54 prison->regions[i].cell = RB_ROOT; in dm_bio_prison_create()
184 spin_lock_irq(&prison->regions[l].lock); in bio_detain()
186 spin_unlock_irq(&prison->regions[l].lock); in bio_detain()
223 spin_lock_irq(&prison->regions[l].lock); in dm_cell_release()
225 spin_unlock_irq(&prison->regions[l].lock); in dm_cell_release()
247 spin_lock_irqsave(&prison->regions[l].lock, flags); in dm_cell_release_no_holder()
275 spin_lock_irq(&prison->regions[l].lock); in dm_cell_visit_release()
277 rb_erase(&cell->node, &prison->regions[l].cell); in dm_cell_visit_release()
[all …]
/drivers/virt/nitro_enclaves/
A Dne_misc_dev_test.c117 phys_contig_mem_regions.regions = kunit_kcalloc(test, MAX_PHYS_REGIONS, in ne_misc_dev_test_merge_phys_contig_memory_regions()
118 sizeof(*phys_contig_mem_regions.regions), in ne_misc_dev_test_merge_phys_contig_memory_regions()
120 KUNIT_ASSERT_TRUE(test, phys_contig_mem_regions.regions); in ne_misc_dev_test_merge_phys_contig_memory_regions()
135 KUNIT_EXPECT_EQ(test, phys_contig_mem_regions.regions[num - 1].start, in ne_misc_dev_test_merge_phys_contig_memory_regions()
137 KUNIT_EXPECT_EQ(test, range_len(&phys_contig_mem_regions.regions[num - 1]), in ne_misc_dev_test_merge_phys_contig_memory_regions()
141 kunit_kfree(test, phys_contig_mem_regions.regions); in ne_misc_dev_test_merge_phys_contig_memory_regions()
A Dne_misc_dev.c136 struct range *regions; member
894 phys_contig_regions->regions[num - 1].end += page_size; in ne_merge_phys_contig_memory_regions()
896 phys_contig_regions->regions[num].start = page_paddr; in ne_merge_phys_contig_memory_regions()
945 phys_contig_mem_regions.regions = kcalloc(max_nr_pages, in ne_set_user_memory_region_ioctl()
946 sizeof(*phys_contig_mem_regions.regions), in ne_set_user_memory_region_ioctl()
948 if (!phys_contig_mem_regions.regions) { in ne_set_user_memory_region_ioctl()
1005 u64 phys_region_addr = phys_contig_mem_regions.regions[i].start; in ne_set_user_memory_region_ioctl()
1023 slot_add_mem_req.paddr = phys_contig_mem_regions.regions[i].start; in ne_set_user_memory_region_ioctl()
1033 kfree(phys_contig_mem_regions.regions); in ne_set_user_memory_region_ioctl()
1046 kfree(phys_contig_mem_regions.regions); in ne_set_user_memory_region_ioctl()
[all …]
/drivers/soc/qcom/
A Dsmem.c468 header = smem->regions[0].virt_base; in qcom_smem_alloc_global()
557 header = smem->regions[0].virt_base; in qcom_smem_get_global()
565 region = &smem->regions[i]; in qcom_smem_get_global()
749 if (ret > __smem->regions[0].size) in qcom_smem_get_free_space()
795 area = &__smem->regions[i]; in qcom_smem_virt_to_phys()
868 header = smem->regions[0].virt_base; in qcom_smem_get_sbl_version()
1087 smem->regions[0].size = size; in qcom_smem_map_global()
1090 if (!smem->regions[0].virt_base) in qcom_smem_map_global()
1149 smem->regions[0].size = rmem->size; in qcom_smem_probe()
1174 smem->regions[i].size); in qcom_smem_probe()
[all …]
A Docmem.c64 struct ocmem_region *regions; member
119 struct ocmem_region *region = &ocmem->regions[i]; in update_ocmem()
131 struct ocmem_region *region = &ocmem->regions[i]; in update_ocmem()
169 struct ocmem_region *region = &ocmem->regions[i]; in update_range()
373 ocmem->regions = devm_kcalloc(dev, ocmem->config->num_regions, in ocmem_dev_probe()
375 if (!ocmem->regions) { in ocmem_dev_probe()
381 struct ocmem_region *region = &ocmem->regions[i]; in ocmem_dev_probe()
/drivers/net/dsa/sja1105/
A Dsja1105_devlink.c85 priv->regions = kcalloc(num_regions, sizeof(struct devlink_region *), in sja1105_setup_devlink_regions()
87 if (!priv->regions) in sja1105_setup_devlink_regions()
97 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_setup_devlink_regions()
99 kfree(priv->regions); in sja1105_setup_devlink_regions()
103 priv->regions[i] = region; in sja1105_setup_devlink_regions()
115 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_teardown_devlink_regions()
117 kfree(priv->regions); in sja1105_teardown_devlink_regions()
/drivers/perf/
A Dmarvell_cn10k_tad_pmu.c32 struct tad_region *regions; member
62 new += readq(tad_pmu->regions[i].base + in tad_pmu_event_counter_read()
80 writeq_relaxed(0, tad_pmu->regions[i].base + in tad_pmu_event_counter_stop()
101 writeq_relaxed(0, tad_pmu->regions[i].base + in tad_pmu_event_counter_start()
293 struct tad_region *regions; in tad_pmu_probe() local
342 regions = devm_kcalloc(&pdev->dev, tad_cnt, in tad_pmu_probe()
343 sizeof(*regions), GFP_KERNEL); in tad_pmu_probe()
344 if (!regions) in tad_pmu_probe()
349 regions[i].base = devm_ioremap(&pdev->dev, in tad_pmu_probe()
352 if (!regions[i].base) { in tad_pmu_probe()
[all …]
/drivers/net/wireless/ath/ath10k/
A Dcoredump.c1300 .regions = qca6174_hw10_mem_regions,
1309 .regions = qca6174_hw10_mem_regions,
1318 .regions = qca6174_hw10_mem_regions,
1327 .regions = qca6174_hw21_mem_regions,
1336 .regions = qca6174_hw30_mem_regions,
1345 .regions = qca6174_hw30_mem_regions,
1363 .regions = qca6174_hw30_mem_regions,
1372 .regions = qca988x_hw20_mem_regions,
1381 .regions = qca9984_hw10_mem_regions,
1390 .regions = qca9984_hw10_mem_regions,
[all …]
/drivers/gpu/drm/i915/gem/selftests/
A Di915_gem_dmabuf.c96 struct intel_memory_region *lmem = i915->mm.regions[INTEL_REGION_LMEM_0]; in igt_dmabuf_import_same_driver_lmem()
220 struct intel_memory_region **regions, in igt_dmabuf_import_same_driver() argument
234 regions, num_regions); in igt_dmabuf_import_same_driver()
278 if (obj->mm.region != i915->mm.regions[INTEL_REGION_SMEM]) { in igt_dmabuf_import_same_driver()
326 struct intel_memory_region *smem = i915->mm.regions[INTEL_REGION_SMEM]; in igt_dmabuf_import_same_driver_smem()
334 struct intel_memory_region *regions[2]; in igt_dmabuf_import_same_driver_lmem_smem() local
336 if (!i915->mm.regions[INTEL_REGION_LMEM_0]) in igt_dmabuf_import_same_driver_lmem_smem()
339 regions[0] = i915->mm.regions[INTEL_REGION_LMEM_0]; in igt_dmabuf_import_same_driver_lmem_smem()
340 regions[1] = i915->mm.regions[INTEL_REGION_SMEM]; in igt_dmabuf_import_same_driver_lmem_smem()
341 return igt_dmabuf_import_same_driver(i915, regions, 2); in igt_dmabuf_import_same_driver_lmem_smem()
/drivers/gpu/drm/nouveau/nvkm/subdev/acr/
A Dgp102.c203 desc->regions.no_regions = 2; in gp102_acr_load_setup()
204 desc->regions.region_props[0].start_addr = acr->wpr_start >> 8; in gp102_acr_load_setup()
205 desc->regions.region_props[0].end_addr = acr->wpr_end >> 8; in gp102_acr_load_setup()
206 desc->regions.region_props[0].region_id = 1; in gp102_acr_load_setup()
207 desc->regions.region_props[0].read_mask = 0xf; in gp102_acr_load_setup()
208 desc->regions.region_props[0].write_mask = 0xc; in gp102_acr_load_setup()
209 desc->regions.region_props[0].client_mask = 0x2; in gp102_acr_load_setup()
210 desc->regions.region_props[0].shadow_mem_start_addr = acr->shadow_start >> 8; in gp102_acr_load_setup()
A Dgm200.c288 desc->regions.no_regions = 2; in gm200_acr_load_setup()
289 desc->regions.region_props[0].start_addr = acr->wpr_start >> 8; in gm200_acr_load_setup()
290 desc->regions.region_props[0].end_addr = acr->wpr_end >> 8; in gm200_acr_load_setup()
291 desc->regions.region_props[0].region_id = 1; in gm200_acr_load_setup()
292 desc->regions.region_props[0].read_mask = 0xf; in gm200_acr_load_setup()
293 desc->regions.region_props[0].write_mask = 0xc; in gm200_acr_load_setup()
294 desc->regions.region_props[0].client_mask = 0x2; in gm200_acr_load_setup()
/drivers/virt/acrn/
A Dmm.c21 struct vm_memory_region_batch *regions; in modify_region() local
24 regions = kzalloc(sizeof(*regions), GFP_KERNEL); in modify_region()
25 if (!regions) in modify_region()
28 regions->vmid = vm->vmid; in modify_region()
29 regions->regions_num = 1; in modify_region()
30 regions->regions_gpa = virt_to_phys(region); in modify_region()
32 ret = hcall_set_memory_regions(virt_to_phys(regions)); in modify_region()
37 kfree(regions); in modify_region()
/drivers/bus/fsl-mc/
A Dfsl-mc-bus.c666 struct resource *regions; in fsl_mc_device_get_mmio_regions() local
685 regions = kmalloc_array(obj_desc->region_count, in fsl_mc_device_get_mmio_regions()
686 sizeof(regions[0]), GFP_KERNEL); in fsl_mc_device_get_mmio_regions()
687 if (!regions) in fsl_mc_device_get_mmio_regions()
714 &regions[i].start); in fsl_mc_device_get_mmio_regions()
740 regions[i].end = regions[i].start + region_desc.size - 1; in fsl_mc_device_get_mmio_regions()
743 regions[i].flags |= IORESOURCE_MEM; in fsl_mc_device_get_mmio_regions()
746 mc_dev->regions = regions; in fsl_mc_device_get_mmio_regions()
750 kfree(regions); in fsl_mc_device_get_mmio_regions()
771 kfree(mc_dev->regions); in fsl_mc_device_release()
[all …]
/drivers/soc/fsl/dpio/
A Ddpio-driver.c204 desc.regs_cena = devm_memremap(dev, dpio_dev->regions[1].start, in dpaa2_dpio_probe()
205 resource_size(&dpio_dev->regions[1]), in dpaa2_dpio_probe()
208 desc.regs_cena = devm_memremap(dev, dpio_dev->regions[2].start, in dpaa2_dpio_probe()
209 resource_size(&dpio_dev->regions[2]), in dpaa2_dpio_probe()
219 desc.regs_cinh = devm_ioremap(dev, dpio_dev->regions[1].start, in dpaa2_dpio_probe()
220 resource_size(&dpio_dev->regions[1])); in dpaa2_dpio_probe()
/drivers/fpga/
A Ddfl-afu-region.c21 INIT_LIST_HEAD(&afu->regions); in afu_mmio_region_init()
25 list_for_each_entry((region), &(afu)->regions, node)
80 list_add(&region->node, &afu->regions); in afu_mmio_region_add()
102 list_for_each_entry_safe(region, tmp, &afu->regions, node) in afu_mmio_region_destroy()
/drivers/vfio/pci/pds/
A Ddirty.c98 if (!dirty->regions) in pds_vfio_dirty_free_bitmaps()
102 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_dirty_free_bitmaps()
132 if (!dirty->regions) in pds_vfio_dirty_free_sgl()
136 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_dirty_free_sgl()
176 vfree(dirty->regions); in pds_vfio_dirty_free_regions()
177 dirty->regions = NULL; in pds_vfio_dirty_free_regions()
190 dirty->regions = vcalloc(num_regions, sizeof(struct pds_vfio_region)); in pds_vfio_dirty_alloc_regions()
191 if (!dirty->regions) in pds_vfio_dirty_alloc_regions()
197 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_dirty_alloc_regions()
511 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_get_region()
/drivers/soc/aspeed/
A Daspeed-p2a-ctrl.c52 struct region regions[P2A_REGION_COUNT]; member
140 const struct region *curr = &ctrl->config->regions[i]; in aspeed_p2a_region_acquire()
283 bits |= priv->parent->config->regions[i].bit; in aspeed_p2a_release()
324 value |= p2a_ctrl->config->regions[i].bit; in aspeed_p2a_disable_all()
399 .regions = {
410 .regions = {
/drivers/gpu/drm/i915/
A Dintel_memory_region.c342 for (i = 0; i < ARRAY_SIZE(i915->mm.regions); i++) { in intel_memory_regions_hw_probe()
384 i915->mm.regions[i] = mem; in intel_memory_regions_hw_probe()
388 for (i = 0; i < ARRAY_SIZE(i915->mm.regions); i++) { in intel_memory_regions_hw_probe()
389 struct intel_memory_region *mem = i915->mm.regions[i]; in intel_memory_regions_hw_probe()
417 for (i = 0; i < ARRAY_SIZE(i915->mm.regions); i++) { in intel_memory_regions_driver_release()
419 fetch_and_zero(&i915->mm.regions[i]); in intel_memory_regions_driver_release()

Completed in 81 milliseconds

123456