| /drivers/mtd/chips/ |
| A D | jedec_probe.c | 307 .regions = { 319 .regions = { 334 .regions = { 349 .regions = { 364 .regions = { 379 .regions = { 395 .regions = { 412 .regions = { 429 .regions = { 444 .regions = { [all …]
|
| A D | cfi_util.c | 365 struct mtd_erase_region_info *regions = mtd->eraseregions; in cfi_varsize_frob() local 379 while (i < mtd->numeraseregions && ofs >= regions[i].offset) in cfi_varsize_frob() 389 if (ofs & (regions[i].erasesize-1)) in cfi_varsize_frob() 399 while (i<mtd->numeraseregions && (ofs + len) >= regions[i].offset) in cfi_varsize_frob() 407 if ((ofs + len) & (regions[i].erasesize-1)) in cfi_varsize_frob() 416 int size = regions[i].erasesize; in cfi_varsize_frob() 427 if (ofs == regions[i].offset + size * regions[i].numblocks) in cfi_varsize_frob()
|
| /drivers/mtd/devices/ |
| A D | mtd_intel_dg.c | 205 if ((nvm->regions[i].offset + nvm->regions[i].size - 1) >= from && in idg_nvm_get_region() 438 u8 id = nvm->regions[i].id; in intel_dg_nvm_init() 467 nvm->regions[i].name, in intel_dg_nvm_init() 468 nvm->regions[i].id, in intel_dg_nvm_init() 469 nvm->regions[i].offset, in intel_dg_nvm_init() 470 nvm->regions[i].size, in intel_dg_nvm_init() 568 nvm->regions[idx].id, nvm->regions[idx].name, from, len); in intel_dg_mtd_read() 607 nvm->regions[idx].id, nvm->regions[idx].name, to, len); in intel_dg_mtd_write() 726 if (invm->regions[i].name) in intel_dg_mtd_probe() 743 if (!invm->regions[i].name) in intel_dg_mtd_probe() [all …]
|
| /drivers/vfio/platform/ |
| A D | vfio_platform_common.c | 146 if (!vdev->regions) in vfio_platform_regions_init() 153 vdev->regions[i].addr = res->start; in vfio_platform_regions_init() 154 vdev->regions[i].size = resource_size(res); in vfio_platform_regions_init() 155 vdev->regions[i].flags = 0; in vfio_platform_regions_init() 162 vdev->regions[i].flags |= in vfio_platform_regions_init() 170 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_platform_regions_init() 171 vdev->regions[i].flags |= in vfio_platform_regions_init() 187 kfree(vdev->regions); in vfio_platform_regions_init() 196 iounmap(vdev->regions[i].ioaddr); in vfio_platform_regions_cleanup() 199 kfree(vdev->regions); in vfio_platform_regions_cleanup() [all …]
|
| /drivers/gpu/drm/nouveau/nvkm/nvfw/ |
| A D | acr.c | 130 hdr->regions.no_regions); in flcn_acr_desc_dump() 135 hdr->regions.region_props[i].start_addr); in flcn_acr_desc_dump() 137 hdr->regions.region_props[i].end_addr); in flcn_acr_desc_dump() 139 hdr->regions.region_props[i].region_id); in flcn_acr_desc_dump() 141 hdr->regions.region_props[i].read_mask); in flcn_acr_desc_dump() 143 hdr->regions.region_props[i].write_mask); in flcn_acr_desc_dump() 173 hdr->regions.no_regions); in flcn_acr_desc_v1_dump() 178 hdr->regions.region_props[i].start_addr); in flcn_acr_desc_v1_dump() 180 hdr->regions.region_props[i].end_addr); in flcn_acr_desc_v1_dump() 182 hdr->regions.region_props[i].region_id); in flcn_acr_desc_v1_dump() [all …]
|
| /drivers/vfio/cdx/ |
| A D | main.c | 21 if (!vdev->regions) in vfio_cdx_open_device() 27 vdev->regions[i].addr = res->start; in vfio_cdx_open_device() 28 vdev->regions[i].size = resource_size(res); in vfio_cdx_open_device() 29 vdev->regions[i].type = res->flags; in vfio_cdx_open_device() 34 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_cdx_open_device() 35 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_cdx_open_device() 36 vdev->regions[i].flags |= in vfio_cdx_open_device() 44 kfree(vdev->regions); in vfio_cdx_open_device() 45 vdev->regions = NULL; in vfio_cdx_open_device() 62 kfree(vdev->regions); in vfio_cdx_close_device() [all …]
|
| /drivers/vfio/fsl-mc/ |
| A D | vfio_fsl_mc.c | 32 if (!vdev->regions) in vfio_fsl_mc_open_device() 36 struct resource *res = &mc_dev->regions[i]; in vfio_fsl_mc_open_device() 39 vdev->regions[i].addr = res->start; in vfio_fsl_mc_open_device() 40 vdev->regions[i].size = resource_size(res); in vfio_fsl_mc_open_device() 41 vdev->regions[i].type = mc_dev->regions[i].flags & IORESOURCE_BITS; in vfio_fsl_mc_open_device() 47 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_fsl_mc_open_device() 48 vdev->regions[i].flags |= in vfio_fsl_mc_open_device() 64 iounmap(vdev->regions[i].ioaddr); in vfio_fsl_mc_regions_cleanup() 65 kfree(vdev->regions); in vfio_fsl_mc_regions_cleanup() 255 region = &vdev->regions[index]; in vfio_fsl_mc_read() [all …]
|
| /drivers/md/ |
| A D | dm-bio-prison-v1.c | 29 struct prison_region regions[] __counted_by(num_locks); 53 spin_lock_init(&prison->regions[i].lock); in dm_bio_prison_create() 54 prison->regions[i].cell = RB_ROOT; in dm_bio_prison_create() 184 spin_lock_irq(&prison->regions[l].lock); in bio_detain() 186 spin_unlock_irq(&prison->regions[l].lock); in bio_detain() 223 spin_lock_irq(&prison->regions[l].lock); in dm_cell_release() 225 spin_unlock_irq(&prison->regions[l].lock); in dm_cell_release() 247 spin_lock_irqsave(&prison->regions[l].lock, flags); in dm_cell_release_no_holder() 275 spin_lock_irq(&prison->regions[l].lock); in dm_cell_visit_release() 277 rb_erase(&cell->node, &prison->regions[l].cell); in dm_cell_visit_release() [all …]
|
| /drivers/virt/nitro_enclaves/ |
| A D | ne_misc_dev_test.c | 117 phys_contig_mem_regions.regions = kunit_kcalloc(test, MAX_PHYS_REGIONS, in ne_misc_dev_test_merge_phys_contig_memory_regions() 118 sizeof(*phys_contig_mem_regions.regions), in ne_misc_dev_test_merge_phys_contig_memory_regions() 120 KUNIT_ASSERT_TRUE(test, phys_contig_mem_regions.regions); in ne_misc_dev_test_merge_phys_contig_memory_regions() 135 KUNIT_EXPECT_EQ(test, phys_contig_mem_regions.regions[num - 1].start, in ne_misc_dev_test_merge_phys_contig_memory_regions() 137 KUNIT_EXPECT_EQ(test, range_len(&phys_contig_mem_regions.regions[num - 1]), in ne_misc_dev_test_merge_phys_contig_memory_regions() 141 kunit_kfree(test, phys_contig_mem_regions.regions); in ne_misc_dev_test_merge_phys_contig_memory_regions()
|
| A D | ne_misc_dev.c | 136 struct range *regions; member 894 phys_contig_regions->regions[num - 1].end += page_size; in ne_merge_phys_contig_memory_regions() 896 phys_contig_regions->regions[num].start = page_paddr; in ne_merge_phys_contig_memory_regions() 945 phys_contig_mem_regions.regions = kcalloc(max_nr_pages, in ne_set_user_memory_region_ioctl() 946 sizeof(*phys_contig_mem_regions.regions), in ne_set_user_memory_region_ioctl() 948 if (!phys_contig_mem_regions.regions) { in ne_set_user_memory_region_ioctl() 1005 u64 phys_region_addr = phys_contig_mem_regions.regions[i].start; in ne_set_user_memory_region_ioctl() 1023 slot_add_mem_req.paddr = phys_contig_mem_regions.regions[i].start; in ne_set_user_memory_region_ioctl() 1033 kfree(phys_contig_mem_regions.regions); in ne_set_user_memory_region_ioctl() 1046 kfree(phys_contig_mem_regions.regions); in ne_set_user_memory_region_ioctl() [all …]
|
| /drivers/soc/qcom/ |
| A D | smem.c | 468 header = smem->regions[0].virt_base; in qcom_smem_alloc_global() 557 header = smem->regions[0].virt_base; in qcom_smem_get_global() 565 region = &smem->regions[i]; in qcom_smem_get_global() 749 if (ret > __smem->regions[0].size) in qcom_smem_get_free_space() 795 area = &__smem->regions[i]; in qcom_smem_virt_to_phys() 868 header = smem->regions[0].virt_base; in qcom_smem_get_sbl_version() 1087 smem->regions[0].size = size; in qcom_smem_map_global() 1090 if (!smem->regions[0].virt_base) in qcom_smem_map_global() 1149 smem->regions[0].size = rmem->size; in qcom_smem_probe() 1174 smem->regions[i].size); in qcom_smem_probe() [all …]
|
| A D | ocmem.c | 64 struct ocmem_region *regions; member 119 struct ocmem_region *region = &ocmem->regions[i]; in update_ocmem() 131 struct ocmem_region *region = &ocmem->regions[i]; in update_ocmem() 169 struct ocmem_region *region = &ocmem->regions[i]; in update_range() 373 ocmem->regions = devm_kcalloc(dev, ocmem->config->num_regions, in ocmem_dev_probe() 375 if (!ocmem->regions) { in ocmem_dev_probe() 381 struct ocmem_region *region = &ocmem->regions[i]; in ocmem_dev_probe()
|
| /drivers/net/dsa/sja1105/ |
| A D | sja1105_devlink.c | 85 priv->regions = kcalloc(num_regions, sizeof(struct devlink_region *), in sja1105_setup_devlink_regions() 87 if (!priv->regions) in sja1105_setup_devlink_regions() 97 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_setup_devlink_regions() 99 kfree(priv->regions); in sja1105_setup_devlink_regions() 103 priv->regions[i] = region; in sja1105_setup_devlink_regions() 115 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_teardown_devlink_regions() 117 kfree(priv->regions); in sja1105_teardown_devlink_regions()
|
| /drivers/perf/ |
| A D | marvell_cn10k_tad_pmu.c | 32 struct tad_region *regions; member 62 new += readq(tad_pmu->regions[i].base + in tad_pmu_event_counter_read() 80 writeq_relaxed(0, tad_pmu->regions[i].base + in tad_pmu_event_counter_stop() 101 writeq_relaxed(0, tad_pmu->regions[i].base + in tad_pmu_event_counter_start() 293 struct tad_region *regions; in tad_pmu_probe() local 342 regions = devm_kcalloc(&pdev->dev, tad_cnt, in tad_pmu_probe() 343 sizeof(*regions), GFP_KERNEL); in tad_pmu_probe() 344 if (!regions) in tad_pmu_probe() 349 regions[i].base = devm_ioremap(&pdev->dev, in tad_pmu_probe() 352 if (!regions[i].base) { in tad_pmu_probe() [all …]
|
| /drivers/net/wireless/ath/ath10k/ |
| A D | coredump.c | 1300 .regions = qca6174_hw10_mem_regions, 1309 .regions = qca6174_hw10_mem_regions, 1318 .regions = qca6174_hw10_mem_regions, 1327 .regions = qca6174_hw21_mem_regions, 1336 .regions = qca6174_hw30_mem_regions, 1345 .regions = qca6174_hw30_mem_regions, 1363 .regions = qca6174_hw30_mem_regions, 1372 .regions = qca988x_hw20_mem_regions, 1381 .regions = qca9984_hw10_mem_regions, 1390 .regions = qca9984_hw10_mem_regions, [all …]
|
| /drivers/gpu/drm/i915/gem/selftests/ |
| A D | i915_gem_dmabuf.c | 96 struct intel_memory_region *lmem = i915->mm.regions[INTEL_REGION_LMEM_0]; in igt_dmabuf_import_same_driver_lmem() 220 struct intel_memory_region **regions, in igt_dmabuf_import_same_driver() argument 234 regions, num_regions); in igt_dmabuf_import_same_driver() 278 if (obj->mm.region != i915->mm.regions[INTEL_REGION_SMEM]) { in igt_dmabuf_import_same_driver() 326 struct intel_memory_region *smem = i915->mm.regions[INTEL_REGION_SMEM]; in igt_dmabuf_import_same_driver_smem() 334 struct intel_memory_region *regions[2]; in igt_dmabuf_import_same_driver_lmem_smem() local 336 if (!i915->mm.regions[INTEL_REGION_LMEM_0]) in igt_dmabuf_import_same_driver_lmem_smem() 339 regions[0] = i915->mm.regions[INTEL_REGION_LMEM_0]; in igt_dmabuf_import_same_driver_lmem_smem() 340 regions[1] = i915->mm.regions[INTEL_REGION_SMEM]; in igt_dmabuf_import_same_driver_lmem_smem() 341 return igt_dmabuf_import_same_driver(i915, regions, 2); in igt_dmabuf_import_same_driver_lmem_smem()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/acr/ |
| A D | gp102.c | 203 desc->regions.no_regions = 2; in gp102_acr_load_setup() 204 desc->regions.region_props[0].start_addr = acr->wpr_start >> 8; in gp102_acr_load_setup() 205 desc->regions.region_props[0].end_addr = acr->wpr_end >> 8; in gp102_acr_load_setup() 206 desc->regions.region_props[0].region_id = 1; in gp102_acr_load_setup() 207 desc->regions.region_props[0].read_mask = 0xf; in gp102_acr_load_setup() 208 desc->regions.region_props[0].write_mask = 0xc; in gp102_acr_load_setup() 209 desc->regions.region_props[0].client_mask = 0x2; in gp102_acr_load_setup() 210 desc->regions.region_props[0].shadow_mem_start_addr = acr->shadow_start >> 8; in gp102_acr_load_setup()
|
| A D | gm200.c | 288 desc->regions.no_regions = 2; in gm200_acr_load_setup() 289 desc->regions.region_props[0].start_addr = acr->wpr_start >> 8; in gm200_acr_load_setup() 290 desc->regions.region_props[0].end_addr = acr->wpr_end >> 8; in gm200_acr_load_setup() 291 desc->regions.region_props[0].region_id = 1; in gm200_acr_load_setup() 292 desc->regions.region_props[0].read_mask = 0xf; in gm200_acr_load_setup() 293 desc->regions.region_props[0].write_mask = 0xc; in gm200_acr_load_setup() 294 desc->regions.region_props[0].client_mask = 0x2; in gm200_acr_load_setup()
|
| /drivers/virt/acrn/ |
| A D | mm.c | 21 struct vm_memory_region_batch *regions; in modify_region() local 24 regions = kzalloc(sizeof(*regions), GFP_KERNEL); in modify_region() 25 if (!regions) in modify_region() 28 regions->vmid = vm->vmid; in modify_region() 29 regions->regions_num = 1; in modify_region() 30 regions->regions_gpa = virt_to_phys(region); in modify_region() 32 ret = hcall_set_memory_regions(virt_to_phys(regions)); in modify_region() 37 kfree(regions); in modify_region()
|
| /drivers/bus/fsl-mc/ |
| A D | fsl-mc-bus.c | 666 struct resource *regions; in fsl_mc_device_get_mmio_regions() local 685 regions = kmalloc_array(obj_desc->region_count, in fsl_mc_device_get_mmio_regions() 686 sizeof(regions[0]), GFP_KERNEL); in fsl_mc_device_get_mmio_regions() 687 if (!regions) in fsl_mc_device_get_mmio_regions() 714 ®ions[i].start); in fsl_mc_device_get_mmio_regions() 740 regions[i].end = regions[i].start + region_desc.size - 1; in fsl_mc_device_get_mmio_regions() 743 regions[i].flags |= IORESOURCE_MEM; in fsl_mc_device_get_mmio_regions() 746 mc_dev->regions = regions; in fsl_mc_device_get_mmio_regions() 750 kfree(regions); in fsl_mc_device_get_mmio_regions() 771 kfree(mc_dev->regions); in fsl_mc_device_release() [all …]
|
| /drivers/soc/fsl/dpio/ |
| A D | dpio-driver.c | 204 desc.regs_cena = devm_memremap(dev, dpio_dev->regions[1].start, in dpaa2_dpio_probe() 205 resource_size(&dpio_dev->regions[1]), in dpaa2_dpio_probe() 208 desc.regs_cena = devm_memremap(dev, dpio_dev->regions[2].start, in dpaa2_dpio_probe() 209 resource_size(&dpio_dev->regions[2]), in dpaa2_dpio_probe() 219 desc.regs_cinh = devm_ioremap(dev, dpio_dev->regions[1].start, in dpaa2_dpio_probe() 220 resource_size(&dpio_dev->regions[1])); in dpaa2_dpio_probe()
|
| /drivers/fpga/ |
| A D | dfl-afu-region.c | 21 INIT_LIST_HEAD(&afu->regions); in afu_mmio_region_init() 25 list_for_each_entry((region), &(afu)->regions, node) 80 list_add(®ion->node, &afu->regions); in afu_mmio_region_add() 102 list_for_each_entry_safe(region, tmp, &afu->regions, node) in afu_mmio_region_destroy()
|
| /drivers/vfio/pci/pds/ |
| A D | dirty.c | 98 if (!dirty->regions) in pds_vfio_dirty_free_bitmaps() 102 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_dirty_free_bitmaps() 132 if (!dirty->regions) in pds_vfio_dirty_free_sgl() 136 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_dirty_free_sgl() 176 vfree(dirty->regions); in pds_vfio_dirty_free_regions() 177 dirty->regions = NULL; in pds_vfio_dirty_free_regions() 190 dirty->regions = vcalloc(num_regions, sizeof(struct pds_vfio_region)); in pds_vfio_dirty_alloc_regions() 191 if (!dirty->regions) in pds_vfio_dirty_alloc_regions() 197 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_dirty_alloc_regions() 511 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_get_region()
|
| /drivers/soc/aspeed/ |
| A D | aspeed-p2a-ctrl.c | 52 struct region regions[P2A_REGION_COUNT]; member 140 const struct region *curr = &ctrl->config->regions[i]; in aspeed_p2a_region_acquire() 283 bits |= priv->parent->config->regions[i].bit; in aspeed_p2a_release() 324 value |= p2a_ctrl->config->regions[i].bit; in aspeed_p2a_disable_all() 399 .regions = { 410 .regions = {
|
| /drivers/gpu/drm/i915/ |
| A D | intel_memory_region.c | 342 for (i = 0; i < ARRAY_SIZE(i915->mm.regions); i++) { in intel_memory_regions_hw_probe() 384 i915->mm.regions[i] = mem; in intel_memory_regions_hw_probe() 388 for (i = 0; i < ARRAY_SIZE(i915->mm.regions); i++) { in intel_memory_regions_hw_probe() 389 struct intel_memory_region *mem = i915->mm.regions[i]; in intel_memory_regions_hw_probe() 417 for (i = 0; i < ARRAY_SIZE(i915->mm.regions); i++) { in intel_memory_regions_driver_release() 419 fetch_and_zero(&i915->mm.regions[i]); in intel_memory_regions_driver_release()
|