| /drivers/md/dm-vdo/indexer/ |
| A D | geometry.c | 62 struct index_geometry *geometry; in uds_make_index_geometry() local 77 geometry->records_per_chapter = geometry->records_per_page * record_pages_per_chapter; in uds_make_index_geometry() 78 geometry->records_per_volume = (u64) geometry->records_per_chapter * chapters_per_volume; in uds_make_index_geometry() 89 geometry->delta_lists_per_chapter = 1 << geometry->chapter_delta_list_bits; in uds_make_index_geometry() 91 geometry->chapter_address_bits = in uds_make_index_geometry() 102 geometry->pages_per_chapter = geometry->index_pages_per_chapter + record_pages_per_chapter; in uds_make_index_geometry() 103 geometry->pages_per_volume = geometry->pages_per_chapter * chapters_per_volume; in uds_make_index_geometry() 104 geometry->bytes_per_volume = in uds_make_index_geometry() 107 *geometry_ptr = geometry; in uds_make_index_geometry() 124 vdo_free(geometry); in uds_free_index_geometry() [all …]
|
| A D | config.c | 32 struct index_geometry *geometry = user->geometry; in are_matching_configurations() local 38 geometry->record_pages_per_chapter); in are_matching_configurations() 45 geometry->chapters_per_volume); in are_matching_configurations() 52 geometry->sparse_chapters_per_volume); in are_matching_configurations() 98 struct index_geometry geometry; in uds_validate_config_contents() local 133 geometry.bytes_per_page = bytes_per_page; in uds_validate_config_contents() 144 user_config->geometry->remapped_virtual = 0; in uds_validate_config_contents() 178 struct index_geometry *geometry = config->geometry; in uds_write_config_contents() local 334 0, 0, &config->geometry); in uds_make_configuration() 358 uds_free_index_geometry(config->geometry); in uds_free_configuration() [all …]
|
| A D | chapter-index.c | 31 memory_size = ((geometry->index_pages_per_chapter + 1) * geometry->bytes_per_page); in uds_make_open_chapter_index() 32 index->geometry = geometry; in uds_make_open_chapter_index() 35 geometry->delta_lists_per_chapter, in uds_make_open_chapter_index() 36 geometry->chapter_mean_delta, in uds_make_open_chapter_index() 37 geometry->chapter_payload_bits, in uds_make_open_chapter_index() 82 const struct index_geometry *geometry = chapter_index->geometry; in uds_put_open_chapter_index_record() local 131 const struct index_geometry *geometry = chapter_index->geometry; in uds_pack_open_chapter_index_page() local 140 geometry->bytes_per_page, in uds_pack_open_chapter_index_page() 214 geometry->chapter_mean_delta, in uds_initialize_chapter_index_page() 215 geometry->chapter_payload_bits, in uds_initialize_chapter_index_page() [all …]
|
| A D | volume.c | 425 struct index_geometry *geometry = volume->geometry; in init_chapter_index_page() local 885 struct index_geometry *geometry = volume->geometry; in uds_search_cached_record_page() local 928 const struct index_geometry *geometry = volume->geometry; in uds_prefetch_volume_chapter() local 940 const struct index_geometry *geometry = volume->geometry; in uds_read_chapter_index_from_volume() local 999 struct index_geometry *geometry = volume->geometry; in uds_search_volume_page_cache_for_rebuild() local 1100 struct index_geometry *geometry = volume->geometry; in write_index_pages() local 1211 struct index_geometry *geometry = volume->geometry; in write_record_pages() local 1272 const struct index_geometry *geometry = volume->geometry; in probe_chapter() local 1356 struct index_geometry *geometry = volume->geometry; in find_chapter_limits() local 1557 result = uds_copy_index_geometry(config->geometry, &volume->geometry); in uds_make_volume() [all …]
|
| A D | index-page-map.c | 30 static inline u32 get_entry_count(const struct index_geometry *geometry) in get_entry_count() argument 32 return geometry->chapters_per_volume * (geometry->index_pages_per_chapter - 1); in get_entry_count() 35 int uds_make_index_page_map(const struct index_geometry *geometry, in uds_make_index_page_map() argument 45 map->geometry = geometry; in uds_make_index_page_map() 46 map->entries_per_chapter = geometry->index_pages_per_chapter - 1; in uds_make_index_page_map() 83 u32 delta_list_number = uds_hash_to_chapter_delta_list(name, map->geometry); in uds_find_index_page_number() 105 map->geometry->delta_lists_per_chapter - 1); in uds_get_list_number_bounds() 118 u64 saved_size = uds_compute_index_page_map_save_size(map->geometry); in uds_write_index_page_map() 128 for (i = 0; i < get_entry_count(map->geometry); i++) in uds_write_index_page_map() 145 u64 saved_size = uds_compute_index_page_map_save_size(map->geometry); in uds_read_index_page_map() [all …]
|
| A D | geometry.h | 107 void uds_free_index_geometry(struct index_geometry *geometry); 109 u32 __must_check uds_map_to_physical_chapter(const struct index_geometry *geometry, 117 uds_is_reduced_index_geometry(const struct index_geometry *geometry) in uds_is_reduced_index_geometry() argument 119 return !!(geometry->chapters_per_volume & 1); in uds_is_reduced_index_geometry() 123 uds_is_sparse_index_geometry(const struct index_geometry *geometry) in uds_is_sparse_index_geometry() argument 125 return geometry->sparse_chapters_per_volume > 0; in uds_is_sparse_index_geometry() 128 bool __must_check uds_has_sparse_chapters(const struct index_geometry *geometry, 132 bool __must_check uds_is_chapter_sparse(const struct index_geometry *geometry, 137 u32 __must_check uds_chapters_to_expire(const struct index_geometry *geometry,
|
| A D | hash-utils.h | 47 const struct index_geometry *geometry) in uds_hash_to_chapter_delta_list() argument 49 return ((uds_extract_chapter_index_bytes(name) >> geometry->chapter_address_bits) & in uds_hash_to_chapter_delta_list() 50 ((1 << geometry->chapter_delta_list_bits) - 1)); in uds_hash_to_chapter_delta_list() 55 const struct index_geometry *geometry) in uds_hash_to_chapter_delta_address() argument 57 return uds_extract_chapter_index_bytes(name) & ((1 << geometry->chapter_address_bits) - 1); in uds_hash_to_chapter_delta_address()
|
| A D | chapter-index.h | 25 const struct index_geometry *geometry; member 33 const struct index_geometry *geometry, 50 const struct index_geometry *geometry, 54 const struct index_geometry *geometry); 57 const struct index_geometry *geometry,
|
| A D | index.c | 80 return uds_is_chapter_sparse(zone->index->volume->geometry, in is_zone_chapter_sparse() 648 const struct index_geometry *geometry) in initialize_index_queues() argument 723 uds_chapters_to_expire(index->volume->geometry, in close_chapters() 785 index->volume->geometry, in make_chapter_writer() 829 struct index_geometry *geometry = index->volume->geometry; in rebuild_index_page_map() local 830 u32 chapter = uds_map_to_physical_chapter(geometry, vcn); in rebuild_index_page_map() 976 const struct index_geometry *geometry; in replay_chapter() local 985 geometry = index->volume->geometry; in replay_chapter() 997 for (i = 0; i < geometry->record_pages_per_chapter; i++) { in replay_chapter() 1009 for (j = 0; j < geometry->records_per_page; j++) { in replay_chapter() [all …]
|
| A D | sparse-cache.c | 155 const struct index_geometry *geometry; member 218 const struct index_geometry *geometry) in initialize_cached_chapter_index() argument 223 chapter->index_pages_count = geometry->index_pages_per_chapter; in initialize_cached_chapter_index() 258 int uds_make_sparse_cache(const struct index_geometry *geometry, unsigned int capacity, in uds_make_sparse_cache() argument 271 cache->geometry = geometry; in uds_make_sparse_cache() 285 result = initialize_cached_chapter_index(&cache->chapters[i], geometry); in uds_make_sparse_cache() 559 const struct index_geometry *geometry, in search_cached_chapter_index() argument 565 uds_map_to_physical_chapter(geometry, chapter->virtual_chapter); in search_cached_chapter_index() 571 return uds_search_chapter_index_page(index_page, geometry, name, in search_cached_chapter_index() 596 result = search_cached_chapter_index(chapter, cache->geometry, in uds_search_sparse_cache()
|
| A D | index-page-map.h | 19 const struct index_geometry *geometry; member 25 int __must_check uds_make_index_page_map(const struct index_geometry *geometry, 48 u64 uds_compute_index_page_map_save_size(const struct index_geometry *geometry);
|
| A D | open-chapter.c | 63 int uds_make_open_chapter(const struct index_geometry *geometry, unsigned int zone_count, in uds_make_open_chapter() argument 68 size_t capacity = geometry->records_per_chapter / zone_count; in uds_make_open_chapter() 233 records_per_chapter = index->geometry->records_per_chapter; in fill_delta_chapter_index() 234 records_per_page = index->geometry->records_per_page; in fill_delta_chapter_index() 348 u64 uds_compute_saved_open_chapter_size(struct index_geometry *geometry) in uds_compute_saved_open_chapter_size() argument 350 unsigned int records_per_chapter = geometry->records_per_chapter; in uds_compute_saved_open_chapter_size()
|
| A D | volume-index.c | 210 struct index_geometry *geometry = config->geometry; in compute_volume_sub_index_parameters() local 211 u64 records_per_chapter = geometry->records_per_chapter; in compute_volume_sub_index_parameters() 213 params->chapter_count = geometry->chapters_per_volume; in compute_volume_sub_index_parameters() 220 if (uds_is_reduced_index_geometry(geometry)) in compute_volume_sub_index_parameters() 326 split->hook_geometry = *config->geometry; in split_configuration() 327 split->hook_config.geometry = &split->hook_geometry; in split_configuration() 329 split->non_hook_geometry = *config->geometry; in split_configuration() 330 split->non_hook_config.geometry = &split->non_hook_geometry; in split_configuration() 333 sparse_chapters = config->geometry->sparse_chapters_per_volume; in split_configuration() 354 if (!uds_is_sparse_index_geometry(config->geometry)) in compute_volume_index_save_bytes() [all …]
|
| A D | open-chapter.h | 46 int __must_check uds_make_open_chapter(const struct index_geometry *geometry, 77 u64 uds_compute_saved_open_chapter_size(struct index_geometry *geometry);
|
| /drivers/iommu/ |
| A D | sprd-iommu.c | 131 return ((domain->geometry.aperture_end - in sprd_iommu_pgt_size() 132 domain->geometry.aperture_start + 1) >> in sprd_iommu_pgt_size() 148 dom->domain.geometry.aperture_start = 0; in sprd_iommu_domain_alloc_paging() 149 dom->domain.geometry.aperture_end = SZ_256M - 1; in sprd_iommu_domain_alloc_paging() 150 dom->domain.geometry.force_aperture = true; in sprd_iommu_domain_alloc_paging() 181 val = (dom->domain.geometry.aperture_end - in sprd_iommu_vpn_range() 296 unsigned long start = domain->geometry.aperture_start; in sprd_iommu_map() 297 unsigned long end = domain->geometry.aperture_end; in sprd_iommu_map() 331 unsigned long start = domain->geometry.aperture_start; in sprd_iommu_unmap() 332 unsigned long end = domain->geometry.aperture_end; in sprd_iommu_unmap() [all …]
|
| A D | s390-iommu.c | 561 s390_domain->domain.geometry.force_aperture = true; in s390_domain_alloc_paging() 562 s390_domain->domain.geometry.aperture_start = 0; in s390_domain_alloc_paging() 563 s390_domain->domain.geometry.aperture_end = max_tbl_size(s390_domain); in s390_domain_alloc_paging() 691 if (WARN_ON(domain->geometry.aperture_start > zdev->end_dma || in s390_iommu_attach_device() 692 domain->geometry.aperture_end < zdev->start_dma)) in s390_iommu_attach_device() 905 if (iova < s390_domain->domain.geometry.aperture_start || in s390_iommu_map_pages() 906 (iova + size - 1) > s390_domain->domain.geometry.aperture_end) in s390_iommu_map_pages() 979 if (iova < domain->geometry.aperture_start || in s390_iommu_iova_to_phys() 980 iova > domain->geometry.aperture_end) in s390_iommu_iova_to_phys() 1015 if (WARN_ON(iova < s390_domain->domain.geometry.aperture_start || in s390_iommu_unmap_pages() [all …]
|
| A D | fsl_pamu_domain.c | 175 if (iova < domain->geometry.aperture_start || in fsl_pamu_iova_to_phys() 176 iova > domain->geometry.aperture_end) in fsl_pamu_iova_to_phys() 218 dma_domain->iommu_domain. geometry.aperture_start = 0; in fsl_pamu_domain_alloc() 219 dma_domain->iommu_domain.geometry.aperture_end = (1ULL << 36) - 1; in fsl_pamu_domain_alloc() 220 dma_domain->iommu_domain.geometry.force_aperture = true; in fsl_pamu_domain_alloc()
|
| A D | virtio-iommu.c | 47 struct iommu_domain_geometry geometry; member 403 u64 iova = vdomain->domain.geometry.aperture_start; in viommu_domain_map_identity() 404 u64 limit = vdomain->domain.geometry.aperture_end; in viommu_domain_map_identity() 690 vdomain->domain.geometry = viommu->geometry; in viommu_domain_alloc_paging() 1205 viommu->geometry = (struct iommu_domain_geometry) { in viommu_probe() 1237 order_base_2(viommu->geometry.aperture_end)); in viommu_probe()
|
| /drivers/gpu/drm/imagination/ |
| A D | pvr_context.c | 173 pvr_queue_destroy(ctx->queues.geometry); in pvr_context_destroy_queues() 202 ctx->queues.geometry = pvr_queue_create(ctx, DRM_PVR_JOB_TYPE_GEOMETRY, in pvr_context_create_queues() 204 if (IS_ERR(ctx->queues.geometry)) { in pvr_context_create_queues() 205 err = PTR_ERR(ctx->queues.geometry); in pvr_context_create_queues() 206 ctx->queues.geometry = NULL; in pvr_context_create_queues() 260 pvr_queue_kill(ctx->queues.geometry); in pvr_context_kill_queues()
|
| A D | pvr_context.h | 76 struct pvr_queue *geometry; member 98 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.geometry : NULL; in pvr_context_get_queue_for_job()
|
| /drivers/md/dm-vdo/ |
| A D | encodings.h | 801 vdo_get_index_region_start(struct volume_geometry geometry) in vdo_get_index_region_start() argument 803 return geometry.regions[VDO_INDEX_REGION].start_block; in vdo_get_index_region_start() 813 vdo_get_data_region_start(struct volume_geometry geometry) in vdo_get_data_region_start() argument 815 return geometry.regions[VDO_DATA_REGION].start_block; in vdo_get_data_region_start() 825 vdo_get_index_region_size(struct volume_geometry geometry) in vdo_get_index_region_size() argument 827 return vdo_get_data_region_start(geometry) - in vdo_get_index_region_size() 828 vdo_get_index_region_start(geometry); in vdo_get_index_region_size() 832 struct volume_geometry *geometry); 1281 struct volume_geometry *geometry,
|
| A D | encodings.c | 249 struct volume_geometry *geometry, u32 version) in decode_volume_geometry() argument 259 geometry->unused = unused; in decode_volume_geometry() 262 geometry->nonce = nonce; in decode_volume_geometry() 264 memcpy((unsigned char *) &geometry->uuid, buffer + *offset, sizeof(uuid_t)); in decode_volume_geometry() 269 geometry->bio_offset = bio_offset; in decode_volume_geometry() 278 geometry->regions[id] = (struct volume_region) { in decode_volume_geometry() 288 geometry->index_config = (struct index_config) { in decode_volume_geometry() 321 decode_volume_geometry(block, &offset, geometry, header.version.major_version); in vdo_parse_geometry_block() 1299 struct volume_geometry *geometry, in decode_components() argument 1337 int vdo_decode_component_states(u8 *buffer, struct volume_geometry *geometry, in vdo_decode_component_states() argument [all …]
|
| /drivers/gpu/drm/rockchip/ |
| A D | rockchip_drm_drv.c | 104 struct iommu_domain_geometry *geometry; in rockchip_drm_init_iommu() local 118 geometry = &private->domain->geometry; in rockchip_drm_init_iommu() 119 start = geometry->aperture_start; in rockchip_drm_init_iommu() 120 end = geometry->aperture_end; in rockchip_drm_init_iommu()
|
| /drivers/staging/media/ipu3/ |
| A D | ipu3-mmu.c | 63 struct imgu_mmu_info geometry; member 68 return container_of(info, struct imgu_mmu, geometry); in to_imgu_mmu() 481 mmu->geometry.aperture_start = 0; in imgu_mmu_init() 482 mmu->geometry.aperture_end = DMA_BIT_MASK(IPU3_MMU_ADDRESS_BITS); in imgu_mmu_init() 484 return &mmu->geometry; in imgu_mmu_init()
|
| /drivers/gpu/host1x/ |
| A D | dev.c | 399 struct iommu_domain_geometry *geometry; in host1x_iommu_attach() local 422 geometry = &host->domain->geometry; in host1x_iommu_attach() 423 start = geometry->aperture_start & host->info->dma_mask; in host1x_iommu_attach() 424 end = geometry->aperture_end & host->info->dma_mask; in host1x_iommu_attach()
|