| /drivers/md/ |
| A D | dm-cache-target.c | 489 mg->cache = cache; in alloc_migration() 497 struct cache *cache = mg->cache; in free_migration() local 1103 struct cache *cache = mg->cache; in copy() local 1131 struct cache *cache = mg->cache; in overwrite_endio() local 1177 struct cache *cache = mg->cache; in mg_complete() local 1246 struct cache *cache = mg->cache; in mg_update_metadata() local 1348 struct cache *cache = mg->cache; in mg_full_copy() local 1402 struct cache *cache = mg->cache; in mg_lock_writes() local 1462 struct cache *cache = mg->cache; in invalidate_complete() local 1513 struct cache *cache = mg->cache; in invalidate_remove() local [all …]
|
| /drivers/acpi/acpica/ |
| A D | utcache.c | 47 if (!cache) { in acpi_os_create_cache() 58 *return_cache = cache; in acpi_os_create_cache() 81 if (!cache) { in acpi_os_purge_cache() 100 cache->current_depth--; in acpi_os_purge_cache() 135 acpi_os_free(cache); in acpi_os_delete_cache() 165 if (cache->current_depth >= cache->max_depth) { in acpi_os_release_object() 215 if (!cache) { in acpi_os_acquire_object() 228 if (cache->list_head) { in acpi_os_acquire_object() 257 if ((cache->total_allocated - cache->total_freed) > in acpi_os_acquire_object() 259 cache->max_occupied = in acpi_os_acquire_object() [all …]
|
| /drivers/md/dm-vdo/indexer/ |
| A D | sparse-cache.c | 248 list->capacity = cache->capacity; in make_search_list() 263 struct sparse_cache *cache; in uds_make_sparse_cache() local 271 cache->geometry = geometry; in uds_make_sparse_cache() 272 cache->capacity = capacity; in uds_make_sparse_cache() 273 cache->zone_count = zone_count; in uds_make_sparse_cache() 291 result = make_search_list(cache, &cache->search_lists[i]); in uds_make_sparse_cache() 302 *cache_ptr = cache; in uds_make_sparse_cache() 305 uds_free_sparse_cache(cache); in uds_make_sparse_cache() 349 if (cache == NULL) in uds_free_sparse_cache() 352 vdo_free(cache->scratch_entries); in uds_free_sparse_cache() [all …]
|
| A D | volume.c | 234 if (cache->cache[i].read_pending) in select_victim_in_cache() 237 last_used = READ_ONCE(cache->cache[i].last_used); in select_victim_in_cache() 244 page = &cache->cache[oldest_index]; in select_victim_in_cache() 278 WRITE_ONCE(cache->index[physical_page], page - cache->cache); in put_page_in_cache() 667 *page_ptr = &cache->cache[index]; in get_page_and_index() 1528 &cache->cache); in initialize_page_cache() 1534 cache->index[i] = cache->cache_slots; in initialize_page_cache() 1537 clear_cache_page(cache, &cache->cache[i]); in initialize_page_cache() 1655 if (cache->cache != NULL) { in uninitialize_page_cache() 1657 release_page_buffer(&cache->cache[i]); in uninitialize_page_cache() [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | xe_guc_buf.c | 18 return container_of(cache, struct xe_guc, buf); in cache_to_guc() 23 return guc_to_gt(cache_to_guc(cache)); in cache_to_gt() 35 int xe_guc_buf_cache_init(struct xe_guc_buf_cache *cache) in xe_guc_buf_cache_init() argument 37 struct xe_gt *gt = cache_to_gt(cache); in xe_guc_buf_cache_init() 43 cache->sam = sam; in xe_guc_buf_cache_init() 57 u32 xe_guc_buf_cache_dwords(struct xe_guc_buf_cache *cache) in xe_guc_buf_cache_dwords() argument 59 return cache->sam ? cache->sam->base.size / sizeof(u32) : 0; in xe_guc_buf_cache_dwords() 76 if (cache->sam) in xe_guc_buf_reserve() 99 sa = __xe_sa_bo_new(cache->sam, size, GFP_ATOMIC); in xe_guc_buf_from_data() 162 ptrdiff_t offset = ptr - cache->sam->cpu_ptr; in xe_guc_cache_gpu_addr_from_ptr() [all …]
|
| A D | xe_guc_buf.h | 14 int xe_guc_buf_cache_init(struct xe_guc_buf_cache *cache); 15 u32 xe_guc_buf_cache_dwords(struct xe_guc_buf_cache *cache); 16 struct xe_guc_buf xe_guc_buf_reserve(struct xe_guc_buf_cache *cache, u32 dwords); 17 struct xe_guc_buf xe_guc_buf_from_data(struct xe_guc_buf_cache *cache, 35 u64 xe_guc_cache_gpu_addr_from_ptr(struct xe_guc_buf_cache *cache, const void *ptr, u32 size); 39 xe_guc_buf_reserve(cache, num), 40 struct xe_guc_buf_cache *cache, u32 num); 44 xe_guc_buf_from_data(cache, data, size), 45 struct xe_guc_buf_cache *cache, const void *data, size_t size);
|
| /drivers/infiniband/hw/hfi1/ |
| A D | platform.c | 196 u8 *cache = ppd->qsfp_info.cache; in qual_power() local 225 u8 *cache = ppd->qsfp_info.cache; in qual_bitrate() local 251 u8 *cache = ppd->qsfp_info.cache; in set_qsfp_high_power() local 288 u8 *cache = ppd->qsfp_info.cache; in apply_rx_cdr() local 339 u8 *cache = ppd->qsfp_info.cache; in apply_tx_cdr() local 388 u8 *cache = ppd->qsfp_info.cache; in apply_cdr_settings() local 401 u8 *cache = ppd->qsfp_info.cache; in apply_tx_eq_auto() local 414 u8 *cache = ppd->qsfp_info.cache; in apply_tx_eq_prog() local 503 u8 *cache = ppd->qsfp_info.cache; in apply_eq_settings() local 620 u8 *cache = ppd->qsfp_info.cache; in aoc_low_power_setting() local [all …]
|
| A D | qsfp.c | 501 u8 *cache = &cp->cache[0]; in refresh_qsfp_cache() local 523 if (!(cache[2] & 4)) { in refresh_qsfp_cache() 525 if ((cache[195] & 0xC0) == 0xC0) { in refresh_qsfp_cache() 542 } else if ((cache[195] & 0x80) == 0x80) { in refresh_qsfp_cache() 730 u8 *cache = &ppd->qsfp_info.cache[0]; in qsfp_dump() local 735 u8 *atten = &cache[QSFP_ATTEN_OFFS]; in qsfp_dump() 736 u8 *vendor_oui = &cache[QSFP_VOUI_OFFS]; in qsfp_dump() 746 cache[QSFP_MOD_LEN_OFFS]); in qsfp_dump() 748 power_byte = cache[QSFP_MOD_PWR_OFFS]; in qsfp_dump() 763 QSFP_PN_LEN, &cache[QSFP_PN_OFFS]); in qsfp_dump() [all …]
|
| /drivers/gpu/drm/xe/tests/ |
| A D | xe_guc_buf_kunit.c | 86 buf = xe_guc_buf_reserve(cache, 1); in test_smallest() 100 buf = xe_guc_buf_reserve(cache, xe_guc_buf_cache_dwords(cache)); in test_largest() 115 dwords = xe_guc_buf_cache_dwords(cache); in test_granular() 120 bufs[n] = xe_guc_buf_reserve(cache, 1); in test_granular() 135 dwords = xe_guc_buf_cache_dwords(cache); in test_unique() 140 bufs[n] = xe_guc_buf_reserve(cache, 1); in test_unique() 164 b1 = xe_guc_buf_reserve(cache, dwords); in test_overlap() 165 b2 = xe_guc_buf_reserve(cache, dwords); in test_overlap() 196 b1 = xe_guc_buf_reserve(cache, xe_guc_buf_cache_dwords(cache)); in test_reusable() 202 b2 = xe_guc_buf_reserve(cache, xe_guc_buf_cache_dwords(cache)); in test_reusable() [all …]
|
| /drivers/md/dm-vdo/ |
| A D | block-map.c | 158 struct vdo_page_cache *cache = info->cache; in get_page_buffer() local 185 for (info = cache->infos; info < cache->infos + cache->page_count; info++) { in initialize_info() 188 info->cache = cache; in initialize_info() 376 struct vdo_page_cache *cache = info->cache; in set_info_pbn() local 575 for (info = cache->infos; info < cache->infos + cache->page_count; info++) { in set_persistent_error() 669 struct vdo_page_cache *cache = info->cache; in handle_load_error() local 694 struct vdo_page_cache *cache = info->cache; in page_is_loaded() local 1158 cache = page_completion->cache; in vdo_release_page_completion() 1224 .cache = cache, in vdo_get_page() 1334 for (info = cache->infos; info < cache->infos + cache->page_count; info++) { in vdo_invalidate_page_cache() [all …]
|
| /drivers/base/regmap/ |
| A D | regcache-flat.c | 24 unsigned int *cache; in regcache_flat_init() local 29 map->cache = kcalloc(regcache_flat_get_index(map, map->max_register) in regcache_flat_init() 31 if (!map->cache) in regcache_flat_init() 34 cache = map->cache; in regcache_flat_init() 40 cache[index] = map->reg_defaults[i].def; in regcache_flat_init() 48 kfree(map->cache); in regcache_flat_exit() 49 map->cache = NULL; in regcache_flat_exit() 57 unsigned int *cache = map->cache; in regcache_flat_read() local 60 *value = cache[index]; in regcache_flat_read() 68 unsigned int *cache = map->cache; in regcache_flat_write() local [all …]
|
| /drivers/net/wireless/silabs/wfx/ |
| A D | data_tx.c | 79 return it - cache->cache; in wfx_tx_policy_find() 82 return it - cache->cache; in wfx_tx_policy_find() 110 spin_lock_bh(&cache->lock); in wfx_tx_policy_get() 126 idx = entry - cache->cache; in wfx_tx_policy_get() 128 wfx_tx_policy_use(cache, &cache->cache[idx]); in wfx_tx_policy_get() 129 if (list_empty(&cache->free)) in wfx_tx_policy_get() 142 spin_lock_bh(&cache->lock); in wfx_tx_policy_put() 144 usage = wfx_tx_policy_release(cache, &cache->cache[idx]); in wfx_tx_policy_put() 188 memset(cache, 0, sizeof(*cache)); in wfx_tx_policy_init() 194 for (i = 0; i < ARRAY_SIZE(cache->cache); ++i) in wfx_tx_policy_init() [all …]
|
| /drivers/gpu/drm/vmwgfx/ |
| A D | vmw_surface_cache.h | 389 struct vmw_surface_cache *cache) in vmw_surface_setup_cache() argument 394 memset(cache, 0, sizeof(*cache)); in vmw_surface_setup_cache() 396 cache->num_mip_levels = num_mip_levels; in vmw_surface_setup_cache() 397 cache->num_layers = num_layers; in vmw_surface_setup_cache() 418 cache->mip_chain_bytes += mip->bytes; in vmw_surface_setup_cache() 420 cache->sheet_bytes = cache->mip_chain_bytes * num_layers; in vmw_surface_setup_cache() 421 if (!cache->sheet_bytes) in vmw_surface_setup_cache() 448 loc->sheet = offset / cache->sheet_bytes; in vmw_surface_get_loc() 451 layer = offset / cache->mip_chain_bytes; in vmw_surface_get_loc() 452 offset -= layer * cache->mip_chain_bytes; in vmw_surface_get_loc() [all …]
|
| /drivers/block/ |
| A D | ps3vram.c | 311 struct ps3vram_cache *cache = &priv->cache; in ps3vram_cache_evict() local 323 entry * cache->page_size, cache->tags[entry].address, in ps3vram_cache_evict() 324 cache->page_size); in ps3vram_cache_evict() 333 struct ps3vram_cache *cache = &priv->cache; in ps3vram_cache_load() local 342 address, entry * cache->page_size, cache->page_size); in ps3vram_cache_load() 353 struct ps3vram_cache *cache = &priv->cache; in ps3vram_cache_flush() local 367 struct ps3vram_cache *cache = &priv->cache; in ps3vram_cache_match() local 380 cache->hit++; in ps3vram_cache_match() 394 cache->miss++; in ps3vram_cache_match() 407 if (!priv->cache.tags) in ps3vram_cache_init() [all …]
|
| /drivers/net/ethernet/netronome/nfp/nfpcore/ |
| A D | nfp_cppcore.c | 169 if (cache->id) in nfp_cpp_free() 172 kfree(cache); in nfp_cpp_free() 805 cache = kzalloc(sizeof(*cache), GFP_KERNEL); in nfp_cpp_area_cache_add() 854 addr + length <= cache->addr + cache->size) in area_cache_get() 877 cache->addr = addr & ~(u64)(cache->size - 1); in area_cache_get() 882 id, cache->addr, cache->size); in area_cache_get() 907 if (!cache) in area_cache_put() 926 if (cache) { in __nfp_cpp_read() 942 if (cache) in __nfp_cpp_read() 995 if (cache) { in __nfp_cpp_write() [all …]
|
| A D | nfp_rtsym.c | 101 struct nfp_rtsym_table *cache; in __nfp_rtsym_table_read() local 121 size = sizeof(*cache); in __nfp_rtsym_table_read() 124 cache = kmalloc(size, GFP_KERNEL); in __nfp_rtsym_table_read() 125 if (!cache) in __nfp_rtsym_table_read() 128 cache->cpp = cpp; in __nfp_rtsym_table_read() 130 cache->strtab = (void *)&cache->symtab[cache->num]; in __nfp_rtsym_table_read() 139 cache->strtab[strtab_size] = '\0'; in __nfp_rtsym_table_read() 141 for (n = 0; n < cache->num; n++) in __nfp_rtsym_table_read() 143 &cache->symtab[n], &rtsymtab[n]); in __nfp_rtsym_table_read() 147 return cache; in __nfp_rtsym_table_read() [all …]
|
| /drivers/net/ethernet/microchip/vcap/ |
| A D | vcap_api_debugfs_kunit.c | 121 keystr = &admin->cache.keystream[start]; in test_cache_read() 122 mskstr = &admin->cache.maskstream[start]; in test_cache_read() 162 keystr = &admin->cache.keystream[start]; in test_cache_write() 163 mskstr = &admin->cache.maskstream[start]; in test_cache_write() 307 .cache = { in vcap_api_addr_keyset_test() 325 admin.cache.keystream = &keydata[idx]; in vcap_api_addr_keyset_test() 326 admin.cache.maskstream = &mskdata[idx]; in vcap_api_addr_keyset_test() 333 admin.cache.keystream = &keydata[idx]; in vcap_api_addr_keyset_test() 334 admin.cache.maskstream = &mskdata[idx]; in vcap_api_addr_keyset_test() 354 .cache = { in vcap_api_show_admin_raw_test() [all …]
|
| /drivers/hwmon/ |
| A D | bt1-pvt.c | 208 struct pvt_cache *cache; in pvt_soft_isr() local 223 cache = &pvt->cache[pvt->sensor]; in pvt_soft_isr() 293 struct pvt_cache *cache = &pvt->cache[type]; in pvt_read_data() local 299 data = cache->data; in pvt_read_data() 411 struct pvt_cache *cache; in pvt_hard_isr() local 431 cache = &pvt->cache[pvt->sensor]; in pvt_hard_isr() 435 complete(&cache->conversion); in pvt_hard_isr() 455 struct pvt_cache *cache = &pvt->cache[type]; in pvt_read_data() local 493 data = READ_ONCE(cache->data); in pvt_read_data() 662 ktime_t kt, cache; in pvt_write_timeout() local [all …]
|
| /drivers/net/wireless/st/cw1200/ |
| A D | txrx.c | 224 return it - cache->cache; in tx_policy_find() 229 return it - cache->cache; in tx_policy_find() 257 spin_lock_bh(&cache->lock); in tx_policy_clean() 261 entry = &cache->cache[idx]; in tx_policy_clean() 286 memset(cache, 0, sizeof(*cache)); in tx_policy_init() 293 list_add(&cache->cache[i].link, &cache->free); in tx_policy_init() 306 spin_lock_bh(&cache->lock); in tx_policy_get() 324 idx = entry - cache->cache; in tx_policy_get() 328 tx_policy_use(cache, &cache->cache[idx]); in tx_policy_get() 344 usage = tx_policy_release(cache, &cache->cache[idx]); in tx_policy_put() [all …]
|
| /drivers/xen/ |
| A D | grant-table.c | 908 cache->pages = NULL; in cache_init() 913 return !cache->pages; in cache_empty() 920 page = cache->pages; in cache_deq() 929 cache->pages = page; in cache_enq() 961 cache_init(cache); in gnttab_page_cache_init() 962 cache->num_pages = 0; in gnttab_page_cache_init() 972 if (cache_empty(cache)) { in gnttab_page_cache_get() 977 page[0] = cache_deq(cache); in gnttab_page_cache_get() 978 cache->num_pages--; in gnttab_page_cache_get() 996 cache->num_pages += num; in gnttab_page_cache_put() [all …]
|
| /drivers/net/ethernet/microchip/lan966x/ |
| A D | lan966x_vcap_impl.c | 393 memset(&admin->cache.counter, 0, sizeof(admin->cache.counter)); in lan966x_vcap_cache_erase() 412 admin->cache.counter = counter; in lan966x_es0_read_esdx_counter() 459 admin->cache.sticky = admin->cache.counter > 0; in lan966x_vcap_cache_write() 501 admin->cache.counter = in lan966x_vcap_cache_read() 503 admin->cache.sticky = admin->cache.counter > 0; in lan966x_vcap_cache_read() 601 kfree(admin->cache.keystream); in lan966x_vcap_admin_free() 602 kfree(admin->cache.maskstream); in lan966x_vcap_admin_free() 603 kfree(admin->cache.actionstream); in lan966x_vcap_admin_free() 638 if (!admin->cache.keystream || in lan966x_vcap_admin_alloc() 639 !admin->cache.maskstream || in lan966x_vcap_admin_alloc() [all …]
|
| /drivers/md/bcache/ |
| A D | alloc.c | 88 struct cache *ca; in bch_rescale_priorities() 106 ca = c->cache; in bch_rescale_priorities() 177 static void invalidate_buckets_lru(struct cache *ca) in invalidate_buckets_lru() 263 static void invalidate_buckets(struct cache *ca) in invalidate_buckets() 317 struct cache *ca = arg; in bch_allocator_thread() 488 struct cache *ca; in __bch_bucket_alloc_set() 499 ca = c->cache; in __bch_bucket_alloc_set() 578 ret->sectors_free = c->cache->sb.bucket_size; in pick_data_bucket() 669 &c->cache->sectors_written); in bch_alloc_sectors() 672 if (b->sectors_free < c->cache->sb.block_size) in bch_alloc_sectors() [all …]
|
| /drivers/net/ethernet/microchip/sparx5/ |
| A D | sparx5_vcap_impl.c | 899 memset(&admin->cache.counter, 0, sizeof(admin->cache.counter)); in sparx5_vcap_cache_erase() 1143 admin->cache.counter = in sparx5_vcap_is0_cache_read() 1145 admin->cache.sticky = in sparx5_vcap_is0_cache_read() 1180 admin->cache.counter = in sparx5_vcap_is2_cache_read() 1183 admin->cache.counter = in sparx5_vcap_is2_cache_read() 1185 admin->cache.sticky = in sparx5_vcap_is2_cache_read() 1233 admin->cache.counter = in sparx5_vcap_es0_cache_read() 1235 admin->cache.sticky = admin->cache.counter; in sparx5_vcap_es0_cache_read() 1269 admin->cache.counter = in sparx5_vcap_es2_cache_read() 1271 admin->cache.sticky = in sparx5_vcap_es2_cache_read() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| A D | pat_arg.c | 88 *cache = new_cache; in mlx5hws_pat_init_pattern_cache() 95 mutex_destroy(&cache->lock); in mlx5hws_pat_uninit_pattern_cache() 96 kfree(cache); in mlx5hws_pat_uninit_pattern_cache() 129 mlx5hws_pat_find_cached_pattern(struct mlx5hws_pattern_cache *cache, in mlx5hws_pat_find_cached_pattern() argument 135 list_for_each_entry(cached_pat, &cache->ptrn_list, ptrn_list_node) { in mlx5hws_pat_find_cached_pattern() 156 list_move(&cached_pattern->ptrn_list_node, &cache->ptrn_list); in mlx5hws_pat_get_existing_cached_pattern() 164 mlx5hws_pat_add_pattern_to_cache(struct mlx5hws_pattern_cache *cache, in mlx5hws_pat_add_pattern_to_cache() argument 182 list_add(&cached_pattern->ptrn_list_node, &cache->ptrn_list); in mlx5hws_pat_add_pattern_to_cache() 217 struct mlx5hws_pattern_cache *cache = ctx->pattern_cache; in mlx5hws_pat_put_pattern() local 220 mutex_lock(&cache->lock); in mlx5hws_pat_put_pattern() [all …]
|
| /drivers/net/xen-netback/ |
| A D | hash.c | 50 spin_lock_irqsave(&vif->hash.cache.lock, flags); in xenvif_add_hash() 55 lockdep_is_held(&vif->hash.cache.lock)) { in xenvif_add_hash() 65 new->seq = atomic_inc_return(&vif->hash.cache.seq); in xenvif_add_hash() 66 list_add_rcu(&new->link, &vif->hash.cache.list); in xenvif_add_hash() 70 vif->hash.cache.count--; in xenvif_add_hash() 75 spin_unlock_irqrestore(&vif->hash.cache.lock, flags); in xenvif_add_hash() 104 spin_lock_irqsave(&vif->hash.cache.lock, flags); in xenvif_flush_hash() 108 vif->hash.cache.count--; in xenvif_flush_hash() 458 BUG_ON(vif->hash.cache.count); in xenvif_init_hash() 460 spin_lock_init(&vif->hash.cache.lock); in xenvif_init_hash() [all …]
|