Home
last modified time | relevance | path

Searched refs:cache (Results 1 – 25 of 211) sorted by relevance

123456789

/drivers/md/
A Ddm-cache-target.c489 mg->cache = cache; in alloc_migration()
497 struct cache *cache = mg->cache; in free_migration() local
1103 struct cache *cache = mg->cache; in copy() local
1131 struct cache *cache = mg->cache; in overwrite_endio() local
1177 struct cache *cache = mg->cache; in mg_complete() local
1246 struct cache *cache = mg->cache; in mg_update_metadata() local
1348 struct cache *cache = mg->cache; in mg_full_copy() local
1402 struct cache *cache = mg->cache; in mg_lock_writes() local
1462 struct cache *cache = mg->cache; in invalidate_complete() local
1513 struct cache *cache = mg->cache; in invalidate_remove() local
[all …]
/drivers/acpi/acpica/
A Dutcache.c47 if (!cache) { in acpi_os_create_cache()
58 *return_cache = cache; in acpi_os_create_cache()
81 if (!cache) { in acpi_os_purge_cache()
100 cache->current_depth--; in acpi_os_purge_cache()
135 acpi_os_free(cache); in acpi_os_delete_cache()
165 if (cache->current_depth >= cache->max_depth) { in acpi_os_release_object()
215 if (!cache) { in acpi_os_acquire_object()
228 if (cache->list_head) { in acpi_os_acquire_object()
257 if ((cache->total_allocated - cache->total_freed) > in acpi_os_acquire_object()
259 cache->max_occupied = in acpi_os_acquire_object()
[all …]
/drivers/md/dm-vdo/indexer/
A Dsparse-cache.c248 list->capacity = cache->capacity; in make_search_list()
263 struct sparse_cache *cache; in uds_make_sparse_cache() local
271 cache->geometry = geometry; in uds_make_sparse_cache()
272 cache->capacity = capacity; in uds_make_sparse_cache()
273 cache->zone_count = zone_count; in uds_make_sparse_cache()
291 result = make_search_list(cache, &cache->search_lists[i]); in uds_make_sparse_cache()
302 *cache_ptr = cache; in uds_make_sparse_cache()
305 uds_free_sparse_cache(cache); in uds_make_sparse_cache()
349 if (cache == NULL) in uds_free_sparse_cache()
352 vdo_free(cache->scratch_entries); in uds_free_sparse_cache()
[all …]
A Dvolume.c234 if (cache->cache[i].read_pending) in select_victim_in_cache()
237 last_used = READ_ONCE(cache->cache[i].last_used); in select_victim_in_cache()
244 page = &cache->cache[oldest_index]; in select_victim_in_cache()
278 WRITE_ONCE(cache->index[physical_page], page - cache->cache); in put_page_in_cache()
667 *page_ptr = &cache->cache[index]; in get_page_and_index()
1528 &cache->cache); in initialize_page_cache()
1534 cache->index[i] = cache->cache_slots; in initialize_page_cache()
1537 clear_cache_page(cache, &cache->cache[i]); in initialize_page_cache()
1655 if (cache->cache != NULL) { in uninitialize_page_cache()
1657 release_page_buffer(&cache->cache[i]); in uninitialize_page_cache()
[all …]
/drivers/gpu/drm/xe/
A Dxe_guc_buf.c18 return container_of(cache, struct xe_guc, buf); in cache_to_guc()
23 return guc_to_gt(cache_to_guc(cache)); in cache_to_gt()
35 int xe_guc_buf_cache_init(struct xe_guc_buf_cache *cache) in xe_guc_buf_cache_init() argument
37 struct xe_gt *gt = cache_to_gt(cache); in xe_guc_buf_cache_init()
43 cache->sam = sam; in xe_guc_buf_cache_init()
57 u32 xe_guc_buf_cache_dwords(struct xe_guc_buf_cache *cache) in xe_guc_buf_cache_dwords() argument
59 return cache->sam ? cache->sam->base.size / sizeof(u32) : 0; in xe_guc_buf_cache_dwords()
76 if (cache->sam) in xe_guc_buf_reserve()
99 sa = __xe_sa_bo_new(cache->sam, size, GFP_ATOMIC); in xe_guc_buf_from_data()
162 ptrdiff_t offset = ptr - cache->sam->cpu_ptr; in xe_guc_cache_gpu_addr_from_ptr()
[all …]
A Dxe_guc_buf.h14 int xe_guc_buf_cache_init(struct xe_guc_buf_cache *cache);
15 u32 xe_guc_buf_cache_dwords(struct xe_guc_buf_cache *cache);
16 struct xe_guc_buf xe_guc_buf_reserve(struct xe_guc_buf_cache *cache, u32 dwords);
17 struct xe_guc_buf xe_guc_buf_from_data(struct xe_guc_buf_cache *cache,
35 u64 xe_guc_cache_gpu_addr_from_ptr(struct xe_guc_buf_cache *cache, const void *ptr, u32 size);
39 xe_guc_buf_reserve(cache, num),
40 struct xe_guc_buf_cache *cache, u32 num);
44 xe_guc_buf_from_data(cache, data, size),
45 struct xe_guc_buf_cache *cache, const void *data, size_t size);
/drivers/infiniband/hw/hfi1/
A Dplatform.c196 u8 *cache = ppd->qsfp_info.cache; in qual_power() local
225 u8 *cache = ppd->qsfp_info.cache; in qual_bitrate() local
251 u8 *cache = ppd->qsfp_info.cache; in set_qsfp_high_power() local
288 u8 *cache = ppd->qsfp_info.cache; in apply_rx_cdr() local
339 u8 *cache = ppd->qsfp_info.cache; in apply_tx_cdr() local
388 u8 *cache = ppd->qsfp_info.cache; in apply_cdr_settings() local
401 u8 *cache = ppd->qsfp_info.cache; in apply_tx_eq_auto() local
414 u8 *cache = ppd->qsfp_info.cache; in apply_tx_eq_prog() local
503 u8 *cache = ppd->qsfp_info.cache; in apply_eq_settings() local
620 u8 *cache = ppd->qsfp_info.cache; in aoc_low_power_setting() local
[all …]
A Dqsfp.c501 u8 *cache = &cp->cache[0]; in refresh_qsfp_cache() local
523 if (!(cache[2] & 4)) { in refresh_qsfp_cache()
525 if ((cache[195] & 0xC0) == 0xC0) { in refresh_qsfp_cache()
542 } else if ((cache[195] & 0x80) == 0x80) { in refresh_qsfp_cache()
730 u8 *cache = &ppd->qsfp_info.cache[0]; in qsfp_dump() local
735 u8 *atten = &cache[QSFP_ATTEN_OFFS]; in qsfp_dump()
736 u8 *vendor_oui = &cache[QSFP_VOUI_OFFS]; in qsfp_dump()
746 cache[QSFP_MOD_LEN_OFFS]); in qsfp_dump()
748 power_byte = cache[QSFP_MOD_PWR_OFFS]; in qsfp_dump()
763 QSFP_PN_LEN, &cache[QSFP_PN_OFFS]); in qsfp_dump()
[all …]
/drivers/gpu/drm/xe/tests/
A Dxe_guc_buf_kunit.c86 buf = xe_guc_buf_reserve(cache, 1); in test_smallest()
100 buf = xe_guc_buf_reserve(cache, xe_guc_buf_cache_dwords(cache)); in test_largest()
115 dwords = xe_guc_buf_cache_dwords(cache); in test_granular()
120 bufs[n] = xe_guc_buf_reserve(cache, 1); in test_granular()
135 dwords = xe_guc_buf_cache_dwords(cache); in test_unique()
140 bufs[n] = xe_guc_buf_reserve(cache, 1); in test_unique()
164 b1 = xe_guc_buf_reserve(cache, dwords); in test_overlap()
165 b2 = xe_guc_buf_reserve(cache, dwords); in test_overlap()
196 b1 = xe_guc_buf_reserve(cache, xe_guc_buf_cache_dwords(cache)); in test_reusable()
202 b2 = xe_guc_buf_reserve(cache, xe_guc_buf_cache_dwords(cache)); in test_reusable()
[all …]
/drivers/md/dm-vdo/
A Dblock-map.c158 struct vdo_page_cache *cache = info->cache; in get_page_buffer() local
185 for (info = cache->infos; info < cache->infos + cache->page_count; info++) { in initialize_info()
188 info->cache = cache; in initialize_info()
376 struct vdo_page_cache *cache = info->cache; in set_info_pbn() local
575 for (info = cache->infos; info < cache->infos + cache->page_count; info++) { in set_persistent_error()
669 struct vdo_page_cache *cache = info->cache; in handle_load_error() local
694 struct vdo_page_cache *cache = info->cache; in page_is_loaded() local
1158 cache = page_completion->cache; in vdo_release_page_completion()
1224 .cache = cache, in vdo_get_page()
1334 for (info = cache->infos; info < cache->infos + cache->page_count; info++) { in vdo_invalidate_page_cache()
[all …]
/drivers/base/regmap/
A Dregcache-flat.c24 unsigned int *cache; in regcache_flat_init() local
29 map->cache = kcalloc(regcache_flat_get_index(map, map->max_register) in regcache_flat_init()
31 if (!map->cache) in regcache_flat_init()
34 cache = map->cache; in regcache_flat_init()
40 cache[index] = map->reg_defaults[i].def; in regcache_flat_init()
48 kfree(map->cache); in regcache_flat_exit()
49 map->cache = NULL; in regcache_flat_exit()
57 unsigned int *cache = map->cache; in regcache_flat_read() local
60 *value = cache[index]; in regcache_flat_read()
68 unsigned int *cache = map->cache; in regcache_flat_write() local
[all …]
/drivers/net/wireless/silabs/wfx/
A Ddata_tx.c79 return it - cache->cache; in wfx_tx_policy_find()
82 return it - cache->cache; in wfx_tx_policy_find()
110 spin_lock_bh(&cache->lock); in wfx_tx_policy_get()
126 idx = entry - cache->cache; in wfx_tx_policy_get()
128 wfx_tx_policy_use(cache, &cache->cache[idx]); in wfx_tx_policy_get()
129 if (list_empty(&cache->free)) in wfx_tx_policy_get()
142 spin_lock_bh(&cache->lock); in wfx_tx_policy_put()
144 usage = wfx_tx_policy_release(cache, &cache->cache[idx]); in wfx_tx_policy_put()
188 memset(cache, 0, sizeof(*cache)); in wfx_tx_policy_init()
194 for (i = 0; i < ARRAY_SIZE(cache->cache); ++i) in wfx_tx_policy_init()
[all …]
/drivers/gpu/drm/vmwgfx/
A Dvmw_surface_cache.h389 struct vmw_surface_cache *cache) in vmw_surface_setup_cache() argument
394 memset(cache, 0, sizeof(*cache)); in vmw_surface_setup_cache()
396 cache->num_mip_levels = num_mip_levels; in vmw_surface_setup_cache()
397 cache->num_layers = num_layers; in vmw_surface_setup_cache()
418 cache->mip_chain_bytes += mip->bytes; in vmw_surface_setup_cache()
420 cache->sheet_bytes = cache->mip_chain_bytes * num_layers; in vmw_surface_setup_cache()
421 if (!cache->sheet_bytes) in vmw_surface_setup_cache()
448 loc->sheet = offset / cache->sheet_bytes; in vmw_surface_get_loc()
451 layer = offset / cache->mip_chain_bytes; in vmw_surface_get_loc()
452 offset -= layer * cache->mip_chain_bytes; in vmw_surface_get_loc()
[all …]
/drivers/block/
A Dps3vram.c311 struct ps3vram_cache *cache = &priv->cache; in ps3vram_cache_evict() local
323 entry * cache->page_size, cache->tags[entry].address, in ps3vram_cache_evict()
324 cache->page_size); in ps3vram_cache_evict()
333 struct ps3vram_cache *cache = &priv->cache; in ps3vram_cache_load() local
342 address, entry * cache->page_size, cache->page_size); in ps3vram_cache_load()
353 struct ps3vram_cache *cache = &priv->cache; in ps3vram_cache_flush() local
367 struct ps3vram_cache *cache = &priv->cache; in ps3vram_cache_match() local
380 cache->hit++; in ps3vram_cache_match()
394 cache->miss++; in ps3vram_cache_match()
407 if (!priv->cache.tags) in ps3vram_cache_init()
[all …]
/drivers/net/ethernet/netronome/nfp/nfpcore/
A Dnfp_cppcore.c169 if (cache->id) in nfp_cpp_free()
172 kfree(cache); in nfp_cpp_free()
805 cache = kzalloc(sizeof(*cache), GFP_KERNEL); in nfp_cpp_area_cache_add()
854 addr + length <= cache->addr + cache->size) in area_cache_get()
877 cache->addr = addr & ~(u64)(cache->size - 1); in area_cache_get()
882 id, cache->addr, cache->size); in area_cache_get()
907 if (!cache) in area_cache_put()
926 if (cache) { in __nfp_cpp_read()
942 if (cache) in __nfp_cpp_read()
995 if (cache) { in __nfp_cpp_write()
[all …]
A Dnfp_rtsym.c101 struct nfp_rtsym_table *cache; in __nfp_rtsym_table_read() local
121 size = sizeof(*cache); in __nfp_rtsym_table_read()
124 cache = kmalloc(size, GFP_KERNEL); in __nfp_rtsym_table_read()
125 if (!cache) in __nfp_rtsym_table_read()
128 cache->cpp = cpp; in __nfp_rtsym_table_read()
130 cache->strtab = (void *)&cache->symtab[cache->num]; in __nfp_rtsym_table_read()
139 cache->strtab[strtab_size] = '\0'; in __nfp_rtsym_table_read()
141 for (n = 0; n < cache->num; n++) in __nfp_rtsym_table_read()
143 &cache->symtab[n], &rtsymtab[n]); in __nfp_rtsym_table_read()
147 return cache; in __nfp_rtsym_table_read()
[all …]
/drivers/net/ethernet/microchip/vcap/
A Dvcap_api_debugfs_kunit.c121 keystr = &admin->cache.keystream[start]; in test_cache_read()
122 mskstr = &admin->cache.maskstream[start]; in test_cache_read()
162 keystr = &admin->cache.keystream[start]; in test_cache_write()
163 mskstr = &admin->cache.maskstream[start]; in test_cache_write()
307 .cache = { in vcap_api_addr_keyset_test()
325 admin.cache.keystream = &keydata[idx]; in vcap_api_addr_keyset_test()
326 admin.cache.maskstream = &mskdata[idx]; in vcap_api_addr_keyset_test()
333 admin.cache.keystream = &keydata[idx]; in vcap_api_addr_keyset_test()
334 admin.cache.maskstream = &mskdata[idx]; in vcap_api_addr_keyset_test()
354 .cache = { in vcap_api_show_admin_raw_test()
[all …]
/drivers/hwmon/
A Dbt1-pvt.c208 struct pvt_cache *cache; in pvt_soft_isr() local
223 cache = &pvt->cache[pvt->sensor]; in pvt_soft_isr()
293 struct pvt_cache *cache = &pvt->cache[type]; in pvt_read_data() local
299 data = cache->data; in pvt_read_data()
411 struct pvt_cache *cache; in pvt_hard_isr() local
431 cache = &pvt->cache[pvt->sensor]; in pvt_hard_isr()
435 complete(&cache->conversion); in pvt_hard_isr()
455 struct pvt_cache *cache = &pvt->cache[type]; in pvt_read_data() local
493 data = READ_ONCE(cache->data); in pvt_read_data()
662 ktime_t kt, cache; in pvt_write_timeout() local
[all …]
/drivers/net/wireless/st/cw1200/
A Dtxrx.c224 return it - cache->cache; in tx_policy_find()
229 return it - cache->cache; in tx_policy_find()
257 spin_lock_bh(&cache->lock); in tx_policy_clean()
261 entry = &cache->cache[idx]; in tx_policy_clean()
286 memset(cache, 0, sizeof(*cache)); in tx_policy_init()
293 list_add(&cache->cache[i].link, &cache->free); in tx_policy_init()
306 spin_lock_bh(&cache->lock); in tx_policy_get()
324 idx = entry - cache->cache; in tx_policy_get()
328 tx_policy_use(cache, &cache->cache[idx]); in tx_policy_get()
344 usage = tx_policy_release(cache, &cache->cache[idx]); in tx_policy_put()
[all …]
/drivers/xen/
A Dgrant-table.c908 cache->pages = NULL; in cache_init()
913 return !cache->pages; in cache_empty()
920 page = cache->pages; in cache_deq()
929 cache->pages = page; in cache_enq()
961 cache_init(cache); in gnttab_page_cache_init()
962 cache->num_pages = 0; in gnttab_page_cache_init()
972 if (cache_empty(cache)) { in gnttab_page_cache_get()
977 page[0] = cache_deq(cache); in gnttab_page_cache_get()
978 cache->num_pages--; in gnttab_page_cache_get()
996 cache->num_pages += num; in gnttab_page_cache_put()
[all …]
/drivers/net/ethernet/microchip/lan966x/
A Dlan966x_vcap_impl.c393 memset(&admin->cache.counter, 0, sizeof(admin->cache.counter)); in lan966x_vcap_cache_erase()
412 admin->cache.counter = counter; in lan966x_es0_read_esdx_counter()
459 admin->cache.sticky = admin->cache.counter > 0; in lan966x_vcap_cache_write()
501 admin->cache.counter = in lan966x_vcap_cache_read()
503 admin->cache.sticky = admin->cache.counter > 0; in lan966x_vcap_cache_read()
601 kfree(admin->cache.keystream); in lan966x_vcap_admin_free()
602 kfree(admin->cache.maskstream); in lan966x_vcap_admin_free()
603 kfree(admin->cache.actionstream); in lan966x_vcap_admin_free()
638 if (!admin->cache.keystream || in lan966x_vcap_admin_alloc()
639 !admin->cache.maskstream || in lan966x_vcap_admin_alloc()
[all …]
/drivers/md/bcache/
A Dalloc.c88 struct cache *ca; in bch_rescale_priorities()
106 ca = c->cache; in bch_rescale_priorities()
177 static void invalidate_buckets_lru(struct cache *ca) in invalidate_buckets_lru()
263 static void invalidate_buckets(struct cache *ca) in invalidate_buckets()
317 struct cache *ca = arg; in bch_allocator_thread()
488 struct cache *ca; in __bch_bucket_alloc_set()
499 ca = c->cache; in __bch_bucket_alloc_set()
578 ret->sectors_free = c->cache->sb.bucket_size; in pick_data_bucket()
669 &c->cache->sectors_written); in bch_alloc_sectors()
672 if (b->sectors_free < c->cache->sb.block_size) in bch_alloc_sectors()
[all …]
/drivers/net/ethernet/microchip/sparx5/
A Dsparx5_vcap_impl.c899 memset(&admin->cache.counter, 0, sizeof(admin->cache.counter)); in sparx5_vcap_cache_erase()
1143 admin->cache.counter = in sparx5_vcap_is0_cache_read()
1145 admin->cache.sticky = in sparx5_vcap_is0_cache_read()
1180 admin->cache.counter = in sparx5_vcap_is2_cache_read()
1183 admin->cache.counter = in sparx5_vcap_is2_cache_read()
1185 admin->cache.sticky = in sparx5_vcap_is2_cache_read()
1233 admin->cache.counter = in sparx5_vcap_es0_cache_read()
1235 admin->cache.sticky = admin->cache.counter; in sparx5_vcap_es0_cache_read()
1269 admin->cache.counter = in sparx5_vcap_es2_cache_read()
1271 admin->cache.sticky = in sparx5_vcap_es2_cache_read()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
A Dpat_arg.c88 *cache = new_cache; in mlx5hws_pat_init_pattern_cache()
95 mutex_destroy(&cache->lock); in mlx5hws_pat_uninit_pattern_cache()
96 kfree(cache); in mlx5hws_pat_uninit_pattern_cache()
129 mlx5hws_pat_find_cached_pattern(struct mlx5hws_pattern_cache *cache, in mlx5hws_pat_find_cached_pattern() argument
135 list_for_each_entry(cached_pat, &cache->ptrn_list, ptrn_list_node) { in mlx5hws_pat_find_cached_pattern()
156 list_move(&cached_pattern->ptrn_list_node, &cache->ptrn_list); in mlx5hws_pat_get_existing_cached_pattern()
164 mlx5hws_pat_add_pattern_to_cache(struct mlx5hws_pattern_cache *cache, in mlx5hws_pat_add_pattern_to_cache() argument
182 list_add(&cached_pattern->ptrn_list_node, &cache->ptrn_list); in mlx5hws_pat_add_pattern_to_cache()
217 struct mlx5hws_pattern_cache *cache = ctx->pattern_cache; in mlx5hws_pat_put_pattern() local
220 mutex_lock(&cache->lock); in mlx5hws_pat_put_pattern()
[all …]
/drivers/net/xen-netback/
A Dhash.c50 spin_lock_irqsave(&vif->hash.cache.lock, flags); in xenvif_add_hash()
55 lockdep_is_held(&vif->hash.cache.lock)) { in xenvif_add_hash()
65 new->seq = atomic_inc_return(&vif->hash.cache.seq); in xenvif_add_hash()
66 list_add_rcu(&new->link, &vif->hash.cache.list); in xenvif_add_hash()
70 vif->hash.cache.count--; in xenvif_add_hash()
75 spin_unlock_irqrestore(&vif->hash.cache.lock, flags); in xenvif_add_hash()
104 spin_lock_irqsave(&vif->hash.cache.lock, flags); in xenvif_flush_hash()
108 vif->hash.cache.count--; in xenvif_flush_hash()
458 BUG_ON(vif->hash.cache.count); in xenvif_init_hash()
460 spin_lock_init(&vif->hash.cache.lock); in xenvif_init_hash()
[all …]

Completed in 90 milliseconds

123456789