| /linux/tools/lib/api/fs/ |
| A D | cgroup.c | 17 static struct cgroupfs_cache_entry *cached; variable 27 if (cached && !strcmp(cached->subsys, subsys)) { in cgroupfs_find_mountpoint() 28 if (strlen(cached->mountpoint) < maxlen) { in cgroupfs_find_mountpoint() 29 strcpy(buf, cached->mountpoint); in cgroupfs_find_mountpoint() 94 if (!cached) in cgroupfs_find_mountpoint() 95 cached = calloc(1, sizeof(*cached)); in cgroupfs_find_mountpoint() 97 if (cached) { in cgroupfs_find_mountpoint() 98 strncpy(cached->subsys, subsys, sizeof(cached->subsys) - 1); in cgroupfs_find_mountpoint() 99 strcpy(cached->mountpoint, mountpoint); in cgroupfs_find_mountpoint()
|
| /linux/fs/btrfs/ |
| A D | extent-io-tree.h | 141 struct extent_state **cached); 143 struct extent_state **cached); 146 struct extent_state **cached) in lock_extent() argument 152 u64 end, struct extent_state **cached) in try_lock_extent() argument 172 u32 bits, struct extent_state **cached, 177 struct extent_state **cached) in clear_extent_bit() argument 183 struct extent_state **cached) in unlock_extent() argument 207 u64 end, struct extent_state **cached) in clear_extent_dirty() argument 211 EXTENT_DO_ACCOUNTING, cached); in clear_extent_dirty() 229 u64 end, struct extent_state **cached) in lock_dio_extent() argument [all …]
|
| A D | extent-io-tree.c | 664 cached = *cached_state; in __clear_extent_bit() 671 if (cached && extent_state_in_tree(cached) && in __clear_extent_bit() 672 cached->start <= start && cached->end > start) { in __clear_extent_bit() 675 state = cached; in __clear_extent_bit() 1659 cached = *cached_state; in count_range_bits() 1664 if (cached->start <= cur_start && cur_start <= cached->end) { in count_range_bits() 1665 state = cached; in count_range_bits() 1678 state = cached; in count_range_bits() 1767 if (cached && extent_state_in_tree(cached) && cached->start <= start && in test_range_bit() 1768 cached->end > start) in test_range_bit() [all …]
|
| A D | fiemap.c | 72 bool cached; member 117 if (!cache->cached) in emit_fiemap_extent() 252 cache->cached = false; in emit_fiemap_extent() 266 cache->cached = false; in emit_fiemap_extent() 270 cache->cached = true; in emit_fiemap_extent() 295 if (!cache->cached) in emit_last_fiemap_cache() 300 cache->cached = false; in emit_last_fiemap_cache() 822 if (cache.cached && cache.offset + cache.len >= last_extent_end) { in extent_fiemap()
|
| A D | block-group.h | 165 int cached; member 375 return cache->cached == BTRFS_CACHE_FINISHED || in btrfs_block_group_done() 376 cache->cached == BTRFS_CACHE_ERROR; in btrfs_block_group_done()
|
| /linux/fs/lockd/ |
| A D | mon.c | 343 if (cached != NULL) { in nsm_get_handle() 344 refcount_inc(&cached->sm_count); in nsm_get_handle() 348 "cnt %d\n", cached->sm_name, in nsm_get_handle() 349 cached->sm_addrbuf, in nsm_get_handle() 351 return cached; in nsm_get_handle() 382 struct nsm_handle *cached; in nsm_reboot_lookup() local 388 if (unlikely(cached == NULL)) { in nsm_reboot_lookup() 392 return cached; in nsm_reboot_lookup() 395 refcount_inc(&cached->sm_count); in nsm_reboot_lookup() 399 cached->sm_name, cached->sm_addrbuf, in nsm_reboot_lookup() [all …]
|
| /linux/lib/zstd/compress/ |
| A D | hist.c | 91 { U32 cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() local 93 U32 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 98 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 103 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 108 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp()
|
| /linux/drivers/mtd/maps/ |
| A D | pxa2xx-flash.c | 26 unsigned long start = (unsigned long)map->cached + from; in pxa2xx_map_inval_cache() 70 info->map.cached = ioremap_cache(info->map.phys, info->map.size); in pxa2xx_flash_probe() 71 if (!info->map.cached) in pxa2xx_flash_probe() 87 if (info->map.cached) in pxa2xx_flash_probe() 88 iounmap(info->map.cached); in pxa2xx_flash_probe() 109 if (info->map.cached) in pxa2xx_flash_remove() 110 iounmap(info->map.cached); in pxa2xx_flash_remove()
|
| /linux/tools/perf/util/ |
| A D | smt.c | 9 static bool cached; in smt_on() local 13 if (cached) in smt_on() 21 cached = true; in smt_on()
|
| /linux/tools/perf/arch/x86/util/ |
| A D | topdown.c | 14 static bool cached; in topdown_sys_has_perf_metrics() local 17 if (cached) in topdown_sys_has_perf_metrics() 30 cached = true; in topdown_sys_has_perf_metrics()
|
| A D | tsc.c | 61 static bool cached; in arch_get_tsc_freq() local 65 if (cached) in arch_get_tsc_freq() 68 cached = true; in arch_get_tsc_freq()
|
| /linux/fs/proc/ |
| A D | meminfo.c | 38 long cached; in meminfo_proc_show() local 48 cached = global_node_page_state(NR_FILE_PAGES) - in meminfo_proc_show() 50 if (cached < 0) in meminfo_proc_show() 51 cached = 0; in meminfo_proc_show() 64 show_val_kb(m, "Cached: ", cached); in meminfo_proc_show()
|
| /linux/drivers/gpu/drm/msm/dsi/phy/ |
| A D | dsi_phy_10nm.c | 474 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_10nm_pll_save_state() local 478 cached->pll_out_div = readl(pll_10nm->phy->pll_base + in dsi_10nm_pll_save_state() 480 cached->pll_out_div &= 0x3; in dsi_10nm_pll_save_state() 483 cached->bit_clk_div = cmn_clk_cfg0 & 0xf; in dsi_10nm_pll_save_state() 484 cached->pix_clk_div = (cmn_clk_cfg0 & 0xf0) >> 4; in dsi_10nm_pll_save_state() 487 cached->pll_mux = cmn_clk_cfg1 & 0x3; in dsi_10nm_pll_save_state() 490 pll_10nm->phy->id, cached->pll_out_div, cached->bit_clk_div, in dsi_10nm_pll_save_state() 491 cached->pix_clk_div, cached->pll_mux); in dsi_10nm_pll_save_state() 504 val |= cached->pll_out_div; in dsi_10nm_pll_restore_state() 507 writel(cached->bit_clk_div | (cached->pix_clk_div << 4), in dsi_10nm_pll_restore_state() [all …]
|
| A D | dsi_phy_7nm.c | 544 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_save_state() local 548 cached->pll_out_div = readl(pll_7nm->phy->pll_base + in dsi_7nm_pll_save_state() 550 cached->pll_out_div &= 0x3; in dsi_7nm_pll_save_state() 553 cached->bit_clk_div = cmn_clk_cfg0 & 0xf; in dsi_7nm_pll_save_state() 554 cached->pix_clk_div = (cmn_clk_cfg0 & 0xf0) >> 4; in dsi_7nm_pll_save_state() 557 cached->pll_mux = cmn_clk_cfg1 & 0x3; in dsi_7nm_pll_save_state() 560 pll_7nm->phy->id, cached->pll_out_div, cached->bit_clk_div, in dsi_7nm_pll_save_state() 561 cached->pix_clk_div, cached->pll_mux); in dsi_7nm_pll_save_state() 574 val |= cached->pll_out_div; in dsi_7nm_pll_restore_state() 577 writel(cached->bit_clk_div | (cached->pix_clk_div << 4), in dsi_7nm_pll_restore_state() [all …]
|
| /linux/drivers/mtd/devices/ |
| A D | phram.c | 37 bool cached; member 88 if (phram->cached) in phram_map() 104 if (phram->cached) { in phram_unmap() 128 bool cached = np ? !of_property_read_bool(np, "no-map") : false; in register_device() local 136 new->cached = cached; in register_device()
|
| /linux/tools/testing/selftests/drivers/net/lib/py/ |
| A D | env.py | 215 cached = self._required_cmd.get(comm, {}) 216 if cached.get(key) is None: 217 cached[key] = cmd("command -v -- " + comm, fail=False, 219 self._required_cmd[comm] = cached 220 return cached[key]
|
| /linux/Documentation/filesystems/ |
| A D | fuse-io.rst | 10 - cached 21 In cached mode reads may be satisfied from the page cache, and data may be 25 The cached mode has two sub modes controlling how writes are handled. The 31 WRITE requests, as well as updating any cached pages (and caching previously
|
| /linux/Documentation/admin-guide/device-mapper/ |
| A D | writecache.rst | 6 doesn't cache reads because reads are supposed to be cached in page cache 17 2. the underlying device that will be cached 58 new writes (however, writes to already cached blocks are 60 writes) and it will gradually writeback any cached 62 process with "dmsetup status". When the number of cached
|
| /linux/arch/arm/mach-omap2/ |
| A D | sram.c | 184 int cached = 1; in omap2_map_sram() local 194 cached = 0; in omap2_map_sram() 201 omap_sram_base = __arm_ioremap_exec(omap_sram_start, omap_sram_size, cached); in omap2_map_sram()
|
| /linux/Documentation/ABI/testing/ |
| A D | sysfs-class-iommu-intel-iommu | 15 The cached hardware capability register value 23 The cached hardware extended capability register
|
| /linux/fs/bcachefs/ |
| A D | extents.c | 135 if (!ret && !p.ptr.cached) in bch2_bkey_pick_read_device() 628 if (!p.ptr.cached && crc_is_compressed(p.crc)) in bch2_bkey_sectors_compressed() 654 if (p.ptr.cached) in bch2_bkey_replicas() 669 if (p->ptr.cached) in __extent_ptr_durability() 884 (!ptr->cached || in bch2_bkey_has_target() 1018 ptr->cached = true; in bch2_extent_ptr_set_cached() 1041 ptr->cached && in bch2_extent_normalize() 1068 if (ptr->cached) { in bch2_extent_normalize_by_opts() 1088 ptr->cached ? " cached" : ""); in bch2_extent_ptr_to_text() 1097 if (ptr->cached) in bch2_extent_ptr_to_text() [all …]
|
| A D | alloc_foreground.h | 184 bool cached) in bch2_alloc_sectors_append_ptrs_inlined() argument 197 ptr.cached = cached || in bch2_alloc_sectors_append_ptrs_inlined()
|
| /linux/drivers/gpu/drm/ttm/ |
| A D | ttm_agp_backend.c | 55 int ret, cached = ttm->caching == ttm_cached; in ttm_agp_bind() local 77 mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY; in ttm_agp_bind()
|
| /linux/arch/s390/appldata/ |
| A D | appldata_mem.c | 54 u64 cached; /* size of (used) cache, w/o buffers */ member 102 mem_data->cached = P2K(global_node_page_state(NR_FILE_PAGES) in appldata_get_mem_data()
|
| /linux/drivers/block/ |
| A D | ps3vram.c | 428 unsigned int cached, count; in ps3vram_read() local 449 cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; in ps3vram_read() 453 (unsigned int)from, cached, offset, avail, count); in ps3vram_read() 457 memcpy(buf, priv->xdr_buf + cached, avail); in ps3vram_read() 472 unsigned int cached, count; in ps3vram_write() local 490 cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; in ps3vram_write() 494 cached, offset, avail, count); in ps3vram_write() 498 memcpy(priv->xdr_buf + cached, buf, avail); in ps3vram_write()
|