| /linux/tools/lib/api/fs/ |
| H A D | cgroup.c | 17 static struct cgroupfs_cache_entry *cached; variable 27 if (cached && !strcmp(cached->subsys, subsys)) { in cgroupfs_find_mountpoint() 28 if (strlen(cached->mountpoint) < maxlen) { in cgroupfs_find_mountpoint() 29 strcpy(buf, cached->mountpoint); in cgroupfs_find_mountpoint() 94 if (!cached) in cgroupfs_find_mountpoint() 95 cached = calloc(1, sizeof(*cached)); in cgroupfs_find_mountpoint() 97 if (cached) { in cgroupfs_find_mountpoint() 98 strncpy(cached->subsys, subsys, sizeof(cached->subsys) - 1); in cgroupfs_find_mountpoint() 99 strcpy(cached->mountpoint, mountpoint); in cgroupfs_find_mountpoint()
|
| /linux/fs/btrfs/ |
| H A D | extent-io-tree.h | 145 struct extent_state **cached); 147 u32 bits, struct extent_state **cached); 150 struct extent_state **cached) in btrfs_lock_extent() argument 152 return btrfs_lock_extent_bits(tree, start, end, EXTENT_LOCKED, cached); in btrfs_lock_extent() 156 u64 end, struct extent_state **cached) in btrfs_try_lock_extent() argument 158 return btrfs_try_lock_extent_bits(tree, start, end, EXTENT_LOCKED, cached); in btrfs_try_lock_extent() 178 u32 bits, struct extent_state **cached, 183 struct extent_state **cached) in btrfs_clear_extent_bit() argument 185 return btrfs_clear_extent_bit_changeset(tree, start, end, bits, cached, NULL); in btrfs_clear_extent_bit() 189 struct extent_state **cached) in btrfs_unlock_extent() argument [all …]
|
| H A D | extent-io-tree.c | 617 struct extent_state *cached; in btrfs_clear_extent_bit_changeset() local 652 cached = *cached_state; in btrfs_clear_extent_bit_changeset() 659 if (cached && extent_state_in_tree(cached) && in btrfs_clear_extent_bit_changeset() 660 cached->start <= start && cached->end > start) { in btrfs_clear_extent_bit_changeset() 662 refcount_dec(&cached->refs); in btrfs_clear_extent_bit_changeset() 663 state = cached; in btrfs_clear_extent_bit_changeset() 667 btrfs_free_extent_state(cached); in btrfs_clear_extent_bit_changeset() 1671 struct extent_state *cached; in btrfs_count_range_bits() local 1685 cached = *cached_state; in btrfs_count_range_bits() 1687 if (!extent_state_in_tree(cached)) in btrfs_count_range_bits() [all …]
|
| H A D | fiemap.c | 72 bool cached; member 117 if (!cache->cached) in emit_fiemap_extent() 252 cache->cached = false; in emit_fiemap_extent() 266 cache->cached = false; in emit_fiemap_extent() 270 cache->cached = true; in emit_fiemap_extent() 295 if (!cache->cached) in emit_last_fiemap_cache() 300 cache->cached = false; in emit_last_fiemap_cache() 822 if (cache.cached && cache.offset + cache.len >= last_extent_end) { in extent_fiemap()
|
| /linux/fs/lockd/ |
| H A D | mon.c | 322 struct nsm_handle *cached, *new = NULL; in nsm_get_handle() local 338 cached = nsm_lookup_hostname(&ln->nsm_handles, in nsm_get_handle() 341 cached = nsm_lookup_addr(&ln->nsm_handles, sap); in nsm_get_handle() 343 if (cached != NULL) { in nsm_get_handle() 344 refcount_inc(&cached->sm_count); in nsm_get_handle() 348 "cnt %d\n", cached->sm_name, in nsm_get_handle() 349 cached->sm_addrbuf, in nsm_get_handle() 350 refcount_read(&cached->sm_count)); in nsm_get_handle() 351 return cached; in nsm_get_handle() 382 struct nsm_handle *cached; in nsm_reboot_lookup() local [all …]
|
| /linux/lib/zstd/compress/ |
| H A D | hist.c | 102 { U32 cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() local 104 U32 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 109 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 114 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp() 119 c = cached; cached = MEM_read32(ip); ip += 4; in HIST_count_parallel_wksp()
|
| /linux/drivers/mtd/maps/ |
| H A D | pxa2xx-flash.c | 26 unsigned long start = (unsigned long)map->cached + from; in pxa2xx_map_inval_cache() 70 info->map.cached = ioremap_cache(info->map.phys, info->map.size); in pxa2xx_flash_probe() 71 if (!info->map.cached) in pxa2xx_flash_probe() 87 if (info->map.cached) in pxa2xx_flash_probe() 88 iounmap(info->map.cached); in pxa2xx_flash_probe() 109 if (info->map.cached) in pxa2xx_flash_remove() 110 iounmap(info->map.cached); in pxa2xx_flash_remove()
|
| /linux/tools/perf/util/ |
| H A D | smt.c | 9 static bool cached; in smt_on() local 13 if (cached) in smt_on() 21 cached = true; in smt_on()
|
| H A D | addr2line.c | 101 static bool cached; in addr2line_configure() local 104 if (!cached) { in addr2line_configure() 117 cached = true; in addr2line_configure() 122 cached = true; in addr2line_configure()
|
| H A D | tool_pmu.c | 342 static bool has_pmem, cached; in has_pmem() local 346 if (!cached) { in has_pmem() 349 cached = true; in has_pmem()
|
| /linux/tools/perf/arch/x86/util/ |
| H A D | topdown.c | 17 static bool cached; in topdown_sys_has_perf_metrics() local 20 if (cached) in topdown_sys_has_perf_metrics() 33 cached = true; in topdown_sys_has_perf_metrics()
|
| /linux/Documentation/filesystems/fuse/ |
| H A D | fuse-io.rst | 10 - cached 21 In cached mode reads may be satisfied from the page cache, and data may be 25 The cached mode has two sub modes controlling how writes are handled. The 31 WRITE requests, as well as updating any cached pages (and caching previously
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-fs-erofs | 24 currently used to drop in-memory pclusters and cached 27 - 1 : invalidate cached compressed folios 29 - 3 : drop in-memory pclusters and cached compressed folios
|
| H A D | sysfs-class-iommu-intel-iommu | 15 The cached hardware capability register value 23 The cached hardware extended capability register
|
| /linux/tools/testing/selftests/drivers/net/lib/py/ |
| H A D | env.py | 255 cached = self._required_cmd.get(comm, {}) 256 if cached.get(key) is None: 257 cached[key] = cmd("command -v -- " + comm, fail=False, 259 self._required_cmd[comm] = cached 260 return cached[key]
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | writecache.rst | 6 doesn't cache reads because reads are supposed to be cached in page cache 17 2. the underlying device that will be cached 58 new writes (however, writes to already cached blocks are 60 writes) and it will gradually writeback any cached 62 process with "dmsetup status". When the number of cached
|
| /linux/drivers/gpu/drm/msm/dsi/phy/ |
| H A D | dsi_phy_7nm.c | 634 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_save_state() local 639 cached->pll_out_div = readl(pll_7nm->phy->pll_base + in dsi_7nm_pll_save_state() 641 cached->pll_out_div &= 0x3; in dsi_7nm_pll_save_state() 644 cached->bit_clk_div = FIELD_GET(DSI_7nm_PHY_CMN_CLK_CFG0_DIV_CTRL_3_0__MASK, cmn_clk_cfg0); in dsi_7nm_pll_save_state() 645 cached->pix_clk_div = FIELD_GET(DSI_7nm_PHY_CMN_CLK_CFG0_DIV_CTRL_7_4__MASK, cmn_clk_cfg0); in dsi_7nm_pll_save_state() 648 cached->pll_mux = FIELD_GET(DSI_7nm_PHY_CMN_CLK_CFG1_DSICLK_SEL__MASK, cmn_clk_cfg1); in dsi_7nm_pll_save_state() 652 pll_7nm->phy->id, cached->pll_out_div, cached->bit_clk_div, in dsi_7nm_pll_save_state() 653 cached->pix_clk_div, cached->pll_mux); in dsi_7nm_pll_save_state() 659 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_7nm_pll_restore_state() local 665 val |= cached->pll_out_div; in dsi_7nm_pll_restore_state() [all …]
|
| /linux/arch/arm/mach-omap2/ |
| H A D | sram.c | 184 int cached = 1; in omap2_map_sram() local 194 cached = 0; in omap2_map_sram() 201 omap_sram_base = __arm_ioremap_exec(omap_sram_start, omap_sram_size, cached); in omap2_map_sram()
|
| /linux/drivers/block/ |
| H A D | ps3vram.c | 428 unsigned int cached, count; in ps3vram_read() local 449 cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; in ps3vram_read() 453 (unsigned int)from, cached, offset, avail, count); in ps3vram_read() 457 memcpy(buf, priv->xdr_buf + cached, avail); in ps3vram_read() 472 unsigned int cached, count; in ps3vram_write() local 490 cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; in ps3vram_write() 494 cached, offset, avail, count); in ps3vram_write() 498 memcpy(priv->xdr_buf + cached, buf, avail); in ps3vram_write()
|
| /linux/arch/s390/appldata/ |
| H A D | appldata_mem.c | 54 u64 cached; /* size of (used) cache, w/o buffers */ member 102 mem_data->cached = P2K(global_node_page_state(NR_FILE_PAGES) in appldata_get_mem_data()
|
| /linux/tools/perf/Documentation/ |
| H A D | perf-buildid-cache.txt | 45 Remove a cached binary which has same build-id of specified file 49 Purge all cached binaries including older caches which have specified 53 Purge all cached binaries. This will flush out entire cache.
|
| /linux/sound/soc/codecs/ |
| H A D | sigmadsp.c | 37 bool cached; member 148 ctrl->cached = true; in sigmadsp_ctrl_put() 165 if (!ctrl->cached) { in sigmadsp_ctrl_get() 172 ctrl->cached = true; in sigmadsp_ctrl_get() 682 if (ctrl->cached) in sigmadsp_activate_ctrl()
|
| /linux/Documentation/mm/ |
| H A D | page_frags.rst | 18 and tracks allows multiple calls to make use of a cached page. The 35 fragments, but the page fragments are cached at the ring or descriptor
|
| /linux/drivers/net/wireless/ath/ath11k/ |
| H A D | dp.c | 109 if (ring->cached) in ath11k_dp_srng_cleanup() 232 bool cached = false; in ath11k_dp_srng_setup() local 247 cached = true; in ath11k_dp_srng_setup() 250 cached = false; in ath11k_dp_srng_setup() 254 if (cached) in ath11k_dp_srng_setup() 323 if (cached) { in ath11k_dp_srng_setup() 325 ring->cached = 1; in ath11k_dp_srng_setup()
|
| /linux/sound/pci/asihpi/ |
| H A D | hpicmn.c | 179 int cached = 0; in control_cache_alloc_check() local 228 cached++; in control_cache_alloc_check() 237 cached, pC->p_info[info->control_index], in control_cache_alloc_check() 262 pC->init = (u16)cached; in control_cache_alloc_check()
|