| /linux/drivers/md/ |
| H A D | dm-cache-policy-internal.h | 16 static inline int policy_lookup(struct dm_cache_policy *p, dm_oblock_t oblock, dm_cblock_t *cblock, in policy_lookup() argument 19 return p->lookup(p, oblock, cblock, data_dir, fast_copy, background_queued); in policy_lookup() 23 dm_oblock_t oblock, dm_cblock_t *cblock, in policy_lookup_with_work() argument 29 return p->lookup(p, oblock, cblock, data_dir, fast_copy, NULL); in policy_lookup_with_work() 32 return p->lookup_with_work(p, oblock, cblock, data_dir, fast_copy, work); in policy_lookup_with_work() 48 static inline void policy_set_dirty(struct dm_cache_policy *p, dm_cblock_t cblock) in policy_set_dirty() argument 50 p->set_dirty(p, cblock); in policy_set_dirty() 53 static inline void policy_clear_dirty(struct dm_cache_policy *p, dm_cblock_t cblock) in policy_clear_dirty() argument 55 p->clear_dirty(p, cblock); in policy_clear_dirty() 59 dm_oblock_t oblock, dm_cblock_t cblock, in policy_load_mapping() argument [all …]
|
| H A D | dm-cache-policy.h | 33 dm_cblock_t cblock; member 57 int (*lookup)(struct dm_cache_policy *p, dm_oblock_t oblock, dm_cblock_t *cblock, 70 dm_oblock_t oblock, dm_cblock_t *cblock, 89 void (*set_dirty)(struct dm_cache_policy *p, dm_cblock_t cblock); 90 void (*clear_dirty)(struct dm_cache_policy *p, dm_cblock_t cblock); 97 dm_cblock_t cblock, bool dirty, 104 int (*invalidate_mapping)(struct dm_cache_policy *p, dm_cblock_t cblock); 110 uint32_t (*get_hint)(struct dm_cache_policy *p, dm_cblock_t cblock);
|
| H A D | dm-cache-target.c | 36 * cblock: index of a cache block 618 static void set_dirty(struct cache *cache, dm_cblock_t cblock) in set_dirty() argument 620 if (!test_and_set_bit(from_cblock(cblock), cache->dirty_bitset)) { in set_dirty() 622 policy_set_dirty(cache->policy, cblock); in set_dirty() 630 static void force_set_dirty(struct cache *cache, dm_cblock_t cblock) in force_set_dirty() argument 632 if (!test_and_set_bit(from_cblock(cblock), cache->dirty_bitset)) in force_set_dirty() 634 policy_set_dirty(cache->policy, cblock); in force_set_dirty() 637 static void force_clear_dirty(struct cache *cache, dm_cblock_t cblock) in force_clear_dirty() argument 639 if (test_and_clear_bit(from_cblock(cblock), cache->dirty_bitset)) { in force_clear_dirty() 644 policy_clear_dirty(cache->policy, cblock); in force_clear_dirty() 730 remap_to_cache(struct cache * cache,struct bio * bio,dm_cblock_t cblock) remap_to_cache() argument 771 remap_to_cache_dirty(struct cache * cache,struct bio * bio,dm_oblock_t oblock,dm_cblock_t cblock) remap_to_cache_dirty() argument 834 remap_to_origin_and_cache(struct cache * cache,struct bio * bio,dm_oblock_t oblock,dm_cblock_t cblock) remap_to_origin_and_cache() argument 1179 dm_cblock_t cblock = op->cblock; mg_complete() local 1484 invalidate_cblock(struct cache * cache,dm_cblock_t cblock) invalidate_cblock() argument 1564 invalidate_start(struct cache * cache,dm_cblock_t cblock,dm_oblock_t oblock,struct bio * bio) invalidate_start() argument 1623 dm_cblock_t cblock; map_bio() local 2807 load_mapping(void * context,dm_oblock_t oblock,dm_cblock_t cblock,bool dirty,uint32_t hint,bool hint_valid) load_mapping() argument 2821 load_filtered_mapping(void * context,dm_oblock_t oblock,dm_cblock_t cblock,bool dirty,uint32_t hint,bool hint_valid) load_filtered_mapping() argument [all...] |
| H A D | dm-cache-policy-smq.c | 746 * This assumes the cblock hasn't already been allocated. 1198 work.cblock = infer_cblock(mq, e); in queue_writeback() 1229 work.cblock = infer_cblock(mq, e); in queue_demotion() 1261 * We allocate the entry now to reserve the cblock. If the in queue_promotion() 1269 work.cblock = infer_cblock(mq, e); in queue_promotion() 1376 static int __lookup(struct smq_policy *mq, dm_oblock_t oblock, dm_cblock_t *cblock, in __lookup() argument 1390 *cblock = infer_cblock(mq, e); in __lookup() 1411 static int smq_lookup(struct dm_cache_policy *p, dm_oblock_t oblock, dm_cblock_t *cblock, in smq_lookup() argument 1420 r = __lookup(mq, oblock, cblock, in smq_lookup() 1429 dm_oblock_t oblock, dm_cblock_t *cblock, in smq_lookup_with_work() argument 1528 __smq_set_clear_dirty(struct smq_policy * mq,dm_cblock_t cblock,bool set) __smq_set_clear_dirty() argument 1541 smq_set_dirty(struct dm_cache_policy * p,dm_cblock_t cblock) smq_set_dirty() argument 1551 smq_clear_dirty(struct dm_cache_policy * p,dm_cblock_t cblock) smq_clear_dirty() argument 1561 random_level(dm_cblock_t cblock) random_level() argument 1567 smq_load_mapping(struct dm_cache_policy * p,dm_oblock_t oblock,dm_cblock_t cblock,bool dirty,uint32_t hint,bool hint_valid) smq_load_mapping() argument 1588 smq_invalidate_mapping(struct dm_cache_policy * p,dm_cblock_t cblock) smq_invalidate_mapping() argument 1603 smq_get_hint(struct dm_cache_policy * p,dm_cblock_t cblock) smq_get_hint() argument [all...] |
| H A D | dm-cache-metadata.c | 729 * little-endian format. The index is the cblock, the high 48bits of the 1227 static int __remove(struct dm_cache_metadata *cmd, dm_cblock_t cblock) in __remove() 1233 r = dm_array_set_value(&cmd->info, cmd->root, from_cblock(cblock), in __remove() 1242 int dm_cache_remove_mapping(struct dm_cache_metadata *cmd, dm_cblock_t cblock) in dm_cache_remove_mapping() 1247 r = __remove(cmd, cblock); in __insert() 1254 dm_cblock_t cblock, dm_oblock_t oblock) in __insert() 1261 r = dm_array_set_value(&cmd->info, cmd->root, from_cblock(cblock), in __insert() 1271 dm_cblock_t cblock, dm_oblock_t oblock) in dm_cache_insert_mapping() 1276 r = __insert(cmd, cblock, oblock); in policy_unchanged() 1518 static int __dirty(struct dm_cache_metadata *cmd, dm_cblock_t cblock, boo in __dirty() 1221 __remove(struct dm_cache_metadata * cmd,dm_cblock_t cblock) __remove() argument 1236 dm_cache_remove_mapping(struct dm_cache_metadata * cmd,dm_cblock_t cblock) dm_cache_remove_mapping() argument 1248 __insert(struct dm_cache_metadata * cmd,dm_cblock_t cblock,dm_oblock_t oblock) __insert() argument 1265 dm_cache_insert_mapping(struct dm_cache_metadata * cmd,dm_cblock_t cblock,dm_oblock_t oblock) dm_cache_insert_mapping() argument 1512 __dirty(struct dm_cache_metadata * cmd,dm_cblock_t cblock,bool dirty) __dirty() argument [all...] |
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | cache.rst | 310 message, which takes an arbitrary number of cblock ranges. Each cblock 312 of values from 5 to 9. Each cblock must be expressed as a decimal 313 value, in the future a variant message that takes cblock ranges 318 invalidate_cblocks [<cblock>|<cblock begin>-<cblock end>]*
|
| /linux/tools/perf/ |
| H A D | builtin-record.c | 292 static int record__aio_write(struct aiocb *cblock, int trace_fd, in record__aio_write() argument 297 cblock->aio_fildes = trace_fd; in record__aio_write() 298 cblock->aio_buf = buf; in record__aio_write() 299 cblock->aio_nbytes = size; in record__aio_write() 300 cblock->aio_offset = off; in record__aio_write() 301 cblock->aio_sigevent.sigev_notify = SIGEV_NONE; in record__aio_write() 304 rc = aio_write(cblock); in record__aio_write() 308 cblock->aio_fildes = -1; in record__aio_write() 317 static int record__aio_complete(struct mmap *md, struct aiocb *cblock) in record__aio_complete() argument 325 aio_errno = aio_error(cblock); in record__aio_complete() [all...] |