| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_page_dirty.c | 57 return vbo->dirty && (vbo->dirty->start < vbo->dirty->end); in vmw_bo_is_dirty() 70 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_pagetable() local 77 offset, dirty->bitmap_size, in vmw_bo_dirty_scan_pagetable() 78 offset, &dirty->bitmap[0], in vmw_bo_dirty_scan_pagetable() 79 &dirty->start, &dirty->end); in vmw_bo_dirty_scan_pagetable() 81 dirty->change_count++; in vmw_bo_dirty_scan_pagetable() 83 dirty->change_count = 0; in vmw_bo_dirty_scan_pagetable() 85 if (dirty->change_count > VMW_DIRTY_NUM_CHANGE_TRIGGERS) { in vmw_bo_dirty_scan_pagetable() 86 dirty->change_count = 0; in vmw_bo_dirty_scan_pagetable() 87 dirty->method = VMW_BO_DIRTY_MKWRITE; in vmw_bo_dirty_scan_pagetable() [all …]
|
| H A D | vmwgfx_scrn.c | 990 static void vmw_sou_surface_fifo_commit(struct vmw_kms_dirty *dirty) in vmw_sou_surface_fifo_commit() argument 993 container_of(dirty, typeof(*sdirty), base); in vmw_sou_surface_fifo_commit() 994 struct vmw_kms_sou_dirty_cmd *cmd = dirty->cmd; in vmw_sou_surface_fifo_commit() 995 s32 trans_x = dirty->unit->crtc.x - sdirty->dst_x; in vmw_sou_surface_fifo_commit() 996 s32 trans_y = dirty->unit->crtc.y - sdirty->dst_y; in vmw_sou_surface_fifo_commit() 997 size_t region_size = dirty->num_hits * sizeof(SVGASignedRect); in vmw_sou_surface_fifo_commit() 1001 if (!dirty->num_hits) { in vmw_sou_surface_fifo_commit() 1002 vmw_cmd_commit(dirty->dev_priv, 0); in vmw_sou_surface_fifo_commit() 1024 cmd->body.destScreenId = dirty->unit->unit; in vmw_sou_surface_fifo_commit() 1027 for (i = 0; i < dirty->num_hits; ++i, ++blit) { in vmw_sou_surface_fifo_commit() [all …]
|
| H A D | vmwgfx_kms.c | 433 if (!bo->dirty && surf && surf->res.dirty) in vmw_framebuffer_surface_destroy() 497 .dirty = drm_atomic_helper_dirtyfb, 603 .dirty = drm_atomic_helper_dirtyfb, 774 if (surface->res.dirty == NULL) in vmw_kms_fb_create() 1512 struct vmw_kms_dirty *dirty) in vmw_kms_helper_dirty() argument 1519 dirty->dev_priv = dev_priv; in vmw_kms_helper_dirty() 1522 if (dirty->crtc) { in vmw_kms_helper_dirty() 1523 units[num_units++] = vmw_crtc_to_du(dirty->crtc); in vmw_kms_helper_dirty() 1543 dirty->unit = unit; in vmw_kms_helper_dirty() 1544 if (dirty->fifo_reserve_size > 0) { in vmw_kms_helper_dirty() [all …]
|
| H A D | vmwgfx_stdu.c | 466 static void vmw_stdu_bo_cpu_clip(struct vmw_kms_dirty *dirty) in vmw_stdu_bo_cpu_clip() argument 469 container_of(dirty, struct vmw_stdu_dirty, base); in vmw_stdu_bo_cpu_clip() 471 dirty->num_hits = 1; in vmw_stdu_bo_cpu_clip() 474 ddirty->left = min_t(s32, ddirty->left, dirty->unit_x1); in vmw_stdu_bo_cpu_clip() 475 ddirty->top = min_t(s32, ddirty->top, dirty->unit_y1); in vmw_stdu_bo_cpu_clip() 476 ddirty->right = max_t(s32, ddirty->right, dirty->unit_x2); in vmw_stdu_bo_cpu_clip() 477 ddirty->bottom = max_t(s32, ddirty->bottom, dirty->unit_y2); in vmw_stdu_bo_cpu_clip() 484 ddirty->fb_left = min_t(s32, ddirty->fb_left, dirty->fb_x); in vmw_stdu_bo_cpu_clip() 485 ddirty->fb_top = min_t(s32, ddirty->fb_top, dirty->fb_y); in vmw_stdu_bo_cpu_clip() 497 static void vmw_stdu_bo_cpu_commit(struct vmw_kms_dirty *dirty) in vmw_stdu_bo_cpu_commit() argument [all …]
|
| H A D | vmwgfx_surface.c | 643 WARN_ON(res->dirty); in vmw_user_surface_free() 1291 if (res->guest_memory_bo->dirty && res->guest_memory_dirty) { in vmw_gb_surface_bind() 1766 static void vmw_subres_dirty_add(struct vmw_surface_dirty *dirty, in vmw_subres_dirty_add() argument 1770 const struct vmw_surface_cache *cache = &dirty->cache; in vmw_subres_dirty_add() 1771 SVGA3dBox *box = &dirty->boxes[loc_start->sub_resource]; in vmw_subres_dirty_add() 1776 if (WARN_ON(loc_start->sub_resource >= dirty->num_subres)) in vmw_subres_dirty_add() 1814 static void vmw_subres_dirty_full(struct vmw_surface_dirty *dirty, u32 subres) in vmw_subres_dirty_full() argument 1816 const struct vmw_surface_cache *cache = &dirty->cache; in vmw_subres_dirty_full() 1819 SVGA3dBox *box = &dirty->boxes[subres]; in vmw_subres_dirty_full() 1836 struct vmw_surface_dirty *dirty = in vmw_surface_tex_dirty_range_add() local [all …]
|
| H A D | vmwgfx_validation.c | 82 u32 dirty : 1; member 288 u32 dirty, in vmw_validation_add_resource() argument 335 if (dirty) { in vmw_validation_add_resource() 338 node->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0; in vmw_validation_add_resource() 358 void *val_private, u32 dirty) in vmw_validation_res_set_dirty() argument 362 if (!dirty) in vmw_validation_res_set_dirty() 368 val->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0; in vmw_validation_res_set_dirty() 481 val->dirty, in vmw_validation_res_unreserve() 565 if (vbo->dirty) in vmw_validation_bo_validate() 593 val->dirty); in vmw_validation_res_validate()
|
| /linux/sound/pci/ctxfi/ |
| H A D | cthw20k1.c | 101 union src_dirty dirty; member 124 union src_mgr_dirty dirty; member 148 union srcimp_mgr_dirty dirty; member 181 ctl->dirty.bf.ctl = 1; in src_set_state() 190 ctl->dirty.bf.ctl = 1; in src_set_bm() 199 ctl->dirty.bf.ctl = 1; in src_set_rsr() 208 ctl->dirty.bf.ctl = 1; in src_set_sf() 217 ctl->dirty.bf.ctl = 1; in src_set_wr() 226 ctl->dirty.bf.ctl = 1; in src_set_pm() 235 ctl->dirty.bf.ctl = 1; in src_set_rom() [all …]
|
| H A D | cthw20k2.c | 101 union src_dirty dirty; member 124 union src_mgr_dirty dirty; member 148 union srcimp_mgr_dirty dirty; member 181 ctl->dirty.bf.ctl = 1; in src_set_state() 190 ctl->dirty.bf.ctl = 1; in src_set_bm() 199 ctl->dirty.bf.ctl = 1; in src_set_rsr() 208 ctl->dirty.bf.ctl = 1; in src_set_sf() 217 ctl->dirty.bf.ctl = 1; in src_set_wr() 226 ctl->dirty.bf.ctl = 1; in src_set_pm() 235 ctl->dirty.bf.ctl = 1; in src_set_rom() [all …]
|
| /linux/fs/ubifs/ |
| H A D | lprops.c | 32 return lprops->free + lprops->dirty; in get_heap_comp_val() 34 return lprops->dirty; in get_heap_comp_val() 407 if (lprops->free + lprops->dirty == c->leb_size) { in ubifs_categorize_lprops() 415 if (lprops->dirty + lprops->free >= c->min_idx_node_sz) in ubifs_categorize_lprops() 418 if (lprops->dirty >= c->dead_wm && in ubifs_categorize_lprops() 419 lprops->dirty > lprops->free) in ubifs_categorize_lprops() 522 int free, int dirty, int flags, in ubifs_change_lp() argument 532 lprops->lnum, free, dirty, flags); in ubifs_change_lp() 545 ubifs_assert(c, dirty == LPROPS_NC || dirty >= 0); in ubifs_change_lp() 554 ubifs_assert(c, !(lprops->free & 7) && !(lprops->dirty & 7)); in ubifs_change_lp() [all …]
|
| H A D | find.c | 54 if (lprops->free + lprops->dirty >= c->dark_wm) in valuable() 97 if (lprops->free + lprops->dirty < data->min_space) in scan_for_dirty_cb() 103 if (lprops->free + lprops->dirty == c->leb_size) { in scan_for_dirty_cb() 107 } else if (lprops->dirty < c->dead_wm) in scan_for_dirty_cb() 138 if (lprops->free + lprops->dirty < min_space) in scan_for_dirty() 140 if (lprops->dirty < c->dead_wm) in scan_for_dirty() 154 if (lprops->free + lprops->dirty < min_space) in scan_for_dirty() 158 if (lprops->dirty < c->dead_wm) in scan_for_dirty() 180 ubifs_assert(c, lprops->free + lprops->dirty >= min_space); in scan_for_dirty() 181 ubifs_assert(c, lprops->dirty >= c->dead_wm || in scan_for_dirty() [all …]
|
| H A D | lpt.c | 353 pack_bits(c, &addr, &pos, pnode->lprops[i].dirty >> 3, in ubifs_pack_pnode() 415 pack_bits(c, &addr, &pos, ltab[i].dirty, c->lpt_spc_bits); in ubifs_pack_ltab() 452 void ubifs_add_lpt_dirt(struct ubifs_info *c, int lnum, int dirty) in ubifs_add_lpt_dirt() argument 454 if (!dirty || !lnum) in ubifs_add_lpt_dirt() 457 lnum, dirty, c->ltab[lnum - c->lpt_first].dirty); in ubifs_add_lpt_dirt() 459 c->ltab[lnum - c->lpt_first].dirty += dirty; in ubifs_add_lpt_dirt() 469 static void set_ltab(struct ubifs_info *c, int lnum, int free, int dirty) in set_ltab() argument 473 c->ltab[lnum - c->lpt_first].dirty, free, dirty); in set_ltab() 476 c->ltab[lnum - c->lpt_first].dirty = dirty; in set_ltab() 644 ltab[i].dirty = 0; in ubifs_create_dflt_lpt() [all …]
|
| H A D | replay.c | 76 int dirty; member 91 int err = 0, dirty; in set_bud_lprops() local 101 dirty = lp->dirty; in set_bud_lprops() 102 if (b->bud->start == 0 && (lp->free != c->leb_size || lp->dirty != 0)) { in set_bud_lprops() 123 lp->free, lp->dirty); in set_bud_lprops() 125 lp->free, lp->dirty); in set_bud_lprops() 126 dirty -= c->leb_size - lp->free; in set_bud_lprops() 134 if (dirty != 0) in set_bud_lprops() 136 b->bud->lnum, lp->free, lp->dirty, b->free, in set_bud_lprops() 137 b->dirty); in set_bud_lprops() [all …]
|
| H A D | lpt_commit.c | 129 static void upd_ltab(struct ubifs_info *c, int lnum, int free, int dirty) in upd_ltab() argument 133 c->ltab[lnum - c->lpt_first].dirty, free, dirty); in upd_ltab() 136 c->ltab[lnum - c->lpt_first].dirty += dirty; in upd_ltab() 691 else if (c->ltab[i].free + c->ltab[i].dirty == c->leb_size) in need_write_all() 715 if (c->ltab[i].dirty > 0 && in lpt_tgc_start() 716 c->ltab[i].free + c->ltab[i].dirty == c->leb_size) { in lpt_tgc_start() 719 c->ltab[i].dirty = 0; in lpt_tgc_start() 1152 int i, lnum = -1, dirty = 0; in lpt_gc() local 1158 c->ltab[i].free + c->ltab[i].dirty == c->leb_size) in lpt_gc() 1160 if (c->ltab[i].dirty > dirty) { in lpt_gc() [all …]
|
| /linux/drivers/vfio/pci/pds/ |
| H A D | dirty.c | 21 return pds_vfio->dirty.is_enabled; in pds_vfio_dirty_is_enabled() 26 pds_vfio->dirty.is_enabled = true; in pds_vfio_dirty_set_enabled() 31 pds_vfio->dirty.is_enabled = false; in pds_vfio_dirty_set_disabled() 94 static void pds_vfio_dirty_free_bitmaps(struct pds_vfio_dirty *dirty) in pds_vfio_dirty_free_bitmaps() argument 96 if (!dirty->regions) in pds_vfio_dirty_free_bitmaps() 99 for (int i = 0; i < dirty->num_regions; i++) { in pds_vfio_dirty_free_bitmaps() 100 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_dirty_free_bitmaps() 128 struct pds_vfio_dirty *dirty = &pds_vfio->dirty; in pds_vfio_dirty_free_sgl() local 130 if (!dirty->regions) in pds_vfio_dirty_free_sgl() 133 for (int i = 0; i < dirty->num_regions; i++) { in pds_vfio_dirty_free_sgl() [all …]
|
| /linux/Documentation/mm/ |
| H A D | arch_pgtable_helpers.rst | 25 | pte_dirty | Tests a dirty PTE | 33 | pte_soft_dirty | Tests a soft dirty PTE | 35 | pte_swp_soft_dirty | Tests a soft dirty swapped PTE | 41 | pte_mkdirty | Creates a dirty PTE | 55 | pte_mksoft_dirty | Creates a soft dirty PTE | 57 | pte_clear_soft_dirty | Clears a soft dirty PTE | 59 | pte_swp_mksoft_dirty | Creates a soft dirty swapped PTE | 61 | pte_swp_clear_soft_dirty | Clears a soft dirty swapped PTE | 95 | pmd_dirty | Tests a dirty PMD | 103 | pmd_soft_dirty | Tests a soft dirty PMD | [all …]
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | soft-dirty.rst | 5 The soft-dirty is a bit on a PTE which helps to track which pages a task 8 1. Clear soft-dirty bits from the task's PTEs. 15 3. Read soft-dirty bits from the PTEs. 18 64-bit qword is the soft-dirty one. If set, the respective PTE was 23 when the soft-dirty bit is cleared. So, after this, when the task tries to 25 the soft-dirty bit on the respective PTE. 28 soft-dirty bits clear, the #PF-s that occur after that are processed fast. 30 the kernel does is finds this fact out and puts both writable and soft-dirty 34 there is still a scenario when we can lose soft dirty bits -- a task 37 including soft dirty bits. To notify user space application about such [all …]
|
| /linux/fs/jffs2/ |
| H A D | debug.c | 163 uint32_t free = 0, dirty = 0, used = 0, wasted = 0, in __jffs2_dbg_superblock_counts() local 171 dirty += c->gcblock->dirty_size; in __jffs2_dbg_superblock_counts() 179 dirty += c->nextblock->dirty_size; in __jffs2_dbg_superblock_counts() 187 dirty += jeb->dirty_size; in __jffs2_dbg_superblock_counts() 195 dirty += jeb->dirty_size; in __jffs2_dbg_superblock_counts() 203 dirty += jeb->dirty_size; in __jffs2_dbg_superblock_counts() 211 dirty += jeb->dirty_size; in __jffs2_dbg_superblock_counts() 219 dirty += jeb->dirty_size; in __jffs2_dbg_superblock_counts() 227 dirty += jeb->dirty_size; in __jffs2_dbg_superblock_counts() 235 dirty += jeb->dirty_size; in __jffs2_dbg_superblock_counts() [all …]
|
| /linux/mm/ |
| H A D | page-writeback.c | 8 * Contains functions related to writing back dirty pages at the 58 * Estimate write bandwidth or update dirty limit at 200ms intervals. 90 * The generator of dirty data starts writeback at this percentage 108 * The longest time for which data is allowed to remain dirty 227 * user-configurable dirty ratio is the effective number of pages that 231 * Because the user is allowed to specify the dirty limit globally as 232 * absolute number of bytes, calculating the per-zone dirty limit can 241 * Return: the node's number of pages potentially available for dirty 242 * page cache. This is the base value for the per-node dirty limits. 314 * Return: the global number of pages potentially available for dirty in highmem_dirtyable_memory() 450 unsigned long dirty; node_dirty_limit() local 980 pos_ratio_polynom(unsigned long setpoint,unsigned long dirty,unsigned long limit) pos_ratio_polynom() argument 1343 unsigned long dirty = dtc->dirty; wb_update_dirty_ratelimit() local 1565 dirty_poll_interval(unsigned long dirty,unsigned long thresh) dirty_poll_interval() argument 1712 unsigned long dirty, thresh; domain_poll_intv() local 1737 unsigned long dirty, thresh, bg_thresh; domain_dirty_freerun() local [all...] |
| /linux/drivers/misc/sgi-gru/ |
| H A D | gruhandles.c | 152 unsigned long vaddr, int asid, int dirty, in tfh_write_only() argument 159 tfh->dirty = dirty; in tfh_write_only() 168 unsigned long vaddr, int asid, int dirty, in tfh_write_restart() argument 175 tfh->dirty = dirty; in tfh_write_restart()
|
| /linux/Documentation/dev-tools/ |
| H A D | checkuapi.rst | 24 will use any dirty changes in tree to UAPI files. If there are no 25 dirty changes, HEAD will be used. 76 Installing user-facing UAPI headers from dirty tree... OK 78 Checking changes to UAPI headers between HEAD and dirty tree... 100 Installing user-facing UAPI headers from dirty tree... OK 102 Checking changes to UAPI headers between HEAD and dirty tree... 103 ==== ABI differences detected in include/linux/bpf.h from HEAD -> dirty tree ==== 124 Installing user-facing UAPI headers from dirty tree... OK 126 Checking changes to UAPI headers between HEAD and dirty tree... 150 Installing user-facing UAPI headers from dirty tree... OK [all …]
|
| /linux/fs/ntfs3/ |
| H A D | record.c | 58 mi->dirty = true; in mi_new_attt_id() 140 mi->dirty = true; in mi_read() 167 mi->dirty = true; in mi_read() 396 if (!mi->dirty) in mi_write() 409 mi->dirty = false; in mi_write() 450 mi->dirty = true; in mi_format_new() 530 mi->dirty = true; in mi_insert_attr() 557 ni->mi.dirty = true; in mi_remove_attr() 564 mi->dirty = true; in mi_remove_attr() 614 mi->dirty = true; in mi_resize_attr() [all …]
|
| /linux/drivers/char/xilinx_hwicap/ |
| H A D | buffer_icap.c | 272 bool dirty = false; in buffer_icap_set_configuration() local 281 dirty = true; in buffer_icap_set_configuration() 300 dirty = false; in buffer_icap_set_configuration() 304 if (dirty) { in buffer_icap_set_configuration()
|
| /linux/drivers/md/ |
| H A D | dm-cache-policy-smq.c | 44 bool dirty:1; member 725 e->dirty = true; /* FIXME: audit */ in init_entry() 828 struct queue dirty; member 889 struct queue *q = &mq->dirty; in __update_writeback_sentinels() 935 q_push(&mq->dirty, sentinel); in __sentinels_init() 961 q_del(e->dirty ? &mq->dirty : &mq->clean, e); in del_queue() 966 if (e->dirty) in push_queue() 967 q_push(&mq->dirty, e); in push_queue() 982 if (e->dirty) in push_queue_front() 983 q_push_front(&mq->dirty, e); in push_queue_front() [all …]
|
| /linux/drivers/md/persistent-data/ |
| H A D | dm-bitset.c | 112 if (!info->current_index_set || !info->dirty) in dm_bitset_flush() 124 info->dirty = false; in dm_bitset_flush() 143 info->dirty = false; in read_bits() 177 info->dirty = true; in dm_bitset_set_bit() 194 info->dirty = true; in dm_bitset_clear_bit()
|
| /linux/virt/kvm/ |
| H A D | Kconfig | 37 # Allow enabling both the dirty bitmap and dirty ring. Only architectures 38 # that need to dirty memory outside of a vCPU context should select this.
|