| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_page_dirty.c | 55 bool vmw_bo_is_dirty(struct vmw_bo *vbo) in vmw_bo_is_dirty() argument 57 return vbo->dirty && (vbo->dirty->start < vbo->dirty->end); in vmw_bo_is_dirty() 68 static void vmw_bo_dirty_scan_pagetable(struct vmw_bo *vbo) in vmw_bo_dirty_scan_pagetable() argument 70 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_pagetable() 71 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_pagetable() 72 struct address_space *mapping = vbo->tbo.bdev->dev_mapping; in vmw_bo_dirty_scan_pagetable() 106 static void vmw_bo_dirty_scan_mkwrite(struct vmw_bo *vbo) in vmw_bo_dirty_scan_mkwrite() argument 108 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_mkwrite() 109 unsigned long offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_mkwrite() 110 struct address_space *mapping = vbo->tbo.bdev->dev_mapping; in vmw_bo_dirty_scan_mkwrite() [all …]
|
| H A D | vmwgfx_bo.c | 43 struct vmw_bo *vbo = to_vmw_bo(&bo->base); in vmw_bo_free() local 45 WARN_ON(kref_read(&vbo->tbo.base.refcount) != 0); in vmw_bo_free() 46 vmw_bo_unmap(vbo); in vmw_bo_free() 48 xa_destroy(&vbo->detached_resources); in vmw_bo_free() 49 WARN_ON(vbo->is_dumb && !vbo->dumb_surface); in vmw_bo_free() 50 if (vbo->is_dumb && vbo->dumb_surface) { in vmw_bo_free() 51 res = &vbo->dumb_surface->res; in vmw_bo_free() 52 WARN_ON(vbo != res->guest_memory_bo); in vmw_bo_free() 69 vmw_surface_unreference(&vbo->dumb_surface); in vmw_bo_free() 71 WARN_ON(!RB_EMPTY_ROOT(&vbo->res_tree)); in vmw_bo_free() [all …]
|
| H A D | vmwgfx_validation.c | 158 struct vmw_bo *vbo) in vmw_validation_find_bo_dup() argument 167 unsigned long key = (unsigned long) vbo; in vmw_validation_find_bo_dup() 179 if (entry->base.bo == &vbo->tbo) { in vmw_validation_find_bo_dup() 247 struct vmw_bo *vbo) in vmw_validation_add_bo() argument 251 bo_node = vmw_validation_find_bo_dup(ctx, vbo); in vmw_validation_add_bo() 260 bo_node->hash.key = (unsigned long) vbo; in vmw_validation_add_bo() 265 vmw_bo_reference(vbo); in vmw_validation_add_bo() 266 val_buf->bo = &vbo->tbo; in vmw_validation_add_bo() 384 struct vmw_bo *vbo, in vmw_validation_res_switch_backup() argument 395 val->new_guest_memory_bo = vbo; in vmw_validation_res_switch_backup() [all …]
|
| H A D | vmwgfx_prime.c | 95 struct vmw_bo *vbo; in vmw_prime_handle_to_fd() local 102 ret = vmw_user_bo_lookup(file_priv, handle, &vbo); in vmw_prime_handle_to_fd() 105 if (vbo && vbo->is_dumb) { in vmw_prime_handle_to_fd() 110 vbo, in vmw_prime_handle_to_fd() 120 vmw_user_bo_unref(&vbo); in vmw_prime_handle_to_fd()
|
| H A D | vmwgfx_cursor_plane.c | 153 static void vmw_cursor_mob_destroy(struct vmw_bo **vbo) in vmw_cursor_mob_destroy() argument 155 if (!(*vbo)) in vmw_cursor_mob_destroy() 158 ttm_bo_unpin(&(*vbo)->tbo); in vmw_cursor_mob_destroy() 159 vmw_bo_unreference(vbo); in vmw_cursor_mob_destroy() 174 struct vmw_bo *vbo = vps->cursor.mob; in vmw_cursor_mob_unmap() local 176 if (!vbo || !vbo->map.virtual) in vmw_cursor_mob_unmap() 179 ret = ttm_bo_reserve(&vbo->tbo, true, false, NULL); in vmw_cursor_mob_unmap() 181 vmw_bo_unmap(vbo); in vmw_cursor_mob_unmap() 182 ttm_bo_unreserve(&vbo->tbo); in vmw_cursor_mob_unmap() 434 struct vmw_bo *vbo = vps->cursor.mob; in vmw_cursor_mob_map() local [all …]
|
| H A D | vmwgfx_ttm_buffer.c | 568 struct vmw_bo *vbo; in vmw_bo_create_and_populate() local 579 ret = vmw_bo_create(dev_priv, &bo_params, &vbo); in vmw_bo_create_and_populate() 583 ret = vmw_ttm_populate(vbo->tbo.bdev, vbo->tbo.ttm, &ctx); in vmw_bo_create_and_populate() 586 container_of(vbo->tbo.ttm, struct vmw_ttm_tt, dma_ttm); in vmw_bo_create_and_populate() 590 ttm_bo_unreserve(&vbo->tbo); in vmw_bo_create_and_populate() 593 *bo_p = vbo; in vmw_bo_create_and_populate()
|
| H A D | vmwgfx_validation.h | 155 struct vmw_bo *vbo); 171 struct vmw_bo *vbo,
|
| H A D | vmwgfx_drv.h | 815 void vmw_resource_unbind_list(struct vmw_bo *vbo); 821 int vmw_resources_clean(struct vmw_bo *vbo, pgoff_t start, 1385 bool vmw_bo_is_dirty(struct vmw_bo *vbo); 1386 void vmw_bo_dirty_scan(struct vmw_bo *vbo); 1387 int vmw_bo_dirty_add(struct vmw_bo *vbo); 1388 void vmw_bo_dirty_clear(struct vmw_bo *vbo); 1391 void vmw_bo_dirty_release(struct vmw_bo *vbo); 1392 void vmw_bo_dirty_unmap(struct vmw_bo *vbo,
|
| H A D | vmwgfx_drv.c | 377 struct vmw_bo *vbo; in vmw_dummy_query_bo_create() local 395 ret = vmw_bo_create(dev_priv, &bo_params, &vbo); in vmw_dummy_query_bo_create() 399 ret = ttm_bo_kmap(&vbo->tbo, 0, 1, &map); in vmw_dummy_query_bo_create() 407 vmw_bo_pin_reserved(vbo, false); in vmw_dummy_query_bo_create() 408 ttm_bo_unreserve(&vbo->tbo); in vmw_dummy_query_bo_create() 412 vmw_bo_unreference(&vbo); in vmw_dummy_query_bo_create() 414 dev_priv->dummy_query_bo = vbo; in vmw_dummy_query_bo_create()
|
| H A D | vmwgfx_surface.c | 2241 struct vmw_bo *vbo = NULL; in vmw_dumb_create() local 2276 &vbo); in vmw_dumb_create() 2278 drm_gem_object_put(&vbo->tbo.base); in vmw_dumb_create() 2320 vbo = res->guest_memory_bo; in vmw_dumb_create() 2321 vbo->is_dumb = true; in vmw_dumb_create() 2322 vbo->dumb_surface = vmw_res_to_srf(res); in vmw_dumb_create() 2323 drm_gem_object_put(&vbo->tbo.base); in vmw_dumb_create() 2328 struct vmw_user_surface *usurf = container_of(vbo->dumb_surface, in vmw_dumb_create()
|
| H A D | vmwgfx_execbuf.c | 51 struct vmw_bo *vbo; member 1160 reloc->vbo = vmw_bo; in vmw_translate_mob_ptr() 1217 reloc->vbo = vmw_bo; in vmw_translate_guest_ptr() 1703 struct vmw_bo *vbo; in vmw_cmd_res_switch_backup() local 1711 ret = vmw_translate_mob_ptr(dev_priv, sw_context, buf_id, &vbo); in vmw_cmd_res_switch_backup() 1715 vmw_validation_res_switch_backup(sw_context->ctx, info, vbo, in vmw_cmd_res_switch_backup() 3762 bo = &reloc->vbo->tbo; in vmw_apply_relocations()
|
| /linux/fs/ntfs3/ |
| H A D | file.c | 449 static long ntfs_fallocate(struct file *file, int mode, loff_t vbo, loff_t len) in ntfs_fallocate() argument 456 loff_t end = vbo + len; in ntfs_fallocate() 457 loff_t vbo_down = round_down(vbo, max_t(unsigned long, in ntfs_fallocate() 519 err = attr_punch_hole(ni, vbo, len, &frame_size); in ntfs_fallocate() 532 vbo_a = (vbo + mask) & ~mask; in ntfs_fallocate() 536 from = min_t(loff_t, ni->i_valid, vbo); in ntfs_fallocate() 556 if (vbo < end_a && end_a < end) { in ntfs_fallocate() 568 err = filemap_write_and_wait_range(mapping, vbo_down, vbo); in ntfs_fallocate() 583 err = attr_collapse_range(ni, vbo, len); in ntfs_fallocate() 601 err = attr_insert_range(ni, vbo, len); in ntfs_fallocate() [all …]
|
| H A D | ntfs_fs.h | 490 int attr_collapse_range(struct ntfs_inode *ni, u64 vbo, u64 bytes); 491 int attr_insert_range(struct ntfs_inode *ni, u64 vbo, u64 bytes); 492 int attr_punch_hole(struct ntfs_inode *ni, u64 vbo, u64 bytes, u32 *frame_size); 658 u64 vbo, const void *buf, size_t bytes, int sync); 660 const struct runs_tree *run, u64 vbo); 662 u64 vbo, void *buf, u32 bytes, struct ntfs_buffers *nb, 665 const struct runs_tree *run, u64 vbo, in ntfs_read_run_nb() argument 669 return ntfs_read_run_nb_ra(sbi, run, vbo, buf, bytes, nb, NULL); in ntfs_read_run_nb() 672 u64 vbo, struct NTFS_RECORD_HEADER *rhdr, u32 bytes, 675 const struct runs_tree *run, u64 vbo, in ntfs_read_bh() argument [all …]
|
| H A D | fslog.c | 911 static inline void *alloc_rsttbl_from_idx(struct RESTART_TABLE **tbl, u32 vbo) in alloc_rsttbl_from_idx() argument 920 if (vbo >= bytes) { in alloc_rsttbl_from_idx() 925 u32 bytes2idx = vbo - bytes; in alloc_rsttbl_from_idx() 937 e = Add2Ptr(rt, vbo); in alloc_rsttbl_from_idx() 949 if (off == vbo) { in alloc_rsttbl_from_idx() 971 if (off == vbo) { in alloc_rsttbl_from_idx() 1001 u32 vbo; member 1083 u32 vbo = (lsn << log->seq_num_bits) >> (log->seq_num_bits - 3); in lsn_to_vbo() local 1085 return vbo; in lsn_to_vbo() 1144 static int read_log_page(struct ntfs_log *log, u32 vbo, in read_log_page() argument [all …]
|
| H A D | index.c | 208 size_t data_size, valid_size, vbo, off = bit >> 3; in bmp_buf_get() local 261 vbo = off & ~(size_t)sbi->block_mask; in bmp_buf_get() 263 bbuf->new_valid = vbo + blocksize; in bmp_buf_get() 269 if (vbo >= valid_size) { in bmp_buf_get() 271 } else if (vbo + blocksize > valid_size) { in bmp_buf_get() 373 size_t vbo = from >> 3; in scan_nres_bitmap() local 374 sector_t blk = (vbo & sbi->cluster_mask) >> sb->s_blocksize_bits; in scan_nres_bitmap() 375 sector_t vblock = vbo >> sb->s_blocksize_bits; in scan_nres_bitmap() 388 vcn = vbo >> sbi->cluster_bits; in scan_nres_bitmap() 420 vbo = (u64)vblock << sb->s_blocksize_bits; in scan_nres_bitmap() [all …]
|
| H A D | bitmap.c | 508 size_t wpos, wbit, iw, vbo; in wnd_rescan() local 519 vbo = 0; in wnd_rescan() 537 vbo * 8 - prev_tail, in wnd_rescan() 552 u32 off = vbo & sbi->cluster_mask; in wnd_rescan() 554 if (!run_lookup_entry(&wnd->run, vbo >> cluster_bits, in wnd_rescan() 585 wbit = vbo * 8; in wnd_rescan() 629 vbo += blocksize; in wnd_rescan() 699 size_t vbo; in wnd_map() local 707 vbo = (u64)iw << sb->s_blocksize_bits; in wnd_map() 709 if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen, in wnd_map() [all …]
|
| H A D | record.c | 121 u64 vbo = (u64)mi->rno << sbi->record_bits; in mi_read() local 133 err = ntfs_read_bh(sbi, run, vbo, &rec->rhdr, bpr, &mi->nb); in mi_read() 152 vbo >> sbi->cluster_bits); in mi_read() 162 err = ntfs_read_bh(sbi, run, vbo, &rec->rhdr, bpr, &mi->nb); in mi_read() 420 u64 vbo = (u64)rno << sbi->record_bits; in mi_format_new() local 461 err = ntfs_get_bh(sbi, &ni->file.run, vbo, sbi->record_size, in mi_format_new()
|
| H A D | frecord.c | 902 u64 vbo; in ni_ins_attr_ext() local 933 vbo = is_mft_data ? ((u64)svcn << sbi->cluster_bits) : 0; in ni_ins_attr_ext() 949 vbo <= ((u64)mi->rno << sbi->record_bits))) { in ni_ins_attr_ext() 996 if (is_mft_data && vbo <= ((u64)rno << sbi->record_bits)) { in ni_ins_attr_ext() 1890 u64 frame_vbo, vbo = folio_pos(folio); in ni_read_folio_cmpr() local 1898 if (vbo >= i_size_read(&ni->vfs_inode)) { in ni_read_folio_cmpr() 1913 frame = vbo >> frame_bits; in ni_read_folio_cmpr() 1915 idx = (vbo - frame_vbo) >> PAGE_SHIFT; in ni_read_folio_cmpr() 1980 u64 vbo; in ni_decompress_file() local 2011 for (vbo = 0; vbo < i_size; vbo += bytes) { in ni_decompress_file() [all …]
|
| /linux/drivers/gpu/drm/imx/ipuv3/ |
| H A D | ipuv3-plane.c | 378 unsigned long eba, ubo, vbo, old_ubo, old_vbo, alpha_eba; in ipu_plane_atomic_check() local 466 vbo = drm_plane_state_to_vbo(new_state); in ipu_plane_atomic_check() 468 if (vbo & 0x7 || vbo > 0xfffff8) in ipu_plane_atomic_check() 473 if (vbo != old_vbo) in ipu_plane_atomic_check() 586 unsigned long eba, ubo, vbo; in ipu_plane_atomic_update() local 710 vbo = drm_plane_state_to_vbo(new_state); in ipu_plane_atomic_update() 714 swap(ubo, vbo); in ipu_plane_atomic_update() 717 fb->pitches[1], ubo, vbo); in ipu_plane_atomic_update() 720 "phy = %lu %lu %lu, x = %d, y = %d", eba, ubo, vbo, in ipu_plane_atomic_update()
|