| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_ttm_buffer.c | 473 struct ttm_resource *new_mem) in vmw_move_notify() argument 475 vmw_bo_move_notify(bo, new_mem); in vmw_move_notify() 476 vmw_query_move_notify(bo, old_mem, new_mem); in vmw_move_notify() 499 struct ttm_resource *new_mem, in vmw_move() argument 506 new_man = ttm_manager_type(bo->bdev, new_mem->mem_type); in vmw_move() 510 if (new_man->use_tt && !vmw_memtype_is_system(new_mem->mem_type)) { in vmw_move() 511 ret = vmw_ttm_bind(bo->bdev, bo->ttm, new_mem); in vmw_move() 518 ttm_bo_move_null(bo, new_mem); in vmw_move() 522 vmw_move_notify(bo, bo->resource, new_mem); in vmw_move() 526 ttm_bo_move_null(bo, new_mem); in vmw_move() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_ttm.c | 373 struct ttm_resource *new_mem, in amdgpu_move_blit() argument 385 dst.mem = new_mem; in amdgpu_move_blit() 390 new_mem->size, in amdgpu_move_blit() 414 r = ttm_bo_move_accel_cleanup(bo, fence, true, false, new_mem); in amdgpu_move_blit() 416 r = ttm_bo_move_accel_cleanup(bo, fence, evict, true, new_mem); in amdgpu_move_blit() 486 struct ttm_resource *new_mem, in amdgpu_bo_move() argument 494 if (new_mem->mem_type == TTM_PL_TT || in amdgpu_bo_move() 495 new_mem->mem_type == AMDGPU_PL_PREEMPT) { in amdgpu_bo_move() 496 r = amdgpu_ttm_backend_bind(bo->bdev, bo->ttm, new_mem); in amdgpu_bo_move() 506 amdgpu_bo_move_notify(bo, evict, new_mem); in amdgpu_bo_move() [all …]
|
| H A D | amdgpu_vm.c | 2312 void amdgpu_vm_bo_move(struct amdgpu_bo *bo, struct ttm_resource *new_mem, in amdgpu_vm_bo_move() argument 2322 amdgpu_vm_update_stats_locked(bo_base, new_mem, +1); in amdgpu_vm_bo_move()
|
| /linux/drivers/gpu/drm/ttm/ |
| H A D | ttm_bo_util.c | 710 struct ttm_resource *new_mem) in ttm_bo_move_accel_cleanup() argument 714 struct ttm_resource_manager *man = ttm_manager_type(bdev, new_mem->mem_type); in ttm_bo_move_accel_cleanup() 728 ttm_bo_assign_mem(bo, new_mem); in ttm_bo_move_accel_cleanup() 744 struct ttm_resource *new_mem) in ttm_bo_move_sync_cleanup() argument 747 struct ttm_resource_manager *man = ttm_manager_type(bdev, new_mem->mem_type); in ttm_bo_move_sync_cleanup() 754 ttm_bo_assign_mem(bo, new_mem); in ttm_bo_move_sync_cleanup()
|
| /linux/drivers/pci/hotplug/ |
| H A D | ibmphp_res.c | 188 struct resource_node *new_mem = NULL; in ibmphp_rsrc_init() local 286 new_mem = alloc_resources(curr); in ibmphp_rsrc_init() 287 if (!new_mem) in ibmphp_rsrc_init() 289 new_mem->type = MEM; in ibmphp_rsrc_init() 297 if (ibmphp_add_resource(new_mem) < 0) { in ibmphp_rsrc_init() 301 newbus->firstMem = new_mem; in ibmphp_rsrc_init() 303 new_mem->rangeno = -1; in ibmphp_rsrc_init() 305 …y resource for device %x, bus %x, [%x - %x]\n", new_mem->devfunc, new_mem->busno, new_mem->start, … in ibmphp_rsrc_init()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_bo.c | 908 struct nouveau_mem *new_mem = nouveau_mem(reg); in nouveau_bo_move_prep() local 917 ret = nvif_vmm_get(vmm, LAZY, false, new_mem->mem.page, 0, in nouveau_bo_move_prep() 918 new_mem->mem.size, &old_mem->vma[1]); in nouveau_bo_move_prep() 926 ret = nouveau_mem_map(new_mem, vmm, &old_mem->vma[1]); in nouveau_bo_move_prep()
|
| /linux/drivers/vfio/pci/ |
| H A D | vfio_pci_config.c | 576 bool phys_mem, virt_mem, new_mem, phys_io, virt_io, new_io; in vfio_basic_config_write() local 591 new_mem = !!(new_cmd & PCI_COMMAND_MEMORY); in vfio_basic_config_write() 593 if (!new_mem) { in vfio_basic_config_write() 609 if ((new_mem && virt_mem && !phys_mem && in vfio_basic_config_write()
|
| /linux/tools/power/x86/turbostat/ |
| H A D | turbostat.c | 1974 struct pmt_domain_info *new_mem; in pmt_counter_resize_() local 1976 …new_mem = (struct pmt_domain_info *)reallocarray(pcounter->domains, new_size, sizeof(*pcounter->do… in pmt_counter_resize_() 1977 if (!new_mem) { in pmt_counter_resize_() 1985 memset(&new_mem[pcounter->num_domains], 0, num_new_domains * sizeof(*pcounter->domains)); in pmt_counter_resize_() 1988 pcounter->domains = new_mem; in pmt_counter_resize_()
|