Searched refs:pdei (Results 1 – 4 of 4) sorted by relevance
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
| H A D | vmmnv50.c | 145 nv50_vmm_pgd_pde(struct nvkm_vmm *vmm, struct nvkm_vmm_pt *pgd, u32 pdei) in nv50_vmm_pgd_pde() argument 148 u32 pdeo = vmm->mmu->func->vmm.pd_offset + (pdei * 8); in nv50_vmm_pgd_pde() 151 if (!nv50_vmm_pde(vmm, pgd->pde[pdei], &data)) in nv50_vmm_pgd_pde() 347 u32 pdei; in nv50_vmm_join() local 355 for (pdei = vmm->start >> 29; pdei <= (vmm->limit - 1) >> 29; pdei++) { in nv50_vmm_join() 356 if (!nv50_vmm_pde(vmm, vmm->pd->pde[pdei], &data)) { in nv50_vmm_join() 360 nvkm_wo64(join->inst, pd_offset + (pdei * 8), data); in nv50_vmm_join()
|
| H A D | vmm.c | 146 u32 pdei = it->pte[it->lvl + 1]; in nvkm_vmm_unref_pdes() local 159 func->sparse(vmm, pgd->pt[0], pdei, 1); in nvkm_vmm_unref_pdes() 160 pgd->pde[pdei] = NVKM_VMM_PDE_SPARSE; in nvkm_vmm_unref_pdes() 162 func->unmap(vmm, pgd->pt[0], pdei, 1); in nvkm_vmm_unref_pdes() 163 pgd->pde[pdei] = NULL; in nvkm_vmm_unref_pdes() 170 func->pde(vmm, pgd, pdei); in nvkm_vmm_unref_pdes() 171 pgd->pde[pdei] = NULL; in nvkm_vmm_unref_pdes() 177 func->pde(vmm, pgd, pdei); in nvkm_vmm_unref_pdes() 437 nvkm_vmm_ref_hwpt(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgd, u32 pdei) in nvkm_vmm_ref_hwpt() argument 441 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in nvkm_vmm_ref_hwpt() [all …]
|
| H A D | vmmgm200.c | 54 struct nvkm_mmu_pt *pt, u32 pdei, u32 pdes) in gm200_vmm_pgd_sparse() argument 57 VMM_FO064(pt, vmm, pdei * 8, BIT_ULL(35) /* VOL_BIG. */, pdes); in gm200_vmm_pgd_sparse()
|
| H A D | vmmgf100.c | 106 gf100_vmm_pgd_pde(struct nvkm_vmm *vmm, struct nvkm_vmm_pt *pgd, u32 pdei) in gf100_vmm_pgd_pde() argument 108 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gf100_vmm_pgd_pde() 142 VMM_WO064(pd, vmm, pdei * 8, data); in gf100_vmm_pgd_pde()
|