Searched refs:dummy_page (Results 1 – 15 of 15) sorted by relevance
54 void *dummy_page; member447 mmu->dummy_page = (void *)__get_free_page(GFP_KERNEL); in imgu_mmu_init()448 if (!mmu->dummy_page) in imgu_mmu_init()450 pteval = IPU3_ADDR2PTE(virt_to_phys(mmu->dummy_page)); in imgu_mmu_init()491 free_page((unsigned long)mmu->dummy_page); in imgu_mmu_init()514 free_page((unsigned long)mmu->dummy_page); in imgu_mmu_exit()
174 pd->dummy_page = alloc_page(GFP_DMA32); in psb_mmu_alloc_pd()175 if (!pd->dummy_page) in psb_mmu_alloc_pd()181 pd->invalid_pte = psb_mmu_mask_pte(page_to_pfn(pd->dummy_page), in psb_mmu_alloc_pd()200 clear_page(kmap(pd->dummy_page)); in psb_mmu_alloc_pd()201 kunmap(pd->dummy_page); in psb_mmu_alloc_pd()214 __free_page(pd->dummy_page); in psb_mmu_alloc_pd()254 __free_page(pd->dummy_page); in psb_mmu_free_pagedir()
593 static void *dummy_page[2]; variable600 if (dummy_page[i]) { in free_fake_buffer()601 free_page((unsigned long)dummy_page[i]); in free_fake_buffer()602 dummy_page[i] = NULL; in free_fake_buffer()614 dummy_page[i] = (void *)get_zeroed_page(GFP_KERNEL); in alloc_fake_buffer()615 if (!dummy_page[i]) { in alloc_fake_buffer()640 return virt_to_page(dummy_page[substream->stream]); /* the same page */ in dummy_pcm_page()
127 iommu->dummy_page = (unsigned long) page_address(page); in iommu_table_init()128 memset((void *)iommu->dummy_page, 0, PAGE_SIZE); in iommu_table_init()129 iommu->dummy_page_pa = (unsigned long) __pa(iommu->dummy_page); in iommu_table_init()146 free_page(iommu->dummy_page); in iommu_table_init()147 iommu->dummy_page = 0UL; in iommu_table_init()
68 unsigned long dummy_page; member
56 static struct page *dummy_page; variable330 dummy_tag_frag.page = dummy_page; in tls_device_record_close()1414 dummy_page = alloc_page(GFP_KERNEL); in tls_device_init()1415 if (!dummy_page) in tls_device_init()1433 put_page(dummy_page); in tls_device_init()1442 put_page(dummy_page); in tls_device_cleanup()
158 struct page *dummy_page; member
758 if (omap_dmm->dummy_page) in omap_dmm_remove()759 __free_page(omap_dmm->dummy_page); in omap_dmm_remove()863 omap_dmm->dummy_page = alloc_page(GFP_KERNEL | __GFP_DMA32); in omap_dmm_probe()864 if (!omap_dmm->dummy_page) { in omap_dmm_probe()875 omap_dmm->dummy_pa = page_to_phys(omap_dmm->dummy_page); in omap_dmm_probe()
77 struct page *dummy_page = ttm_glob.dummy_read_page; in amdgpu_gart_dummy_page_init() local81 adev->dummy_page_addr = dma_map_page_attrs(&adev->pdev->dev, dummy_page, 0, in amdgpu_gart_dummy_page_init()
155 mmu_info->dummy_page = pt; in get_dummy_page()170 free_page((unsigned long)mmu_info->dummy_page); in free_dummy_page()
138 mmu_info->dummy_page = pt; in get_dummy_page()153 free_page((unsigned long)mmu_info->dummy_page); in free_dummy_page()
269 struct page *dummy_page = (struct page *)res; in barrier_release_dummy_page() local271 __free_page(dummy_page); in barrier_release_dummy_page()
1283 (u32)(rdev->dummy_page.addr >> 12)); in cayman_pcie_gart_enable()1307 (u32)(rdev->dummy_page.addr >> 12)); in cayman_pcie_gart_enable()
932 (u32)(rdev->dummy_page.addr >> 12)); in rv770_pcie_gart_enable()
1174 (u32)(rdev->dummy_page.addr >> 12)); in r600_pcie_gart_enable()3697 WREG32(INTERRUPT_CNTL2, rdev->dummy_page.addr >> 8); in r600_irq_init()