Home
last modified time | relevance | path

Searched refs:base_pfn (Results 1 – 17 of 17) sorted by relevance

/linux/mm/
H A Dcma.c42 return PFN_PHYS(cma->ranges[0].base_pfn); in cma_get_base()
71 return (cmr->base_pfn & ((1UL << align_order) - 1)) in cma_bitmap_aligned_offset()
87 bitmap_no = (pfn - cmr->base_pfn) >> cma->order_per_bit; in cma_clear_bitmap()
104 unsigned long base_pfn; in cma_validate_zones() local
120 base_pfn = cmr->base_pfn; in cma_validate_zones()
127 WARN_ON_ONCE(!pfn_valid(base_pfn)); in cma_validate_zones()
128 if (pfn_range_intersects_zones(cma->nid, base_pfn, cmr->count)) { in cma_validate_zones()
160 if (early_pfn[r] != cmr->base_pfn) { in cma_activate_area()
161 count = early_pfn[r] - cmr->base_pfn; in cma_activate_area()
166 for (pfn = early_pfn[r]; pfn < cmr->base_pfn + cmr->count; in cma_activate_area()
[all …]
H A Dcma.h27 unsigned long base_pfn; member
H A Dnuma_emulation.c207 unsigned long base_pfn = PHYS_PFN(base); in uniform_size() local
210 return PFN_PHYS((max_pfn - base_pfn - hole_pfns) / nr_nodes); in uniform_size()
H A Dsparse-vmemmap.c103 return altmap->base_pfn + altmap->reserve + altmap->alloc in vmem_altmap_next_pfn()
H A Dmm_init.c899 if (start_pfn == altmap->base_pfn) in memmap_init_range()
901 end_pfn = altmap->base_pfn + vmem_altmap_offset(altmap); in memmap_init_range()
1133 start_pfn = altmap->base_pfn + vmem_altmap_offset(altmap); in memmap_init_zone_device()
H A Dmemory_hotplug.c406 if (altmap->base_pfn != pfn in __add_pages()
1449 .base_pfn = PHYS_PFN(cur_start), in create_altmaps_and_memory_blocks()
/linux/drivers/hv/
H A Dhv_proc.c28 u64 base_pfn; in hv_call_deposit_pages() local
83 base_pfn = page_to_pfn(pages[i]); in hv_call_deposit_pages()
85 input_page->gpa_page_list[page_count] = base_pfn + j; in hv_call_deposit_pages()
101 base_pfn = page_to_pfn(pages[i]); in hv_call_deposit_pages()
103 __free_page(pfn_to_page(base_pfn + j)); in hv_call_deposit_pages()
/linux/arch/powerpc/kvm/
H A Dbook3s_hv_uvmem.c229 unsigned long base_pfn; member
261 p->base_pfn = slot->base_gfn; in kvmppc_uvmem_slot_init()
279 if (p->base_pfn == slot->base_gfn) { in kvmppc_uvmem_slot_free()
295 if (gfn >= p->base_pfn && gfn < p->base_pfn + p->nr_pfns) { in kvmppc_mark_gfn()
296 unsigned long index = gfn - p->base_pfn; in kvmppc_mark_gfn()
339 if (gfn >= p->base_pfn && gfn < p->base_pfn + p->nr_pfns) { in kvmppc_gfn_is_uvmem_pfn()
340 unsigned long index = gfn - p->base_pfn; in kvmppc_gfn_is_uvmem_pfn()
369 if (*gfn >= iter->base_pfn && *gfn < iter->base_pfn + iter->nr_pfns) { in kvmppc_next_nontransitioned_gfn()
379 for (i = *gfn; i < p->base_pfn + p->nr_pfns; i++) { in kvmppc_next_nontransitioned_gfn()
380 unsigned long index = i - p->base_pfn; in kvmppc_next_nontransitioned_gfn()
/linux/arch/x86/kernel/cpu/mtrr/
H A Dcleanup.c24 unsigned long base_pfn; member
61 base = range_state[i].base_pfn; in x86_get_mtrr_mem_range()
81 base = range_state[i].base_pfn; in x86_get_mtrr_mem_range()
181 range_state[reg].base_pfn = basek >> (PAGE_SHIFT - 10); in save_var_mtrr()
193 basek = range_state[reg].base_pfn << (PAGE_SHIFT - 10); in set_var_mtrr_all()
383 set_var_mtrr_range(struct var_mtrr_state *state, unsigned long base_pfn, in set_var_mtrr_range() argument
392 basek = base_pfn << (PAGE_SHIFT - 10); in set_var_mtrr_range()
515 start_base = range_state[i].base_pfn << (PAGE_SHIFT - 10); in print_out_mtrr_range_state()
681 range_state[i].base_pfn = base; in mtrr_cleanup()
881 range_state[i].base_pfn = base; in mtrr_trim_uncached_memory()
[all …]
/linux/drivers/staging/media/ipu3/
H A Dipu3-dmamap.c233 unsigned long order, base_pfn; in imgu_dmamap_init() local
240 base_pfn = max_t(unsigned long, 1, imgu->mmu->aperture_start >> order); in imgu_dmamap_init()
241 init_iova_domain(&imgu->iova_domain, 1UL << order, base_pfn); in imgu_dmamap_init()
/linux/drivers/nvdimm/
H A Dpfn_devs.c658 unsigned long base_pfn = PHYS_PFN(base); in init_altmap_base() local
660 return SUBSECTION_ALIGN_DOWN(base_pfn); in init_altmap_base()
666 unsigned long base_pfn = PHYS_PFN(base); in init_altmap_reserve() local
668 reserve += base_pfn - SUBSECTION_ALIGN_DOWN(base_pfn); in init_altmap_reserve()
686 .base_pfn = init_altmap_base(base), in __nvdimm_setup_pfn()
/linux/arch/x86/hyperv/
H A Dhv_crash.c620 if (cda_info.base_pfn == 0) { in hv_root_crash_init()
625 hv_cda = phys_to_virt(cda_info.base_pfn << HV_HYP_PAGE_SHIFT); in hv_root_crash_init()
/linux/arch/powerpc/mm/book3s64/
H A Dradix_pgtable.c762 unsigned long base_pfn = page_to_pfn(page); in free_vmemmap_pages() local
769 alt_start = altmap->base_pfn; in free_vmemmap_pages()
770 alt_end = altmap->base_pfn + altmap->reserve + altmap->free; in free_vmemmap_pages()
772 if (base_pfn >= alt_start && base_pfn < alt_end) { in free_vmemmap_pages()
/linux/include/linux/
H A Dmemremap.h22 unsigned long base_pfn; member
/linux/drivers/iommu/
H A Ddma-iommu.c664 unsigned long order, base_pfn; in iommu_dma_init_domain() local
675 base_pfn = 1; in iommu_dma_init_domain()
686 base_pfn = max_t(unsigned long, base_pfn, in iommu_dma_init_domain()
692 base_pfn != iovad->start_pfn) { in iommu_dma_init_domain()
700 init_iova_domain(iovad, 1UL << order, base_pfn); in iommu_dma_init_domain()
/linux/kernel/liveupdate/
H A Dkexec_handover.c1373 unsigned long base_pfn = PHYS_PFN(kho_scratch[i].addr); in kho_init() local
1386 for (pfn = base_pfn; pfn < base_pfn + count; in kho_init()
/linux/include/hyperv/
H A Dhvhdk_mini.h154 u64 base_pfn : 64 - HV_PFN_RANGE_PGBITS; member