Lines Matching +full:boot +full:- +full:pages

1 // SPDX-License-Identifier: GPL-2.0-only
65 * in non-PAE compilation mode, since the middle layer is folded.
187 BUG_ON(vaddr < fix_to_virt(FIX_KMAP_BEGIN - 1) in page_table_kmap_check()
245 * of max_low_pfn pages, by creating page tables starting from address
270 * First iteration will setup identity mapping using large/small pages in kernel_physical_mapping_init()
279 * "Software should not write to a paging-structure entry in a way in kernel_physical_mapping_init()
309 * Map with big pages if possible, otherwise in kernel_physical_mapping_init()
324 addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE + in kernel_physical_mapping_init()
325 PAGE_OFFSET + PAGE_SIZE-1; in kernel_physical_mapping_init()
428 * in the 32-bit EFI stub. in sync_initial_page_table()
446 * memory from the boot time page table. in native_pagetable_init()
447 * In virtual address space, we should have at least two pages in native_pagetable_init()
453 for (pfn = max_low_pfn; pfn < 1<<(32-PAGE_SHIFT); pfn++) { in native_pagetable_init()
486 * the boot process.
502 * created - mappings will be set by set_fixmap(): in early_ioremap_page_table_range_init()
504 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in early_ioremap_page_table_range_init()
505 end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK; in early_ioremap_page_table_range_init()
523 /* Used in PAGE_KERNEL_* macros which are reasonably used out-of-tree: */
526 /* user-defined highmem size */
527 static unsigned int highmem_pages = -1;
537 return -EINVAL; in parse_highmem()
545 "highmem size (%luMB) is bigger than pages available (%luMB)!\n"
550 * All of RAM fits into lowmem - but if user wants highmem
551 * artificially via the highmem=x boot parameter then create
559 if (highmem_pages == -1) in lowmem_pfn_init()
568 if (max_low_pfn - highmem_pages < 64*1024*1024/PAGE_SIZE) { in lowmem_pfn_init()
573 max_low_pfn -= highmem_pages; in lowmem_pfn_init()
577 printk(KERN_ERR "ignoring highmem size on non-highmem kernel!\n"); in lowmem_pfn_init()
582 "only %luMB highmem pages available, ignoring highmem size of %luMB!\n"
587 * We have more RAM than fits into lowmem - we try to put it into
588 * highmem, also taking the highmem=x boot parameter into account:
594 if (highmem_pages == -1) in highmem_pfn_init()
595 highmem_pages = max_pfn - MAXMEM_PFN; in highmem_pfn_init()
602 pages_to_mb(max_pfn - MAXMEM_PFN), in highmem_pfn_init()
645 pages_to_mb(highend_pfn - highstart_pfn)); in initmem_init()
646 high_memory = (void *) __va(highstart_pfn * PAGE_SIZE - 1) + 1; in initmem_init()
648 high_memory = (void *) __va(max_low_pfn * PAGE_SIZE - 1) + 1; in initmem_init()
667 printk(KERN_INFO " mapped low ram: 0 - %08lx\n", in setup_bootmem_allocator()
669 printk(KERN_INFO " low ram: 0 - %08lx\n", max_low_pfn<<PAGE_SHIFT); in setup_bootmem_allocator()
673 * paging_init() sets up the page tables - note that the first 8MB are
677 * that we can trap those pesky NULL-reference errors in the kernel.
725 * With CONFIG_DEBUG_PAGEALLOC initialization of highmem pages has to in mem_init()
728 * pages that was already passed to the buddy allocator, hence marked as in mem_init()
745 #define __FIXADDR_TOP (-PAGE_SIZE) in mem_init()
750 #define high_memory (-128UL << 20) in mem_init()
777 unsigned long size = (((unsigned long)__init_end + HPAGE_SIZE) & HPAGE_MASK) - start; in mark_nxdata_nx()
780 printk(KERN_INFO "NX-protecting the kernel data: %luk\n", size >> 10); in mark_nxdata_nx()
787 unsigned long size = (unsigned long)__end_rodata - start; in mark_rodata_ro()
790 pr_info("Write protecting kernel text and read-only data: %luk\n", in mark_rodata_ro()
796 pr_info("Testing CPA: Reverting %lx-%lx\n", start, start + size); in mark_rodata_ro()