1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* 3 * Copyright (C) 2000 - 2003 Jeff Dike (jdike@addtoit.com) 4 * Copyright 2003 PathScale, Inc. 5 */ 6 7 #ifndef __UM_PAGE_H 8 #define __UM_PAGE_H 9 10 #include <linux/const.h> 11 12 /* PAGE_SHIFT determines the page size */ 13 #define PAGE_SHIFT CONFIG_PAGE_SHIFT 14 #define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT) 15 #define PAGE_MASK (~(PAGE_SIZE-1)) 16 17 #ifndef __ASSEMBLY__ 18 19 struct page; 20 21 #include <linux/pfn.h> 22 #include <linux/types.h> 23 #include <asm/vm-flags.h> 24 25 /* 26 * These are used to make use of C type-checking.. 27 */ 28 29 #define clear_page(page) memset((void *)(page), 0, PAGE_SIZE) 30 #define copy_page(to,from) memcpy((void *)(to), (void *)(from), PAGE_SIZE) 31 32 #define clear_user_page(page, vaddr, pg) clear_page(page) 33 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) 34 35 typedef struct { unsigned long pte; } pte_t; 36 typedef struct { unsigned long pgd; } pgd_t; 37 38 #ifdef CONFIG_3_LEVEL_PGTABLES 39 typedef struct { unsigned long pmd; } pmd_t; 40 #define pmd_val(x) ((x).pmd) 41 #define __pmd(x) ((pmd_t) { (x) } ) 42 #endif 43 44 #define pte_val(x) ((x).pte) 45 46 47 #define pte_get_bits(p, bits) ((p).pte & (bits)) 48 #define pte_set_bits(p, bits) ((p).pte |= (bits)) 49 #define pte_clear_bits(p, bits) ((p).pte &= ~(bits)) 50 #define pte_copy(to, from) ((to).pte = (from).pte) 51 #define pte_is_zero(p) (!((p).pte & ~_PAGE_NEWPAGE)) 52 #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot)) 53 54 typedef unsigned long phys_t; 55 56 typedef struct { unsigned long pgprot; } pgprot_t; 57 58 typedef struct page *pgtable_t; 59 60 #define pgd_val(x) ((x).pgd) 61 #define pgprot_val(x) ((x).pgprot) 62 63 #define __pte(x) ((pte_t) { (x) } ) 64 #define __pgd(x) ((pgd_t) { (x) } ) 65 #define __pgprot(x) ((pgprot_t) { (x) } ) 66 67 extern unsigned long uml_physmem; 68 69 #define PAGE_OFFSET (uml_physmem) 70 #define KERNELBASE PAGE_OFFSET 71 72 #define __va_space (8*1024*1024) 73 74 #include <mem.h> 75 76 /* Cast to unsigned long before casting to void * to avoid a warning from 77 * mmap_kmem about cutting a long long down to a void *. Not sure that 78 * casting is the right thing, but 32-bit UML can't have 64-bit virtual 79 * addresses 80 */ 81 #define __pa(virt) uml_to_phys((void *) (unsigned long) (virt)) 82 #define __va(phys) uml_to_virt((unsigned long) (phys)) 83 84 #define phys_to_pfn(p) ((p) >> PAGE_SHIFT) 85 #define pfn_to_phys(pfn) PFN_PHYS(pfn) 86 87 #define virt_addr_valid(v) pfn_valid(phys_to_pfn(__pa(v))) 88 89 #include <asm-generic/memory_model.h> 90 #include <asm-generic/getorder.h> 91 92 #endif /* __ASSEMBLY__ */ 93 94 #ifdef CONFIG_X86_32 95 #define __HAVE_ARCH_GATE_AREA 1 96 #endif 97 98 #endif /* __UM_PAGE_H */ 99