Home
last modified time | relevance | path

Searched refs:PAGE_OFFSET (Results 1 – 25 of 184) sorted by relevance

12345678

/linux/arch/alpha/include/asm/
H A Dsetup.h23 #define KERNEL_START (PAGE_OFFSET+KERNEL_START_PHYS)
25 #define INIT_STACK (PAGE_OFFSET+KERNEL_START_PHYS+0x02000)
26 #define EMPTY_PGT (PAGE_OFFSET+KERNEL_START_PHYS+0x04000)
27 #define EMPTY_PGE (PAGE_OFFSET+KERNEL_START_PHYS+0x08000)
28 #define ZERO_PGE (PAGE_OFFSET+KERNEL_START_PHYS+0x0A000)
30 #define START_ADDR (PAGE_OFFSET+KERNEL_START_PHYS+0x10000)
H A Dpage.h64 #define PAGE_OFFSET 0xffff800000000000UL macro
66 #define PAGE_OFFSET 0xfffffc0000000000UL macro
72 #define PAGE_OFFSET 0xffff800000000000 macro
74 #define PAGE_OFFSET 0xfffffc0000000000 macro
79 #define __pa(x) ((unsigned long) (x) - PAGE_OFFSET)
80 #define __va(x) ((void *)((unsigned long) (x) + PAGE_OFFSET))
/linux/arch/powerpc/include/asm/
H A Dpage.h74 #define PAGE_OFFSET ASM_CONST(CONFIG_PAGE_OFFSET) macro
108 #define MEMORY_START (PHYSICAL_START + PAGE_OFFSET - KERNELBASE)
203 VIRTUAL_WARN_ON((unsigned long)(x) >= PAGE_OFFSET); \
204 (void *)(unsigned long)((phys_addr_t)(x) | PAGE_OFFSET); \
209 VIRTUAL_WARN_ON((unsigned long)(x) < PAGE_OFFSET); \
214 #define __va(x) ((void *)(unsigned long)((phys_addr_t)(x) + PAGE_OFFSET - MEMORY_START))
215 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + MEMORY_START)
234 _addr >= PAGE_OFFSET && _addr < (unsigned long)high_memory && \
259 #define is_kernel_addr(x) ((x) >= PAGE_OFFSET)
H A Dkdump.h26 #define KDUMP_TRAMPOLINE_START (0x0100 + PAGE_OFFSET)
27 #define KDUMP_TRAMPOLINE_END (0x3000 + PAGE_OFFSET)
H A Dkasan.h23 #define KASAN_KERN_START ALIGN_DOWN(PAGE_OFFSET - SZ_256M, SZ_256M)
25 #define KASAN_KERN_START PAGE_OFFSET
/linux/arch/openrisc/include/asm/
H A Dpage.h20 #define PAGE_OFFSET 0xc0000000 macro
21 #define KERNELBASE PAGE_OFFSET
63 #define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET))
64 #define __pa(x) ((unsigned long) (x) - PAGE_OFFSET)
72 (mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT))
/linux/arch/sh/include/asm/
H A Dpage.h110 #define PAGE_OFFSET CONFIG_PAGE_OFFSET macro
122 #define ___pa(x) ((x)-PAGE_OFFSET+__MEMORY_START)
123 #define ___va(x) ((x)+PAGE_OFFSET-__MEMORY_START)
125 #define ___pa(x) ((x)-PAGE_OFFSET)
126 #define ___va(x) ((x)+PAGE_OFFSET)
139 #define UNCAC_ADDR(addr) ((addr) - PAGE_OFFSET + uncached_start)
140 #define CAC_ADDR(addr) ((addr) - uncached_start + PAGE_OFFSET)
/linux/arch/csky/include/asm/
H A Dpage.h23 #define PAGE_OFFSET CONFIG_PAGE_OFFSET macro
33 #define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && \
68 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + va_pa_offset)
69 #define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET - va_pa_offset))
78 #define MAP_NR(x) PFN_DOWN((unsigned long)(x) - PAGE_OFFSET - \
/linux/arch/riscv/include/asm/
H A Dpage.h33 #define PAGE_OFFSET PAGE_OFFSET_L3 macro
35 #define PAGE_OFFSET kernel_map.page_offset macro
38 #define PAGE_OFFSET _AC(0xc0000000, UL) macro
41 #define PAGE_OFFSET ((unsigned long)phys_ram_base) macro
128 ((x) >= PAGE_OFFSET && (!IS_ENABLED(CONFIG_64BIT) || (x) < PAGE_OFFSET + KERN_VIRT_SIZE))
206 (unsigned long)(_addr) >= PAGE_OFFSET && pfn_valid(virt_to_pfn(_addr)); \
/linux/arch/powerpc/mm/book3s32/
H A Dhash_low.S53 lis r8, (mmu_hash_lock - PAGE_OFFSET)@h
54 ori r8, r8, (mmu_hash_lock - PAGE_OFFSET)@l
154 lis r8, (mmu_hash_lock - PAGE_OFFSET)@ha
156 stw r0, (mmu_hash_lock - PAGE_OFFSET)@l(r8)
163 lis r8, (mmu_hash_lock - PAGE_OFFSET)@ha
165 stw r0, (mmu_hash_lock - PAGE_OFFSET)@l(r8)
204 lis r6, (mmu_hash_lock - PAGE_OFFSET)@ha
205 addi r6, r6, (mmu_hash_lock - PAGE_OFFSET)@l
256 lis r6, (mmu_hash_lock - PAGE_OFFSET)@ha
257 addi r6, r6, (mmu_hash_lock - PAGE_OFFSET)@l
[all …]
H A Dmmu.c151 setbat(idx, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X); in __mmu_mapin_ram()
161 unsigned long border = (unsigned long)__srwx_boundary - PAGE_OFFSET; in mmu_mapin_ram()
164 size = roundup_pow_of_two((unsigned long)_einittext - PAGE_OFFSET); in mmu_mapin_ram()
165 setibat(0, PAGE_OFFSET, 0, size, PAGE_KERNEL_X); in mmu_mapin_ram()
200 unsigned long base = (unsigned long)_stext - PAGE_OFFSET; in mmu_mark_initmem_nx()
201 unsigned long top = ALIGN((unsigned long)_etext - PAGE_OFFSET, SZ_128K); in mmu_mark_initmem_nx()
202 unsigned long border = (unsigned long)__init_begin - PAGE_OFFSET; in mmu_mark_initmem_nx()
207 setibat(i++, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X); in mmu_mark_initmem_nx()
218 setibat(i++, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X); in mmu_mark_initmem_nx()
396 unsigned int hash = (unsigned int)Hash - PAGE_OFFSET; in MMU_init_hw_patch()
/linux/arch/x86/kernel/
H A Dhead32.c45 *(pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = (pte_t) {.pte = 0}; in zap_early_initrd_mapping()
115 *(*pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = pl2; in init_map()
145 *ptr = (unsigned long)ptep + PAGE_OFFSET; in mk_early_pgtbl_32()
154 *ptr = (pte.pte & PTE_PFN_MASK) + PAGE_OFFSET; in mk_early_pgtbl_32()
159 *ptr = (unsigned long)pl2p + PAGE_OFFSET; in mk_early_pgtbl_32()
168 *ptr = (unsigned long)pl2p + PAGE_OFFSET; in mk_early_pgtbl_32()
/linux/arch/nios2/include/asm/
H A Dpage.h26 #define PAGE_OFFSET \ macro
75 ((unsigned long)(x) - PAGE_OFFSET + PHYS_OFFSET)
77 ((void *)((unsigned long)(x) + PAGE_OFFSET - PHYS_OFFSET))
80 ((void *)(((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET)
/linux/arch/xtensa/include/asm/
H A Dpage.h24 #define PAGE_OFFSET XCHAL_KSEG_CACHED_VADDR macro
29 #define PAGE_OFFSET _AC(CONFIG_DEFAULT_MEM_START, UL) macro
145 unsigned long off = va - PAGE_OFFSET; in ___pa()
166 ((unsigned long) (x) - PAGE_OFFSET + PHYS_OFFSET)
169 ((void *)((unsigned long) (x) - PHYS_OFFSET + PAGE_OFFSET))
/linux/tools/testing/selftests/powerpc/mm/
H A Dbad_accesses.c24 #define PAGE_OFFSET (0xcul << 60) macro
60 (fault_addr < PAGE_OFFSET || fault_addr >= kernel_virt_end)); in bad_access()
92 kernel_virt_end = PAGE_OFFSET + (7 * (512ul << 40)); in test()
97 kernel_virt_end = PAGE_OFFSET + (7 * (64ul << 40)); in test()
/linux/arch/m68k/include/asm/
H A Dpage_mm.h99 if (x >= PAGE_OFFSET) in ___pa()
100 return (x - PAGE_OFFSET); in ___pa()
111 return (void *)(x + PAGE_OFFSET); in __va()
144 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (…
H A Dpage_no.h32 #define virt_to_page(addr) (mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT))
33 #define page_to_virt(page) __va(((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET))
35 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET) && \
/linux/arch/hexagon/include/asm/
H A Dmem-layout.h19 #define PAGE_OFFSET _AC(0xc0000000, UL) macro
41 #define TASK_SIZE (PAGE_OFFSET)
54 #define MIN_KERNEL_SEG (PAGE_OFFSET >> PGDIR_SHIFT) /* L1 shift is 22 bits */
/linux/drivers/pci/controller/
H A Dpci-versatile.c94 writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem)); in versatile_pci_probe()
131 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_0); in versatile_pci_probe()
132 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_1); in versatile_pci_probe()
133 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_2); in versatile_pci_probe()
/linux/arch/powerpc/kernel/
H A Dhead_8xx.S216 0: lwz r10, (itlb_miss_counter - PAGE_OFFSET)@l(0)
218 stw r10, (itlb_miss_counter - PAGE_OFFSET)@l(0)
264 0: lwz r10, (dtlb_miss_counter - PAGE_OFFSET)@l(0)
266 stw r10, (dtlb_miss_counter - PAGE_OFFSET)@l(0)
328 cmplwi cr1, r11, (.Ldtlbie - PAGE_OFFSET)@l
329 cmplwi cr7, r11, (.Litlbie - PAGE_OFFSET)@l
349 lwz r10, (instruction_counter - PAGE_OFFSET)@l(0)
351 stw r10, (instruction_counter - PAGE_OFFSET)@l(0)
376 oris r10, r10, (swapper_pg_dir - PAGE_OFFSET)@ha
377 lwz r10, (swapper_pg_dir - PAGE_OFFSET)@l(r10) /* Get the level 1 entry */
[all …]
/linux/arch/arm/mach-zynq/
H A Dplatsmp.c46 if (__pa(PAGE_OFFSET)) { in zynq_cpun_start()
53 zero = (__force u8 __iomem *)PAGE_OFFSET; in zynq_cpun_start()
70 if (__pa(PAGE_OFFSET)) in zynq_cpun_start()
/linux/arch/hexagon/kernel/
H A Dhead.S39 r1.h = #HI(PAGE_OFFSET);
40 r1.l = #LO(PAGE_OFFSET);
95 R1.H = #HI(PAGE_OFFSET >> (22 - 2))
96 R1.L = #LO(PAGE_OFFSET >> (22 - 2))
/linux/arch/arm/mm/
H A Dphysaddr.c21 if (!high_memory && x >= PAGE_OFFSET) in __virt_addr_valid()
24 if (high_memory && x >= PAGE_OFFSET && x < (unsigned long)high_memory) in __virt_addr_valid()
/linux/arch/openrisc/kernel/
H A Dvmlinux.lds.S22 # define LOAD_OFFSET PAGE_OFFSET
23 # define LOAD_BASE PAGE_OFFSET
/linux/arch/parisc/include/asm/
H A Dpage.h131 #define PAGE_OFFSET ((unsigned long)__PAGE_OFFSET) macro
146 #define __pa(x) ((unsigned long)(x)-PAGE_OFFSET)
147 #define __va(x) ((void *)((unsigned long)(x)+PAGE_OFFSET))

12345678