| /linux/arch/alpha/include/asm/ |
| H A D | setup.h | 23 #define KERNEL_START (PAGE_OFFSET+KERNEL_START_PHYS) 25 #define INIT_STACK (PAGE_OFFSET+KERNEL_START_PHYS+0x02000) 26 #define EMPTY_PGT (PAGE_OFFSET+KERNEL_START_PHYS+0x04000) 27 #define EMPTY_PGE (PAGE_OFFSET+KERNEL_START_PHYS+0x08000) 28 #define ZERO_PGE (PAGE_OFFSET+KERNEL_START_PHYS+0x0A000) 30 #define START_ADDR (PAGE_OFFSET+KERNEL_START_PHYS+0x10000)
|
| /linux/arch/sh/include/asm/ |
| H A D | page.h | 110 #define PAGE_OFFSET CONFIG_PAGE_OFFSET macro 122 #define ___pa(x) ((x)-PAGE_OFFSET+__MEMORY_START) 123 #define ___va(x) ((x)+PAGE_OFFSET-__MEMORY_START) 125 #define ___pa(x) ((x)-PAGE_OFFSET) 126 #define ___va(x) ((x)+PAGE_OFFSET) 139 #define UNCAC_ADDR(addr) ((addr) - PAGE_OFFSET + uncached_start) 140 #define CAC_ADDR(addr) ((addr) - uncached_start + PAGE_OFFSET)
|
| /linux/arch/powerpc/mm/book3s32/ |
| H A D | hash_low.S | 53 lis r8, (mmu_hash_lock - PAGE_OFFSET)@h 54 ori r8, r8, (mmu_hash_lock - PAGE_OFFSET)@l 154 lis r8, (mmu_hash_lock - PAGE_OFFSET)@ha 156 stw r0, (mmu_hash_lock - PAGE_OFFSET)@l(r8) 163 lis r8, (mmu_hash_lock - PAGE_OFFSET)@ha 165 stw r0, (mmu_hash_lock - PAGE_OFFSET)@l(r8) 204 lis r6, (mmu_hash_lock - PAGE_OFFSET)@ha 205 addi r6, r6, (mmu_hash_lock - PAGE_OFFSET)@l 256 lis r6, (mmu_hash_lock - PAGE_OFFSET)@ha 257 addi r6, r6, (mmu_hash_lock - PAGE_OFFSET)@l [all …]
|
| /linux/arch/csky/include/asm/ |
| H A D | page.h | 23 #define PAGE_OFFSET CONFIG_PAGE_OFFSET macro 33 #define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && \ 68 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + va_pa_offset) 69 #define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET - va_pa_offset)) 78 #define MAP_NR(x) PFN_DOWN((unsigned long)(x) - PAGE_OFFSET - \
|
| /linux/arch/x86/kernel/ |
| H A D | head32.c | 45 *(pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = (pte_t) {.pte = 0}; in zap_early_initrd_mapping() 115 *(*pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = pl2; in init_map() 145 *ptr = (unsigned long)ptep + PAGE_OFFSET; in mk_early_pgtbl_32() 154 *ptr = (pte.pte & PTE_PFN_MASK) + PAGE_OFFSET; in mk_early_pgtbl_32() 159 *ptr = (unsigned long)pl2p + PAGE_OFFSET; in mk_early_pgtbl_32() 168 *ptr = (unsigned long)pl2p + PAGE_OFFSET; in mk_early_pgtbl_32()
|
| H A D | doublefault_32.c | 14 #define ptr_ok(x) ((x) > PAGE_OFFSET && (x) < PAGE_OFFSET + MAXMEM)
|
| /linux/arch/riscv/include/asm/ |
| H A D | page.h | 32 #define PAGE_OFFSET kernel_map.page_offset macro 34 #define PAGE_OFFSET _AC(0xc0000000, UL) macro 37 #define PAGE_OFFSET ((unsigned long)phys_ram_base) macro 115 ((x) >= PAGE_OFFSET && (!IS_ENABLED(CONFIG_64BIT) || (x) < PAGE_OFFSET + KERN_VIRT_SIZE)) 178 (unsigned long)(_addr) >= PAGE_OFFSET && pfn_valid(virt_to_pfn(_addr)); \
|
| /linux/tools/testing/selftests/powerpc/mm/ |
| H A D | bad_accesses.c | 24 #define PAGE_OFFSET (0xcul << 60) macro 60 (fault_addr < PAGE_OFFSET || fault_addr >= kernel_virt_end)); in bad_access() 92 kernel_virt_end = PAGE_OFFSET + (7 * (512ul << 40)); in test() 97 kernel_virt_end = PAGE_OFFSET + (7 * (64ul << 40)); in test()
|
| /linux/arch/m68k/include/asm/ |
| H A D | page_mm.h | 99 if (x >= PAGE_OFFSET) in ___pa() 100 return (x - PAGE_OFFSET); in ___pa() 111 return (void *)(x + PAGE_OFFSET); in __va() 144 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (…
|
| H A D | mmu_context.h | 114 if (mmuar >= PAGE_OFFSET) { in load_ksp_mmu() 140 pte = (mmuar >= PAGE_OFFSET) ? pte_offset_kernel(pmd, mmuar) in load_ksp_mmu() 147 if (!pte_dirty(*pte) && mmuar <= PAGE_OFFSET) in load_ksp_mmu() 164 if (pte && mmuar < PAGE_OFFSET) in load_ksp_mmu()
|
| /linux/arch/hexagon/include/asm/ |
| H A D | mem-layout.h | 19 #define PAGE_OFFSET _AC(0xc0000000, UL) macro 41 #define TASK_SIZE (PAGE_OFFSET) 54 #define MIN_KERNEL_SEG (PAGE_OFFSET >> PGDIR_SHIFT) /* L1 shift is 22 bits */
|
| /linux/arch/powerpc/kernel/ |
| H A D | head_8xx.S | 216 0: lwz r10, (itlb_miss_counter - PAGE_OFFSET)@l(0) 218 stw r10, (itlb_miss_counter - PAGE_OFFSET)@l(0) 264 0: lwz r10, (dtlb_miss_counter - PAGE_OFFSET)@l(0) 266 stw r10, (dtlb_miss_counter - PAGE_OFFSET)@l(0) 328 cmplwi cr1, r11, (.Ldtlbie - PAGE_OFFSET)@l 329 cmplwi cr7, r11, (.Litlbie - PAGE_OFFSET)@l 349 lwz r10, (instruction_counter - PAGE_OFFSET)@l(0) 351 stw r10, (instruction_counter - PAGE_OFFSET)@l(0) 376 oris r10, r10, (swapper_pg_dir - PAGE_OFFSET)@ha 377 lwz r10, (swapper_pg_dir - PAGE_OFFSET)@l(r10) /* Get the level 1 entry */ [all …]
|
| H A D | head_44x.S | 186 lis r5,PAGE_OFFSET@h 187 ori r5,r5,PAGE_OFFSET@l 316 lis r11, PAGE_OFFSET@h 430 lis r11, PAGE_OFFSET@h 570 lis r11,PAGE_OFFSET@h 671 lis r11,PAGE_OFFSET@h 898 lis r3,PAGE_OFFSET@h 899 ori r3,r3,PAGE_OFFSET@l 1110 lis r3,PAGE_OFFSET@h 1111 ori r3,r3,PAGE_OFFSET@l [all …]
|
| /linux/drivers/pci/controller/ |
| H A D | pci-versatile.c | 94 writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem)); in versatile_pci_probe() 131 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_0); in versatile_pci_probe() 132 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_1); in versatile_pci_probe() 133 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_2); in versatile_pci_probe()
|
| /linux/arch/arm/mach-zynq/ |
| H A D | platsmp.c | 46 if (__pa(PAGE_OFFSET)) { in zynq_cpun_start() 53 zero = (__force u8 __iomem *)PAGE_OFFSET; in zynq_cpun_start() 70 if (__pa(PAGE_OFFSET)) in zynq_cpun_start()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | kdump.h | 26 #define KDUMP_TRAMPOLINE_START (0x0100 + PAGE_OFFSET) 27 #define KDUMP_TRAMPOLINE_END (0x3000 + PAGE_OFFSET)
|
| H A D | kasan.h | 23 #define KASAN_KERN_START ALIGN_DOWN(PAGE_OFFSET - SZ_256M, SZ_256M) 25 #define KASAN_KERN_START PAGE_OFFSET
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 41 self.PAGE_OFFSET = int(gdb.parse_and_eval("page_offset_base")) 156 return pa + self.PAGE_OFFSET 162 return va - self.PAGE_OFFSET 264 self.PAGE_OFFSET = self._PAGE_OFFSET(self.VA_BITS) 268 self.VMEMMAP_RANGE = self._PAGE_END(self.VA_BITS_MIN) - self.PAGE_OFFSET 366 if (addr - self.PAGE_OFFSET) < (self.PAGE_END - self.PAGE_OFFSET): 371 return addr - self.PAGE_OFFSET + self.PHYS_OFFSET 399 return (pa - self.PHYS_OFFSET) | self.PAGE_OFFSET 445 return self.PAGE_OFFSET + (__idx * self.PAGE_SIZE) 451 __idx = int(self.kasan_reset_tag(va) - self.PAGE_OFFSET) // self.PAGE_SIZE
|
| /linux/arch/arm/mm/ |
| H A D | physaddr.c | 21 if (!high_memory && x >= PAGE_OFFSET) in __virt_addr_valid() 24 if (high_memory && x >= PAGE_OFFSET && x < (unsigned long)high_memory) in __virt_addr_valid()
|
| /linux/arch/openrisc/kernel/ |
| H A D | vmlinux.lds.S | 22 # define LOAD_OFFSET PAGE_OFFSET 23 # define LOAD_BASE PAGE_OFFSET
|
| /linux/arch/arm/include/asm/ |
| H A D | dma.h | 13 arm_dma_zone_size && arm_dma_zone_size < (0x100000000ULL - PAGE_OFFSET) ? \ 14 (PAGE_OFFSET + arm_dma_zone_size) : 0xffffffffUL; })
|
| /linux/arch/mips/include/asm/mach-generic/ |
| H A D | spaces.h | 94 #ifndef PAGE_OFFSET 95 #define PAGE_OFFSET (CAC_BASE + PHYS_OFFSET) macro
|
| /linux/Documentation/arch/arm/ |
| H A D | porting.rst | 14 phys = virt - PAGE_OFFSET + PHYS_OFFSET 64 PAGE_OFFSET 66 boot phase, virtual address PAGE_OFFSET will be mapped to physical 84 hence PAGE_OFFSET). 87 Virtual start address of kernel, normally PAGE_OFFSET + 0x8000.
|
| /linux/arch/loongarch/include/asm/ |
| H A D | addrspace.h | 70 #ifndef PAGE_OFFSET 71 #define PAGE_OFFSET (CACHE_BASE + PHYS_OFFSET) macro
|
| /linux/arch/powerpc/mm/nohash/ |
| H A D | e500.c | 226 unsigned long virt = PAGE_OFFSET; in map_mem_in_cams() 240 return tlbcam_addrs[tlbcam_index - 1].limit - PAGE_OFFSET + 1; in mmu_mapin_ram() 325 virt_phys_offset = PAGE_OFFSET - memstart_addr; in relocate_init() 368 map_mem_in_cams_addr(start, PAGE_OFFSET + offset, in relocate_init()
|