Searched refs:vaddr_end (Results 1 – 8 of 8) sorted by relevance
| /linux/arch/x86/mm/ |
| H A D | mem_encrypt_amd.c | 259 unsigned long vaddr_end = vaddr + size; in enc_dec_hypercall() local 261 while (vaddr < vaddr_end) { in enc_dec_hypercall() 395 unsigned long vaddr_end, vaddr_next, start; in early_set_memory_enc_dec() local 403 vaddr_end = vaddr + size; in early_set_memory_enc_dec() 405 for (; vaddr < vaddr_end; vaddr = vaddr_next) { in early_set_memory_enc_dec() 428 ((vaddr_end - vaddr) >= psize)) { in early_set_memory_enc_dec() 450 __pa((vaddr_end & pmask) + psize), in early_set_memory_enc_dec() 545 unsigned long vaddr, vaddr_end, npages; in mem_encrypt_free_decrypted_mem() local 549 vaddr_end = (unsigned long)__end_bss_decrypted; in mem_encrypt_free_decrypted_mem() 550 npages = (vaddr_end - vaddr) >> PAGE_SHIFT; in mem_encrypt_free_decrypted_mem() [all …]
|
| H A D | kaslr.c | 41 static const unsigned long vaddr_end = CPU_ENTRY_AREA_BASE; variable 96 BUILD_BUG_ON(vaddr_start >= vaddr_end); in kernel_randomize_memory() 97 BUILD_BUG_ON(vaddr_end != CPU_ENTRY_AREA_BASE); in kernel_randomize_memory() 98 BUILD_BUG_ON(vaddr_end > __START_KERNEL_map); in kernel_randomize_memory() 135 remain_entropy = vaddr_end - vaddr_start; in kernel_randomize_memory()
|
| /linux/arch/x86/coco/sev/ |
| H A D | core.c | 211 unsigned long vaddr_end = vaddr + PMD_SIZE; in pval_pages() local 213 for (; vaddr < vaddr_end; vaddr += PAGE_SIZE, pfn++) { in pval_pages() 320 unsigned long vaddr_end, int op) in __set_pages_state() argument 337 while (vaddr < vaddr_end && i < ARRAY_SIZE(data->entries)) { in __set_pages_state() 352 (vaddr_end - vaddr) >= PMD_SIZE) { in __set_pages_state() 394 unsigned long vaddr_end; in set_pages_state() local 404 vaddr_end = vaddr + (npages << PAGE_SHIFT); in set_pages_state() 406 while (vaddr < vaddr_end) in set_pages_state() 407 vaddr = __set_pages_state(&desc, vaddr, vaddr_end, op); in set_pages_state()
|
| /linux/Documentation/arch/x86/x86_64/ |
| H A D | mm.rst | 70 | | | | vaddr_end for KASLR 140 | | | | vaddr_end for KASLR
|
| /linux/arch/x86/xen/ |
| H A D | mmu_pv.c | 1068 unsigned long vaddr_end) in xen_cleanhighmap() argument 1075 for (; vaddr <= vaddr_end && (pmd < (level2_kernel_pgt + PTRS_PER_PMD)); in xen_cleanhighmap() 1093 void *vaddr_end = vaddr + size; in xen_free_ro_pages() local 1095 for (; vaddr < vaddr_end; vaddr += PAGE_SIZE) in xen_free_ro_pages()
|
| /linux/mm/ |
| H A D | huge_memory.c | 4584 unsigned long vaddr_end, unsigned int new_order, in split_huge_pages_pid() argument 4594 vaddr_end &= PAGE_MASK; in split_huge_pages_pid() 4612 pid, vaddr_start, vaddr_end, new_order, in_folio_offset); in split_huge_pages_pid() 4619 for (addr = vaddr_start; addr < vaddr_end; addr += PAGE_SIZE) { in split_huge_pages_pid() 4781 unsigned long vaddr_start, vaddr_end; in split_huge_pages_write() local 4827 &vaddr_end, &new_order, &in_folio_offset); in split_huge_pages_write() 4837 ret = split_huge_pages_pid(pid, vaddr_start, vaddr_end, new_order, in split_huge_pages_write()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | sev.c | 810 unsigned long vaddr, vaddr_end, next_vaddr, npages, pages, size, i; in sev_launch_update_data() local 824 vaddr_end = vaddr + size; in sev_launch_update_data() 840 for (i = 0; vaddr < vaddr_end; vaddr = next_vaddr, i += pages) { in sev_launch_update_data() 1290 unsigned long vaddr, vaddr_end, next_vaddr; in sev_dbg_crypt() local 1311 vaddr_end = vaddr + size; in sev_dbg_crypt() 1314 for (; vaddr < vaddr_end; vaddr = next_vaddr) { in sev_dbg_crypt()
|
| /linux/drivers/vfio/ |
| H A D | vfio_iommu_type1.c | 1688 unsigned long vaddr_end; in vfio_dma_do_map() local 1702 check_add_overflow(vaddr, size - 1, &vaddr_end)) in vfio_dma_do_map()
|