| /linux/fs/minix/ |
| H A D | dir.c | 102 char *p, *kaddr, *limit; in minix_readdir() local 105 kaddr = dir_get_folio(inode, n, &folio); in minix_readdir() 106 if (IS_ERR(kaddr)) in minix_readdir() 108 p = kaddr+offset; in minix_readdir() 109 limit = kaddr + minix_last_byte(inode, n) - chunk_size; in minix_readdir() 132 folio_release_kmap(folio, kaddr); in minix_readdir() 169 char *kaddr, *limit; in minix_find_entry() local 171 kaddr = dir_get_folio(dir, n, foliop); in minix_find_entry() 172 if (IS_ERR(kaddr)) in minix_find_entry() 175 limit = kaddr + minix_last_byte(dir, n) - sbi->s_dirsize; in minix_find_entry() [all …]
|
| /linux/fs/btrfs/ |
| H A D | accessors.c | 59 char *kaddr = folio_address(eb->folios[idx]) + oif; \ 69 return get_unaligned_le##bits(kaddr); \ 72 lebytes[0] = *kaddr; \ 73 kaddr = folio_address(eb->folios[idx + 1]); \ 74 lebytes[1] = *kaddr; \ 76 memcpy_split_src(lebytes, kaddr, \ 89 char *kaddr = folio_address(eb->folios[idx]) + oif; \ 99 put_unaligned_le##bits(val, kaddr); \ 104 *kaddr = lebytes[0]; \ 105 kaddr = folio_address(eb->folios[idx + 1]); \ [all …]
|
| H A D | lzo.c | 205 char *kaddr; in copy_compressed_data_to_bio() local 227 kaddr = kmap_local_folio(*out_folio, offset_in_folio(*out_folio, *total_out)); in copy_compressed_data_to_bio() 228 write_compress_length(kaddr, compressed_size); in copy_compressed_data_to_bio() 229 kunmap_local(kaddr); in copy_compressed_data_to_bio() 253 kaddr = kmap_local_folio(*out_folio, foffset); in copy_compressed_data_to_bio() 254 memcpy(kaddr, compressed_data + *total_out - copy_start, copy_len); in copy_compressed_data_to_bio() 255 kunmap_local(kaddr); in copy_compressed_data_to_bio() 435 char *kaddr; in lzo_decompress_bio() local 451 kaddr = kmap_local_folio(fi.folio, 0); in lzo_decompress_bio() 452 len_in = read_compress_length(kaddr); in lzo_decompress_bio() [all …]
|
| /linux/arch/m68k/sun3/ |
| H A D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument 29 j = *(volatile unsigned long *)kaddr; in dvma_page() 30 *(volatile unsigned long *)kaddr = j; in dvma_page() 32 ptep = pfn_pte(virt_to_pfn((void *)kaddr), PAGE_KERNEL); in dvma_page() 40 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page() 44 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument 56 dvma_page(kaddr, vaddr); in dvma_map_iommu() 57 kaddr += PAGE_SIZE; in dvma_map_iommu()
|
| /linux/drivers/acpi/ |
| H A D | nvs.c | 78 void *kaddr; member 138 if (entry->kaddr) { in suspend_nvs_free() 140 iounmap(entry->kaddr); in suspend_nvs_free() 143 acpi_os_unmap_iomem(entry->kaddr, in suspend_nvs_free() 146 entry->kaddr = NULL; in suspend_nvs_free() 182 entry->kaddr = acpi_os_get_iomem(phys, size); in suspend_nvs_save() 183 if (!entry->kaddr) { in suspend_nvs_save() 184 entry->kaddr = acpi_os_ioremap(phys, size); in suspend_nvs_save() 185 entry->unmap = !!entry->kaddr; in suspend_nvs_save() 187 if (!entry->kaddr) { in suspend_nvs_save() [all …]
|
| /linux/arch/m68k/sun3x/ |
| H A D | dvma.c | 78 inline int dvma_map_cpu(unsigned long kaddr, in dvma_map_cpu() argument 87 kaddr &= PAGE_MASK; in dvma_map_cpu() 92 pr_debug("dvma: mapping kern %08lx to virt %08lx\n", kaddr, vaddr); in dvma_map_cpu() 127 __pa(kaddr), vaddr); in dvma_map_cpu() 128 set_pte(pte, pfn_pte(virt_to_pfn((void *)kaddr), in dvma_map_cpu() 131 kaddr += PAGE_SIZE; in dvma_map_cpu() 146 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, int len) in dvma_map_iommu() argument 162 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu() 167 kaddr += DVMA_PAGE_SIZE; in dvma_map_iommu()
|
| /linux/arch/arm/include/asm/ |
| H A D | tlbflush.h | 472 static inline void __local_flush_tlb_kernel_page(unsigned long kaddr) in __local_flush_tlb_kernel_page() argument 477 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page() 478 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page() 479 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page() 483 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page() 484 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page() 485 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page() 488 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument 492 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page() 497 __local_flush_tlb_kernel_page(kaddr); in local_flush_tlb_kernel_page() [all …]
|
| /linux/tools/testing/nvdimm/ |
| H A D | pmem-dax.c | 12 long nr_pages, enum dax_access_mode mode, void **kaddr, in __pmem_direct_access() argument 28 if (kaddr) in __pmem_direct_access() 29 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access() 39 if (kaddr) in __pmem_direct_access() 40 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | kfence.h | 20 char *kaddr, *vaddr; in arch_kfence_init_pool() local 40 kaddr = kfence_pool; in arch_kfence_init_pool() 42 while (kaddr < kfence_pool + KFENCE_POOL_SIZE) { in arch_kfence_init_pool() 43 set_page_address(virt_to_page(kaddr), vaddr); in arch_kfence_init_pool() 44 kaddr += PAGE_SIZE; in arch_kfence_init_pool()
|
| H A D | io.h | 71 #define virt_to_phys(kaddr) \ argument 73 (likely((unsigned long)kaddr < vm_map_base)) ? __pa((unsigned long)kaddr) : \ 74 page_to_phys(tlb_virt_to_page((unsigned long)kaddr)) + offset_in_page((unsigned long)kaddr);\
|
| /linux/fs/erofs/ |
| H A D | xattr.c | 15 void *kaddr; member 124 it->kaddr = erofs_bread(&it->buf, it->pos, true); in erofs_xattr_copy_to_buffer() 125 if (IS_ERR(it->kaddr)) in erofs_xattr_copy_to_buffer() 126 return PTR_ERR(it->kaddr); in erofs_xattr_copy_to_buffer() 130 memcpy(it->buffer + it->buffer_ofs, it->kaddr, slice); in erofs_xattr_copy_to_buffer() 145 entry = *(struct erofs_xattr_entry *)it->kaddr; in erofs_listxattr_foreach() 195 entry = *(struct erofs_xattr_entry *)it->kaddr; in erofs_getxattr_foreach() 226 it->kaddr = erofs_bread(&it->buf, it->pos, true); in erofs_getxattr_foreach() 227 if (IS_ERR(it->kaddr)) in erofs_getxattr_foreach() 228 return PTR_ERR(it->kaddr); in erofs_getxattr_foreach() [all …]
|
| /linux/arch/csky/include/asm/ |
| H A D | page.h | 33 #define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && \ argument 34 (void *)(kaddr) < high_memory) 73 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument 75 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
|
| /linux/arch/m68k/include/asm/ |
| H A D | page_mm.h | 123 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument 125 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn() 144 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (… argument
|
| /linux/drivers/tee/amdtee/ |
| H A D | shm_pool.c | 26 shm->kaddr = (void *)va; in pool_op_alloc() 34 shm->kaddr = NULL; in pool_op_alloc() 45 free_pages((unsigned long)shm->kaddr, get_order(shm->size)); in pool_op_free() 46 shm->kaddr = NULL; in pool_op_free()
|
| /linux/drivers/misc/ |
| H A D | open-dice.c | 45 void *kaddr; in open_dice_wipe() local 48 kaddr = devm_memremap(drvdata->misc.this_device, drvdata->rmem->base, in open_dice_wipe() 50 if (IS_ERR(kaddr)) { in open_dice_wipe() 52 return PTR_ERR(kaddr); in open_dice_wipe() 55 memset(kaddr, 0, drvdata->rmem->size); in open_dice_wipe() 56 devm_memunmap(drvdata->misc.this_device, kaddr); in open_dice_wipe()
|
| /linux/arch/sh/mm/ |
| H A D | cache.c | 130 void *kaddr = kmap_atomic(page); in clear_user_highpage() local 132 clear_page(kaddr); in clear_user_highpage() 134 if (pages_do_alias((unsigned long)kaddr, vaddr & PAGE_MASK)) in clear_user_highpage() 135 __flush_purge_region(kaddr, PAGE_SIZE); in clear_user_highpage() 137 kunmap_atomic(kaddr); in clear_user_highpage() 166 void *kaddr; in __flush_anon_page() local 168 kaddr = kmap_coherent(page, vmaddr); in __flush_anon_page() 171 kunmap_coherent(kaddr); in __flush_anon_page()
|
| H A D | uncached.c | 24 int virt_addr_uncached(unsigned long kaddr) in virt_addr_uncached() argument 26 return (kaddr >= uncached_start) && (kaddr < uncached_end); in virt_addr_uncached()
|
| /linux/drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
| H A D | smu8_smumgr.c | 333 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_populate_single_scratch_task() 355 (struct smu8_ih_meta_data *)smu8_smu->scratch_buffer[i].kaddr; in smu8_smu_populate_single_scratch_task() 370 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_populate_single_ucode_load_task() 409 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_initialize_toc_empty_job_list() 420 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_construct_toc_for_vddgfx_enter() 438 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_construct_toc_for_vddgfx_exit() 590 entry->kaddr = (char *) smu8_smu->smu_buffer.kaddr + in smu8_smu_populate_single_scratch_entry() 611 *table = (struct SMU8_Fusion_ClkTable *)smu8_smu->scratch_buffer[i].kaddr; in smu8_download_pptable_settings() 781 &smu8_smu->toc_buffer.kaddr); in smu8_smu_init() 791 &smu8_smu->smu_buffer.kaddr); in smu8_smu_init() [all …]
|
| /linux/drivers/misc/lkdtm/ |
| H A D | usercopy.c | 351 static void do_usercopy_page_span(const char *name, void *kaddr) in do_usercopy_page_span() argument 363 memset(kaddr, 0xAA, PAGE_SIZE); in do_usercopy_page_span() 366 kaddr += PAGE_SIZE / 2; in do_usercopy_page_span() 369 name, kaddr); in do_usercopy_page_span() 370 if (copy_to_user((void __user *)uaddr, kaddr, in do_usercopy_page_span() 377 name, kaddr); in do_usercopy_page_span() 378 if (copy_to_user((void __user *)uaddr, kaddr, unconst + PAGE_SIZE)) { in do_usercopy_page_span()
|
| /linux/arch/arm/mm/ |
| H A D | flush.c | 128 void __flush_ptrace_access(struct page *page, unsigned long uaddr, void *kaddr, in __flush_ptrace_access() argument 133 unsigned long addr = (unsigned long)kaddr; in __flush_ptrace_access() 147 unsigned long addr = (unsigned long)kaddr; in __flush_ptrace_access() 160 unsigned long uaddr, void *kaddr, unsigned long len) in flush_ptrace_access() argument 167 __flush_ptrace_access(page, uaddr, kaddr, len, flags); in flush_ptrace_access() 171 void *kaddr, unsigned long len) in flush_uprobe_xol_access() argument 175 __flush_ptrace_access(page, uaddr, kaddr, len, flags); in flush_uprobe_xol_access()
|
| H A D | copypage-v4wt.c | 62 void *ptr, *kaddr = kmap_atomic(page); in v4wt_clear_user_highpage() local 77 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wt_clear_user_highpage() 79 kunmap_atomic(kaddr); in v4wt_clear_user_highpage()
|
| H A D | copypage-fa.c | 57 void *ptr, *kaddr = kmap_atomic(page); in fa_clear_user_highpage() local 74 : "0" (kaddr), "I" (PAGE_SIZE / 32) in fa_clear_user_highpage() 76 kunmap_atomic(kaddr); in fa_clear_user_highpage()
|
| H A D | copypage-v4wb.c | 67 void *ptr, *kaddr = kmap_atomic(page); in v4wb_clear_user_highpage() local 84 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wb_clear_user_highpage() 86 kunmap_atomic(kaddr); in v4wb_clear_user_highpage()
|
| H A D | copypage-xsc3.c | 82 void *ptr, *kaddr = kmap_atomic(page); in xsc3_mc_clear_user_highpage() local 96 : "0" (kaddr), "I" (PAGE_SIZE / 32) in xsc3_mc_clear_user_highpage() 98 kunmap_atomic(kaddr); in xsc3_mc_clear_user_highpage()
|
| H A D | copypage-feroceon.c | 81 void *ptr, *kaddr = kmap_atomic(page); in feroceon_clear_user_highpage() local 99 : "0" (kaddr), "I" (PAGE_SIZE / 32) in feroceon_clear_user_highpage() 101 kunmap_atomic(kaddr); in feroceon_clear_user_highpage()
|