| /linux/sound/soc/codecs/ |
| H A D | lpass-va-macro.c | 460 static int va_clk_rsc_fs_gen_request(struct va_macro *va, bool enable) in va_clk_rsc_fs_gen_request() argument 462 struct regmap *regmap = va->regmap; in va_clk_rsc_fs_gen_request() 493 static int va_macro_mclk_enable(struct va_macro *va, bool mclk_enable) in va_macro_mclk_enable() argument 495 struct regmap *regmap = va->regmap; in va_macro_mclk_enable() 498 va_clk_rsc_fs_gen_request(va, true); in va_macro_mclk_enable() 502 va_clk_rsc_fs_gen_request(va, false); in va_macro_mclk_enable() 512 struct va_macro *va = snd_soc_component_get_drvdata(comp); in va_macro_mclk_event() local 516 return clk_prepare_enable(va->fsgen); in va_macro_mclk_event() 518 clk_disable_unprepare(va->fsgen); in va_macro_mclk_event() 573 struct va_macro *va = snd_soc_component_get_drvdata(component); in va_macro_tx_mixer_get() local [all …]
|
| /linux/mm/ |
| H A D | vmalloc.c | 1049 va_size(struct vmap_area *va) in va_size() argument 1051 return (va->va_end - va->va_start); in va_size() 1057 struct vmap_area *va; in get_subtree_max_size() local 1059 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size() 1060 return va ? va->subtree_max_size : 0; in get_subtree_max_size() 1086 struct vmap_area *va; in __find_vmap_area() local 1088 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area() 1089 if (addr < va->va_start) in __find_vmap_area() 1091 else if (addr >= va->va_end) in __find_vmap_area() 1094 return va; in __find_vmap_area() [all …]
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_uvmm.c | 70 } va; member 174 u64 addr = reg->va.addr; in nouveau_uvma_region_sparse_unref() 175 u64 range = reg->va.range; in nouveau_uvma_region_sparse_unref() 183 u64 addr = uvma->va.va.addr; in nouveau_uvma_vmm_put() 184 u64 range = uvma->va.va.range; in nouveau_uvma_vmm_put() 194 u64 addr = uvma->va.va.addr; in nouveau_uvma_map() 195 u64 offset = uvma->va.gem.offset; in nouveau_uvma_map() 196 u64 range = uvma->va.va.range; in nouveau_uvma_map() 207 u64 addr = uvma->va.va.addr; in nouveau_uvma_unmap() 208 u64 range = uvma->va.va.range; in nouveau_uvma_unmap() [all …]
|
| /linux/drivers/scsi/qedi/ |
| H A D | qedi_dbg.c | 14 va_list va; in qedi_dbg_err() local 17 va_start(va, fmt); in qedi_dbg_err() 20 vaf.va = &va; in qedi_dbg_err() 28 va_end(va); in qedi_dbg_err() 35 va_list va; in qedi_dbg_warn() local 38 va_start(va, fmt); in qedi_dbg_warn() 41 vaf.va = &va; in qedi_dbg_warn() 53 va_end(va); in qedi_dbg_warn() 60 va_list va; in qedi_dbg_notice() local 63 va_start(va, fmt); in qedi_dbg_notice() [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_gpuvm.c | 966 #define GPUVA_START(node) ((node)->va.addr) 967 #define GPUVA_LAST(node) ((node)->va.addr + (node)->va.range - 1) 977 struct drm_gpuva *va); 978 static void __drm_gpuva_remove(struct drm_gpuva *va); 1009 u64 kstart = gpuvm->kernel_alloc_node.va.addr; in drm_gpuvm_in_kernel_node() 1010 u64 krange = gpuvm->kernel_alloc_node.va.range; in drm_gpuvm_in_kernel_node() 1128 gpuvm->kernel_alloc_node.va.addr = reserve_offset; in drm_gpuvm_init() 1129 gpuvm->kernel_alloc_node.va.range = reserve_range; in drm_gpuvm_init() 1143 if (gpuvm->kernel_alloc_node.va.range) in drm_gpuvm_fini() 1318 struct drm_gpuva *va; in drm_gpuvm_prepare_range() local [all …]
|
| /linux/drivers/scsi/qedf/ |
| H A D | qedf_dbg.c | 13 va_list va; in qedf_dbg_err() local 16 va_start(va, fmt); in qedf_dbg_err() 19 vaf.va = &va; in qedf_dbg_err() 27 va_end(va); in qedf_dbg_err() 34 va_list va; in qedf_dbg_warn() local 37 va_start(va, fmt); in qedf_dbg_warn() 40 vaf.va = &va; in qedf_dbg_warn() 52 va_end(va); in qedf_dbg_warn() 59 va_list va; in qedf_dbg_notice() local 62 va_start(va, fmt); in qedf_dbg_notice() [all …]
|
| /linux/drivers/iommu/generic_pt/ |
| H A D | pt_iter.h | 36 prefix = fvalog2_div(range->va, range->max_vasz_lg2 - 1) ? in pt_check_range() 43 if (!fvalog2_div_eq(range->va, prefix, range->max_vasz_lg2) || in pt_check_range() 61 pts->range->va = fvalog2_set_mod(pts->range->va, lower_va, in pt_index_to_va() 91 if (log2_mod(pts->range->va, oasz_lg2)) in pt_entry_fully_covered() 95 if (!log2_div_eq(range->va, range->last_va, oasz_lg2)) in pt_entry_fully_covered() 114 return log2_div(fvalog2_mod(pts->range->va, in pt_range_to_index() 117 return log2_mod(log2_div(pts->range->va, isz_lg2), in pt_range_to_index() 133 if (range->va == range->last_va) in pt_range_to_end_index() 145 if (log2_div_eq(range->va, range->last_va, num_entries_lg2 + isz_lg2)) in pt_range_to_end_index() 237 range.va = fvalog2_set_mod(pt_full_va_prefix(common), 0, max_vasz_lg2); in _pt_top_range() [all …]
|
| H A D | kunit_generic_pt.h | 11 static void do_map(struct kunit *test, pt_vaddr_t va, pt_oaddr_t pa, in do_map() argument 19 ret = iommu_map(&priv->domain, va, pa, len, IOMMU_READ | IOMMU_WRITE, in do_map() 119 range.last_va = fvalog2_set_mod_max(range.va, in check_all_levels() 128 range.va = range.last_va - (priv->smallest_pgsz - 1); in check_all_levels() 129 do_map(test, range.va, 0, priv->smallest_pgsz); in check_all_levels() 131 range = pt_make_range(priv->common, range.va, range.last_va); in check_all_levels() 189 static unsigned int ref_best_pgsize(pt_vaddr_t pgsz_bitmap, pt_vaddr_t va, in ref_best_pgsize() argument 197 log2_mod(va, pgsz_lg2) == 0 && in ref_best_pgsize() 199 va + log2_to_int(pgsz_lg2) - 1 <= last_va && in ref_best_pgsize() 200 log2_div_eq(va, va + log2_to_int(pgsz_lg2) - 1, pgsz_lg2) && in ref_best_pgsize() [all …]
|
| /linux/drivers/dio/ |
| H A D | dio.c | 127 void *va; in dio_find() local 139 va = (void *)(pa + DIO_VIRADDRBASE); in dio_find() 141 va = ioremap(pa, PAGE_SIZE); in dio_find() 144 (unsigned char *)va + DIO_IDOFF, 1)) { in dio_find() 146 iounmap(va); in dio_find() 150 prid = DIO_ID(va); in dio_find() 153 secid = DIO_SECID(va); in dio_find() 160 iounmap(va); in dio_find() 200 u_char *va; in dio_init() local 212 va = (void *)(pa + DIO_VIRADDRBASE); in dio_init() [all …]
|
| /linux/drivers/gpu/drm/imagination/ |
| H A D | pvr_vm.c | 119 #define to_pvr_vm_gpuva(va) container_of_const(va, struct pvr_vm_gpuva, base) argument 191 .map.va.addr = bind_op->device_addr, in pvr_vm_bind_op_exec() 192 .map.va.range = bind_op->size, in pvr_vm_bind_op_exec() 360 if ((op->map.gem.offset | op->map.va.range) & ~PVR_DEVICE_PAGE_MASK) in pvr_vm_gpuva_map() 363 err = pvr_mmu_map(ctx->mmu_op_ctx, op->map.va.range, pvr_gem->flags, in pvr_vm_gpuva_map() 364 op->map.va.addr); in pvr_vm_gpuva_map() 392 int err = pvr_mmu_unmap(ctx->mmu_op_ctx, op->unmap.va->va.addr, in pvr_vm_gpuva_unmap() 393 op->unmap.va->va.range); in pvr_vm_gpuva_unmap() 399 drm_gpuva_unlink(op->unmap.va); in pvr_vm_gpuva_unmap() 400 kfree(to_pvr_vm_gpuva(op->unmap.va)); in pvr_vm_gpuva_unmap() [all …]
|
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | ucall_common.c | 86 va_list va; in ucall_assert() local 95 va_start(va, fmt); in ucall_assert() 96 guest_vsnprintf(uc->buffer, UCALL_BUFFER_LEN, fmt, va); in ucall_assert() 97 va_end(va); in ucall_assert() 107 va_list va; in ucall_fmt() local 112 va_start(va, fmt); in ucall_fmt() 113 guest_vsnprintf(uc->buffer, UCALL_BUFFER_LEN, fmt, va); in ucall_fmt() 114 va_end(va); in ucall_fmt() 124 va_list va; in ucall() local 133 va_start(va, nargs); in ucall() [all …]
|
| /linux/arch/alpha/kernel/ |
| H A D | traps.c | 368 unsigned long count, va, pc; member 377 do_entUna(void * va, unsigned long opcode, unsigned long reg, in do_entUna() argument 386 unaligned[0].va = (unsigned long) va; in do_entUna() 404 : "r"(va), "0"(0)); in do_entUna() 420 : "r"(va), "0"(0)); in do_entUna() 436 : "r"(va), "0"(0)); in do_entUna() 464 : "r"(va), "r"(una_reg(reg)), "0"(0)); in do_entUna() 488 : "r"(va), "r"(una_reg(reg)), "0"(0)); in do_entUna() 512 : "r"(va), "r"(una_reg(reg)), "0"(0)); in do_entUna() 519 pc, va, opcode, reg); in do_entUna() [all …]
|
| /linux/fs/ceph/ |
| H A D | ceph_frag.c | 10 unsigned va = ceph_frag_value(a); in ceph_frag_compare() local 12 if (va < vb) in ceph_frag_compare() 14 if (va > vb) in ceph_frag_compare() 16 va = ceph_frag_bits(a); in ceph_frag_compare() 18 if (va < vb) in ceph_frag_compare() 20 if (va > vb) in ceph_frag_compare()
|
| /linux/lib/ |
| H A D | test_debug_virtual.c | 25 void *va; in test_debug_virtual_init() local 27 va = (void *)VMALLOC_START; in test_debug_virtual_init() 28 pa = virt_to_phys(va); in test_debug_virtual_init() 30 pr_info("PA: %pa for VA: 0x%lx\n", &pa, (unsigned long)va); in test_debug_virtual_init() 37 va = foo; in test_debug_virtual_init() 38 pr_info("PA: %pa for VA: 0x%lx\n", &pa, (unsigned long)va); in test_debug_virtual_init()
|
| /linux/arch/riscv/include/asm/ |
| H A D | sections.h | 18 static inline bool is_va_kernel_text(uintptr_t va) in is_va_kernel_text() argument 23 return va >= start && va < end; in is_va_kernel_text() 26 static inline bool is_va_kernel_lm_alias_text(uintptr_t va) in is_va_kernel_lm_alias_text() argument 31 return va >= start && va < end; in is_va_kernel_lm_alias_text()
|
| /linux/tools/testing/selftests/proc/ |
| H A D | proc-self-map-files-002.c | 55 unsigned long va; in main() local 64 for (va = 0; va < va_max; va += PAGE_SIZE) { in main() 65 p = mmap((void *)va, PAGE_SIZE, PROT_NONE, MAP_PRIVATE|MAP_FILE|MAP_FIXED, fd, 0); in main() 66 if (p == (void *)va) in main() 69 if (va == va_max) { in main()
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 158 def __virt_to_phys(self, va): argument 159 if va >= self.START_KERNEL_map: 160 return va - self.START_KERNEL_map + self.PHYS_BASE 162 return va - self.PAGE_OFFSET 164 def virt_to_phys(self, va): argument 165 return self.__virt_to_phys(va) 167 def virt_to_page(self, va): argument 168 return self.pfn_to_page(self.virt_to_pfn(va)) 170 def __pa(self, va): argument 171 return self.__virt_to_phys(va) [all …]
|
| /linux/arch/arm/mach-omap1/ |
| H A D | irq.c | 59 void __iomem *va; member 71 return readl_relaxed(irq_banks[bank].va + offset); in irq_bank_readl() 75 writel_relaxed(value, irq_banks[bank].va + offset); in irq_bank_writel() 81 writel_relaxed(0x1, irq_banks[1].va + IRQ_CONTROL_REG_OFFSET); in omap_ack_irq() 83 writel_relaxed(0x1, irq_banks[0].va + IRQ_CONTROL_REG_OFFSET); in omap_ack_irq() 137 void __iomem *l1 = irq_banks[0].va; in omap1_handle_irq() 138 void __iomem *l2 = irq_banks[1].va; in omap1_handle_irq() 208 irq_banks[i].va = ioremap(irq_banks[i].base_reg, 0xff); in omap1_init_irq() 209 if (WARN_ON(!irq_banks[i].va)) in omap1_init_irq() 247 omap_alloc_gc(irq_banks[i].va, irq_base + i * 32, 32); in omap1_init_irq()
|
| /linux/arch/powerpc/math-emu/ |
| H A D | math_efp.c | 183 union dw_union vc, va, vb; in do_spe_mathemu() local 202 va.wp[0] = current->thread.evr[fa]; in do_spe_mathemu() 203 va.wp[1] = regs->gpr[fa]; in do_spe_mathemu() 211 pr_debug("va: %08x %08x\n", va.wp[0], va.wp[1]); in do_spe_mathemu() 221 FP_UNPACK_SP(SA, va.wp + 1); in do_spe_mathemu() 227 FP_UNPACK_SP(SA, va.wp + 1); in do_spe_mathemu() 236 vc.wp[1] = va.wp[1] & ~SIGN_BIT_S; in do_spe_mathemu() 240 vc.wp[1] = va.wp[1] | SIGN_BIT_S; in do_spe_mathemu() 244 vc.wp[1] = va.wp[1] ^ SIGN_BIT_S; in do_spe_mathemu() 350 FP_UNPACK_DP(DA, va.dp); in do_spe_mathemu() [all …]
|
| /linux/arch/arm64/crypto/ |
| H A D | aes-ce-ccm-core.S | 26 .macro dround, va, vb, vk 27 aese \va\().16b, \vk\().16b 28 aesmc \va\().16b, \va\().16b 33 .macro aes_encrypt, va, vb, nr 35 dround \va, \vb, v10 36 dround \va, \vb, v11 38 dround \va, \vb, v12 39 dround \va, \vb, v13 41 dround \va, \vb, \v 43 aese \va\().16b, v4.16b
|
| /linux/arch/parisc/include/asm/ |
| H A D | special_insns.h | 5 #define lpa(va) ({ \ argument 14 : "r" (va) \ 20 #define lpa_user(va) ({ \ argument 29 : "r" (va) \ 47 #define prober_user(sr, va) ({ \ argument 57 : "i" (sr), "r" (va), "i" (PRIV_USER) \
|
| /linux/drivers/tee/amdtee/ |
| H A D | shm_pool.c | 15 unsigned long va; in pool_op_alloc() local 22 va = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in pool_op_alloc() 23 if (!va) in pool_op_alloc() 26 shm->kaddr = (void *)va; in pool_op_alloc() 27 shm->paddr = __psp_pa((void *)va); in pool_op_alloc() 33 free_pages(va, order); in pool_op_alloc()
|
| /linux/arch/powerpc/mm/book3s64/ |
| H A D | radix_tlb.c | 158 static __always_inline void __tlbiel_va(unsigned long va, unsigned long pid, in __tlbiel_va() argument 163 rb = va & ~(PPC_BITMASK(52, 63)); in __tlbiel_va() 174 static __always_inline void __tlbie_va(unsigned long va, unsigned long pid, in __tlbie_va() argument 179 rb = va & ~(PPC_BITMASK(52, 63)); in __tlbie_va() 190 static __always_inline void __tlbie_lpid_va(unsigned long va, unsigned long lpid, in __tlbie_lpid_va() argument 195 rb = va & ~(PPC_BITMASK(52, 63)); in __tlbie_lpid_va() 207 static inline void fixup_tlbie_va(unsigned long va, unsigned long pid, in fixup_tlbie_va() argument 212 __tlbie_va(va, 0, ap, RIC_FLUSH_TLB); in fixup_tlbie_va() 217 __tlbie_va(va, pid, ap, RIC_FLUSH_TLB); in fixup_tlbie_va() 221 static inline void fixup_tlbie_va_range(unsigned long va, unsigned long pid, in fixup_tlbie_va_range() argument [all …]
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_mr.c | 329 u8 *va; in rxe_mr_copy_xarray() local 341 va = kmap_local_page(info->page); in rxe_mr_copy_xarray() 344 memcpy(addr, va + page_offset, bytes); in rxe_mr_copy_xarray() 346 memcpy(va + page_offset, addr, bytes); in rxe_mr_copy_xarray() 347 kunmap_local(va); in rxe_mr_copy_xarray() 363 u8 *va; in rxe_mr_copy_dma() local 369 va = kmap_local_page(page); in rxe_mr_copy_dma() 372 memcpy(va + page_offset, addr, bytes); in rxe_mr_copy_dma() 374 memcpy(addr, va + page_offset, bytes); in rxe_mr_copy_dma() 376 kunmap_local(va); in rxe_mr_copy_dma() [all …]
|
| /linux/arch/arm64/kvm/hyp/vhe/ |
| H A D | tlb.c | 232 int __kvm_tlbi_s1e2(struct kvm_s2_mmu *mmu, u64 va, u64 sys_encoding) in __kvm_tlbi_s1e2() argument 274 __tlbi(vae1is, va); in __kvm_tlbi_s1e2() 288 __tlbi(vale1is, va); in __kvm_tlbi_s1e2() 296 __tlbi(aside1is, va); in __kvm_tlbi_s1e2() 304 __tlbi(vaae1is, va); in __kvm_tlbi_s1e2() 312 __tlbi(vaale1is, va); in __kvm_tlbi_s1e2() 326 __tlbi(rvae1is, va); in __kvm_tlbi_s1e2() 340 __tlbi(rvale1is, va); in __kvm_tlbi_s1e2() 348 __tlbi(rvaae1is, va); in __kvm_tlbi_s1e2() 356 __tlbi(rvaale1is, va); in __kvm_tlbi_s1e2()
|