Home
last modified time | relevance | path

Searched full:va (Results 1 – 25 of 845) sorted by relevance

12345678910>>...34

/linux/sound/soc/codecs/
H A Dlpass-va-macro.c22 /* VA macro registers */
274 /* VA macro */
298 /* VA core */
460 static int va_clk_rsc_fs_gen_request(struct va_macro *va, bool enable) in va_clk_rsc_fs_gen_request() argument
462 struct regmap *regmap = va->regmap; in va_clk_rsc_fs_gen_request()
493 static int va_macro_mclk_enable(struct va_macro *va, bool mclk_enable) in va_macro_mclk_enable() argument
495 struct regmap *regmap = va->regmap; in va_macro_mclk_enable()
498 va_clk_rsc_fs_gen_request(va, true); in va_macro_mclk_enable()
502 va_clk_rsc_fs_gen_request(va, false); in va_macro_mclk_enable()
512 struct va_macro *va = snd_soc_component_get_drvdata(comp); in va_macro_mclk_event() local
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_uvmm.c6 * The uvmm mutex protects any operations on the GPU VA space provided by the
7 * DRM GPU VA manager.
70 } va; member
174 u64 addr = reg->va.addr; in nouveau_uvma_region_sparse_unref()
175 u64 range = reg->va.range; in nouveau_uvma_region_sparse_unref()
183 u64 addr = uvma->va.va.addr; in nouveau_uvma_vmm_put()
184 u64 range = uvma->va.va.range; in nouveau_uvma_vmm_put()
194 u64 addr = uvma->va.va.addr; in nouveau_uvma_map()
195 u64 offset = uvma->va.gem.offset; in nouveau_uvma_map()
196 u64 range = uvma->va.va.range; in nouveau_uvma_map()
[all …]
/linux/mm/
H A Dvmalloc.c890 * All vmap_area objects in this tree are sorted by va->va_start
1049 va_size(struct vmap_area *va) in va_size() argument
1051 return (va->va_end - va->va_start); in va_size()
1057 struct vmap_area *va; in get_subtree_max_size() local
1059 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size()
1060 return va ? va->subtree_max_size : 0; in get_subtree_max_size()
1086 struct vmap_area *va; in __find_vmap_area() local
1088 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area()
1089 if (addr < va->va_start) in __find_vmap_area()
1091 else if (addr >= va->va_end) in __find_vmap_area()
[all …]
/linux/drivers/scsi/qedi/
H A Dqedi_dbg.c14 va_list va; in qedi_dbg_err() local
17 va_start(va, fmt); in qedi_dbg_err()
20 vaf.va = &va; in qedi_dbg_err()
28 va_end(va); in qedi_dbg_err()
35 va_list va; in qedi_dbg_warn() local
38 va_start(va, fmt); in qedi_dbg_warn()
41 vaf.va = &va; in qedi_dbg_warn()
53 va_end(va); in qedi_dbg_warn()
60 va_list va; in qedi_dbg_notice() local
63 va_start(va, fmt); in qedi_dbg_notice()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_gpuvm.c38 * The DRM GPU VA Manager, represented by struct drm_gpuvm keeps track of a
39 * GPU's virtual address (VA) space and manages the corresponding virtual
44 * all existing GPU VA mappings using this &drm_gem_object as backing buffer.
49 * The GPU VA manager internally uses a rb-tree to manage the
53 * portion of VA space reserved by the kernel. This node is initialized together
54 * with the GPU VA manager instance and removed when the GPU VA manager is
107 * Besides its capability to manage and represent a GPU VA space, the
108 * GPU VA manager also provides functions to let the &drm_gpuvm calculate a
111 * Therefore the DRM GPU VA manager provides an algorithm implementing splitting
112 * and merging of existing GPU VA mappings with the ones that are requested to
[all …]
/linux/drivers/iommu/generic_pt/
H A Dpt_iter.h24 * Check that VA and last_va fall within the permitted range of VAs. If the
36 prefix = fvalog2_div(range->va, range->max_vasz_lg2 - 1) ? in pt_check_range()
43 if (!fvalog2_div_eq(range->va, prefix, range->max_vasz_lg2) || in pt_check_range()
50 * pt_index_to_va() - Update range->va to the current pts->index
53 * Adjust range->va to match the current index. This is done in a lazy manner
54 * since computing the VA takes several instructions and is rarely required.
61 pts->range->va = fvalog2_set_mod(pts->range->va, lower_va, in pt_index_to_va()
66 * Add index_count_lg2 number of entries to pts's VA and index. The VA will be
91 if (log2_mod(pts->range->va, oasz_lg2)) in pt_entry_fully_covered()
95 if (!log2_div_eq(range->va, range->last_va, oasz_lg2)) in pt_entry_fully_covered()
[all …]
H A Dkunit_generic_pt.h11 static void do_map(struct kunit *test, pt_vaddr_t va, pt_oaddr_t pa, in do_map() argument
19 ret = iommu_map(&priv->domain, va, pa, len, IOMMU_READ | IOMMU_WRITE, in do_map()
48 * If we were able to use the full VA space this should always be the in __check_all_levels()
119 range.last_va = fvalog2_set_mod_max(range.va, in check_all_levels()
123 * Map a page at the highest VA, this will populate all the levels so we in check_all_levels()
128 range.va = range.last_va - (priv->smallest_pgsz - 1); in check_all_levels()
129 do_map(test, range.va, 0, priv->smallest_pgsz); in check_all_levels()
131 range = pt_make_range(priv->common, range.va, range.last_va); in check_all_levels()
189 static unsigned int ref_best_pgsize(pt_vaddr_t pgsz_bitmap, pt_vaddr_t va, in ref_best_pgsize() argument
197 log2_mod(va, pgsz_lg2) == 0 && in ref_best_pgsize()
[all …]
H A Dkunit_iommu_pt.h10 static void do_map(struct kunit *test, pt_vaddr_t va, pt_oaddr_t pa,
77 static void do_unmap(struct kunit *test, pt_vaddr_t va, pt_vaddr_t len) in do_unmap() argument
82 ret = iommu_unmap(&priv->domain, va, len); in do_unmap()
86 static void check_iova(struct kunit *test, pt_vaddr_t va, pt_oaddr_t pa, in check_iova() argument
90 pt_vaddr_t pfn = log2_div(va, priv->smallest_pgsz_lg2); in check_iova()
122 if (top_range.va == 0) in test_increase_level()
126 do_map(test, top_range.va - priv->smallest_pgsz, 0, in test_increase_level()
146 cur_va = range.va + priv->smallest_pgsz * 256; in test_map_simple()
174 cur_va = range.va + priv->smallest_pgsz * 256; in test_map_simple()
215 cur_va = ALIGN(range.va + priv->smallest_pgsz * 256, in test_map_table_to_oa()
[all …]
/linux/drivers/scsi/qedf/
H A Dqedf_dbg.c13 va_list va; in qedf_dbg_err() local
16 va_start(va, fmt); in qedf_dbg_err()
19 vaf.va = &va; in qedf_dbg_err()
27 va_end(va); in qedf_dbg_err()
34 va_list va; in qedf_dbg_warn() local
37 va_start(va, fmt); in qedf_dbg_warn()
40 vaf.va = &va; in qedf_dbg_warn()
52 va_end(va); in qedf_dbg_warn()
59 va_list va; in qedf_dbg_notice() local
62 va_start(va, fmt); in qedf_dbg_notice()
[all …]
/linux/drivers/dio/
H A Ddio.c127 void *va; in dio_find() local
139 va = (void *)(pa + DIO_VIRADDRBASE); in dio_find()
141 va = ioremap(pa, PAGE_SIZE); in dio_find()
144 (unsigned char *)va + DIO_IDOFF, 1)) { in dio_find()
146 iounmap(va); in dio_find()
150 prid = DIO_ID(va); in dio_find()
153 secid = DIO_SECID(va); in dio_find()
160 iounmap(va); in dio_find()
200 u_char *va; in dio_init() local
212 va = (void *)(pa + DIO_VIRADDRBASE); in dio_init()
[all …]
/linux/tools/testing/selftests/kvm/lib/
H A Ducall_common.c86 va_list va; in ucall_assert() local
95 va_start(va, fmt); in ucall_assert()
96 guest_vsnprintf(uc->buffer, UCALL_BUFFER_LEN, fmt, va); in ucall_assert()
97 va_end(va); in ucall_assert()
107 va_list va; in ucall_fmt() local
112 va_start(va, fmt); in ucall_fmt()
113 guest_vsnprintf(uc->buffer, UCALL_BUFFER_LEN, fmt, va); in ucall_fmt()
114 va_end(va); in ucall_fmt()
124 va_list va; in ucall() local
133 va_start(va, nargs); in ucall()
[all …]
/linux/arch/arm64/kvm/
H A Dva_layout.c17 * The LSB of the HYP VA tag
21 * The HYP VA tag value with the region bit
27 * Compute HYP VA by using the same computation as kern_hyp_va().
37 * Store a hyp VA <-> PA offset into a EL2-owned variable.
43 /* Compute the offset from the hyp VA and PA of a random symbol. */ in init_hyp_physvirt_offset()
50 * Calculate the actual VA size used by the hypervisor
56 * be different from the number of VA bits used by the regular kernel in kvm_hyp_va_bits()
64 * So use the maximum of the idmap VA bits and the regular kernel stage in kvm_hyp_va_bits()
65 * 1 VA bits as the hypervisor VA size to assure that the hypervisor can in kvm_hyp_va_bits()
72 * We want to generate a hyp VA with the following format (with V ==
[all …]
/linux/scripts/gdb/linux/
H A Dmm.py158 def __virt_to_phys(self, va): argument
159 if va >= self.START_KERNEL_map:
160 return va - self.START_KERNEL_map + self.PHYS_BASE
162 return va - self.PAGE_OFFSET
164 def virt_to_phys(self, va): argument
165 return self.__virt_to_phys(va)
167 def virt_to_page(self, va): argument
168 return self.pfn_to_page(self.virt_to_pfn(va))
170 def __pa(self, va): argument
171 return self.__virt_to_phys(va)
[all …]
/linux/lib/
H A Dtest_debug_virtual.c25 void *va; in test_debug_virtual_init() local
27 va = (void *)VMALLOC_START; in test_debug_virtual_init()
28 pa = virt_to_phys(va); in test_debug_virtual_init()
30 pr_info("PA: %pa for VA: 0x%lx\n", &pa, (unsigned long)va); in test_debug_virtual_init()
37 va = foo; in test_debug_virtual_init()
38 pr_info("PA: %pa for VA: 0x%lx\n", &pa, (unsigned long)va); in test_debug_virtual_init()
/linux/drivers/gpu/drm/panthor/
H A Dpanthor_mmu.c176 /** @va: Virtual range targeted by the VM operation. */
178 /** @va.addr: Start address. */
181 /** @va.range: Range size. */
183 } va; member
202 * @map.new_vma: The new VMA object that will be inserted to the VA tree.
229 * We delegate all the VA management to the common drm_gpuvm framework
278 * @mm: Memory management object representing the auto-VA/kernel-VA.
280 * Used to auto-allocate VA space for kernel-managed objects (tiler
283 * For the MCU VM, this is managing the VA range that's used to map
287 * exceed half of the VA space addressable.
[all …]
/linux/drivers/gpu/drm/imagination/
H A Dpvr_vm.c119 #define to_pvr_vm_gpuva(va) container_of_const(va, struct pvr_vm_gpuva, base) argument
148 * @new_va: Prealloced VA mapping object (init in callback).
154 * @prev_va: Prealloced VA mapping object (init in callback).
161 * @next_va: Prealloced VA mapping object (init in callback).
191 .map.va.addr = bind_op->device_addr, in pvr_vm_bind_op_exec()
192 .map.va.range = bind_op->size, in pvr_vm_bind_op_exec()
360 if ((op->map.gem.offset | op->map.va.range) & ~PVR_DEVICE_PAGE_MASK) in pvr_vm_gpuva_map()
363 err = pvr_mmu_map(ctx->mmu_op_ctx, op->map.va.range, pvr_gem->flags, in pvr_vm_gpuva_map()
364 op->map.va.addr); in pvr_vm_gpuva_map()
392 int err = pvr_mmu_unmap(ctx->mmu_op_ctx, op->unmap.va->va.addr, in pvr_vm_gpuva_unmap()
[all …]
/linux/drivers/infiniband/hw/usnic/
H A Dusnic_uiom.c59 usnic_err("Device %s iommu fault domain 0x%p va 0x%lx flags 0x%x\n", in usnic_uiom_dma_fault()
165 usnic_dbg("va: 0x%lx pa: %pa\n", in usnic_uiom_get_pages()
193 long unsigned va, size; in usnic_uiom_unmap_sorted_intervals() local
196 va = interval->start << PAGE_SHIFT; in usnic_uiom_unmap_sorted_intervals()
200 usnic_dbg("va 0x%lx size 0x%lx", va, PAGE_SIZE); in usnic_uiom_unmap_sorted_intervals()
201 iommu_unmap(pd->domain, va, PAGE_SIZE); in usnic_uiom_unmap_sorted_intervals()
202 va += PAGE_SIZE; in usnic_uiom_unmap_sorted_intervals()
219 vpn_start = (uiomr->va & PAGE_MASK) >> PAGE_SHIFT; in __usnic_uiom_reg_release()
250 long int va = uiomr->va & PAGE_MASK; in usnic_uiom_map_sorted_intervals() local
258 for (i = 0; i < chunk->nents; i++, va += PAGE_SIZE) { in usnic_uiom_map_sorted_intervals()
[all …]
/linux/Documentation/devicetree/bindings/sound/
H A Dcs42l56.txt9 - VA-supply, VCP-supply, VLDO-supply : power supplies for the device,
31 0 = 0.5 x VA
32 1 = 0.6 x VA
33 2 = 0.7 x VA
34 3 = 0.8 x VA
35 4 = 0.83 x VA
36 5 = 0.91 x VA
62 VA-supply = <&reg_audio>;
/linux/arch/parisc/kernel/
H A Dentry.S176 va = r8 /* virtual address for which the trap occurred */ define
189 mfctl %pcoq, va
206 mfctl %pcoq, va
220 mfctl %ior,va
238 mfctl %ior,va
252 mfctl %ior, va
270 mfctl %ior, va
282 mfctl %ior,va
298 mfctl %ior,va
312 mfctl %ior,va
[all …]
/linux/fs/ceph/
H A Dceph_frag.c10 unsigned va = ceph_frag_value(a); in ceph_frag_compare() local
12 if (va < vb) in ceph_frag_compare()
14 if (va > vb) in ceph_frag_compare()
16 va = ceph_frag_bits(a); in ceph_frag_compare()
18 if (va < vb) in ceph_frag_compare()
20 if (va > vb) in ceph_frag_compare()
/linux/drivers/gpu/drm/lima/
H A Dlima_vm.c29 #define LIMA_PDE(va) (va >> LIMA_VM_PD_SHIFT) argument
30 #define LIMA_PTE(va) ((va & LIMA_VM_PT_MASK) >> LIMA_VM_PT_SHIFT) argument
31 #define LIMA_PBE(va) (va >> LIMA_VM_PB_SHIFT) argument
32 #define LIMA_BTE(va) ((va & LIMA_VM_BT_MASK) >> LIMA_VM_BT_SHIFT) argument
47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() argument
49 u32 pbe = LIMA_PBE(va); in lima_vm_map_page()
50 u32 bte = LIMA_BTE(va); in lima_vm_map_page()
81 list_for_each_entry(bo_va, &bo->va, list) { in lima_vm_bo_find()
138 list_add_tail(&bo_va->list, &bo->va); in lima_vm_bo_add()
/linux/arch/riscv/include/asm/
H A Dsections.h18 static inline bool is_va_kernel_text(uintptr_t va) in is_va_kernel_text() argument
23 return va >= start && va < end; in is_va_kernel_text()
26 static inline bool is_va_kernel_lm_alias_text(uintptr_t va) in is_va_kernel_lm_alias_text() argument
31 return va >= start && va < end; in is_va_kernel_lm_alias_text()
/linux/drivers/misc/vmw_vmci/
H A Dvmci_queue_pair.h35 u64 ppn_va; /* Start VA of queue pair PPNs. */
44 u64 va; /* Start VA of queue pair PPNs. */ member
54 * pass down the VA of the mapped file. Before host support was added
59 * provide the VA of the mapped files.
80 u64 produce_page_file; /* User VA. */
81 u64 consume_page_file; /* User VA. */
86 u64 produce_va; /* User VA of the mapped file. */
87 u64 consume_va; /* User VA of the mapped file. */
101 * this is a list of PPNs, and on hosted, it is a user VA where the
/linux/drivers/media/platform/mediatek/vcodec/common/
H A Dmtk_vcodec_util.c66 mem->va = dma_alloc_attrs(&plat_dev->dev, mem->size, &mem->dma_addr, in mtk_vcodec_mem_alloc()
68 if (!mem->va) { in mtk_vcodec_mem_alloc()
74 mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va, in mtk_vcodec_mem_alloc()
99 if (!mem->va) { in mtk_vcodec_mem_free()
100 mtk_v4l2_err(plat_dev, "%s: Tried to free a NULL VA", __func__); in mtk_vcodec_mem_free()
106 mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va, in mtk_vcodec_mem_free()
109 dma_free_coherent(&plat_dev->dev, mem->size, mem->va, mem->dma_addr); in mtk_vcodec_mem_free()
110 mem->va = NULL; in mtk_vcodec_mem_free()
/linux/arch/arm64/crypto/
H A Daes-ce-ccm-core.S26 .macro dround, va, vb, vk
27 aese \va\().16b, \vk\().16b
28 aesmc \va\().16b, \va\().16b
33 .macro aes_encrypt, va, vb, nr
35 dround \va, \vb, v10
36 dround \va, \vb, v11
38 dround \va, \vb, v12
39 dround \va, \vb, v13
41 dround \va, \vb, \v
43 aese \va\().16b, v4.16b

12345678910>>...34