| /linux/arch/m68k/include/asm/ |
| H A D | bitops.h | 31 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) in bset_reg_set_bit() argument 33 char *p = (char *)vaddr + (nr ^ 31) / 8; in bset_reg_set_bit() 41 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) in bset_mem_set_bit() argument 43 char *p = (char *)vaddr + (nr ^ 31) / 8; in bset_mem_set_bit() 50 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) in bfset_mem_set_bit() argument 54 : "d" (nr ^ 31), "o" (*vaddr) in bfset_mem_set_bit() 59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) argument 61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) argument 63 #define set_bit(nr, vaddr) (__builtin_constant_p(nr) ? \ argument 64 bset_mem_set_bit(nr, vaddr) : \ [all …]
|
| /linux/arch/riscv/mm/ |
| H A D | kasan_init.c | 29 static void __init kasan_populate_pte(pmd_t *pmd, unsigned long vaddr, unsigned long end) in kasan_populate_pte() argument 39 ptep = pte_offset_kernel(pmd, vaddr); in kasan_populate_pte() 47 } while (ptep++, vaddr += PAGE_SIZE, vaddr != end); in kasan_populate_pte() 50 static void __init kasan_populate_pmd(pud_t *pud, unsigned long vaddr, unsigned long end) in kasan_populate_pmd() argument 61 pmdp = pmd_offset(pud, vaddr); in kasan_populate_pmd() 64 next = pmd_addr_end(vaddr, end); in kasan_populate_pmd() 66 if (pmd_none(pmdp_get(pmdp)) && IS_ALIGNED(vaddr, PMD_SIZE) && in kasan_populate_pmd() 67 (next - vaddr) >= PMD_SIZE) { in kasan_populate_pmd() 76 kasan_populate_pte(pmdp, vaddr, next); in kasan_populate_pmd() 77 } while (pmdp++, vaddr = next, vaddr != end); in kasan_populate_pmd() [all …]
|
| /linux/arch/parisc/kernel/ |
| H A D | pci-dma.c | 77 unsigned long vaddr, in map_pte_uncached() argument 81 unsigned long orig_vaddr = vaddr; in map_pte_uncached() 83 vaddr &= ~PMD_MASK; in map_pte_uncached() 84 end = vaddr + size; in map_pte_uncached() 96 vaddr += PAGE_SIZE; in map_pte_uncached() 100 } while (vaddr < end); in map_pte_uncached() 104 static inline int map_pmd_uncached(pmd_t * pmd, unsigned long vaddr, in map_pmd_uncached() argument 108 unsigned long orig_vaddr = vaddr; in map_pmd_uncached() 110 vaddr &= ~PGDIR_MASK; in map_pmd_uncached() 111 end = vaddr + size; in map_pmd_uncached() [all …]
|
| /linux/arch/arm/mm/ |
| H A D | cache-xsc3l2.c | 88 unsigned long vaddr; in xsc3_l2_inv_range() local 95 vaddr = -1; /* to force the first mapping */ in xsc3_l2_inv_range() 101 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); in xsc3_l2_inv_range() 102 xsc3_l2_clean_mva(vaddr); in xsc3_l2_inv_range() 103 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() 111 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range() 112 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() 120 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range() 121 xsc3_l2_clean_mva(vaddr); in xsc3_l2_inv_range() 122 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() [all …]
|
| /linux/arch/s390/include/asm/ |
| H A D | idals.h | 33 static inline bool idal_is_needed(void *vaddr, unsigned int length) in idal_is_needed() argument 35 dma64_t paddr = virt_to_dma64(vaddr); in idal_is_needed() 43 static inline unsigned int idal_nr_words(void *vaddr, unsigned int length) in idal_nr_words() argument 47 cidaw = (unsigned long)vaddr & (IDA_BLOCK_SIZE - 1); in idal_nr_words() 56 static inline unsigned int idal_2k_nr_words(void *vaddr, unsigned int length) in idal_2k_nr_words() argument 60 cidaw = (unsigned long)vaddr & (IDA_2K_BLOCK_SIZE - 1); in idal_2k_nr_words() 69 static inline dma64_t *idal_create_words(dma64_t *idaws, void *vaddr, unsigned int length) in idal_create_words() argument 71 dma64_t paddr = virt_to_dma64(vaddr); in idal_create_words() 75 cidaw = idal_nr_words(vaddr, length); in idal_create_words() 88 static inline int set_normalized_cda(struct ccw1 *ccw, void *vaddr) in set_normalized_cda() argument 136 void *vaddr; idal_buffer_alloc() local 172 void *vaddr; idal_buffer_free() local 274 void *vaddr; idal_buffer_set_cda() local 298 void *vaddr; idal_buffer_to_user() local 320 void *vaddr; idal_buffer_from_user() local [all...] |
| /linux/arch/x86/mm/ |
| H A D | mem_encrypt_amd.c | 156 static void __init __sme_early_map_unmap_mem(void *vaddr, unsigned long size, in __sme_early_map_unmap_mem() argument 159 unsigned long paddr = (unsigned long)vaddr - __PAGE_OFFSET; in __sme_early_map_unmap_mem() 167 __early_make_pgtable((unsigned long)vaddr, pmd); in __sme_early_map_unmap_mem() 169 vaddr += PMD_SIZE; in __sme_early_map_unmap_mem() 256 static void enc_dec_hypercall(unsigned long vaddr, unsigned long size, bool enc) in enc_dec_hypercall() argument 259 unsigned long vaddr_end = vaddr + size; in enc_dec_hypercall() 261 while (vaddr < vaddr_end) { in enc_dec_hypercall() 266 kpte = lookup_address(vaddr, &level); in enc_dec_hypercall() 281 vaddr = (vaddr & pmask) + psize; in enc_dec_hypercall() 286 static int amd_enc_status_change_prepare(unsigned long vaddr, int npages, bool enc) in amd_enc_status_change_prepare() argument [all …]
|
| H A D | kaslr.c | 82 unsigned long vaddr_start, vaddr; in kernel_randomize_memory() local 89 vaddr = vaddr_start; in kernel_randomize_memory() 151 vaddr += entropy; in kernel_randomize_memory() 152 *kaslr_regions[i].base = vaddr; in kernel_randomize_memory() 155 vaddr += get_padding(&kaslr_regions[i]); in kernel_randomize_memory() 163 *kaslr_regions[i].end = __pa_nodebug(vaddr - 1); in kernel_randomize_memory() 166 vaddr = round_up(vaddr + 1, PUD_SIZE); in kernel_randomize_memory() 175 unsigned long paddr, vaddr; in init_trampoline_kaslr() local 188 vaddr = (unsigned long)__va(paddr); in init_trampoline_kaslr() 189 pgd = pgd_offset_k(vaddr); in init_trampoline_kaslr() [all …]
|
| H A D | pgtable_32.c | 27 void set_pte_vaddr(unsigned long vaddr, pte_t pteval) in set_pte_vaddr() argument 35 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pte_vaddr() 40 p4d = p4d_offset(pgd, vaddr); in set_pte_vaddr() 45 pud = pud_offset(p4d, vaddr); in set_pte_vaddr() 50 pmd = pmd_offset(pud, vaddr); in set_pte_vaddr() 55 pte = pte_offset_kernel(pmd, vaddr); in set_pte_vaddr() 57 set_pte_at(&init_mm, vaddr, pte, pteval); in set_pte_vaddr() 59 pte_clear(&init_mm, vaddr, pte); in set_pte_vaddr() 65 flush_tlb_one_kernel(vaddr); in set_pte_vaddr()
|
| /linux/arch/parisc/mm/ |
| H A D | fixmap.c | 15 unsigned long vaddr = __fix_to_virt(idx); in set_fixmap() local 16 pgd_t *pgd = pgd_offset_k(vaddr); in set_fixmap() 17 p4d_t *p4d = p4d_offset(pgd, vaddr); in set_fixmap() 18 pud_t *pud = pud_offset(p4d, vaddr); in set_fixmap() 19 pmd_t *pmd = pmd_offset(pud, vaddr); in set_fixmap() 22 pte = pte_offset_kernel(pmd, vaddr); in set_fixmap() 23 set_pte_at(&init_mm, vaddr, pte, __mk_pte(phys, PAGE_KERNEL_RWX)); in set_fixmap() 24 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); in set_fixmap() 29 unsigned long vaddr = __fix_to_virt(idx); in clear_fixmap() local 30 pte_t *pte = virt_to_kpte(vaddr); in clear_fixmap() [all …]
|
| /linux/mm/ |
| H A D | highmem.c | 165 struct page *__kmap_to_page(void *vaddr) in __kmap_to_page() argument 167 unsigned long base = (unsigned long) vaddr & PAGE_MASK; in __kmap_to_page() 169 unsigned long addr = (unsigned long)vaddr; in __kmap_to_page() 193 return virt_to_page(vaddr); in __kmap_to_page() 248 unsigned long vaddr; in map_new_virtual() 290 vaddr = PKMAP_ADDR(last_pkmap_nr); in map_new_virtual() 291 set_pte_at(&init_mm, vaddr, in map_new_virtual() 295 set_page_address(page, (void *)vaddr); in map_new_virtual() 297 return vaddr; in map_new_virtual() 310 unsigned long vaddr; in kmap_high() 247 unsigned long vaddr; map_new_virtual() local 309 unsigned long vaddr; kmap_high() local 339 unsigned long vaddr, flags; kmap_high_get() local 361 unsigned long vaddr; kunmap_high() local 495 arch_kmap_local_post_map(vaddr,pteval) global() argument 499 arch_kmap_local_pre_unmap(vaddr) global() argument 503 arch_kmap_local_post_unmap(vaddr) global() argument 507 arch_kmap_local_unmap_idx(idx,vaddr) global() argument 518 arch_kmap_local_set_pte(mm,vaddr,ptep,ptev) global() argument 523 kmap_high_unmap_local(unsigned long vaddr) kmap_high_unmap_local() argument 536 kmap_get_pte(unsigned long vaddr,int idx) kmap_get_pte() argument 552 unsigned long vaddr; __kmap_local_pfn_prot() local 596 kunmap_local_indexed(const void * vaddr) kunmap_local_indexed() argument [all...] |
| /linux/arch/m68k/sun3x/ |
| H A D | dvma.c | 79 unsigned long vaddr, int len) in dvma_map_cpu() argument 88 vaddr &= PAGE_MASK; in dvma_map_cpu() 90 end = PAGE_ALIGN(vaddr + len); in dvma_map_cpu() 92 pr_debug("dvma: mapping kern %08lx to virt %08lx\n", kaddr, vaddr); in dvma_map_cpu() 93 pgd = pgd_offset_k(vaddr); in dvma_map_cpu() 94 p4d = p4d_offset(pgd, vaddr); in dvma_map_cpu() 95 pud = pud_offset(p4d, vaddr); in dvma_map_cpu() 101 if((pmd = pmd_alloc(&init_mm, pud, vaddr)) == NULL) { in dvma_map_cpu() 106 if((end & PGDIR_MASK) > (vaddr & PGDIR_MASK)) in dvma_map_cpu() 107 end2 = (vaddr + (PGDIR_SIZE-1)) & PGDIR_MASK; in dvma_map_cpu() [all …]
|
| /linux/drivers/media/common/videobuf2/ |
| H A D | videobuf2-vmalloc.c | 26 void *vaddr; member 47 buf->vaddr = vmalloc_user(buf->size); in vb2_vmalloc_alloc() 48 if (!buf->vaddr) { in vb2_vmalloc_alloc() 68 vfree(buf->vaddr); in vb2_vmalloc_put() 74 unsigned long vaddr, unsigned long size) in vb2_vmalloc_get_userptr() argument 86 offset = vaddr & ~PAGE_MASK; in vb2_vmalloc_get_userptr() 88 vec = vb2_create_framevec(vaddr, size, in vb2_vmalloc_get_userptr() 107 buf->vaddr = (__force void *) in vb2_vmalloc_get_userptr() 110 buf->vaddr = vm_map_ram(frame_vector_pages(vec), n_pages, -1); in vb2_vmalloc_get_userptr() 113 if (!buf->vaddr) in vb2_vmalloc_get_userptr() 129 unsigned long vaddr = (unsigned long)buf->vaddr & PAGE_MASK; vb2_vmalloc_put_userptr() local 220 void *vaddr = buf->vaddr; vb2_vmalloc_dmabuf_ops_attach() local [all...] |
| /linux/drivers/net/ethernet/freescale/fman/ |
| H A D | fman_muram.c | 19 unsigned long vaddr) in fman_muram_vbase_to_offset() argument 21 return vaddr - (unsigned long)muram->vbase; in fman_muram_vbase_to_offset() 40 void __iomem *vaddr; in fman_muram_init() local 53 vaddr = ioremap(base, size); in fman_muram_init() 54 if (!vaddr) { in fman_muram_init() 59 ret = gen_pool_add_virt(muram->pool, (unsigned long)vaddr, in fman_muram_init() 63 iounmap(vaddr); in fman_muram_init() 67 memset_io(vaddr, 0, (int)size); in fman_muram_init() 69 muram->vbase = vaddr; in fman_muram_init() 106 unsigned long vaddr; in fman_muram_alloc() local [all...] |
| /linux/drivers/media/platform/chips-media/wave5/ |
| H A D | wave5-vdi.c | 18 if (!vpu_dev->common_mem.vaddr) { in wave5_vdi_allocate_common_memory() 34 &vpu_dev->common_mem.daddr, vpu_dev->common_mem.size, vpu_dev->common_mem.vaddr); in wave5_vdi_allocate_common_memory() 90 if (!vb || !vb->vaddr) { in wave5_vdi_clear_memory() 95 memset(vb->vaddr, 0, vb->size); in wave5_vdi_clear_memory() 102 if (!vb || !vb->vaddr) { in wave5_vdi_write_memory() 112 memcpy(vb->vaddr + offset, data, len); in wave5_vdi_write_memory() 119 void *vaddr; in wave5_vdi_allocate_dma_memory() local 127 vaddr = dma_alloc_coherent(vpu_dev->dev, vb->size, &daddr, GFP_KERNEL); in wave5_vdi_allocate_dma_memory() 128 if (!vaddr) in wave5_vdi_allocate_dma_memory() 130 vb->vaddr = vaddr; in wave5_vdi_allocate_dma_memory() [all …]
|
| /linux/arch/m68k/sun3/ |
| H A D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument 35 if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { in dvma_page() 36 sun3_put_pte(vaddr, pte); in dvma_page() 37 ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; in dvma_page() 40 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page() 49 unsigned long vaddr; in dvma_map_iommu() local 51 vaddr = dvma_btov(baddr); in dvma_map_iommu() 53 end = vaddr + len; in dvma_map_iommu() 55 while(vaddr < end) { in dvma_map_iommu() 56 dvma_page(kaddr, vaddr); in dvma_map_iommu() [all …]
|
| /linux/arch/sh/mm/ |
| H A D | kmap.c | 21 unsigned long vaddr; in kmap_coherent_init() local 24 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); in kmap_coherent_init() 25 kmap_coherent_pte = virt_to_kpte(vaddr); in kmap_coherent_init() 32 unsigned long vaddr; in kmap_coherent() local 43 vaddr = __fix_to_virt(idx); in kmap_coherent() 48 return (void *)vaddr; in kmap_coherent() 54 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; in kunmap_coherent() local 55 enum fixed_addresses idx = __virt_to_fix(vaddr); in kunmap_coherent() 58 __flush_purge_region((void *)vaddr, PAGE_SIZE); in kunmap_coherent() 60 pte_clear(&init_mm, vaddr, kmap_coherent_pte - idx); in kunmap_coherent() [all …]
|
| /linux/arch/mips/mm/ |
| H A D | pgtable-32.c | 43 unsigned long vaddr; in pagetable_init() local 62 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1); in pagetable_init() 63 fixrange_init(vaddr & PMD_MASK, vaddr + FIXADDR_SIZE, pgd_base); in pagetable_init() 69 vaddr = PKMAP_BASE; in pagetable_init() 70 fixrange_init(vaddr & PMD_MASK, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in pagetable_init() 72 pgd = swapper_pg_dir + pgd_index(vaddr); in pagetable_init() 73 p4d = p4d_offset(pgd, vaddr); in pagetable_init() 74 pud = pud_offset(p4d, vaddr); in pagetable_init() 75 pmd = pmd_offset(pud, vaddr); in pagetable_init() 76 pte = pte_offset_kernel(pmd, vaddr); in pagetable_init()
|
| /linux/drivers/misc/sgi-gru/ |
| H A D | grufault.c | 48 struct vm_area_struct *gru_find_vma(unsigned long vaddr) in gru_find_vma() argument 52 vma = vma_lookup(current->mm, vaddr); in gru_find_vma() 66 static struct gru_thread_state *gru_find_lock_gts(unsigned long vaddr) in gru_find_lock_gts() argument 73 vma = gru_find_vma(vaddr); in gru_find_lock_gts() 75 gts = gru_find_thread_state(vma, TSID(vaddr, vma)); in gru_find_lock_gts() 83 static struct gru_thread_state *gru_alloc_locked_gts(unsigned long vaddr) in gru_alloc_locked_gts() argument 90 vma = gru_find_vma(vaddr); in gru_alloc_locked_gts() 94 gts = gru_alloc_thread_state(vma, TSID(vaddr, vma)); in gru_alloc_locked_gts() 178 unsigned long vaddr, int write, in non_atomic_pte_lookup() argument 188 if (get_user_pages(vaddr, 1, write ? FOLL_WRITE : 0, &page) <= 0) in non_atomic_pte_lookup() [all …]
|
| /linux/arch/xtensa/mm/ |
| H A D | cache.c | 60 unsigned long vaddr) in kmap_invalidate_coherent() argument 62 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { in kmap_invalidate_coherent() 82 unsigned long vaddr, unsigned long *paddr) in coherent_kvaddr() argument 85 return (void *)(base + (vaddr & DCACHE_ALIAS_MASK)); in coherent_kvaddr() 88 void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument 92 void *kvaddr = coherent_kvaddr(page, TLBTEMP_BASE_1, vaddr, &paddr); in clear_user_highpage() 95 kmap_invalidate_coherent(page, vaddr); in clear_user_highpage() 103 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument 107 void *dst_vaddr = coherent_kvaddr(dst, TLBTEMP_BASE_1, vaddr, in copy_user_highpage() 109 void *src_vaddr = coherent_kvaddr(src, TLBTEMP_BASE_2, vaddr, in copy_user_highpage() [all …]
|
| /linux/arch/x86/kernel/ |
| H A D | uprobes.c | 607 *sr = utask->vaddr + auprobe->defparam.ilen; in riprel_pre_xol() 636 unsigned long vaddr; member 639 static bool is_reachable_by_call(unsigned long vtramp, unsigned long vaddr) in is_reachable_by_call() argument 641 long delta = (long)(vaddr + 5 - vtramp); in is_reachable_by_call() 646 static unsigned long find_nearest_trampoline(unsigned long vaddr) in find_nearest_trampoline() argument 654 unsigned long call_end = vaddr + 5; in find_nearest_trampoline() 685 static struct uprobe_trampoline *create_uprobe_trampoline(unsigned long vaddr) in create_uprobe_trampoline() argument 695 vaddr = find_nearest_trampoline(vaddr); in create_uprobe_trampoline() 696 if (IS_ERR_VALUE(vaddr)) in create_uprobe_trampoline() 714 get_uprobe_trampoline(unsigned long vaddr,bool * new) get_uprobe_trampoline() argument 929 verify_insn(struct page * page,unsigned long vaddr,uprobe_opcode_t * new_opcode,int nbytes,void * data) verify_insn() argument 966 int3_update(struct arch_uprobe * auprobe,struct vm_area_struct * vma,unsigned long vaddr,char * insn,bool optimize) int3_update() argument 1018 swbp_optimize(struct arch_uprobe * auprobe,struct vm_area_struct * vma,unsigned long vaddr,unsigned long tramp) swbp_optimize() argument 1028 swbp_unoptimize(struct arch_uprobe * auprobe,struct vm_area_struct * vma,unsigned long vaddr) swbp_unoptimize() argument 1033 copy_from_vaddr(struct mm_struct * mm,unsigned long vaddr,void * dst,int len) copy_from_vaddr() argument 1047 __is_optimized(uprobe_opcode_t * insn,unsigned long vaddr) __is_optimized() argument 1059 is_optimized(struct mm_struct * mm,unsigned long vaddr) is_optimized() argument 1077 set_swbp(struct arch_uprobe * auprobe,struct vm_area_struct * vma,unsigned long vaddr) set_swbp() argument 1095 set_orig_insn(struct arch_uprobe * auprobe,struct vm_area_struct * vma,unsigned long vaddr) set_orig_insn() argument 1112 __arch_uprobe_optimize(struct arch_uprobe * auprobe,struct mm_struct * mm,unsigned long vaddr) __arch_uprobe_optimize() argument 1131 arch_uprobe_optimize(struct arch_uprobe * auprobe,unsigned long vaddr) arch_uprobe_optimize() argument 1161 can_optimize(struct insn * insn,unsigned long vaddr) can_optimize() argument 1185 can_optimize(struct insn * insn,unsigned long vaddr) can_optimize() argument 1831 unsigned long vaddr; arch_uprobe_get_xol_area() local [all...] |
| /linux/arch/csky/mm/ |
| H A D | tcm.c | 29 unsigned long vaddr, paddr; local 42 vaddr = __fix_to_virt(FIX_TCM - i); 45 pte_offset_kernel((pmd_t *)pgd_offset_k(vaddr), vaddr); 49 flush_tlb_one(vaddr); 61 vaddr = __fix_to_virt(FIX_TCM - CONFIG_ITCM_NR_PAGES - i); 64 pte_offset_kernel((pmd_t *) pgd_offset_k(vaddr), vaddr); 68 flush_tlb_one(vaddr); 111 unsigned long vaddr; local 116 vaddr = gen_pool_alloc(tcm_pool, len); 117 if (!vaddr) [all …]
|
| /linux/drivers/scsi/ |
| H A D | hpsa.h | 172 void __iomem *vaddr; member 423 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command() 424 (void) readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command() 430 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_no_read() 436 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_ioaccel2() 448 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 449 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 453 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 454 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 465 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() [all …]
|
| /linux/arch/nios2/mm/ |
| H A D | dma-mapping.c | 24 void *vaddr = phys_to_virt(paddr); in arch_sync_dma_for_device() local 28 invalidate_dcache_range((unsigned long)vaddr, in arch_sync_dma_for_device() 29 (unsigned long)(vaddr + size)); in arch_sync_dma_for_device() 37 flush_dcache_range((unsigned long)vaddr, in arch_sync_dma_for_device() 38 (unsigned long)(vaddr + size)); in arch_sync_dma_for_device() 48 void *vaddr = phys_to_virt(paddr); in arch_sync_dma_for_cpu() local 53 invalidate_dcache_range((unsigned long)vaddr, in arch_sync_dma_for_cpu() 54 (unsigned long)(vaddr + size)); in arch_sync_dma_for_cpu()
|
| /linux/include/linux/ |
| H A D | iosys-map.h | 112 void *vaddr; member 123 .vaddr = (vaddr_), \ 183 static inline void iosys_map_set_vaddr(struct iosys_map *map, void *vaddr) in iosys_map_set_vaddr() argument 185 map->vaddr = vaddr; in iosys_map_set_vaddr() 222 return lhs->vaddr == rhs->vaddr; in iosys_map_is_equal() 239 return !map->vaddr; in iosys_map_is_null() 287 memcpy(dst->vaddr + dst_offset, src, len); in iosys_map_memcpy_to() 307 memcpy(dst, src->vaddr + src_offset, len); in iosys_map_memcpy_from() 323 map->vaddr += incr; in iosys_map_incr() 342 memset(dst->vaddr + offset, value, len); in iosys_map_memset() [all …]
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | shmem_utils.c | 63 void *vaddr; in shmem_pin_map() local 77 vaddr = vmap(pages, n_pages, VM_MAP_PUT_PAGES, PAGE_KERNEL); in shmem_pin_map() 78 if (!vaddr) in shmem_pin_map() 81 return vaddr; in shmem_pin_map() 105 void *vaddr; in __shmem_rw() local 112 vaddr = kmap_local_page(page); in __shmem_rw() 114 memcpy(vaddr + offset_in_page(off), ptr, this); in __shmem_rw() 117 memcpy(ptr, vaddr + offset_in_page(off), this); in __shmem_rw() 120 kunmap_local(vaddr); in __shmem_rw() 140 void *vaddr; in shmem_read_to_iosys_map() local [all...] |