| /linux/drivers/iommu/iommufd/ |
| H A D | iova_bitmap.c | 103 struct iova_bitmap_map mapped; member 133 return (iova >> bitmap->mapped.pgshift) / in iova_bitmap_offset_to_index() 143 unsigned long pgshift = bitmap->mapped.pgshift; in iova_bitmap_index_to_offset() 167 struct iova_bitmap_map *mapped = &bitmap->mapped; in iova_bitmap_get() local 196 FOLL_WRITE, mapped->pages); in iova_bitmap_get() 200 mapped->npages = (unsigned long)ret; in iova_bitmap_get() 202 mapped->iova = iova_bitmap_mapped_iova(bitmap); in iova_bitmap_get() 209 mapped->pgoff = offset_in_page(addr); in iova_bitmap_get() 210 mapped->length = iova_bitmap_mapped_length(bitmap); in iova_bitmap_get() 221 struct iova_bitmap_map *mapped = &bitmap->mapped; in iova_bitmap_put() local [all …]
|
| /linux/drivers/iommu/ |
| H A D | io-pgtable-arm-selftests.c | 60 size_t size, mapped; in arm_lpae_run_tests() local 94 GFP_KERNEL, &mapped)) in arm_lpae_run_tests() 100 GFP_KERNEL, &mapped)) in arm_lpae_run_tests() 122 IOMMU_WRITE, GFP_KERNEL, &mapped)) in arm_lpae_run_tests() 135 mapped = 0; in arm_lpae_run_tests() 141 GFP_KERNEL, &mapped)) in arm_lpae_run_tests() 143 if (mapped != size) in arm_lpae_run_tests()
|
| /linux/drivers/gpu/drm/sti/ |
| H A D | NOTES | 39 These IP are mapped to the DRM objects as following: 40 - The CRTCs are mapped to the Compositor Main and Aux Mixers 41 - The Framebuffers and planes are mapped to the Compositor GDP (non video 43 - The Cursor is mapped to the Compositor Cursor 44 - The Encoders are mapped to the TVOut 45 - The Bridges/Connectors are mapped to the HDMI / DVO / HD Analog / SD analog
|
| /linux/Documentation/userspace-api/media/dvb/ |
| H A D | dmx-munmap.rst | 31 Address of the mapped buffer as returned by the 35 Length of the mapped buffer. This must be the same value as given to 41 Unmaps a previously with the :c:func:`mmap()` function mapped 52 mapped yet.
|
| H A D | dmx-reqbufs.rst | 36 This ioctl is used to initiate a memory mapped or DMABUF based demux I/O. 38 Memory mapped buffers are located in device memory and must be allocated 39 with this ioctl before they can be mapped into the application's address 63 buffers, however this cannot succeed when any buffers are still mapped.
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | pagemap.rst | 12 physical frame each virtual page is mapped to. It contains one 64-bit 21 * Bit 56 page exclusively mapped (since 4.2) 38 precisely which pages are mapped (or in swap) and comparing mapped 41 Traditionally, bit 56 indicates that a page is mapped exactly once and bit 42 56 is clear when a page is mapped multiple times, even when mapped in the 46 mapped in the same process, even if the page is mapped multiple times in that 48 is *maybe* mapped in a different process. In some cases, a large allocation 49 might be treated as "maybe mapped by multiple processes" even though this 53 determine which areas of memory are actually mapped and llseek to 57 times each page is mapped, indexed by PFN. Some kernel configurations do [all …]
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | func-munmap.rst | 29 Address of the mapped buffer as returned by the 33 Length of the mapped buffer. This must be the same value as given to 43 Unmaps a previously with the :c:func:`mmap()` function mapped 54 mapped yet.
|
| /linux/mm/kmsan/ |
| H A D | shadow.c | 222 int nr, mapped, err = 0; in kmsan_vmap_pages_range_noflush() local 248 mapped = __vmap_pages_range_noflush(shadow_start, shadow_end, prot, in kmsan_vmap_pages_range_noflush() 251 if (mapped) { in kmsan_vmap_pages_range_noflush() 252 err = mapped; in kmsan_vmap_pages_range_noflush() 256 mapped = __vmap_pages_range_noflush(origin_start, origin_end, prot, in kmsan_vmap_pages_range_noflush() 259 if (mapped) { in kmsan_vmap_pages_range_noflush() 260 err = mapped; in kmsan_vmap_pages_range_noflush()
|
| H A D | hooks.c | 156 int nr, err = 0, clean = 0, mapped; in kmsan_ioremap_page_range() local 170 mapped = __vmap_pages_range_noflush( in kmsan_ioremap_page_range() 174 if (mapped) { in kmsan_ioremap_page_range() 175 err = mapped; in kmsan_ioremap_page_range() 179 mapped = __vmap_pages_range_noflush( in kmsan_ioremap_page_range() 183 if (mapped) { in kmsan_ioremap_page_range() 187 err = mapped; in kmsan_ioremap_page_range()
|
| /linux/fs/jfs/ |
| H A D | ioctl.c | 45 long mapped=0; in jfs_map_ext2() local 50 mapped |= jfs_map[index].jfs_flag; in jfs_map_ext2() 53 mapped |= jfs_map[index].ext2_flag; in jfs_map_ext2() 57 return mapped; in jfs_map_ext2()
|
| /linux/Documentation/devicetree/bindings/display/ |
| H A D | st,stih4xx.txt | 6 - reg: Physical base address of the IP registers and length of memory mapped region. 14 - reg: Physical base address of the IP registers and length of memory mapped region. 32 - reg: Physical base address of the IP registers and length of memory mapped region. 48 - reg: Physical base address of the IP registers and length of memory mapped region. 49 - reg-names: names of the mapped memory regions listed in regs property in 60 - reg: Physical base address of the IP registers and length of memory mapped region. 61 - reg-names: names of the mapped memory regions listed in regs property in 76 - reg: Physical base address of the IP registers and length of memory mapped region. 77 - reg-names: names of the mapped memory regions listed in regs property in 89 - reg: Physical base address of the IP registers and length of memory mapped region. [all …]
|
| /linux/Documentation/arch/loongarch/ |
| H A D | introduction.rst | 293 LoongArch supports direct-mapped virtual memory and page-mapped virtual memory. 295 Direct-mapped virtual memory is configured by CSR.DMWn (n=0~3), it has a simple 300 Page-mapped virtual memory has arbitrary relationship between VA and PA, which 309 ``UVRANGE`` ``0x00000000 - 0x7FFFFFFF`` Page-mapped, Cached, PLV0~3 310 ``KPRANGE0`` ``0x80000000 - 0x9FFFFFFF`` Direct-mapped, Uncached, PLV0 311 ``KPRANGE1`` ``0xA0000000 - 0xBFFFFFFF`` Direct-mapped, Cached, PLV0 312 ``KVRANGE`` ``0xC0000000 - 0xFFFFFFFF`` Page-mapped, Cached, PLV0 315 User mode (PLV3) can only access UVRANGE. For direct-mapped KPRANGE0 and 317 direct-mapped VA of 0x00001000 is 0x80001000, and the cached direct-mapped 325 ``XUVRANGE`` ``0x0000000000000000 - Page-mapped, Cached, PLV0~3 [all …]
|
| /linux/Documentation/scsi/ |
| H A D | g_NCR5380.rst | 16 memory mapped modes. 38 base=xx[,...] the port or base address(es) (for port or memory mapped, resp.) 55 mapped, resp.) 71 E.g. a port mapped NCR5380 board, driver to probe for IRQ:: 79 E.g. a memory mapped NCR53C400 board with no IRQ::
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_gem_vma.c | 274 if (!msm_vma->mapped) in msm_gem_vma_unmap() 294 msm_vma->mapped = false; in msm_gem_vma_unmap() 308 if (msm_vma->mapped) in msm_gem_vma_map() 311 msm_vma->mapped = true; in msm_gem_vma_map() 339 msm_vma->mapped = false; in msm_gem_vma_map() 350 GEM_WARN_ON(msm_vma->mapped); in msm_gem_vma_close() 407 vma->mapped = false; in msm_gem_vma_new() 534 to_msm_vma(vma)->mapped = true; in msm_gem_vm_sm_step_map() 547 bool mapped = to_msm_vma(orig_vma)->mapped; in msm_gem_vm_sm_step_remap() local 554 if (mapped) { in msm_gem_vm_sm_step_remap() [all …]
|
| H A D | msm_iommu.c | 143 size_t mapped = 0; in msm_iommu_pagetable_map_prr() local 147 ret = ops->map_pages(ops, addr, phys, size, 1, prot, GFP_KERNEL, &mapped); in msm_iommu_pagetable_map_prr() 152 addr += mapped; in msm_iommu_pagetable_map_prr() 153 len -= mapped; in msm_iommu_pagetable_map_prr() 195 size_t pgsize, count, mapped = 0; in msm_iommu_pagetable_map() local 201 prot, GFP_KERNEL, &mapped); in msm_iommu_pagetable_map() 206 phys += mapped; in msm_iommu_pagetable_map() 207 addr += mapped; in msm_iommu_pagetable_map() 208 size -= mapped; in msm_iommu_pagetable_map() 209 len -= mapped; in msm_iommu_pagetable_map()
|
| /linux/Documentation/mm/ |
| H A D | arch_pgtable_helpers.rst | 21 | pte_present | Tests a valid mapped PTE | 63 | pte_mknotpresent | Invalidates a mapped PTE | 85 | pmd_bad | Tests a non-table mapped PMD | 87 | pmd_leaf | Tests a leaf mapped PMD | 158 | pud_bad | Tests a non-table mapped PUD | 160 | pud_leaf | Tests a leaf mapped PUD | 164 | pud_present | Tests a valid mapped PUD |
|
| H A D | vmalloced-kernel-stacks.rst | 25 Virtually mapped kernel stacks with guard pages cause kernel stack 30 support for virtually mapped stacks with guard pages. This feature 61 mapped task stacks. This option depends on HAVE_ARCH_VMAP_STACK. 63 - Enable this if you want the use virtually-mapped kernel stacks 88 pages are mapped into contiguous kernel virtual space with PAGE_KERNEL 114 virtually mapped kernel stacks are enabled.
|
| /linux/drivers/mtd/maps/ |
| H A D | Kconfig | 17 ROM driver code to communicate with chips which are mapped 43 are mapped on your particular target board. Refer to the 74 and RAM driver code to communicate with chips which are mapped 140 tristate "CFI Flash device mapped on AMD SC520 CDP" 148 tristate "CFI Flash device mapped on AMD NetSc520" 156 tristate "JEDEC Flash device mapped on Technologic Systems TS-5500" 174 tristate "CFI Flash device mapped on Arcom SBC-GXx boards" 185 tristate "CFI Flash device mapped on Intel XScale PXA2xx based boards" 191 tristate "Flash device mapped with DOCCS on NatSemi SCx200" 194 Enable support for a flash chip mapped using the DOCCS signal on a [all …]
|
| /linux/fs/nilfs2/ |
| H A D | file.c | 71 goto mapped; in nilfs_page_mkwrite() 87 goto mapped; in nilfs_page_mkwrite() 109 mapped: in nilfs_page_mkwrite()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | user_exp_rcv.c | 137 bool mapped) in unpin_rcv_pages() argument 143 if (mapped) { in unpin_rcv_pages() 152 hfi1_release_user_pages(mm, pages, npages, mapped); in unpin_rcv_pages() 253 tididx = 0, mapped, mapped_pages = 0; in hfi1_user_exp_rcv_setup() local 335 tidlist, &tididx, &mapped); in hfi1_user_exp_rcv_setup() 350 mapped_pages += mapped; in hfi1_user_exp_rcv_setup() 380 &tididx, &mapped); in hfi1_user_exp_rcv_setup() 391 mapped_pages += mapped; in hfi1_user_exp_rcv_setup() 666 int mapped = 0; in program_rcvarray() local 707 mapped += npages; in program_rcvarray() [all …]
|
| /linux/drivers/sh/intc/ |
| H A D | virq.c | 149 struct intc_map_entry *mapped; in intc_subgroup_init_one() local 154 mapped = radix_tree_lookup(&d->tree, subgroup->parent_id); in intc_subgroup_init_one() 155 if (!mapped) { in intc_subgroup_init_one() 160 pirq = mapped - intc_irq_xlate; in intc_subgroup_init_one()
|
| /linux/arch/powerpc/mm/book3s64/ |
| H A D | iommu_api.c | 31 atomic64_t mapped; member 125 atomic64_set(&mem->mapped, 1); in mm_iommu_do_alloc() 267 if (atomic64_cmpxchg(&mem->mapped, 1, 0) != 1) { in mm_iommu_put() 385 if (atomic64_inc_not_zero(&mem->mapped)) in mm_iommu_mapped_inc() 395 atomic64_add_unless(&mem->mapped, -1, 1); in mm_iommu_mapped_dec()
|
| /linux/drivers/dma-buf/heaps/ |
| H A D | system_heap.c | 38 bool mapped; member 92 a->mapped = false; in system_heap_attach() 128 a->mapped = true; in system_heap_map_dma_buf() 138 a->mapped = false; in system_heap_unmap_dma_buf() 154 if (!a->mapped) in system_heap_dma_buf_begin_cpu_access() 175 if (!a->mapped) in system_heap_dma_buf_end_cpu_access()
|
| /linux/drivers/xen/ |
| H A D | xlate_mmu.c | 74 int mapped; member 129 info->mapped++; in remap_pte_fn() 166 data.mapped = 0; in xen_xlate_remap_gfn_array() 170 return err < 0 ? err : data.mapped; in xen_xlate_remap_gfn_array()
|
| /linux/drivers/block/rnbd/ |
| H A D | README | 11 transport. After being mapped, the remote block devices can be accessed 39 mapped from the server side. After the session to the server machine is 40 established, the mapped device will appear on the client side under 50 When a device is mapped from the client, the server generates the path
|