| /linux/Documentation/mm/ |
| H A D | highmem.rst | 15 at all times. This means the kernel needs to start using temporary mappings of 51 The kernel contains several ways of creating temporary mappings. The following 55 short term mappings. They can be invoked from any context (including 56 interrupts) but the mappings can only be used in the context which acquired 64 These mappings are thread-local and CPU-local, meaning that the mapping 89 mappings, the local mappings are only valid in the context of the caller 94 Most code can be designed to use thread local mappings. User should 99 Nesting kmap_local_page() and kmap_atomic() mappings is allowed to a certain 103 mappings. 110 effects of atomic mappings, i.e. disabling page faults or preemption, or both. [all …]
|
| H A D | hugetlbfs_reserv.rst | 87 of mappings. Location differences are: 89 - For private mappings, the reservation map hangs off the VMA structure. 92 - For shared mappings, the reservation map hangs off the inode. Specifically, 93 inode->i_mapping->private_data. Since shared mappings are always backed 121 One of the big differences between PRIVATE and SHARED mappings is the way 124 - For shared mappings, an entry in the reservation map indicates a reservation 127 - For private mappings, the lack of an entry in the reservation map indicates 133 For private mappings, hugetlb_reserve_pages() creates the reservation map and 138 are needed for the current mapping/segment. For private mappings, this is 139 always the value (to - from). However, for shared mappings it is possible that [all …]
|
| /linux/drivers/gpu/drm/tegra/ |
| H A D | submit.c | 150 xa_lock(&context->mappings); in tegra_drm_mapping_get() 152 mapping = xa_load(&context->mappings, id); in tegra_drm_mapping_get() 156 xa_unlock(&context->mappings); in tegra_drm_mapping_get() 261 struct tegra_drm_used_mapping *mappings; in submit_process_bufs() local 273 mappings = kzalloc_objs(*mappings, args->num_bufs); in submit_process_bufs() 274 if (!mappings) { in submit_process_bufs() 303 mappings[i].mapping = mapping; in submit_process_bufs() 304 mappings[i].flags = buf->flags; in submit_process_bufs() 307 job_data->used_mappings = mappings; in submit_process_bufs() 316 tegra_drm_mapping_put(mappings[i].mapping); in submit_process_bufs() [all …]
|
| H A D | uapi.c | 39 xa_for_each(&context->mappings, id, mapping) in tegra_drm_channel_context_close() 42 xa_destroy(&context->mappings); in tegra_drm_channel_context_close() 144 xa_init_flags(&context->mappings, XA_FLAGS_ALLOC1); in tegra_drm_ioctl_channel_open() 255 err = xa_alloc(&context->mappings, &args->mapping, mapping, XA_LIMIT(1, U32_MAX), in tegra_drm_ioctl_channel_map() 290 mapping = xa_erase(&context->mappings, args->mapping); in tegra_drm_ioctl_channel_unmap()
|
| /linux/drivers/gpu/drm/panfrost/ |
| H A D | panfrost_gem.c | 86 WARN_ON_ONCE(!list_empty(&bo->mappings.list)); in panfrost_gem_free_object() 115 mutex_lock(&bo->mappings.lock); in panfrost_gem_mapping_get() 116 list_for_each_entry(iter, &bo->mappings.list, node) { in panfrost_gem_mapping_get() 123 mutex_unlock(&bo->mappings.lock); in panfrost_gem_mapping_get() 164 list_for_each_entry(mapping, &bo->mappings.list, node) in panfrost_gem_teardown_mappings_locked() 212 mutex_lock(&bo->mappings.lock); in panfrost_gem_open() 214 list_add_tail(&mapping->node, &bo->mappings.list); in panfrost_gem_open() 215 mutex_unlock(&bo->mappings.lock); in panfrost_gem_open() 229 mutex_lock(&bo->mappings.lock); in panfrost_gem_close() 230 list_for_each_entry(iter, &bo->mappings.list, node) { in panfrost_gem_close() [all …]
|
| H A D | panfrost_drv.c | 210 job->mappings = kvmalloc_objs(struct panfrost_gem_mapping *, in panfrost_lookup_bos() 212 if (!job->mappings) in panfrost_lookup_bos() 226 job->mappings[i] = mapping; in panfrost_lookup_bos() 482 mutex_lock(&bo->mappings.lock); in panfrost_ioctl_madvise() 486 first = list_first_entry(&bo->mappings.list, in panfrost_ioctl_madvise() 498 if (!list_is_singular(&bo->mappings.list) || in panfrost_ioctl_madvise() 516 mutex_unlock(&bo->mappings.lock); in panfrost_ioctl_madvise()
|
| /linux/Documentation/arch/arm/ |
| H A D | memory.rst | 62 Machine specific static mappings are also 72 PKMAP_BASE PAGE_OFFSET-1 Permanent kernel mappings 78 placed here using dynamic mappings. 85 00001000 TASK_SIZE-1 User space mappings 86 Per-thread mappings are placed here via 96 Please note that mappings which collide with the above areas may result 103 must set up their own mappings using open() and mmap().
|
| /linux/Documentation/driver-api/ |
| H A D | io-mapping.rst | 44 used with mappings created by io_mapping_create_wc() 46 Temporary mappings are only valid in the context of the caller. The mapping 56 Nested mappings need to be undone in reverse order because the mapping 65 The mappings are released with:: 83 The mappings are released with::
|
| /linux/drivers/soc/aspeed/ |
| H A D | Kconfig | 13 Control LPC firmware cycle mappings through ioctl()s. The driver 43 Control ASPEED P2A VGA MMIO to BMC mappings through ioctl()s. The 44 driver also provides an interface for userspace mappings to a
|
| /linux/drivers/regulator/ |
| H A D | rpi-panel-attiny-regulator.c | 60 static const struct gpio_signal_mappings mappings[NUM_GPIO] = { variable 215 last_val = attiny_get_port_state(state, mappings[off].reg); in attiny_gpio_set() 217 last_val |= mappings[off].mask; in attiny_gpio_set() 219 last_val &= ~mappings[off].mask; in attiny_gpio_set() 221 attiny_set_port_state(state, mappings[off].reg, last_val); in attiny_gpio_set()
|
| /linux/Documentation/driver-api/usb/ |
| H A D | dma.rst | 19 manage dma mappings for existing dma-ready buffers (see below). 27 don't manage dma mappings for URBs. 41 IOMMU to manage the DMA mappings. It can cost MUCH more to set up and 42 tear down the IOMMU mappings with each request than perform the I/O! 64 "streaming" DMA mappings.)
|
| /linux/Documentation/userspace-api/ |
| H A D | mseal.rst | 133 - System mappings: 134 The system mappings are created by the kernel and includes vdso, vvar, 137 Those system mappings are readonly only or execute only, memory sealing can 143 the CONFIG_MSEAL_SYSTEM_MAPPINGS seals all system mappings of this 150 or unmapping system mappings. Known broken software at the time 174 the heap manager's ability to manage the mappings; the outcome is 188 attributes, such as protection bits (RWX). Sealed mappings doesn't mean the
|
| /linux/Documentation/arch/x86/x86_64/ |
| H A D | mm.rst | 38 … | | | | starting offset of kernel mappings. 49 … | | | | starting offset of kernel mappings. 93 from 0.125 PB to 64 PB. All kernel mappings shift down to the -64 PB starting 109 … | | | | starting offset of kernel mappings. 119 … | | | | starting offset of kernel mappings. 168 The mappings are not part of any other kernel PGD and are only available
|
| /linux/drivers/cxl/ |
| H A D | pmem.c | 397 struct nd_mapping_desc mappings[CXL_DECODER_MAX_INTERLEAVE]; in cxl_pmem_region_probe() local 409 memset(&mappings, 0, sizeof(mappings)); in cxl_pmem_region_probe() 478 mappings[i] = (struct nd_mapping_desc) { in cxl_pmem_region_probe() 486 ndr_desc.mapping = mappings; in cxl_pmem_region_probe()
|
| /linux/tools/testing/nvdimm/test/ |
| H A D | ndtest.c | 414 struct nd_mapping_desc mappings[NDTEST_MAX_MAPPING]; in ndtest_create_region() local 422 memset(&mappings, 0, sizeof(mappings)); in ndtest_create_region() 430 ndr_desc->mapping = mappings; in ndtest_create_region() 451 mappings[i].start = region->mapping[i].start; in ndtest_create_region() 452 mappings[i].size = region->mapping[i].size; in ndtest_create_region() 453 mappings[i].position = region->mapping[i].position; in ndtest_create_region() 454 mappings[i].nvdimm = p->config->dimms[ndimm].nvdimm; in ndtest_create_region()
|
| /linux/mm/ |
| H A D | Kconfig.debug | 123 bool "Check for invalid mappings in user page tables" 187 bool "Warn on W+X mappings at boot" 193 Generate a warning if any W+X mappings are found at boot. 196 mappings after applying NX, as such mappings are a security risk. 200 <arch>/mm: Checked W+X mappings: passed, no W+X pages found. 204 <arch>/mm: Checked W+X mappings: failed, <N> W+X pages found. 207 still fine, as W+X mappings are not a security hole in
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-firmware-efi-runtime-map | 10 the same physical to virtual address mappings as the first 11 kernel. The mappings are exported to sysfs so userspace tools
|
| /linux/Documentation/arch/arm64/ |
| H A D | hugetlbpage.rst | 15 1) Block mappings at the pud/pmd level 20 mappings reduce the depth of page table walk needed to translate hugepage
|
| /linux/Documentation/filesystems/iomap/ |
| H A D | design.rst | 24 This layer tries to obtain mappings of each file ranges to storage 28 2. An upper layer that acts upon the space mappings provided by the 31 The iteration can involve mappings of file's logical offset ranges to 62 units (generally memory pages or blocks) and looks up space mappings on 64 largest space mappings that it can create for a given file operation and 69 Larger space mappings improve runtime performance by amortizing the cost 237 * **IOMAP_F_MERGED**: Multiple contiguous block mappings were 287 ``IOMAP_INLINE`` mappings. 295 that should be used to detect stale mappings. 299 Filesystems with completely static mappings need not set this value. [all …]
|
| /linux/Documentation/gpu/rfc/ |
| H A D | i915_vm_bind.rst | 9 specified address space (VM). These mappings (also referred to as persistent 10 mappings) will be persistent across multiple GPU submissions (execbuf calls) 12 mappings during each submission (as required by older execbuf mode). 27 * Multiple Virtual Address (VA) mappings can map to the same physical pages 30 * Support capture of persistent mappings in the dump upon GPU error. 90 path (where required mappings are already bound) submission latency is O(1) 201 execbuf. VM_BIND allows bind/unbind of mappings required for the directly 231 mapped objects. Page table pages are similar to persistent mappings of a
|
| /linux/Documentation/arch/riscv/ |
| H A D | vm-layout.rst | 44 … | | | | starting offset of kernel mappings. 81 | | | | starting offset of kernel mappings. 117 | | | | starting offset of kernel mappings.
|
| /linux/drivers/nvdimm/ |
| H A D | region_devs.c | 314 static DEVICE_ATTR_RO(mappings); 543 u32 mappings, remainder; in align_store() local 557 mappings = max_t(u32, 1, nd_region->ndr_mappings); in align_store() 558 dpa = div_u64_rem(val, mappings, &remainder); in align_store() 966 int mappings; in default_align() local 972 mappings = max_t(u16, 1, nd_region->ndr_mappings); in default_align() 973 div_u64_rem(align, mappings, &remainder); in default_align() 975 align *= mappings; in default_align()
|
| /linux/drivers/gpu/drm/etnaviv/ |
| H A D | etnaviv_mmu.c | 164 list_for_each_entry(free, &context->mappings, mmu_node) { in etnaviv_iommu_find_iova() 287 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem() 312 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem() 440 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_get_suballoc_va()
|
| /linux/drivers/rapidio/devices/ |
| H A D | rio_mport_cdev.c | 156 struct list_head mappings; member 367 list_add_tail(&map->node, &md->mappings); in rio_mport_create_outbound_mapping() 383 list_for_each_entry(map, &md->mappings, node) { in rio_mport_get_outbound_mapping() 457 list_for_each_entry_safe(map, _map, &md->mappings, node) { in rio_mport_obw_free() 884 list_for_each_entry(map, &md->mappings, node) { in rio_dma_transfer() 1089 list_add_tail(&map->node, &md->mappings); in rio_mport_create_dma_mapping() 1136 list_for_each_entry_safe(map, _map, &md->mappings, node) { in rio_mport_free_dma() 1216 list_add_tail(&map->node, &md->mappings); in rio_mport_create_inbound_mapping() 1241 list_for_each_entry(map, &md->mappings, node) { in rio_mport_get_inbound_mapping() 1322 list_for_each_entry_safe(map, _map, &md->mappings, node) { in rio_mport_inbound_free() [all …]
|
| /linux/Documentation/netlabel/ |
| H A D | cipso_ipv4.rst | 46 appropriate for the host. These mappings are defined as part of a CIPSO 55 mappings from the network labels to the corresponding LSM identifiers. The
|