/linux/tools/testing/selftests/mm/ |
H A D | thp_settings.c | 203 unsigned long orders = thp_supported_orders(); in thp_read_settings() local 230 if (!((1 << i) & orders)) { in thp_read_settings() 255 unsigned long orders = thp_supported_orders(); in thp_write_settings() local 281 if (!((1 << i) & orders)) in thp_write_settings() 352 unsigned long orders = 0; in __thp_supported_orders() local 369 orders |= 1UL << i; in __thp_supported_orders() 372 return orders; in __thp_supported_orders()
|
H A D | cow.c | 52 unsigned long orders; in detect_thp_sizes() local 60 orders = 1UL << sz2ord(pmdsize); in detect_thp_sizes() 61 orders |= thp_supported_orders(); in detect_thp_sizes() 63 for (i = 0; orders && count < max; i++) { in detect_thp_sizes() 64 if (!(orders & (1UL << i))) in detect_thp_sizes() 66 orders &= ~(1UL << i); in detect_thp_sizes()
|
/linux/include/linux/ |
H A D | folio_queue.h | 31 u8 orders[PAGEVEC_SIZE]; /* Order of each folio */ member 244 folioq->orders[slot] = __folio_order(folio); in folioq_append() 266 folioq->orders[slot] = __folio_order(folio); in folioq_append_mark() 297 return folioq->orders[slot]; in folioq_folio_order()
|
/linux/drivers/gpu/drm/ttm/tests/ |
H A D | ttm_pool_test.c | 250 pt = &pool->caching[caching].orders[order]; in ttm_pool_alloc_order_caching_match() 279 pt_pool = &pool->caching[pool_caching].orders[order]; in ttm_pool_alloc_caching_mismatch() 280 pt_tt = &pool->caching[tt_caching].orders[order]; in ttm_pool_alloc_caching_mismatch() 313 pt_pool = &pool->caching[caching].orders[order]; in ttm_pool_alloc_order_mismatch() 314 pt_tt = &pool->caching[caching].orders[0]; in ttm_pool_alloc_order_mismatch() 354 pt = &pool->caching[caching].orders[order]; in ttm_pool_free_dma_alloc() 385 pt = &pool->caching[caching].orders[order]; in ttm_pool_free_no_dma_alloc() 405 pt = &pool->caching[caching].orders[order]; in ttm_pool_fini_basic()
|
H A D | ttm_device_test.c | 179 pt = pool->caching[i].orders[j]; in ttm_device_init_pools()
|
/linux/drivers/gpu/drm/ttm/ |
H A D | ttm_pool.c | 292 return &pool->caching[caching].orders[order]; in ttm_pool_select_type() 298 return &pool->caching[caching].orders[order]; in ttm_pool_select_type() 306 return &pool->caching[caching].orders[order]; in ttm_pool_select_type() 581 if (pt != &pool->caching[i].orders[j]) in ttm_pool_init() 619 if (pt != &pool->caching[i].orders[j]) in ttm_pool_fini() 755 ttm_pool_debugfs_orders(pool->caching[i].orders, m); in ttm_pool_debugfs()
|
/linux/include/drm/ttm/ |
H A D | ttm_pool.h | 78 struct ttm_pool_type orders[NR_PAGE_ORDERS]; member
|
/linux/mm/ |
H A D | memory.c | 4183 unsigned long orders) in thp_swap_suitable_orders() argument 4187 order = highest_order(orders); in thp_swap_suitable_orders() 4194 while (orders) { in thp_swap_suitable_orders() 4198 order = next_order(&orders, order); in thp_swap_suitable_orders() 4201 return orders; in thp_swap_suitable_orders() 4207 unsigned long orders; in alloc_swap_folio() local 4236 orders = thp_vma_allowable_orders(vma, vma->vm_flags, in alloc_swap_folio() 4238 orders = thp_vma_suitable_orders(vma, vmf->address, orders); in alloc_swap_folio() 4239 orders = thp_swap_suitable_orders(swp_offset(entry), in alloc_swap_folio() 4240 vmf->address, orders); in alloc_swap_folio() [all …]
|
H A D | huge_memory.c | 104 unsigned long orders) in __thp_vma_allowable_orders() argument 119 orders &= supported_orders; in __thp_vma_allowable_orders() 120 if (!orders) in __thp_vma_allowable_orders() 131 return in_pf ? orders : 0; in __thp_vma_allowable_orders() 149 int order = highest_order(orders); in __thp_vma_allowable_orders() 152 while (orders) { in __thp_vma_allowable_orders() 156 order = next_order(&orders, order); in __thp_vma_allowable_orders() 159 if (!orders) in __thp_vma_allowable_orders() 188 return orders; in __thp_vma_allowable_orders() 191 return orders; in __thp_vma_allowable_orders() [all …]
|
H A D | shmem.c | 1812 unsigned long orders) in shmem_suitable_orders() argument 1820 orders = thp_vma_suitable_orders(vma, vmf->address, orders); in shmem_suitable_orders() 1821 if (!orders) in shmem_suitable_orders() 1826 order = highest_order(orders); in shmem_suitable_orders() 1827 while (orders) { in shmem_suitable_orders() 1841 order = next_order(&orders, order); in shmem_suitable_orders() 1844 return orders; in shmem_suitable_orders() 1849 unsigned long orders) in shmem_suitable_orders() argument 1871 struct mm_struct *fault_mm, unsigned long orders) in shmem_alloc_and_add_folio() argument 1881 orders = 0; in shmem_alloc_and_add_folio() [all …]
|
/linux/tools/memory-model/Documentation/ |
H A D | glossary.txt | 29 a special operation that includes a load and which orders that 117 Fully Ordered: An operation such as smp_mb() that orders all of 120 that orders all of its CPU's prior accesses, itself, and 167 a special operation that includes a store and which orders that
|
H A D | recipes.txt | 232 The smp_store_release() macro orders any prior accesses against the 233 store, while the smp_load_acquire macro orders the load against any 273 smp_store_release(), but the rcu_dereference() macro orders the load only 310 The smp_wmb() macro orders prior stores against later stores, and the 311 smp_rmb() macro orders prior loads against later loads. Therefore, if
|
H A D | ordering.txt | 11 1. Barriers (also known as "fences"). A barrier orders some or 67 First, the smp_mb() full memory barrier orders all of the CPU's prior 115 synchronize_srcu() and so on. However, these primitives have orders
|
/linux/Documentation/userspace-api/media/v4l/ |
H A D | pixfmt-bayer.rst | 15 orders. See also `the Wikipedia article on Bayer filter
|
H A D | field-order.rst | 80 If multiple field orders are possible the 81 driver must choose one of the possible field orders during
|
/linux/Documentation/ |
H A D | atomic_t.txt | 194 smp_mb__before_atomic() orders all earlier accesses against the RMW op 195 itself and all accesses following it, and smp_mb__after_atomic() orders all 226 a RELEASE because it orders preceding instructions against both the read
|
/linux/fs/netfs/ |
H A D | rolling_buffer.c | 139 fq->orders[ix] = order; in rolling_buffer_load_from_ra()
|
/linux/arch/arm/boot/dts/marvell/ |
H A D | armada-385-synology-ds116.dts | 53 * and takes single-character orders :
|
/linux/Documentation/admin-guide/mm/ |
H A D | transhuge.rst | 579 lower orders or small pages. 583 instead falls back to using huge pages with lower orders or 596 and instead falls back to using huge pages with lower orders or 601 falls back to using huge pages with lower orders or small pages 628 smaller orders. This can happen for a variety of reasons but a
|
H A D | pagemap.rst | 99 The buddy system organizes free memory in blocks of various orders.
|
/linux/tools/virtio/virtio-trace/ |
H A D | README | 9 - controlled by start/stop orders from a Host
|
/linux/Documentation/core-api/ |
H A D | genalloc.rst | 49 that state, so one of the first orders of business is usually to add memory
|
H A D | memory-allocation.rst | 130 This might be really dangerous especially for larger orders.
|
/linux/tools/memory-model/ |
H A D | lock.cat | 8 * Generate coherence orders and handle lock operations
|
H A D | linux-kernel.cat | 158 (* rb orders instructions just as pb does *)
|