| /linux/arch/powerpc/mm/nohash/ |
| H A D | kaslr_booke.c | 23 struct regions { struct 38 struct regions __initdata regions; argument 113 if (regions.reserved_mem < 0) in overlaps_reserved_region() 117 for (subnode = fdt_first_subnode(fdt, regions.reserved_mem); in overlaps_reserved_region() 125 while (len >= (regions.reserved_mem_addr_cells + in overlaps_reserved_region() 126 regions.reserved_mem_size_cells)) { in overlaps_reserved_region() 128 if (regions.reserved_mem_addr_cells == 2) in overlaps_reserved_region() 131 reg += regions.reserved_mem_addr_cells; in overlaps_reserved_region() 132 len -= 4 * regions.reserved_mem_addr_cells; in overlaps_reserved_region() 135 if (regions.reserved_mem_size_cells == 2) in overlaps_reserved_region() [all …]
|
| /linux/drivers/mtd/devices/ |
| H A D | mtd_intel_dg.c | 44 } regions[] __counted_by(nregions); member 167 * [3:0] regions 12-15 read state 168 * [7:4] regions 12-15 write state 169 * [19:8] regions 0-11 read state 170 * [31:20] regions 0-11 write state 209 if ((nvm->regions[i].offset + nvm->regions[i].size - 1) >= from && in idg_nvm_get_region() 210 nvm->regions[i].offset <= from && in idg_nvm_get_region() 211 nvm->regions[i].size != 0) in idg_nvm_get_region() 444 u8 id = nvm->regions[i].id; in intel_dg_nvm_init() 454 id, nvm->regions[i].name, region, base, limit); in intel_dg_nvm_init() [all …]
|
| /linux/drivers/mtd/chips/ |
| H A D | jedec_probe.c | 275 const uint32_t regions[6]; member 307 .regions = { 319 .regions = { 334 .regions = { 349 .regions = { 364 .regions = { 379 .regions = { 395 .regions = { 412 .regions = { 429 .regions = { [all …]
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | dm-clone.rst | 58 3. A small metadata device - it records which regions are already valid in the 59 destination device, i.e., which regions have already been hydrated, or have 65 Regions section in Design 68 dm-clone divides the source and destination devices in fixed sized regions. 69 Regions are the unit of hydration, i.e., the minimum amount of data copied from 77 Reads and writes from/to hydrated regions are serviced from the destination 93 as a hint to skip hydration of the regions covered by the request, i.e., it 111 A message `hydration_threshold <#regions>` can be used to set the maximum number 112 of regions being copied, the default being 1 region. 116 region size. A message `hydration_batch_size <#regions>` can be used to tune the [all …]
|
| /linux/tools/testing/selftests/damon/ |
| H A D | damon_nr_regions.py | 11 Create process of the given 'real_nr_regions' regions, monitor it using 21 # stat every monitored regions 46 print('tried regions update failed: %s' % err) 52 print('tried regions is not collected') 58 print('tried regions is not created') 72 print('number of regions that collected are:') 79 # test min_nr_regions larger than real nr regions 82 # test max_nr_regions smaller than real nr regions 85 # test online-tuned max_nr_regions that smaller than real nr regions 89 # stat every monitored regions [all …]
|
| H A D | access_memory_even.c | 5 * Receives number of regions and size of each region from user. Allocate the 6 * regions and repeatedly access even numbered (starting from zero) regions. 15 char **regions; in main() local 28 regions = malloc(sizeof(*regions) * nr_regions); in main() 30 regions[i] = malloc(sz_region); in main() 35 memset(regions[i], i, sz_region); in main()
|
| H A D | access_memory.c | 18 char **regions; in main() 40 regions = malloc(sizeof(*regions) * nr_regions); in main() 42 regions[i] = malloc(sz_region); 49 memset(regions[i], i, sz_region); 13 char **regions; main() local
|
| H A D | damos_tried_regions.py | 10 # repeatedly access even-numbered ones in 14 regions of 10 MiB size 14 # stat every monitored regions 36 print('tried regions update failed: %s' % err) 42 print('tried regions is not collected') 48 print('tried regions is not created')
|
| /linux/Documentation/networking/devlink/ |
| H A D | devlink-region.rst | 7 ``devlink`` regions enable access to driver defined address regions using 10 Each device can create and register its own supported address regions. The 15 Regions may optionally support triggering snapshots on demand. 22 address regions that are otherwise inaccessible to the user. 24 Regions may also be used to provide an additional way to debug complex error 27 Regions may optionally support capturing a snapshot on demand via the 34 Regions may optionally allow directly reading from their contents without a 54 # Show all of the exposed regions with region sizes: 81 As regions are likely very device or driver specific, no generic regions are 83 specific regions a driver supports.
|
| /linux/drivers/vfio/platform/ |
| H A D | vfio_platform_common.c | 144 vdev->regions = kzalloc_objs(struct vfio_platform_region, cnt, in vfio_platform_regions_init() 146 if (!vdev->regions) in vfio_platform_regions_init() 153 vdev->regions[i].addr = res->start; in vfio_platform_regions_init() 154 vdev->regions[i].size = resource_size(res); in vfio_platform_regions_init() 155 vdev->regions[i].flags = 0; in vfio_platform_regions_init() 159 vdev->regions[i].type = VFIO_PLATFORM_REGION_TYPE_MMIO; in vfio_platform_regions_init() 160 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ; in vfio_platform_regions_init() 162 vdev->regions[i].flags |= in vfio_platform_regions_init() 166 * Only regions addressed with PAGE granularity may be in vfio_platform_regions_init() 169 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_platform_regions_init() [all …]
|
| /linux/tools/testing/memblock/tests/ |
| H A D | basic_api.c | 17 ASSERT_NE(memblock.memory.regions, NULL); in memblock_initialization_check() 22 ASSERT_NE(memblock.reserved.regions, NULL); in memblock_initialization_check() 37 * and size to the collection of available memory regions (memblock.memory). 45 rgn = &memblock.memory.regions[0]; in memblock_add_simple_check() 70 * NUMA node and memory flags to the collection of available memory regions. 78 rgn = &memblock.memory.regions[0]; in memblock_add_node_simple_check() 114 * available memory regions (memblock.memory). The total size and 121 rgn1 = &memblock.memory.regions[0]; in memblock_add_disjoint_check() 122 rgn2 = &memblock.memory.regions[1]; in memblock_add_disjoint_check() 167 * and has size of two regions minus their intersection. The total size of [all …]
|
| H A D | alloc_nid_api.c | 66 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_simple_check() 118 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_end_misaligned_check() 169 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_exact_address_generic_check() 221 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_narrow_range_check() 307 * Expect a merge of both regions. Only the region size gets updated. 311 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_min_reserved_generic_check() 359 * Expect a merge of regions. Only the region size gets updated. 363 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_nid_max_reserved_generic_check() 399 * there are two reserved regions at the borders, with a gap big enough to fit 416 struct memblock_region *rgn1 = &memblock.reserved.regions[1]; in alloc_nid_top_down_reserved_with_space_check() [all …]
|
| H A D | alloc_exact_nid_api.c | 30 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_top_down_numa_simple_check() 31 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_simple_check() 82 struct memblock_region *new_rgn = &memblock.reserved.regions[1]; in alloc_exact_nid_top_down_numa_part_reserved_check() 83 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_part_reserved_check() 143 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_top_down_numa_split_range_low_check() 144 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_split_range_low_check() 200 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_top_down_numa_no_overlap_split_check() 201 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_no_overlap_split_check() 202 struct memblock_region *node2 = &memblock.memory.regions[6]; in alloc_exact_nid_top_down_numa_no_overlap_split_check() 258 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_exact_nid_top_down_numa_no_overlap_low_check() [all …]
|
| H A D | alloc_api.c | 26 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_simple_check() 73 struct memblock_region *rgn1 = &memblock.reserved.regions[1]; in alloc_top_down_disjoint_check() 74 struct memblock_region *rgn2 = &memblock.reserved.regions[0]; in alloc_top_down_disjoint_check() 121 * Expect a merge of both regions. Only the region size gets updated. 125 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_before_check() 163 * Expect a merge of both regions. Both the base address and size of the region 168 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_after_check() 204 * A test that tries to allocate memory when there are two reserved regions with 217 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_top_down_second_fit_check() 254 * A test that tries to allocate memory when there are two reserved regions with [all …]
|
| H A D | alloc_helpers_api.c | 20 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_simple_generic_check() 63 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_misaligned_generic_check() 110 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_high_addr_check() 149 * regions get merged into one. 153 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_no_space_above_check() 186 * Expect successful allocation and merge of both regions. 190 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_top_down_min_addr_cap_check() 236 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_high_addr_check() 278 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_no_space_above_check() 314 struct memblock_region *rgn = &memblock.reserved.regions[0]; in alloc_from_bottom_up_min_addr_cap_check()
|
| /linux/drivers/virt/nitro_enclaves/ |
| H A D | ne_misc_dev_test.c | 23 * regions = {} 34 * regions = {} 45 * regions = { 58 * regions = { 72 * regions = { 87 * regions = { 102 * regions = { 117 phys_contig_mem_regions.regions = kunit_kcalloc(test, MAX_PHYS_REGIONS, in ne_misc_dev_test_merge_phys_contig_memory_regions() 118 sizeof(*phys_contig_mem_regions.regions), in ne_misc_dev_test_merge_phys_contig_memory_regions() 120 KUNIT_ASSERT_TRUE(test, phys_contig_mem_regions.regions); in ne_misc_dev_test_merge_phys_contig_memory_regions() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/nvfw/ |
| H A D | acr.c | 130 hdr->regions.no_regions); in flcn_acr_desc_dump() 132 for (i = 0; i < ARRAY_SIZE(hdr->regions.region_props); i++) { in flcn_acr_desc_dump() 135 hdr->regions.region_props[i].start_addr); in flcn_acr_desc_dump() 137 hdr->regions.region_props[i].end_addr); in flcn_acr_desc_dump() 139 hdr->regions.region_props[i].region_id); in flcn_acr_desc_dump() 141 hdr->regions.region_props[i].read_mask); in flcn_acr_desc_dump() 143 hdr->regions.region_props[i].write_mask); in flcn_acr_desc_dump() 145 hdr->regions.region_props[i].client_mask); in flcn_acr_desc_dump() 173 hdr->regions.no_regions); in flcn_acr_desc_v1_dump() 175 for (i = 0; i < ARRAY_SIZE(hdr->regions.region_props); i++) { in flcn_acr_desc_v1_dump() [all …]
|
| /linux/drivers/net/dsa/sja1105/ |
| H A D | sja1105_devlink.c | 7 /* Since devlink regions have a fixed size and the static config has a variable 85 priv->regions = kzalloc_objs(struct devlink_region *, num_regions); in sja1105_setup_devlink_regions() 86 if (!priv->regions) in sja1105_setup_devlink_regions() 96 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_setup_devlink_regions() 98 kfree(priv->regions); in sja1105_setup_devlink_regions() 102 priv->regions[i] = region; in sja1105_setup_devlink_regions() 114 dsa_devlink_region_destroy(priv->regions[i]); in sja1105_teardown_devlink_regions() 116 kfree(priv->regions); in sja1105_teardown_devlink_regions()
|
| /linux/drivers/md/ |
| H A D | dm-bio-prison-v1.c | 29 struct prison_region regions[] __counted_by(num_locks); 47 prison = kzalloc_flex(*prison, regions, num_locks); in dm_bio_prison_create() 53 spin_lock_init(&prison->regions[i].lock); in dm_bio_prison_create() 54 prison->regions[i].cell = RB_ROOT; in dm_bio_prison_create() 184 spin_lock_irq(&prison->regions[l].lock); in bio_detain() 185 r = __bio_detain(&prison->regions[l].cell, key, inmate, cell_prealloc, cell_result); in bio_detain() 186 spin_unlock_irq(&prison->regions[l].lock); in bio_detain() 223 spin_lock_irq(&prison->regions[l].lock); in dm_cell_release() 224 __cell_release(&prison->regions[l].cell, cell, bios); in dm_cell_release() 225 spin_unlock_irq(&prison->regions[l].lock); in dm_cell_release() [all …]
|
| /linux/include/linux/ |
| H A D | damon.h | 75 * regions are merged into a new region, both @nr_accesses and @age of the new 76 * region are set as region size-weighted average of those of the two regions. 93 * @nr_regions: Number of monitoring target regions of this target. 94 * @regions_list: Head of the monitoring target regions of this target. 126 * @DAMOS_MIGRATE_HOT: Migrate the regions prioritizing warmer regions. 127 * @DAMOS_MIGRATE_COLD: Migrate the regions prioritizing colder regions. 265 * For selecting regions within the quota, DAMON prioritizes current scheme's 266 * target memory regions using the &struct damon_operations->get_scheme_score. 324 * appropriate memory regions. Else, DAMON checks &metric of the system for at 344 * @nr_tried: Total number of regions that the scheme is tried to be applied. [all …]
|
| H A D | kcsan.h | 28 * We distinguish between: (a) nestable atomic regions that may contain 29 * other nestable regions; and (b) flat atomic regions that do not keep 36 * atomic regions, but reader critical sections as (b) flat atomic 37 * regions, but have encountered cases where seqlock reader critical
|
| /linux/Documentation/devicetree/bindings/fpga/ |
| H A D | fpga-region.yaml | 24 FPGA Regions represent FPGA's and partial reconfiguration regions of FPGA's in 25 the Device Tree. FPGA Regions provide a way to program FPGAs under device tree 60 * A persona may create more regions. 70 will be used to gate the busses. Traffic to other regions is not affected. 74 * An FPGA image may create a set of reprogrammable regions, each having its 84 * A base image may set up a set of partial reconfiguration regions that may 104 Figure 1: An FPGA set up with a base image that created three regions. Each 127 FPGA Regions represent FPGA's and FPGA PR regions in the device tree. An FPGA 154 These FPGA regions are children of FPGA bridges which are then children of the 163 FPGA Regions do not inherit their ancestor FPGA regions' bridges. This prevents [all …]
|
| /linux/drivers/soc/qcom/ |
| H A D | smem.c | 43 * the partition and holds properties for the two internal memory regions. The 44 * two regions are cached and non-cached memory respectively. Each region 273 * @num_regions: number of @regions 274 * @regions: list of the memory regions defining the shared memory 288 struct smem_region regions[] __counted_by(num_regions); 472 header = smem->regions[0].virt_base; in qcom_smem_alloc_global() 561 header = smem->regions[0].virt_base; in qcom_smem_get_global() 569 region = &smem->regions[i]; in qcom_smem_get_global() 750 header = __smem->regions[0].virt_base; in qcom_smem_get_free_space() 753 if (ret > __smem->regions[0].size) in qcom_smem_get_free_space() [all …]
|
| /linux/drivers/virt/acrn/ |
| H A D | mm.c | 21 struct vm_memory_region_batch *regions; in modify_region() local 24 regions = kzalloc_obj(*regions); in modify_region() 25 if (!regions) in modify_region() 28 regions->vmid = vm->vmid; in modify_region() 29 regions->regions_num = 1; in modify_region() 30 regions->regions_gpa = virt_to_phys(region); in modify_region() 32 ret = hcall_set_memory_regions(virt_to_phys(regions)); in modify_region() 37 kfree(regions); in modify_region() 322 "Failed to set regions, VM[%u]!\n", vm->vmid); in acrn_vm_ram_map()
|
| /linux/arch/powerpc/platforms/powernv/ |
| H A D | opal-fadump.c | 57 * Preserve memory only if kernel memory regions are registered in opal_fadump_dt_scan() 98 pr_debug("Boot memory regions count: %d\n", be16_to_cpu(fdm->region_cnt)); in opal_fadump_update_config() 102 * destination address of boot memory regions. in opal_fadump_update_config() 105 pr_debug("Destination address of boot memory regions: %#016llx\n", in opal_fadump_update_config() 128 pr_debug("Boot memory regions:\n"); in opal_fadump_get_config() 150 * boot memory regions are registered for MPIPL. In such in opal_fadump_get_config() 159 pr_warn("Not all memory regions were saved!!!\n"); in opal_fadump_get_config() 160 pr_warn(" Unsaved memory regions:\n"); in opal_fadump_get_config() 169 …pr_warn("If the unsaved regions only contain pages that are filtered out (eg. free/user pages), th… in opal_fadump_get_config() 170 pr_warn("WARNING: If the unsaved regions contain kernel pages, the vmcore will be corrupted.\n"); in opal_fadump_get_config() [all …]
|