| /linux/drivers/hv/ |
| H A D | mshv_regions.c | 77 static long mshv_region_process_chunk(struct mshv_mem_region *region, in mshv_region_process_chunk() argument 80 int (*handler)(struct mshv_mem_region *region, in mshv_region_process_chunk() argument 86 u64 gfn = region->start_gfn + page_offset; in mshv_region_process_chunk() 91 page = region->mreg_pages[page_offset]; in mshv_region_process_chunk() 101 page = region->mreg_pages[page_offset + count]; in mshv_region_process_chunk() 113 ret = handler(region, flags, page_offset, count, stride > 1); in mshv_region_process_chunk() 139 static int mshv_region_process_range(struct mshv_mem_region *region, in mshv_region_process_range() argument 142 int (*handler)(struct mshv_mem_region *region, in mshv_region_process_range() argument 150 if (page_offset + page_count > region->nr_pages) in mshv_region_process_range() 155 if (!region->mreg_pages[page_offset]) { in mshv_region_process_range() [all …]
|
| /linux/drivers/fpga/ |
| H A D | dfl-afu-dma-region.c | 35 struct dfl_afu_dma_region *region) in afu_dma_pin_pages() argument 37 int npages = region->length >> PAGE_SHIFT; in afu_dma_pin_pages() 45 region->pages = kzalloc_objs(struct page *, npages); in afu_dma_pin_pages() 46 if (!region->pages) { in afu_dma_pin_pages() 51 pinned = pin_user_pages_fast(region->user_addr, npages, FOLL_WRITE, in afu_dma_pin_pages() 52 region->pages); in afu_dma_pin_pages() 66 unpin_user_pages(region->pages, pinned); in afu_dma_pin_pages() 68 kfree(region->pages); in afu_dma_pin_pages() 83 struct dfl_afu_dma_region *region) in afu_dma_unpin_pages() argument 85 long npages = region->length >> PAGE_SHIFT; in afu_dma_unpin_pages() [all …]
|
| H A D | fpga-region.c | 46 static struct fpga_region *fpga_region_get(struct fpga_region *region) in fpga_region_get() argument 48 struct device *dev = ®ion->dev; in fpga_region_get() 50 if (!mutex_trylock(®ion->mutex)) { in fpga_region_get() 56 if (!try_module_get(region->ops_owner)) { in fpga_region_get() 58 mutex_unlock(®ion->mutex); in fpga_region_get() 64 return region; in fpga_region_get() 72 static void fpga_region_put(struct fpga_region *region) in fpga_region_put() argument 74 struct device *dev = ®ion->dev; in fpga_region_put() 78 module_put(region->ops_owner); in fpga_region_put() 80 mutex_unlock(®ion->mutex); in fpga_region_put() [all …]
|
| /linux/net/devlink/ |
| H A D | region.c | 29 struct devlink_region *region; member 37 struct devlink_region *region; in devlink_region_get_by_name() local 39 list_for_each_entry(region, &devlink->region_list, list) in devlink_region_get_by_name() 40 if (!strcmp(region->ops->name, region_name)) in devlink_region_get_by_name() 41 return region; in devlink_region_get_by_name() 50 struct devlink_region *region; in devlink_port_region_get_by_name() local 52 list_for_each_entry(region, &port->region_list, list) in devlink_port_region_get_by_name() 53 if (!strcmp(region->port_ops->name, region_name)) in devlink_port_region_get_by_name() 54 return region; in devlink_port_region_get_by_name() 60 devlink_region_snapshot_get_by_id(struct devlink_region *region, u32 id) in devlink_region_snapshot_get_by_id() argument [all …]
|
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | kvm_util.c | 266 * range addressed by a single page table into a low and high region in ____vm_create() 268 * the VA region spans [0, 2^(va_bits - 1)), [-(2^(va_bits - 1), -1]. in ____vm_create() 424 * maximum page table size for a memory region will be when the in kvm_set_files_rlimit() 496 * Force GUEST_MEMFD for the primary memory region if necessary, e.g. 513 * MMIO region would prevent silently clobbering the MMIO region. 516 ucall_init(vm, slot0->region.guest_phys_addr + slot0->region.memory_size); in __vm_create_with_vcpus() 596 struct userspace_mem_region *region; in __pin_task_to_cpu() 602 hash_for_each(vmp->regions.slot_hash, ctr, region, slot_nod in __pin_task_to_cpu() 563 struct userspace_mem_region *region; kvm_vm_restart() local 690 struct userspace_mem_region *region = userspace_mem_region_find() local 773 __vm_mem_region_delete(struct kvm_vm * vm,struct userspace_mem_region * region) __vm_mem_region_delete() argument 800 struct userspace_mem_region *region; kvm_vm_free() local 837 vm_userspace_mem_region_gpa_insert(struct rb_root * gpa_tree,struct userspace_mem_region * region) vm_userspace_mem_region_gpa_insert() argument 863 vm_userspace_mem_region_hva_insert(struct rb_root * hva_tree,struct userspace_mem_region * region) vm_userspace_mem_region_hva_insert() argument 891 struct kvm_userspace_memory_region region = { __vm_set_user_memory_region() local 919 struct kvm_userspace_memory_region2 region = { __vm_set_user_memory_region2() local 952 struct userspace_mem_region *region; vm_mem_add() local 1142 struct userspace_mem_region *region; memslot2region() local 1174 struct userspace_mem_region *region; vm_mem_region_set_flags() local 1189 struct userspace_mem_region *region = memslot2region(vm, slot); vm_mem_region_reload() local 1213 struct userspace_mem_region *region; vm_mem_region_move() local 1242 struct userspace_mem_region *region = memslot2region(vm, slot); vm_mem_region_delete() local 1254 struct userspace_mem_region *region; vm_guest_mem_fallocate() local 1590 struct userspace_mem_region *region; addr_gpa2hva() local 1626 struct userspace_mem_region *region = addr_hva2gpa() local 1666 struct userspace_mem_region *region; addr_gpa2alias() local 1936 struct userspace_mem_region *region; vm_dump() local 2072 struct userspace_mem_region *region; __vm_phy_pages_alloc() local 2337 struct userspace_mem_region *region; vm_is_gpa_protected() local [all...] |
| /linux/tools/testing/selftests/vfio/ |
| H A D | vfio_dma_mapping_test.c | 138 struct dma_region region; in TEST_F() local 144 region.vaddr = mmap(NULL, size, PROT_READ | PROT_WRITE, flags, -1, 0); in TEST_F() 147 if (flags & MAP_HUGETLB && region.vaddr == MAP_FAILED) in TEST_F() 150 ASSERT_NE(region.vaddr, MAP_FAILED); in TEST_F() 152 region.iova = iova_allocator_alloc(self->iova_allocator, size); in TEST_F() 153 region.size = size; in TEST_F() 155 iommu_map(self->iommu, ®ion); in TEST_F() 156 printf("Mapped HVA %p (size 0x%lx) at IOVA 0x%lx\n", region.vaddr, size, region.iova); in TEST_F() 158 ASSERT_EQ(region.iova, to_iova(self->device, region.vaddr)); in TEST_F() 160 rc = iommu_mapping_get(device_bdf, region.iova, &mapping); in TEST_F() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| H A D | spectrum1_acl_tcam.c | 14 struct mlxsw_sp_acl_tcam_region *region; member 62 struct mlxsw_sp1_acl_tcam_region *region) in mlxsw_sp1_acl_ctcam_region_catchall_add() argument 67 mlxsw_sp_acl_ctcam_chunk_init(®ion->cregion, in mlxsw_sp1_acl_ctcam_region_catchall_add() 68 ®ion->catchall.cchunk, in mlxsw_sp1_acl_ctcam_region_catchall_add() 81 err = mlxsw_sp_acl_ctcam_entry_add(mlxsw_sp, ®ion->cregion, in mlxsw_sp1_acl_ctcam_region_catchall_add() 82 ®ion->catchall.cchunk, in mlxsw_sp1_acl_ctcam_region_catchall_add() 83 ®ion->catchall.centry, in mlxsw_sp1_acl_ctcam_region_catchall_add() 87 region->catchall.rulei = rulei; in mlxsw_sp1_acl_ctcam_region_catchall_add() 95 mlxsw_sp_acl_ctcam_chunk_fini(®ion->catchall.cchunk); in mlxsw_sp1_acl_ctcam_region_catchall_add() 101 struct mlxsw_sp1_acl_tcam_region *region) in mlxsw_sp1_acl_ctcam_region_catchall_del() argument [all …]
|
| H A D | spectrum_acl_ctcam.c | 15 struct mlxsw_sp_acl_tcam_region *region, in mlxsw_sp_acl_ctcam_region_resize() argument 21 region->key_type, new_size, region->id, in mlxsw_sp_acl_ctcam_region_resize() 22 region->tcam_region_info); in mlxsw_sp_acl_ctcam_region_resize() 28 struct mlxsw_sp_acl_tcam_region *region, in mlxsw_sp_acl_ctcam_region_move() argument 34 region->tcam_region_info, src_offset, in mlxsw_sp_acl_ctcam_region_move() 35 region->tcam_region_info, dst_offset, size); in mlxsw_sp_acl_ctcam_region_move() 46 struct mlxsw_sp_acl_tcam_region *region = cregion->region; in mlxsw_sp_acl_ctcam_region_entry_insert() local 61 region->tcam_region_info, in mlxsw_sp_acl_ctcam_region_entry_insert() 65 mlxsw_afk_encode(afk, region->key_info, &rulei->values, key, mask); in mlxsw_sp_acl_ctcam_region_entry_insert() 94 cregion->region->tcam_region_info, in mlxsw_sp_acl_ctcam_region_entry_remove() [all …]
|
| H A D | spectrum2_acl_tcam.c | 18 struct mlxsw_sp_acl_tcam_region *region; member 155 struct mlxsw_sp2_acl_tcam_region *region = region_priv; in mlxsw_sp2_acl_tcam_region_init() local 158 region->region = _region; in mlxsw_sp2_acl_tcam_region_init() 161 ®ion->aregion, in mlxsw_sp2_acl_tcam_region_init() 169 struct mlxsw_sp2_acl_tcam_region *region = region_priv; in mlxsw_sp2_acl_tcam_region_fini() local 171 mlxsw_sp_acl_atcam_region_fini(®ion->aregion); in mlxsw_sp2_acl_tcam_region_fini() 176 struct mlxsw_sp_acl_tcam_region *region) in mlxsw_sp2_acl_tcam_region_associate() argument 178 return mlxsw_sp_acl_atcam_region_associate(mlxsw_sp, region->id); in mlxsw_sp2_acl_tcam_region_associate() 183 struct mlxsw_sp2_acl_tcam_region *region = region_priv; in mlxsw_sp2_acl_tcam_region_rehash_hints_get() local 185 return mlxsw_sp_acl_atcam_rehash_hints_get(®ion->aregion); in mlxsw_sp2_acl_tcam_region_rehash_hints_get() [all …]
|
| H A D | spectrum_acl_tcam.c | 149 struct mlxsw_sp_acl_tcam_region *region; member 169 struct mlxsw_sp_acl_tcam_region *region; member 210 struct mlxsw_sp_acl_tcam_region *region; in mlxsw_sp_acl_tcam_group_update() local 215 list_for_each_entry(region, &group->region_list, list) { in mlxsw_sp_acl_tcam_group_update() 219 if (region->list.next != &group->region_list && in mlxsw_sp_acl_tcam_group_update() 220 list_next_entry(region, list)->vregion == region->vregion) in mlxsw_sp_acl_tcam_group_update() 223 region->id, multi); in mlxsw_sp_acl_tcam_group_update() 385 struct mlxsw_sp_acl_tcam_region *region, in mlxsw_sp_acl_tcam_group_region_attach() argument 413 list_add_tail(®ion->list, pos); in mlxsw_sp_acl_tcam_group_region_attach() 414 region->group = group; in mlxsw_sp_acl_tcam_group_region_attach() [all …]
|
| /linux/drivers/vfio/pci/pds/ |
| H A D | dirty.c | 72 static int pds_vfio_dirty_alloc_bitmaps(struct pds_vfio_region *region, in pds_vfio_dirty_alloc_bitmaps() argument 87 region->host_seq = host_seq_bmp; in pds_vfio_dirty_alloc_bitmaps() 88 region->host_ack = host_ack_bmp; in pds_vfio_dirty_alloc_bitmaps() 89 region->bmp_bytes = bytes; in pds_vfio_dirty_alloc_bitmaps() 100 struct pds_vfio_region *region = &dirty->regions[i]; in pds_vfio_dirty_free_bitmaps() local 102 vfree(region->host_seq); in pds_vfio_dirty_free_bitmaps() 103 vfree(region->host_ack); in pds_vfio_dirty_free_bitmaps() 104 region->host_seq = NULL; in pds_vfio_dirty_free_bitmaps() 105 region->host_ack = NULL; in pds_vfio_dirty_free_bitmaps() 106 region->bmp_bytes = 0; in pds_vfio_dirty_free_bitmaps() [all …]
|
| /linux/drivers/platform/x86/intel/pmt/ |
| H A D | discovery-kunit.c | 27 struct telemetry_region *region = &feature_group->regions[i]; in validate_pmt_regions() local 30 i, region->plat_info.cdie_mask, region->plat_info.package_id, in validate_pmt_regions() 31 region->plat_info.partition, region->plat_info.segment); in validate_pmt_regions() 33 region->plat_info.bus_number, region->plat_info.device_number, in validate_pmt_regions() 34 region->plat_info.function_number, region->guid); in validate_pmt_regions() 35 kunit_info(test, "\t\taddr=%p, size=%zu, num_rmids=%u", region->addr, region->size, in validate_pmt_regions() 36 region->num_rmids); in validate_pmt_regions() 39 KUNIT_ASSERT_GE(test, region->plat_info.cdie_mask, 0); in validate_pmt_regions() 40 KUNIT_ASSERT_GE(test, region->plat_info.package_id, 0); in validate_pmt_regions() 41 KUNIT_ASSERT_GE(test, region->plat_info.partition, 0); in validate_pmt_regions() [all …]
|
| /linux/kernel/cgroup/ |
| H A D | dmem.c | 63 struct dmem_cgroup_region *region; member 137 kref_put(&pool->region->ref, dmemcg_free_region); in dmemcg_pool_free_rcu() 234 find_cg_pool_locked(struct dmemcg_state *dmemcs, struct dmem_cgroup_region *region) in find_cg_pool_locked() argument 239 if (pool->region == region) in find_cg_pool_locked() 271 if (pool->region == limit_pool->region) { in dmem_cgroup_calculate_protection() 355 alloc_pool_single(struct dmemcg_state *dmemcs, struct dmem_cgroup_region *region, in alloc_pool_single() argument 370 pool->region = region; in alloc_pool_single() 374 ppool = find_cg_pool_locked(parent, region); in alloc_pool_single() 380 kref_get(®ion->ref); in alloc_pool_single() 387 list_add_tail(&pool->region_node, ®ion->pools); in alloc_pool_single() [all …]
|
| /linux/mm/ |
| H A D | nommu.c | 440 struct vm_region *region, *last; in validate_nommu_regions() local 452 region = rb_entry(p, struct vm_region, vm_rb); in validate_nommu_regions() 455 BUG_ON(region->vm_end <= region->vm_start); in validate_nommu_regions() 456 BUG_ON(region->vm_top < region->vm_end); in validate_nommu_regions() 457 BUG_ON(region->vm_start < last->vm_top); in validate_nommu_regions() 471 static void add_nommu_region(struct vm_region *region) in add_nommu_region() argument 483 if (region->vm_start < pregion->vm_start) in add_nommu_region() 485 else if (region->vm_start > pregion->vm_start) in add_nommu_region() 487 else if (pregion == region) in add_nommu_region() 493 rb_link_node(®ion->vm_rb, parent, p); in add_nommu_region() [all …]
|
| /linux/arch/x86/boot/compressed/ |
| H A D | kaslr.c | 463 static void store_slot_info(struct mem_vector *region, unsigned long image_size) in store_slot_info() argument 470 slot_area.addr = region->start; in store_slot_info() 471 slot_area.num = 1 + (region->size - image_size) / CONFIG_PHYSICAL_ALIGN; in store_slot_info() 482 process_gb_huge_pages(struct mem_vector *region, unsigned long image_size) in process_gb_huge_pages() argument 489 store_slot_info(region, image_size); in process_gb_huge_pages() 494 pud_start = ALIGN(region->start, PUD_SIZE); in process_gb_huge_pages() 495 pud_end = ALIGN_DOWN(region->start + region->size, PUD_SIZE); in process_gb_huge_pages() 499 store_slot_info(region, image_size); in process_gb_huge_pages() 504 if (pud_start >= region->start + image_size) { in process_gb_huge_pages() 505 tmp.start = region->start; in process_gb_huge_pages() [all …]
|
| /linux/drivers/s390/cio/ |
| H A D | vfio_ccw_chp.c | 22 struct ccw_schib_region *region; in vfio_ccw_schib_region_read() local 25 if (pos + count > sizeof(*region)) in vfio_ccw_schib_region_read() 29 region = private->region[i].data; in vfio_ccw_schib_region_read() 36 memcpy(region, &sch->schib, sizeof(*region)); in vfio_ccw_schib_region_read() 38 if (copy_to_user(buf, (void *)region + pos, count)) { in vfio_ccw_schib_region_read() 59 struct vfio_ccw_region *region) in vfio_ccw_schib_region_release() argument 86 struct ccw_crw_region *region; in vfio_ccw_crw_region_read() local 90 if (pos + count > sizeof(*region)) in vfio_ccw_crw_region_read() 100 region = private->region[i].data; in vfio_ccw_crw_region_read() 103 memcpy(®ion->crw, &crw->crw, sizeof(region->crw)); in vfio_ccw_crw_region_read() [all …]
|
| H A D | vfio_ccw_async.c | 20 struct ccw_cmd_region *region; in vfio_ccw_async_region_read() local 23 if (pos + count > sizeof(*region)) in vfio_ccw_async_region_read() 27 region = private->region[i].data; in vfio_ccw_async_region_read() 28 if (copy_to_user(buf, (void *)region + pos, count)) in vfio_ccw_async_region_read() 42 struct ccw_cmd_region *region; in vfio_ccw_async_region_write() local 45 if (pos + count > sizeof(*region)) in vfio_ccw_async_region_write() 51 region = private->region[i].data; in vfio_ccw_async_region_write() 52 if (copy_from_user((void *)region + pos, buf, count)) { in vfio_ccw_async_region_write() 59 ret = region->ret_code ? region->ret_code : count; in vfio_ccw_async_region_write() 67 struct vfio_ccw_region *region) in vfio_ccw_async_region_release() argument
|
| /linux/tools/testing/selftests/mm/ |
| H A D | rmap.c | 60 char *region; member 199 if (data->region != MAP_FAILED) in FIXTURE_TEARDOWN() 200 munmap(data->region, data->mapsize); in FIXTURE_TEARDOWN() 201 data->region = MAP_FAILED; in FIXTURE_TEARDOWN() 227 FORCE_READ(*data->region); in access_region() 230 int try_to_move_page(char *region) in try_to_move_page() argument 237 ret = move_pages(0, 1, (void **)®ion, NULL, &status, MPOL_MF_MOVE_ALL); in try_to_move_page() 255 ret = move_pages(0, 1, (void **)®ion, &node, &status, MPOL_MF_MOVE_ALL); in try_to_move_page() 280 ret = try_to_move_page(data->region); in move_region() 287 *data->expected_pfn = pagemap_get_pfn(pagemap_fd, data->region); in move_region() [all …]
|
| /linux/drivers/acpi/pmic/ |
| H A D | Kconfig | 4 bool "PMIC (Power Management Integrated Circuit) operation region support" 7 region of the PMIC chip. The operation region can be used 14 bool "ACPI operation region support for Bay Trail Crystal Cove PMIC" 17 This config adds ACPI operation region support for the Bay Trail 21 bool "ACPI operation region support for Cherry Trail Crystal Cove PMIC" 24 This config adds ACPI operation region support for the Cherry Trail 28 bool "ACPI operation region support for XPower AXP288 PMIC" 31 This config adds ACPI operation region support for XPower AXP288 PMIC. 34 bool "ACPI operation region support for BXT WhiskeyCove PMIC" 37 This config adds ACPI operation region support for BXT WhiskeyCove PMIC. [all …]
|
| /linux/drivers/mtd/maps/ |
| H A D | pismo.c | 89 struct pismo_mem *region, const char *name, in pismo_add_device() argument 94 phys_addr_t base = region->base; in pismo_add_device() 101 res.end = base + region->size - 1; in pismo_add_device() 131 struct pismo_mem *region) in pismo_add_nor() argument 134 .width = region->width, in pismo_add_nor() 140 return pismo_add_device(pismo, i, region, "physmap-flash", in pismo_add_nor() 145 struct pismo_mem *region) in pismo_add_sram() argument 148 .bankwidth = region->width, in pismo_add_sram() 151 return pismo_add_device(pismo, i, region, "mtd-ram", in pismo_add_sram() 159 struct pismo_mem region; in pismo_add_one() local [all …]
|
| /linux/tools/testing/selftests/vfio/lib/ |
| H A D | iommu.c | 71 struct dma_region *region; in __iommu_hva2iova() local 73 list_for_each_entry(region, &iommu->dma_regions, link) { in __iommu_hva2iova() 74 if (vaddr < region->vaddr) in __iommu_hva2iova() 77 if (vaddr >= region->vaddr + region->size) in __iommu_hva2iova() 81 *iova = region->iova + (vaddr - region->vaddr); in __iommu_hva2iova() 100 static int vfio_iommu_map(struct iommu *iommu, struct dma_region *region) in vfio_iommu_map() argument 105 .vaddr = (u64)region->vaddr, in vfio_iommu_map() 106 .iova = region->iova, in vfio_iommu_map() 107 .size = region->size, in vfio_iommu_map() 116 static int iommufd_map(struct iommu *iommu, struct dma_region *region) in iommufd_map() argument [all …]
|
| /linux/Documentation/networking/devlink/ |
| H A D | iosm.rst | 61 data required for flashing. The data like region count and address of each region 109 - The summary of exception details logged as part of this region. 111 - This region contains the details related to the exception occurred in the 114 - This region contains the logs related to the modem CDD driver. 116 - This region contains the eeprom logs. 118 - This region contains the current instance of bootloader logs. 120 - This region contains the previous instance of bootloader logs. 126 $ devlink region show 128 $ devlink region new pci/0000:02:00.0/report.json 130 $ devlink region dump pci/0000:02:00.0/report.json snapshot 0 [all …]
|
| H A D | devlink-region.rst | 11 region can then be accessed via the devlink region interface. 17 Snapshot identifiers are scoped to the devlink instance, not a region. 21 The major benefit to creating a region is to provide access to internal 29 requested snapshots must implement the ``.snapshot`` callback for the region 38 region should implement the ``.read`` callback in the ``devlink_region_ops`` 48 $ devlink region help 49 $ devlink region show [ DEV/REGION ] 50 $ devlink region del DEV/REGION snapshot SNAPSHOT_ID 51 $ devlink region dump DEV/REGION [ snapshot SNAPSHOT_ID ] 52 $ devlink region read DEV/REGION [ snapshot SNAPSHOT_ID ] address ADDRESS length LENGTH [all …]
|
| /linux/drivers/fpga/tests/ |
| H A D | fpga-region-test.c | 32 struct fpga_region *region; member 90 static int fake_region_get_bridges(struct fpga_region *region) in fake_region_get_bridges() argument 92 struct fpga_bridge *bridge = region->priv; in fake_region_get_bridges() 94 return fpga_bridge_get_to_list(bridge->dev.parent, region->info, ®ion->bridge_list); in fake_region_get_bridges() 105 struct fpga_region *region; in fpga_region_test_class_find() local 107 region = fpga_region_class_find(NULL, ctx->region_dev, fake_region_match); in fpga_region_test_class_find() 108 KUNIT_EXPECT_PTR_EQ(test, region, ctx->region); in fpga_region_test_class_find() 110 put_device(®ion->dev); in fpga_region_test_class_find() 133 ctx->region->info = img_info; in fpga_region_test_program_fpga() 134 ret = fpga_region_program_fpga(ctx->region); in fpga_region_test_program_fpga() [all …]
|
| /linux/drivers/virt/acrn/ |
| H A D | mm.c | 19 static int modify_region(struct acrn_vm *vm, struct vm_memory_region_op *region) in modify_region() argument 30 regions->regions_gpa = virt_to_phys(region); in modify_region() 55 struct vm_memory_region_op *region; in acrn_mm_region_add() local 58 region = kzalloc_obj(*region); in acrn_mm_region_add() 59 if (!region) in acrn_mm_region_add() 62 region->type = ACRN_MEM_REGION_ADD; in acrn_mm_region_add() 63 region->user_vm_pa = user_gpa; in acrn_mm_region_add() 64 region->service_vm_pa = service_gpa; in acrn_mm_region_add() 65 region->size = size; in acrn_mm_region_add() 66 region->attr = ((mem_type & ACRN_MEM_TYPE_MASK) | in acrn_mm_region_add() [all …]
|