| /linux/mm/ |
| H A D | migrate_device.c | 9 #include <linux/migrate.h> 25 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_skip() local 29 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip() 30 migrate->src[migrate->npages++] = 0; in migrate_vma_collect_skip() 41 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_hole() local 49 (migrate->flags & MIGRATE_VMA_SELECT_COMPOUND) && in migrate_vma_collect_hole() 52 migrate->src[migrate in migrate_vma_collect_hole() 142 struct migrate_vma *migrate = walk->private; migrate_vma_collect_huge_pmd() local 252 struct migrate_vma *migrate = walk->private; migrate_vma_collect_pmd() local 508 migrate_vma_collect(struct migrate_vma * migrate) migrate_vma_collect() argument 665 migrate_vma_unmap(struct migrate_vma * migrate) migrate_vma_unmap() argument 793 migrate_vma_insert_huge_pmd_page(struct migrate_vma * migrate,unsigned long addr,struct page * page,unsigned long * src,pmd_t * pmdp) migrate_vma_insert_huge_pmd_page() argument 908 migrate_vma_split_unmapped_folio(struct migrate_vma * migrate,unsigned long idx,unsigned long addr,struct folio * folio) migrate_vma_split_unmapped_folio() argument 930 migrate_vma_insert_huge_pmd_page(struct migrate_vma * migrate,unsigned long addr,struct page * page,unsigned long * src,pmd_t * pmdp) migrate_vma_insert_huge_pmd_page() argument 939 migrate_vma_split_unmapped_folio(struct migrate_vma * migrate,unsigned long idx,unsigned long addr,struct folio * folio) migrate_vma_split_unmapped_folio() argument 968 migrate_vma_insert_page(struct migrate_vma * migrate,unsigned long addr,unsigned long * dst,unsigned long * src) migrate_vma_insert_page() argument 1103 __migrate_device_pages(unsigned long * src_pfns,unsigned long * dst_pfns,unsigned long npages,struct migrate_vma * migrate) __migrate_device_pages() argument 1265 migrate_vma_pages(struct migrate_vma * migrate) migrate_vma_pages() argument 1351 migrate_vma_finalize(struct migrate_vma * migrate) migrate_vma_finalize() argument [all...] |
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | selftest_migrate.c | 36 static int copy(struct intel_migrate *migrate, in copy() argument 37 int (*fn)(struct intel_migrate *migrate, in copy() argument 44 struct drm_i915_private *i915 = migrate->context->engine->i915; in copy() 90 err = fn(migrate, &ww, src, dst, &rq); in copy() 151 GEM_BUG_ON(ce->vm != ce->engine->gt->migrate.context->vm); in intel_context_copy_ccs() 255 static int clear(struct intel_migrate *migrate, in clear() argument 256 int (*fn)(struct intel_migrate *migrate, in clear() argument 263 struct drm_i915_private *i915 = migrate->context->engine->i915; in clear() 301 err = intel_migrate_ccs_copy(migrate, &ww, NULL, in clear() 318 err = fn(migrate, &ww, obj, val, &rq); in clear() [all …]
|
| H A D | intel_gt_types.h | 237 struct intel_migrate migrate; member
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_pagemap.c | 479 struct migrate_vma migrate = { in drm_pagemap_migrate_to_devmem() local 528 buf = kvcalloc(npages, 2 * sizeof(*migrate.src) + sizeof(*pagemap_addr) + in drm_pagemap_migrate_to_devmem() 534 pagemap_addr = buf + (2 * sizeof(*migrate.src) * npages); in drm_pagemap_migrate_to_devmem() 535 pages = buf + (2 * sizeof(*migrate.src) + sizeof(*pagemap_addr)) * npages; in drm_pagemap_migrate_to_devmem() 545 migrate.vma = vas; in drm_pagemap_migrate_to_devmem() 546 migrate.src = buf; in drm_pagemap_migrate_to_devmem() 547 migrate.dst = migrate.src + npages; in drm_pagemap_migrate_to_devmem() 549 err = migrate_vma_setup(&migrate); in drm_pagemap_migrate_to_devmem() 553 if (!migrate.cpages) { in drm_pagemap_migrate_to_devmem() 559 if (migrate.cpages != npages) { in drm_pagemap_migrate_to_devmem() [all …]
|
| /linux/tools/testing/selftests/mm/ |
| H A D | migration.c | 66 int migrate(uint64_t *ptr, int n1, int n2) in migrate() function 141 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 176 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 206 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 248 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 273 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 308 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
|
| H A D | rmap.c | 154 FIXTURE(migrate) in FIXTURE() argument 159 FIXTURE_SETUP(migrate) in FIXTURE_SETUP() argument 195 FIXTURE_TEARDOWN(migrate) in FIXTURE_TEARDOWN() argument 311 TEST_F(migrate, anon) in TEST_F() argument 328 TEST_F(migrate, shm) in TEST_F() argument 353 TEST_F(migrate, file) in TEST_F() argument 410 TEST_F(migrate, ksm) in TEST_F() argument
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_dmem.c | 83 struct nouveau_dmem_migrate migrate; member 171 if (drm->dmem->migrate.copy_func(drm, folio_nr_pages(sfolio), in nouveau_dmem_copy_folio() 267 nouveau_fence_new(&fence, dmem->migrate.chan); in nouveau_dmem_migrate_to_ram() 517 nouveau_fence_new(&fence, chunk->drm->dmem->migrate.chan); in nouveau_dmem_evict_chunk() 558 struct nvif_push *push = &drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_copy() 631 struct nvif_push *push = &drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_clear() 691 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy; in nouveau_dmem_migrate_init() 692 drm->dmem->migrate.clear_func = nvc0b5_migrate_clear; in nouveau_dmem_migrate_init() 693 drm->dmem->migrate.chan = drm->ttm.chan; in nouveau_dmem_migrate_init() 753 if (drm->dmem->migrate.copy_func(drm, folio_nr_pages(page_folio(spage)), in nouveau_dmem_migrate_copy_one() [all …]
|
| /linux/include/linux/ |
| H A D | psci.h | 29 int (*migrate)(unsigned long cpuid); member 41 u32 migrate; member
|
| H A D | migrate.h | 186 void migrate_vma_pages(struct migrate_vma *migrate); 187 void migrate_vma_finalize(struct migrate_vma *migrate);
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | xapic_ipi_test.c | 359 bool *migrate, int *delay_usecs) in get_cmdline_args() argument 371 *migrate = true; in get_cmdline_args() 397 bool migrate = false; in main() local 403 get_cmdline_args(argc, argv, &run_secs, &migrate, &delay_usecs); in main() 457 if (!migrate) in main()
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | vas.c | 652 bool migrate) in reconfig_open_windows() argument 680 if ((vcaps->nr_close_wins > creds) && !migrate) in reconfig_open_windows() 694 if (migrate) in reconfig_open_windows() 762 bool migrate) in reconfig_close_windows() argument 769 if (migrate) in reconfig_close_windows() 827 if (rc && !migrate) in reconfig_close_windows() 839 if (!migrate && !--excess_creds) in reconfig_close_windows()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_ttm_move.c | 202 if (!to_gt(i915)->migrate.context || intel_gt_is_wedged(to_gt(i915))) in i915_ttm_accel_move() 215 intel_engine_pm_get(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move() 216 ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, deps, in i915_ttm_accel_move() 229 intel_engine_pm_get(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move() 230 ret = intel_context_migrate_copy(to_gt(i915)->migrate.context, in i915_ttm_accel_move() 242 intel_engine_pm_put(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_migrate.h | 142 struct xe_lrc *xe_migrate_lrc(struct xe_migrate *migrate); 143 struct xe_exec_queue *xe_migrate_exec_queue(struct xe_migrate *migrate);
|
| H A D | xe_bo.c | 173 return tile->migrate; in mem_type_to_migrate() 848 struct xe_migrate *migrate = NULL; in xe_bo_move() local 958 migrate = bo->tile->migrate; in xe_bo_move() 960 migrate = mem_type_to_migrate(xe, new_mem->mem_type); in xe_bo_move() 962 migrate = mem_type_to_migrate(xe, old_mem_type); in xe_bo_move() 964 migrate = xe->tiles[0].migrate; in xe_bo_move() 966 xe_assert(xe, migrate); in xe_bo_move() 987 fence = xe_migrate_clear(migrate, bo, new_mem, flags); in xe_bo_move() 989 fence = xe_migrate_copy(migrate, bo, bo, old_mem, new_mem, in xe_bo_move() 1268 struct xe_migrate *migrate; in xe_bo_evict_pinned_copy() local [all …]
|
| H A D | xe_svm.c | 575 gt = xe_migrate_exec_queue(vr->migrate)->gt; in xe_svm_copy() 623 __fence = xe_migrate_from_vram(vr->migrate, in xe_svm_copy() 633 __fence = xe_migrate_to_vram(vr->migrate, in xe_svm_copy() 664 __fence = xe_migrate_from_vram(vr->migrate, 1, in xe_svm_copy() 672 __fence = xe_migrate_to_vram(vr->migrate, 1, in xe_svm_copy() 1155 DECL_SVM_RANGE_COUNT_STATS(migrate, MIGRATE) in DECL_SVM_RANGE_COUNT_STATS() 1180 DECL_SVM_RANGE_US_STATS(migrate, MIGRATE)
|
| /linux/Documentation/gpu/rfc/ |
| H A D | gpusvm.rst | 21 migrate can actually migrate, with only the faulting page guaranteed 22 to migrate). 110 migrate device layer are a performance bottleneck, having compound
|
| /linux/drivers/gpu/drm/i915/selftests/ |
| H A D | i915_perf_selftests.h | 20 selftest(migrate, intel_migrate_perf_selftests)
|
| H A D | i915_live_selftests.h | 30 selftest(migrate, intel_migrate_live_selftests)
|
| /linux/tools/perf/scripts/python/ |
| H A D | sched-migration.py | 191 def migrate(self, ts_list, new, old_cpu, new_cpu): member in TimeSlice 351 def migrate(self, headers, pid, prio, orig_cpu, dest_cpu): member in SchedEventProxy 353 ts.migrate(self.timeslices, pid, orig_cpu, dest_cpu) 418 parser.migrate(headers, pid, prio, orig_cpu, dest_cpu)
|
| /linux/fs/ocfs2/dlm/ |
| H A D | dlmmaster.c | 3016 struct dlm_migrate_request migrate; in dlm_do_migrate_request() local 3020 memset(&migrate, 0, sizeof(migrate)); in dlm_do_migrate_request() 3021 migrate.namelen = res->lockname.len; in dlm_do_migrate_request() 3022 memcpy(migrate.name, res->lockname.name, migrate.namelen); in dlm_do_migrate_request() 3023 migrate.new_master = new_master; in dlm_do_migrate_request() 3024 migrate.master = master; in dlm_do_migrate_request() 3044 &migrate, sizeof(migrate), nodenum, in dlm_do_migrate_request() 3049 migrate.namelen, migrate.name, ret, nodenum); in dlm_do_migrate_request() 3093 struct dlm_migrate_request *migrate = (struct dlm_migrate_request *) msg->buf; in dlm_migrate_request_handler() local 3102 name = migrate->name; in dlm_migrate_request_handler() [all …]
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | psci-relay.c | 53 is_psci_0_1(migrate, func_id)); in is_psci_0_1_call() 229 if (is_psci_0_1(cpu_off, func_id) || is_psci_0_1(migrate, func_id)) in psci_0_1_handler()
|
| /linux/fs/ext4/ |
| H A D | Makefile | 10 indirect.o inline.o inode.o ioctl.o mballoc.o migrate.o \
|
| /linux/kernel/sched/ |
| H A D | stats.h | 229 static inline void psi_enqueue(struct task_struct *p, bool migrate) {} in psi_enqueue() argument 230 static inline void psi_dequeue(struct task_struct *p, bool migrate) {} in psi_dequeue() argument
|
| /linux/Documentation/trace/ |
| H A D | hwlat_detector.rst | 81 By default, one hwlat detector's kernel thread will migrate across each CPU 87 - round-robin: migrate across each CPU specified in cpumask [default]
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-kernel-mm-numa | 17 systems to migrate pages from fast tiers to slow tiers
|