| /linux/mm/ |
| H A D | migrate_device.c | 9 #include <linux/migrate.h> 25 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_skip() local 29 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip() 30 migrate->src[migrate->npages++] = 0; in migrate_vma_collect_skip() 41 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_hole() local 49 (migrate->flags & MIGRATE_VMA_SELECT_COMPOUND) && in migrate_vma_collect_hole() 52 migrate->src[migrate->npages] = MIGRATE_PFN_MIGRATE | in migrate_vma_collect_hole() 54 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole() 55 migrate->npages++; in migrate_vma_collect_hole() 56 migrate->cpages++; in migrate_vma_collect_hole() [all …]
|
| H A D | migrate.c | 3 * Memory Migration functionality - linux/mm/migrate.c 16 #include <linux/migrate.h> 49 #include <trace/events/migrate.h> 211 * migrate_movable_ops_page - migrate an isolated movable_ops page 216 * Migrate an isolated movable_ops page. 802 * For memory tiering mode, when migrate between slow and fast in folio_migrate_flags() 883 * @dst: The folio to migrate the data to. 885 * @mode: How to migrate the page. 887 * Common logic to directly migrate a single LRU folio suitable for 1008 * @dst: The folio to migrate to. [all …]
|
| H A D | compaction.c | 13 #include <linux/migrate.h> 346 * should be skipped for page isolation when the migrate and free page scanner 374 /* Update the migrate PFN */ in __reset_isolation_suitable() 820 * isolate_migratepages_block() - isolate all migrate-able pages within 891 * to failure. We should migrate the pages now and in isolate_migratepages_block() 1048 * It's possible to migrate LRU and non-lru movable pages. in isolate_migratepages_block() 1088 * Only allow to migrate anonymous pages in GFP_NOFS context in isolate_migratepages_block() 1107 * it will be able to migrate without blocking - clean pages in isolate_migratepages_block() 1122 * a ->migrate_folio callback are possible to migrate in isolate_migratepages_block() 1311 * isolate_migratepages_range() - isolate migrate-able pages in a PFN range [all …]
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | selftest_migrate.c | 36 static int copy(struct intel_migrate *migrate, in copy() argument 37 int (*fn)(struct intel_migrate *migrate, in copy() argument 44 struct drm_i915_private *i915 = migrate->context->engine->i915; in copy() 90 err = fn(migrate, &ww, src, dst, &rq); in copy() 151 GEM_BUG_ON(ce->vm != ce->engine->gt->migrate.context->vm); in intel_context_copy_ccs() 255 static int clear(struct intel_migrate *migrate, in clear() argument 256 int (*fn)(struct intel_migrate *migrate, in clear() argument 263 struct drm_i915_private *i915 = migrate->context->engine->i915; in clear() 301 err = intel_migrate_ccs_copy(migrate, &ww, NULL, in clear() 318 err = fn(migrate, &ww, obj, val, &rq); in clear() [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_pagemap.c | 8 #include <linux/migrate.h> 21 * to migrate memory back and forth between device memory and system RAM and 30 * population requests and after that migrate all device pages to system ram. 42 * and it can evict the range to system before trying to migrate. Ideally an 43 * implementation would just try to migrate the missing part of the range and 56 * migrate to / from device memory. 171 * @migrate_pfn: Array of migrate page frame numbers 293 * @migrate_pfn: Array of migrate pfns set up for the mapped pages. Used to 448 * drm_pagemap_migrate_to_devmem() - Migrate a struct mm_struct range to device memory 449 * @devmem_allocation: The device memory allocation to migrate to. [all …]
|
| /linux/tools/testing/selftests/mm/ |
| H A D | migration.c | 66 int migrate(uint64_t *ptr, int n1, int n2) in migrate() function 90 printf("Didn't migrate %d pages\n", ret); in migrate() 93 perror("Couldn't migrate pages"); in migrate() 141 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 176 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 206 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 248 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 273 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0); 308 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
|
| H A D | rmap.c | 154 FIXTURE(migrate) in FIXTURE() argument 159 FIXTURE_SETUP(migrate) in FIXTURE_SETUP() argument 195 FIXTURE_TEARDOWN(migrate) in FIXTURE_TEARDOWN() argument 257 /* migrate successfully */ in try_to_move_page() 311 TEST_F(migrate, anon) in TEST_F() argument 328 TEST_F(migrate, shm) in TEST_F() argument 353 TEST_F(migrate, file) in TEST_F() argument 410 TEST_F(migrate, ksm) in TEST_F() argument
|
| H A D | hmm-tests.c | 547 int ret, use_thp, migrate; in TEST_F() local 549 for (migrate = 0; migrate < 2; ++migrate) { in TEST_F() 586 if (migrate) { in TEST_F() 612 if (!migrate) { in TEST_F() 628 if (!migrate) { in TEST_F() 992 * Migrate anonymous memory to device private memory. 994 TEST_F(hmm, migrate) in TEST_F() argument 1025 /* Migrate memory to device. */ in TEST_F() 1038 * Migrate anonymous memory to device private memory and fault some of it back 1073 /* Migrate memory to device. */ in TEST_F() [all …]
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | migrate_reuseport.c | 3 * Check if we can migrate child sockets. 8 * 4. update a map to migrate all child sockets 11 * and migrate the requests in the accept queue 15 * and migrate the requests in the accept queue 19 * and migrate the requests in the accept queue 355 /* Migrate TCP_ESTABLISHED and TCP_SYN_RECV requests in migrate_dance() 364 /* No dance for TCP_NEW_SYN_RECV to migrate based on eBPF */ in migrate_dance() 385 /* Migrate from the last listener to the second one. in migrate_dance() 399 /* Migrate back to the last one from the second one */ in migrate_dance() 504 /* Migrate the requests in the accept queue only. in run_test() [all …]
|
| /linux/include/linux/ |
| H A D | migrate.h | 31 * If the driver cannot migrate the page at the moment, it can return 118 return -EAGAIN; /* can't migrate now */ in migrate_misplaced_folio_prepare() 122 return -EAGAIN; /* can't migrate now */ in migrate_misplaced_folio() 187 * Set to vmf->page if this is being called to migrate a page as part of 194 void migrate_vma_pages(struct migrate_vma *migrate); 195 void migrate_vma_finalize(struct migrate_vma *migrate);
|
| H A D | psci.h | 29 int (*migrate)(unsigned long cpuid); member 41 u32 migrate; member
|
| /linux/tools/testing/selftests/net/tcp_ao/ |
| H A D | restore.c | 94 try_server_run("TCP-AO migrate to another socket (server)", port++, in server_fn() 128 test_fail("pre-migrate verify failed"); in test_get_sk_checkpoint() 169 test_ok("%s: post-migrate connection is broken", tst_name); in test_sk_restore() 171 test_fail("%s: post-migrate connection is working", tst_name); in test_sk_restore() 174 test_fail("%s: post-migrate connection is working", tst_name); in test_sk_restore() 176 test_ok("%s: post-migrate connection is alive", tst_name); in test_sk_restore() 205 test_sk_restore("TCP-AO migrate to another socket (client)", port++, in client_fn()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_migrate.c | 46 * struct xe_migrate - migrate context. 129 * migrate offset in xe_migrate_vram_ofs() 372 * migrate jobs servicing the faults gets stuck behind the job that faulted. 397 * xe_migrate_alloc - Allocate a migrate struct for a given &xe_tile 432 * xe_migrate_init() - Initialize a migrate context 507 drm_dbg(&xe->drm, "Migrate min chunk size is 0x%08llx\n", in xe_migrate_init() 1045 * xe_migrate_lrc() - Get the LRC from migrate context. 1046 * @migrate: Migrate context. 1050 struct xe_lrc *xe_migrate_lrc(struct xe_migrate *migrate) in xe_migrate_lrc() argument 1052 return migrate->q->lrc[0]; in xe_migrate_lrc() [all …]
|
| H A D | xe_bo.c | 173 return tile->migrate; in mem_type_to_migrate() 848 struct xe_migrate *migrate = NULL; in xe_bo_move() local 958 migrate = bo->tile->migrate; in xe_bo_move() 960 migrate = mem_type_to_migrate(xe, new_mem->mem_type); in xe_bo_move() 962 migrate = mem_type_to_migrate(xe, old_mem_type); in xe_bo_move() 964 migrate = xe->tiles[0].migrate; in xe_bo_move() 966 xe_assert(xe, migrate); in xe_bo_move() 987 fence = xe_migrate_clear(migrate, bo, new_mem, flags); in xe_bo_move() 989 fence = xe_migrate_copy(migrate, bo, bo, old_mem, new_mem, in xe_bo_move() 1268 struct xe_migrate *migrate; in xe_bo_evict_pinned_copy() local [all …]
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_dmem.c | 42 #include <linux/migrate.h> 83 struct nouveau_dmem_migrate migrate; member 171 if (drm->dmem->migrate.copy_func(drm, folio_nr_pages(sfolio), in nouveau_dmem_copy_folio() 232 * FIXME what we really want is to find some heuristic to migrate more in nouveau_dmem_migrate_to_ram() 267 nouveau_fence_new(&fence, dmem->migrate.chan); in nouveau_dmem_migrate_to_ram() 517 nouveau_fence_new(&fence, chunk->drm->dmem->migrate.chan); in nouveau_dmem_evict_chunk() 558 struct nvif_push *push = &drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_copy() 631 struct nvif_push *push = &drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_clear() 691 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy; in nouveau_dmem_migrate_init() 692 drm->dmem->migrate.clear_func = nvc0b5_migrate_clear; in nouveau_dmem_migrate_init() [all …]
|
| /linux/tools/perf/pmu-events/arch/arm64/hisilicon/hip08/ |
| H A D | uncore-hha.json | 148 "EventName": "sdir-home-migrate", 149 "BriefDescription": "Count of the number of HHA S-Dir read home migrate operations", 154 "EventName": "edir-home-migrate", 155 "BriefDescription": "Count of the number of HHA E-Dir read home migrate operations",
|
| /linux/Documentation/gpu/rfc/ |
| H A D | gpusvm.rst | 21 migrate can actually migrate, with only the faulting page guaranteed 22 to migrate). 110 migrate device layer are a performance bottleneck, having compound
|
| /linux/include/uapi/linux/ |
| H A D | mempolicy.h | 55 #define MPOL_MF_LAZY (1<<3) /* UNSUPPORTED FLAG: Lazy migrate on fault */ 68 #define MPOL_F_MOF (1 << 3) /* this policy wants migrate on fault */ 69 #define MPOL_F_MORON (1 << 4) /* Migrate On protnone Reference On Node */
|
| /linux/Documentation/gpu/xe/ |
| H A D | xe_migrate.rst | 4 Migrate Layer 8 :doc: Migrate Layer
|
| /linux/scripts/ |
| H A D | generate_rust_analyzer.py | 79 # > 0e071c2c6a58 ("Migrate core to Rust 2024") 80 # > f505d4e8e380 ("Migrate alloc to Rust 2024") 81 # > 0b2489c226c3 ("Migrate proc_macro to Rust 2024") 82 # > 993359e70112 ("Migrate std to Rust 2024")
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | xapic_ipi_test.c | 306 * migrate_pages with PID=0 will migrate all pages of this in do_migrations() 359 bool *migrate, int *delay_usecs) in get_cmdline_args() argument 371 *migrate = true; in get_cmdline_args() 397 bool migrate = false; in main() local 403 get_cmdline_args(argc, argv, &run_secs, &migrate, &delay_usecs); in main() 457 if (!migrate) in main()
|
| /linux/fs/ocfs2/dlm/ |
| H A D | dlmmaster.c | 824 * of the MIGRATION mle: either the migrate finished or in dlm_get_lock_resource() 1506 mlog(0, "node %u is master, but trying to migrate to " in dlm_master_request_handler() 1510 "node is trying to migrate it to %u?!\n", in dlm_master_request_handler() 2594 mlog(ML_ERROR, "tried to migrate %.*s to %u, but " in dlm_migrate_lockres() 2766 /* Wheee! Migrate lockres here! Will sleep so drop spinlock. */ in dlm_empty_lockres() 2771 mlog(0, "%s: res %.*s, Migrate to node %u failed with %d\n", in dlm_empty_lockres() 2961 * Pick a node to migrate the lock resource to. This function selects a 3016 struct dlm_migrate_request migrate; in dlm_do_migrate_request() local 3020 memset(&migrate, 0, sizeof(migrate)); in dlm_do_migrate_request() 3021 migrate.namelen = res->lockname.len; in dlm_do_migrate_request() [all …]
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_ttm_move.c | 202 if (!to_gt(i915)->migrate.context || intel_gt_is_wedged(to_gt(i915))) in i915_ttm_accel_move() 215 intel_engine_pm_get(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move() 216 ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, deps, in i915_ttm_accel_move() 229 intel_engine_pm_get(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move() 230 ret = intel_context_migrate_copy(to_gt(i915)->migrate.context, in i915_ttm_accel_move() 242 intel_engine_pm_put(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move() 661 ret = -EIO; /* never feed non-migrate fences into ttm */ in i915_ttm_move()
|
| /linux/kernel/irq/ |
| H A D | cpuhotplug.c | 68 pr_debug("IRQ %u: Unable to migrate away\n", d->irq); in migrate_one_irq() 162 * irq_migrate_all_off_this_cpu - Migrate irqs away from offline cpu 164 * The current CPU has been marked offline. Migrate IRQs off this CPU.
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_migrate_reuseport.c | 3 * Check if we can migrate child sockets. 97 SEC("sk_reuseport/migrate")
|