Home
last modified time | relevance | path

Searched full:migrate (Results 1 – 25 of 273) sorted by relevance

1234567891011

/linux/mm/
H A Dmigrate_device.c9 #include <linux/migrate.h>
25 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_skip() local
29 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip()
30 migrate->src[migrate->npages++] = 0; in migrate_vma_collect_skip()
41 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_hole() local
49 (migrate->flags & MIGRATE_VMA_SELECT_COMPOUND) && in migrate_vma_collect_hole()
52 migrate->src[migrate->npages] = MIGRATE_PFN_MIGRATE | in migrate_vma_collect_hole()
54 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole()
55 migrate->npages++; in migrate_vma_collect_hole()
56 migrate->cpages++; in migrate_vma_collect_hole()
[all …]
H A Dmigrate.c3 * Memory Migration functionality - linux/mm/migrate.c
16 #include <linux/migrate.h>
49 #include <trace/events/migrate.h>
211 * migrate_movable_ops_page - migrate an isolated movable_ops page
216 * Migrate an isolated movable_ops page.
801 * For memory tiering mode, when migrate between slow and fast in folio_migrate_flags()
882 * @dst: The folio to migrate the data to.
884 * @mode: How to migrate the page.
886 * Common logic to directly migrate a single LRU folio suitable for
1007 * @dst: The folio to migrate to.
[all …]
H A Dcompaction.c13 #include <linux/migrate.h>
346 * should be skipped for page isolation when the migrate and free page scanner
374 /* Update the migrate PFN */ in __reset_isolation_suitable()
820 * isolate_migratepages_block() - isolate all migrate-able pages within
891 * to failure. We should migrate the pages now and in isolate_migratepages_block()
1048 * It's possible to migrate LRU and non-lru movable pages. in isolate_migratepages_block()
1088 * Only allow to migrate anonymous pages in GFP_NOFS context in isolate_migratepages_block()
1107 * it will be able to migrate without blocking - clean pages in isolate_migratepages_block()
1122 * a ->migrate_folio callback are possible to migrate in isolate_migratepages_block()
1311 * isolate_migratepages_range() - isolate migrate-able pages in a PFN range
[all …]
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_migrate.c26 #include <linux/migrate.h>
181 * @mfence: migrate fence
263 static unsigned long svm_migrate_successful_pages(struct migrate_vma *migrate) in svm_migrate_successful_pages() argument
268 for (i = 0; i < migrate->npages; i++) { in svm_migrate_successful_pages()
269 if (migrate->dst[i] & MIGRATE_PFN_VALID && in svm_migrate_successful_pages()
270 migrate->src[i] & MIGRATE_PFN_MIGRATE) in svm_migrate_successful_pages()
278 struct migrate_vma *migrate, struct dma_fence **mfence, in svm_migrate_copy_to_vram() argument
281 u64 npages = migrate->npages; in svm_migrate_copy_to_vram()
299 for (i = j = 0; (i < npages) && (mpages < migrate->cpages); i++) { in svm_migrate_copy_to_vram()
302 if (migrate->src[i] & MIGRATE_PFN_MIGRATE) { in svm_migrate_copy_to_vram()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_pagemap.c8 #include <linux/migrate.h>
19 * to migrate memory back and forth between device memory and system RAM and
28 * population requests and after that migrate all device pages to system ram.
40 * and it can evict the range to system before trying to migrate. Ideally an
41 * implementation would just try to migrate the missing part of the range and
54 * migrate to / from device memory.
167 * @migrate_pfn: Array of migrate page frame numbers
207 * @migrate_pfn: Array of migrate page frame numbers to map
293 * drm_pagemap_migrate_to_devmem() - Migrate a struct mm_struct range to device memory
294 * @devmem_allocation: The device memory allocation to migrate to.
[all …]
/linux/tools/testing/selftests/mm/
H A Dmigration.c66 int migrate(uint64_t *ptr, int n1, int n2) in migrate() function
90 printf("Didn't migrate %d pages\n", ret); in migrate()
93 perror("Couldn't migrate pages"); in migrate()
141 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
176 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
206 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
248 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
273 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
308 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
H A Drmap.c154 FIXTURE(migrate) in FIXTURE() argument
159 FIXTURE_SETUP(migrate) in FIXTURE_SETUP() argument
195 FIXTURE_TEARDOWN(migrate) in FIXTURE_TEARDOWN() argument
257 /* migrate successfully */ in try_to_move_page()
311 TEST_F(migrate, anon) in TEST_F() argument
328 TEST_F(migrate, shm) in TEST_F() argument
353 TEST_F(migrate, file) in TEST_F() argument
410 TEST_F(migrate, ksm) in TEST_F() argument
H A Dhmm-tests.c547 int ret, use_thp, migrate; in TEST_F() local
549 for (migrate = 0; migrate < 2; ++migrate) { in TEST_F()
586 if (migrate) { in TEST_F()
612 if (!migrate) { in TEST_F()
628 if (!migrate) { in TEST_F()
992 * Migrate anonymous memory to device private memory.
994 TEST_F(hmm, migrate) in TEST_F() argument
1025 /* Migrate memory to device. */ in TEST_F()
1038 * Migrate anonymous memory to device private memory and fault some of it back
1073 /* Migrate memory to device. */ in TEST_F()
[all …]
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dmigrate_reuseport.c3 * Check if we can migrate child sockets.
8 * 4. update a map to migrate all child sockets
11 * and migrate the requests in the accept queue
15 * and migrate the requests in the accept queue
19 * and migrate the requests in the accept queue
355 /* Migrate TCP_ESTABLISHED and TCP_SYN_RECV requests in migrate_dance()
364 /* No dance for TCP_NEW_SYN_RECV to migrate based on eBPF */ in migrate_dance()
385 /* Migrate from the last listener to the second one. in migrate_dance()
399 /* Migrate back to the last one from the second one */ in migrate_dance()
504 /* Migrate the requests in the accept queue only. in run_test()
[all …]
/linux/include/linux/
H A Dmigrate.h31 * If the driver cannot migrate the page at the moment, it can return
110 return -EAGAIN; /* can't migrate now */ in migrate_misplaced_folio_prepare()
114 return -EAGAIN; /* can't migrate now */ in migrate_misplaced_folio()
179 * Set to vmf->page if this is being called to migrate a page as part of
186 void migrate_vma_pages(struct migrate_vma *migrate);
187 void migrate_vma_finalize(struct migrate_vma *migrate);
H A Dpsci.h29 int (*migrate)(unsigned long cpuid); member
41 u32 migrate; member
/linux/tools/testing/selftests/net/tcp_ao/
H A Drestore.c94 try_server_run("TCP-AO migrate to another socket (server)", port++, in server_fn()
128 test_fail("pre-migrate verify failed"); in test_get_sk_checkpoint()
169 test_ok("%s: post-migrate connection is broken", tst_name); in test_sk_restore()
171 test_fail("%s: post-migrate connection is working", tst_name); in test_sk_restore()
174 test_fail("%s: post-migrate connection is working", tst_name); in test_sk_restore()
176 test_ok("%s: post-migrate connection is alive", tst_name); in test_sk_restore()
205 test_sk_restore("TCP-AO migrate to another socket (client)", port++, in client_fn()
/linux/drivers/gpu/drm/xe/
H A Dxe_migrate.c44 * struct xe_migrate - migrate context.
127 * migrate offset in xe_migrate_vram_ofs()
370 * migrate jobs servicing the faults gets stuck behind the job that faulted.
395 * xe_migrate_alloc - Allocate a migrate struct for a given &xe_tile
430 * xe_migrate_init() - Initialize a migrate context
504 drm_dbg(&xe->drm, "Migrate min chunk size is 0x%08llx\n", in xe_migrate_init()
1042 * xe_migrate_lrc() - Get the LRC from migrate context.
1043 * @migrate: Migrate context.
1047 struct xe_lrc *xe_migrate_lrc(struct xe_migrate *migrate) in xe_migrate_lrc() argument
1049 return migrate->q->lrc[0]; in xe_migrate_lrc()
[all …]
H A Dxe_bo.c173 return tile->migrate; in mem_type_to_migrate()
849 struct xe_migrate *migrate = NULL; in xe_bo_move() local
959 migrate = bo->tile->migrate; in xe_bo_move()
961 migrate = mem_type_to_migrate(xe, new_mem->mem_type); in xe_bo_move()
963 migrate = mem_type_to_migrate(xe, old_mem_type); in xe_bo_move()
965 migrate = xe->tiles[0].migrate; in xe_bo_move()
967 xe_assert(xe, migrate); in xe_bo_move()
988 fence = xe_migrate_clear(migrate, bo, new_mem, flags); in xe_bo_move()
990 fence = xe_migrate_copy(migrate, bo, bo, old_mem, new_mem, in xe_bo_move()
1264 struct xe_migrate *migrate; in xe_bo_evict_pinned_copy() local
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_dmem.c42 #include <linux/migrate.h>
83 struct nouveau_dmem_migrate migrate; member
171 if (drm->dmem->migrate.copy_func(drm, folio_nr_pages(sfolio), in nouveau_dmem_copy_folio()
232 * FIXME what we really want is to find some heuristic to migrate more in nouveau_dmem_migrate_to_ram()
267 nouveau_fence_new(&fence, dmem->migrate.chan); in nouveau_dmem_migrate_to_ram()
517 nouveau_fence_new(&fence, chunk->drm->dmem->migrate.chan); in nouveau_dmem_evict_chunk()
558 struct nvif_push *push = &drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_copy()
631 struct nvif_push *push = &drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_clear()
691 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy; in nouveau_dmem_migrate_init()
692 drm->dmem->migrate.clear_func = nvc0b5_migrate_clear; in nouveau_dmem_migrate_init()
[all …]
/linux/tools/perf/pmu-events/arch/arm64/hisilicon/hip08/
H A Duncore-hha.json148 "EventName": "sdir-home-migrate",
149 "BriefDescription": "Count of the number of HHA S-Dir read home migrate operations",
154 "EventName": "edir-home-migrate",
155 "BriefDescription": "Count of the number of HHA E-Dir read home migrate operations",
/linux/Documentation/gpu/rfc/
H A Dgpusvm.rst21 migrate can actually migrate, with only the faulting page guaranteed
22 to migrate).
110 migrate device layer are a performance bottleneck, having compound
/linux/include/uapi/linux/
H A Dmempolicy.h52 #define MPOL_MF_LAZY (1<<3) /* UNSUPPORTED FLAG: Lazy migrate on fault */
65 #define MPOL_F_MOF (1 << 3) /* this policy wants migrate on fault */
66 #define MPOL_F_MORON (1 << 4) /* Migrate On protnone Reference On Node */
/linux/Documentation/gpu/xe/
H A Dxe_migrate.rst4 Migrate Layer
8 :doc: Migrate Layer
/linux/tools/testing/selftests/kvm/x86/
H A Dxapic_ipi_test.c306 * migrate_pages with PID=0 will migrate all pages of this in do_migrations()
359 bool *migrate, int *delay_usecs) in get_cmdline_args() argument
371 *migrate = true; in get_cmdline_args()
397 bool migrate = false; in main() local
403 get_cmdline_args(argc, argv, &run_secs, &migrate, &delay_usecs); in main()
457 if (!migrate) in main()
/linux/fs/ocfs2/dlm/
H A Ddlmmaster.c824 * of the MIGRATION mle: either the migrate finished or in dlm_get_lock_resource()
1506 mlog(0, "node %u is master, but trying to migrate to " in dlm_master_request_handler()
1510 "node is trying to migrate it to %u?!\n", in dlm_master_request_handler()
2594 mlog(ML_ERROR, "tried to migrate %.*s to %u, but " in dlm_migrate_lockres()
2766 /* Wheee! Migrate lockres here! Will sleep so drop spinlock. */ in dlm_empty_lockres()
2771 mlog(0, "%s: res %.*s, Migrate to node %u failed with %d\n", in dlm_empty_lockres()
2961 * Pick a node to migrate the lock resource to. This function selects a
3016 struct dlm_migrate_request migrate; in dlm_do_migrate_request() local
3020 memset(&migrate, 0, sizeof(migrate)); in dlm_do_migrate_request()
3021 migrate.namelen = res->lockname.len; in dlm_do_migrate_request()
[all …]
/linux/kernel/irq/
H A Dcpuhotplug.c68 pr_debug("IRQ %u: Unable to migrate away\n", d->irq); in migrate_one_irq()
162 * irq_migrate_all_off_this_cpu - Migrate irqs away from offline cpu
164 * The current CPU has been marked offline. Migrate IRQs off this CPU.
/linux/tools/testing/selftests/bpf/progs/
H A Dtest_migrate_reuseport.c3 * Check if we can migrate child sockets.
97 SEC("sk_reuseport/migrate")
/linux/arch/arm/kernel/
H A Dpsci_smp.c37 * migrate Migrate the context to a different CPU
/linux/tools/testing/selftests/kvm/
H A Darch_timer.c17 * even more, an option to migrate the vCPUs across pCPUs (-m), at a
109 "Failed to migrate the vCPU:%u to pCPU: %u; ret: %d", in test_migrate_vcpu()
186 …pr_info("\t-m: Frequency (in ms) of vCPUs to migrate to different pCPU. 0 to turn off (default: %u… in test_print_help()

1234567891011