Home
last modified time | relevance | path

Searched full:migrate (Results 1 – 25 of 529) sorted by relevance

12345678910>>...22

/kernel/linux/linux-6.6/mm/
Dmigrate_device.c9 #include <linux/migrate.h>
24 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_skip() local
28 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip()
29 migrate->src[migrate->npages++] = 0; in migrate_vma_collect_skip()
40 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_hole() local
48 migrate->src[migrate->npages] = MIGRATE_PFN_MIGRATE; in migrate_vma_collect_hole()
49 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole()
50 migrate->npages++; in migrate_vma_collect_hole()
51 migrate->cpages++; in migrate_vma_collect_hole()
62 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_pmd() local
[all …]
Dmigrate.c3 * Memory Migration functionality - linux/mm/migrate.c
16 #include <linux/migrate.h>
56 #include <trace/events/migrate.h>
598 * For memory tiering mode, when migrate between slow and fast in folio_migrate_flags()
679 * @dst: The folio to migrate the data to.
681 * @mode: How to migrate the page.
683 * Common logic to directly migrate a single LRU folio suitable for
809 * @dst: The folio to migrate to.
810 * @src: The folio to migrate from.
811 * @mode: How to migrate the folio.
[all …]
Dpage_isolation.c12 #include <linux/migrate.h>
158 * We assume the caller intended to SET migrate type to isolate. in set_migratetype_isolate()
291 * @migratetype: migrate type to set in error recovery.
387 * migrate compound pages then let the free page handling code in isolate_single_pageblock()
487 * @migratetype: Migrate type to set in error recovery.
500 * future will not be allocated again. If specified range includes migrate types
577 * @migratetype: New migrate type to set on the range
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdkfd/
Dkfd_migrate.c27 #include <linux/migrate.h>
181 * @mfence: migrate fence
263 static unsigned long svm_migrate_successful_pages(struct migrate_vma *migrate) in svm_migrate_successful_pages() argument
268 for (i = 0; i < migrate->npages; i++) { in svm_migrate_successful_pages()
269 if (migrate->src[i] & MIGRATE_PFN_VALID && in svm_migrate_successful_pages()
270 migrate->src[i] & MIGRATE_PFN_MIGRATE) in svm_migrate_successful_pages()
276 static unsigned long svm_migrate_unsuccessful_pages(struct migrate_vma *migrate) in svm_migrate_unsuccessful_pages() argument
281 for (i = 0; i < migrate->npages; i++) { in svm_migrate_unsuccessful_pages()
282 if (migrate->src[i] & MIGRATE_PFN_VALID && in svm_migrate_unsuccessful_pages()
283 !(migrate->src[i] & MIGRATE_PFN_MIGRATE)) in svm_migrate_unsuccessful_pages()
[all …]
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/
Dselftest_migrate.c36 static int copy(struct intel_migrate *migrate, in copy() argument
37 int (*fn)(struct intel_migrate *migrate, in copy() argument
44 struct drm_i915_private *i915 = migrate->context->engine->i915; in copy()
90 err = fn(migrate, &ww, src, dst, &rq); in copy()
151 GEM_BUG_ON(ce->vm != ce->engine->gt->migrate.context->vm); in intel_context_copy_ccs()
255 static int clear(struct intel_migrate *migrate, in clear() argument
256 int (*fn)(struct intel_migrate *migrate, in clear() argument
263 struct drm_i915_private *i915 = migrate->context->engine->i915; in clear()
301 err = intel_migrate_ccs_copy(migrate, &ww, NULL, in clear()
318 err = fn(migrate, &ww, obj, val, &rq); in clear()
[all …]
/kernel/linux/linux-5.10/mm/
Dmigrate.c3 * Memory Migration functionality - linux/mm/migrate.c
16 #include <linux/migrate.h>
56 #include <trace/events/migrate.h>
70 * drained them. Those pages will fail to migrate like other in migrate_prep()
639 /* Migrate the page's xpm state */ in migrate_page_states()
721 * Common logic to directly migrate a single LRU page suitable for
1014 * cannot try to migrate this page. in move_to_new_page()
1183 * Obtain the lock on page, remove all ptes and migrate the page
1229 * Compaction can migrate also non-LRU pages which are in unmap_and_move()
1284 * This means that when we try to migrate hugepage whose subpages are
[all …]
Dpage_isolation.c12 #include <linux/migrate.h>
27 * We assume the caller intended to SET migrate type to isolate. in set_migratetype_isolate()
149 * @migratetype: Migrate type to set in error recovery.
160 * future will not be allocated again. If specified range includes migrate types
/kernel/linux/linux-6.6/net/xfrm/
DKconfig61 bool "Transformation migrate database"
122 bool "PF_KEY MIGRATE"
126 Add a PF_KEY MIGRATE message to PF_KEYv2 socket family.
127 The PF_KEY MIGRATE message is used to dynamically update
133 <draft-sugimoto-mip6-pfkey-migrate>.
/kernel/linux/linux-5.10/net/xfrm/
DKconfig61 bool "Transformation migrate database"
122 bool "PF_KEY MIGRATE"
126 Add a PF_KEY MIGRATE message to PF_KEYv2 socket family.
127 The PF_KEY MIGRATE message is used to dynamically update
133 <draft-sugimoto-mip6-pfkey-migrate>.
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/
Dnouveau_dmem.c42 #include <linux/migrate.h>
82 struct nouveau_dmem_migrate migrate; member
152 if (drm->dmem->migrate.copy_func(drm, 1, NOUVEAU_APER_HOST, *dma_addr, in nouveau_dmem_copy_one()
183 * FIXME what we really want is to find some heuristic to migrate more in nouveau_dmem_migrate_to_ram()
212 nouveau_fence_new(&fence, dmem->migrate.chan); in nouveau_dmem_migrate_to_ram()
405 nouveau_fence_new(&fence, chunk->drm->dmem->migrate.chan); in nouveau_dmem_evict_chunk()
446 struct nvif_push *push = drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_copy()
519 struct nvif_push *push = drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_clear()
579 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy; in nouveau_dmem_migrate_init()
580 drm->dmem->migrate.clear_func = nvc0b5_migrate_clear; in nouveau_dmem_migrate_init()
[all …]
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/prog_tests/
Dmigrate_reuseport.c3 * Check if we can migrate child sockets.
8 * 4. update a map to migrate all child sockets
11 * and migrate the requests in the accept queue
15 * and migrate the requests in the accept queue
19 * and migrate the requests in the accept queue
355 /* Migrate TCP_ESTABLISHED and TCP_SYN_RECV requests in migrate_dance()
364 /* No dance for TCP_NEW_SYN_RECV to migrate based on eBPF */ in migrate_dance()
385 /* Migrate from the last listener to the second one. in migrate_dance()
399 /* Migrate back to the last one from the second one */ in migrate_dance()
504 /* Migrate the requests in the accept queue only. in run_test()
[all …]
/kernel/linux/linux-6.6/tools/testing/selftests/mm/
Dmigration.c63 int migrate(uint64_t *ptr, int n1, int n2) in migrate() function
83 printf("Didn't migrate %d pages\n", ret); in migrate()
85 perror("Couldn't migrate pages"); in migrate()
135 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
170 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
197 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
/kernel/linux/linux-6.6/arch/um/drivers/
DKconfig168 migrate to UML_NET_VECTOR.
185 migrate to UML_NET_VECTOR.
206 migrate to UML_NET_VECTOR.
227 migrate to UML_NET_VECTOR.
273 migrate to UML_NET_VECTOR.
296 migrate to UML_NET_VECTOR.
316 migrate to UML_NET_VECTOR.
342 migrate to UML_NET_VECTOR.
/kernel/linux/linux-5.10/arch/um/drivers/
DKconfig161 migrate to UML_NET_VECTOR.
178 migrate to UML_NET_VECTOR.
199 migrate to UML_NET_VECTOR.
220 migrate to UML_NET_VECTOR.
257 migrate to UML_NET_VECTOR.
280 migrate to UML_NET_VECTOR.
299 migrate to UML_NET_VECTOR.
325 migrate to UML_NET_VECTOR.
/kernel/linux/linux-5.10/include/linux/
Dmigrate.h33 /* In mm/debug.c; also keep sync with include/trace/events/migrate.h */
117 return -EAGAIN; /* can't migrate now */ in migrate_misplaced_page()
198 * Set to vmf->page if this is being called to migrate a page as part of
205 void migrate_vma_pages(struct migrate_vma *migrate);
206 void migrate_vma_finalize(struct migrate_vma *migrate);
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/
Dnouveau_dmem.c81 struct nouveau_dmem_migrate migrate; member
164 if (drm->dmem->migrate.copy_func(drm, 1, NOUVEAU_APER_HOST, *dma_addr, in nouveau_dmem_fault_copy_one()
199 * FIXME what we really want is to find some heuristic to migrate more in nouveau_dmem_migrate_to_ram()
212 nouveau_fence_new(dmem->migrate.chan, false, &fence); in nouveau_dmem_migrate_to_ram()
399 struct nvif_push *push = drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_copy()
472 struct nvif_push *push = drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_clear()
532 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy; in nouveau_dmem_migrate_init()
533 drm->dmem->migrate.clear_func = nvc0b5_migrate_clear; in nouveau_dmem_migrate_init()
534 drm->dmem->migrate.chan = drm->ttm.chan; in nouveau_dmem_migrate_init()
590 if (drm->dmem->migrate.copy_func(drm, 1, in nouveau_dmem_migrate_copy_one()
[all …]
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gem/
Di915_gem_object.c663 * @obj: The object to migrate
664 * @id: The region intended to migrate to
667 * given region. Note that pinning may affect the ability to migrate as
671 * possibility to migrate objects and might be slightly less permissive
701 if (!obj->ops->migrate) in i915_gem_object_can_migrate()
719 * i915_gem_object_migrate - Migrate an object to the desired region id
720 * @obj: The object to migrate.
723 * @id: The region id to migrate to.
725 * Attempt to migrate the object to the desired memory region. The
751 * __i915_gem_object_migrate - Migrate an object to the desired region id, with
[all …]
/kernel/linux/linux-5.10/include/uapi/linux/
Dmempolicy.h48 #define MPOL_MF_LAZY (1<<3) /* Modifies '_MOVE: lazy migrate on fault */
62 #define MPOL_F_MOF (1 << 3) /* this policy wants migrate on fault */
63 #define MPOL_F_MORON (1 << 4) /* Migrate On protnone Reference On Node */
/kernel/linux/linux-6.6/include/linux/
Dmigrate.h40 * If the driver cannot migrate the page at the moment, it can return
151 return -EAGAIN; /* can't migrate now */ in migrate_misplaced_page()
214 * Set to vmf->page if this is being called to migrate a page as part of
221 void migrate_vma_pages(struct migrate_vma *migrate);
222 void migrate_vma_finalize(struct migrate_vma *migrate);
Dpsci.h29 int (*migrate)(unsigned long cpuid); member
41 u32 migrate; member
/kernel/linux/linux-6.6/include/uapi/linux/
Dmempolicy.h51 #define MPOL_MF_LAZY (1<<3) /* Modifies '_MOVE: lazy migrate on fault */
64 #define MPOL_F_MOF (1 << 3) /* this policy wants migrate on fault */
65 #define MPOL_F_MORON (1 << 4) /* Migrate On protnone Reference On Node */
/kernel/linux/linux-6.6/drivers/firmware/psci/
Dpsci.c46 * calls to its resident CPU, so we must avoid issuing those. We never migrate
244 return __psci_migrate(psci_0_1_function_ids.migrate, cpuid); in psci_0_1_migrate()
249 return __psci_migrate(PSCI_FN_NATIVE(0_2, MIGRATE), cpuid); in psci_0_2_migrate()
354 PSCI_ID_NATIVE(0_2, MIGRATE),
623 .migrate = psci_0_2_migrate, in psci_0_2_set_functions()
720 if (!of_property_read_u32(np, "migrate", &id)) { in psci_0_1_init()
721 psci_0_1_function_ids.migrate = id; in psci_0_1_init()
722 psci_ops.migrate = psci_0_1_migrate; in psci_0_1_init()
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/x86_64/
Dxapic_ipi_test.c308 * migrate_pages with PID=0 will migrate all pages of this in do_migrations()
361 bool *migrate, int *delay_usecs) in get_cmdline_args() argument
373 *migrate = true; in get_cmdline_args()
399 bool migrate = false; in main() local
405 get_cmdline_args(argc, argv, &run_secs, &migrate, &delay_usecs); in main()
461 if (!migrate) in main()
/kernel/linux/linux-5.10/drivers/firmware/psci/
Dpsci.c44 * calls to its resident CPU, so we must avoid issuing those. We never migrate
439 psci_function_id[PSCI_FN_MIGRATE] = PSCI_FN_NATIVE(0_2, MIGRATE); in psci_0_2_set_functions()
440 psci_ops.migrate = psci_migrate; in psci_0_2_set_functions()
535 if (!of_property_read_u32(np, "migrate", &id)) { in psci_0_1_init()
537 psci_ops.migrate = psci_migrate; in psci_0_1_init()
/kernel/linux/linux-6.6/Documentation/devicetree/bindings/arm/
Dpsci.yaml85 migrate:
87 description: Function ID for MIGRATE operation
153 migrate = <0x95c10003>;

12345678910>>...22