Home
last modified time | relevance | path

Searched refs:migrate (Results 1 – 25 of 99) sorted by relevance

1234

/linux/mm/
H A Dmigrate_device.c9 #include <linux/migrate.h>
24 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_skip() local
28 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip()
29 migrate->src[migrate->npages++] = 0; in migrate_vma_collect_skip()
40 struct migrate_vma *migrate = walk->private; in migrate_vma_collect_hole() local
48 migrate->src[migrate->npages] = MIGRATE_PFN_MIGRATE; in migrate_vma_collect_hole()
49 migrate in migrate_vma_collect_hole()
62 struct migrate_vma *migrate = walk->private; migrate_vma_collect_pmd() local
310 migrate_vma_collect(struct migrate_vma * migrate) migrate_vma_collect() argument
466 migrate_vma_unmap(struct migrate_vma * migrate) migrate_vma_unmap() argument
586 migrate_vma_insert_page(struct migrate_vma * migrate,unsigned long addr,struct page * page,unsigned long * src) migrate_vma_insert_page() argument
704 __migrate_device_pages(unsigned long * src_pfns,unsigned long * dst_pfns,unsigned long npages,struct migrate_vma * migrate) __migrate_device_pages() argument
815 migrate_vma_pages(struct migrate_vma * migrate) migrate_vma_pages() argument
901 migrate_vma_finalize(struct migrate_vma * migrate) migrate_vma_finalize() argument
[all...]
/linux/drivers/gpu/drm/
H A Ddrm_pagemap.c7 #include <linux/migrate.h>
18 * to migrate memory back and forth between device memory and system RAM and
27 * population requests and after that migrate all device pages to system ram.
39 * and it can evict the range to system before trying to migrate. Ideally an
40 * implementation would just try to migrate the missing part of the range and
53 * migrate to / from device memory.
166 * @migrate_pfn: Array of migrate page frame numbers
206 * @migrate_pfn: Array of migrate page frame numbers to map
276 * @devmem_allocation: The device memory allocation to migrate to.
281 * @start: Start of the virtual address range to migrate
307 struct migrate_vma migrate = { drm_pagemap_migrate_to_devmem() local
603 struct migrate_vma migrate = { __drm_pagemap_migrate_to_ram() local
[all...]
/linux/drivers/gpu/drm/i915/gt/
H A Dselftest_migrate.c36 static int copy(struct intel_migrate *migrate, in copy() argument
37 int (*fn)(struct intel_migrate *migrate, in copy() argument
44 struct drm_i915_private *i915 = migrate->context->engine->i915; in copy()
90 err = fn(migrate, &ww, src, dst, &rq); in copy()
151 GEM_BUG_ON(ce->vm != ce->engine->gt->migrate.context->vm); in intel_context_copy_ccs()
255 static int clear(struct intel_migrate *migrate, in clear() argument
256 int (*fn)(struct intel_migrate *migrate, in clear() argument
263 struct drm_i915_private *i915 = migrate->context->engine->i915; in clear()
301 err = intel_migrate_ccs_copy(migrate, &ww, NULL, in clear()
318 err = fn(migrate, in clear()
410 __migrate_copy(struct intel_migrate * migrate,struct i915_gem_ww_ctx * ww,struct drm_i915_gem_object * src,struct drm_i915_gem_object * dst,struct i915_request ** out) __migrate_copy() argument
424 __global_copy(struct intel_migrate * migrate,struct i915_gem_ww_ctx * ww,struct drm_i915_gem_object * src,struct drm_i915_gem_object * dst,struct i915_request ** out) __global_copy() argument
439 migrate_copy(struct intel_migrate * migrate,u32 sz,struct rnd_state * prng) migrate_copy() argument
445 global_copy(struct intel_migrate * migrate,u32 sz,struct rnd_state * prng) global_copy() argument
450 __migrate_clear(struct intel_migrate * migrate,struct i915_gem_ww_ctx * ww,struct drm_i915_gem_object * obj,u32 value,struct i915_request ** out) __migrate_clear() argument
463 __global_clear(struct intel_migrate * migrate,struct i915_gem_ww_ctx * ww,struct drm_i915_gem_object * obj,u32 value,struct i915_request ** out) __global_clear() argument
477 migrate_clear(struct intel_migrate * migrate,u32 sz,struct rnd_state * prng) migrate_clear() argument
483 global_clear(struct intel_migrate * migrate,u32 sz,struct rnd_state * prng) global_clear() argument
491 struct intel_migrate *migrate = &gt->migrate; live_migrate_copy() local
513 struct intel_migrate *migrate = &gt->migrate; live_migrate_clear() local
549 struct intel_migrate *migrate = &gt->migrate; live_emit_pte_full_ring() local
677 struct intel_migrate *migrate; global() member
682 threaded_migrate(struct intel_migrate * migrate,int (* fn)(void * arg),unsigned int flags) threaded_migrate() argument
741 struct intel_migrate *migrate = &gt->migrate; thread_migrate_copy() local
756 struct intel_migrate *migrate = &gt->migrate; thread_global_copy() local
778 struct intel_migrate *migrate = &gt->migrate; thread_migrate_clear() local
786 struct intel_migrate *migrate = &gt->migrate; thread_global_clear() local
[all...]
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_migrate.c27 #include <linux/migrate.h>
181 * @mfence: migrate fence
263 static unsigned long svm_migrate_unsuccessful_pages(struct migrate_vma *migrate) in svm_migrate_unsuccessful_pages() argument
268 for (i = 0; i < migrate->npages; i++) { in svm_migrate_unsuccessful_pages()
269 if (migrate->src[i] & MIGRATE_PFN_VALID && in svm_migrate_unsuccessful_pages()
270 !(migrate->src[i] & MIGRATE_PFN_MIGRATE)) in svm_migrate_unsuccessful_pages()
278 struct migrate_vma *migrate, struct dma_fence **mfence, in svm_migrate_copy_to_vram() argument
281 uint64_t npages = migrate->npages; in svm_migrate_copy_to_vram()
299 for (i = j = 0; (i < npages) && (mpages < migrate->cpages); i++) { in svm_migrate_copy_to_vram()
302 if (migrate in svm_migrate_copy_to_vram()
396 struct migrate_vma migrate = { 0 }; svm_migrate_vma_to_vram() local
582 svm_migrate_copy_to_ram(struct amdgpu_device * adev,struct svm_range * prange,struct migrate_vma * migrate,struct dma_fence ** mfence,dma_addr_t * scratch,uint64_t npages) svm_migrate_copy_to_ram() argument
697 struct migrate_vma migrate = { 0 }; svm_migrate_vma_to_ram() local
[all...]
/linux/tools/testing/selftests/mm/
H A Dmigration.c66 int migrate(uint64_t *ptr, int n1, int n2) in migrate() function
90 printf("Didn't migrate %d pages\n", ret); in migrate()
93 perror("Couldn't migrate pages"); in migrate()
141 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
176 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
206 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
248 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
273 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
308 ASSERT_EQ(migrate(ptr, self->n1, self->n2), 0);
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_dmem.c42 #include <linux/migrate.h>
82 struct nouveau_dmem_migrate migrate; member
153 if (drm->dmem->migrate.copy_func(drm, 1, NOUVEAU_APER_HOST, *dma_addr, in nouveau_dmem_copy_one()
184 * FIXME what we really want is to find some heuristic to migrate more in nouveau_dmem_migrate_to_ram()
213 nouveau_fence_new(&fence, dmem->migrate.chan); in nouveau_dmem_migrate_to_ram()
399 nouveau_fence_new(&fence, chunk->drm->dmem->migrate.chan); in nouveau_dmem_evict_chunk()
439 struct nvif_push *push = &drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_copy()
512 struct nvif_push *push = &drm->dmem->migrate.chan->chan.push; in nvc0b5_migrate_clear()
572 drm->dmem->migrate.copy_func = nvc0b5_migrate_copy; in nouveau_dmem_migrate_init()
573 drm->dmem->migrate in nouveau_dmem_migrate_init()
[all...]
/linux/include/linux/
H A Dpsci.h29 int (*migrate)(unsigned long cpuid); member
41 u32 migrate; member
H A Dmigrate.h40 * If the driver cannot migrate the page at the moment, it can return
114 return -EAGAIN; /* can't migrate now */ in migrate_misplaced_folio_prepare()
118 return -EAGAIN; /* can't migrate now */ in migrate_misplaced_folio()
181 * Set to vmf->page if this is being called to migrate a page as part of
188 void migrate_vma_pages(struct migrate_vma *migrate);
189 void migrate_vma_finalize(struct migrate_vma *migrate);
/linux/tools/testing/selftests/kvm/x86/
H A Dxapic_ipi_test.c307 * migrate_pages with PID=0 will migrate all pages of this in do_migrations()
360 bool *migrate, int *delay_usecs) in get_cmdline_args() argument
372 *migrate = true; in get_cmdline_args()
398 bool migrate = false; in main() local
404 get_cmdline_args(argc, argv, &run_secs, &migrate, &delay_usecs); in main()
458 if (!migrate) in main()
/linux/arch/powerpc/platforms/pseries/
H A Dvas.c652 bool migrate) in reconfig_open_windows() argument
680 if ((vcaps->nr_close_wins > creds) && !migrate) in reconfig_open_windows()
694 if (migrate) in reconfig_open_windows()
762 bool migrate) in reconfig_close_windows() argument
769 if (migrate) in reconfig_close_windows()
827 if (rc && !migrate) in reconfig_close_windows()
839 if (!migrate && !--excess_creds) in reconfig_close_windows()
/linux/drivers/gpu/drm/xe/
H A Dxe_bo.c165 return tile->migrate; in mem_type_to_migrate()
796 struct xe_migrate *migrate = NULL; in xe_bo_move() local
906 migrate = bo->tile->migrate; in xe_bo_move()
908 migrate = mem_type_to_migrate(xe, new_mem->mem_type); in xe_bo_move()
910 migrate = mem_type_to_migrate(xe, old_mem_type); in xe_bo_move()
912 migrate = xe->tiles[0].migrate; in xe_bo_move()
914 xe_assert(xe, migrate); in xe_bo_move()
935 fence = xe_migrate_clear(migrate, b in xe_bo_move()
1234 struct xe_migrate *migrate; xe_bo_evict_pinned() local
1326 struct xe_migrate *migrate; xe_bo_restore_pinned() local
1569 struct xe_migrate *migrate = xe_ttm_access_memory() local
[all...]
H A Dxe_tile.c192 xe_migrate_wait(tile->migrate); in xe_tile_migrate_wait()
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_ttm_move.c202 if (!to_gt(i915)->migrate.context || intel_gt_is_wedged(to_gt(i915))) in i915_ttm_accel_move()
215 intel_engine_pm_get(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move()
216 ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, deps, in i915_ttm_accel_move()
229 intel_engine_pm_get(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move()
230 ret = intel_context_migrate_copy(to_gt(i915)->migrate.context, in i915_ttm_accel_move()
242 intel_engine_pm_put(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move()
661 ret = -EIO; /* never feed non-migrate fences into ttm */ in i915_ttm_move()
/linux/Documentation/gpu/rfc/
H A Dgpusvm.rst21 migrate can actually migrate, with only the faulting page guaranteed
22 to migrate).
110 migrate device layer are a performance bottleneck, having compound
/linux/drivers/gpu/drm/i915/selftests/
H A Di915_perf_selftests.h20 selftest(migrate, intel_migrate_perf_selftests)
H A Di915_live_selftests.h30 selftest(migrate, intel_migrate_live_selftests)
/linux/drivers/firmware/psci/
H A Dpsci.c46 * calls to its resident CPU, so we must avoid issuing those. We never migrate
245 return __psci_migrate(psci_0_1_function_ids.migrate, cpuid); in psci_0_1_migrate()
667 .migrate = psci_0_2_migrate, in psci_0_2_set_functions()
765 if (!of_property_read_u32(np, "migrate", &id)) { in psci_0_1_init()
766 psci_0_1_function_ids.migrate = id; in psci_0_1_init()
767 psci_ops.migrate = psci_0_1_migrate; in psci_0_1_init()
/linux/arch/arm64/kvm/hyp/nvhe/
H A Dpsci-relay.c53 is_psci_0_1(migrate, func_id)); in is_psci_0_1_call()
229 if (is_psci_0_1(cpu_off, func_id) || is_psci_0_1(migrate, func_id)) in psci_0_1_handler()
/linux/fs/bcachefs/
H A Dbcachefs_ioctl.h187 x(migrate, 2) \
200 * scrub, rereplicate, migrate).
224 } migrate; member
/linux/tools/perf/scripts/python/
H A Dsched-migration.py191 def migrate(self, ts_list, new, old_cpu, new_cpu): member in TimeSlice
351 def migrate(self, headers, pid, prio, orig_cpu, dest_cpu): member in SchedEventProxy
353 ts.migrate(self.timeslices, pid, orig_cpu, dest_cpu)
418 parser.migrate(headers, pid, prio, orig_cpu, dest_cpu)
/linux/fs/ocfs2/dlm/
H A Ddlmmaster.c824 * of the MIGRATION mle: either the migrate finished or in dlm_get_lock_resource()
1510 mlog(0, "node %u is master, but trying to migrate to " in dlm_master_request_handler()
1514 "node is trying to migrate it to %u?!\n", in dlm_master_request_handler()
2605 mlog(ML_ERROR, "tried to migrate %.*s to %u, but " in dlm_migrate_lockres()
2972 * Pick a node to migrate the lock resource to. This function selects a
3027 struct dlm_migrate_request migrate; in dlm_do_migrate_request() local
3031 memset(&migrate, 0, sizeof(migrate)); in dlm_do_migrate_request()
3032 migrate.namelen = res->lockname.len; in dlm_do_migrate_request()
3033 memcpy(migrate in dlm_do_migrate_request()
3104 struct dlm_migrate_request *migrate = (struct dlm_migrate_request *) msg->buf; dlm_migrate_request_handler() local
[all...]
/linux/fs/ext4/
H A DMakefile10 indirect.o inline.o inode.o ioctl.o mballoc.o migrate.o \
/linux/kernel/sched/
H A Dstats.h224 static inline void psi_enqueue(struct task_struct *p, bool migrate) {} in psi_enqueue() argument
225 static inline void psi_dequeue(struct task_struct *p, bool migrate) {} in psi_dequeue() argument
/linux/net/xfrm/
H A DKconfig62 bool "Transformation migrate database"
134 <draft-sugimoto-mip6-pfkey-migrate>.
/linux/Documentation/trace/
H A Dhwlat_detector.rst81 By default, one hwlat detector's kernel thread will migrate across each CPU
87 - round-robin: migrate across each CPU specified in cpumask [default]

1234