Home
last modified time | relevance | path

Searched refs:devmem (Results 1 – 11 of 11) sorted by relevance

/linux/lib/
H A Dtest_hmm.c499 struct dmirror_chunk *devmem; in dmirror_allocate_chunk() local
507 devmem = kzalloc_obj(*devmem); in dmirror_allocate_chunk()
508 if (!devmem) in dmirror_allocate_chunk()
517 devmem->pagemap.range.start = res->start; in dmirror_allocate_chunk()
518 devmem->pagemap.range.end = res->end; in dmirror_allocate_chunk()
519 devmem->pagemap.type = MEMORY_DEVICE_PRIVATE; in dmirror_allocate_chunk()
522 devmem->pagemap.range.start = (MINOR(mdevice->cdevice.dev) - 2) ? in dmirror_allocate_chunk()
525 devmem->pagemap.range.end = devmem->pagemap.range.start + in dmirror_allocate_chunk()
527 devmem->pagemap.type = MEMORY_DEVICE_COHERENT; in dmirror_allocate_chunk()
534 devmem->pagemap.nr_range = 1; in dmirror_allocate_chunk()
[all …]
/linux/Documentation/networking/
H A Ddevmem.rst11 Device memory TCP (devmem TCP) enables receiving data directly into device
79 Header split, flow steering, & RSS are required features for devmem TCP.
84 Flow steering & RSS are used to ensure that only flows targeting devmem land on
85 an RX queue bound to devmem.
134 devmem TCP, even if the dmabuf is not actually backed by devmem. An example of
135 this is udmabuf, which wraps user memory (non-devmem) in a dmabuf.
150 devmem data by passing the MSG_SOCK_DEVMEM flag to recvmsg::
155 on devmem data.
210 Applications may receive no SO_DEVMEM_* cmsgs. That indicates non-devmem,
277 devmem TCP, even if the dmabuf is not actually backed by devmem. An example of
[all …]
H A Dindex.rst52 devmem
/linux/drivers/gpu/drm/
H A Ddrm_pagemap.c132 struct drm_pagemap_devmem *devmem = zdd->devmem_allocation; in drm_pagemap_zdd_destroy() local
135 if (devmem) { in drm_pagemap_zdd_destroy()
136 complete_all(&devmem->detached); in drm_pagemap_zdd_destroy()
137 if (devmem->ops->devmem_release) in drm_pagemap_zdd_destroy()
138 devmem->ops->devmem_release(devmem); in drm_pagemap_zdd_destroy()
339 drm_pagemap_migrate_remote_to_local(struct drm_pagemap_devmem *devmem, in drm_pagemap_migrate_remote_to_local() argument
358 devmem->pre_migrate_fence); in drm_pagemap_migrate_remote_to_local()
366 drm_pagemap_migrate_sys_to_dev(struct drm_pagemap_devmem *devmem, in drm_pagemap_migrate_sys_to_dev() argument
374 int err = drm_pagemap_migrate_map_pages(devmem->dev, devmem->dpagemap, in drm_pagemap_migrate_sys_to_dev()
382 devmem->pre_migrate_fence); in drm_pagemap_migrate_sys_to_dev()
[all …]
/linux/tools/testing/selftests/drivers/net/hw/
H A DMakefile24 devmem.py \
/linux/net/core/
H A DMakefile49 obj-$(CONFIG_NET_DEVMEM) += devmem.o
H A Dskbuff.c1730 bool devmem) in msg_zerocopy_alloc() argument
1745 if (likely(!devmem) && mm_account_pinned_pages(&uarg->mmp, size)) { in msg_zerocopy_alloc()
1768 struct ubuf_info *uarg, bool devmem) in msg_zerocopy_realloc() argument
1798 if (likely(!devmem) && in msg_zerocopy_realloc()
1814 return msg_zerocopy_alloc(sk, size, devmem); in msg_zerocopy_realloc()
/linux/Documentation/translations/zh_CN/networking/
H A Dindex.rst76 * devmem
/linux/drivers/gpu/drm/xe/
H A Dxe_migrate.c635 bool devmem = false; in emit_pte() local
649 devmem = true; in emit_pte()
654 0, devmem, flags); in emit_pte()
H A Dxe_vm.c1377 u32 pt_level, bool devmem, u64 flags) in xelp_pte_encode_addr() argument
1389 if (devmem) in xelp_pte_encode_addr()
/linux/include/linux/
H A Dskbuff.h1756 struct ubuf_info *uarg, bool devmem);