/linux/crypto/async_tx/ |
H A D | async_xor.c | 23 do_async_xor(struct dma_chan *chan, struct dmaengine_unmap_data *unmap, in do_async_xor() argument 32 int src_cnt = unmap->to_cnt; in do_async_xor() 34 dma_addr_t dma_dest = unmap->addr[unmap->to_cnt]; in do_async_xor() 35 dma_addr_t *src_list = unmap->addr; in do_async_xor() 63 if (src_list > unmap->addr) in do_async_xor() 66 xor_src_cnt, unmap->len, in do_async_xor() 77 xor_src_cnt, unmap->len, in do_async_xor() 82 dma_set_unmap(tx, unmap); in do_async_xor() 190 struct dmaengine_unmap_data *unmap in async_xor_offs() local 323 struct dmaengine_unmap_data *unmap = NULL; async_xor_val_offs() local [all...] |
H A D | async_memcpy.c | 40 struct dmaengine_unmap_data *unmap = NULL; in async_memcpy() local 43 unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT); in async_memcpy() 45 if (unmap && is_dma_copy_aligned(device, src_offset, dest_offset, len)) { in async_memcpy() 53 unmap->to_cnt = 1; in async_memcpy() 54 unmap->addr[0] = dma_map_page(device->dev, src, src_offset, len, in async_memcpy() 56 unmap->from_cnt = 1; in async_memcpy() 57 unmap->addr[1] = dma_map_page(device->dev, dest, dest_offset, len, in async_memcpy() 59 unmap->len = len; in async_memcpy() 61 tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], in async_memcpy() 62 unmap in async_memcpy() [all...] |
H A D | async_pq.c | 37 struct dmaengine_unmap_data *unmap, in do_async_gen_syndrome() argument 76 dma_dest[0] = unmap->addr[disks - 2]; in do_async_gen_syndrome() 77 dma_dest[1] = unmap->addr[disks - 1]; in do_async_gen_syndrome() 79 &unmap->addr[src_off], in do_async_gen_syndrome() 81 &scfs[src_off], unmap->len, in do_async_gen_syndrome() 89 dma_set_unmap(tx, unmap); in do_async_gen_syndrome() 185 struct dmaengine_unmap_data *unmap = NULL; in async_gen_syndrome() local 190 unmap = dmaengine_get_unmap_data(device->dev, disks, GFP_NOWAIT); in async_gen_syndrome() 193 if (unmap && !(submit->flags & ASYNC_TX_PQ_XOR_DST) && in async_gen_syndrome() 209 unmap in async_gen_syndrome() 307 struct dmaengine_unmap_data *unmap = NULL; async_syndrome_val() local [all...] |
H A D | async_raid6_recov.c | 25 struct dmaengine_unmap_data *unmap = NULL; in async_sum_product() local 31 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_sum_product() 33 if (unmap) { in async_sum_product() 41 unmap->addr[0] = dma_map_page(dev, srcs[0], src_offs[0], in async_sum_product() 43 unmap->addr[1] = dma_map_page(dev, srcs[1], src_offs[1], in async_sum_product() 45 unmap->to_cnt = 2; in async_sum_product() 47 unmap->addr[2] = dma_map_page(dev, dest, d_off, in async_sum_product() 49 unmap->bidi_cnt = 1; in async_sum_product() 51 pq[1] = unmap->addr[2]; in async_sum_product() 53 unmap in async_sum_product() 94 struct dmaengine_unmap_data *unmap = NULL; async_mult() local [all...] |
/linux/tools/testing/selftests/mm/ |
H A D | ksm_functional_tests.c | 186 goto unmap; in __mmap_and_merge_range() 195 goto unmap; in __mmap_and_merge_range() 204 goto unmap; in __mmap_and_merge_range() 207 goto unmap; in __mmap_and_merge_range() 213 goto unmap; in __mmap_and_merge_range() 223 goto unmap; in __mmap_and_merge_range() 232 goto unmap; in __mmap_and_merge_range() 236 unmap: in __mmap_and_merge_range() 271 goto unmap; in test_unmerge() 276 unmap in test_unmerge() [all...] |
/linux/drivers/xen/xenbus/ |
H A D | xenbus_client.c | 80 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; member 92 int (*unmap)(struct xenbus_device *dev, void *vaddr); member 593 gnttab_set_unmap_op(&info->unmap[j], in __xenbus_map_ring() 600 BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, info->unmap, j)); in __xenbus_map_ring() 604 if (info->unmap[i].status != GNTST_okay) { in __xenbus_map_ring() 614 * xenbus_unmap_ring - unmap memory from another domain 618 * @vaddrs: addresses to unmap 628 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; in xenbus_unmap_ring() local 636 gnttab_set_unmap_op(&unmap[i], vaddrs[i], in xenbus_unmap_ring() 639 BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, unmap, in xenbus_unmap_ring() 802 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; xenbus_unmap_ring_pv() local [all...] |
/linux/arch/arm/xen/ |
H A D | p2m.c | 98 struct gnttab_unmap_grant_ref unmap; in set_foreign_p2m_mapping() local 112 unmap.host_addr = map_ops[i].host_addr; in set_foreign_p2m_mapping() 113 unmap.handle = map_ops[i].handle; in set_foreign_p2m_mapping() 116 unmap.dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping() 118 unmap.dev_bus_addr = 0; in set_foreign_p2m_mapping() 124 unmap.status = 1; in set_foreign_p2m_mapping() 127 &unmap, 1); in set_foreign_p2m_mapping() 128 if (rc || unmap.status != GNTST_okay) in set_foreign_p2m_mapping() 129 pr_err_once("gnttab unmap failed: rc=%d st=%d\n", in set_foreign_p2m_mapping() 130 rc, unmap in set_foreign_p2m_mapping() [all...] |
/linux/arch/um/kernel/ |
H A D | tlb.c | 24 int (*unmap)(struct mm_id *mm_idp, member 85 ret = ops->unmap(ops->mm_idp, addr, PAGE_SIZE); in update_pte_range() 105 ret = ops->unmap(ops->mm_idp, addr, in update_pmd_range() 128 ret = ops->unmap(ops->mm_idp, addr, in update_pud_range() 151 ret = ops->unmap(ops->mm_idp, addr, in update_p4d_range() 174 ops.unmap = kern_unmap; in um_tlb_sync() 177 ops.unmap = unmap; in um_tlb_sync() 185 ret = ops.unmap(ops.mm_idp, addr, in um_tlb_sync()
|
/linux/drivers/iommu/iommufd/ |
H A D | vfio_compat.c | 208 struct vfio_iommu_type1_dma_unmap unmap; in iommufd_vfio_unmap_dma() local 213 if (copy_from_user(&unmap, arg, minsz)) in iommufd_vfio_unmap_dma() 216 if (unmap.argsz < minsz || unmap.flags & ~supported_flags) in iommufd_vfio_unmap_dma() 223 if (unmap.flags & VFIO_DMA_UNMAP_FLAG_ALL) { in iommufd_vfio_unmap_dma() 224 if (unmap.iova != 0 || unmap.size != 0) { in iommufd_vfio_unmap_dma() 236 unsigned long iovas[] = { unmap.iova + unmap.size - 1, in iommufd_vfio_unmap_dma() 237 unmap in iommufd_vfio_unmap_dma() [all...] |
/linux/drivers/net/ethernet/brocade/bna/ |
H A D | bnad.c | 94 struct bnad_tx_unmap *unmap; in bnad_tx_buff_unmap() local 98 unmap = &unmap_q[index]; in bnad_tx_buff_unmap() 99 nvecs = unmap->nvecs; in bnad_tx_buff_unmap() 101 skb = unmap->skb; in bnad_tx_buff_unmap() 102 unmap->skb = NULL; in bnad_tx_buff_unmap() 103 unmap->nvecs = 0; in bnad_tx_buff_unmap() 105 dma_unmap_addr(&unmap->vectors[0], dma_addr), in bnad_tx_buff_unmap() 107 dma_unmap_addr_set(&unmap->vectors[0], dma_addr, 0); in bnad_tx_buff_unmap() 116 unmap = &unmap_q[index]; in bnad_tx_buff_unmap() 120 dma_unmap_addr(&unmap in bnad_tx_buff_unmap() 165 struct bnad_tx_unmap *unmap; bnad_txcmpl_process() local 293 bnad_rxq_cleanup_page(struct bnad * bnad,struct bnad_rx_unmap * unmap) bnad_rxq_cleanup_page() argument 308 bnad_rxq_cleanup_skb(struct bnad * bnad,struct bnad_rx_unmap * unmap) bnad_rxq_cleanup_skb() argument 329 struct bnad_rx_unmap *unmap = &unmap_q->unmap[i]; bnad_rxq_cleanup() local 344 struct bnad_rx_unmap *unmap, *prev; bnad_rxq_refill_page() local 418 struct bnad_rx_unmap *unmap; bnad_rxq_refill_skb() local 504 struct bnad_rx_unmap *unmap; bnad_cq_drop_packet() local 542 struct bnad_rx_unmap *unmap; bnad_cq_setup_skb_frags() local 572 bnad_cq_setup_skb(struct bnad * bnad,struct sk_buff * skb,struct bnad_rx_unmap * unmap,u32 len) bnad_cq_setup_skb() argument 593 struct bnad_rx_unmap *unmap = NULL; bnad_cq_process() local 2919 struct bnad_tx_unmap *unmap_q, *unmap, *head_unmap; bnad_start_xmit() local [all...] |
/linux/drivers/dma/ |
H A D | dmaengine.c | 1351 #define __UNMAP_POOL(x) { .size = x, .name = "dmaengine-unmap-" __stringify(x) } 1384 struct dmaengine_unmap_data *unmap = container_of(kref, typeof(*unmap), kref); in dmaengine_unmap() local 1385 struct device *dev = unmap->dev; in dmaengine_unmap() 1388 cnt = unmap->to_cnt; in dmaengine_unmap() 1390 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap() 1392 cnt += unmap->from_cnt; in dmaengine_unmap() 1394 dma_unmap_page(dev, unmap->addr[i], unmap in dmaengine_unmap() 1407 dmaengine_unmap_put(struct dmaengine_unmap_data * unmap) dmaengine_unmap_put() argument 1458 struct dmaengine_unmap_data *unmap; dmaengine_get_unmap_data() local [all...] |
H A D | mv_xor.c | 769 struct dmaengine_unmap_data *unmap; in mv_chan_memcpy_self_test() local 792 unmap = dmaengine_get_unmap_data(dma_chan->device->dev, 2, GFP_KERNEL); in mv_chan_memcpy_self_test() 793 if (!unmap) { in mv_chan_memcpy_self_test() 801 unmap->addr[0] = src_dma; in mv_chan_memcpy_self_test() 808 unmap->to_cnt = 1; in mv_chan_memcpy_self_test() 813 unmap->addr[1] = dest_dma; in mv_chan_memcpy_self_test() 820 unmap->from_cnt = 1; in mv_chan_memcpy_self_test() 821 unmap->len = PAGE_SIZE; in mv_chan_memcpy_self_test() 862 dmaengine_unmap_put(unmap); in mv_chan_memcpy_self_test() 880 struct dmaengine_unmap_data *unmap; in mv_chan_xor_self_test() local [all...] |
/linux/arch/x86/xen/ |
H A D | p2m.c | 701 struct gnttab_unmap_grant_ref unmap[2]; in set_foreign_p2m_mapping() local 728 unmap[0].host_addr = map_ops[i].host_addr; in set_foreign_p2m_mapping() 729 unmap[0].handle = map_ops[i].handle; in set_foreign_p2m_mapping() 732 unmap[0].dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping() 734 unmap[0].dev_bus_addr = 0; in set_foreign_p2m_mapping() 738 unmap[1].host_addr = kmap_ops[i].host_addr; in set_foreign_p2m_mapping() 739 unmap[1].handle = kmap_ops[i].handle; in set_foreign_p2m_mapping() 742 unmap[1].dev_bus_addr = kmap_ops[i].dev_bus_addr; in set_foreign_p2m_mapping() 744 unmap[1].dev_bus_addr = 0; in set_foreign_p2m_mapping() 751 unmap[ in set_foreign_p2m_mapping() [all...] |
/linux/include/xen/ |
H A D | grant_table.h | 179 gnttab_set_unmap_op(struct gnttab_unmap_grant_ref *unmap, phys_addr_t addr, in gnttab_set_unmap_op() argument 183 unmap->host_addr = addr; in gnttab_set_unmap_op() 185 unmap->host_addr = __pa(addr); in gnttab_set_unmap_op() 187 unmap->host_addr = addr; in gnttab_set_unmap_op() 189 unmap->handle = handle; in gnttab_set_unmap_op() 190 unmap->dev_bus_addr = 0; in gnttab_set_unmap_op()
|
/linux/drivers/acpi/ |
H A D | nvs.c | 80 bool unmap; member 139 if (entry->unmap) { in suspend_nvs_free() 141 entry->unmap = false; in suspend_nvs_free() 185 entry->unmap = !!entry->kaddr; in suspend_nvs_save()
|
/linux/drivers/firmware/efi/ |
H A D | memattr.c | 40 goto unmap; in efi_memattr_init() 55 goto unmap; in efi_memattr_init() 62 unmap: in efi_memattr_init()
|
/linux/drivers/gpu/drm/ |
H A D | drm_gpuvm.c | 108 * sequence of operations to satisfy a given map or unmap request. 117 * containing map, unmap and remap operations for a given newly requested 124 * of unmap operations, a maximum of two remap operations and a single map 132 * &drm_gpuva to unmap is physically contiguous with the original mapping 139 * one unmap operation and one or two map operations, such that drivers can 147 * call back into the driver in order to unmap a range of GPU VA space. The 149 * enclosed by the given range unmap operations are created. For mappings which 468 * &drm_gpuva_op in order to satisfy a given map or unmap request and how to 521 * va = op->remap.unmap->va; 551 * va = op->unmap 2075 op_remap_cb(const struct drm_gpuvm_ops * fn,void * priv,struct drm_gpuva_op_map * prev,struct drm_gpuva_op_map * next,struct drm_gpuva_op_unmap * unmap) op_remap_cb() argument 2284 struct drm_gpuva_op_unmap unmap = { .va = va }; __drm_gpuvm_sm_unmap() local [all...] |
/linux/arch/x86/platform/geode/ |
H A D | net5501.c | 66 goto unmap; in net5501_present() 80 unmap: in net5501_present()
|
/linux/drivers/irqchip/ |
H A D | irq-mchp-eic.c | 220 goto unmap; in mchp_eic_init() 226 goto unmap; in mchp_eic_init() 231 goto unmap; in mchp_eic_init() 268 unmap: in mchp_eic_init()
|
/linux/drivers/gpu/drm/imagination/ |
H A D | pvr_mmu.c | 112 * a series of maps. However, a single unmap should always be immediately 1433 /** @unmap: Data specifically for unmap operations. */ 1436 * @l1_free_tables: Collects page table objects freed by unmap 1442 * @l0_free_tables: Collects page table objects freed by unmap 1446 } unmap; member 1521 op_ctx->curr_page.l1_table->next_free = op_ctx->unmap.l1_free_tables; in pvr_page_table_l2_remove() 1522 op_ctx->unmap.l1_free_tables = op_ctx->curr_page.l1_table; in pvr_page_table_l2_remove() 1584 op_ctx->curr_page.l0_table->next_free = op_ctx->unmap.l0_free_tables; in pvr_page_table_l1_remove() 1585 op_ctx->unmap in pvr_page_table_l1_remove() [all...] |
/linux/drivers/net/wireless/mediatek/mt76/ |
H A D | wed.c | 48 goto unmap; in mt76_wed_init_rx_buf() 52 goto unmap; in mt76_wed_init_rx_buf() 59 goto unmap; in mt76_wed_init_rx_buf() 72 unmap: in mt76_wed_init_rx_buf()
|
/linux/include/linux/ |
H A D | dmaengine.h | 603 * @unmap: hook for generic DMA unmap data 624 struct dmaengine_unmap_data *unmap; member 636 struct dmaengine_unmap_data *unmap) in dma_set_unmap() argument 638 kref_get(&unmap->kref); in dma_set_unmap() 639 tx->unmap = unmap; in dma_set_unmap() 644 void dmaengine_unmap_put(struct dmaengine_unmap_data *unmap); 647 struct dmaengine_unmap_data *unmap) in dma_set_unmap() argument 655 static inline void dmaengine_unmap_put(struct dmaengine_unmap_data *unmap) in dmaengine_unmap_put() argument [all...] |
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmgm200.c | 37 .unmap = gf100_vmm_pgt_unmap, 47 .unmap = gf100_vmm_pgt_unmap, 62 .unmap = gf100_vmm_pgt_unmap,
|
/linux/drivers/xen/ |
H A D | arm-device.c | 70 goto unmap; in xen_map_device_mmio() 94 goto unmap; in xen_map_device_mmio() 99 unmap: in xen_map_device_mmio() 126 (action == BUS_NOTIFY_DEL_DEVICE ? "unmap" : "?"), in xen_platform_notifier() 168 (action == BUS_NOTIFY_DEL_DEVICE ? "unmap" : "?"), in xen_amba_notifier()
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_uvmm.c | 12 * Actual map/unmap operations within the fence signalling critical path are 15 * list in order to map/unmap it's entries, can't occur concurrently. 472 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare_unwind() 484 op_unmap_prepare_unwind(op->unmap.va); in nouveau_uvmm_sm_prepare_unwind() 509 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare_unwind() 526 struct drm_gpuva_op_unmap *u = &op->unmap; in nouveau_uvmm_sm_prepare_unwind() 647 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare() 656 op_unmap_prepare(r->unmap); in nouveau_uvmm_sm_prepare() 687 struct drm_gpuva_op_unmap *u = &op->unmap; in nouveau_uvmm_sm_prepare() 769 return op->remap.unmap in op_gem_obj() 871 nouveau_uvmm_sm_cleanup(struct nouveau_uvmm * uvmm,struct nouveau_uvma_prealloc * new,struct drm_gpuva_ops * ops,bool unmap) nouveau_uvmm_sm_cleanup() argument [all...] |