Home
last modified time | relevance | path

Searched full:unmap (Results 1 – 25 of 935) sorted by relevance

12345678910>>...38

/linux-6.14.4/crypto/async_tx/
Dasync_xor.c23 do_async_xor(struct dma_chan *chan, struct dmaengine_unmap_data *unmap, in do_async_xor() argument
32 int src_cnt = unmap->to_cnt; in do_async_xor()
34 dma_addr_t dma_dest = unmap->addr[unmap->to_cnt]; in do_async_xor()
35 dma_addr_t *src_list = unmap->addr; in do_async_xor()
63 if (src_list > unmap->addr) in do_async_xor()
66 xor_src_cnt, unmap->len, in do_async_xor()
77 xor_src_cnt, unmap->len, in do_async_xor()
82 dma_set_unmap(tx, unmap); in do_async_xor()
190 struct dmaengine_unmap_data *unmap = NULL; in async_xor_offs() local
195 unmap = dmaengine_get_unmap_data(device->dev, src_cnt+1, GFP_NOWAIT); in async_xor_offs()
[all …]
Dasync_pq.c37 struct dmaengine_unmap_data *unmap, in do_async_gen_syndrome() argument
76 dma_dest[0] = unmap->addr[disks - 2]; in do_async_gen_syndrome()
77 dma_dest[1] = unmap->addr[disks - 1]; in do_async_gen_syndrome()
79 &unmap->addr[src_off], in do_async_gen_syndrome()
81 &scfs[src_off], unmap->len, in do_async_gen_syndrome()
89 dma_set_unmap(tx, unmap); in do_async_gen_syndrome()
185 struct dmaengine_unmap_data *unmap = NULL; in async_gen_syndrome() local
190 unmap = dmaengine_get_unmap_data(device->dev, disks, GFP_NOWAIT); in async_gen_syndrome()
193 if (unmap && !(submit->flags & ASYNC_TX_PQ_XOR_DST) && in async_gen_syndrome()
209 unmap->len = len; in async_gen_syndrome()
[all …]
Dasync_memcpy.c40 struct dmaengine_unmap_data *unmap = NULL; in async_memcpy() local
43 unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT); in async_memcpy()
45 if (unmap && is_dma_copy_aligned(device, src_offset, dest_offset, len)) { in async_memcpy()
53 unmap->to_cnt = 1; in async_memcpy()
54 unmap->addr[0] = dma_map_page(device->dev, src, src_offset, len, in async_memcpy()
56 unmap->from_cnt = 1; in async_memcpy()
57 unmap->addr[1] = dma_map_page(device->dev, dest, dest_offset, len, in async_memcpy()
59 unmap->len = len; in async_memcpy()
61 tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], in async_memcpy()
62 unmap->addr[0], len, in async_memcpy()
[all …]
Dasync_raid6_recov.c25 struct dmaengine_unmap_data *unmap = NULL; in async_sum_product() local
31 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_sum_product()
33 if (unmap) { in async_sum_product()
41 unmap->addr[0] = dma_map_page(dev, srcs[0], src_offs[0], in async_sum_product()
43 unmap->addr[1] = dma_map_page(dev, srcs[1], src_offs[1], in async_sum_product()
45 unmap->to_cnt = 2; in async_sum_product()
47 unmap->addr[2] = dma_map_page(dev, dest, d_off, in async_sum_product()
49 unmap->bidi_cnt = 1; in async_sum_product()
51 pq[1] = unmap->addr[2]; in async_sum_product()
53 unmap->len = len; in async_sum_product()
[all …]
/linux-6.14.4/tools/testing/selftests/mm/
Dksm_functional_tests.c186 goto unmap; in __mmap_and_merge_range()
195 goto unmap; in __mmap_and_merge_range()
204 goto unmap; in __mmap_and_merge_range()
207 goto unmap; in __mmap_and_merge_range()
213 goto unmap; in __mmap_and_merge_range()
223 goto unmap; in __mmap_and_merge_range()
232 goto unmap; in __mmap_and_merge_range()
236 unmap: in __mmap_and_merge_range()
271 goto unmap; in test_unmerge()
276 unmap: in test_unmerge()
[all …]
Dmremap_dontunmap.c59 "unable to unmap destination mapping"); in kernel_support_for_mremap_dontunmap()
63 "unable to unmap source mapping"); in kernel_support_for_mremap_dontunmap()
122 "unable to unmap destination mapping"); in mremap_dontunmap_simple()
124 "unable to unmap source mapping"); in mremap_dontunmap_simple()
155 "unable to unmap source mapping"); in mremap_dontunmap_simple_shmem()
174 "unable to unmap destination mapping"); in mremap_dontunmap_simple_shmem()
176 "unable to unmap source mapping"); in mremap_dontunmap_simple_shmem()
209 // The dest mapping will have been unmap by mremap so we expect the Xs in mremap_dontunmap_simple_fixed()
221 "unable to unmap destination mapping"); in mremap_dontunmap_simple_fixed()
223 "unable to unmap source mapping"); in mremap_dontunmap_simple_fixed()
[all …]
/linux-6.14.4/drivers/xen/xenbus/
Dxenbus_client.c80 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; member
92 int (*unmap)(struct xenbus_device *dev, void *vaddr); member
591 gnttab_set_unmap_op(&info->unmap[j], in __xenbus_map_ring()
598 BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, info->unmap, j)); in __xenbus_map_ring()
602 if (info->unmap[i].status != GNTST_okay) { in __xenbus_map_ring()
612 * xenbus_unmap_ring - unmap memory from another domain
616 * @vaddrs: addresses to unmap
618 * Unmap memory in this domain that was imported from another domain.
626 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; in xenbus_unmap_ring() local
634 gnttab_set_unmap_op(&unmap[i], vaddrs[i], in xenbus_unmap_ring()
[all …]
/linux-6.14.4/arch/arm/xen/
Dp2m.c98 struct gnttab_unmap_grant_ref unmap; in set_foreign_p2m_mapping() local
112 unmap.host_addr = map_ops[i].host_addr; in set_foreign_p2m_mapping()
113 unmap.handle = map_ops[i].handle; in set_foreign_p2m_mapping()
116 unmap.dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping()
118 unmap.dev_bus_addr = 0; in set_foreign_p2m_mapping()
124 unmap.status = 1; in set_foreign_p2m_mapping()
127 &unmap, 1); in set_foreign_p2m_mapping()
128 if (rc || unmap.status != GNTST_okay) in set_foreign_p2m_mapping()
129 pr_err_once("gnttab unmap failed: rc=%d st=%d\n", in set_foreign_p2m_mapping()
130 rc, unmap.status); in set_foreign_p2m_mapping()
/linux-6.14.4/drivers/iommu/iommufd/
Dvfio_compat.c208 struct vfio_iommu_type1_dma_unmap unmap; in iommufd_vfio_unmap_dma() local
213 if (copy_from_user(&unmap, arg, minsz)) in iommufd_vfio_unmap_dma()
216 if (unmap.argsz < minsz || unmap.flags & ~supported_flags) in iommufd_vfio_unmap_dma()
223 if (unmap.flags & VFIO_DMA_UNMAP_FLAG_ALL) { in iommufd_vfio_unmap_dma()
224 if (unmap.iova != 0 || unmap.size != 0) { in iommufd_vfio_unmap_dma()
236 unsigned long iovas[] = { unmap.iova + unmap.size - 1, in iommufd_vfio_unmap_dma()
237 unmap.iova - 1 }; in iommufd_vfio_unmap_dma()
240 unmap.iova ? 2 : 1); in iommufd_vfio_unmap_dma()
244 rc = iopt_unmap_iova(&ioas->iopt, unmap.iova, unmap.size, in iommufd_vfio_unmap_dma()
247 unmap.size = unmapped; in iommufd_vfio_unmap_dma()
[all …]
/linux-6.14.4/drivers/net/ethernet/brocade/bna/
Dbnad.c93 struct bnad_tx_unmap *unmap; in bnad_tx_buff_unmap() local
97 unmap = &unmap_q[index]; in bnad_tx_buff_unmap()
98 nvecs = unmap->nvecs; in bnad_tx_buff_unmap()
100 skb = unmap->skb; in bnad_tx_buff_unmap()
101 unmap->skb = NULL; in bnad_tx_buff_unmap()
102 unmap->nvecs = 0; in bnad_tx_buff_unmap()
104 dma_unmap_addr(&unmap->vectors[0], dma_addr), in bnad_tx_buff_unmap()
106 dma_unmap_addr_set(&unmap->vectors[0], dma_addr, 0); in bnad_tx_buff_unmap()
115 unmap = &unmap_q[index]; in bnad_tx_buff_unmap()
119 dma_unmap_addr(&unmap->vectors[vector], dma_addr), in bnad_tx_buff_unmap()
[all …]
/linux-6.14.4/arch/um/kernel/
Dtlb.c24 int (*unmap)(struct mm_id *mm_idp, member
85 ret = ops->unmap(ops->mm_idp, addr, PAGE_SIZE); in update_pte_range()
105 ret = ops->unmap(ops->mm_idp, addr, in update_pmd_range()
128 ret = ops->unmap(ops->mm_idp, addr, in update_pud_range()
151 ret = ops->unmap(ops->mm_idp, addr, in update_p4d_range()
174 ops.unmap = kern_unmap; in um_tlb_sync()
177 ops.unmap = unmap; in um_tlb_sync()
185 ret = ops.unmap(ops.mm_idp, addr, in um_tlb_sync()
/linux-6.14.4/drivers/xen/
Dxen-front-pgdir-shbuf.c60 /* Unmap grant references of the buffer. */
61 int (*unmap)(struct xen_front_pgdir_shbuf *buf); member
106 * Unmap granted references of the shared buffer.
110 * shared by the frontend itself) or unmap the provided granted
118 if (buf->ops && buf->ops->unmap) in xen_front_pgdir_shbuf_unmap()
119 return buf->ops->unmap(buf); in xen_front_pgdir_shbuf_unmap()
121 /* No need to unmap own grant references. */ in xen_front_pgdir_shbuf_unmap()
194 * Unmap the buffer previously mapped with grant references
227 "Failed to unmap page %d: %d\n", in backend_unmap()
233 "Failed to unmap grant references, ret %d", ret); in backend_unmap()
[all …]
Darm-device.c70 goto unmap; in xen_map_device_mmio()
94 goto unmap; in xen_map_device_mmio()
99 unmap: in xen_map_device_mmio()
126 (action == BUS_NOTIFY_DEL_DEVICE ? "unmap" : "?"), in xen_platform_notifier()
168 (action == BUS_NOTIFY_DEL_DEVICE ? "unmap" : "?"), in xen_amba_notifier()
/linux-6.14.4/include/linux/
Dzpool.h24 * changed memory back out on unmap. Write-only does not copy
30 ZPOOL_MM_RO, /* read-only (no copy-out at unmap time) */
69 * @unmap: unmap a handle.
92 void (*unmap)(void *pool, unsigned long handle); member
/linux-6.14.4/include/net/libeth/
Dtx.h17 * @LIBETH_SQE_SLAB: kmalloc-allocated buffer, unmap and kfree()
18 * @LIBETH_SQE_FRAG: mapped skb frag, only unmap DMA
19 * @LIBETH_SQE_SKB: &sk_buff, unmap and napi_consume_skb(), update stats
35 * @dma: DMA address to unmap
36 * @len: length of the mapped region to unmap
/linux-6.14.4/drivers/dma/
Ddmaengine.c1321 #define __UNMAP_POOL(x) { .size = x, .name = "dmaengine-unmap-" __stringify(x) }
1354 struct dmaengine_unmap_data *unmap = container_of(kref, typeof(*unmap), kref); in dmaengine_unmap() local
1355 struct device *dev = unmap->dev; in dmaengine_unmap()
1358 cnt = unmap->to_cnt; in dmaengine_unmap()
1360 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
1362 cnt += unmap->from_cnt; in dmaengine_unmap()
1364 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
1366 cnt += unmap->bidi_cnt; in dmaengine_unmap()
1368 if (unmap->addr[i] == 0) in dmaengine_unmap()
1370 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
[all …]
Dmv_xor.c769 struct dmaengine_unmap_data *unmap; in mv_chan_memcpy_self_test() local
792 unmap = dmaengine_get_unmap_data(dma_chan->device->dev, 2, GFP_KERNEL); in mv_chan_memcpy_self_test()
793 if (!unmap) { in mv_chan_memcpy_self_test()
801 unmap->addr[0] = src_dma; in mv_chan_memcpy_self_test()
808 unmap->to_cnt = 1; in mv_chan_memcpy_self_test()
813 unmap->addr[1] = dest_dma; in mv_chan_memcpy_self_test()
820 unmap->from_cnt = 1; in mv_chan_memcpy_self_test()
821 unmap->len = PAGE_SIZE; in mv_chan_memcpy_self_test()
862 dmaengine_unmap_put(unmap); in mv_chan_memcpy_self_test()
880 struct dmaengine_unmap_data *unmap; in mv_chan_xor_self_test() local
[all …]
/linux-6.14.4/include/drm/
Ddrm_gpuvm.h807 * @DRM_GPUVA_OP_UNMAP: the unmap op type
857 * struct drm_gpuva_op_unmap - GPU VA unmap operation
859 * This structure represents a single unmap operation generated by the
864 * @va: the &drm_gpuva to unmap
888 * mapping(s), hence it consists of a maximum of two map and one unmap
891 * The @unmap operation takes care of removing the original existing mapping.
899 * unmap and map operations, is to give drivers the chance of extracting driver
900 * specific data for creating the new mappings from the unmap operations's
916 * @unmap: the unmap operation for the original existing mapping
918 struct drm_gpuva_op_unmap *unmap; member
[all …]
/linux-6.14.4/drivers/gpu/drm/
Ddrm_gpuvm.c107 * sequence of operations to satisfy a given map or unmap request.
116 * containing map, unmap and remap operations for a given newly requested
123 * of unmap operations, a maximum of two remap operations and a single map
131 * &drm_gpuva to unmap is physically contiguous with the original mapping
138 * one unmap operation and one or two map operations, such that drivers can
146 * call back into the driver in order to unmap a range of GPU VA space. The
148 * enclosed by the given range unmap operations are created. For mappings which
467 * &drm_gpuva_op in order to satisfy a given map or unmap request and how to
520 * va = op->remap.unmap->va;
550 * va = op->unmap->va;
[all …]
/linux-6.14.4/arch/x86/xen/
Dp2m.c701 struct gnttab_unmap_grant_ref unmap[2]; in set_foreign_p2m_mapping() local
728 unmap[0].host_addr = map_ops[i].host_addr; in set_foreign_p2m_mapping()
729 unmap[0].handle = map_ops[i].handle; in set_foreign_p2m_mapping()
732 unmap[0].dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping()
734 unmap[0].dev_bus_addr = 0; in set_foreign_p2m_mapping()
738 unmap[1].host_addr = kmap_ops[i].host_addr; in set_foreign_p2m_mapping()
739 unmap[1].handle = kmap_ops[i].handle; in set_foreign_p2m_mapping()
742 unmap[1].dev_bus_addr = kmap_ops[i].dev_bus_addr; in set_foreign_p2m_mapping()
744 unmap[1].dev_bus_addr = 0; in set_foreign_p2m_mapping()
751 unmap[0].status = 1; in set_foreign_p2m_mapping()
[all …]
/linux-6.14.4/drivers/scsi/
Dsd.h56 SD_LBP_UNMAP, /* Use UNMAP command */
57 SD_LBP_WS16, /* Use WRITE SAME(16) with UNMAP bit */
58 SD_LBP_WS10, /* Use WRITE SAME(10) with UNMAP bit */
66 SD_ZERO_WS16_UNMAP, /* Use WRITE SAME(16) with UNMAP */
67 SD_ZERO_WS10_UNMAP, /* Use WRITE SAME(10) with UNMAP */
196 case UNMAP: in scsi_medium_access_command()
/linux-6.14.4/drivers/gpu/drm/nouveau/
Dnouveau_exec.c25 * and unmap memory. Mappings may be flagged as sparse. Sparse mappings are not
35 * Requests to unmap a sparse mapping that still contains memory backed mappings
38 * Unmap requests are not bound to the range of existing mappings and can even
40 * make sure to unmap all memory backed mappings within the given range,
42 * within the given range. Unmap requests with the sparse flag set must match
50 * - unmap non-existent sparse mappings
51 * - unmap a sparse mapping and map a new sparse mapping overlapping the range
53 * - unmap a sparse mapping and map new memory backed mappings overlapping the
Dnouveau_uvmm.c12 * Actual map/unmap operations within the fence signalling critical path are
15 * list in order to map/unmap it's entries, can't occur concurrently.
472 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare_unwind()
484 op_unmap_prepare_unwind(op->unmap.va); in nouveau_uvmm_sm_prepare_unwind()
491 /* Unmap operation don't allocate page tables, hence skip the following in nouveau_uvmm_sm_prepare_unwind()
509 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare_unwind()
526 struct drm_gpuva_op_unmap *u = &op->unmap; in nouveau_uvmm_sm_prepare_unwind()
647 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare()
656 op_unmap_prepare(r->unmap); in nouveau_uvmm_sm_prepare()
687 struct drm_gpuva_op_unmap *u = &op->unmap; in nouveau_uvmm_sm_prepare()
[all …]
/linux-6.14.4/drivers/vfio/
Dvfio_iommu_type1.c990 * Sync if the number of fast-unmap regions hits the limit in unmap_unpin_fast()
1065 * largest contiguous physical memory chunk to unmap. in vfio_unmap_unpin()
1075 * First, try to use fast unmap/unpin. In case of failure, in vfio_unmap_unpin()
1076 * switch to slow unmap/unpin path. in vfio_unmap_unpin()
1130 * That way the user will be able to map/unmap buffers whose size/ in vfio_update_pgsize_bitmap()
1235 * and unmap iovas within the range we're about to unmap. Drivers MUST unpin
1263 struct vfio_iommu_type1_dma_unmap *unmap, in vfio_dma_do_unmap() argument
1270 dma_addr_t iova = unmap->iova; in vfio_dma_do_unmap()
1271 u64 size = unmap->size; in vfio_dma_do_unmap()
1272 bool unmap_all = unmap->flags & VFIO_DMA_UNMAP_FLAG_ALL; in vfio_dma_do_unmap()
[all …]
/linux-6.14.4/drivers/gpu/drm/msm/
Dmsm_gem_vma.c41 /* Actually unmap memory for the vma */
51 aspace->mmu->funcs->unmap(aspace->mmu, vma->iova, size); in msm_gem_vma_purge()
77 * a lock across map/unmap which is also used in the job_run() in msm_gem_vma_map()
82 * for the pgtable in map/unmap ops. in msm_gem_vma_map()

12345678910>>...38