Lines Matching full:buffers

29 		dma_unmap_single(dev, bl->buffers[i].addr,  in qat_bl_free_bufl()
30 bl->buffers[i].len, bl_dma_dir); in qat_bl_free_bufl()
39 dma_unmap_single(dev, blout->buffers[i].addr, in qat_bl_free_bufl()
40 blout->buffers[i].len, in qat_bl_free_bufl()
68 size_t sz_out, sz = struct_size(bufl, buffers, n); in __qat_bl_sgl_to_bufl()
93 bufl->buffers[i].addr = DMA_MAPPING_ERROR; in __qat_bl_sgl_to_bufl()
107 bufl->buffers[y].addr = dma_map_single(dev, sg_virt(sg) + left, in __qat_bl_sgl_to_bufl()
110 bufl->buffers[y].len = sg->length; in __qat_bl_sgl_to_bufl()
111 if (unlikely(dma_mapping_error(dev, bufl->buffers[y].addr))) in __qat_bl_sgl_to_bufl()
115 bufl->buffers[y].len -= left; in __qat_bl_sgl_to_bufl()
128 struct qat_alg_buf *buffers; in __qat_bl_sgl_to_bufl() local
133 sz_out = struct_size(buflout, buffers, n); in __qat_bl_sgl_to_bufl()
149 buffers = buflout->buffers; in __qat_bl_sgl_to_bufl()
151 buffers[i].addr = DMA_MAPPING_ERROR; in __qat_bl_sgl_to_bufl()
163 buffers[y].addr = dma_map_single(dev, sg_virt(sg) + left, in __qat_bl_sgl_to_bufl()
166 if (unlikely(dma_mapping_error(dev, buffers[y].addr))) in __qat_bl_sgl_to_bufl()
168 buffers[y].len = sg->length; in __qat_bl_sgl_to_bufl()
171 buffers[y].len -= left; in __qat_bl_sgl_to_bufl()
176 buffers[sg_nctr].addr = extra_dst_buff; in __qat_bl_sgl_to_bufl()
177 buffers[sg_nctr].len = sz_extra_dst_buff; in __qat_bl_sgl_to_bufl()
202 if (buflout->buffers[i].addr == extra_dst_buff) in __qat_bl_sgl_to_bufl()
204 if (!dma_mapping_error(dev, buflout->buffers[i].addr)) in __qat_bl_sgl_to_bufl()
205 dma_unmap_single(dev, buflout->buffers[i].addr, in __qat_bl_sgl_to_bufl()
206 buflout->buffers[i].len, in __qat_bl_sgl_to_bufl()
219 if (!dma_mapping_error(dev, bufl->buffers[i].addr)) in __qat_bl_sgl_to_bufl()
220 dma_unmap_single(dev, bufl->buffers[i].addr, in __qat_bl_sgl_to_bufl()
221 bufl->buffers[i].len, in __qat_bl_sgl_to_bufl()
263 if (!dma_mapping_error(dev, bl->buffers[i].addr)) in qat_bl_sgl_unmap()
264 dma_unmap_single(dev, bl->buffers[i].addr, in qat_bl_sgl_unmap()
265 bl->buffers[i].len, DMA_FROM_DEVICE); in qat_bl_sgl_unmap()
280 sz = struct_size(bufl, buffers, n); in qat_bl_sgl_map()
286 bufl->buffers[i].addr = DMA_MAPPING_ERROR; in qat_bl_sgl_map()
295 bufl->buffers[y].addr = dma_map_single(dev, sg_virt(sg), in qat_bl_sgl_map()
298 bufl->buffers[y].len = sg->length; in qat_bl_sgl_map()
299 if (unlikely(dma_mapping_error(dev, bufl->buffers[y].addr))) in qat_bl_sgl_map()
312 if (!dma_mapping_error(dev, bufl->buffers[i].addr)) in qat_bl_sgl_map()
313 dma_unmap_single(dev, bufl->buffers[i].addr, in qat_bl_sgl_map()
314 bufl->buffers[i].len, in qat_bl_sgl_map()
383 new_bl_size = struct_size(new_bl, buffers, new_bl->num_bufs); in qat_bl_realloc_map_new_dst()