/linux-4.19.296/crypto/ |
D | scatterwalk.c | 89 src = sg_next(src); in scatterwalk_ffwd() 94 scatterwalk_crypto_chain(dst, sg_next(src), 2); in scatterwalk_ffwd()
|
D | ccm.c | 311 err = crypto_ccm_auth(req, sg_next(pctx->src), cryptlen); in crypto_ccm_encrypt() 328 scatterwalk_map_and_copy(odata, sg_next(dst), cryptlen, in crypto_ccm_encrypt() 345 dst = sg_next(req->src == req->dst ? pctx->src : pctx->dst); in crypto_ccm_decrypt_done() 375 scatterwalk_map_and_copy(authtag, sg_next(pctx->src), cryptlen, in crypto_ccm_decrypt() 392 err = crypto_ccm_auth(req, sg_next(dst), cryptlen); in crypto_ccm_decrypt()
|
D | lrw.c | 196 scatterwalk_crypto_chain(rctx->dst, sg_next(sg), 2); in post_crypt() 273 scatterwalk_crypto_chain(rctx->src, sg_next(sg), 2); in pre_crypt()
|
D | xts.c | 141 scatterwalk_crypto_chain(rctx->dst, sg_next(sg), 2); in post_crypt() 207 scatterwalk_crypto_chain(rctx->src, sg_next(sg), 2); in pre_crypt()
|
D | keywrap.c | 123 sg = sg_next(sg); in crypto_kw_scatterlist_ff()
|
D | ablkcipher.c | 86 scatterwalk_start(&walk->out, sg_next(walk->out.sg)); in ablkcipher_done_slow()
|
D | ahash.c | 128 walk->sg = sg_next(walk->sg); in crypto_hash_walk_done()
|
D | gcm.c | 451 gctx->src = sg_next(req->src == req->dst ? pctx->src : pctx->dst); in gcm_encrypt_continue() 536 gctx->src = sg_next(pctx->src); in crypto_gcm_decrypt()
|
/linux-4.19.296/lib/ |
D | scatterlist.c | 25 struct scatterlist *sg_next(struct scatterlist *sg) in sg_next() function 36 EXPORT_SYMBOL(sg_next); 50 for (nents = 0; sg; sg = sg_next(sg)) in sg_nents() 78 for (nents = 0, total = 0; sg; sg = sg_next(sg)) { in sg_nents_for_len() 515 sg = sg_next(sg); in sgl_alloc_order() 619 piter->sg = sg_next(piter->sg); in __sg_page_iter_next()
|
D | sg_split.c | 98 in_sg = sg_next(in_sg); in sg_split_phys() 121 in_sg = sg_next(in_sg); in sg_split_mapped()
|
D | kfifo.c | 336 sgl = sg_next(sgl); in setup_sgl_buf()
|
/linux-4.19.296/include/linux/ |
D | scatterlist.h | 152 for (__i = 0, sg = (sglist); __i < (nr); __i++, sg = sg_next(sg)) 256 struct scatterlist *sg_next(struct scatterlist *);
|
/linux-4.19.296/include/crypto/ |
D | scatterwalk.h | 98 scatterwalk_start(walk, sg_next(walk->sg)); in scatterwalk_pagedone()
|
/linux-4.19.296/drivers/s390/scsi/ |
D | zfcp_qdio.h | 235 for (; sg; sg = sg_next(sg)) in zfcp_qdio_real_bytes()
|
D | zfcp_qdio.c | 195 for (; sg; sg = sg_next(sg)) { in zfcp_qdio_sbals_from_sg()
|
D | zfcp_dbf.c | 454 for (; sg && pay_sum < cap_len; sg = sg_next(sg)) { in zfcp_dbf_san()
|
/linux-4.19.296/drivers/media/common/videobuf2/ |
D | videobuf2-dma-sg.c | 402 rd = sg_next(rd); in vb2_dma_sg_dmabuf_ops_attach() 403 wr = sg_next(wr); in vb2_dma_sg_dmabuf_ops_attach()
|
D | videobuf2-dma-contig.c | 253 rd = sg_next(rd); in vb2_dc_dmabuf_ops_attach() 254 wr = sg_next(wr); in vb2_dc_dmabuf_ops_attach()
|
/linux-4.19.296/drivers/virtio/ |
D | virtio_ring.c | 339 for (sg = sgs[n]; sg; sg = sg_next(sg)) { in virtqueue_add() 352 for (sg = sgs[n]; sg; sg = sg_next(sg)) { in virtqueue_add() 463 for (sg = sgs[i]; sg; sg = sg_next(sg)) in virtqueue_add_sgs()
|
/linux-4.19.296/block/ |
D | blk-merge.c | 396 *sg = sg_next(*sg); in __blk_segment_map_sg() 460 sg = sg_next(sg); in blk_rq_map_sg()
|
D | blk-integrity.c | 114 sg = sg_next(sg); in blk_rq_map_integrity_sg()
|
/linux-4.19.296/drivers/memstick/core/ |
D | ms_block.c | 51 sg_from = sg_next(sg_from); in msb_sg_copy() 65 sg_to = sg_next(sg_to); in msb_sg_copy() 67 sg_from = sg_next(sg_from); in msb_sg_copy() 80 sg_from = sg_next(sg_from); in msb_sg_copy() 81 sg_to = sg_next(sg_to); in msb_sg_copy()
|
/linux-4.19.296/drivers/message/fusion/ |
D | mptscsih.c | 239 sg = sg_next(sg); in mptscsih_AddSGE() 248 sg = sg_next(sg); in mptscsih_AddSGE()
|
/linux-4.19.296/drivers/misc/mic/scif/ |
D | scif_rma.c | 572 while ((sg = sg_next(sg)) && sg_dma_address(sg) == last_da) { in scif_iommu_map()
|