Lines Matching refs:sec4_sg_bytes
902 int sec4_sg_bytes; member
929 int sec4_sg_bytes; member
940 int sec4_sg_bytes) in caam_unmap() argument
953 if (sec4_sg_bytes) in caam_unmap()
954 dma_unmap_single(dev, sec4_sg_dma, sec4_sg_bytes, in caam_unmap()
964 edesc->sec4_sg_dma, edesc->sec4_sg_bytes); in aead_unmap()
976 edesc->sec4_sg_dma, edesc->sec4_sg_bytes); in skcipher_unmap()
1014 return PTR_ALIGN((u8 *)edesc->sec4_sg + edesc->sec4_sg_bytes, in skcipher_edesc_iv()
1318 int sec4_sg_index, sec4_sg_len, sec4_sg_bytes; in aead_edesc_alloc() local
1396 sec4_sg_bytes = sec4_sg_len * sizeof(struct sec4_sg_entry); in aead_edesc_alloc()
1399 edesc = kzalloc(sizeof(*edesc) + desc_bytes + sec4_sg_bytes, flags); in aead_edesc_alloc()
1428 if (!sec4_sg_bytes) in aead_edesc_alloc()
1432 sec4_sg_bytes, DMA_TO_DEVICE); in aead_edesc_alloc()
1440 edesc->sec4_sg_bytes = sec4_sg_bytes; in aead_edesc_alloc()
1626 int dst_sg_idx, sec4_sg_ents, sec4_sg_bytes; in skcipher_edesc_alloc() local
1696 sec4_sg_bytes = sec4_sg_ents * sizeof(struct sec4_sg_entry); in skcipher_edesc_alloc()
1701 aligned_size = sizeof(*edesc) + desc_bytes + sec4_sg_bytes; in skcipher_edesc_alloc()
1718 edesc->sec4_sg_bytes = sec4_sg_bytes; in skcipher_edesc_alloc()
1755 if (sec4_sg_bytes) { in skcipher_edesc_alloc()
1757 sec4_sg_bytes, in skcipher_edesc_alloc()
1772 sec4_sg_bytes, 1); in skcipher_edesc_alloc()