Lines Matching refs:rctx
23 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_free_data_blocks() local
29 list_for_each_entry_safe(block, tmp, &rctx->blocks, list) { in eip93_hash_free_data_blocks()
34 if (!list_empty(&rctx->blocks)) in eip93_hash_free_data_blocks()
35 INIT_LIST_HEAD(&rctx->blocks); in eip93_hash_free_data_blocks()
37 if (rctx->finalize) in eip93_hash_free_data_blocks()
38 dma_unmap_single(eip93->dev, rctx->data_dma, in eip93_hash_free_data_blocks()
39 rctx->data_used, in eip93_hash_free_data_blocks()
45 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_free_sa_record() local
51 dma_unmap_single(eip93->dev, rctx->sa_record_hmac_base, in eip93_hash_free_sa_record()
52 sizeof(rctx->sa_record_hmac), DMA_TO_DEVICE); in eip93_hash_free_sa_record()
54 dma_unmap_single(eip93->dev, rctx->sa_record_base, in eip93_hash_free_sa_record()
55 sizeof(rctx->sa_record), DMA_TO_DEVICE); in eip93_hash_free_sa_record()
61 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_handle_result() local
65 struct sa_state *sa_state = &rctx->sa_state; in eip93_hash_handle_result()
69 dma_unmap_single(eip93->dev, rctx->sa_state_base, in eip93_hash_handle_result()
76 if (rctx->partial_hash) in eip93_hash_handle_result()
79 if (rctx->finalize || rctx->partial_hash) { in eip93_hash_handle_result()
137 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_export_sa_state() local
138 struct sa_state *sa_state = &rctx->sa_state; in eip93_hash_export_sa_state()
147 if (!rctx->len) in eip93_hash_export_sa_state()
154 state->len = rctx->len; in eip93_hash_export_sa_state()
155 state->data_used = rctx->data_used; in eip93_hash_export_sa_state()
160 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in __eip93_hash_init() local
163 struct sa_record *sa_record = &rctx->sa_record; in __eip93_hash_init()
191 struct sa_record *sa_record_hmac = &rctx->sa_record_hmac; in __eip93_hash_init()
201 rctx->len = 0; in __eip93_hash_init()
202 rctx->data_used = 0; in __eip93_hash_init()
203 rctx->partial_hash = false; in __eip93_hash_init()
204 rctx->finalize = false; in __eip93_hash_init()
205 INIT_LIST_HEAD(&rctx->blocks); in __eip93_hash_init()
212 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_send_hash_req() local
228 cdesc.sa_addr = rctx->sa_record_base; in eip93_send_hash_req()
231 cdesc.state_addr = rctx->sa_state_base; in eip93_send_hash_req()
243 if (rctx->finalize && !rctx->partial_hash) { in eip93_send_hash_req()
246 struct sa_record *sa_record_hmac = &rctx->sa_record_hmac; in eip93_send_hash_req()
248 rctx->sa_record_hmac_base = dma_map_single(eip93->dev, in eip93_send_hash_req()
252 ret = dma_mapping_error(eip93->dev, rctx->sa_record_hmac_base); in eip93_send_hash_req()
256 cdesc.sa_addr = rctx->sa_record_hmac_base; in eip93_send_hash_req()
288 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_init() local
291 struct sa_state *sa_state = &rctx->sa_state; in eip93_hash_init()
301 memcpy(rctx->data, ctx->ipad, SHA256_BLOCK_SIZE); in eip93_hash_init()
303 rctx->data_used = SHA256_BLOCK_SIZE; in eip93_hash_init()
304 rctx->len += SHA256_BLOCK_SIZE; in eip93_hash_init()
317 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in __eip93_hash_update() local
327 offset = rctx->data_used; in __eip93_hash_update()
350 memcpy(block->data, rctx->data, offset); in __eip93_hash_update()
355 list_add(&block->list, &rctx->blocks); in __eip93_hash_update()
362 rctx->data + offset, to_consume, in __eip93_hash_update()
364 rctx->data_used = offset + read; in __eip93_hash_update()
367 rctx->len += read + consumed; in __eip93_hash_update()
370 list_for_each_entry_reverse(block, &rctx->blocks, list) { in __eip93_hash_update()
372 list_is_first(&block->list, &rctx->blocks); in __eip93_hash_update()
391 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_update() local
394 struct sa_record *sa_record = &rctx->sa_record; in eip93_hash_update()
395 struct sa_state *sa_state = &rctx->sa_state; in eip93_hash_update()
402 rctx->sa_state_base = dma_map_single(eip93->dev, sa_state, in eip93_hash_update()
405 ret = dma_mapping_error(eip93->dev, rctx->sa_state_base); in eip93_hash_update()
409 rctx->sa_record_base = dma_map_single(eip93->dev, sa_record, in eip93_hash_update()
412 ret = dma_mapping_error(eip93->dev, rctx->sa_record_base); in eip93_hash_update()
423 dma_unmap_single(eip93->dev, rctx->sa_record_base, in eip93_hash_update()
427 dma_unmap_single(eip93->dev, rctx->sa_state_base, in eip93_hash_update()
439 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in __eip93_hash_final() local
443 struct sa_record *sa_record = &rctx->sa_record; in __eip93_hash_final()
444 struct sa_state *sa_state = &rctx->sa_state; in __eip93_hash_final()
449 if (!rctx->len && !IS_HMAC(ctx->flags)) { in __eip93_hash_final()
475 rctx->finalize = true; in __eip93_hash_final()
478 rctx->sa_state_base = dma_map_single(eip93->dev, sa_state, in __eip93_hash_final()
481 ret = dma_mapping_error(eip93->dev, rctx->sa_state_base); in __eip93_hash_final()
485 rctx->sa_record_base = dma_map_single(eip93->dev, sa_record, in __eip93_hash_final()
488 ret = dma_mapping_error(eip93->dev, rctx->sa_record_base); in __eip93_hash_final()
494 ret = eip93_send_hash_req(async, rctx->data, &rctx->data_dma, in __eip93_hash_final()
495 rctx->data_used, true); in __eip93_hash_final()
504 dma_unmap_single(eip93->dev, rctx->sa_record_base, in __eip93_hash_final()
508 dma_unmap_single(eip93->dev, rctx->sa_state_base, in __eip93_hash_final()
521 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_finup() local
524 struct sa_record *sa_record = &rctx->sa_record; in eip93_hash_finup()
525 struct sa_state *sa_state = &rctx->sa_state; in eip93_hash_finup()
529 if (rctx->len + req->nbytes || IS_HMAC(ctx->flags)) { in eip93_hash_finup()
530 rctx->sa_state_base = dma_map_single(eip93->dev, sa_state, in eip93_hash_finup()
533 ret = dma_mapping_error(eip93->dev, rctx->sa_state_base); in eip93_hash_finup()
537 rctx->sa_record_base = dma_map_single(eip93->dev, sa_record, in eip93_hash_finup()
540 ret = dma_mapping_error(eip93->dev, rctx->sa_record_base); in eip93_hash_finup()
552 dma_unmap_single(eip93->dev, rctx->sa_record_base, in eip93_hash_finup()
555 dma_unmap_single(eip93->dev, rctx->sa_state_base, in eip93_hash_finup()
600 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_import() local
602 struct sa_state *sa_state = &rctx->sa_state; in eip93_hash_import()
609 rctx->len = state->len; in eip93_hash_import()
610 rctx->data_used = state->data_used; in eip93_hash_import()
613 if (rctx->len) in eip93_hash_import()
614 memcpy(rctx->data, state->data, rctx->data_used); in eip93_hash_import()
621 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_export() local
625 if (rctx->len) in eip93_hash_export()
626 memcpy(state->data, rctx->data, rctx->data_used); in eip93_hash_export()