| /crypto/ |
| A D | cts.c | 61 unsigned offset; member 106 unsigned int offset; in cts_cbc_encrypt() local 109 offset = rctx->offset; in cts_cbc_encrypt() 151 unsigned int offset; in crypto_cts_encrypt() local 168 rctx->offset = offset; in crypto_cts_encrypt() 173 offset, req->iv); in crypto_cts_encrypt() 187 unsigned int offset; in cts_cbc_decrypt() local 191 offset = rctx->offset; in cts_cbc_decrypt() 245 unsigned int offset; in crypto_cts_decrypt() local 268 rctx->offset = offset; in crypto_cts_decrypt() [all …]
|
| A D | scatterwalk.c | 37 nbytes += walk->offset - sg->offset; in scatterwalk_skip() 44 walk->offset = sg->offset + nbytes; in scatterwalk_skip() 143 sg_set_page(dst, sg_page(src), src->length - len, src->offset + len); in scatterwalk_ffwd() 201 diff = offset_in_page(walk->in.offset) - in skcipher_next_fast() 202 offset_in_page(walk->out.offset); in skcipher_next_fast() 203 diff |= (u8 *)(sg_page(walk->in.sg) + (walk->in.offset >> PAGE_SHIFT)) - in skcipher_next_fast() 204 (u8 *)(sg_page(walk->out.sg) + (walk->out.offset >> PAGE_SHIFT)); in skcipher_next_fast() 236 if (unlikely((walk->in.offset | walk->out.offset) & walk->alignmask)) { in skcipher_walk_next()
|
| A D | ahash.c | 75 unsigned int offset = walk->offset; in hash_walk_next() local 77 ((unsigned int)(PAGE_SIZE)) - offset); in hash_walk_next() 80 walk->data += offset; in hash_walk_next() 90 walk->offset = sg->offset; in hash_walk_new_entry() 92 walk->offset = offset_in_page(walk->offset); in hash_walk_new_entry() 130 walk->data -= walk->offset; in crypto_hash_walk_done() 139 walk->offset = 0; in crypto_hash_walk_done() 209 unsigned int offset; in shash_ahash_digest() local 224 offset = sg->offset; in shash_ahash_digest() 225 data = lowmem_page_address(page) + offset; in shash_ahash_digest() [all …]
|
| A D | xcbc.c | 93 unsigned int offset = 0; in crypto_xcbc_digest_finup() local 98 offset += bs; in crypto_xcbc_digest_finup() 100 crypto_xor(prev, &tctx->consts[offset], bs); in crypto_xcbc_digest_finup()
|
| A D | cmac.c | 125 unsigned int offset = 0; in crypto_cmac_digest_finup() local 130 offset += bs; in crypto_cmac_digest_finup() 132 crypto_xor(prev, (const u8 *)tctx->consts + offset, bs); in crypto_cmac_digest_finup()
|
| A D | af_alg.c | 646 unsigned int af_alg_count_tsgl(struct sock *sk, size_t bytes, size_t offset) in af_alg_count_tsgl() argument 664 if (offset >= sg[i].length) { in af_alg_count_tsgl() 665 offset -= sg[i].length; in af_alg_count_tsgl() 670 bytes_count = sg[i].length - offset; in af_alg_count_tsgl() 672 offset = 0; in af_alg_count_tsgl() 736 sg[i].offset + dst_offset); in af_alg_pull_tsgl() 743 sg[i].offset += plen; in af_alg_pull_tsgl() 1004 PAGE_SIZE - sg->offset - sg->length); in af_alg_sendmsg() 1007 sg->offset + sg->length, in af_alg_sendmsg() 1013 ctx->merge = (sg->offset + sg->length) & in af_alg_sendmsg()
|
| A D | md5.c | 163 unsigned int offset; in md5_finup() local 169 offset = len; in md5_finup() 170 p = (char *)block + offset; in md5_finup() 171 padding = 56 - (offset + 1); in md5_finup()
|
| A D | testmgr.c | 276 unsigned int offset; member 356 .offset = 1, 394 .offset = PAGE_SIZE - 32 397 .offset = PAGE_SIZE - 7 426 .offset = 1, 458 .offset = PAGE_SIZE - 7, 635 unsigned int offset = partitions[i].div->offset; in build_test_sglist() local 639 offset += alignmask; in build_test_sglist() 645 offset /= 2; in build_test_sglist() 697 unsigned int offset = sg->offset; in verify_correct_output() local [all …]
|
| A D | md4.c | 187 const unsigned int offset = mctx->byte_count & 0x3f; in md4_final() local 188 char *p = (char *)mctx->block + offset; in md4_final() 189 int padding = 56 - (offset + 1); in md4_final()
|
| A D | acompress.c | 509 walk->in.offset); in acomp_walk_next_src() 526 walk->out.offset); in acomp_walk_next_dst() 556 walk->in.offset = 0; in acomp_walk_virt() 561 walk->out.offset = 0; in acomp_walk_virt()
|
| A D | adiantum.c | 301 if (dst_nents == 1 && dst->offset + req->cryptlen <= PAGE_SIZE) { in adiantum_finish() 304 void *virt = kmap_local_page(page) + dst->offset; in adiantum_finish() 362 if (src_nents == 1 && src->offset + req->cryptlen <= PAGE_SIZE) { in adiantum_crypt() 364 void *virt = kmap_local_page(sg_page(src)) + src->offset; in adiantum_crypt()
|
| A D | xts.c | 164 int offset = req->cryptlen & ~(XTS_BLOCK_SIZE - 1); in xts_cts_final() local 172 offset - XTS_BLOCK_SIZE); in xts_cts_final() 176 scatterwalk_map_and_copy(b, req->src, offset, tail, 0); in xts_cts_final()
|
| A D | scompress.c | 197 doff = req->dst->offset; in scomp_acomp_comp_decomp() 219 soff = req->src->offset; in scomp_acomp_comp_decomp()
|
| A D | essiv.c | 235 req->src->offset); in essiv_aead_crypt()
|
| /crypto/async_tx/ |
| A D | async_xor.c | 100 do_sync_xor_offs(struct page *dest, unsigned int offset, in do_sync_xor_offs() argument 119 (src_offs ? src_offs[i] : offset); in do_sync_xor_offs() 122 dest_buf = page_address(dest) + offset; in do_sync_xor_offs() 146 if (!is_dma_xor_aligned(device, offset, 0, len)) in dma_xor_aligned_offsets() 182 async_xor_offs(struct page *dest, unsigned int offset, in async_xor_offs() argument 197 if (unmap && dma_xor_aligned_offsets(device, offset, in async_xor_offs() 211 src_offs ? src_offs[i] : offset, in async_xor_offs() 243 do_sync_xor_offs(dest, offset, src_list, src_offs, in async_xor_offs() 276 return async_xor_offs(dest, offset, src_list, NULL, in async_xor() 283 return !memchr_inv(page_address(p) + offset, 0, len); in page_is_zero() [all …]
|
| /crypto/krb5/ |
| A D | rfc6803_camellia.c | 34 size_t bsize, offset, seg; in rfc6803_calc_KDF_FEEDBACK_CMAC() local 81 offset = 0; in rfc6803_calc_KDF_FEEDBACK_CMAC() 96 seg = min_t(size_t, result->len - offset, K.len); in rfc6803_calc_KDF_FEEDBACK_CMAC() 97 memcpy(result->data + offset, K.data, seg); in rfc6803_calc_KDF_FEEDBACK_CMAC() 98 offset += seg; in rfc6803_calc_KDF_FEEDBACK_CMAC() 99 } while (offset < result->len); in rfc6803_calc_KDF_FEEDBACK_CMAC()
|
| A D | selftest.c | 45 unsigned int off = sg->offset, len = umin(sg->length, limit); 385 size_t offset, len, message_len; in krb5_test_one_mic() local 419 len, &offset); in krb5_test_one_mic() 427 memcpy(buf + offset, plain.data, plain.len); in krb5_test_one_mic() 456 offset = 0; in krb5_test_one_mic() 457 ret = crypto_krb5_verify_mic(krb5, ci, NULL, sg, 1, &offset, &len); in krb5_test_one_mic() 464 if (CHECK(offset != mic.len) || in krb5_test_one_mic() 468 if (memcmp(buf + offset, plain.data, plain.len) != 0) { in krb5_test_one_mic() 471 pr_warn("BUF %*phN\n", plain.len, buf + offset); in krb5_test_one_mic()
|
| A D | rfc3961_simplified.c | 84 size_t offset, size_t len) in crypto_shash_update_sg() argument 92 sg_miter_skip(&miter, offset); in crypto_shash_update_sg() 723 size_t bsize, data_offset, data_len, offset = *_offset, len = *_len; in rfc3961_verify_mic() local 730 data_offset = offset + krb5->cksum_len; in rfc3961_verify_mic() 762 done = sg_pcopy_to_buffer(sg, nr_sg, cksum2, krb5->cksum_len, offset); in rfc3961_verify_mic()
|
| A D | internal.h | 174 size_t offset, size_t len);
|