| /linux/drivers/s390/cio/ |
| A D | airq.c | 132 iv = kzalloc(sizeof(*iv), GFP_KERNEL); in airq_iv_create() 133 if (!iv) in airq_iv_create() 160 iv->end = 0; in airq_iv_create() 181 return iv; in airq_iv_create() 187 if (iv->flags & AIRQ_IV_CACHELINE && iv->vector) in airq_iv_create() 191 kfree(iv); in airq_iv_create() 209 cio_dma_free(iv->vector, iv_size(iv->bits)); in airq_iv_release() 211 kfree(iv); in airq_iv_release() 231 bit = find_first_bit_inv(iv->avail, iv->bits); in airq_iv_alloc() 274 while (iv->end > 0 && !test_bit_inv(iv->end - 1, iv->avail)) in airq_iv_free() [all …]
|
| /linux/arch/s390/include/asm/ |
| A D | airq.h | 53 void airq_iv_release(struct airq_iv *iv); 61 return airq_iv_alloc(iv, 1); in airq_iv_alloc_bit() 66 airq_iv_free(iv, bit, 1); in airq_iv_free_bit() 69 static inline unsigned long airq_iv_end(struct airq_iv *iv) in airq_iv_end() argument 71 return iv->end; in airq_iv_end() 77 bit_spin_lock(bit ^ be_to_le, iv->bitlock); in airq_iv_lock() 83 bit_spin_unlock(bit ^ be_to_le, iv->bitlock); in airq_iv_unlock() 89 iv->data[bit] = data; in airq_iv_set_data() 95 return iv->data[bit]; in airq_iv_get_data() 101 iv->ptr[bit] = ptr; in airq_iv_set_ptr() [all …]
|
| /linux/crypto/ |
| A D | testmgr.h | 68 const char *iv; member 109 const char *iv; member 12199 .iv = zeroed_string, 12231 .iv = zeroed_string, 17149 .iv = "", 17168 .iv = "", 17461 .iv = "", 17482 .iv = "", 21248 .iv = zeroed_string, 21863 .iv = zeroed_string, [all …]
|
| A D | cbc.c | 17 u8 *iv) in crypto_cbc_encrypt_segment() argument 22 crypto_xor(iv, src, bsize); in crypto_cbc_encrypt_segment() 24 memcpy(iv, dst, bsize); in crypto_cbc_encrypt_segment() 34 u8 *iv = oiv; in crypto_cbc_encrypt_inplace() local 40 crypto_xor(src, iv, bsize); in crypto_cbc_encrypt_inplace() 42 iv = src; in crypto_cbc_encrypt_inplace() 47 memcpy(oiv, iv, bsize); in crypto_cbc_encrypt_inplace() 74 const u8 *iv = oiv; in crypto_cbc_decrypt_segment() local 82 iv = src; in crypto_cbc_decrypt_segment() 88 memcpy(oiv, iv, bsize); in crypto_cbc_decrypt_segment() [all …]
|
| A D | pcbc.c | 28 u8 * const iv = walk->iv; in crypto_pcbc_encrypt_segment() local 31 crypto_xor(iv, src, bsize); in crypto_pcbc_encrypt_segment() 32 crypto_cipher_encrypt_one(tfm, dst, iv); in crypto_pcbc_encrypt_segment() 33 crypto_xor_cpy(iv, dst, src, bsize); in crypto_pcbc_encrypt_segment() 49 u8 * const iv = walk->iv; in crypto_pcbc_encrypt_inplace() local 54 crypto_xor(iv, src, bsize); in crypto_pcbc_encrypt_inplace() 95 u8 * const iv = walk->iv; in crypto_pcbc_decrypt_segment() local 99 crypto_xor(dst, iv, bsize); in crypto_pcbc_decrypt_segment() 100 crypto_xor_cpy(iv, dst, src, bsize); in crypto_pcbc_decrypt_segment() 116 u8 * const iv = walk->iv; in crypto_pcbc_decrypt_inplace() local [all …]
|
| /linux/arch/riscv/crypto/ |
| A D | chacha-riscv64-glue.c | 17 size_t len, const u32 iv[4]); 21 u32 iv[CHACHA_IV_SIZE / sizeof(u32)]; in riscv64_chacha20_crypt() local 30 iv[0] = get_unaligned_le32(req->iv); in riscv64_chacha20_crypt() 31 iv[1] = get_unaligned_le32(req->iv + 4); in riscv64_chacha20_crypt() 32 iv[2] = get_unaligned_le32(req->iv + 8); in riscv64_chacha20_crypt() 33 iv[3] = get_unaligned_le32(req->iv + 12); in riscv64_chacha20_crypt() 42 walk.dst.virt.addr, nbytes, iv); in riscv64_chacha20_crypt() 43 iv[0] += nbytes / CHACHA_BLOCK_SIZE; in riscv64_chacha20_crypt() 49 CHACHA_BLOCK_SIZE, iv); in riscv64_chacha20_crypt()
|
| /linux/drivers/crypto/nx/ |
| A D | nx-aes-ccm.c | 126 if (1 > iv[0] || iv[0] > 7) in crypto_ccm_check_iv() 138 memcpy(b0, iv, 16); in generate_b0() 169 memset(iv + 15 - iv[0], 0, iv[0] + 1); in generate_pat() 327 u8 *iv, in ccm_nx_decrypt() argument 405 u8 *iv, in ccm_nx_encrypt() argument 481 u8 *iv = rctx->iv; in ccm4309_aes_nx_encrypt() local 483 iv[0] = 3; in ccm4309_aes_nx_encrypt() 485 memcpy(iv + 4, req->iv, 8); in ccm4309_aes_nx_encrypt() 505 u8 *iv = rctx->iv; in ccm4309_aes_nx_decrypt() local 507 iv[0] = 3; in ccm4309_aes_nx_decrypt() [all …]
|
| A D | nx-aes-ctr.c | 72 static int ctr_aes_nx_crypt(struct skcipher_request *req, u8 *iv) in ctr_aes_nx_crypt() argument 86 rc = nx_build_sg_lists(nx_ctx, iv, req->dst, req->src, in ctr_aes_nx_crypt() 88 csbcpb->cpb.aes_ctr.iv); in ctr_aes_nx_crypt() 102 memcpy(iv, csbcpb->cpb.aes_cbc.cv, AES_BLOCK_SIZE); in ctr_aes_nx_crypt() 119 u8 iv[16]; in ctr3686_aes_nx_crypt() local 121 memcpy(iv, nx_ctx->priv.ctr.nonce, CTR_RFC3686_NONCE_SIZE); in ctr3686_aes_nx_crypt() 122 memcpy(iv + CTR_RFC3686_NONCE_SIZE, req->iv, CTR_RFC3686_IV_SIZE); in ctr3686_aes_nx_crypt() 123 iv[12] = iv[13] = iv[14] = 0; in ctr3686_aes_nx_crypt() 124 iv[15] = 1; in ctr3686_aes_nx_crypt() 126 return ctr_aes_nx_crypt(req, iv); in ctr3686_aes_nx_crypt()
|
| A D | nx-aes-gcm.c | 326 rc = gcm_empty(req, rctx->iv, enc); in gcm_aes_nx_crypt() 328 rc = gmac(req, rctx->iv, assoclen); in gcm_aes_nx_crypt() 419 char *iv = rctx->iv; in gcm_aes_nx_encrypt() local 421 memcpy(iv, req->iv, GCM_AES_IV_SIZE); in gcm_aes_nx_encrypt() 429 char *iv = rctx->iv; in gcm_aes_nx_decrypt() local 431 memcpy(iv, req->iv, GCM_AES_IV_SIZE); in gcm_aes_nx_decrypt() 441 char *iv = rctx->iv; in gcm4106_aes_nx_encrypt() local 444 memcpy(iv, nonce, NX_GCM4106_NONCE_LEN); in gcm4106_aes_nx_encrypt() 445 memcpy(iv + NX_GCM4106_NONCE_LEN, req->iv, 8); in gcm4106_aes_nx_encrypt() 458 char *iv = rctx->iv; in gcm4106_aes_nx_decrypt() local [all …]
|
| /linux/lib/crypto/ |
| A D | aescfb.c | 43 int len, const u8 iv[AES_BLOCK_SIZE]) in aescfb_encrypt() 46 const u8 *v = iv; in aescfb_encrypt() 72 int len, const u8 iv[AES_BLOCK_SIZE]) in aescfb_decrypt() 76 aescfb_encrypt_block(ctx, ks[0], iv); in aescfb_decrypt() 113 u8 iv[AES_BLOCK_SIZE]; member 122 .iv = "\x00\x01\x02\x03\x04\x05\x06\x07" 146 .iv = "\x00\x01\x02\x03\x04\x05\x06\x07" 171 .iv = "\x00\x01\x02\x03\x04\x05\x06\x07" 194 .iv = "\x00\x01\x02\x03\x04\x05\x06\x07" 207 .iv = "\x00\x01\x02\x03\x04\x05\x06\x07" [all …]
|
| A D | chacha20poly1305.c | 39 u8 iv[CHACHA_IV_SIZE]; in xchacha_init() local 41 memset(iv, 0, 8); in xchacha_init() 53 memzero_explicit(iv, sizeof(iv)); in xchacha_init() 97 __le64 iv[2]; in chacha20poly1305_encrypt() local 101 iv[0] = 0; in chacha20poly1305_encrypt() 107 memzero_explicit(iv, sizeof(iv)); in chacha20poly1305_encrypt() 175 __le64 iv[2]; in chacha20poly1305_decrypt() local 180 iv[0] = 0; in chacha20poly1305_decrypt() 188 memzero_explicit(iv, sizeof(iv)); in chacha20poly1305_decrypt() 226 __le64 iv[2]; in chacha20poly1305_crypt_sg_inplace() member [all …]
|
| A D | aesgcm.c | 150 const u8 iv[GCM_AES_IV_SIZE], u8 *authtag) in aesgcm_encrypt() 154 memcpy(ctr, iv, GCM_AES_IV_SIZE); in aesgcm_encrypt() 186 memcpy(ctr, iv, GCM_AES_IV_SIZE); in aesgcm_decrypt() 563 u8 iv[GCM_AES_IV_SIZE]; member 585 .iv = "\xca\xfe\xba\xbe\xfa\xce\xdb\xad" 595 .iv = "\xca\xfe\xba\xbe\xfa\xce\xdb\xad" 620 .iv = "\xca\xfe\xba\xbe\xfa\xce\xdb\xad" 642 .iv = "\xca\xfe\xba\xbe\xfa\xce\xdb\xad" 654 .iv = "\xca\xfe\xba\xbe\xfa\xce\xdb\xad" 669 .iv = "\xca\xfe\xba\xbe\xfa\xce\xdb\xad" [all …]
|
| /linux/arch/x86/crypto/ |
| A D | sm4_aesni_avx_glue.c | 27 const u8 *src, u8 *iv); 29 const u8 *src, u8 *iv); 102 const u8 *iv = walk.iv; in sm4_cbc_encrypt() local 109 iv = dst; in sm4_cbc_encrypt() 114 if (iv != walk.iv) in sm4_cbc_encrypt() 115 memcpy(walk.iv, iv, SM4_BLOCK_SIZE); in sm4_cbc_encrypt() 150 u8 iv[SM4_BLOCK_SIZE]; in sm4_avx_cbc_decrypt() local 169 memcpy(walk.iv, iv, SM4_BLOCK_SIZE); in sm4_avx_cbc_decrypt() 220 walk.iv, SM4_BLOCK_SIZE); in sm4_avx_ctr_crypt() 221 crypto_inc(walk.iv, SM4_BLOCK_SIZE); in sm4_avx_ctr_crypt() [all …]
|
| /linux/block/ |
| A D | blk-integrity.c | 29 struct bio_vec iv, ivprv = { NULL }; in blk_rq_count_integrity_sg() local 35 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_count_integrity_sg() 38 if (!biovec_phys_mergeable(q, &ivprv, &iv)) in blk_rq_count_integrity_sg() 43 seg_size += iv.bv_len; in blk_rq_count_integrity_sg() 47 seg_size = iv.bv_len; in blk_rq_count_integrity_sg() 51 ivprv = iv; in blk_rq_count_integrity_sg() 69 struct bio_vec iv, ivprv = { NULL }; in blk_rq_map_integrity_sg() local 77 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_map_integrity_sg() 84 sg->length += iv.bv_len; in blk_rq_map_integrity_sg() 94 sg_set_page(sg, iv.bv_page, iv.bv_len, iv.bv_offset); in blk_rq_map_integrity_sg() [all …]
|
| /linux/include/crypto/ |
| A D | chacha.h | 65 void chacha_init_arch(u32 *state, const u32 *key, const u8 *iv); 66 static inline void chacha_init_generic(u32 *state, const u32 *key, const u8 *iv) in chacha_init_generic() argument 77 state[12] = get_unaligned_le32(iv + 0); in chacha_init_generic() 78 state[13] = get_unaligned_le32(iv + 4); in chacha_init_generic() 79 state[14] = get_unaligned_le32(iv + 8); in chacha_init_generic() 80 state[15] = get_unaligned_le32(iv + 12); in chacha_init_generic() 83 static inline void chacha_init(u32 *state, const u32 *key, const u8 *iv) in chacha_init() argument 86 chacha_init_arch(state, key, iv); in chacha_init() 88 chacha_init_generic(state, key, iv); in chacha_init()
|
| /linux/arch/arm64/crypto/ |
| A D | sm4-neon-glue.c | 24 u8 *iv, unsigned int nblocks); 26 u8 *iv, unsigned int nblocks); 91 const u8 *iv = walk.iv; in sm4_cbc_encrypt() local 96 crypto_xor_cpy(dst, src, iv, SM4_BLOCK_SIZE); in sm4_cbc_encrypt() 98 iv = dst; in sm4_cbc_encrypt() 103 if (iv != walk.iv) in sm4_cbc_encrypt() 104 memcpy(walk.iv, iv, SM4_BLOCK_SIZE); in sm4_cbc_encrypt() 132 walk.iv, nblocks); in sm4_cbc_decrypt() 163 walk.iv, nblocks); in sm4_ctr_crypt() 176 sm4_crypt_block(ctx->rkey_enc, keystream, walk.iv); in sm4_ctr_crypt() [all …]
|
| A D | ghash-ce-glue.c | 374 dg, iv, ctx->aes_key.key_enc, nrounds, in gcm_encrypt() 444 dg, iv, ctx->aes_key.key_enc, in gcm_decrypt() 466 u8 iv[AES_BLOCK_SIZE]; in gcm_aes_encrypt() local 468 memcpy(iv, req->iv, GCM_AES_IV_SIZE); in gcm_aes_encrypt() 469 return gcm_encrypt(req, iv, req->assoclen); in gcm_aes_encrypt() 474 u8 iv[AES_BLOCK_SIZE]; in gcm_aes_decrypt() local 476 memcpy(iv, req->iv, GCM_AES_IV_SIZE); in gcm_aes_decrypt() 504 u8 iv[AES_BLOCK_SIZE]; in rfc4106_encrypt() local 507 memcpy(iv + RFC4106_NONCE_SIZE, req->iv, GCM_RFC4106_IV_SIZE); in rfc4106_encrypt() 517 u8 iv[AES_BLOCK_SIZE]; in rfc4106_decrypt() local [all …]
|
| /linux/drivers/crypto/ccp/ |
| A D | ccp-crypto-aes.c | 33 memcpy(req->iv, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_complete() 83 if (!req->iv) in ccp_aes_crypt() 86 memcpy(rctx->iv, req->iv, AES_BLOCK_SIZE); in ccp_aes_crypt() 101 rctx->cmd.u.aes.iv = iv_sg; in ccp_aes_crypt() 139 req->iv = rctx->rfc3686_info; in ccp_aes_rfc3686_complete() 163 u8 *iv; in ccp_aes_rfc3686_crypt() local 166 iv = rctx->rfc3686_iv; in ccp_aes_rfc3686_crypt() 169 iv += CTR_RFC3686_NONCE_SIZE; in ccp_aes_rfc3686_crypt() 170 memcpy(iv, req->iv, CTR_RFC3686_IV_SIZE); in ccp_aes_rfc3686_crypt() 172 iv += CTR_RFC3686_IV_SIZE; in ccp_aes_rfc3686_crypt() [all …]
|
| /linux/net/mac80211/ |
| A D | wep.c | 39 if ((iv & 0xff00) == 0xff00) { in ieee80211_wep_weak_iv() 40 u8 B = (iv >> 16) & 0xff; in ieee80211_wep_weak_iv() 49 int keylen, int keyidx, u8 *iv) in ieee80211_wep_get_iv() argument 55 if (!iv) in ieee80211_wep_get_iv() 58 *iv++ = (local->wep_iv >> 16) & 0xff; in ieee80211_wep_get_iv() 59 *iv++ = (local->wep_iv >> 8) & 0xff; in ieee80211_wep_get_iv() 60 *iv++ = local->wep_iv & 0xff; in ieee80211_wep_get_iv() 61 *iv++ = keyidx << 6; in ieee80211_wep_get_iv() 136 u8 *iv; in ieee80211_wep_encrypt() local 144 if (!iv) in ieee80211_wep_encrypt() [all …]
|
| /linux/tools/testing/crypto/chacha20-s390/ |
| A D | test-cipher.c | 54 u8 iv[16], key[32]; in test_lib_chacha() local 58 memset(iv, 'I', sizeof(iv)); in test_lib_chacha() 65 16, 1, iv, 16, 1); in test_lib_chacha() 69 chacha_init_arch(chacha_state, (u32*)key, iv); in test_lib_chacha() 84 chacha_init_arch(chacha_state, (u32 *)key, iv); in test_lib_chacha() 131 u8 iv[16], key[32]; in test_skcipher() local 153 memset(iv, 'I', sizeof(iv)); in test_skcipher() 166 16, 1, iv, 16, 1); in test_skcipher() 176 data_size, iv); in test_skcipher() 195 memset(iv, 'I', sizeof(iv)); in test_skcipher() [all …]
|
| /linux/net/qrtr/ |
| A D | ns.c | 114 struct kvec iv; in service_announce_new() local 119 iv.iov_base = &pkt; in service_announce_new() 140 struct kvec iv; in service_announce_del() local 146 iv.iov_base = &pkt; in service_announce_del() 171 struct kvec iv; in lookup_notify() local 174 iv.iov_base = &pkt; in lookup_notify() 303 struct kvec iv; in say_hello() local 306 iv.iov_base = &pkt; in say_hello() 343 struct kvec iv; in ctrl_cmd_bye() local 396 struct kvec iv; in ctrl_cmd_del_client() local [all …]
|
| /linux/drivers/crypto/amcc/ |
| A D | crypto4xx_alg.c | 75 __le32 iv[AES_IV_SIZE]; in crypto4xx_crypt() local 81 crypto4xx_memcpy_to_le32(iv, req->iv, ivlen); in crypto4xx_crypt() 331 if (is_ccm && !(req->iv[0] == 1 || req->iv[0] == 3)) in crypto4xx_aead_need_fallback() 346 req->iv); in crypto4xx_aead_fallback() 432 __le32 iv[16]; in crypto4xx_crypt_aes_ccm() local 446 if (req->iv[0] == 1) { in crypto4xx_crypt_aes_ccm() 452 crypto4xx_memcpy_to_le32(iv, req->iv, 16 - (req->iv[0] + 1)); in crypto4xx_crypt_aes_ccm() 455 len, iv, sizeof(iv), in crypto4xx_crypt_aes_ccm() 578 __le32 iv[4]; in crypto4xx_crypt_aes_gcm() local 587 crypto4xx_memcpy_to_le32(iv, req->iv, GCM_AES_IV_SIZE); in crypto4xx_crypt_aes_gcm() [all …]
|
| /linux/arch/powerpc/crypto/ |
| A D | aes-gcm-p10-glue.c | 37 void *rkey, u8 *iv, void *Xi); 39 void *rkey, u8 *iv, void *Xi); 50 u8 iv[16]; member 122 *((__be32 *)(iv+12)) = counter; in gcmp10_init() 129 aes_p10_encrypt(iv, gctx->ivtag, rdkey); in gcmp10_init() 132 *((__be32 *)(iv+12)) = counter; in gcmp10_init() 133 memcpy(gctx->iv, iv, 16); in gcmp10_init() 216 unsigned char *iv = PTR_ALIGN((void *)ivbuf, PPC_ALIGN); in p10_aes_gcm_crypt() local 225 memcpy(iv, req->iv, GCM_IV_SIZE); in p10_aes_gcm_crypt() 267 &ctx->enc_key, gctx->iv, hash->Htable); in p10_aes_gcm_crypt() [all …]
|
| /linux/include/uapi/linux/ |
| A D | tls.h | 127 unsigned char iv[TLS_CIPHER_AES_GCM_128_IV_SIZE]; member 135 unsigned char iv[TLS_CIPHER_AES_GCM_256_IV_SIZE]; member 143 unsigned char iv[TLS_CIPHER_AES_CCM_128_IV_SIZE]; member 151 unsigned char iv[TLS_CIPHER_CHACHA20_POLY1305_IV_SIZE]; member 159 unsigned char iv[TLS_CIPHER_SM4_GCM_IV_SIZE]; member 167 unsigned char iv[TLS_CIPHER_SM4_CCM_IV_SIZE]; member 175 unsigned char iv[TLS_CIPHER_ARIA_GCM_128_IV_SIZE]; member 183 unsigned char iv[TLS_CIPHER_ARIA_GCM_256_IV_SIZE]; member
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | linked_list_fail.c | 11 struct map_value *v, *v2, *iv, *iv2; \ 25 iv = bpf_map_lookup_elem(map, &(int){ 0 }); \ 26 if (!iv) \ 66 CHECK(inner_map, pop_front, &iv->head); 67 CHECK(inner_map, pop_back, &iv->head); 89 CHECK(inner_map, push_front, &iv->head, &f->node2); 90 CHECK(inner_map, push_back, &iv->head, &f->node2); 109 CHECK(kptr_inner_map, op, &f1->lock, &iv->head); \ 114 CHECK(global_inner_map, op, &glock, &iv->head); \ 119 CHECK(map_inner_map, op, &v->lock, &iv->head); \ [all …]
|