| /arch/x86/crypto/ |
| A D | ecb_cbc_helpers.h | 32 #define ECB_WALK_ADVANCE(blocks) do { \ argument 33 dst += (blocks) * __bsize; \ 34 src += (blocks) * __bsize; \ 35 nbytes -= (blocks) * __bsize; \ 38 #define ECB_BLOCK(blocks, func) do { \ argument 39 const int __blocks = (blocks); \ 46 ECB_WALK_ADVANCE(blocks); \ 61 #define CBC_DEC_BLOCK(blocks, func) do { \ argument 62 const int __blocks = (blocks); \ 68 const u8 *__iv = src + ((blocks) - 1) * __bsize; \ [all …]
|
| A D | Kconfig | 98 Processes 16 blocks in parallel. 115 Processes eight blocks in parallel. 129 Processes one or three blocks in parallel. 144 Processes eight blocks in parallel. 159 Processes four blocks in parallel. 175 Processes eight blocks in parallel. 188 Processes 16 blocks in parallel. 283 Processes eight blocks in parallel. 300 Processes 16 blocks in parallel. 318 Processes 32 blocks in parallel. [all …]
|
| /arch/arm64/crypto/ |
| A D | aes-neonbs-glue.c | 30 int rounds, int blocks); 32 int rounds, int blocks); 47 int rounds, int blocks); 97 int rounds, int blocks)) in __ecb_crypt() argument 110 blocks = round_down(blocks, in __ecb_crypt() 115 ctx->rounds, blocks); in __ecb_crypt() 193 blocks = round_down(blocks, in cbc_decrypt() 224 if (blocks >= 8) { in ctr_encrypt() 226 blocks, walk.iv); in ctr_encrypt() 324 if (blocks >= 8) { in __xts_crypt() [all …]
|
| A D | sha3-ce-glue.c | 31 asmlinkage int sha3_ce_transform(u64 *st, const u8 *data, int blocks, 40 int blocks; in sha3_update() local 44 blocks = len / bs; in sha3_update() 45 len -= blocks * bs; in sha3_update() 50 rem = sha3_ce_transform(sctx->st, data, blocks, ds); in sha3_update() 52 data += (blocks - rem) * bs; in sha3_update() 53 blocks = rem; in sha3_update() 54 } while (blocks); in sha3_update()
|
| A D | aes-glue.c | 79 int rounds, int blocks); 185 unsigned int blocks; in ecb_encrypt() local 205 unsigned int blocks; in ecb_decrypt() local 225 unsigned int blocks; in cbc_encrypt_walk() local 254 unsigned int blocks; in cbc_decrypt_walk() local 413 unsigned int blocks; in essiv_cbc_encrypt() local 418 if (blocks) { in essiv_cbc_encrypt() 435 unsigned int blocks; in essiv_cbc_decrypt() local 440 if (blocks) { in essiv_cbc_decrypt() 887 blocks = rem; in mac_do_update() [all …]
|
| A D | ghash-ce-glue.c | 72 void (*simd_update)(int blocks, u64 dg[], in ghash_do_simd_update() argument 78 simd_update(blocks, dg, src, key->h, head); in ghash_do_simd_update() 90 int blocks; in ghash_update() local 92 blocks = len / GHASH_BLOCK_SIZE; in ghash_update() 93 len -= blocks * GHASH_BLOCK_SIZE; in ghash_update() 96 int chunk = min(blocks, MAX_BLOCKS); in ghash_update() 100 blocks -= chunk; in ghash_update() 102 } while (unlikely(blocks > 0)); in ghash_update() 252 int blocks = count / GHASH_BLOCK_SIZE; in gcm_update_mac() local 254 ghash_do_simd_update(blocks, dg, src, &ctx->ghash_key, in gcm_update_mac() [all …]
|
| A D | aes-ce-ccm-glue.c | 36 int blocks, u8 dg[], int enc_before, 106 u32 blocks = abytes / AES_BLOCK_SIZE; in ce_aes_ccm_auth_data() local 108 if (macp == AES_BLOCK_SIZE || (!macp && blocks > 0)) { in ce_aes_ccm_auth_data() 109 u32 rem = ce_aes_mac_update(in, rk, rounds, blocks, mac, in ce_aes_ccm_auth_data() 111 u32 adv = (blocks - rem) * AES_BLOCK_SIZE; in ce_aes_ccm_auth_data()
|
| A D | sm3-neon-glue.c | 18 int blocks);
|
| A D | sm3-ce-glue.c | 21 int blocks);
|
| /arch/m68k/emu/ |
| A D | nfblock.c | 40 static inline s32 nfhd_get_capacity(u32 major, u32 minor, u32 *blocks, in nfhd_get_capacity() argument 44 virt_to_phys(blocks), virt_to_phys(blocksize)); in nfhd_get_capacity() 55 u32 blocks, bsize; member 84 geo->cylinders = dev->blocks >> (6 - dev->bshift); in nfhd_getgeo() 97 static int __init nfhd_init_one(int id, u32 blocks, u32 bsize) in nfhd_init_one() argument 108 blocks, bsize); in nfhd_init_one() 120 dev->blocks = blocks; in nfhd_init_one() 136 set_capacity(dev->disk, (sector_t)blocks * (bsize / 512)); in nfhd_init_one() 155 u32 blocks, bsize; in nfhd_init() local 173 if (nfhd_get_capacity(i, 0, &blocks, &bsize)) in nfhd_init() [all …]
|
| /arch/arm/crypto/ |
| A D | aes-neonbs-glue.c | 29 int rounds, int blocks); 31 int rounds, int blocks); 34 int rounds, int blocks, u8 iv[]); 37 int rounds, int blocks, u8 ctr[]); 40 int rounds, int blocks, u8 iv[], int); 42 int rounds, int blocks, u8 iv[], int); 82 int rounds, int blocks)) in __ecb_crypt() argument 95 blocks = round_down(blocks, in __ecb_crypt() 100 ctx->rounds, blocks); in __ecb_crypt() 181 blocks = round_down(blocks, in cbc_decrypt() [all …]
|
| A D | aes-ce-glue.c | 28 int rounds, int blocks); 30 int rounds, int blocks); 33 int rounds, int blocks, u8 iv[]); 35 int rounds, int blocks, u8 iv[]); 170 unsigned int blocks; in ecb_encrypt() local 190 unsigned int blocks; in ecb_decrypt() local 210 unsigned int blocks; in cbc_encrypt_walk() local 240 unsigned int blocks; in cbc_decrypt_walk() local 386 int err, blocks; in ctr_encrypt() local 407 blocks = -1; in ctr_encrypt() [all …]
|
| A D | blake2b-neon-glue.c | 24 const size_t blocks = min_t(size_t, nblocks, in blake2b_compress_arch() local 28 blake2b_compress_neon(state, block, blocks, inc); in blake2b_compress_arch() 31 nblocks -= blocks; in blake2b_compress_arch() 32 block += blocks * BLAKE2B_BLOCK_SIZE; in blake2b_compress_arch()
|
| A D | ghash-ce-glue.c | 74 pmull_ghash_update_p64(blocks, dg, src, key->h, head); in ghash_do_update() 76 pmull_ghash_update_p8(blocks, dg, src, key->h, head); in ghash_do_update() 85 int blocks; in ghash_update() local 87 blocks = len / GHASH_BLOCK_SIZE; in ghash_update() 88 ghash_do_update(blocks, ctx->digest, src, key, NULL); in ghash_update() 89 return len - blocks * GHASH_BLOCK_SIZE; in ghash_update() 186 void pmull_gcm_encrypt(int blocks, u64 dg[], const char *src, 190 void pmull_gcm_enc_final(int blocks, u64 dg[], char *tag, 256 int blocks = count / GHASH_BLOCK_SIZE; in gcm_update_mac() local 258 pmull_ghash_update_p64(blocks, dg, src, ctx->h, in gcm_update_mac() [all …]
|
| A D | Kconfig | 75 fastest AES implementation for single blocks. For multiple 76 blocks, the NEON bit-sliced implementation is usually faster. 104 sliced AES code cannot process single blocks efficiently, in certain
|
| /arch/powerpc/kernel/ |
| A D | rtas_flash.c | 110 struct flash_block blocks[FLASH_BLOCKS_PER_NODE]; member 173 if (f->blocks[i].data == NULL) { in flash_list_valid() 176 block_size = f->blocks[i].length; in flash_list_valid() 201 kmem_cache_free(flash_block_cache, f->blocks[i].data); in free_flash_list() 356 fl->blocks[next_free].data = p; in rtas_flash_write() 357 fl->blocks[next_free].length = count; in rtas_flash_write() 586 f->blocks[i].data = (char *)cpu_to_be64(__pa(f->blocks[i].data)); in rtas_flash_firmware() 587 image_size += f->blocks[i].length; in rtas_flash_firmware() 588 f->blocks[i].length = cpu_to_be64(f->blocks[i].length); in rtas_flash_firmware()
|
| /arch/powerpc/crypto/ |
| A D | aes-gcm-p10.S | 36 # AES is implemented with 8x blocks and GHASH is using 2 4x blocks. 691 # Process different blocks 700 # Process 8x AES/GCM blocks 705 divdu 12, 5, 10 # n 128 bytes-blocks 759 # Loop 8x blocks and compute ghash 899 # Done 8x blocks 969 # Process different blocks 978 # Process 8x AES/GCM blocks 983 divdu 12, 5, 10 # n 128 bytes-blocks 1037 # Loop 8x blocks and compute ghash [all …]
|
| A D | md5-glue.c | 16 extern void ppc_md5_transform(u32 *state, const u8 *src, u32 blocks);
|
| /arch/arm/boot/dts/microchip/ |
| A D | sama5d3_tcb1.dtsi | 4 * 2 TC blocks.
|
| /arch/powerpc/platforms/pseries/ |
| A D | papr_scm.c | 56 uint64_t blocks; member 143 p->blocks, BIND_ANY_ADDR, token); in drc_pmem_bind() 210 p->drc_index, p->blocks - 1); in drc_pmem_query_n_bind() 215 if ((end_addr - start_addr) != ((p->blocks - 1) * p->block_size)) in drc_pmem_query_n_bind() 1253 mapping.size = p->blocks * p->block_size; // XXX: potential overflow? in papr_scm_nvdimm_init() 1364 u64 blocks, block_size; in papr_scm_probe() local 1383 if (of_property_read_u64(dn, "ibm,number-of-blocks", &blocks)) { in papr_scm_probe() 1414 p->blocks = blocks; in papr_scm_probe() 1460 p->res.end = p->bound_addr + p->blocks * p->block_size - 1; in papr_scm_probe()
|
| /arch/x86/kernel/cpu/mce/ |
| A D | amd.c | 244 struct threshold_block *blocks; member 949 first_block = bp[bank]->blocks; in amd_threshold_interrupt() 1187 if (tb->blocks) in allocate_threshold_blocks() 1188 list_add(&b->miscj, &tb->blocks->miscj); in allocate_threshold_blocks() 1190 tb->blocks = b; in allocate_threshold_blocks() 1265 list_for_each_entry_safe(pos, tmp, &bank->blocks->miscj, miscj) { in deallocate_threshold_blocks() 1270 kobject_put(&bank->blocks->kobj); in deallocate_threshold_blocks() 1275 if (!bank->blocks) in threshold_remove_bank()
|
| /arch/mips/cavium-octeon/ |
| A D | Kconfig | 22 This selects the size of CVMSEG LM, which is in cache blocks. The 23 legally range is from zero to 54 cache blocks (i.e. CVMSEG LM is
|
| /arch/arm/boot/dts/intel/socfpga/ |
| A D | socfpga_cyclone5_mcvevk.dts | 55 blocks = <0x5>;
|
| /arch/arm64/boot/dts/freescale/ |
| A D | fsl-ls1043-post.dtsi | 10 /* include used FMan blocks */
|
| /arch/arm/boot/dts/broadcom/ |
| A D | bcm47094-dlink-dir-890l.dts | 109 * 64KB blocks; total size 2MB, same that can be 137 /* Spansion S34ML01G2, 128MB with 128KB erase blocks */
|