Home
last modified time | relevance | path

Searched refs:key_enc (Results 1 – 19 of 19) sorted by relevance

/linux/lib/crypto/
H A Daes.c199 ctx->key_enc[i] = get_unaligned_le32(in_key + i * sizeof(u32)); in aes_expandkey()
202 u32 *rki = ctx->key_enc + (i * kwords); in aes_expandkey()
231 ctx->key_dec[0] = ctx->key_enc[key_len + 24]; in aes_expandkey()
232 ctx->key_dec[1] = ctx->key_enc[key_len + 25]; in aes_expandkey()
233 ctx->key_dec[2] = ctx->key_enc[key_len + 26]; in aes_expandkey()
234 ctx->key_dec[3] = ctx->key_enc[key_len + 27]; in aes_expandkey()
237 ctx->key_dec[i] = inv_mix_columns(ctx->key_enc[j]); in aes_expandkey()
238 ctx->key_dec[i + 1] = inv_mix_columns(ctx->key_enc[j + 1]); in aes_expandkey()
239 ctx->key_dec[i + 2] = inv_mix_columns(ctx->key_enc[j + 2]); in aes_expandkey()
240 ctx->key_dec[i + 3] = inv_mix_columns(ctx->key_enc[j + 3]); in aes_expandkey()
[all …]
/linux/arch/powerpc/crypto/
H A Daes-spe-glue.c42 u32 key_enc[AES_MAX_KEYLENGTH_U32]; member
48 u32 key_enc[AES_MAX_KEYLENGTH_U32]; member
54 extern void ppc_encrypt_aes(u8 *out, const u8 *in, u32 *key_enc, u32 rounds);
56 extern void ppc_encrypt_ecb(u8 *out, const u8 *in, u32 *key_enc, u32 rounds,
60 extern void ppc_encrypt_cbc(u8 *out, const u8 *in, u32 *key_enc, u32 rounds,
64 extern void ppc_crypt_ctr (u8 *out, const u8 *in, u32 *key_enc, u32 rounds,
66 extern void ppc_encrypt_xts(u8 *out, const u8 *in, u32 *key_enc, u32 rounds,
71 extern void ppc_expand_key_128(u32 *key_enc, const u8 *key);
72 extern void ppc_expand_key_192(u32 *key_enc, const u8 *key);
73 extern void ppc_expand_key_256(u32 *key_enc, const u8 *key);
[all …]
/linux/arch/arm64/crypto/
H A Daes-ce-glue.c56 __aes_ce_encrypt(ctx->key_enc, dst, src, num_rounds(ctx)); in aes_cipher_encrypt()
85 struct aes_block *key_enc, *key_dec; in ce_aes_expandkey() local
95 ctx->key_enc[i] = get_unaligned_le32(in_key + i * sizeof(u32)); in ce_aes_expandkey()
99 u32 *rki = ctx->key_enc + (i * kwords); in ce_aes_expandkey()
128 key_enc = (struct aes_block *)ctx->key_enc; in ce_aes_expandkey()
132 key_dec[0] = key_enc[j]; in ce_aes_expandkey()
134 __aes_ce_invert(key_dec + i, key_enc + j); in ce_aes_expandkey()
135 key_dec[i] = key_enc[0]; in ce_aes_expandkey()
H A Daes-glue.c192 ctx->key_enc, rounds, blocks); in ecb_encrypt()
230 ctx->key_enc, rounds, blocks, walk->iv); in cbc_encrypt_walk()
328 ctx->key_enc, rounds, walk.nbytes, walk.iv); in cts_cbc_encrypt()
421 ctx->key1.key_enc, rounds, blocks, in essiv_cbc_encrypt()
422 req->iv, ctx->key2.key_enc); in essiv_cbc_encrypt()
444 req->iv, ctx->key2.key_enc); in essiv_cbc_decrypt()
482 aes_xctr_encrypt(dst, src, ctx->key_enc, rounds, nbytes, in xctr_encrypt()
527 aes_ctr_encrypt(dst, src, ctx->key_enc, rounds, nbytes, in ctr_encrypt()
584 ctx->key1.key_enc, rounds, nbytes, in xts_encrypt()
585 ctx->key2.key_enc, walk.iv, first); in xts_encrypt()
[all …]
H A Daes-neonbs-glue.c89 aesbs_convert_key(ctx->rk, rk.key_enc, ctx->rounds); in aesbs_setkey()
147 memcpy(ctx->enc, rk.key_enc, sizeof(ctx->enc)); in aesbs_cbc_ctr_setkey()
150 aesbs_convert_key(ctx->key.rk, rk.key_enc, ctx->key.rounds); in aesbs_cbc_ctr_setkey()
272 memcpy(ctx->twkey, rk.key_enc, sizeof(ctx->twkey)); in aesbs_xts_setkey()
339 neon_aes_xts_encrypt(out, in, ctx->cts.key_enc, in __xts_crypt()
373 neon_aes_xts_encrypt(out, in, ctx->cts.key_enc, ctx->key.rounds, in __xts_crypt()
H A Daes-ce-ccm-glue.c155 ctx->key_enc, num_rounds(ctx)); in ccm_calculate_auth_mac()
168 macp = ce_aes_ccm_auth_data(mac, p, n, macp, ctx->key_enc, in ccm_calculate_auth_mac()
222 ctx->key_enc, num_rounds(ctx), in ccm_encrypt()
289 ctx->key_enc, num_rounds(ctx), in ccm_decrypt()
H A Daes-cipher-glue.c20 __aes_arm64_encrypt(ctx->key_enc, out, in, rounds); in aes_arm64_encrypt()
H A Dghash-ce-glue.c374 dg, iv, ctx->aes_key.key_enc, nrounds, in gcm_encrypt()
444 dg, iv, ctx->aes_key.key_enc, in gcm_decrypt()
/linux/arch/arm/crypto/
H A Daes-ce-glue.c80 struct aes_block *key_enc, *key_dec; in ce_aes_expandkey() local
90 ctx->key_enc[i] = get_unaligned_le32(in_key + i * sizeof(u32)); in ce_aes_expandkey()
94 u32 *rki = ctx->key_enc + (i * kwords); in ce_aes_expandkey()
124 key_enc = (struct aes_block *)ctx->key_enc; in ce_aes_expandkey()
128 key_dec[0] = key_enc[j]; in ce_aes_expandkey()
130 ce_aes_invert(key_dec + i, key_enc + j); in ce_aes_expandkey()
131 key_dec[i] = key_enc[0]; in ce_aes_expandkey()
180 ctx->key_enc, num_rounds(ctx), blocks); in ecb_encrypt()
218 ctx->key_enc, num_rounds(ctx), blocks, in cbc_encrypt_walk()
318 ctx->key_enc, num_rounds(ctx), walk.nbytes, in cts_cbc_encrypt()
[all …]
H A Daes-neonbs-glue.c81 aesbs_convert_key(ctx->rk, rk.key_enc, ctx->rounds); in aesbs_setkey()
139 aesbs_convert_key(ctx->key.rk, ctx->fallback.key_enc, ctx->key.rounds); in aesbs_cbc_setkey()
162 __aes_arm_encrypt(ctx->fallback.key_enc, in cbc_encrypt()
216 aesbs_convert_key(ctx->key.rk, ctx->fallback.key_enc, ctx->key.rounds); in aesbs_ctr_setkey_sync()
261 __aes_arm_encrypt(ctx->fallback.key_enc, ctx->key.rounds, src, dst); in ctr_encrypt_one()
323 __aes_arm_encrypt(ctx->tweak_key.key_enc, rounds, walk.iv, walk.iv); in __xts_crypt()
355 __aes_arm_encrypt(ctx->fallback.key_enc, rounds, buf, buf); in __xts_crypt()
H A Daes-cipher-glue.c22 __aes_arm_encrypt(ctx->key_enc, rounds, in, out); in aes_arm_encrypt()
H A Dghash-ce-glue.c396 memcpy(ctx->rk, aes_ctx.key_enc, sizeof(ctx->rk)); in gcm_aes_setkey()
/linux/crypto/
H A Daes_generic.c1180 const u32 *kp = ctx->key_enc + 4; in crypto_aes_encrypt()
1183 b0[0] = ctx->key_enc[0] ^ get_unaligned_le32(in); in crypto_aes_encrypt()
1184 b0[1] = ctx->key_enc[1] ^ get_unaligned_le32(in + 4); in crypto_aes_encrypt()
1185 b0[2] = ctx->key_enc[2] ^ get_unaligned_le32(in + 8); in crypto_aes_encrypt()
1186 b0[3] = ctx->key_enc[3] ^ get_unaligned_le32(in + 12); in crypto_aes_encrypt()
/linux/include/crypto/
H A Daes.h26 u32 key_enc[AES_MAX_KEYLENGTH_U32]; member
/linux/drivers/crypto/inside-secure/
H A Dsafexcel_cipher.c386 if (le32_to_cpu(ctx->key[i]) != aes.key_enc[i]) { in safexcel_skcipher_aes_setkey()
394 ctx->key[i] = cpu_to_le32(aes.key_enc[i]); in safexcel_skcipher_aes_setkey()
1375 if (le32_to_cpu(ctx->key[i]) != aes.key_enc[i]) { in safexcel_skcipher_aesctr_setkey()
1383 ctx->key[i] = cpu_to_le32(aes.key_enc[i]); in safexcel_skcipher_aesctr_setkey()
2413 if (le32_to_cpu(ctx->key[i]) != aes.key_enc[i]) { in safexcel_skcipher_aesxts_setkey()
2421 ctx->key[i] = cpu_to_le32(aes.key_enc[i]); in safexcel_skcipher_aesxts_setkey()
2431 aes.key_enc[i]) { in safexcel_skcipher_aesxts_setkey()
2440 cpu_to_le32(aes.key_enc[i]); in safexcel_skcipher_aesxts_setkey()
2522 if (le32_to_cpu(ctx->key[i]) != aes.key_enc[i]) { in safexcel_aead_gcm_setkey()
2530 ctx->key[i] = cpu_to_le32(aes.key_enc[i]); in safexcel_aead_gcm_setkey()
[all …]
/linux/drivers/crypto/
H A Dpadlock-aes.c150 memcpy(ctx->E, gen_aes.key_enc, AES_MAX_KEYLENGTH); in aes_set_key()
/linux/drivers/crypto/intel/qat/qat_common/
H A Dqat_algs.c474 key = (u8 *)aes_expanded.key_enc + (AES_BLOCK_SIZE * nrounds); in qat_alg_xts_reverse_key()
479 key = (u8 *)aes_expanded.key_enc + (AES_BLOCK_SIZE * nrounds); in qat_alg_xts_reverse_key()
/linux/arch/x86/crypto/
H A Daesni-intel_glue.c1204 BUILD_BUG_ON(offsetof(struct aes_gcm_key_aesni, base.aes_key.key_enc) != 0); in gcm_setkey()
1209 BUILD_BUG_ON(offsetof(struct aes_gcm_key_avx10, base.aes_key.key_enc) != 0); in gcm_setkey()
/linux/drivers/crypto/aspeed/
H A Daspeed-hace-crypto.c589 memcpy(ctx->key, gen_aes_key.key_enc, AES_MAX_KEYLENGTH); in aspeed_aes_setkey()