Lines Matching full:iv
75 __le32 iv[AES_IV_SIZE]; in crypto4xx_crypt() local
81 crypto4xx_memcpy_to_le32(iv, req->iv, ivlen); in crypto4xx_crypt()
84 req->cryptlen, iv, ivlen, decrypt ? ctx->sa_in : ctx->sa_out, in crypto4xx_crypt()
212 __le32 iv[AES_IV_SIZE / 4] = { in crypto4xx_rfc3686_encrypt() local
214 cpu_to_le32p((u32 *) req->iv), in crypto4xx_rfc3686_encrypt()
215 cpu_to_le32p((u32 *) (req->iv + 4)), in crypto4xx_rfc3686_encrypt()
219 req->cryptlen, iv, AES_IV_SIZE, in crypto4xx_rfc3686_encrypt()
227 __le32 iv[AES_IV_SIZE / 4] = { in crypto4xx_rfc3686_decrypt() local
229 cpu_to_le32p((u32 *) req->iv), in crypto4xx_rfc3686_decrypt()
230 cpu_to_le32p((u32 *) (req->iv + 4)), in crypto4xx_rfc3686_decrypt()
234 req->cryptlen, iv, AES_IV_SIZE, in crypto4xx_rfc3686_decrypt()
244 unsigned int counter = be32_to_cpup((__be32 *)(req->iv + iv_len - 4)); in crypto4xx_ctr_crypt()
251 * the whole IV is a counter. So fallback if the counter is going to in crypto4xx_ctr_crypt()
262 req->cryptlen, req->iv); in crypto4xx_ctr_crypt()
331 if (is_ccm && !(req->iv[0] == 1 || req->iv[0] == 3)) in crypto4xx_aead_need_fallback()
346 req->iv); in crypto4xx_aead_fallback()
432 __le32 iv[16]; in crypto4xx_crypt_aes_ccm() local
446 if (req->iv[0] == 1) { in crypto4xx_crypt_aes_ccm()
451 iv[3] = cpu_to_le32(0); in crypto4xx_crypt_aes_ccm()
452 crypto4xx_memcpy_to_le32(iv, req->iv, 16 - (req->iv[0] + 1)); in crypto4xx_crypt_aes_ccm()
455 len, iv, sizeof(iv), in crypto4xx_crypt_aes_ccm()
578 __le32 iv[4]; in crypto4xx_crypt_aes_gcm() local
587 crypto4xx_memcpy_to_le32(iv, req->iv, GCM_AES_IV_SIZE); in crypto4xx_crypt_aes_gcm()
588 iv[3] = cpu_to_le32(1); in crypto4xx_crypt_aes_gcm()
591 len, iv, sizeof(iv), in crypto4xx_crypt_aes_gcm()