Lines Matching refs:iv_size

183 	unsigned int iv_size;  member
364 memset(iv, 0, cc->iv_size); in crypt_iv_plain_gen()
373 memset(iv, 0, cc->iv_size); in crypt_iv_plain64_gen()
382 memset(iv, 0, cc->iv_size); in crypt_iv_plain64be_gen()
384 *(__be64 *)&iv[cc->iv_size - sizeof(u64)] = cpu_to_be64(dmreq->iv_sector); in crypt_iv_plain64be_gen()
396 memset(iv, 0, cc->iv_size); in crypt_iv_essiv_gen()
442 memset(iv, 0, cc->iv_size - sizeof(u64)); /* rest is cleared below */ in crypt_iv_benbi_gen()
445 put_unaligned(val, (__be64 *)(iv + cc->iv_size - sizeof(u64))); in crypt_iv_benbi_gen()
453 memset(iv, 0, cc->iv_size); in crypt_iv_null_gen()
544 memcpy(iv, ctx.state.h, cc->iv_size); in crypt_iv_lmk_one()
559 memset(iv, 0, cc->iv_size); in crypt_iv_lmk_gen()
577 crypto_xor(dst + sg->offset, iv, cc->iv_size); in crypt_iv_lmk_post()
603 if (cc->key_size <= (cc->iv_size + TCW_WHITENING_SIZE)) { in crypt_iv_tcw_ctr()
608 tcw->iv_seed = kzalloc(cc->iv_size, GFP_KERNEL); in crypt_iv_tcw_ctr()
622 int key_offset = cc->key_size - cc->iv_size - TCW_WHITENING_SIZE; in crypt_iv_tcw_init()
624 memcpy(tcw->iv_seed, &cc->key[key_offset], cc->iv_size); in crypt_iv_tcw_init()
625 memcpy(tcw->whitening, &cc->key[key_offset + cc->iv_size], in crypt_iv_tcw_init()
635 memset(tcw->iv_seed, 0, cc->iv_size); in crypt_iv_tcw_wipe()
683 if (cc->iv_size > 8) in crypt_iv_tcw_gen()
685 cc->iv_size - 8); in crypt_iv_tcw_gen()
712 get_random_bytes(iv, cc->iv_size); in crypt_iv_random_gen()
724 if (crypto_skcipher_blocksize(any_tfm(cc)) != cc->iv_size) { in crypt_iv_eboiv_ctr()
746 req = kmalloc(reqsize + cc->iv_size, GFP_NOIO); in crypt_iv_eboiv_gen()
753 memset(buf, 0, cc->iv_size); in crypt_iv_eboiv_gen()
756 sg_init_one(&src, page_address(ZERO_PAGE(0)), cc->iv_size); in crypt_iv_eboiv_gen()
757 sg_init_one(&dst, iv, cc->iv_size); in crypt_iv_eboiv_gen()
758 skcipher_request_set_crypt(req, &src, &dst, cc->iv_size, buf); in crypt_iv_eboiv_gen()
1249 return iv_of_dmreq(cc, dmreq) + cc->iv_size; in org_iv_of_dmreq()
1255 u8 *ptr = iv_of_dmreq(cc, dmreq) + cc->iv_size + cc->iv_size; in org_sector_of_dmreq()
1263 u8 *ptr = iv_of_dmreq(cc, dmreq) + cc->iv_size + in org_tag_of_dmreq()
1264 cc->iv_size + sizeof(uint64_t); in org_tag_of_dmreq()
1297 BUG_ON(cc->integrity_iv_size && cc->integrity_iv_size != cc->iv_size); in crypt_convert_block_aead()
1326 sg_set_buf(&dmreq->sg_in[1], org_iv, cc->iv_size); in crypt_convert_block_aead()
1332 sg_set_buf(&dmreq->sg_out[1], org_iv, cc->iv_size); in crypt_convert_block_aead()
1339 memcpy(org_iv, tag_iv, cc->iv_size); in crypt_convert_block_aead()
1346 memcpy(tag_iv, org_iv, cc->iv_size); in crypt_convert_block_aead()
1349 memcpy(iv, org_iv, cc->iv_size); in crypt_convert_block_aead()
1352 aead_request_set_ad(req, sizeof(uint64_t) + cc->iv_size); in crypt_convert_block_aead()
1445 memcpy(iv, org_iv, cc->iv_size); in crypt_convert_block_skcipher()
2793 cc->iv_size = crypto_aead_ivsize(any_tfm_aead(cc)); in crypt_ctr_ivmode()
2795 cc->iv_size = crypto_skcipher_ivsize(any_tfm(cc)); in crypt_ctr_ivmode()
2797 if (cc->iv_size) in crypt_ctr_ivmode()
2799 cc->iv_size = max(cc->iv_size, in crypt_ctr_ivmode()
2845 cc->key_extra_size = cc->iv_size + TCW_WHITENING_SIZE; in crypt_ctr_ivmode()
2849 cc->integrity_iv_size = cc->iv_size; in crypt_ctr_ivmode()
2963 cc->iv_size = crypto_aead_ivsize(any_tfm_aead(cc)); in crypt_ctr_cipher_new()
2965 cc->iv_size = crypto_skcipher_ivsize(any_tfm(cc)); in crypt_ctr_cipher_new()
3297 iv_size_padding + cc->iv_size + in crypt_ctr()
3298 cc->iv_size + in crypt_ctr()