Lines Matching refs:AES_BLOCK_SIZE
60 u8 rk[13 * (8 * AES_BLOCK_SIZE) + 32];
62 } __aligned(AES_BLOCK_SIZE);
106 while (walk.nbytes >= AES_BLOCK_SIZE) {
107 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE;
111 walk.stride / AES_BLOCK_SIZE);
118 walk.nbytes - blocks * AES_BLOCK_SIZE);
166 while (walk.nbytes >= AES_BLOCK_SIZE) {
167 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE;
175 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE);
189 while (walk.nbytes >= AES_BLOCK_SIZE) {
190 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE;
194 walk.stride / AES_BLOCK_SIZE);
202 walk.nbytes - blocks * AES_BLOCK_SIZE);
218 int blocks = (walk.nbytes / AES_BLOCK_SIZE) & ~7;
219 int nbytes = walk.nbytes % (8 * AES_BLOCK_SIZE);
227 dst += blocks * AES_BLOCK_SIZE;
228 src += blocks * AES_BLOCK_SIZE;
231 u8 buf[AES_BLOCK_SIZE];
234 if (unlikely(nbytes < AES_BLOCK_SIZE))
241 if (unlikely(nbytes < AES_BLOCK_SIZE))
283 int tail = req->cryptlen % (8 * AES_BLOCK_SIZE);
293 if (req->cryptlen < AES_BLOCK_SIZE)
297 if (unlikely(tail > 0 && tail < AES_BLOCK_SIZE)) {
299 AES_BLOCK_SIZE) - 2;
306 xts_blocks * AES_BLOCK_SIZE,
317 while (walk.nbytes >= AES_BLOCK_SIZE) {
318 int blocks = (walk.nbytes / AES_BLOCK_SIZE) & ~7;
334 out += blocks * AES_BLOCK_SIZE;
335 in += blocks * AES_BLOCK_SIZE;
336 nbytes -= blocks * AES_BLOCK_SIZE;
361 skcipher_request_set_crypt(req, src, dst, AES_BLOCK_SIZE + tail,
398 .base.cra_blocksize = AES_BLOCK_SIZE,
404 .walksize = 8 * AES_BLOCK_SIZE,
412 .base.cra_blocksize = AES_BLOCK_SIZE,
418 .walksize = 8 * AES_BLOCK_SIZE,
419 .ivsize = AES_BLOCK_SIZE,
433 .chunksize = AES_BLOCK_SIZE,
434 .walksize = 8 * AES_BLOCK_SIZE,
435 .ivsize = AES_BLOCK_SIZE,
443 .base.cra_blocksize = AES_BLOCK_SIZE,
449 .walksize = 8 * AES_BLOCK_SIZE,
450 .ivsize = AES_BLOCK_SIZE,