Lines Matching refs:chunksize

137 			    unsigned int chunksize, u32 stat_pad_len)  in spu_skcipher_rx_sg_create()  argument
161 rctx->dst_nents, chunksize); in spu_skcipher_rx_sg_create()
162 if (datalen < chunksize) { in spu_skcipher_rx_sg_create()
164 __func__, chunksize, datalen); in spu_skcipher_rx_sg_create()
199 u8 tx_frag_num, unsigned int chunksize, u32 pad_len) in spu_skcipher_tx_sg_create() argument
224 rctx->src_nents, chunksize); in spu_skcipher_tx_sg_create()
225 if (unlikely(datalen < chunksize)) { in spu_skcipher_tx_sg_create()
307 unsigned int chunksize; /* Num bytes of request to submit */ in handle_skcipher_req() local
338 chunksize = ctx->max_payload; in handle_skcipher_req()
340 chunksize = remaining; in handle_skcipher_req()
342 rctx->src_sent += chunksize; in handle_skcipher_req()
346 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, chunksize); in handle_skcipher_req()
347 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_skcipher_req()
384 add_to_ctr(rctx->msg_buf.iv_ctr, chunksize >> 4); in handle_skcipher_req()
394 rctx->src_sent, chunk_start, remaining, chunksize); in handle_skcipher_req()
402 &cipher_parms, chunksize); in handle_skcipher_req()
404 atomic64_add(chunksize, &iproc_priv.bytes_out); in handle_skcipher_req()
406 stat_pad_len = spu->spu_wordalign_padlen(chunksize); in handle_skcipher_req()
420 dump_sg(rctx->src_sg, rctx->src_skip, chunksize); in handle_skcipher_req()
438 err = spu_skcipher_rx_sg_create(mssg, rctx, rx_frag_num, chunksize, in handle_skcipher_req()
452 err = spu_skcipher_tx_sg_create(mssg, rctx, tx_frag_num, chunksize, in handle_skcipher_req()
662 unsigned int chunksize = 0; /* length of hash carry + new data */ in handle_ahash_req() local
723 chunksize = nbytes_to_hash; in handle_ahash_req()
725 (chunksize > ctx->max_payload)) in handle_ahash_req()
726 chunksize = ctx->max_payload; in handle_ahash_req()
737 rem = chunksize % blocksize; /* remainder */ in handle_ahash_req()
740 chunksize -= rem; in handle_ahash_req()
741 if (chunksize == 0) { in handle_ahash_req()
762 new_data_len = chunksize - local_nbuf; in handle_ahash_req()
779 rctx->total_sent += chunksize; in handle_ahash_req()
786 chunksize, in handle_ahash_req()
799 atomic64_add(chunksize, &iproc_priv.bytes_out); in handle_ahash_req()
809 flow_log("chunk_start: %u chunk_size: %u\n", chunk_start, chunksize); in handle_ahash_req()
830 data_pad_len = spu->spu_gcm_ccm_pad_len(ctx->cipher.mode, chunksize); in handle_ahash_req()
1165 unsigned int chunksize, in spu_aead_tx_sg_create() argument
1206 datalen = chunksize; in spu_aead_tx_sg_create()
1207 if ((chunksize > ctx->digestsize) && incl_icv) in spu_aead_tx_sg_create()
1261 unsigned int chunksize; in handle_aead_req() local
1282 chunksize = rctx->total_todo; in handle_aead_req()
1284 flow_log("%s: chunksize %u\n", __func__, chunksize); in handle_aead_req()
1338 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, chunksize); in handle_aead_req()
1339 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_aead_req()
1346 rctx->total_sent = chunksize; in handle_aead_req()
1347 rctx->src_sent = chunksize; in handle_aead_req()
1366 chunksize); in handle_aead_req()
1384 chunksize - digestsize); in handle_aead_req()
1388 chunksize, rctx->is_encrypt, in handle_aead_req()
1401 aead_parms.assoc_size + chunksize - in handle_aead_req()
1406 aead_parms.assoc_size + chunksize); in handle_aead_req()
1420 atomic64_add(chunksize, &iproc_priv.bytes_out); in handle_aead_req()
1422 flow_log("%s()-sent chunksize:%u\n", __func__, chunksize); in handle_aead_req()
1430 &aead_parms, chunksize); in handle_aead_req()
1434 chunksize, aead_parms.aad_pad_len, in handle_aead_req()
1455 dump_sg(rctx->src_sg, rctx->src_skip, chunksize); in handle_aead_req()
1468 resp_len = chunksize; in handle_aead_req()
1507 assoc_nents, aead_parms.iv_len, chunksize, in handle_aead_req()