Lines Matching +full:0 +full:- +full:19

35 	ctx->vtable = &br_aes_pwr8_ctr_vtable;  in br_aes_pwr8_ctr_init()
36 ctx->num_rounds = br_aes_pwr8_keysched(ctx->skey.skni, key, len); in br_aes_pwr8_ctr_init()
47 0x03020100, 0x07060504, 0x0B0A0908, 0x0F0E0D0C in ctr_128()
51 0, 0, 0, 4 in ctr_128()
54 cc0 = 0; in ctr_128()
84 li(%[cc0], 0) in ctr_128()
90 lxvw4x(47, 0, %[idx2be]) in ctr_128()
95 lxvw4x(60, 0, %[ctrinc]) in ctr_128()
108 vperm(19, 19, 19, 15) in ctr_128()
119 vadduwm(27, 19, 28) in ctr_128()
133 vxor(16, 16, 0) in ctr_128()
134 vxor(17, 17, 0) in ctr_128()
135 vxor(18, 18, 0) in ctr_128()
136 vxor(19, 19, 0) in ctr_128()
140 vcipher(19, 19, 1) in ctr_128()
144 vcipher(19, 19, 2) in ctr_128()
148 vcipher(19, 19, 3) in ctr_128()
152 vcipher(19, 19, 4) in ctr_128()
156 vcipher(19, 19, 5) in ctr_128()
160 vcipher(19, 19, 6) in ctr_128()
164 vcipher(19, 19, 7) in ctr_128()
168 vcipher(19, 19, 8) in ctr_128()
172 vcipher(19, 19, 9) in ctr_128()
176 vcipherlast(19, 19, 10) in ctr_128()
182 vperm(19, 19, 19, 15) in ctr_128()
191 vxor(19, 23, 19) in ctr_128()
205 vand(19, 27, 27) in ctr_128()
231 0x03020100, 0x07060504, 0x0B0A0908, 0x0F0E0D0C in ctr_192()
235 0, 0, 0, 4 in ctr_192()
238 cc0 = 0; in ctr_192()
272 li(%[cc0], 0) in ctr_192()
278 lxvw4x(47, 0, %[idx2be]) in ctr_192()
283 lxvw4x(60, 0, %[ctrinc]) in ctr_192()
296 vperm(19, 19, 19, 15) in ctr_192()
307 vadduwm(27, 19, 28) in ctr_192()
321 vxor(16, 16, 0) in ctr_192()
322 vxor(17, 17, 0) in ctr_192()
323 vxor(18, 18, 0) in ctr_192()
324 vxor(19, 19, 0) in ctr_192()
328 vcipher(19, 19, 1) in ctr_192()
332 vcipher(19, 19, 2) in ctr_192()
336 vcipher(19, 19, 3) in ctr_192()
340 vcipher(19, 19, 4) in ctr_192()
344 vcipher(19, 19, 5) in ctr_192()
348 vcipher(19, 19, 6) in ctr_192()
352 vcipher(19, 19, 7) in ctr_192()
356 vcipher(19, 19, 8) in ctr_192()
360 vcipher(19, 19, 9) in ctr_192()
364 vcipher(19, 19, 10) in ctr_192()
368 vcipher(19, 19, 11) in ctr_192()
372 vcipherlast(19, 19, 12) in ctr_192()
378 vperm(19, 19, 19, 15) in ctr_192()
387 vxor(19, 23, 19) in ctr_192()
401 vand(19, 27, 27) in ctr_192()
427 0x03020100, 0x07060504, 0x0B0A0908, 0x0F0E0D0C in ctr_256()
431 0, 0, 0, 4 in ctr_256()
434 cc0 = 0; in ctr_256()
472 li(%[cc0], 0) in ctr_256()
478 lxvw4x(47, 0, %[idx2be]) in ctr_256()
483 lxvw4x(60, 0, %[ctrinc]) in ctr_256()
496 vperm(19, 19, 19, 15) in ctr_256()
507 vadduwm(27, 19, 28) in ctr_256()
521 vxor(16, 16, 0) in ctr_256()
522 vxor(17, 17, 0) in ctr_256()
523 vxor(18, 18, 0) in ctr_256()
524 vxor(19, 19, 0) in ctr_256()
528 vcipher(19, 19, 1) in ctr_256()
532 vcipher(19, 19, 2) in ctr_256()
536 vcipher(19, 19, 3) in ctr_256()
540 vcipher(19, 19, 4) in ctr_256()
544 vcipher(19, 19, 5) in ctr_256()
548 vcipher(19, 19, 6) in ctr_256()
552 vcipher(19, 19, 7) in ctr_256()
556 vcipher(19, 19, 8) in ctr_256()
560 vcipher(19, 19, 9) in ctr_256()
564 vcipher(19, 19, 10) in ctr_256()
568 vcipher(19, 19, 11) in ctr_256()
572 vcipher(19, 19, 12) in ctr_256()
576 vcipher(19, 19, 13) in ctr_256()
580 vcipherlast(19, 19, 14) in ctr_256()
586 vperm(19, 19, 19, 15) in ctr_256()
595 vxor(19, 23, 19) in ctr_256()
609 vand(19, 27, 27) in ctr_256()
636 memcpy(ivbuf + 0, iv, 12); in br_aes_pwr8_ctr_run()
641 br_enc32be(ivbuf + 12, cc + 0); in br_aes_pwr8_ctr_run()
645 switch (ctx->num_rounds) { in br_aes_pwr8_ctr_run()
647 ctr_128(ctx->skey.skni, ivbuf, buf, in br_aes_pwr8_ctr_run()
651 ctr_192(ctx->skey.skni, ivbuf, buf, in br_aes_pwr8_ctr_run()
655 ctr_256(ctx->skey.skni, ivbuf, buf, in br_aes_pwr8_ctr_run()
663 if (len > 0) { in br_aes_pwr8_ctr_run()
667 memset(tmp + len, 0, (sizeof tmp) - len); in br_aes_pwr8_ctr_run()
668 br_enc32be(ivbuf + 12, cc + 0); in br_aes_pwr8_ctr_run()
672 switch (ctx->num_rounds) { in br_aes_pwr8_ctr_run()
674 ctr_128(ctx->skey.skni, ivbuf, tmp, 4); in br_aes_pwr8_ctr_run()
677 ctr_192(ctx->skey.skni, ivbuf, tmp, 4); in br_aes_pwr8_ctr_run()
680 ctr_256(ctx->skey.skni, ivbuf, tmp, 4); in br_aes_pwr8_ctr_run()