Lines Matching +full:8 +full:- +full:15

1 /* SPDX-License-Identifier: GPL-2.0-or-later */
5 # Copyright 2023- IBM Corp. All rights reserved
13 # do rounds, 8 quarter rounds
16 # 3. a += b; d ^= a; d <<<= 8;
21 # row1 = (row1 + row2), row4 = row1 xor row4, row4 rotate each word by 8
43 #include <asm/asm-offsets.h>
44 #include <asm/asm-compat.h>
81 stdu 1,-752(1)
84 SAVE_GPR 15, 120, 1
117 SAVE_VSX 15, 208, 9
152 RESTORE_VSX 15, 208, 9
171 RESTORE_GPR 15, 120, 1
210 vpermxor 15, 15, 3, 25
216 vadduwm 8, 8, 12
219 vadduwm 11, 11, 15
224 vxor 4, 4, 8
258 vpermxor 15, 15, 3, 25
264 vadduwm 8, 8, 12
267 vadduwm 11, 11, 15
274 vxor 4, 4, 8
304 vpermxor 15, 15, 0, 25
314 vadduwm 10, 10, 15
316 vadduwm 8, 8, 13
324 vxor 7, 7, 8
354 vpermxor 15, 15, 0, 25
364 vadduwm 10, 10, 15
366 vadduwm 8, 8, 13
377 vxor 7, 7, 8
403 vpermxor 15, 15, 3, 20
404 vadduwm 8, 8, 12
407 vadduwm 11, 11, 15
408 vxor 4, 4, 8
423 vpermxor 15, 15, 3, 22
424 vadduwm 8, 8, 12
427 vadduwm 11, 11, 15
428 vxor 4, 4, 8
442 vpermxor 15, 15, 0, 20
446 vadduwm 10, 10, 15
448 vadduwm 8, 8, 13
452 vxor 7, 7, 8
462 vpermxor 15, 15, 0, 22
466 vadduwm 10, 10, 15
468 vadduwm 8, 8, 13
472 vxor 7, 7, 8
494 vadduwm \S+0, \S+0, 16-\S
495 vadduwm \S+4, \S+4, 17-\S
496 vadduwm \S+8, \S+8, 18-\S
497 vadduwm \S+12, \S+12, 19-\S
499 vadduwm \S+1, \S+1, 16-\S
500 vadduwm \S+5, \S+5, 17-\S
501 vadduwm \S+9, \S+9, 18-\S
502 vadduwm \S+13, \S+13, 19-\S
504 vadduwm \S+2, \S+2, 16-\S
505 vadduwm \S+6, \S+6, 17-\S
506 vadduwm \S+10, \S+10, 18-\S
507 vadduwm \S+14, \S+14, 19-\S
509 vadduwm \S+3, \S+3, 16-\S
510 vadduwm \S+7, \S+7, 17-\S
511 vadduwm \S+11, \S+11, 18-\S
512 vadduwm \S+15, \S+15, 19-\S
529 lxvw4x 8, 24, 9
536 lxvw4x 15, 31, 9
546 xxlxor \S+34, \S+34, 8
553 xxlxor \S+47, \S+47, 15
587 # r17 - r31 mainly for Write_256 macro.
604 mr 15, 6 # len
619 vsldoi 30, 4, 5, 8 # vr30 counter, 4 (0, 1, 2, 3)
629 sradi 8, 7, 1
631 mtctr 8
640 vspltisw 26, 8
667 xxspltw 32+8, 18, 0
674 xxspltw 32+15, 19, 3
708 TP_4x 8, 9, 10, 11
709 TP_4x 12, 13, 14, 15
726 addi 15, 15, -256 # len -=256
734 TP_4x 16+8, 16+9, 16+10, 16+11
735 TP_4x 16+12, 16+13, 16+14, 16+15
744 addi 15, 15, -256 # len +=256
754 cmpdi 15, 0
757 cmpdi 15, 512
760 mtctr 8
776 sradi 8, 7, 1
777 mtctr 8
789 vspltw 8, 18, 0
797 vspltw 15, 19, 3
808 TP_4x 8, 9, 10, 11
809 TP_4x 12, 13, 14, 15
814 addi 15, 15, -256 # len += 256
820 cmpdi 15, 0
822 cmpdi 15, 256
825 mtctr 8