Searched refs:inc_le128 (Results 1 – 7 of 7) sorted by relevance
| /linux/arch/arm64/crypto/ |
| H A D | sm4-ce-ccm-core.S | 31 #define inc_le128(vctr) \ macro 132 inc_le128(v8) /* +0 */ 133 inc_le128(v9) /* +1 */ 134 inc_le128(v10) /* +2 */ 135 inc_le128(v11) /* +3 */ 164 inc_le128(v8) 179 inc_le128(v8) 244 inc_le128(v8) /* +0 */ 245 inc_le128(v9) /* +1 */ 246 inc_le128(v10) /* +2 */ [all …]
|
| H A D | sm4-neon-core.S | 459 #define inc_le128(vctr) \ macro 467 inc_le128(v0) /* +0 */ 468 inc_le128(v1) /* +1 */ 469 inc_le128(v2) /* +2 */ 470 inc_le128(v3) /* +3 */ 471 inc_le128(v4) /* +4 */ 472 inc_le128(v5) /* +5 */ 473 inc_le128(v6) /* +6 */ 474 inc_le128(v7) /* +7 */ 506 inc_le128(v0) /* +0 */ [all …]
|
| H A D | sm4-ce-core.S | 424 #define inc_le128(vctr) \ macro 432 inc_le128(v0) /* +0 */ 433 inc_le128(v1) /* +1 */ 434 inc_le128(v2) /* +2 */ 435 inc_le128(v3) /* +3 */ 436 inc_le128(v4) /* +4 */ 437 inc_le128(v5) /* +5 */ 438 inc_le128(v6) /* +6 */ 439 inc_le128(v7) /* +7 */ 469 inc_le128(v0) /* +0 */ [all …]
|
| /linux/arch/x86/crypto/ |
| H A D | sm4-aesni-avx2-asm_64.S | 272 #define inc_le128(x, minus_one, tmp) \ macro 305 inc_le128(RTMP4x, RNOTx, RTMP1x); 335 inc_le128(RTMP0, RNOT, RTMP1); 336 inc_le128(RTMP0, RNOT, RTMP1); 338 inc_le128(RTMP0, RNOT, RTMP1); 339 inc_le128(RTMP0, RNOT, RTMP1); 341 inc_le128(RTMP0, RNOT, RTMP1); 342 inc_le128(RTMP0, RNOT, RTMP1); 344 inc_le128(RTMP0, RNOT, RTMP1); 345 inc_le128(RTMP0, RNOT, RTMP1); [all …]
|
| H A D | aria-aesni-avx2-asm_64.S | 54 #define inc_le128(x, minus_one, tmp) \ macro 1036 inc_le128(%xmm7, %xmm0, %xmm4); 1099 inc_le128(%ymm3, %ymm0, %ymm4); 1100 inc_le128(%ymm3, %ymm0, %ymm4); 1102 inc_le128(%ymm3, %ymm0, %ymm4); 1103 inc_le128(%ymm3, %ymm0, %ymm4); 1105 inc_le128(%ymm3, %ymm0, %ymm4); 1106 inc_le128(%ymm3, %ymm0, %ymm4); 1108 inc_le128(%ymm3, %ymm0, %ymm4); 1109 inc_le128(%ymm3, %ymm0, %ymm4); [all …]
|
| H A D | sm4-aesni-avx-asm_64.S | 437 #define inc_le128(x, minus_one, tmp) \ macro 444 inc_le128(RTMP0, RNOT, RTMP2); /* +1 */ 446 inc_le128(RTMP0, RNOT, RTMP2); /* +2 */ 448 inc_le128(RTMP0, RNOT, RTMP2); /* +3 */ 450 inc_le128(RTMP0, RNOT, RTMP2); /* +4 */ 452 inc_le128(RTMP0, RNOT, RTMP2); /* +5 */ 454 inc_le128(RTMP0, RNOT, RTMP2); /* +6 */ 456 inc_le128(RTMP0, RNOT, RTMP2); /* +7 */ 458 inc_le128(RTMP0, RNOT, RTMP2); /* +8 */
|
| H A D | aria-aesni-avx-asm_64.S | 38 #define inc_le128(x, minus_one, tmp) \ macro 1035 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1037 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1039 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1041 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1043 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1045 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1047 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1058 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1060 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ [all …]
|