1/* $FreeBSD$ */ 2/* Do not modify. This file is auto-generated from aesni-sha256-x86_64.pl. */ 3.text 4 5 6.globl aesni_cbc_sha256_enc 7.type aesni_cbc_sha256_enc,@function 8.align 16 9aesni_cbc_sha256_enc: 10.cfi_startproc 11 leaq OPENSSL_ia32cap_P(%rip),%r11 12 movl $1,%eax 13 cmpq $0,%rdi 14 je .Lprobe 15 movl 0(%r11),%eax 16 movq 4(%r11),%r10 17 btq $61,%r10 18 jc aesni_cbc_sha256_enc_shaext 19 movq %r10,%r11 20 shrq $32,%r11 21 22 testl $2048,%r10d 23 jnz aesni_cbc_sha256_enc_xop 24 andl $296,%r11d 25 cmpl $296,%r11d 26 je aesni_cbc_sha256_enc_avx2 27 andl $268435456,%r10d 28 jnz aesni_cbc_sha256_enc_avx 29 ud2 30 xorl %eax,%eax 31 cmpq $0,%rdi 32 je .Lprobe 33 ud2 34.Lprobe: 35 .byte 0xf3,0xc3 36.cfi_endproc 37.size aesni_cbc_sha256_enc,.-aesni_cbc_sha256_enc 38 39.align 64 40.type K256,@object 41K256: 42.long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5 43.long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5 44.long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5 45.long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5 46.long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3 47.long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3 48.long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174 49.long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174 50.long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc 51.long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc 52.long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da 53.long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da 54.long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7 55.long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7 56.long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967 57.long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967 58.long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13 59.long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13 60.long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85 61.long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85 62.long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3 63.long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3 64.long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070 65.long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070 66.long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5 67.long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5 68.long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3 69.long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3 70.long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208 71.long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208 72.long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2 73.long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2 74 75.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f 76.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f 77.long 0,0,0,0, 0,0,0,0, -1,-1,-1,-1 78.long 0,0,0,0, 0,0,0,0 79.byte 65,69,83,78,73,45,67,66,67,43,83,72,65,50,53,54,32,115,116,105,116,99,104,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 80.align 64 81.type aesni_cbc_sha256_enc_xop,@function 82.align 64 83aesni_cbc_sha256_enc_xop: 84.cfi_startproc 85.Lxop_shortcut: 86 movq 8(%rsp),%r10 87 movq %rsp,%rax 88.cfi_def_cfa_register %rax 89 pushq %rbx 90.cfi_offset %rbx,-16 91 pushq %rbp 92.cfi_offset %rbp,-24 93 pushq %r12 94.cfi_offset %r12,-32 95 pushq %r13 96.cfi_offset %r13,-40 97 pushq %r14 98.cfi_offset %r14,-48 99 pushq %r15 100.cfi_offset %r15,-56 101 subq $128,%rsp 102 andq $-64,%rsp 103 104 shlq $6,%rdx 105 subq %rdi,%rsi 106 subq %rdi,%r10 107 addq %rdi,%rdx 108 109 110 movq %rsi,64+8(%rsp) 111 movq %rdx,64+16(%rsp) 112 113 movq %r8,64+32(%rsp) 114 movq %r9,64+40(%rsp) 115 movq %r10,64+48(%rsp) 116 movq %rax,120(%rsp) 117.cfi_escape 0x0f,0x06,0x77,0xf8,0x00,0x06,0x23,0x08 118.Lprologue_xop: 119 vzeroall 120 121 movq %rdi,%r12 122 leaq 128(%rcx),%rdi 123 leaq K256+544(%rip),%r13 124 movl 240-128(%rdi),%r14d 125 movq %r9,%r15 126 movq %r10,%rsi 127 vmovdqu (%r8),%xmm8 128 subq $9,%r14 129 130 movl 0(%r15),%eax 131 movl 4(%r15),%ebx 132 movl 8(%r15),%ecx 133 movl 12(%r15),%edx 134 movl 16(%r15),%r8d 135 movl 20(%r15),%r9d 136 movl 24(%r15),%r10d 137 movl 28(%r15),%r11d 138 139 vmovdqa 0(%r13,%r14,8),%xmm14 140 vmovdqa 16(%r13,%r14,8),%xmm13 141 vmovdqa 32(%r13,%r14,8),%xmm12 142 vmovdqu 0-128(%rdi),%xmm10 143 jmp .Lloop_xop 144.align 16 145.Lloop_xop: 146 vmovdqa K256+512(%rip),%xmm7 147 vmovdqu 0(%rsi,%r12,1),%xmm0 148 vmovdqu 16(%rsi,%r12,1),%xmm1 149 vmovdqu 32(%rsi,%r12,1),%xmm2 150 vmovdqu 48(%rsi,%r12,1),%xmm3 151 vpshufb %xmm7,%xmm0,%xmm0 152 leaq K256(%rip),%rbp 153 vpshufb %xmm7,%xmm1,%xmm1 154 vpshufb %xmm7,%xmm2,%xmm2 155 vpaddd 0(%rbp),%xmm0,%xmm4 156 vpshufb %xmm7,%xmm3,%xmm3 157 vpaddd 32(%rbp),%xmm1,%xmm5 158 vpaddd 64(%rbp),%xmm2,%xmm6 159 vpaddd 96(%rbp),%xmm3,%xmm7 160 vmovdqa %xmm4,0(%rsp) 161 movl %eax,%r14d 162 vmovdqa %xmm5,16(%rsp) 163 movl %ebx,%esi 164 vmovdqa %xmm6,32(%rsp) 165 xorl %ecx,%esi 166 vmovdqa %xmm7,48(%rsp) 167 movl %r8d,%r13d 168 jmp .Lxop_00_47 169 170.align 16 171.Lxop_00_47: 172 subq $-32*4,%rbp 173 vmovdqu (%r12),%xmm9 174 movq %r12,64+0(%rsp) 175 vpalignr $4,%xmm0,%xmm1,%xmm4 176 rorl $14,%r13d 177 movl %r14d,%eax 178 vpalignr $4,%xmm2,%xmm3,%xmm7 179 movl %r9d,%r12d 180 xorl %r8d,%r13d 181.byte 143,232,120,194,236,14 182 rorl $9,%r14d 183 xorl %r10d,%r12d 184 vpsrld $3,%xmm4,%xmm4 185 rorl $5,%r13d 186 xorl %eax,%r14d 187 vpaddd %xmm7,%xmm0,%xmm0 188 andl %r8d,%r12d 189 vpxor %xmm10,%xmm9,%xmm9 190 vmovdqu 16-128(%rdi),%xmm10 191 xorl %r8d,%r13d 192 addl 0(%rsp),%r11d 193 movl %eax,%r15d 194.byte 143,232,120,194,245,11 195 rorl $11,%r14d 196 xorl %r10d,%r12d 197 vpxor %xmm5,%xmm4,%xmm4 198 xorl %ebx,%r15d 199 rorl $6,%r13d 200 addl %r12d,%r11d 201 andl %r15d,%esi 202.byte 143,232,120,194,251,13 203 xorl %eax,%r14d 204 addl %r13d,%r11d 205 vpxor %xmm6,%xmm4,%xmm4 206 xorl %ebx,%esi 207 addl %r11d,%edx 208 vpsrld $10,%xmm3,%xmm6 209 rorl $2,%r14d 210 addl %esi,%r11d 211 vpaddd %xmm4,%xmm0,%xmm0 212 movl %edx,%r13d 213 addl %r11d,%r14d 214.byte 143,232,120,194,239,2 215 rorl $14,%r13d 216 movl %r14d,%r11d 217 vpxor %xmm6,%xmm7,%xmm7 218 movl %r8d,%r12d 219 xorl %edx,%r13d 220 rorl $9,%r14d 221 xorl %r9d,%r12d 222 vpxor %xmm5,%xmm7,%xmm7 223 rorl $5,%r13d 224 xorl %r11d,%r14d 225 andl %edx,%r12d 226 vpxor %xmm8,%xmm9,%xmm9 227 xorl %edx,%r13d 228 vpsrldq $8,%xmm7,%xmm7 229 addl 4(%rsp),%r10d 230 movl %r11d,%esi 231 rorl $11,%r14d 232 xorl %r9d,%r12d 233 vpaddd %xmm7,%xmm0,%xmm0 234 xorl %eax,%esi 235 rorl $6,%r13d 236 addl %r12d,%r10d 237 andl %esi,%r15d 238.byte 143,232,120,194,248,13 239 xorl %r11d,%r14d 240 addl %r13d,%r10d 241 vpsrld $10,%xmm0,%xmm6 242 xorl %eax,%r15d 243 addl %r10d,%ecx 244.byte 143,232,120,194,239,2 245 rorl $2,%r14d 246 addl %r15d,%r10d 247 vpxor %xmm6,%xmm7,%xmm7 248 movl %ecx,%r13d 249 addl %r10d,%r14d 250 rorl $14,%r13d 251 movl %r14d,%r10d 252 vpxor %xmm5,%xmm7,%xmm7 253 movl %edx,%r12d 254 xorl %ecx,%r13d 255 rorl $9,%r14d 256 xorl %r8d,%r12d 257 vpslldq $8,%xmm7,%xmm7 258 rorl $5,%r13d 259 xorl %r10d,%r14d 260 andl %ecx,%r12d 261 vaesenc %xmm10,%xmm9,%xmm9 262 vmovdqu 32-128(%rdi),%xmm10 263 xorl %ecx,%r13d 264 vpaddd %xmm7,%xmm0,%xmm0 265 addl 8(%rsp),%r9d 266 movl %r10d,%r15d 267 rorl $11,%r14d 268 xorl %r8d,%r12d 269 vpaddd 0(%rbp),%xmm0,%xmm6 270 xorl %r11d,%r15d 271 rorl $6,%r13d 272 addl %r12d,%r9d 273 andl %r15d,%esi 274 xorl %r10d,%r14d 275 addl %r13d,%r9d 276 xorl %r11d,%esi 277 addl %r9d,%ebx 278 rorl $2,%r14d 279 addl %esi,%r9d 280 movl %ebx,%r13d 281 addl %r9d,%r14d 282 rorl $14,%r13d 283 movl %r14d,%r9d 284 movl %ecx,%r12d 285 xorl %ebx,%r13d 286 rorl $9,%r14d 287 xorl %edx,%r12d 288 rorl $5,%r13d 289 xorl %r9d,%r14d 290 andl %ebx,%r12d 291 vaesenc %xmm10,%xmm9,%xmm9 292 vmovdqu 48-128(%rdi),%xmm10 293 xorl %ebx,%r13d 294 addl 12(%rsp),%r8d 295 movl %r9d,%esi 296 rorl $11,%r14d 297 xorl %edx,%r12d 298 xorl %r10d,%esi 299 rorl $6,%r13d 300 addl %r12d,%r8d 301 andl %esi,%r15d 302 xorl %r9d,%r14d 303 addl %r13d,%r8d 304 xorl %r10d,%r15d 305 addl %r8d,%eax 306 rorl $2,%r14d 307 addl %r15d,%r8d 308 movl %eax,%r13d 309 addl %r8d,%r14d 310 vmovdqa %xmm6,0(%rsp) 311 vpalignr $4,%xmm1,%xmm2,%xmm4 312 rorl $14,%r13d 313 movl %r14d,%r8d 314 vpalignr $4,%xmm3,%xmm0,%xmm7 315 movl %ebx,%r12d 316 xorl %eax,%r13d 317.byte 143,232,120,194,236,14 318 rorl $9,%r14d 319 xorl %ecx,%r12d 320 vpsrld $3,%xmm4,%xmm4 321 rorl $5,%r13d 322 xorl %r8d,%r14d 323 vpaddd %xmm7,%xmm1,%xmm1 324 andl %eax,%r12d 325 vaesenc %xmm10,%xmm9,%xmm9 326 vmovdqu 64-128(%rdi),%xmm10 327 xorl %eax,%r13d 328 addl 16(%rsp),%edx 329 movl %r8d,%r15d 330.byte 143,232,120,194,245,11 331 rorl $11,%r14d 332 xorl %ecx,%r12d 333 vpxor %xmm5,%xmm4,%xmm4 334 xorl %r9d,%r15d 335 rorl $6,%r13d 336 addl %r12d,%edx 337 andl %r15d,%esi 338.byte 143,232,120,194,248,13 339 xorl %r8d,%r14d 340 addl %r13d,%edx 341 vpxor %xmm6,%xmm4,%xmm4 342 xorl %r9d,%esi 343 addl %edx,%r11d 344 vpsrld $10,%xmm0,%xmm6 345 rorl $2,%r14d 346 addl %esi,%edx 347 vpaddd %xmm4,%xmm1,%xmm1 348 movl %r11d,%r13d 349 addl %edx,%r14d 350.byte 143,232,120,194,239,2 351 rorl $14,%r13d 352 movl %r14d,%edx 353 vpxor %xmm6,%xmm7,%xmm7 354 movl %eax,%r12d 355 xorl %r11d,%r13d 356 rorl $9,%r14d 357 xorl %ebx,%r12d 358 vpxor %xmm5,%xmm7,%xmm7 359 rorl $5,%r13d 360 xorl %edx,%r14d 361 andl %r11d,%r12d 362 vaesenc %xmm10,%xmm9,%xmm9 363 vmovdqu 80-128(%rdi),%xmm10 364 xorl %r11d,%r13d 365 vpsrldq $8,%xmm7,%xmm7 366 addl 20(%rsp),%ecx 367 movl %edx,%esi 368 rorl $11,%r14d 369 xorl %ebx,%r12d 370 vpaddd %xmm7,%xmm1,%xmm1 371 xorl %r8d,%esi 372 rorl $6,%r13d 373 addl %r12d,%ecx 374 andl %esi,%r15d 375.byte 143,232,120,194,249,13 376 xorl %edx,%r14d 377 addl %r13d,%ecx 378 vpsrld $10,%xmm1,%xmm6 379 xorl %r8d,%r15d 380 addl %ecx,%r10d 381.byte 143,232,120,194,239,2 382 rorl $2,%r14d 383 addl %r15d,%ecx 384 vpxor %xmm6,%xmm7,%xmm7 385 movl %r10d,%r13d 386 addl %ecx,%r14d 387 rorl $14,%r13d 388 movl %r14d,%ecx 389 vpxor %xmm5,%xmm7,%xmm7 390 movl %r11d,%r12d 391 xorl %r10d,%r13d 392 rorl $9,%r14d 393 xorl %eax,%r12d 394 vpslldq $8,%xmm7,%xmm7 395 rorl $5,%r13d 396 xorl %ecx,%r14d 397 andl %r10d,%r12d 398 vaesenc %xmm10,%xmm9,%xmm9 399 vmovdqu 96-128(%rdi),%xmm10 400 xorl %r10d,%r13d 401 vpaddd %xmm7,%xmm1,%xmm1 402 addl 24(%rsp),%ebx 403 movl %ecx,%r15d 404 rorl $11,%r14d 405 xorl %eax,%r12d 406 vpaddd 32(%rbp),%xmm1,%xmm6 407 xorl %edx,%r15d 408 rorl $6,%r13d 409 addl %r12d,%ebx 410 andl %r15d,%esi 411 xorl %ecx,%r14d 412 addl %r13d,%ebx 413 xorl %edx,%esi 414 addl %ebx,%r9d 415 rorl $2,%r14d 416 addl %esi,%ebx 417 movl %r9d,%r13d 418 addl %ebx,%r14d 419 rorl $14,%r13d 420 movl %r14d,%ebx 421 movl %r10d,%r12d 422 xorl %r9d,%r13d 423 rorl $9,%r14d 424 xorl %r11d,%r12d 425 rorl $5,%r13d 426 xorl %ebx,%r14d 427 andl %r9d,%r12d 428 vaesenc %xmm10,%xmm9,%xmm9 429 vmovdqu 112-128(%rdi),%xmm10 430 xorl %r9d,%r13d 431 addl 28(%rsp),%eax 432 movl %ebx,%esi 433 rorl $11,%r14d 434 xorl %r11d,%r12d 435 xorl %ecx,%esi 436 rorl $6,%r13d 437 addl %r12d,%eax 438 andl %esi,%r15d 439 xorl %ebx,%r14d 440 addl %r13d,%eax 441 xorl %ecx,%r15d 442 addl %eax,%r8d 443 rorl $2,%r14d 444 addl %r15d,%eax 445 movl %r8d,%r13d 446 addl %eax,%r14d 447 vmovdqa %xmm6,16(%rsp) 448 vpalignr $4,%xmm2,%xmm3,%xmm4 449 rorl $14,%r13d 450 movl %r14d,%eax 451 vpalignr $4,%xmm0,%xmm1,%xmm7 452 movl %r9d,%r12d 453 xorl %r8d,%r13d 454.byte 143,232,120,194,236,14 455 rorl $9,%r14d 456 xorl %r10d,%r12d 457 vpsrld $3,%xmm4,%xmm4 458 rorl $5,%r13d 459 xorl %eax,%r14d 460 vpaddd %xmm7,%xmm2,%xmm2 461 andl %r8d,%r12d 462 vaesenc %xmm10,%xmm9,%xmm9 463 vmovdqu 128-128(%rdi),%xmm10 464 xorl %r8d,%r13d 465 addl 32(%rsp),%r11d 466 movl %eax,%r15d 467.byte 143,232,120,194,245,11 468 rorl $11,%r14d 469 xorl %r10d,%r12d 470 vpxor %xmm5,%xmm4,%xmm4 471 xorl %ebx,%r15d 472 rorl $6,%r13d 473 addl %r12d,%r11d 474 andl %r15d,%esi 475.byte 143,232,120,194,249,13 476 xorl %eax,%r14d 477 addl %r13d,%r11d 478 vpxor %xmm6,%xmm4,%xmm4 479 xorl %ebx,%esi 480 addl %r11d,%edx 481 vpsrld $10,%xmm1,%xmm6 482 rorl $2,%r14d 483 addl %esi,%r11d 484 vpaddd %xmm4,%xmm2,%xmm2 485 movl %edx,%r13d 486 addl %r11d,%r14d 487.byte 143,232,120,194,239,2 488 rorl $14,%r13d 489 movl %r14d,%r11d 490 vpxor %xmm6,%xmm7,%xmm7 491 movl %r8d,%r12d 492 xorl %edx,%r13d 493 rorl $9,%r14d 494 xorl %r9d,%r12d 495 vpxor %xmm5,%xmm7,%xmm7 496 rorl $5,%r13d 497 xorl %r11d,%r14d 498 andl %edx,%r12d 499 vaesenc %xmm10,%xmm9,%xmm9 500 vmovdqu 144-128(%rdi),%xmm10 501 xorl %edx,%r13d 502 vpsrldq $8,%xmm7,%xmm7 503 addl 36(%rsp),%r10d 504 movl %r11d,%esi 505 rorl $11,%r14d 506 xorl %r9d,%r12d 507 vpaddd %xmm7,%xmm2,%xmm2 508 xorl %eax,%esi 509 rorl $6,%r13d 510 addl %r12d,%r10d 511 andl %esi,%r15d 512.byte 143,232,120,194,250,13 513 xorl %r11d,%r14d 514 addl %r13d,%r10d 515 vpsrld $10,%xmm2,%xmm6 516 xorl %eax,%r15d 517 addl %r10d,%ecx 518.byte 143,232,120,194,239,2 519 rorl $2,%r14d 520 addl %r15d,%r10d 521 vpxor %xmm6,%xmm7,%xmm7 522 movl %ecx,%r13d 523 addl %r10d,%r14d 524 rorl $14,%r13d 525 movl %r14d,%r10d 526 vpxor %xmm5,%xmm7,%xmm7 527 movl %edx,%r12d 528 xorl %ecx,%r13d 529 rorl $9,%r14d 530 xorl %r8d,%r12d 531 vpslldq $8,%xmm7,%xmm7 532 rorl $5,%r13d 533 xorl %r10d,%r14d 534 andl %ecx,%r12d 535 vaesenc %xmm10,%xmm9,%xmm9 536 vmovdqu 160-128(%rdi),%xmm10 537 xorl %ecx,%r13d 538 vpaddd %xmm7,%xmm2,%xmm2 539 addl 40(%rsp),%r9d 540 movl %r10d,%r15d 541 rorl $11,%r14d 542 xorl %r8d,%r12d 543 vpaddd 64(%rbp),%xmm2,%xmm6 544 xorl %r11d,%r15d 545 rorl $6,%r13d 546 addl %r12d,%r9d 547 andl %r15d,%esi 548 xorl %r10d,%r14d 549 addl %r13d,%r9d 550 xorl %r11d,%esi 551 addl %r9d,%ebx 552 rorl $2,%r14d 553 addl %esi,%r9d 554 movl %ebx,%r13d 555 addl %r9d,%r14d 556 rorl $14,%r13d 557 movl %r14d,%r9d 558 movl %ecx,%r12d 559 xorl %ebx,%r13d 560 rorl $9,%r14d 561 xorl %edx,%r12d 562 rorl $5,%r13d 563 xorl %r9d,%r14d 564 andl %ebx,%r12d 565 vaesenclast %xmm10,%xmm9,%xmm11 566 vaesenc %xmm10,%xmm9,%xmm9 567 vmovdqu 176-128(%rdi),%xmm10 568 xorl %ebx,%r13d 569 addl 44(%rsp),%r8d 570 movl %r9d,%esi 571 rorl $11,%r14d 572 xorl %edx,%r12d 573 xorl %r10d,%esi 574 rorl $6,%r13d 575 addl %r12d,%r8d 576 andl %esi,%r15d 577 xorl %r9d,%r14d 578 addl %r13d,%r8d 579 xorl %r10d,%r15d 580 addl %r8d,%eax 581 rorl $2,%r14d 582 addl %r15d,%r8d 583 movl %eax,%r13d 584 addl %r8d,%r14d 585 vmovdqa %xmm6,32(%rsp) 586 vpalignr $4,%xmm3,%xmm0,%xmm4 587 rorl $14,%r13d 588 movl %r14d,%r8d 589 vpalignr $4,%xmm1,%xmm2,%xmm7 590 movl %ebx,%r12d 591 xorl %eax,%r13d 592.byte 143,232,120,194,236,14 593 rorl $9,%r14d 594 xorl %ecx,%r12d 595 vpsrld $3,%xmm4,%xmm4 596 rorl $5,%r13d 597 xorl %r8d,%r14d 598 vpaddd %xmm7,%xmm3,%xmm3 599 andl %eax,%r12d 600 vpand %xmm12,%xmm11,%xmm8 601 vaesenc %xmm10,%xmm9,%xmm9 602 vmovdqu 192-128(%rdi),%xmm10 603 xorl %eax,%r13d 604 addl 48(%rsp),%edx 605 movl %r8d,%r15d 606.byte 143,232,120,194,245,11 607 rorl $11,%r14d 608 xorl %ecx,%r12d 609 vpxor %xmm5,%xmm4,%xmm4 610 xorl %r9d,%r15d 611 rorl $6,%r13d 612 addl %r12d,%edx 613 andl %r15d,%esi 614.byte 143,232,120,194,250,13 615 xorl %r8d,%r14d 616 addl %r13d,%edx 617 vpxor %xmm6,%xmm4,%xmm4 618 xorl %r9d,%esi 619 addl %edx,%r11d 620 vpsrld $10,%xmm2,%xmm6 621 rorl $2,%r14d 622 addl %esi,%edx 623 vpaddd %xmm4,%xmm3,%xmm3 624 movl %r11d,%r13d 625 addl %edx,%r14d 626.byte 143,232,120,194,239,2 627 rorl $14,%r13d 628 movl %r14d,%edx 629 vpxor %xmm6,%xmm7,%xmm7 630 movl %eax,%r12d 631 xorl %r11d,%r13d 632 rorl $9,%r14d 633 xorl %ebx,%r12d 634 vpxor %xmm5,%xmm7,%xmm7 635 rorl $5,%r13d 636 xorl %edx,%r14d 637 andl %r11d,%r12d 638 vaesenclast %xmm10,%xmm9,%xmm11 639 vaesenc %xmm10,%xmm9,%xmm9 640 vmovdqu 208-128(%rdi),%xmm10 641 xorl %r11d,%r13d 642 vpsrldq $8,%xmm7,%xmm7 643 addl 52(%rsp),%ecx 644 movl %edx,%esi 645 rorl $11,%r14d 646 xorl %ebx,%r12d 647 vpaddd %xmm7,%xmm3,%xmm3 648 xorl %r8d,%esi 649 rorl $6,%r13d 650 addl %r12d,%ecx 651 andl %esi,%r15d 652.byte 143,232,120,194,251,13 653 xorl %edx,%r14d 654 addl %r13d,%ecx 655 vpsrld $10,%xmm3,%xmm6 656 xorl %r8d,%r15d 657 addl %ecx,%r10d 658.byte 143,232,120,194,239,2 659 rorl $2,%r14d 660 addl %r15d,%ecx 661 vpxor %xmm6,%xmm7,%xmm7 662 movl %r10d,%r13d 663 addl %ecx,%r14d 664 rorl $14,%r13d 665 movl %r14d,%ecx 666 vpxor %xmm5,%xmm7,%xmm7 667 movl %r11d,%r12d 668 xorl %r10d,%r13d 669 rorl $9,%r14d 670 xorl %eax,%r12d 671 vpslldq $8,%xmm7,%xmm7 672 rorl $5,%r13d 673 xorl %ecx,%r14d 674 andl %r10d,%r12d 675 vpand %xmm13,%xmm11,%xmm11 676 vaesenc %xmm10,%xmm9,%xmm9 677 vmovdqu 224-128(%rdi),%xmm10 678 xorl %r10d,%r13d 679 vpaddd %xmm7,%xmm3,%xmm3 680 addl 56(%rsp),%ebx 681 movl %ecx,%r15d 682 rorl $11,%r14d 683 xorl %eax,%r12d 684 vpaddd 96(%rbp),%xmm3,%xmm6 685 xorl %edx,%r15d 686 rorl $6,%r13d 687 addl %r12d,%ebx 688 andl %r15d,%esi 689 xorl %ecx,%r14d 690 addl %r13d,%ebx 691 xorl %edx,%esi 692 addl %ebx,%r9d 693 rorl $2,%r14d 694 addl %esi,%ebx 695 movl %r9d,%r13d 696 addl %ebx,%r14d 697 rorl $14,%r13d 698 movl %r14d,%ebx 699 movl %r10d,%r12d 700 xorl %r9d,%r13d 701 rorl $9,%r14d 702 xorl %r11d,%r12d 703 rorl $5,%r13d 704 xorl %ebx,%r14d 705 andl %r9d,%r12d 706 vpor %xmm11,%xmm8,%xmm8 707 vaesenclast %xmm10,%xmm9,%xmm11 708 vmovdqu 0-128(%rdi),%xmm10 709 xorl %r9d,%r13d 710 addl 60(%rsp),%eax 711 movl %ebx,%esi 712 rorl $11,%r14d 713 xorl %r11d,%r12d 714 xorl %ecx,%esi 715 rorl $6,%r13d 716 addl %r12d,%eax 717 andl %esi,%r15d 718 xorl %ebx,%r14d 719 addl %r13d,%eax 720 xorl %ecx,%r15d 721 addl %eax,%r8d 722 rorl $2,%r14d 723 addl %r15d,%eax 724 movl %r8d,%r13d 725 addl %eax,%r14d 726 vmovdqa %xmm6,48(%rsp) 727 movq 64+0(%rsp),%r12 728 vpand %xmm14,%xmm11,%xmm11 729 movq 64+8(%rsp),%r15 730 vpor %xmm11,%xmm8,%xmm8 731 vmovdqu %xmm8,(%r15,%r12,1) 732 leaq 16(%r12),%r12 733 cmpb $0,131(%rbp) 734 jne .Lxop_00_47 735 vmovdqu (%r12),%xmm9 736 movq %r12,64+0(%rsp) 737 rorl $14,%r13d 738 movl %r14d,%eax 739 movl %r9d,%r12d 740 xorl %r8d,%r13d 741 rorl $9,%r14d 742 xorl %r10d,%r12d 743 rorl $5,%r13d 744 xorl %eax,%r14d 745 andl %r8d,%r12d 746 vpxor %xmm10,%xmm9,%xmm9 747 vmovdqu 16-128(%rdi),%xmm10 748 xorl %r8d,%r13d 749 addl 0(%rsp),%r11d 750 movl %eax,%r15d 751 rorl $11,%r14d 752 xorl %r10d,%r12d 753 xorl %ebx,%r15d 754 rorl $6,%r13d 755 addl %r12d,%r11d 756 andl %r15d,%esi 757 xorl %eax,%r14d 758 addl %r13d,%r11d 759 xorl %ebx,%esi 760 addl %r11d,%edx 761 rorl $2,%r14d 762 addl %esi,%r11d 763 movl %edx,%r13d 764 addl %r11d,%r14d 765 rorl $14,%r13d 766 movl %r14d,%r11d 767 movl %r8d,%r12d 768 xorl %edx,%r13d 769 rorl $9,%r14d 770 xorl %r9d,%r12d 771 rorl $5,%r13d 772 xorl %r11d,%r14d 773 andl %edx,%r12d 774 vpxor %xmm8,%xmm9,%xmm9 775 xorl %edx,%r13d 776 addl 4(%rsp),%r10d 777 movl %r11d,%esi 778 rorl $11,%r14d 779 xorl %r9d,%r12d 780 xorl %eax,%esi 781 rorl $6,%r13d 782 addl %r12d,%r10d 783 andl %esi,%r15d 784 xorl %r11d,%r14d 785 addl %r13d,%r10d 786 xorl %eax,%r15d 787 addl %r10d,%ecx 788 rorl $2,%r14d 789 addl %r15d,%r10d 790 movl %ecx,%r13d 791 addl %r10d,%r14d 792 rorl $14,%r13d 793 movl %r14d,%r10d 794 movl %edx,%r12d 795 xorl %ecx,%r13d 796 rorl $9,%r14d 797 xorl %r8d,%r12d 798 rorl $5,%r13d 799 xorl %r10d,%r14d 800 andl %ecx,%r12d 801 vaesenc %xmm10,%xmm9,%xmm9 802 vmovdqu 32-128(%rdi),%xmm10 803 xorl %ecx,%r13d 804 addl 8(%rsp),%r9d 805 movl %r10d,%r15d 806 rorl $11,%r14d 807 xorl %r8d,%r12d 808 xorl %r11d,%r15d 809 rorl $6,%r13d 810 addl %r12d,%r9d 811 andl %r15d,%esi 812 xorl %r10d,%r14d 813 addl %r13d,%r9d 814 xorl %r11d,%esi 815 addl %r9d,%ebx 816 rorl $2,%r14d 817 addl %esi,%r9d 818 movl %ebx,%r13d 819 addl %r9d,%r14d 820 rorl $14,%r13d 821 movl %r14d,%r9d 822 movl %ecx,%r12d 823 xorl %ebx,%r13d 824 rorl $9,%r14d 825 xorl %edx,%r12d 826 rorl $5,%r13d 827 xorl %r9d,%r14d 828 andl %ebx,%r12d 829 vaesenc %xmm10,%xmm9,%xmm9 830 vmovdqu 48-128(%rdi),%xmm10 831 xorl %ebx,%r13d 832 addl 12(%rsp),%r8d 833 movl %r9d,%esi 834 rorl $11,%r14d 835 xorl %edx,%r12d 836 xorl %r10d,%esi 837 rorl $6,%r13d 838 addl %r12d,%r8d 839 andl %esi,%r15d 840 xorl %r9d,%r14d 841 addl %r13d,%r8d 842 xorl %r10d,%r15d 843 addl %r8d,%eax 844 rorl $2,%r14d 845 addl %r15d,%r8d 846 movl %eax,%r13d 847 addl %r8d,%r14d 848 rorl $14,%r13d 849 movl %r14d,%r8d 850 movl %ebx,%r12d 851 xorl %eax,%r13d 852 rorl $9,%r14d 853 xorl %ecx,%r12d 854 rorl $5,%r13d 855 xorl %r8d,%r14d 856 andl %eax,%r12d 857 vaesenc %xmm10,%xmm9,%xmm9 858 vmovdqu 64-128(%rdi),%xmm10 859 xorl %eax,%r13d 860 addl 16(%rsp),%edx 861 movl %r8d,%r15d 862 rorl $11,%r14d 863 xorl %ecx,%r12d 864 xorl %r9d,%r15d 865 rorl $6,%r13d 866 addl %r12d,%edx 867 andl %r15d,%esi 868 xorl %r8d,%r14d 869 addl %r13d,%edx 870 xorl %r9d,%esi 871 addl %edx,%r11d 872 rorl $2,%r14d 873 addl %esi,%edx 874 movl %r11d,%r13d 875 addl %edx,%r14d 876 rorl $14,%r13d 877 movl %r14d,%edx 878 movl %eax,%r12d 879 xorl %r11d,%r13d 880 rorl $9,%r14d 881 xorl %ebx,%r12d 882 rorl $5,%r13d 883 xorl %edx,%r14d 884 andl %r11d,%r12d 885 vaesenc %xmm10,%xmm9,%xmm9 886 vmovdqu 80-128(%rdi),%xmm10 887 xorl %r11d,%r13d 888 addl 20(%rsp),%ecx 889 movl %edx,%esi 890 rorl $11,%r14d 891 xorl %ebx,%r12d 892 xorl %r8d,%esi 893 rorl $6,%r13d 894 addl %r12d,%ecx 895 andl %esi,%r15d 896 xorl %edx,%r14d 897 addl %r13d,%ecx 898 xorl %r8d,%r15d 899 addl %ecx,%r10d 900 rorl $2,%r14d 901 addl %r15d,%ecx 902 movl %r10d,%r13d 903 addl %ecx,%r14d 904 rorl $14,%r13d 905 movl %r14d,%ecx 906 movl %r11d,%r12d 907 xorl %r10d,%r13d 908 rorl $9,%r14d 909 xorl %eax,%r12d 910 rorl $5,%r13d 911 xorl %ecx,%r14d 912 andl %r10d,%r12d 913 vaesenc %xmm10,%xmm9,%xmm9 914 vmovdqu 96-128(%rdi),%xmm10 915 xorl %r10d,%r13d 916 addl 24(%rsp),%ebx 917 movl %ecx,%r15d 918 rorl $11,%r14d 919 xorl %eax,%r12d 920 xorl %edx,%r15d 921 rorl $6,%r13d 922 addl %r12d,%ebx 923 andl %r15d,%esi 924 xorl %ecx,%r14d 925 addl %r13d,%ebx 926 xorl %edx,%esi 927 addl %ebx,%r9d 928 rorl $2,%r14d 929 addl %esi,%ebx 930 movl %r9d,%r13d 931 addl %ebx,%r14d 932 rorl $14,%r13d 933 movl %r14d,%ebx 934 movl %r10d,%r12d 935 xorl %r9d,%r13d 936 rorl $9,%r14d 937 xorl %r11d,%r12d 938 rorl $5,%r13d 939 xorl %ebx,%r14d 940 andl %r9d,%r12d 941 vaesenc %xmm10,%xmm9,%xmm9 942 vmovdqu 112-128(%rdi),%xmm10 943 xorl %r9d,%r13d 944 addl 28(%rsp),%eax 945 movl %ebx,%esi 946 rorl $11,%r14d 947 xorl %r11d,%r12d 948 xorl %ecx,%esi 949 rorl $6,%r13d 950 addl %r12d,%eax 951 andl %esi,%r15d 952 xorl %ebx,%r14d 953 addl %r13d,%eax 954 xorl %ecx,%r15d 955 addl %eax,%r8d 956 rorl $2,%r14d 957 addl %r15d,%eax 958 movl %r8d,%r13d 959 addl %eax,%r14d 960 rorl $14,%r13d 961 movl %r14d,%eax 962 movl %r9d,%r12d 963 xorl %r8d,%r13d 964 rorl $9,%r14d 965 xorl %r10d,%r12d 966 rorl $5,%r13d 967 xorl %eax,%r14d 968 andl %r8d,%r12d 969 vaesenc %xmm10,%xmm9,%xmm9 970 vmovdqu 128-128(%rdi),%xmm10 971 xorl %r8d,%r13d 972 addl 32(%rsp),%r11d 973 movl %eax,%r15d 974 rorl $11,%r14d 975 xorl %r10d,%r12d 976 xorl %ebx,%r15d 977 rorl $6,%r13d 978 addl %r12d,%r11d 979 andl %r15d,%esi 980 xorl %eax,%r14d 981 addl %r13d,%r11d 982 xorl %ebx,%esi 983 addl %r11d,%edx 984 rorl $2,%r14d 985 addl %esi,%r11d 986 movl %edx,%r13d 987 addl %r11d,%r14d 988 rorl $14,%r13d 989 movl %r14d,%r11d 990 movl %r8d,%r12d 991 xorl %edx,%r13d 992 rorl $9,%r14d 993 xorl %r9d,%r12d 994 rorl $5,%r13d 995 xorl %r11d,%r14d 996 andl %edx,%r12d 997 vaesenc %xmm10,%xmm9,%xmm9 998 vmovdqu 144-128(%rdi),%xmm10 999 xorl %edx,%r13d 1000 addl 36(%rsp),%r10d 1001 movl %r11d,%esi 1002 rorl $11,%r14d 1003 xorl %r9d,%r12d 1004 xorl %eax,%esi 1005 rorl $6,%r13d 1006 addl %r12d,%r10d 1007 andl %esi,%r15d 1008 xorl %r11d,%r14d 1009 addl %r13d,%r10d 1010 xorl %eax,%r15d 1011 addl %r10d,%ecx 1012 rorl $2,%r14d 1013 addl %r15d,%r10d 1014 movl %ecx,%r13d 1015 addl %r10d,%r14d 1016 rorl $14,%r13d 1017 movl %r14d,%r10d 1018 movl %edx,%r12d 1019 xorl %ecx,%r13d 1020 rorl $9,%r14d 1021 xorl %r8d,%r12d 1022 rorl $5,%r13d 1023 xorl %r10d,%r14d 1024 andl %ecx,%r12d 1025 vaesenc %xmm10,%xmm9,%xmm9 1026 vmovdqu 160-128(%rdi),%xmm10 1027 xorl %ecx,%r13d 1028 addl 40(%rsp),%r9d 1029 movl %r10d,%r15d 1030 rorl $11,%r14d 1031 xorl %r8d,%r12d 1032 xorl %r11d,%r15d 1033 rorl $6,%r13d 1034 addl %r12d,%r9d 1035 andl %r15d,%esi 1036 xorl %r10d,%r14d 1037 addl %r13d,%r9d 1038 xorl %r11d,%esi 1039 addl %r9d,%ebx 1040 rorl $2,%r14d 1041 addl %esi,%r9d 1042 movl %ebx,%r13d 1043 addl %r9d,%r14d 1044 rorl $14,%r13d 1045 movl %r14d,%r9d 1046 movl %ecx,%r12d 1047 xorl %ebx,%r13d 1048 rorl $9,%r14d 1049 xorl %edx,%r12d 1050 rorl $5,%r13d 1051 xorl %r9d,%r14d 1052 andl %ebx,%r12d 1053 vaesenclast %xmm10,%xmm9,%xmm11 1054 vaesenc %xmm10,%xmm9,%xmm9 1055 vmovdqu 176-128(%rdi),%xmm10 1056 xorl %ebx,%r13d 1057 addl 44(%rsp),%r8d 1058 movl %r9d,%esi 1059 rorl $11,%r14d 1060 xorl %edx,%r12d 1061 xorl %r10d,%esi 1062 rorl $6,%r13d 1063 addl %r12d,%r8d 1064 andl %esi,%r15d 1065 xorl %r9d,%r14d 1066 addl %r13d,%r8d 1067 xorl %r10d,%r15d 1068 addl %r8d,%eax 1069 rorl $2,%r14d 1070 addl %r15d,%r8d 1071 movl %eax,%r13d 1072 addl %r8d,%r14d 1073 rorl $14,%r13d 1074 movl %r14d,%r8d 1075 movl %ebx,%r12d 1076 xorl %eax,%r13d 1077 rorl $9,%r14d 1078 xorl %ecx,%r12d 1079 rorl $5,%r13d 1080 xorl %r8d,%r14d 1081 andl %eax,%r12d 1082 vpand %xmm12,%xmm11,%xmm8 1083 vaesenc %xmm10,%xmm9,%xmm9 1084 vmovdqu 192-128(%rdi),%xmm10 1085 xorl %eax,%r13d 1086 addl 48(%rsp),%edx 1087 movl %r8d,%r15d 1088 rorl $11,%r14d 1089 xorl %ecx,%r12d 1090 xorl %r9d,%r15d 1091 rorl $6,%r13d 1092 addl %r12d,%edx 1093 andl %r15d,%esi 1094 xorl %r8d,%r14d 1095 addl %r13d,%edx 1096 xorl %r9d,%esi 1097 addl %edx,%r11d 1098 rorl $2,%r14d 1099 addl %esi,%edx 1100 movl %r11d,%r13d 1101 addl %edx,%r14d 1102 rorl $14,%r13d 1103 movl %r14d,%edx 1104 movl %eax,%r12d 1105 xorl %r11d,%r13d 1106 rorl $9,%r14d 1107 xorl %ebx,%r12d 1108 rorl $5,%r13d 1109 xorl %edx,%r14d 1110 andl %r11d,%r12d 1111 vaesenclast %xmm10,%xmm9,%xmm11 1112 vaesenc %xmm10,%xmm9,%xmm9 1113 vmovdqu 208-128(%rdi),%xmm10 1114 xorl %r11d,%r13d 1115 addl 52(%rsp),%ecx 1116 movl %edx,%esi 1117 rorl $11,%r14d 1118 xorl %ebx,%r12d 1119 xorl %r8d,%esi 1120 rorl $6,%r13d 1121 addl %r12d,%ecx 1122 andl %esi,%r15d 1123 xorl %edx,%r14d 1124 addl %r13d,%ecx 1125 xorl %r8d,%r15d 1126 addl %ecx,%r10d 1127 rorl $2,%r14d 1128 addl %r15d,%ecx 1129 movl %r10d,%r13d 1130 addl %ecx,%r14d 1131 rorl $14,%r13d 1132 movl %r14d,%ecx 1133 movl %r11d,%r12d 1134 xorl %r10d,%r13d 1135 rorl $9,%r14d 1136 xorl %eax,%r12d 1137 rorl $5,%r13d 1138 xorl %ecx,%r14d 1139 andl %r10d,%r12d 1140 vpand %xmm13,%xmm11,%xmm11 1141 vaesenc %xmm10,%xmm9,%xmm9 1142 vmovdqu 224-128(%rdi),%xmm10 1143 xorl %r10d,%r13d 1144 addl 56(%rsp),%ebx 1145 movl %ecx,%r15d 1146 rorl $11,%r14d 1147 xorl %eax,%r12d 1148 xorl %edx,%r15d 1149 rorl $6,%r13d 1150 addl %r12d,%ebx 1151 andl %r15d,%esi 1152 xorl %ecx,%r14d 1153 addl %r13d,%ebx 1154 xorl %edx,%esi 1155 addl %ebx,%r9d 1156 rorl $2,%r14d 1157 addl %esi,%ebx 1158 movl %r9d,%r13d 1159 addl %ebx,%r14d 1160 rorl $14,%r13d 1161 movl %r14d,%ebx 1162 movl %r10d,%r12d 1163 xorl %r9d,%r13d 1164 rorl $9,%r14d 1165 xorl %r11d,%r12d 1166 rorl $5,%r13d 1167 xorl %ebx,%r14d 1168 andl %r9d,%r12d 1169 vpor %xmm11,%xmm8,%xmm8 1170 vaesenclast %xmm10,%xmm9,%xmm11 1171 vmovdqu 0-128(%rdi),%xmm10 1172 xorl %r9d,%r13d 1173 addl 60(%rsp),%eax 1174 movl %ebx,%esi 1175 rorl $11,%r14d 1176 xorl %r11d,%r12d 1177 xorl %ecx,%esi 1178 rorl $6,%r13d 1179 addl %r12d,%eax 1180 andl %esi,%r15d 1181 xorl %ebx,%r14d 1182 addl %r13d,%eax 1183 xorl %ecx,%r15d 1184 addl %eax,%r8d 1185 rorl $2,%r14d 1186 addl %r15d,%eax 1187 movl %r8d,%r13d 1188 addl %eax,%r14d 1189 movq 64+0(%rsp),%r12 1190 movq 64+8(%rsp),%r13 1191 movq 64+40(%rsp),%r15 1192 movq 64+48(%rsp),%rsi 1193 1194 vpand %xmm14,%xmm11,%xmm11 1195 movl %r14d,%eax 1196 vpor %xmm11,%xmm8,%xmm8 1197 vmovdqu %xmm8,(%r12,%r13,1) 1198 leaq 16(%r12),%r12 1199 1200 addl 0(%r15),%eax 1201 addl 4(%r15),%ebx 1202 addl 8(%r15),%ecx 1203 addl 12(%r15),%edx 1204 addl 16(%r15),%r8d 1205 addl 20(%r15),%r9d 1206 addl 24(%r15),%r10d 1207 addl 28(%r15),%r11d 1208 1209 cmpq 64+16(%rsp),%r12 1210 1211 movl %eax,0(%r15) 1212 movl %ebx,4(%r15) 1213 movl %ecx,8(%r15) 1214 movl %edx,12(%r15) 1215 movl %r8d,16(%r15) 1216 movl %r9d,20(%r15) 1217 movl %r10d,24(%r15) 1218 movl %r11d,28(%r15) 1219 1220 jb .Lloop_xop 1221 1222 movq 64+32(%rsp),%r8 1223 movq 120(%rsp),%rsi 1224.cfi_def_cfa %rsi,8 1225 vmovdqu %xmm8,(%r8) 1226 vzeroall 1227 movq -48(%rsi),%r15 1228.cfi_restore %r15 1229 movq -40(%rsi),%r14 1230.cfi_restore %r14 1231 movq -32(%rsi),%r13 1232.cfi_restore %r13 1233 movq -24(%rsi),%r12 1234.cfi_restore %r12 1235 movq -16(%rsi),%rbp 1236.cfi_restore %rbp 1237 movq -8(%rsi),%rbx 1238.cfi_restore %rbx 1239 leaq (%rsi),%rsp 1240.cfi_def_cfa_register %rsp 1241.Lepilogue_xop: 1242 .byte 0xf3,0xc3 1243.cfi_endproc 1244.size aesni_cbc_sha256_enc_xop,.-aesni_cbc_sha256_enc_xop 1245.type aesni_cbc_sha256_enc_avx,@function 1246.align 64 1247aesni_cbc_sha256_enc_avx: 1248.cfi_startproc 1249.Lavx_shortcut: 1250 movq 8(%rsp),%r10 1251 movq %rsp,%rax 1252.cfi_def_cfa_register %rax 1253 pushq %rbx 1254.cfi_offset %rbx,-16 1255 pushq %rbp 1256.cfi_offset %rbp,-24 1257 pushq %r12 1258.cfi_offset %r12,-32 1259 pushq %r13 1260.cfi_offset %r13,-40 1261 pushq %r14 1262.cfi_offset %r14,-48 1263 pushq %r15 1264.cfi_offset %r15,-56 1265 subq $128,%rsp 1266 andq $-64,%rsp 1267 1268 shlq $6,%rdx 1269 subq %rdi,%rsi 1270 subq %rdi,%r10 1271 addq %rdi,%rdx 1272 1273 1274 movq %rsi,64+8(%rsp) 1275 movq %rdx,64+16(%rsp) 1276 1277 movq %r8,64+32(%rsp) 1278 movq %r9,64+40(%rsp) 1279 movq %r10,64+48(%rsp) 1280 movq %rax,120(%rsp) 1281.cfi_escape 0x0f,0x06,0x77,0xf8,0x00,0x06,0x23,0x08 1282.Lprologue_avx: 1283 vzeroall 1284 1285 movq %rdi,%r12 1286 leaq 128(%rcx),%rdi 1287 leaq K256+544(%rip),%r13 1288 movl 240-128(%rdi),%r14d 1289 movq %r9,%r15 1290 movq %r10,%rsi 1291 vmovdqu (%r8),%xmm8 1292 subq $9,%r14 1293 1294 movl 0(%r15),%eax 1295 movl 4(%r15),%ebx 1296 movl 8(%r15),%ecx 1297 movl 12(%r15),%edx 1298 movl 16(%r15),%r8d 1299 movl 20(%r15),%r9d 1300 movl 24(%r15),%r10d 1301 movl 28(%r15),%r11d 1302 1303 vmovdqa 0(%r13,%r14,8),%xmm14 1304 vmovdqa 16(%r13,%r14,8),%xmm13 1305 vmovdqa 32(%r13,%r14,8),%xmm12 1306 vmovdqu 0-128(%rdi),%xmm10 1307 jmp .Lloop_avx 1308.align 16 1309.Lloop_avx: 1310 vmovdqa K256+512(%rip),%xmm7 1311 vmovdqu 0(%rsi,%r12,1),%xmm0 1312 vmovdqu 16(%rsi,%r12,1),%xmm1 1313 vmovdqu 32(%rsi,%r12,1),%xmm2 1314 vmovdqu 48(%rsi,%r12,1),%xmm3 1315 vpshufb %xmm7,%xmm0,%xmm0 1316 leaq K256(%rip),%rbp 1317 vpshufb %xmm7,%xmm1,%xmm1 1318 vpshufb %xmm7,%xmm2,%xmm2 1319 vpaddd 0(%rbp),%xmm0,%xmm4 1320 vpshufb %xmm7,%xmm3,%xmm3 1321 vpaddd 32(%rbp),%xmm1,%xmm5 1322 vpaddd 64(%rbp),%xmm2,%xmm6 1323 vpaddd 96(%rbp),%xmm3,%xmm7 1324 vmovdqa %xmm4,0(%rsp) 1325 movl %eax,%r14d 1326 vmovdqa %xmm5,16(%rsp) 1327 movl %ebx,%esi 1328 vmovdqa %xmm6,32(%rsp) 1329 xorl %ecx,%esi 1330 vmovdqa %xmm7,48(%rsp) 1331 movl %r8d,%r13d 1332 jmp .Lavx_00_47 1333 1334.align 16 1335.Lavx_00_47: 1336 subq $-32*4,%rbp 1337 vmovdqu (%r12),%xmm9 1338 movq %r12,64+0(%rsp) 1339 vpalignr $4,%xmm0,%xmm1,%xmm4 1340 shrdl $14,%r13d,%r13d 1341 movl %r14d,%eax 1342 movl %r9d,%r12d 1343 vpalignr $4,%xmm2,%xmm3,%xmm7 1344 xorl %r8d,%r13d 1345 shrdl $9,%r14d,%r14d 1346 xorl %r10d,%r12d 1347 vpsrld $7,%xmm4,%xmm6 1348 shrdl $5,%r13d,%r13d 1349 xorl %eax,%r14d 1350 andl %r8d,%r12d 1351 vpaddd %xmm7,%xmm0,%xmm0 1352 vpxor %xmm10,%xmm9,%xmm9 1353 vmovdqu 16-128(%rdi),%xmm10 1354 xorl %r8d,%r13d 1355 addl 0(%rsp),%r11d 1356 movl %eax,%r15d 1357 vpsrld $3,%xmm4,%xmm7 1358 shrdl $11,%r14d,%r14d 1359 xorl %r10d,%r12d 1360 xorl %ebx,%r15d 1361 vpslld $14,%xmm4,%xmm5 1362 shrdl $6,%r13d,%r13d 1363 addl %r12d,%r11d 1364 andl %r15d,%esi 1365 vpxor %xmm6,%xmm7,%xmm4 1366 xorl %eax,%r14d 1367 addl %r13d,%r11d 1368 xorl %ebx,%esi 1369 vpshufd $250,%xmm3,%xmm7 1370 addl %r11d,%edx 1371 shrdl $2,%r14d,%r14d 1372 addl %esi,%r11d 1373 vpsrld $11,%xmm6,%xmm6 1374 movl %edx,%r13d 1375 addl %r11d,%r14d 1376 shrdl $14,%r13d,%r13d 1377 vpxor %xmm5,%xmm4,%xmm4 1378 movl %r14d,%r11d 1379 movl %r8d,%r12d 1380 xorl %edx,%r13d 1381 vpslld $11,%xmm5,%xmm5 1382 shrdl $9,%r14d,%r14d 1383 xorl %r9d,%r12d 1384 shrdl $5,%r13d,%r13d 1385 vpxor %xmm6,%xmm4,%xmm4 1386 xorl %r11d,%r14d 1387 andl %edx,%r12d 1388 vpxor %xmm8,%xmm9,%xmm9 1389 xorl %edx,%r13d 1390 vpsrld $10,%xmm7,%xmm6 1391 addl 4(%rsp),%r10d 1392 movl %r11d,%esi 1393 shrdl $11,%r14d,%r14d 1394 vpxor %xmm5,%xmm4,%xmm4 1395 xorl %r9d,%r12d 1396 xorl %eax,%esi 1397 shrdl $6,%r13d,%r13d 1398 vpsrlq $17,%xmm7,%xmm7 1399 addl %r12d,%r10d 1400 andl %esi,%r15d 1401 xorl %r11d,%r14d 1402 vpaddd %xmm4,%xmm0,%xmm0 1403 addl %r13d,%r10d 1404 xorl %eax,%r15d 1405 addl %r10d,%ecx 1406 vpxor %xmm7,%xmm6,%xmm6 1407 shrdl $2,%r14d,%r14d 1408 addl %r15d,%r10d 1409 movl %ecx,%r13d 1410 vpsrlq $2,%xmm7,%xmm7 1411 addl %r10d,%r14d 1412 shrdl $14,%r13d,%r13d 1413 movl %r14d,%r10d 1414 vpxor %xmm7,%xmm6,%xmm6 1415 movl %edx,%r12d 1416 xorl %ecx,%r13d 1417 shrdl $9,%r14d,%r14d 1418 vpshufd $132,%xmm6,%xmm6 1419 xorl %r8d,%r12d 1420 shrdl $5,%r13d,%r13d 1421 xorl %r10d,%r14d 1422 vpsrldq $8,%xmm6,%xmm6 1423 andl %ecx,%r12d 1424 vaesenc %xmm10,%xmm9,%xmm9 1425 vmovdqu 32-128(%rdi),%xmm10 1426 xorl %ecx,%r13d 1427 addl 8(%rsp),%r9d 1428 vpaddd %xmm6,%xmm0,%xmm0 1429 movl %r10d,%r15d 1430 shrdl $11,%r14d,%r14d 1431 xorl %r8d,%r12d 1432 vpshufd $80,%xmm0,%xmm7 1433 xorl %r11d,%r15d 1434 shrdl $6,%r13d,%r13d 1435 addl %r12d,%r9d 1436 vpsrld $10,%xmm7,%xmm6 1437 andl %r15d,%esi 1438 xorl %r10d,%r14d 1439 addl %r13d,%r9d 1440 vpsrlq $17,%xmm7,%xmm7 1441 xorl %r11d,%esi 1442 addl %r9d,%ebx 1443 shrdl $2,%r14d,%r14d 1444 vpxor %xmm7,%xmm6,%xmm6 1445 addl %esi,%r9d 1446 movl %ebx,%r13d 1447 addl %r9d,%r14d 1448 vpsrlq $2,%xmm7,%xmm7 1449 shrdl $14,%r13d,%r13d 1450 movl %r14d,%r9d 1451 movl %ecx,%r12d 1452 vpxor %xmm7,%xmm6,%xmm6 1453 xorl %ebx,%r13d 1454 shrdl $9,%r14d,%r14d 1455 xorl %edx,%r12d 1456 vpshufd $232,%xmm6,%xmm6 1457 shrdl $5,%r13d,%r13d 1458 xorl %r9d,%r14d 1459 andl %ebx,%r12d 1460 vpslldq $8,%xmm6,%xmm6 1461 vaesenc %xmm10,%xmm9,%xmm9 1462 vmovdqu 48-128(%rdi),%xmm10 1463 xorl %ebx,%r13d 1464 addl 12(%rsp),%r8d 1465 movl %r9d,%esi 1466 vpaddd %xmm6,%xmm0,%xmm0 1467 shrdl $11,%r14d,%r14d 1468 xorl %edx,%r12d 1469 xorl %r10d,%esi 1470 vpaddd 0(%rbp),%xmm0,%xmm6 1471 shrdl $6,%r13d,%r13d 1472 addl %r12d,%r8d 1473 andl %esi,%r15d 1474 xorl %r9d,%r14d 1475 addl %r13d,%r8d 1476 xorl %r10d,%r15d 1477 addl %r8d,%eax 1478 shrdl $2,%r14d,%r14d 1479 addl %r15d,%r8d 1480 movl %eax,%r13d 1481 addl %r8d,%r14d 1482 vmovdqa %xmm6,0(%rsp) 1483 vpalignr $4,%xmm1,%xmm2,%xmm4 1484 shrdl $14,%r13d,%r13d 1485 movl %r14d,%r8d 1486 movl %ebx,%r12d 1487 vpalignr $4,%xmm3,%xmm0,%xmm7 1488 xorl %eax,%r13d 1489 shrdl $9,%r14d,%r14d 1490 xorl %ecx,%r12d 1491 vpsrld $7,%xmm4,%xmm6 1492 shrdl $5,%r13d,%r13d 1493 xorl %r8d,%r14d 1494 andl %eax,%r12d 1495 vpaddd %xmm7,%xmm1,%xmm1 1496 vaesenc %xmm10,%xmm9,%xmm9 1497 vmovdqu 64-128(%rdi),%xmm10 1498 xorl %eax,%r13d 1499 addl 16(%rsp),%edx 1500 movl %r8d,%r15d 1501 vpsrld $3,%xmm4,%xmm7 1502 shrdl $11,%r14d,%r14d 1503 xorl %ecx,%r12d 1504 xorl %r9d,%r15d 1505 vpslld $14,%xmm4,%xmm5 1506 shrdl $6,%r13d,%r13d 1507 addl %r12d,%edx 1508 andl %r15d,%esi 1509 vpxor %xmm6,%xmm7,%xmm4 1510 xorl %r8d,%r14d 1511 addl %r13d,%edx 1512 xorl %r9d,%esi 1513 vpshufd $250,%xmm0,%xmm7 1514 addl %edx,%r11d 1515 shrdl $2,%r14d,%r14d 1516 addl %esi,%edx 1517 vpsrld $11,%xmm6,%xmm6 1518 movl %r11d,%r13d 1519 addl %edx,%r14d 1520 shrdl $14,%r13d,%r13d 1521 vpxor %xmm5,%xmm4,%xmm4 1522 movl %r14d,%edx 1523 movl %eax,%r12d 1524 xorl %r11d,%r13d 1525 vpslld $11,%xmm5,%xmm5 1526 shrdl $9,%r14d,%r14d 1527 xorl %ebx,%r12d 1528 shrdl $5,%r13d,%r13d 1529 vpxor %xmm6,%xmm4,%xmm4 1530 xorl %edx,%r14d 1531 andl %r11d,%r12d 1532 vaesenc %xmm10,%xmm9,%xmm9 1533 vmovdqu 80-128(%rdi),%xmm10 1534 xorl %r11d,%r13d 1535 vpsrld $10,%xmm7,%xmm6 1536 addl 20(%rsp),%ecx 1537 movl %edx,%esi 1538 shrdl $11,%r14d,%r14d 1539 vpxor %xmm5,%xmm4,%xmm4 1540 xorl %ebx,%r12d 1541 xorl %r8d,%esi 1542 shrdl $6,%r13d,%r13d 1543 vpsrlq $17,%xmm7,%xmm7 1544 addl %r12d,%ecx 1545 andl %esi,%r15d 1546 xorl %edx,%r14d 1547 vpaddd %xmm4,%xmm1,%xmm1 1548 addl %r13d,%ecx 1549 xorl %r8d,%r15d 1550 addl %ecx,%r10d 1551 vpxor %xmm7,%xmm6,%xmm6 1552 shrdl $2,%r14d,%r14d 1553 addl %r15d,%ecx 1554 movl %r10d,%r13d 1555 vpsrlq $2,%xmm7,%xmm7 1556 addl %ecx,%r14d 1557 shrdl $14,%r13d,%r13d 1558 movl %r14d,%ecx 1559 vpxor %xmm7,%xmm6,%xmm6 1560 movl %r11d,%r12d 1561 xorl %r10d,%r13d 1562 shrdl $9,%r14d,%r14d 1563 vpshufd $132,%xmm6,%xmm6 1564 xorl %eax,%r12d 1565 shrdl $5,%r13d,%r13d 1566 xorl %ecx,%r14d 1567 vpsrldq $8,%xmm6,%xmm6 1568 andl %r10d,%r12d 1569 vaesenc %xmm10,%xmm9,%xmm9 1570 vmovdqu 96-128(%rdi),%xmm10 1571 xorl %r10d,%r13d 1572 addl 24(%rsp),%ebx 1573 vpaddd %xmm6,%xmm1,%xmm1 1574 movl %ecx,%r15d 1575 shrdl $11,%r14d,%r14d 1576 xorl %eax,%r12d 1577 vpshufd $80,%xmm1,%xmm7 1578 xorl %edx,%r15d 1579 shrdl $6,%r13d,%r13d 1580 addl %r12d,%ebx 1581 vpsrld $10,%xmm7,%xmm6 1582 andl %r15d,%esi 1583 xorl %ecx,%r14d 1584 addl %r13d,%ebx 1585 vpsrlq $17,%xmm7,%xmm7 1586 xorl %edx,%esi 1587 addl %ebx,%r9d 1588 shrdl $2,%r14d,%r14d 1589 vpxor %xmm7,%xmm6,%xmm6 1590 addl %esi,%ebx 1591 movl %r9d,%r13d 1592 addl %ebx,%r14d 1593 vpsrlq $2,%xmm7,%xmm7 1594 shrdl $14,%r13d,%r13d 1595 movl %r14d,%ebx 1596 movl %r10d,%r12d 1597 vpxor %xmm7,%xmm6,%xmm6 1598 xorl %r9d,%r13d 1599 shrdl $9,%r14d,%r14d 1600 xorl %r11d,%r12d 1601 vpshufd $232,%xmm6,%xmm6 1602 shrdl $5,%r13d,%r13d 1603 xorl %ebx,%r14d 1604 andl %r9d,%r12d 1605 vpslldq $8,%xmm6,%xmm6 1606 vaesenc %xmm10,%xmm9,%xmm9 1607 vmovdqu 112-128(%rdi),%xmm10 1608 xorl %r9d,%r13d 1609 addl 28(%rsp),%eax 1610 movl %ebx,%esi 1611 vpaddd %xmm6,%xmm1,%xmm1 1612 shrdl $11,%r14d,%r14d 1613 xorl %r11d,%r12d 1614 xorl %ecx,%esi 1615 vpaddd 32(%rbp),%xmm1,%xmm6 1616 shrdl $6,%r13d,%r13d 1617 addl %r12d,%eax 1618 andl %esi,%r15d 1619 xorl %ebx,%r14d 1620 addl %r13d,%eax 1621 xorl %ecx,%r15d 1622 addl %eax,%r8d 1623 shrdl $2,%r14d,%r14d 1624 addl %r15d,%eax 1625 movl %r8d,%r13d 1626 addl %eax,%r14d 1627 vmovdqa %xmm6,16(%rsp) 1628 vpalignr $4,%xmm2,%xmm3,%xmm4 1629 shrdl $14,%r13d,%r13d 1630 movl %r14d,%eax 1631 movl %r9d,%r12d 1632 vpalignr $4,%xmm0,%xmm1,%xmm7 1633 xorl %r8d,%r13d 1634 shrdl $9,%r14d,%r14d 1635 xorl %r10d,%r12d 1636 vpsrld $7,%xmm4,%xmm6 1637 shrdl $5,%r13d,%r13d 1638 xorl %eax,%r14d 1639 andl %r8d,%r12d 1640 vpaddd %xmm7,%xmm2,%xmm2 1641 vaesenc %xmm10,%xmm9,%xmm9 1642 vmovdqu 128-128(%rdi),%xmm10 1643 xorl %r8d,%r13d 1644 addl 32(%rsp),%r11d 1645 movl %eax,%r15d 1646 vpsrld $3,%xmm4,%xmm7 1647 shrdl $11,%r14d,%r14d 1648 xorl %r10d,%r12d 1649 xorl %ebx,%r15d 1650 vpslld $14,%xmm4,%xmm5 1651 shrdl $6,%r13d,%r13d 1652 addl %r12d,%r11d 1653 andl %r15d,%esi 1654 vpxor %xmm6,%xmm7,%xmm4 1655 xorl %eax,%r14d 1656 addl %r13d,%r11d 1657 xorl %ebx,%esi 1658 vpshufd $250,%xmm1,%xmm7 1659 addl %r11d,%edx 1660 shrdl $2,%r14d,%r14d 1661 addl %esi,%r11d 1662 vpsrld $11,%xmm6,%xmm6 1663 movl %edx,%r13d 1664 addl %r11d,%r14d 1665 shrdl $14,%r13d,%r13d 1666 vpxor %xmm5,%xmm4,%xmm4 1667 movl %r14d,%r11d 1668 movl %r8d,%r12d 1669 xorl %edx,%r13d 1670 vpslld $11,%xmm5,%xmm5 1671 shrdl $9,%r14d,%r14d 1672 xorl %r9d,%r12d 1673 shrdl $5,%r13d,%r13d 1674 vpxor %xmm6,%xmm4,%xmm4 1675 xorl %r11d,%r14d 1676 andl %edx,%r12d 1677 vaesenc %xmm10,%xmm9,%xmm9 1678 vmovdqu 144-128(%rdi),%xmm10 1679 xorl %edx,%r13d 1680 vpsrld $10,%xmm7,%xmm6 1681 addl 36(%rsp),%r10d 1682 movl %r11d,%esi 1683 shrdl $11,%r14d,%r14d 1684 vpxor %xmm5,%xmm4,%xmm4 1685 xorl %r9d,%r12d 1686 xorl %eax,%esi 1687 shrdl $6,%r13d,%r13d 1688 vpsrlq $17,%xmm7,%xmm7 1689 addl %r12d,%r10d 1690 andl %esi,%r15d 1691 xorl %r11d,%r14d 1692 vpaddd %xmm4,%xmm2,%xmm2 1693 addl %r13d,%r10d 1694 xorl %eax,%r15d 1695 addl %r10d,%ecx 1696 vpxor %xmm7,%xmm6,%xmm6 1697 shrdl $2,%r14d,%r14d 1698 addl %r15d,%r10d 1699 movl %ecx,%r13d 1700 vpsrlq $2,%xmm7,%xmm7 1701 addl %r10d,%r14d 1702 shrdl $14,%r13d,%r13d 1703 movl %r14d,%r10d 1704 vpxor %xmm7,%xmm6,%xmm6 1705 movl %edx,%r12d 1706 xorl %ecx,%r13d 1707 shrdl $9,%r14d,%r14d 1708 vpshufd $132,%xmm6,%xmm6 1709 xorl %r8d,%r12d 1710 shrdl $5,%r13d,%r13d 1711 xorl %r10d,%r14d 1712 vpsrldq $8,%xmm6,%xmm6 1713 andl %ecx,%r12d 1714 vaesenc %xmm10,%xmm9,%xmm9 1715 vmovdqu 160-128(%rdi),%xmm10 1716 xorl %ecx,%r13d 1717 addl 40(%rsp),%r9d 1718 vpaddd %xmm6,%xmm2,%xmm2 1719 movl %r10d,%r15d 1720 shrdl $11,%r14d,%r14d 1721 xorl %r8d,%r12d 1722 vpshufd $80,%xmm2,%xmm7 1723 xorl %r11d,%r15d 1724 shrdl $6,%r13d,%r13d 1725 addl %r12d,%r9d 1726 vpsrld $10,%xmm7,%xmm6 1727 andl %r15d,%esi 1728 xorl %r10d,%r14d 1729 addl %r13d,%r9d 1730 vpsrlq $17,%xmm7,%xmm7 1731 xorl %r11d,%esi 1732 addl %r9d,%ebx 1733 shrdl $2,%r14d,%r14d 1734 vpxor %xmm7,%xmm6,%xmm6 1735 addl %esi,%r9d 1736 movl %ebx,%r13d 1737 addl %r9d,%r14d 1738 vpsrlq $2,%xmm7,%xmm7 1739 shrdl $14,%r13d,%r13d 1740 movl %r14d,%r9d 1741 movl %ecx,%r12d 1742 vpxor %xmm7,%xmm6,%xmm6 1743 xorl %ebx,%r13d 1744 shrdl $9,%r14d,%r14d 1745 xorl %edx,%r12d 1746 vpshufd $232,%xmm6,%xmm6 1747 shrdl $5,%r13d,%r13d 1748 xorl %r9d,%r14d 1749 andl %ebx,%r12d 1750 vpslldq $8,%xmm6,%xmm6 1751 vaesenclast %xmm10,%xmm9,%xmm11 1752 vaesenc %xmm10,%xmm9,%xmm9 1753 vmovdqu 176-128(%rdi),%xmm10 1754 xorl %ebx,%r13d 1755 addl 44(%rsp),%r8d 1756 movl %r9d,%esi 1757 vpaddd %xmm6,%xmm2,%xmm2 1758 shrdl $11,%r14d,%r14d 1759 xorl %edx,%r12d 1760 xorl %r10d,%esi 1761 vpaddd 64(%rbp),%xmm2,%xmm6 1762 shrdl $6,%r13d,%r13d 1763 addl %r12d,%r8d 1764 andl %esi,%r15d 1765 xorl %r9d,%r14d 1766 addl %r13d,%r8d 1767 xorl %r10d,%r15d 1768 addl %r8d,%eax 1769 shrdl $2,%r14d,%r14d 1770 addl %r15d,%r8d 1771 movl %eax,%r13d 1772 addl %r8d,%r14d 1773 vmovdqa %xmm6,32(%rsp) 1774 vpalignr $4,%xmm3,%xmm0,%xmm4 1775 shrdl $14,%r13d,%r13d 1776 movl %r14d,%r8d 1777 movl %ebx,%r12d 1778 vpalignr $4,%xmm1,%xmm2,%xmm7 1779 xorl %eax,%r13d 1780 shrdl $9,%r14d,%r14d 1781 xorl %ecx,%r12d 1782 vpsrld $7,%xmm4,%xmm6 1783 shrdl $5,%r13d,%r13d 1784 xorl %r8d,%r14d 1785 andl %eax,%r12d 1786 vpaddd %xmm7,%xmm3,%xmm3 1787 vpand %xmm12,%xmm11,%xmm8 1788 vaesenc %xmm10,%xmm9,%xmm9 1789 vmovdqu 192-128(%rdi),%xmm10 1790 xorl %eax,%r13d 1791 addl 48(%rsp),%edx 1792 movl %r8d,%r15d 1793 vpsrld $3,%xmm4,%xmm7 1794 shrdl $11,%r14d,%r14d 1795 xorl %ecx,%r12d 1796 xorl %r9d,%r15d 1797 vpslld $14,%xmm4,%xmm5 1798 shrdl $6,%r13d,%r13d 1799 addl %r12d,%edx 1800 andl %r15d,%esi 1801 vpxor %xmm6,%xmm7,%xmm4 1802 xorl %r8d,%r14d 1803 addl %r13d,%edx 1804 xorl %r9d,%esi 1805 vpshufd $250,%xmm2,%xmm7 1806 addl %edx,%r11d 1807 shrdl $2,%r14d,%r14d 1808 addl %esi,%edx 1809 vpsrld $11,%xmm6,%xmm6 1810 movl %r11d,%r13d 1811 addl %edx,%r14d 1812 shrdl $14,%r13d,%r13d 1813 vpxor %xmm5,%xmm4,%xmm4 1814 movl %r14d,%edx 1815 movl %eax,%r12d 1816 xorl %r11d,%r13d 1817 vpslld $11,%xmm5,%xmm5 1818 shrdl $9,%r14d,%r14d 1819 xorl %ebx,%r12d 1820 shrdl $5,%r13d,%r13d 1821 vpxor %xmm6,%xmm4,%xmm4 1822 xorl %edx,%r14d 1823 andl %r11d,%r12d 1824 vaesenclast %xmm10,%xmm9,%xmm11 1825 vaesenc %xmm10,%xmm9,%xmm9 1826 vmovdqu 208-128(%rdi),%xmm10 1827 xorl %r11d,%r13d 1828 vpsrld $10,%xmm7,%xmm6 1829 addl 52(%rsp),%ecx 1830 movl %edx,%esi 1831 shrdl $11,%r14d,%r14d 1832 vpxor %xmm5,%xmm4,%xmm4 1833 xorl %ebx,%r12d 1834 xorl %r8d,%esi 1835 shrdl $6,%r13d,%r13d 1836 vpsrlq $17,%xmm7,%xmm7 1837 addl %r12d,%ecx 1838 andl %esi,%r15d 1839 xorl %edx,%r14d 1840 vpaddd %xmm4,%xmm3,%xmm3 1841 addl %r13d,%ecx 1842 xorl %r8d,%r15d 1843 addl %ecx,%r10d 1844 vpxor %xmm7,%xmm6,%xmm6 1845 shrdl $2,%r14d,%r14d 1846 addl %r15d,%ecx 1847 movl %r10d,%r13d 1848 vpsrlq $2,%xmm7,%xmm7 1849 addl %ecx,%r14d 1850 shrdl $14,%r13d,%r13d 1851 movl %r14d,%ecx 1852 vpxor %xmm7,%xmm6,%xmm6 1853 movl %r11d,%r12d 1854 xorl %r10d,%r13d 1855 shrdl $9,%r14d,%r14d 1856 vpshufd $132,%xmm6,%xmm6 1857 xorl %eax,%r12d 1858 shrdl $5,%r13d,%r13d 1859 xorl %ecx,%r14d 1860 vpsrldq $8,%xmm6,%xmm6 1861 andl %r10d,%r12d 1862 vpand %xmm13,%xmm11,%xmm11 1863 vaesenc %xmm10,%xmm9,%xmm9 1864 vmovdqu 224-128(%rdi),%xmm10 1865 xorl %r10d,%r13d 1866 addl 56(%rsp),%ebx 1867 vpaddd %xmm6,%xmm3,%xmm3 1868 movl %ecx,%r15d 1869 shrdl $11,%r14d,%r14d 1870 xorl %eax,%r12d 1871 vpshufd $80,%xmm3,%xmm7 1872 xorl %edx,%r15d 1873 shrdl $6,%r13d,%r13d 1874 addl %r12d,%ebx 1875 vpsrld $10,%xmm7,%xmm6 1876 andl %r15d,%esi 1877 xorl %ecx,%r14d 1878 addl %r13d,%ebx 1879 vpsrlq $17,%xmm7,%xmm7 1880 xorl %edx,%esi 1881 addl %ebx,%r9d 1882 shrdl $2,%r14d,%r14d 1883 vpxor %xmm7,%xmm6,%xmm6 1884 addl %esi,%ebx 1885 movl %r9d,%r13d 1886 addl %ebx,%r14d 1887 vpsrlq $2,%xmm7,%xmm7 1888 shrdl $14,%r13d,%r13d 1889 movl %r14d,%ebx 1890 movl %r10d,%r12d 1891 vpxor %xmm7,%xmm6,%xmm6 1892 xorl %r9d,%r13d 1893 shrdl $9,%r14d,%r14d 1894 xorl %r11d,%r12d 1895 vpshufd $232,%xmm6,%xmm6 1896 shrdl $5,%r13d,%r13d 1897 xorl %ebx,%r14d 1898 andl %r9d,%r12d 1899 vpslldq $8,%xmm6,%xmm6 1900 vpor %xmm11,%xmm8,%xmm8 1901 vaesenclast %xmm10,%xmm9,%xmm11 1902 vmovdqu 0-128(%rdi),%xmm10 1903 xorl %r9d,%r13d 1904 addl 60(%rsp),%eax 1905 movl %ebx,%esi 1906 vpaddd %xmm6,%xmm3,%xmm3 1907 shrdl $11,%r14d,%r14d 1908 xorl %r11d,%r12d 1909 xorl %ecx,%esi 1910 vpaddd 96(%rbp),%xmm3,%xmm6 1911 shrdl $6,%r13d,%r13d 1912 addl %r12d,%eax 1913 andl %esi,%r15d 1914 xorl %ebx,%r14d 1915 addl %r13d,%eax 1916 xorl %ecx,%r15d 1917 addl %eax,%r8d 1918 shrdl $2,%r14d,%r14d 1919 addl %r15d,%eax 1920 movl %r8d,%r13d 1921 addl %eax,%r14d 1922 vmovdqa %xmm6,48(%rsp) 1923 movq 64+0(%rsp),%r12 1924 vpand %xmm14,%xmm11,%xmm11 1925 movq 64+8(%rsp),%r15 1926 vpor %xmm11,%xmm8,%xmm8 1927 vmovdqu %xmm8,(%r15,%r12,1) 1928 leaq 16(%r12),%r12 1929 cmpb $0,131(%rbp) 1930 jne .Lavx_00_47 1931 vmovdqu (%r12),%xmm9 1932 movq %r12,64+0(%rsp) 1933 shrdl $14,%r13d,%r13d 1934 movl %r14d,%eax 1935 movl %r9d,%r12d 1936 xorl %r8d,%r13d 1937 shrdl $9,%r14d,%r14d 1938 xorl %r10d,%r12d 1939 shrdl $5,%r13d,%r13d 1940 xorl %eax,%r14d 1941 andl %r8d,%r12d 1942 vpxor %xmm10,%xmm9,%xmm9 1943 vmovdqu 16-128(%rdi),%xmm10 1944 xorl %r8d,%r13d 1945 addl 0(%rsp),%r11d 1946 movl %eax,%r15d 1947 shrdl $11,%r14d,%r14d 1948 xorl %r10d,%r12d 1949 xorl %ebx,%r15d 1950 shrdl $6,%r13d,%r13d 1951 addl %r12d,%r11d 1952 andl %r15d,%esi 1953 xorl %eax,%r14d 1954 addl %r13d,%r11d 1955 xorl %ebx,%esi 1956 addl %r11d,%edx 1957 shrdl $2,%r14d,%r14d 1958 addl %esi,%r11d 1959 movl %edx,%r13d 1960 addl %r11d,%r14d 1961 shrdl $14,%r13d,%r13d 1962 movl %r14d,%r11d 1963 movl %r8d,%r12d 1964 xorl %edx,%r13d 1965 shrdl $9,%r14d,%r14d 1966 xorl %r9d,%r12d 1967 shrdl $5,%r13d,%r13d 1968 xorl %r11d,%r14d 1969 andl %edx,%r12d 1970 vpxor %xmm8,%xmm9,%xmm9 1971 xorl %edx,%r13d 1972 addl 4(%rsp),%r10d 1973 movl %r11d,%esi 1974 shrdl $11,%r14d,%r14d 1975 xorl %r9d,%r12d 1976 xorl %eax,%esi 1977 shrdl $6,%r13d,%r13d 1978 addl %r12d,%r10d 1979 andl %esi,%r15d 1980 xorl %r11d,%r14d 1981 addl %r13d,%r10d 1982 xorl %eax,%r15d 1983 addl %r10d,%ecx 1984 shrdl $2,%r14d,%r14d 1985 addl %r15d,%r10d 1986 movl %ecx,%r13d 1987 addl %r10d,%r14d 1988 shrdl $14,%r13d,%r13d 1989 movl %r14d,%r10d 1990 movl %edx,%r12d 1991 xorl %ecx,%r13d 1992 shrdl $9,%r14d,%r14d 1993 xorl %r8d,%r12d 1994 shrdl $5,%r13d,%r13d 1995 xorl %r10d,%r14d 1996 andl %ecx,%r12d 1997 vaesenc %xmm10,%xmm9,%xmm9 1998 vmovdqu 32-128(%rdi),%xmm10 1999 xorl %ecx,%r13d 2000 addl 8(%rsp),%r9d 2001 movl %r10d,%r15d 2002 shrdl $11,%r14d,%r14d 2003 xorl %r8d,%r12d 2004 xorl %r11d,%r15d 2005 shrdl $6,%r13d,%r13d 2006 addl %r12d,%r9d 2007 andl %r15d,%esi 2008 xorl %r10d,%r14d 2009 addl %r13d,%r9d 2010 xorl %r11d,%esi 2011 addl %r9d,%ebx 2012 shrdl $2,%r14d,%r14d 2013 addl %esi,%r9d 2014 movl %ebx,%r13d 2015 addl %r9d,%r14d 2016 shrdl $14,%r13d,%r13d 2017 movl %r14d,%r9d 2018 movl %ecx,%r12d 2019 xorl %ebx,%r13d 2020 shrdl $9,%r14d,%r14d 2021 xorl %edx,%r12d 2022 shrdl $5,%r13d,%r13d 2023 xorl %r9d,%r14d 2024 andl %ebx,%r12d 2025 vaesenc %xmm10,%xmm9,%xmm9 2026 vmovdqu 48-128(%rdi),%xmm10 2027 xorl %ebx,%r13d 2028 addl 12(%rsp),%r8d 2029 movl %r9d,%esi 2030 shrdl $11,%r14d,%r14d 2031 xorl %edx,%r12d 2032 xorl %r10d,%esi 2033 shrdl $6,%r13d,%r13d 2034 addl %r12d,%r8d 2035 andl %esi,%r15d 2036 xorl %r9d,%r14d 2037 addl %r13d,%r8d 2038 xorl %r10d,%r15d 2039 addl %r8d,%eax 2040 shrdl $2,%r14d,%r14d 2041 addl %r15d,%r8d 2042 movl %eax,%r13d 2043 addl %r8d,%r14d 2044 shrdl $14,%r13d,%r13d 2045 movl %r14d,%r8d 2046 movl %ebx,%r12d 2047 xorl %eax,%r13d 2048 shrdl $9,%r14d,%r14d 2049 xorl %ecx,%r12d 2050 shrdl $5,%r13d,%r13d 2051 xorl %r8d,%r14d 2052 andl %eax,%r12d 2053 vaesenc %xmm10,%xmm9,%xmm9 2054 vmovdqu 64-128(%rdi),%xmm10 2055 xorl %eax,%r13d 2056 addl 16(%rsp),%edx 2057 movl %r8d,%r15d 2058 shrdl $11,%r14d,%r14d 2059 xorl %ecx,%r12d 2060 xorl %r9d,%r15d 2061 shrdl $6,%r13d,%r13d 2062 addl %r12d,%edx 2063 andl %r15d,%esi 2064 xorl %r8d,%r14d 2065 addl %r13d,%edx 2066 xorl %r9d,%esi 2067 addl %edx,%r11d 2068 shrdl $2,%r14d,%r14d 2069 addl %esi,%edx 2070 movl %r11d,%r13d 2071 addl %edx,%r14d 2072 shrdl $14,%r13d,%r13d 2073 movl %r14d,%edx 2074 movl %eax,%r12d 2075 xorl %r11d,%r13d 2076 shrdl $9,%r14d,%r14d 2077 xorl %ebx,%r12d 2078 shrdl $5,%r13d,%r13d 2079 xorl %edx,%r14d 2080 andl %r11d,%r12d 2081 vaesenc %xmm10,%xmm9,%xmm9 2082 vmovdqu 80-128(%rdi),%xmm10 2083 xorl %r11d,%r13d 2084 addl 20(%rsp),%ecx 2085 movl %edx,%esi 2086 shrdl $11,%r14d,%r14d 2087 xorl %ebx,%r12d 2088 xorl %r8d,%esi 2089 shrdl $6,%r13d,%r13d 2090 addl %r12d,%ecx 2091 andl %esi,%r15d 2092 xorl %edx,%r14d 2093 addl %r13d,%ecx 2094 xorl %r8d,%r15d 2095 addl %ecx,%r10d 2096 shrdl $2,%r14d,%r14d 2097 addl %r15d,%ecx 2098 movl %r10d,%r13d 2099 addl %ecx,%r14d 2100 shrdl $14,%r13d,%r13d 2101 movl %r14d,%ecx 2102 movl %r11d,%r12d 2103 xorl %r10d,%r13d 2104 shrdl $9,%r14d,%r14d 2105 xorl %eax,%r12d 2106 shrdl $5,%r13d,%r13d 2107 xorl %ecx,%r14d 2108 andl %r10d,%r12d 2109 vaesenc %xmm10,%xmm9,%xmm9 2110 vmovdqu 96-128(%rdi),%xmm10 2111 xorl %r10d,%r13d 2112 addl 24(%rsp),%ebx 2113 movl %ecx,%r15d 2114 shrdl $11,%r14d,%r14d 2115 xorl %eax,%r12d 2116 xorl %edx,%r15d 2117 shrdl $6,%r13d,%r13d 2118 addl %r12d,%ebx 2119 andl %r15d,%esi 2120 xorl %ecx,%r14d 2121 addl %r13d,%ebx 2122 xorl %edx,%esi 2123 addl %ebx,%r9d 2124 shrdl $2,%r14d,%r14d 2125 addl %esi,%ebx 2126 movl %r9d,%r13d 2127 addl %ebx,%r14d 2128 shrdl $14,%r13d,%r13d 2129 movl %r14d,%ebx 2130 movl %r10d,%r12d 2131 xorl %r9d,%r13d 2132 shrdl $9,%r14d,%r14d 2133 xorl %r11d,%r12d 2134 shrdl $5,%r13d,%r13d 2135 xorl %ebx,%r14d 2136 andl %r9d,%r12d 2137 vaesenc %xmm10,%xmm9,%xmm9 2138 vmovdqu 112-128(%rdi),%xmm10 2139 xorl %r9d,%r13d 2140 addl 28(%rsp),%eax 2141 movl %ebx,%esi 2142 shrdl $11,%r14d,%r14d 2143 xorl %r11d,%r12d 2144 xorl %ecx,%esi 2145 shrdl $6,%r13d,%r13d 2146 addl %r12d,%eax 2147 andl %esi,%r15d 2148 xorl %ebx,%r14d 2149 addl %r13d,%eax 2150 xorl %ecx,%r15d 2151 addl %eax,%r8d 2152 shrdl $2,%r14d,%r14d 2153 addl %r15d,%eax 2154 movl %r8d,%r13d 2155 addl %eax,%r14d 2156 shrdl $14,%r13d,%r13d 2157 movl %r14d,%eax 2158 movl %r9d,%r12d 2159 xorl %r8d,%r13d 2160 shrdl $9,%r14d,%r14d 2161 xorl %r10d,%r12d 2162 shrdl $5,%r13d,%r13d 2163 xorl %eax,%r14d 2164 andl %r8d,%r12d 2165 vaesenc %xmm10,%xmm9,%xmm9 2166 vmovdqu 128-128(%rdi),%xmm10 2167 xorl %r8d,%r13d 2168 addl 32(%rsp),%r11d 2169 movl %eax,%r15d 2170 shrdl $11,%r14d,%r14d 2171 xorl %r10d,%r12d 2172 xorl %ebx,%r15d 2173 shrdl $6,%r13d,%r13d 2174 addl %r12d,%r11d 2175 andl %r15d,%esi 2176 xorl %eax,%r14d 2177 addl %r13d,%r11d 2178 xorl %ebx,%esi 2179 addl %r11d,%edx 2180 shrdl $2,%r14d,%r14d 2181 addl %esi,%r11d 2182 movl %edx,%r13d 2183 addl %r11d,%r14d 2184 shrdl $14,%r13d,%r13d 2185 movl %r14d,%r11d 2186 movl %r8d,%r12d 2187 xorl %edx,%r13d 2188 shrdl $9,%r14d,%r14d 2189 xorl %r9d,%r12d 2190 shrdl $5,%r13d,%r13d 2191 xorl %r11d,%r14d 2192 andl %edx,%r12d 2193 vaesenc %xmm10,%xmm9,%xmm9 2194 vmovdqu 144-128(%rdi),%xmm10 2195 xorl %edx,%r13d 2196 addl 36(%rsp),%r10d 2197 movl %r11d,%esi 2198 shrdl $11,%r14d,%r14d 2199 xorl %r9d,%r12d 2200 xorl %eax,%esi 2201 shrdl $6,%r13d,%r13d 2202 addl %r12d,%r10d 2203 andl %esi,%r15d 2204 xorl %r11d,%r14d 2205 addl %r13d,%r10d 2206 xorl %eax,%r15d 2207 addl %r10d,%ecx 2208 shrdl $2,%r14d,%r14d 2209 addl %r15d,%r10d 2210 movl %ecx,%r13d 2211 addl %r10d,%r14d 2212 shrdl $14,%r13d,%r13d 2213 movl %r14d,%r10d 2214 movl %edx,%r12d 2215 xorl %ecx,%r13d 2216 shrdl $9,%r14d,%r14d 2217 xorl %r8d,%r12d 2218 shrdl $5,%r13d,%r13d 2219 xorl %r10d,%r14d 2220 andl %ecx,%r12d 2221 vaesenc %xmm10,%xmm9,%xmm9 2222 vmovdqu 160-128(%rdi),%xmm10 2223 xorl %ecx,%r13d 2224 addl 40(%rsp),%r9d 2225 movl %r10d,%r15d 2226 shrdl $11,%r14d,%r14d 2227 xorl %r8d,%r12d 2228 xorl %r11d,%r15d 2229 shrdl $6,%r13d,%r13d 2230 addl %r12d,%r9d 2231 andl %r15d,%esi 2232 xorl %r10d,%r14d 2233 addl %r13d,%r9d 2234 xorl %r11d,%esi 2235 addl %r9d,%ebx 2236 shrdl $2,%r14d,%r14d 2237 addl %esi,%r9d 2238 movl %ebx,%r13d 2239 addl %r9d,%r14d 2240 shrdl $14,%r13d,%r13d 2241 movl %r14d,%r9d 2242 movl %ecx,%r12d 2243 xorl %ebx,%r13d 2244 shrdl $9,%r14d,%r14d 2245 xorl %edx,%r12d 2246 shrdl $5,%r13d,%r13d 2247 xorl %r9d,%r14d 2248 andl %ebx,%r12d 2249 vaesenclast %xmm10,%xmm9,%xmm11 2250 vaesenc %xmm10,%xmm9,%xmm9 2251 vmovdqu 176-128(%rdi),%xmm10 2252 xorl %ebx,%r13d 2253 addl 44(%rsp),%r8d 2254 movl %r9d,%esi 2255 shrdl $11,%r14d,%r14d 2256 xorl %edx,%r12d 2257 xorl %r10d,%esi 2258 shrdl $6,%r13d,%r13d 2259 addl %r12d,%r8d 2260 andl %esi,%r15d 2261 xorl %r9d,%r14d 2262 addl %r13d,%r8d 2263 xorl %r10d,%r15d 2264 addl %r8d,%eax 2265 shrdl $2,%r14d,%r14d 2266 addl %r15d,%r8d 2267 movl %eax,%r13d 2268 addl %r8d,%r14d 2269 shrdl $14,%r13d,%r13d 2270 movl %r14d,%r8d 2271 movl %ebx,%r12d 2272 xorl %eax,%r13d 2273 shrdl $9,%r14d,%r14d 2274 xorl %ecx,%r12d 2275 shrdl $5,%r13d,%r13d 2276 xorl %r8d,%r14d 2277 andl %eax,%r12d 2278 vpand %xmm12,%xmm11,%xmm8 2279 vaesenc %xmm10,%xmm9,%xmm9 2280 vmovdqu 192-128(%rdi),%xmm10 2281 xorl %eax,%r13d 2282 addl 48(%rsp),%edx 2283 movl %r8d,%r15d 2284 shrdl $11,%r14d,%r14d 2285 xorl %ecx,%r12d 2286 xorl %r9d,%r15d 2287 shrdl $6,%r13d,%r13d 2288 addl %r12d,%edx 2289 andl %r15d,%esi 2290 xorl %r8d,%r14d 2291 addl %r13d,%edx 2292 xorl %r9d,%esi 2293 addl %edx,%r11d 2294 shrdl $2,%r14d,%r14d 2295 addl %esi,%edx 2296 movl %r11d,%r13d 2297 addl %edx,%r14d 2298 shrdl $14,%r13d,%r13d 2299 movl %r14d,%edx 2300 movl %eax,%r12d 2301 xorl %r11d,%r13d 2302 shrdl $9,%r14d,%r14d 2303 xorl %ebx,%r12d 2304 shrdl $5,%r13d,%r13d 2305 xorl %edx,%r14d 2306 andl %r11d,%r12d 2307 vaesenclast %xmm10,%xmm9,%xmm11 2308 vaesenc %xmm10,%xmm9,%xmm9 2309 vmovdqu 208-128(%rdi),%xmm10 2310 xorl %r11d,%r13d 2311 addl 52(%rsp),%ecx 2312 movl %edx,%esi 2313 shrdl $11,%r14d,%r14d 2314 xorl %ebx,%r12d 2315 xorl %r8d,%esi 2316 shrdl $6,%r13d,%r13d 2317 addl %r12d,%ecx 2318 andl %esi,%r15d 2319 xorl %edx,%r14d 2320 addl %r13d,%ecx 2321 xorl %r8d,%r15d 2322 addl %ecx,%r10d 2323 shrdl $2,%r14d,%r14d 2324 addl %r15d,%ecx 2325 movl %r10d,%r13d 2326 addl %ecx,%r14d 2327 shrdl $14,%r13d,%r13d 2328 movl %r14d,%ecx 2329 movl %r11d,%r12d 2330 xorl %r10d,%r13d 2331 shrdl $9,%r14d,%r14d 2332 xorl %eax,%r12d 2333 shrdl $5,%r13d,%r13d 2334 xorl %ecx,%r14d 2335 andl %r10d,%r12d 2336 vpand %xmm13,%xmm11,%xmm11 2337 vaesenc %xmm10,%xmm9,%xmm9 2338 vmovdqu 224-128(%rdi),%xmm10 2339 xorl %r10d,%r13d 2340 addl 56(%rsp),%ebx 2341 movl %ecx,%r15d 2342 shrdl $11,%r14d,%r14d 2343 xorl %eax,%r12d 2344 xorl %edx,%r15d 2345 shrdl $6,%r13d,%r13d 2346 addl %r12d,%ebx 2347 andl %r15d,%esi 2348 xorl %ecx,%r14d 2349 addl %r13d,%ebx 2350 xorl %edx,%esi 2351 addl %ebx,%r9d 2352 shrdl $2,%r14d,%r14d 2353 addl %esi,%ebx 2354 movl %r9d,%r13d 2355 addl %ebx,%r14d 2356 shrdl $14,%r13d,%r13d 2357 movl %r14d,%ebx 2358 movl %r10d,%r12d 2359 xorl %r9d,%r13d 2360 shrdl $9,%r14d,%r14d 2361 xorl %r11d,%r12d 2362 shrdl $5,%r13d,%r13d 2363 xorl %ebx,%r14d 2364 andl %r9d,%r12d 2365 vpor %xmm11,%xmm8,%xmm8 2366 vaesenclast %xmm10,%xmm9,%xmm11 2367 vmovdqu 0-128(%rdi),%xmm10 2368 xorl %r9d,%r13d 2369 addl 60(%rsp),%eax 2370 movl %ebx,%esi 2371 shrdl $11,%r14d,%r14d 2372 xorl %r11d,%r12d 2373 xorl %ecx,%esi 2374 shrdl $6,%r13d,%r13d 2375 addl %r12d,%eax 2376 andl %esi,%r15d 2377 xorl %ebx,%r14d 2378 addl %r13d,%eax 2379 xorl %ecx,%r15d 2380 addl %eax,%r8d 2381 shrdl $2,%r14d,%r14d 2382 addl %r15d,%eax 2383 movl %r8d,%r13d 2384 addl %eax,%r14d 2385 movq 64+0(%rsp),%r12 2386 movq 64+8(%rsp),%r13 2387 movq 64+40(%rsp),%r15 2388 movq 64+48(%rsp),%rsi 2389 2390 vpand %xmm14,%xmm11,%xmm11 2391 movl %r14d,%eax 2392 vpor %xmm11,%xmm8,%xmm8 2393 vmovdqu %xmm8,(%r12,%r13,1) 2394 leaq 16(%r12),%r12 2395 2396 addl 0(%r15),%eax 2397 addl 4(%r15),%ebx 2398 addl 8(%r15),%ecx 2399 addl 12(%r15),%edx 2400 addl 16(%r15),%r8d 2401 addl 20(%r15),%r9d 2402 addl 24(%r15),%r10d 2403 addl 28(%r15),%r11d 2404 2405 cmpq 64+16(%rsp),%r12 2406 2407 movl %eax,0(%r15) 2408 movl %ebx,4(%r15) 2409 movl %ecx,8(%r15) 2410 movl %edx,12(%r15) 2411 movl %r8d,16(%r15) 2412 movl %r9d,20(%r15) 2413 movl %r10d,24(%r15) 2414 movl %r11d,28(%r15) 2415 jb .Lloop_avx 2416 2417 movq 64+32(%rsp),%r8 2418 movq 120(%rsp),%rsi 2419.cfi_def_cfa %rsi,8 2420 vmovdqu %xmm8,(%r8) 2421 vzeroall 2422 movq -48(%rsi),%r15 2423.cfi_restore %r15 2424 movq -40(%rsi),%r14 2425.cfi_restore %r14 2426 movq -32(%rsi),%r13 2427.cfi_restore %r13 2428 movq -24(%rsi),%r12 2429.cfi_restore %r12 2430 movq -16(%rsi),%rbp 2431.cfi_restore %rbp 2432 movq -8(%rsi),%rbx 2433.cfi_restore %rbx 2434 leaq (%rsi),%rsp 2435.cfi_def_cfa_register %rsp 2436.Lepilogue_avx: 2437 .byte 0xf3,0xc3 2438.cfi_endproc 2439.size aesni_cbc_sha256_enc_avx,.-aesni_cbc_sha256_enc_avx 2440.type aesni_cbc_sha256_enc_avx2,@function 2441.align 64 2442aesni_cbc_sha256_enc_avx2: 2443.cfi_startproc 2444.Lavx2_shortcut: 2445 movq 8(%rsp),%r10 2446 movq %rsp,%rax 2447.cfi_def_cfa_register %rax 2448 pushq %rbx 2449.cfi_offset %rbx,-16 2450 pushq %rbp 2451.cfi_offset %rbp,-24 2452 pushq %r12 2453.cfi_offset %r12,-32 2454 pushq %r13 2455.cfi_offset %r13,-40 2456 pushq %r14 2457.cfi_offset %r14,-48 2458 pushq %r15 2459.cfi_offset %r15,-56 2460 subq $576,%rsp 2461 andq $-1024,%rsp 2462 addq $448,%rsp 2463 2464 shlq $6,%rdx 2465 subq %rdi,%rsi 2466 subq %rdi,%r10 2467 addq %rdi,%rdx 2468 2469 2470 2471 movq %rdx,64+16(%rsp) 2472 2473 movq %r8,64+32(%rsp) 2474 movq %r9,64+40(%rsp) 2475 movq %r10,64+48(%rsp) 2476 movq %rax,120(%rsp) 2477.cfi_escape 0x0f,0x06,0x77,0xf8,0x00,0x06,0x23,0x08 2478.Lprologue_avx2: 2479 vzeroall 2480 2481 movq %rdi,%r13 2482 vpinsrq $1,%rsi,%xmm15,%xmm15 2483 leaq 128(%rcx),%rdi 2484 leaq K256+544(%rip),%r12 2485 movl 240-128(%rdi),%r14d 2486 movq %r9,%r15 2487 movq %r10,%rsi 2488 vmovdqu (%r8),%xmm8 2489 leaq -9(%r14),%r14 2490 2491 vmovdqa 0(%r12,%r14,8),%xmm14 2492 vmovdqa 16(%r12,%r14,8),%xmm13 2493 vmovdqa 32(%r12,%r14,8),%xmm12 2494 2495 subq $-64,%r13 2496 movl 0(%r15),%eax 2497 leaq (%rsi,%r13,1),%r12 2498 movl 4(%r15),%ebx 2499 cmpq %rdx,%r13 2500 movl 8(%r15),%ecx 2501 cmoveq %rsp,%r12 2502 movl 12(%r15),%edx 2503 movl 16(%r15),%r8d 2504 movl 20(%r15),%r9d 2505 movl 24(%r15),%r10d 2506 movl 28(%r15),%r11d 2507 vmovdqu 0-128(%rdi),%xmm10 2508 jmp .Loop_avx2 2509.align 16 2510.Loop_avx2: 2511 vmovdqa K256+512(%rip),%ymm7 2512 vmovdqu -64+0(%rsi,%r13,1),%xmm0 2513 vmovdqu -64+16(%rsi,%r13,1),%xmm1 2514 vmovdqu -64+32(%rsi,%r13,1),%xmm2 2515 vmovdqu -64+48(%rsi,%r13,1),%xmm3 2516 2517 vinserti128 $1,(%r12),%ymm0,%ymm0 2518 vinserti128 $1,16(%r12),%ymm1,%ymm1 2519 vpshufb %ymm7,%ymm0,%ymm0 2520 vinserti128 $1,32(%r12),%ymm2,%ymm2 2521 vpshufb %ymm7,%ymm1,%ymm1 2522 vinserti128 $1,48(%r12),%ymm3,%ymm3 2523 2524 leaq K256(%rip),%rbp 2525 vpshufb %ymm7,%ymm2,%ymm2 2526 leaq -64(%r13),%r13 2527 vpaddd 0(%rbp),%ymm0,%ymm4 2528 vpshufb %ymm7,%ymm3,%ymm3 2529 vpaddd 32(%rbp),%ymm1,%ymm5 2530 vpaddd 64(%rbp),%ymm2,%ymm6 2531 vpaddd 96(%rbp),%ymm3,%ymm7 2532 vmovdqa %ymm4,0(%rsp) 2533 xorl %r14d,%r14d 2534 vmovdqa %ymm5,32(%rsp) 2535 2536 movq 120(%rsp),%rsi 2537.cfi_def_cfa %rsi,8 2538 leaq -64(%rsp),%rsp 2539 2540 2541 2542 movq %rsi,-8(%rsp) 2543.cfi_escape 0x0f,0x05,0x77,0x78,0x06,0x23,0x08 2544 movl %ebx,%esi 2545 vmovdqa %ymm6,0(%rsp) 2546 xorl %ecx,%esi 2547 vmovdqa %ymm7,32(%rsp) 2548 movl %r9d,%r12d 2549 subq $-32*4,%rbp 2550 jmp .Lavx2_00_47 2551 2552.align 16 2553.Lavx2_00_47: 2554 vmovdqu (%r13),%xmm9 2555 vpinsrq $0,%r13,%xmm15,%xmm15 2556 leaq -64(%rsp),%rsp 2557.cfi_escape 0x0f,0x05,0x77,0x38,0x06,0x23,0x08 2558 2559 pushq 64-8(%rsp) 2560.cfi_escape 0x0f,0x05,0x77,0x00,0x06,0x23,0x08 2561 leaq 8(%rsp),%rsp 2562.cfi_escape 0x0f,0x05,0x77,0x78,0x06,0x23,0x08 2563 vpalignr $4,%ymm0,%ymm1,%ymm4 2564 addl 0+128(%rsp),%r11d 2565 andl %r8d,%r12d 2566 rorxl $25,%r8d,%r13d 2567 vpalignr $4,%ymm2,%ymm3,%ymm7 2568 rorxl $11,%r8d,%r15d 2569 leal (%rax,%r14,1),%eax 2570 leal (%r11,%r12,1),%r11d 2571 vpsrld $7,%ymm4,%ymm6 2572 andnl %r10d,%r8d,%r12d 2573 xorl %r15d,%r13d 2574 rorxl $6,%r8d,%r14d 2575 vpaddd %ymm7,%ymm0,%ymm0 2576 leal (%r11,%r12,1),%r11d 2577 xorl %r14d,%r13d 2578 movl %eax,%r15d 2579 vpsrld $3,%ymm4,%ymm7 2580 rorxl $22,%eax,%r12d 2581 leal (%r11,%r13,1),%r11d 2582 xorl %ebx,%r15d 2583 vpslld $14,%ymm4,%ymm5 2584 rorxl $13,%eax,%r14d 2585 rorxl $2,%eax,%r13d 2586 leal (%rdx,%r11,1),%edx 2587 vpxor %ymm6,%ymm7,%ymm4 2588 andl %r15d,%esi 2589 vpxor %xmm10,%xmm9,%xmm9 2590 vmovdqu 16-128(%rdi),%xmm10 2591 xorl %r12d,%r14d 2592 xorl %ebx,%esi 2593 vpshufd $250,%ymm3,%ymm7 2594 xorl %r13d,%r14d 2595 leal (%r11,%rsi,1),%r11d 2596 movl %r8d,%r12d 2597 vpsrld $11,%ymm6,%ymm6 2598 addl 4+128(%rsp),%r10d 2599 andl %edx,%r12d 2600 rorxl $25,%edx,%r13d 2601 vpxor %ymm5,%ymm4,%ymm4 2602 rorxl $11,%edx,%esi 2603 leal (%r11,%r14,1),%r11d 2604 leal (%r10,%r12,1),%r10d 2605 vpslld $11,%ymm5,%ymm5 2606 andnl %r9d,%edx,%r12d 2607 xorl %esi,%r13d 2608 rorxl $6,%edx,%r14d 2609 vpxor %ymm6,%ymm4,%ymm4 2610 leal (%r10,%r12,1),%r10d 2611 xorl %r14d,%r13d 2612 movl %r11d,%esi 2613 vpsrld $10,%ymm7,%ymm6 2614 rorxl $22,%r11d,%r12d 2615 leal (%r10,%r13,1),%r10d 2616 xorl %eax,%esi 2617 vpxor %ymm5,%ymm4,%ymm4 2618 rorxl $13,%r11d,%r14d 2619 rorxl $2,%r11d,%r13d 2620 leal (%rcx,%r10,1),%ecx 2621 vpsrlq $17,%ymm7,%ymm7 2622 andl %esi,%r15d 2623 vpxor %xmm8,%xmm9,%xmm9 2624 xorl %r12d,%r14d 2625 xorl %eax,%r15d 2626 vpaddd %ymm4,%ymm0,%ymm0 2627 xorl %r13d,%r14d 2628 leal (%r10,%r15,1),%r10d 2629 movl %edx,%r12d 2630 vpxor %ymm7,%ymm6,%ymm6 2631 addl 8+128(%rsp),%r9d 2632 andl %ecx,%r12d 2633 rorxl $25,%ecx,%r13d 2634 vpsrlq $2,%ymm7,%ymm7 2635 rorxl $11,%ecx,%r15d 2636 leal (%r10,%r14,1),%r10d 2637 leal (%r9,%r12,1),%r9d 2638 vpxor %ymm7,%ymm6,%ymm6 2639 andnl %r8d,%ecx,%r12d 2640 xorl %r15d,%r13d 2641 rorxl $6,%ecx,%r14d 2642 vpshufd $132,%ymm6,%ymm6 2643 leal (%r9,%r12,1),%r9d 2644 xorl %r14d,%r13d 2645 movl %r10d,%r15d 2646 vpsrldq $8,%ymm6,%ymm6 2647 rorxl $22,%r10d,%r12d 2648 leal (%r9,%r13,1),%r9d 2649 xorl %r11d,%r15d 2650 vpaddd %ymm6,%ymm0,%ymm0 2651 rorxl $13,%r10d,%r14d 2652 rorxl $2,%r10d,%r13d 2653 leal (%rbx,%r9,1),%ebx 2654 vpshufd $80,%ymm0,%ymm7 2655 andl %r15d,%esi 2656 vaesenc %xmm10,%xmm9,%xmm9 2657 vmovdqu 32-128(%rdi),%xmm10 2658 xorl %r12d,%r14d 2659 xorl %r11d,%esi 2660 vpsrld $10,%ymm7,%ymm6 2661 xorl %r13d,%r14d 2662 leal (%r9,%rsi,1),%r9d 2663 movl %ecx,%r12d 2664 vpsrlq $17,%ymm7,%ymm7 2665 addl 12+128(%rsp),%r8d 2666 andl %ebx,%r12d 2667 rorxl $25,%ebx,%r13d 2668 vpxor %ymm7,%ymm6,%ymm6 2669 rorxl $11,%ebx,%esi 2670 leal (%r9,%r14,1),%r9d 2671 leal (%r8,%r12,1),%r8d 2672 vpsrlq $2,%ymm7,%ymm7 2673 andnl %edx,%ebx,%r12d 2674 xorl %esi,%r13d 2675 rorxl $6,%ebx,%r14d 2676 vpxor %ymm7,%ymm6,%ymm6 2677 leal (%r8,%r12,1),%r8d 2678 xorl %r14d,%r13d 2679 movl %r9d,%esi 2680 vpshufd $232,%ymm6,%ymm6 2681 rorxl $22,%r9d,%r12d 2682 leal (%r8,%r13,1),%r8d 2683 xorl %r10d,%esi 2684 vpslldq $8,%ymm6,%ymm6 2685 rorxl $13,%r9d,%r14d 2686 rorxl $2,%r9d,%r13d 2687 leal (%rax,%r8,1),%eax 2688 vpaddd %ymm6,%ymm0,%ymm0 2689 andl %esi,%r15d 2690 vaesenc %xmm10,%xmm9,%xmm9 2691 vmovdqu 48-128(%rdi),%xmm10 2692 xorl %r12d,%r14d 2693 xorl %r10d,%r15d 2694 vpaddd 0(%rbp),%ymm0,%ymm6 2695 xorl %r13d,%r14d 2696 leal (%r8,%r15,1),%r8d 2697 movl %ebx,%r12d 2698 vmovdqa %ymm6,0(%rsp) 2699 vpalignr $4,%ymm1,%ymm2,%ymm4 2700 addl 32+128(%rsp),%edx 2701 andl %eax,%r12d 2702 rorxl $25,%eax,%r13d 2703 vpalignr $4,%ymm3,%ymm0,%ymm7 2704 rorxl $11,%eax,%r15d 2705 leal (%r8,%r14,1),%r8d 2706 leal (%rdx,%r12,1),%edx 2707 vpsrld $7,%ymm4,%ymm6 2708 andnl %ecx,%eax,%r12d 2709 xorl %r15d,%r13d 2710 rorxl $6,%eax,%r14d 2711 vpaddd %ymm7,%ymm1,%ymm1 2712 leal (%rdx,%r12,1),%edx 2713 xorl %r14d,%r13d 2714 movl %r8d,%r15d 2715 vpsrld $3,%ymm4,%ymm7 2716 rorxl $22,%r8d,%r12d 2717 leal (%rdx,%r13,1),%edx 2718 xorl %r9d,%r15d 2719 vpslld $14,%ymm4,%ymm5 2720 rorxl $13,%r8d,%r14d 2721 rorxl $2,%r8d,%r13d 2722 leal (%r11,%rdx,1),%r11d 2723 vpxor %ymm6,%ymm7,%ymm4 2724 andl %r15d,%esi 2725 vaesenc %xmm10,%xmm9,%xmm9 2726 vmovdqu 64-128(%rdi),%xmm10 2727 xorl %r12d,%r14d 2728 xorl %r9d,%esi 2729 vpshufd $250,%ymm0,%ymm7 2730 xorl %r13d,%r14d 2731 leal (%rdx,%rsi,1),%edx 2732 movl %eax,%r12d 2733 vpsrld $11,%ymm6,%ymm6 2734 addl 36+128(%rsp),%ecx 2735 andl %r11d,%r12d 2736 rorxl $25,%r11d,%r13d 2737 vpxor %ymm5,%ymm4,%ymm4 2738 rorxl $11,%r11d,%esi 2739 leal (%rdx,%r14,1),%edx 2740 leal (%rcx,%r12,1),%ecx 2741 vpslld $11,%ymm5,%ymm5 2742 andnl %ebx,%r11d,%r12d 2743 xorl %esi,%r13d 2744 rorxl $6,%r11d,%r14d 2745 vpxor %ymm6,%ymm4,%ymm4 2746 leal (%rcx,%r12,1),%ecx 2747 xorl %r14d,%r13d 2748 movl %edx,%esi 2749 vpsrld $10,%ymm7,%ymm6 2750 rorxl $22,%edx,%r12d 2751 leal (%rcx,%r13,1),%ecx 2752 xorl %r8d,%esi 2753 vpxor %ymm5,%ymm4,%ymm4 2754 rorxl $13,%edx,%r14d 2755 rorxl $2,%edx,%r13d 2756 leal (%r10,%rcx,1),%r10d 2757 vpsrlq $17,%ymm7,%ymm7 2758 andl %esi,%r15d 2759 vaesenc %xmm10,%xmm9,%xmm9 2760 vmovdqu 80-128(%rdi),%xmm10 2761 xorl %r12d,%r14d 2762 xorl %r8d,%r15d 2763 vpaddd %ymm4,%ymm1,%ymm1 2764 xorl %r13d,%r14d 2765 leal (%rcx,%r15,1),%ecx 2766 movl %r11d,%r12d 2767 vpxor %ymm7,%ymm6,%ymm6 2768 addl 40+128(%rsp),%ebx 2769 andl %r10d,%r12d 2770 rorxl $25,%r10d,%r13d 2771 vpsrlq $2,%ymm7,%ymm7 2772 rorxl $11,%r10d,%r15d 2773 leal (%rcx,%r14,1),%ecx 2774 leal (%rbx,%r12,1),%ebx 2775 vpxor %ymm7,%ymm6,%ymm6 2776 andnl %eax,%r10d,%r12d 2777 xorl %r15d,%r13d 2778 rorxl $6,%r10d,%r14d 2779 vpshufd $132,%ymm6,%ymm6 2780 leal (%rbx,%r12,1),%ebx 2781 xorl %r14d,%r13d 2782 movl %ecx,%r15d 2783 vpsrldq $8,%ymm6,%ymm6 2784 rorxl $22,%ecx,%r12d 2785 leal (%rbx,%r13,1),%ebx 2786 xorl %edx,%r15d 2787 vpaddd %ymm6,%ymm1,%ymm1 2788 rorxl $13,%ecx,%r14d 2789 rorxl $2,%ecx,%r13d 2790 leal (%r9,%rbx,1),%r9d 2791 vpshufd $80,%ymm1,%ymm7 2792 andl %r15d,%esi 2793 vaesenc %xmm10,%xmm9,%xmm9 2794 vmovdqu 96-128(%rdi),%xmm10 2795 xorl %r12d,%r14d 2796 xorl %edx,%esi 2797 vpsrld $10,%ymm7,%ymm6 2798 xorl %r13d,%r14d 2799 leal (%rbx,%rsi,1),%ebx 2800 movl %r10d,%r12d 2801 vpsrlq $17,%ymm7,%ymm7 2802 addl 44+128(%rsp),%eax 2803 andl %r9d,%r12d 2804 rorxl $25,%r9d,%r13d 2805 vpxor %ymm7,%ymm6,%ymm6 2806 rorxl $11,%r9d,%esi 2807 leal (%rbx,%r14,1),%ebx 2808 leal (%rax,%r12,1),%eax 2809 vpsrlq $2,%ymm7,%ymm7 2810 andnl %r11d,%r9d,%r12d 2811 xorl %esi,%r13d 2812 rorxl $6,%r9d,%r14d 2813 vpxor %ymm7,%ymm6,%ymm6 2814 leal (%rax,%r12,1),%eax 2815 xorl %r14d,%r13d 2816 movl %ebx,%esi 2817 vpshufd $232,%ymm6,%ymm6 2818 rorxl $22,%ebx,%r12d 2819 leal (%rax,%r13,1),%eax 2820 xorl %ecx,%esi 2821 vpslldq $8,%ymm6,%ymm6 2822 rorxl $13,%ebx,%r14d 2823 rorxl $2,%ebx,%r13d 2824 leal (%r8,%rax,1),%r8d 2825 vpaddd %ymm6,%ymm1,%ymm1 2826 andl %esi,%r15d 2827 vaesenc %xmm10,%xmm9,%xmm9 2828 vmovdqu 112-128(%rdi),%xmm10 2829 xorl %r12d,%r14d 2830 xorl %ecx,%r15d 2831 vpaddd 32(%rbp),%ymm1,%ymm6 2832 xorl %r13d,%r14d 2833 leal (%rax,%r15,1),%eax 2834 movl %r9d,%r12d 2835 vmovdqa %ymm6,32(%rsp) 2836 leaq -64(%rsp),%rsp 2837.cfi_escape 0x0f,0x05,0x77,0x38,0x06,0x23,0x08 2838 2839 pushq 64-8(%rsp) 2840.cfi_escape 0x0f,0x05,0x77,0x00,0x06,0x23,0x08 2841 leaq 8(%rsp),%rsp 2842.cfi_escape 0x0f,0x05,0x77,0x78,0x06,0x23,0x08 2843 vpalignr $4,%ymm2,%ymm3,%ymm4 2844 addl 0+128(%rsp),%r11d 2845 andl %r8d,%r12d 2846 rorxl $25,%r8d,%r13d 2847 vpalignr $4,%ymm0,%ymm1,%ymm7 2848 rorxl $11,%r8d,%r15d 2849 leal (%rax,%r14,1),%eax 2850 leal (%r11,%r12,1),%r11d 2851 vpsrld $7,%ymm4,%ymm6 2852 andnl %r10d,%r8d,%r12d 2853 xorl %r15d,%r13d 2854 rorxl $6,%r8d,%r14d 2855 vpaddd %ymm7,%ymm2,%ymm2 2856 leal (%r11,%r12,1),%r11d 2857 xorl %r14d,%r13d 2858 movl %eax,%r15d 2859 vpsrld $3,%ymm4,%ymm7 2860 rorxl $22,%eax,%r12d 2861 leal (%r11,%r13,1),%r11d 2862 xorl %ebx,%r15d 2863 vpslld $14,%ymm4,%ymm5 2864 rorxl $13,%eax,%r14d 2865 rorxl $2,%eax,%r13d 2866 leal (%rdx,%r11,1),%edx 2867 vpxor %ymm6,%ymm7,%ymm4 2868 andl %r15d,%esi 2869 vaesenc %xmm10,%xmm9,%xmm9 2870 vmovdqu 128-128(%rdi),%xmm10 2871 xorl %r12d,%r14d 2872 xorl %ebx,%esi 2873 vpshufd $250,%ymm1,%ymm7 2874 xorl %r13d,%r14d 2875 leal (%r11,%rsi,1),%r11d 2876 movl %r8d,%r12d 2877 vpsrld $11,%ymm6,%ymm6 2878 addl 4+128(%rsp),%r10d 2879 andl %edx,%r12d 2880 rorxl $25,%edx,%r13d 2881 vpxor %ymm5,%ymm4,%ymm4 2882 rorxl $11,%edx,%esi 2883 leal (%r11,%r14,1),%r11d 2884 leal (%r10,%r12,1),%r10d 2885 vpslld $11,%ymm5,%ymm5 2886 andnl %r9d,%edx,%r12d 2887 xorl %esi,%r13d 2888 rorxl $6,%edx,%r14d 2889 vpxor %ymm6,%ymm4,%ymm4 2890 leal (%r10,%r12,1),%r10d 2891 xorl %r14d,%r13d 2892 movl %r11d,%esi 2893 vpsrld $10,%ymm7,%ymm6 2894 rorxl $22,%r11d,%r12d 2895 leal (%r10,%r13,1),%r10d 2896 xorl %eax,%esi 2897 vpxor %ymm5,%ymm4,%ymm4 2898 rorxl $13,%r11d,%r14d 2899 rorxl $2,%r11d,%r13d 2900 leal (%rcx,%r10,1),%ecx 2901 vpsrlq $17,%ymm7,%ymm7 2902 andl %esi,%r15d 2903 vaesenc %xmm10,%xmm9,%xmm9 2904 vmovdqu 144-128(%rdi),%xmm10 2905 xorl %r12d,%r14d 2906 xorl %eax,%r15d 2907 vpaddd %ymm4,%ymm2,%ymm2 2908 xorl %r13d,%r14d 2909 leal (%r10,%r15,1),%r10d 2910 movl %edx,%r12d 2911 vpxor %ymm7,%ymm6,%ymm6 2912 addl 8+128(%rsp),%r9d 2913 andl %ecx,%r12d 2914 rorxl $25,%ecx,%r13d 2915 vpsrlq $2,%ymm7,%ymm7 2916 rorxl $11,%ecx,%r15d 2917 leal (%r10,%r14,1),%r10d 2918 leal (%r9,%r12,1),%r9d 2919 vpxor %ymm7,%ymm6,%ymm6 2920 andnl %r8d,%ecx,%r12d 2921 xorl %r15d,%r13d 2922 rorxl $6,%ecx,%r14d 2923 vpshufd $132,%ymm6,%ymm6 2924 leal (%r9,%r12,1),%r9d 2925 xorl %r14d,%r13d 2926 movl %r10d,%r15d 2927 vpsrldq $8,%ymm6,%ymm6 2928 rorxl $22,%r10d,%r12d 2929 leal (%r9,%r13,1),%r9d 2930 xorl %r11d,%r15d 2931 vpaddd %ymm6,%ymm2,%ymm2 2932 rorxl $13,%r10d,%r14d 2933 rorxl $2,%r10d,%r13d 2934 leal (%rbx,%r9,1),%ebx 2935 vpshufd $80,%ymm2,%ymm7 2936 andl %r15d,%esi 2937 vaesenc %xmm10,%xmm9,%xmm9 2938 vmovdqu 160-128(%rdi),%xmm10 2939 xorl %r12d,%r14d 2940 xorl %r11d,%esi 2941 vpsrld $10,%ymm7,%ymm6 2942 xorl %r13d,%r14d 2943 leal (%r9,%rsi,1),%r9d 2944 movl %ecx,%r12d 2945 vpsrlq $17,%ymm7,%ymm7 2946 addl 12+128(%rsp),%r8d 2947 andl %ebx,%r12d 2948 rorxl $25,%ebx,%r13d 2949 vpxor %ymm7,%ymm6,%ymm6 2950 rorxl $11,%ebx,%esi 2951 leal (%r9,%r14,1),%r9d 2952 leal (%r8,%r12,1),%r8d 2953 vpsrlq $2,%ymm7,%ymm7 2954 andnl %edx,%ebx,%r12d 2955 xorl %esi,%r13d 2956 rorxl $6,%ebx,%r14d 2957 vpxor %ymm7,%ymm6,%ymm6 2958 leal (%r8,%r12,1),%r8d 2959 xorl %r14d,%r13d 2960 movl %r9d,%esi 2961 vpshufd $232,%ymm6,%ymm6 2962 rorxl $22,%r9d,%r12d 2963 leal (%r8,%r13,1),%r8d 2964 xorl %r10d,%esi 2965 vpslldq $8,%ymm6,%ymm6 2966 rorxl $13,%r9d,%r14d 2967 rorxl $2,%r9d,%r13d 2968 leal (%rax,%r8,1),%eax 2969 vpaddd %ymm6,%ymm2,%ymm2 2970 andl %esi,%r15d 2971 vaesenclast %xmm10,%xmm9,%xmm11 2972 vaesenc %xmm10,%xmm9,%xmm9 2973 vmovdqu 176-128(%rdi),%xmm10 2974 xorl %r12d,%r14d 2975 xorl %r10d,%r15d 2976 vpaddd 64(%rbp),%ymm2,%ymm6 2977 xorl %r13d,%r14d 2978 leal (%r8,%r15,1),%r8d 2979 movl %ebx,%r12d 2980 vmovdqa %ymm6,0(%rsp) 2981 vpalignr $4,%ymm3,%ymm0,%ymm4 2982 addl 32+128(%rsp),%edx 2983 andl %eax,%r12d 2984 rorxl $25,%eax,%r13d 2985 vpalignr $4,%ymm1,%ymm2,%ymm7 2986 rorxl $11,%eax,%r15d 2987 leal (%r8,%r14,1),%r8d 2988 leal (%rdx,%r12,1),%edx 2989 vpsrld $7,%ymm4,%ymm6 2990 andnl %ecx,%eax,%r12d 2991 xorl %r15d,%r13d 2992 rorxl $6,%eax,%r14d 2993 vpaddd %ymm7,%ymm3,%ymm3 2994 leal (%rdx,%r12,1),%edx 2995 xorl %r14d,%r13d 2996 movl %r8d,%r15d 2997 vpsrld $3,%ymm4,%ymm7 2998 rorxl $22,%r8d,%r12d 2999 leal (%rdx,%r13,1),%edx 3000 xorl %r9d,%r15d 3001 vpslld $14,%ymm4,%ymm5 3002 rorxl $13,%r8d,%r14d 3003 rorxl $2,%r8d,%r13d 3004 leal (%r11,%rdx,1),%r11d 3005 vpxor %ymm6,%ymm7,%ymm4 3006 andl %r15d,%esi 3007 vpand %xmm12,%xmm11,%xmm8 3008 vaesenc %xmm10,%xmm9,%xmm9 3009 vmovdqu 192-128(%rdi),%xmm10 3010 xorl %r12d,%r14d 3011 xorl %r9d,%esi 3012 vpshufd $250,%ymm2,%ymm7 3013 xorl %r13d,%r14d 3014 leal (%rdx,%rsi,1),%edx 3015 movl %eax,%r12d 3016 vpsrld $11,%ymm6,%ymm6 3017 addl 36+128(%rsp),%ecx 3018 andl %r11d,%r12d 3019 rorxl $25,%r11d,%r13d 3020 vpxor %ymm5,%ymm4,%ymm4 3021 rorxl $11,%r11d,%esi 3022 leal (%rdx,%r14,1),%edx 3023 leal (%rcx,%r12,1),%ecx 3024 vpslld $11,%ymm5,%ymm5 3025 andnl %ebx,%r11d,%r12d 3026 xorl %esi,%r13d 3027 rorxl $6,%r11d,%r14d 3028 vpxor %ymm6,%ymm4,%ymm4 3029 leal (%rcx,%r12,1),%ecx 3030 xorl %r14d,%r13d 3031 movl %edx,%esi 3032 vpsrld $10,%ymm7,%ymm6 3033 rorxl $22,%edx,%r12d 3034 leal (%rcx,%r13,1),%ecx 3035 xorl %r8d,%esi 3036 vpxor %ymm5,%ymm4,%ymm4 3037 rorxl $13,%edx,%r14d 3038 rorxl $2,%edx,%r13d 3039 leal (%r10,%rcx,1),%r10d 3040 vpsrlq $17,%ymm7,%ymm7 3041 andl %esi,%r15d 3042 vaesenclast %xmm10,%xmm9,%xmm11 3043 vaesenc %xmm10,%xmm9,%xmm9 3044 vmovdqu 208-128(%rdi),%xmm10 3045 xorl %r12d,%r14d 3046 xorl %r8d,%r15d 3047 vpaddd %ymm4,%ymm3,%ymm3 3048 xorl %r13d,%r14d 3049 leal (%rcx,%r15,1),%ecx 3050 movl %r11d,%r12d 3051 vpxor %ymm7,%ymm6,%ymm6 3052 addl 40+128(%rsp),%ebx 3053 andl %r10d,%r12d 3054 rorxl $25,%r10d,%r13d 3055 vpsrlq $2,%ymm7,%ymm7 3056 rorxl $11,%r10d,%r15d 3057 leal (%rcx,%r14,1),%ecx 3058 leal (%rbx,%r12,1),%ebx 3059 vpxor %ymm7,%ymm6,%ymm6 3060 andnl %eax,%r10d,%r12d 3061 xorl %r15d,%r13d 3062 rorxl $6,%r10d,%r14d 3063 vpshufd $132,%ymm6,%ymm6 3064 leal (%rbx,%r12,1),%ebx 3065 xorl %r14d,%r13d 3066 movl %ecx,%r15d 3067 vpsrldq $8,%ymm6,%ymm6 3068 rorxl $22,%ecx,%r12d 3069 leal (%rbx,%r13,1),%ebx 3070 xorl %edx,%r15d 3071 vpaddd %ymm6,%ymm3,%ymm3 3072 rorxl $13,%ecx,%r14d 3073 rorxl $2,%ecx,%r13d 3074 leal (%r9,%rbx,1),%r9d 3075 vpshufd $80,%ymm3,%ymm7 3076 andl %r15d,%esi 3077 vpand %xmm13,%xmm11,%xmm11 3078 vaesenc %xmm10,%xmm9,%xmm9 3079 vmovdqu 224-128(%rdi),%xmm10 3080 xorl %r12d,%r14d 3081 xorl %edx,%esi 3082 vpsrld $10,%ymm7,%ymm6 3083 xorl %r13d,%r14d 3084 leal (%rbx,%rsi,1),%ebx 3085 movl %r10d,%r12d 3086 vpsrlq $17,%ymm7,%ymm7 3087 addl 44+128(%rsp),%eax 3088 andl %r9d,%r12d 3089 rorxl $25,%r9d,%r13d 3090 vpxor %ymm7,%ymm6,%ymm6 3091 rorxl $11,%r9d,%esi 3092 leal (%rbx,%r14,1),%ebx 3093 leal (%rax,%r12,1),%eax 3094 vpsrlq $2,%ymm7,%ymm7 3095 andnl %r11d,%r9d,%r12d 3096 xorl %esi,%r13d 3097 rorxl $6,%r9d,%r14d 3098 vpxor %ymm7,%ymm6,%ymm6 3099 leal (%rax,%r12,1),%eax 3100 xorl %r14d,%r13d 3101 movl %ebx,%esi 3102 vpshufd $232,%ymm6,%ymm6 3103 rorxl $22,%ebx,%r12d 3104 leal (%rax,%r13,1),%eax 3105 xorl %ecx,%esi 3106 vpslldq $8,%ymm6,%ymm6 3107 rorxl $13,%ebx,%r14d 3108 rorxl $2,%ebx,%r13d 3109 leal (%r8,%rax,1),%r8d 3110 vpaddd %ymm6,%ymm3,%ymm3 3111 andl %esi,%r15d 3112 vpor %xmm11,%xmm8,%xmm8 3113 vaesenclast %xmm10,%xmm9,%xmm11 3114 vmovdqu 0-128(%rdi),%xmm10 3115 xorl %r12d,%r14d 3116 xorl %ecx,%r15d 3117 vpaddd 96(%rbp),%ymm3,%ymm6 3118 xorl %r13d,%r14d 3119 leal (%rax,%r15,1),%eax 3120 movl %r9d,%r12d 3121 vmovdqa %ymm6,32(%rsp) 3122 vmovq %xmm15,%r13 3123 vpextrq $1,%xmm15,%r15 3124 vpand %xmm14,%xmm11,%xmm11 3125 vpor %xmm11,%xmm8,%xmm8 3126 vmovdqu %xmm8,(%r15,%r13,1) 3127 leaq 16(%r13),%r13 3128 leaq 128(%rbp),%rbp 3129 cmpb $0,3(%rbp) 3130 jne .Lavx2_00_47 3131 vmovdqu (%r13),%xmm9 3132 vpinsrq $0,%r13,%xmm15,%xmm15 3133 addl 0+64(%rsp),%r11d 3134 andl %r8d,%r12d 3135 rorxl $25,%r8d,%r13d 3136 rorxl $11,%r8d,%r15d 3137 leal (%rax,%r14,1),%eax 3138 leal (%r11,%r12,1),%r11d 3139 andnl %r10d,%r8d,%r12d 3140 xorl %r15d,%r13d 3141 rorxl $6,%r8d,%r14d 3142 leal (%r11,%r12,1),%r11d 3143 xorl %r14d,%r13d 3144 movl %eax,%r15d 3145 rorxl $22,%eax,%r12d 3146 leal (%r11,%r13,1),%r11d 3147 xorl %ebx,%r15d 3148 rorxl $13,%eax,%r14d 3149 rorxl $2,%eax,%r13d 3150 leal (%rdx,%r11,1),%edx 3151 andl %r15d,%esi 3152 vpxor %xmm10,%xmm9,%xmm9 3153 vmovdqu 16-128(%rdi),%xmm10 3154 xorl %r12d,%r14d 3155 xorl %ebx,%esi 3156 xorl %r13d,%r14d 3157 leal (%r11,%rsi,1),%r11d 3158 movl %r8d,%r12d 3159 addl 4+64(%rsp),%r10d 3160 andl %edx,%r12d 3161 rorxl $25,%edx,%r13d 3162 rorxl $11,%edx,%esi 3163 leal (%r11,%r14,1),%r11d 3164 leal (%r10,%r12,1),%r10d 3165 andnl %r9d,%edx,%r12d 3166 xorl %esi,%r13d 3167 rorxl $6,%edx,%r14d 3168 leal (%r10,%r12,1),%r10d 3169 xorl %r14d,%r13d 3170 movl %r11d,%esi 3171 rorxl $22,%r11d,%r12d 3172 leal (%r10,%r13,1),%r10d 3173 xorl %eax,%esi 3174 rorxl $13,%r11d,%r14d 3175 rorxl $2,%r11d,%r13d 3176 leal (%rcx,%r10,1),%ecx 3177 andl %esi,%r15d 3178 vpxor %xmm8,%xmm9,%xmm9 3179 xorl %r12d,%r14d 3180 xorl %eax,%r15d 3181 xorl %r13d,%r14d 3182 leal (%r10,%r15,1),%r10d 3183 movl %edx,%r12d 3184 addl 8+64(%rsp),%r9d 3185 andl %ecx,%r12d 3186 rorxl $25,%ecx,%r13d 3187 rorxl $11,%ecx,%r15d 3188 leal (%r10,%r14,1),%r10d 3189 leal (%r9,%r12,1),%r9d 3190 andnl %r8d,%ecx,%r12d 3191 xorl %r15d,%r13d 3192 rorxl $6,%ecx,%r14d 3193 leal (%r9,%r12,1),%r9d 3194 xorl %r14d,%r13d 3195 movl %r10d,%r15d 3196 rorxl $22,%r10d,%r12d 3197 leal (%r9,%r13,1),%r9d 3198 xorl %r11d,%r15d 3199 rorxl $13,%r10d,%r14d 3200 rorxl $2,%r10d,%r13d 3201 leal (%rbx,%r9,1),%ebx 3202 andl %r15d,%esi 3203 vaesenc %xmm10,%xmm9,%xmm9 3204 vmovdqu 32-128(%rdi),%xmm10 3205 xorl %r12d,%r14d 3206 xorl %r11d,%esi 3207 xorl %r13d,%r14d 3208 leal (%r9,%rsi,1),%r9d 3209 movl %ecx,%r12d 3210 addl 12+64(%rsp),%r8d 3211 andl %ebx,%r12d 3212 rorxl $25,%ebx,%r13d 3213 rorxl $11,%ebx,%esi 3214 leal (%r9,%r14,1),%r9d 3215 leal (%r8,%r12,1),%r8d 3216 andnl %edx,%ebx,%r12d 3217 xorl %esi,%r13d 3218 rorxl $6,%ebx,%r14d 3219 leal (%r8,%r12,1),%r8d 3220 xorl %r14d,%r13d 3221 movl %r9d,%esi 3222 rorxl $22,%r9d,%r12d 3223 leal (%r8,%r13,1),%r8d 3224 xorl %r10d,%esi 3225 rorxl $13,%r9d,%r14d 3226 rorxl $2,%r9d,%r13d 3227 leal (%rax,%r8,1),%eax 3228 andl %esi,%r15d 3229 vaesenc %xmm10,%xmm9,%xmm9 3230 vmovdqu 48-128(%rdi),%xmm10 3231 xorl %r12d,%r14d 3232 xorl %r10d,%r15d 3233 xorl %r13d,%r14d 3234 leal (%r8,%r15,1),%r8d 3235 movl %ebx,%r12d 3236 addl 32+64(%rsp),%edx 3237 andl %eax,%r12d 3238 rorxl $25,%eax,%r13d 3239 rorxl $11,%eax,%r15d 3240 leal (%r8,%r14,1),%r8d 3241 leal (%rdx,%r12,1),%edx 3242 andnl %ecx,%eax,%r12d 3243 xorl %r15d,%r13d 3244 rorxl $6,%eax,%r14d 3245 leal (%rdx,%r12,1),%edx 3246 xorl %r14d,%r13d 3247 movl %r8d,%r15d 3248 rorxl $22,%r8d,%r12d 3249 leal (%rdx,%r13,1),%edx 3250 xorl %r9d,%r15d 3251 rorxl $13,%r8d,%r14d 3252 rorxl $2,%r8d,%r13d 3253 leal (%r11,%rdx,1),%r11d 3254 andl %r15d,%esi 3255 vaesenc %xmm10,%xmm9,%xmm9 3256 vmovdqu 64-128(%rdi),%xmm10 3257 xorl %r12d,%r14d 3258 xorl %r9d,%esi 3259 xorl %r13d,%r14d 3260 leal (%rdx,%rsi,1),%edx 3261 movl %eax,%r12d 3262 addl 36+64(%rsp),%ecx 3263 andl %r11d,%r12d 3264 rorxl $25,%r11d,%r13d 3265 rorxl $11,%r11d,%esi 3266 leal (%rdx,%r14,1),%edx 3267 leal (%rcx,%r12,1),%ecx 3268 andnl %ebx,%r11d,%r12d 3269 xorl %esi,%r13d 3270 rorxl $6,%r11d,%r14d 3271 leal (%rcx,%r12,1),%ecx 3272 xorl %r14d,%r13d 3273 movl %edx,%esi 3274 rorxl $22,%edx,%r12d 3275 leal (%rcx,%r13,1),%ecx 3276 xorl %r8d,%esi 3277 rorxl $13,%edx,%r14d 3278 rorxl $2,%edx,%r13d 3279 leal (%r10,%rcx,1),%r10d 3280 andl %esi,%r15d 3281 vaesenc %xmm10,%xmm9,%xmm9 3282 vmovdqu 80-128(%rdi),%xmm10 3283 xorl %r12d,%r14d 3284 xorl %r8d,%r15d 3285 xorl %r13d,%r14d 3286 leal (%rcx,%r15,1),%ecx 3287 movl %r11d,%r12d 3288 addl 40+64(%rsp),%ebx 3289 andl %r10d,%r12d 3290 rorxl $25,%r10d,%r13d 3291 rorxl $11,%r10d,%r15d 3292 leal (%rcx,%r14,1),%ecx 3293 leal (%rbx,%r12,1),%ebx 3294 andnl %eax,%r10d,%r12d 3295 xorl %r15d,%r13d 3296 rorxl $6,%r10d,%r14d 3297 leal (%rbx,%r12,1),%ebx 3298 xorl %r14d,%r13d 3299 movl %ecx,%r15d 3300 rorxl $22,%ecx,%r12d 3301 leal (%rbx,%r13,1),%ebx 3302 xorl %edx,%r15d 3303 rorxl $13,%ecx,%r14d 3304 rorxl $2,%ecx,%r13d 3305 leal (%r9,%rbx,1),%r9d 3306 andl %r15d,%esi 3307 vaesenc %xmm10,%xmm9,%xmm9 3308 vmovdqu 96-128(%rdi),%xmm10 3309 xorl %r12d,%r14d 3310 xorl %edx,%esi 3311 xorl %r13d,%r14d 3312 leal (%rbx,%rsi,1),%ebx 3313 movl %r10d,%r12d 3314 addl 44+64(%rsp),%eax 3315 andl %r9d,%r12d 3316 rorxl $25,%r9d,%r13d 3317 rorxl $11,%r9d,%esi 3318 leal (%rbx,%r14,1),%ebx 3319 leal (%rax,%r12,1),%eax 3320 andnl %r11d,%r9d,%r12d 3321 xorl %esi,%r13d 3322 rorxl $6,%r9d,%r14d 3323 leal (%rax,%r12,1),%eax 3324 xorl %r14d,%r13d 3325 movl %ebx,%esi 3326 rorxl $22,%ebx,%r12d 3327 leal (%rax,%r13,1),%eax 3328 xorl %ecx,%esi 3329 rorxl $13,%ebx,%r14d 3330 rorxl $2,%ebx,%r13d 3331 leal (%r8,%rax,1),%r8d 3332 andl %esi,%r15d 3333 vaesenc %xmm10,%xmm9,%xmm9 3334 vmovdqu 112-128(%rdi),%xmm10 3335 xorl %r12d,%r14d 3336 xorl %ecx,%r15d 3337 xorl %r13d,%r14d 3338 leal (%rax,%r15,1),%eax 3339 movl %r9d,%r12d 3340 addl 0(%rsp),%r11d 3341 andl %r8d,%r12d 3342 rorxl $25,%r8d,%r13d 3343 rorxl $11,%r8d,%r15d 3344 leal (%rax,%r14,1),%eax 3345 leal (%r11,%r12,1),%r11d 3346 andnl %r10d,%r8d,%r12d 3347 xorl %r15d,%r13d 3348 rorxl $6,%r8d,%r14d 3349 leal (%r11,%r12,1),%r11d 3350 xorl %r14d,%r13d 3351 movl %eax,%r15d 3352 rorxl $22,%eax,%r12d 3353 leal (%r11,%r13,1),%r11d 3354 xorl %ebx,%r15d 3355 rorxl $13,%eax,%r14d 3356 rorxl $2,%eax,%r13d 3357 leal (%rdx,%r11,1),%edx 3358 andl %r15d,%esi 3359 vaesenc %xmm10,%xmm9,%xmm9 3360 vmovdqu 128-128(%rdi),%xmm10 3361 xorl %r12d,%r14d 3362 xorl %ebx,%esi 3363 xorl %r13d,%r14d 3364 leal (%r11,%rsi,1),%r11d 3365 movl %r8d,%r12d 3366 addl 4(%rsp),%r10d 3367 andl %edx,%r12d 3368 rorxl $25,%edx,%r13d 3369 rorxl $11,%edx,%esi 3370 leal (%r11,%r14,1),%r11d 3371 leal (%r10,%r12,1),%r10d 3372 andnl %r9d,%edx,%r12d 3373 xorl %esi,%r13d 3374 rorxl $6,%edx,%r14d 3375 leal (%r10,%r12,1),%r10d 3376 xorl %r14d,%r13d 3377 movl %r11d,%esi 3378 rorxl $22,%r11d,%r12d 3379 leal (%r10,%r13,1),%r10d 3380 xorl %eax,%esi 3381 rorxl $13,%r11d,%r14d 3382 rorxl $2,%r11d,%r13d 3383 leal (%rcx,%r10,1),%ecx 3384 andl %esi,%r15d 3385 vaesenc %xmm10,%xmm9,%xmm9 3386 vmovdqu 144-128(%rdi),%xmm10 3387 xorl %r12d,%r14d 3388 xorl %eax,%r15d 3389 xorl %r13d,%r14d 3390 leal (%r10,%r15,1),%r10d 3391 movl %edx,%r12d 3392 addl 8(%rsp),%r9d 3393 andl %ecx,%r12d 3394 rorxl $25,%ecx,%r13d 3395 rorxl $11,%ecx,%r15d 3396 leal (%r10,%r14,1),%r10d 3397 leal (%r9,%r12,1),%r9d 3398 andnl %r8d,%ecx,%r12d 3399 xorl %r15d,%r13d 3400 rorxl $6,%ecx,%r14d 3401 leal (%r9,%r12,1),%r9d 3402 xorl %r14d,%r13d 3403 movl %r10d,%r15d 3404 rorxl $22,%r10d,%r12d 3405 leal (%r9,%r13,1),%r9d 3406 xorl %r11d,%r15d 3407 rorxl $13,%r10d,%r14d 3408 rorxl $2,%r10d,%r13d 3409 leal (%rbx,%r9,1),%ebx 3410 andl %r15d,%esi 3411 vaesenc %xmm10,%xmm9,%xmm9 3412 vmovdqu 160-128(%rdi),%xmm10 3413 xorl %r12d,%r14d 3414 xorl %r11d,%esi 3415 xorl %r13d,%r14d 3416 leal (%r9,%rsi,1),%r9d 3417 movl %ecx,%r12d 3418 addl 12(%rsp),%r8d 3419 andl %ebx,%r12d 3420 rorxl $25,%ebx,%r13d 3421 rorxl $11,%ebx,%esi 3422 leal (%r9,%r14,1),%r9d 3423 leal (%r8,%r12,1),%r8d 3424 andnl %edx,%ebx,%r12d 3425 xorl %esi,%r13d 3426 rorxl $6,%ebx,%r14d 3427 leal (%r8,%r12,1),%r8d 3428 xorl %r14d,%r13d 3429 movl %r9d,%esi 3430 rorxl $22,%r9d,%r12d 3431 leal (%r8,%r13,1),%r8d 3432 xorl %r10d,%esi 3433 rorxl $13,%r9d,%r14d 3434 rorxl $2,%r9d,%r13d 3435 leal (%rax,%r8,1),%eax 3436 andl %esi,%r15d 3437 vaesenclast %xmm10,%xmm9,%xmm11 3438 vaesenc %xmm10,%xmm9,%xmm9 3439 vmovdqu 176-128(%rdi),%xmm10 3440 xorl %r12d,%r14d 3441 xorl %r10d,%r15d 3442 xorl %r13d,%r14d 3443 leal (%r8,%r15,1),%r8d 3444 movl %ebx,%r12d 3445 addl 32(%rsp),%edx 3446 andl %eax,%r12d 3447 rorxl $25,%eax,%r13d 3448 rorxl $11,%eax,%r15d 3449 leal (%r8,%r14,1),%r8d 3450 leal (%rdx,%r12,1),%edx 3451 andnl %ecx,%eax,%r12d 3452 xorl %r15d,%r13d 3453 rorxl $6,%eax,%r14d 3454 leal (%rdx,%r12,1),%edx 3455 xorl %r14d,%r13d 3456 movl %r8d,%r15d 3457 rorxl $22,%r8d,%r12d 3458 leal (%rdx,%r13,1),%edx 3459 xorl %r9d,%r15d 3460 rorxl $13,%r8d,%r14d 3461 rorxl $2,%r8d,%r13d 3462 leal (%r11,%rdx,1),%r11d 3463 andl %r15d,%esi 3464 vpand %xmm12,%xmm11,%xmm8 3465 vaesenc %xmm10,%xmm9,%xmm9 3466 vmovdqu 192-128(%rdi),%xmm10 3467 xorl %r12d,%r14d 3468 xorl %r9d,%esi 3469 xorl %r13d,%r14d 3470 leal (%rdx,%rsi,1),%edx 3471 movl %eax,%r12d 3472 addl 36(%rsp),%ecx 3473 andl %r11d,%r12d 3474 rorxl $25,%r11d,%r13d 3475 rorxl $11,%r11d,%esi 3476 leal (%rdx,%r14,1),%edx 3477 leal (%rcx,%r12,1),%ecx 3478 andnl %ebx,%r11d,%r12d 3479 xorl %esi,%r13d 3480 rorxl $6,%r11d,%r14d 3481 leal (%rcx,%r12,1),%ecx 3482 xorl %r14d,%r13d 3483 movl %edx,%esi 3484 rorxl $22,%edx,%r12d 3485 leal (%rcx,%r13,1),%ecx 3486 xorl %r8d,%esi 3487 rorxl $13,%edx,%r14d 3488 rorxl $2,%edx,%r13d 3489 leal (%r10,%rcx,1),%r10d 3490 andl %esi,%r15d 3491 vaesenclast %xmm10,%xmm9,%xmm11 3492 vaesenc %xmm10,%xmm9,%xmm9 3493 vmovdqu 208-128(%rdi),%xmm10 3494 xorl %r12d,%r14d 3495 xorl %r8d,%r15d 3496 xorl %r13d,%r14d 3497 leal (%rcx,%r15,1),%ecx 3498 movl %r11d,%r12d 3499 addl 40(%rsp),%ebx 3500 andl %r10d,%r12d 3501 rorxl $25,%r10d,%r13d 3502 rorxl $11,%r10d,%r15d 3503 leal (%rcx,%r14,1),%ecx 3504 leal (%rbx,%r12,1),%ebx 3505 andnl %eax,%r10d,%r12d 3506 xorl %r15d,%r13d 3507 rorxl $6,%r10d,%r14d 3508 leal (%rbx,%r12,1),%ebx 3509 xorl %r14d,%r13d 3510 movl %ecx,%r15d 3511 rorxl $22,%ecx,%r12d 3512 leal (%rbx,%r13,1),%ebx 3513 xorl %edx,%r15d 3514 rorxl $13,%ecx,%r14d 3515 rorxl $2,%ecx,%r13d 3516 leal (%r9,%rbx,1),%r9d 3517 andl %r15d,%esi 3518 vpand %xmm13,%xmm11,%xmm11 3519 vaesenc %xmm10,%xmm9,%xmm9 3520 vmovdqu 224-128(%rdi),%xmm10 3521 xorl %r12d,%r14d 3522 xorl %edx,%esi 3523 xorl %r13d,%r14d 3524 leal (%rbx,%rsi,1),%ebx 3525 movl %r10d,%r12d 3526 addl 44(%rsp),%eax 3527 andl %r9d,%r12d 3528 rorxl $25,%r9d,%r13d 3529 rorxl $11,%r9d,%esi 3530 leal (%rbx,%r14,1),%ebx 3531 leal (%rax,%r12,1),%eax 3532 andnl %r11d,%r9d,%r12d 3533 xorl %esi,%r13d 3534 rorxl $6,%r9d,%r14d 3535 leal (%rax,%r12,1),%eax 3536 xorl %r14d,%r13d 3537 movl %ebx,%esi 3538 rorxl $22,%ebx,%r12d 3539 leal (%rax,%r13,1),%eax 3540 xorl %ecx,%esi 3541 rorxl $13,%ebx,%r14d 3542 rorxl $2,%ebx,%r13d 3543 leal (%r8,%rax,1),%r8d 3544 andl %esi,%r15d 3545 vpor %xmm11,%xmm8,%xmm8 3546 vaesenclast %xmm10,%xmm9,%xmm11 3547 vmovdqu 0-128(%rdi),%xmm10 3548 xorl %r12d,%r14d 3549 xorl %ecx,%r15d 3550 xorl %r13d,%r14d 3551 leal (%rax,%r15,1),%eax 3552 movl %r9d,%r12d 3553 vpextrq $1,%xmm15,%r12 3554 vmovq %xmm15,%r13 3555 movq 552(%rsp),%r15 3556 addl %r14d,%eax 3557 leaq 448(%rsp),%rbp 3558 3559 vpand %xmm14,%xmm11,%xmm11 3560 vpor %xmm11,%xmm8,%xmm8 3561 vmovdqu %xmm8,(%r12,%r13,1) 3562 leaq 16(%r13),%r13 3563 3564 addl 0(%r15),%eax 3565 addl 4(%r15),%ebx 3566 addl 8(%r15),%ecx 3567 addl 12(%r15),%edx 3568 addl 16(%r15),%r8d 3569 addl 20(%r15),%r9d 3570 addl 24(%r15),%r10d 3571 addl 28(%r15),%r11d 3572 3573 movl %eax,0(%r15) 3574 movl %ebx,4(%r15) 3575 movl %ecx,8(%r15) 3576 movl %edx,12(%r15) 3577 movl %r8d,16(%r15) 3578 movl %r9d,20(%r15) 3579 movl %r10d,24(%r15) 3580 movl %r11d,28(%r15) 3581 3582 cmpq 80(%rbp),%r13 3583 je .Ldone_avx2 3584 3585 xorl %r14d,%r14d 3586 movl %ebx,%esi 3587 movl %r9d,%r12d 3588 xorl %ecx,%esi 3589 jmp .Lower_avx2 3590.align 16 3591.Lower_avx2: 3592 vmovdqu (%r13),%xmm9 3593 vpinsrq $0,%r13,%xmm15,%xmm15 3594 addl 0+16(%rbp),%r11d 3595 andl %r8d,%r12d 3596 rorxl $25,%r8d,%r13d 3597 rorxl $11,%r8d,%r15d 3598 leal (%rax,%r14,1),%eax 3599 leal (%r11,%r12,1),%r11d 3600 andnl %r10d,%r8d,%r12d 3601 xorl %r15d,%r13d 3602 rorxl $6,%r8d,%r14d 3603 leal (%r11,%r12,1),%r11d 3604 xorl %r14d,%r13d 3605 movl %eax,%r15d 3606 rorxl $22,%eax,%r12d 3607 leal (%r11,%r13,1),%r11d 3608 xorl %ebx,%r15d 3609 rorxl $13,%eax,%r14d 3610 rorxl $2,%eax,%r13d 3611 leal (%rdx,%r11,1),%edx 3612 andl %r15d,%esi 3613 vpxor %xmm10,%xmm9,%xmm9 3614 vmovdqu 16-128(%rdi),%xmm10 3615 xorl %r12d,%r14d 3616 xorl %ebx,%esi 3617 xorl %r13d,%r14d 3618 leal (%r11,%rsi,1),%r11d 3619 movl %r8d,%r12d 3620 addl 4+16(%rbp),%r10d 3621 andl %edx,%r12d 3622 rorxl $25,%edx,%r13d 3623 rorxl $11,%edx,%esi 3624 leal (%r11,%r14,1),%r11d 3625 leal (%r10,%r12,1),%r10d 3626 andnl %r9d,%edx,%r12d 3627 xorl %esi,%r13d 3628 rorxl $6,%edx,%r14d 3629 leal (%r10,%r12,1),%r10d 3630 xorl %r14d,%r13d 3631 movl %r11d,%esi 3632 rorxl $22,%r11d,%r12d 3633 leal (%r10,%r13,1),%r10d 3634 xorl %eax,%esi 3635 rorxl $13,%r11d,%r14d 3636 rorxl $2,%r11d,%r13d 3637 leal (%rcx,%r10,1),%ecx 3638 andl %esi,%r15d 3639 vpxor %xmm8,%xmm9,%xmm9 3640 xorl %r12d,%r14d 3641 xorl %eax,%r15d 3642 xorl %r13d,%r14d 3643 leal (%r10,%r15,1),%r10d 3644 movl %edx,%r12d 3645 addl 8+16(%rbp),%r9d 3646 andl %ecx,%r12d 3647 rorxl $25,%ecx,%r13d 3648 rorxl $11,%ecx,%r15d 3649 leal (%r10,%r14,1),%r10d 3650 leal (%r9,%r12,1),%r9d 3651 andnl %r8d,%ecx,%r12d 3652 xorl %r15d,%r13d 3653 rorxl $6,%ecx,%r14d 3654 leal (%r9,%r12,1),%r9d 3655 xorl %r14d,%r13d 3656 movl %r10d,%r15d 3657 rorxl $22,%r10d,%r12d 3658 leal (%r9,%r13,1),%r9d 3659 xorl %r11d,%r15d 3660 rorxl $13,%r10d,%r14d 3661 rorxl $2,%r10d,%r13d 3662 leal (%rbx,%r9,1),%ebx 3663 andl %r15d,%esi 3664 vaesenc %xmm10,%xmm9,%xmm9 3665 vmovdqu 32-128(%rdi),%xmm10 3666 xorl %r12d,%r14d 3667 xorl %r11d,%esi 3668 xorl %r13d,%r14d 3669 leal (%r9,%rsi,1),%r9d 3670 movl %ecx,%r12d 3671 addl 12+16(%rbp),%r8d 3672 andl %ebx,%r12d 3673 rorxl $25,%ebx,%r13d 3674 rorxl $11,%ebx,%esi 3675 leal (%r9,%r14,1),%r9d 3676 leal (%r8,%r12,1),%r8d 3677 andnl %edx,%ebx,%r12d 3678 xorl %esi,%r13d 3679 rorxl $6,%ebx,%r14d 3680 leal (%r8,%r12,1),%r8d 3681 xorl %r14d,%r13d 3682 movl %r9d,%esi 3683 rorxl $22,%r9d,%r12d 3684 leal (%r8,%r13,1),%r8d 3685 xorl %r10d,%esi 3686 rorxl $13,%r9d,%r14d 3687 rorxl $2,%r9d,%r13d 3688 leal (%rax,%r8,1),%eax 3689 andl %esi,%r15d 3690 vaesenc %xmm10,%xmm9,%xmm9 3691 vmovdqu 48-128(%rdi),%xmm10 3692 xorl %r12d,%r14d 3693 xorl %r10d,%r15d 3694 xorl %r13d,%r14d 3695 leal (%r8,%r15,1),%r8d 3696 movl %ebx,%r12d 3697 addl 32+16(%rbp),%edx 3698 andl %eax,%r12d 3699 rorxl $25,%eax,%r13d 3700 rorxl $11,%eax,%r15d 3701 leal (%r8,%r14,1),%r8d 3702 leal (%rdx,%r12,1),%edx 3703 andnl %ecx,%eax,%r12d 3704 xorl %r15d,%r13d 3705 rorxl $6,%eax,%r14d 3706 leal (%rdx,%r12,1),%edx 3707 xorl %r14d,%r13d 3708 movl %r8d,%r15d 3709 rorxl $22,%r8d,%r12d 3710 leal (%rdx,%r13,1),%edx 3711 xorl %r9d,%r15d 3712 rorxl $13,%r8d,%r14d 3713 rorxl $2,%r8d,%r13d 3714 leal (%r11,%rdx,1),%r11d 3715 andl %r15d,%esi 3716 vaesenc %xmm10,%xmm9,%xmm9 3717 vmovdqu 64-128(%rdi),%xmm10 3718 xorl %r12d,%r14d 3719 xorl %r9d,%esi 3720 xorl %r13d,%r14d 3721 leal (%rdx,%rsi,1),%edx 3722 movl %eax,%r12d 3723 addl 36+16(%rbp),%ecx 3724 andl %r11d,%r12d 3725 rorxl $25,%r11d,%r13d 3726 rorxl $11,%r11d,%esi 3727 leal (%rdx,%r14,1),%edx 3728 leal (%rcx,%r12,1),%ecx 3729 andnl %ebx,%r11d,%r12d 3730 xorl %esi,%r13d 3731 rorxl $6,%r11d,%r14d 3732 leal (%rcx,%r12,1),%ecx 3733 xorl %r14d,%r13d 3734 movl %edx,%esi 3735 rorxl $22,%edx,%r12d 3736 leal (%rcx,%r13,1),%ecx 3737 xorl %r8d,%esi 3738 rorxl $13,%edx,%r14d 3739 rorxl $2,%edx,%r13d 3740 leal (%r10,%rcx,1),%r10d 3741 andl %esi,%r15d 3742 vaesenc %xmm10,%xmm9,%xmm9 3743 vmovdqu 80-128(%rdi),%xmm10 3744 xorl %r12d,%r14d 3745 xorl %r8d,%r15d 3746 xorl %r13d,%r14d 3747 leal (%rcx,%r15,1),%ecx 3748 movl %r11d,%r12d 3749 addl 40+16(%rbp),%ebx 3750 andl %r10d,%r12d 3751 rorxl $25,%r10d,%r13d 3752 rorxl $11,%r10d,%r15d 3753 leal (%rcx,%r14,1),%ecx 3754 leal (%rbx,%r12,1),%ebx 3755 andnl %eax,%r10d,%r12d 3756 xorl %r15d,%r13d 3757 rorxl $6,%r10d,%r14d 3758 leal (%rbx,%r12,1),%ebx 3759 xorl %r14d,%r13d 3760 movl %ecx,%r15d 3761 rorxl $22,%ecx,%r12d 3762 leal (%rbx,%r13,1),%ebx 3763 xorl %edx,%r15d 3764 rorxl $13,%ecx,%r14d 3765 rorxl $2,%ecx,%r13d 3766 leal (%r9,%rbx,1),%r9d 3767 andl %r15d,%esi 3768 vaesenc %xmm10,%xmm9,%xmm9 3769 vmovdqu 96-128(%rdi),%xmm10 3770 xorl %r12d,%r14d 3771 xorl %edx,%esi 3772 xorl %r13d,%r14d 3773 leal (%rbx,%rsi,1),%ebx 3774 movl %r10d,%r12d 3775 addl 44+16(%rbp),%eax 3776 andl %r9d,%r12d 3777 rorxl $25,%r9d,%r13d 3778 rorxl $11,%r9d,%esi 3779 leal (%rbx,%r14,1),%ebx 3780 leal (%rax,%r12,1),%eax 3781 andnl %r11d,%r9d,%r12d 3782 xorl %esi,%r13d 3783 rorxl $6,%r9d,%r14d 3784 leal (%rax,%r12,1),%eax 3785 xorl %r14d,%r13d 3786 movl %ebx,%esi 3787 rorxl $22,%ebx,%r12d 3788 leal (%rax,%r13,1),%eax 3789 xorl %ecx,%esi 3790 rorxl $13,%ebx,%r14d 3791 rorxl $2,%ebx,%r13d 3792 leal (%r8,%rax,1),%r8d 3793 andl %esi,%r15d 3794 vaesenc %xmm10,%xmm9,%xmm9 3795 vmovdqu 112-128(%rdi),%xmm10 3796 xorl %r12d,%r14d 3797 xorl %ecx,%r15d 3798 xorl %r13d,%r14d 3799 leal (%rax,%r15,1),%eax 3800 movl %r9d,%r12d 3801 leaq -64(%rbp),%rbp 3802 addl 0+16(%rbp),%r11d 3803 andl %r8d,%r12d 3804 rorxl $25,%r8d,%r13d 3805 rorxl $11,%r8d,%r15d 3806 leal (%rax,%r14,1),%eax 3807 leal (%r11,%r12,1),%r11d 3808 andnl %r10d,%r8d,%r12d 3809 xorl %r15d,%r13d 3810 rorxl $6,%r8d,%r14d 3811 leal (%r11,%r12,1),%r11d 3812 xorl %r14d,%r13d 3813 movl %eax,%r15d 3814 rorxl $22,%eax,%r12d 3815 leal (%r11,%r13,1),%r11d 3816 xorl %ebx,%r15d 3817 rorxl $13,%eax,%r14d 3818 rorxl $2,%eax,%r13d 3819 leal (%rdx,%r11,1),%edx 3820 andl %r15d,%esi 3821 vaesenc %xmm10,%xmm9,%xmm9 3822 vmovdqu 128-128(%rdi),%xmm10 3823 xorl %r12d,%r14d 3824 xorl %ebx,%esi 3825 xorl %r13d,%r14d 3826 leal (%r11,%rsi,1),%r11d 3827 movl %r8d,%r12d 3828 addl 4+16(%rbp),%r10d 3829 andl %edx,%r12d 3830 rorxl $25,%edx,%r13d 3831 rorxl $11,%edx,%esi 3832 leal (%r11,%r14,1),%r11d 3833 leal (%r10,%r12,1),%r10d 3834 andnl %r9d,%edx,%r12d 3835 xorl %esi,%r13d 3836 rorxl $6,%edx,%r14d 3837 leal (%r10,%r12,1),%r10d 3838 xorl %r14d,%r13d 3839 movl %r11d,%esi 3840 rorxl $22,%r11d,%r12d 3841 leal (%r10,%r13,1),%r10d 3842 xorl %eax,%esi 3843 rorxl $13,%r11d,%r14d 3844 rorxl $2,%r11d,%r13d 3845 leal (%rcx,%r10,1),%ecx 3846 andl %esi,%r15d 3847 vaesenc %xmm10,%xmm9,%xmm9 3848 vmovdqu 144-128(%rdi),%xmm10 3849 xorl %r12d,%r14d 3850 xorl %eax,%r15d 3851 xorl %r13d,%r14d 3852 leal (%r10,%r15,1),%r10d 3853 movl %edx,%r12d 3854 addl 8+16(%rbp),%r9d 3855 andl %ecx,%r12d 3856 rorxl $25,%ecx,%r13d 3857 rorxl $11,%ecx,%r15d 3858 leal (%r10,%r14,1),%r10d 3859 leal (%r9,%r12,1),%r9d 3860 andnl %r8d,%ecx,%r12d 3861 xorl %r15d,%r13d 3862 rorxl $6,%ecx,%r14d 3863 leal (%r9,%r12,1),%r9d 3864 xorl %r14d,%r13d 3865 movl %r10d,%r15d 3866 rorxl $22,%r10d,%r12d 3867 leal (%r9,%r13,1),%r9d 3868 xorl %r11d,%r15d 3869 rorxl $13,%r10d,%r14d 3870 rorxl $2,%r10d,%r13d 3871 leal (%rbx,%r9,1),%ebx 3872 andl %r15d,%esi 3873 vaesenc %xmm10,%xmm9,%xmm9 3874 vmovdqu 160-128(%rdi),%xmm10 3875 xorl %r12d,%r14d 3876 xorl %r11d,%esi 3877 xorl %r13d,%r14d 3878 leal (%r9,%rsi,1),%r9d 3879 movl %ecx,%r12d 3880 addl 12+16(%rbp),%r8d 3881 andl %ebx,%r12d 3882 rorxl $25,%ebx,%r13d 3883 rorxl $11,%ebx,%esi 3884 leal (%r9,%r14,1),%r9d 3885 leal (%r8,%r12,1),%r8d 3886 andnl %edx,%ebx,%r12d 3887 xorl %esi,%r13d 3888 rorxl $6,%ebx,%r14d 3889 leal (%r8,%r12,1),%r8d 3890 xorl %r14d,%r13d 3891 movl %r9d,%esi 3892 rorxl $22,%r9d,%r12d 3893 leal (%r8,%r13,1),%r8d 3894 xorl %r10d,%esi 3895 rorxl $13,%r9d,%r14d 3896 rorxl $2,%r9d,%r13d 3897 leal (%rax,%r8,1),%eax 3898 andl %esi,%r15d 3899 vaesenclast %xmm10,%xmm9,%xmm11 3900 vaesenc %xmm10,%xmm9,%xmm9 3901 vmovdqu 176-128(%rdi),%xmm10 3902 xorl %r12d,%r14d 3903 xorl %r10d,%r15d 3904 xorl %r13d,%r14d 3905 leal (%r8,%r15,1),%r8d 3906 movl %ebx,%r12d 3907 addl 32+16(%rbp),%edx 3908 andl %eax,%r12d 3909 rorxl $25,%eax,%r13d 3910 rorxl $11,%eax,%r15d 3911 leal (%r8,%r14,1),%r8d 3912 leal (%rdx,%r12,1),%edx 3913 andnl %ecx,%eax,%r12d 3914 xorl %r15d,%r13d 3915 rorxl $6,%eax,%r14d 3916 leal (%rdx,%r12,1),%edx 3917 xorl %r14d,%r13d 3918 movl %r8d,%r15d 3919 rorxl $22,%r8d,%r12d 3920 leal (%rdx,%r13,1),%edx 3921 xorl %r9d,%r15d 3922 rorxl $13,%r8d,%r14d 3923 rorxl $2,%r8d,%r13d 3924 leal (%r11,%rdx,1),%r11d 3925 andl %r15d,%esi 3926 vpand %xmm12,%xmm11,%xmm8 3927 vaesenc %xmm10,%xmm9,%xmm9 3928 vmovdqu 192-128(%rdi),%xmm10 3929 xorl %r12d,%r14d 3930 xorl %r9d,%esi 3931 xorl %r13d,%r14d 3932 leal (%rdx,%rsi,1),%edx 3933 movl %eax,%r12d 3934 addl 36+16(%rbp),%ecx 3935 andl %r11d,%r12d 3936 rorxl $25,%r11d,%r13d 3937 rorxl $11,%r11d,%esi 3938 leal (%rdx,%r14,1),%edx 3939 leal (%rcx,%r12,1),%ecx 3940 andnl %ebx,%r11d,%r12d 3941 xorl %esi,%r13d 3942 rorxl $6,%r11d,%r14d 3943 leal (%rcx,%r12,1),%ecx 3944 xorl %r14d,%r13d 3945 movl %edx,%esi 3946 rorxl $22,%edx,%r12d 3947 leal (%rcx,%r13,1),%ecx 3948 xorl %r8d,%esi 3949 rorxl $13,%edx,%r14d 3950 rorxl $2,%edx,%r13d 3951 leal (%r10,%rcx,1),%r10d 3952 andl %esi,%r15d 3953 vaesenclast %xmm10,%xmm9,%xmm11 3954 vaesenc %xmm10,%xmm9,%xmm9 3955 vmovdqu 208-128(%rdi),%xmm10 3956 xorl %r12d,%r14d 3957 xorl %r8d,%r15d 3958 xorl %r13d,%r14d 3959 leal (%rcx,%r15,1),%ecx 3960 movl %r11d,%r12d 3961 addl 40+16(%rbp),%ebx 3962 andl %r10d,%r12d 3963 rorxl $25,%r10d,%r13d 3964 rorxl $11,%r10d,%r15d 3965 leal (%rcx,%r14,1),%ecx 3966 leal (%rbx,%r12,1),%ebx 3967 andnl %eax,%r10d,%r12d 3968 xorl %r15d,%r13d 3969 rorxl $6,%r10d,%r14d 3970 leal (%rbx,%r12,1),%ebx 3971 xorl %r14d,%r13d 3972 movl %ecx,%r15d 3973 rorxl $22,%ecx,%r12d 3974 leal (%rbx,%r13,1),%ebx 3975 xorl %edx,%r15d 3976 rorxl $13,%ecx,%r14d 3977 rorxl $2,%ecx,%r13d 3978 leal (%r9,%rbx,1),%r9d 3979 andl %r15d,%esi 3980 vpand %xmm13,%xmm11,%xmm11 3981 vaesenc %xmm10,%xmm9,%xmm9 3982 vmovdqu 224-128(%rdi),%xmm10 3983 xorl %r12d,%r14d 3984 xorl %edx,%esi 3985 xorl %r13d,%r14d 3986 leal (%rbx,%rsi,1),%ebx 3987 movl %r10d,%r12d 3988 addl 44+16(%rbp),%eax 3989 andl %r9d,%r12d 3990 rorxl $25,%r9d,%r13d 3991 rorxl $11,%r9d,%esi 3992 leal (%rbx,%r14,1),%ebx 3993 leal (%rax,%r12,1),%eax 3994 andnl %r11d,%r9d,%r12d 3995 xorl %esi,%r13d 3996 rorxl $6,%r9d,%r14d 3997 leal (%rax,%r12,1),%eax 3998 xorl %r14d,%r13d 3999 movl %ebx,%esi 4000 rorxl $22,%ebx,%r12d 4001 leal (%rax,%r13,1),%eax 4002 xorl %ecx,%esi 4003 rorxl $13,%ebx,%r14d 4004 rorxl $2,%ebx,%r13d 4005 leal (%r8,%rax,1),%r8d 4006 andl %esi,%r15d 4007 vpor %xmm11,%xmm8,%xmm8 4008 vaesenclast %xmm10,%xmm9,%xmm11 4009 vmovdqu 0-128(%rdi),%xmm10 4010 xorl %r12d,%r14d 4011 xorl %ecx,%r15d 4012 xorl %r13d,%r14d 4013 leal (%rax,%r15,1),%eax 4014 movl %r9d,%r12d 4015 vmovq %xmm15,%r13 4016 vpextrq $1,%xmm15,%r15 4017 vpand %xmm14,%xmm11,%xmm11 4018 vpor %xmm11,%xmm8,%xmm8 4019 leaq -64(%rbp),%rbp 4020 vmovdqu %xmm8,(%r15,%r13,1) 4021 leaq 16(%r13),%r13 4022 cmpq %rsp,%rbp 4023 jae .Lower_avx2 4024 4025 movq 552(%rsp),%r15 4026 leaq 64(%r13),%r13 4027 movq 560(%rsp),%rsi 4028 addl %r14d,%eax 4029 leaq 448(%rsp),%rsp 4030 4031 addl 0(%r15),%eax 4032 addl 4(%r15),%ebx 4033 addl 8(%r15),%ecx 4034 addl 12(%r15),%edx 4035 addl 16(%r15),%r8d 4036 addl 20(%r15),%r9d 4037 addl 24(%r15),%r10d 4038 leaq (%rsi,%r13,1),%r12 4039 addl 28(%r15),%r11d 4040 4041 cmpq 64+16(%rsp),%r13 4042 4043 movl %eax,0(%r15) 4044 cmoveq %rsp,%r12 4045 movl %ebx,4(%r15) 4046 movl %ecx,8(%r15) 4047 movl %edx,12(%r15) 4048 movl %r8d,16(%r15) 4049 movl %r9d,20(%r15) 4050 movl %r10d,24(%r15) 4051 movl %r11d,28(%r15) 4052 4053 jbe .Loop_avx2 4054 leaq (%rsp),%rbp 4055 4056 4057.cfi_escape 0x0f,0x06,0x76,0xf8,0x00,0x06,0x23,0x08 4058 4059.Ldone_avx2: 4060 movq 64+32(%rbp),%r8 4061 movq 64+56(%rbp),%rsi 4062.cfi_def_cfa %rsi,8 4063 vmovdqu %xmm8,(%r8) 4064 vzeroall 4065 movq -48(%rsi),%r15 4066.cfi_restore %r15 4067 movq -40(%rsi),%r14 4068.cfi_restore %r14 4069 movq -32(%rsi),%r13 4070.cfi_restore %r13 4071 movq -24(%rsi),%r12 4072.cfi_restore %r12 4073 movq -16(%rsi),%rbp 4074.cfi_restore %rbp 4075 movq -8(%rsi),%rbx 4076.cfi_restore %rbx 4077 leaq (%rsi),%rsp 4078.cfi_def_cfa_register %rsp 4079.Lepilogue_avx2: 4080 .byte 0xf3,0xc3 4081.cfi_endproc 4082.size aesni_cbc_sha256_enc_avx2,.-aesni_cbc_sha256_enc_avx2 4083.type aesni_cbc_sha256_enc_shaext,@function 4084.align 32 4085aesni_cbc_sha256_enc_shaext: 4086.cfi_startproc 4087 movq 8(%rsp),%r10 4088 leaq K256+128(%rip),%rax 4089 movdqu (%r9),%xmm1 4090 movdqu 16(%r9),%xmm2 4091 movdqa 512-128(%rax),%xmm3 4092 4093 movl 240(%rcx),%r11d 4094 subq %rdi,%rsi 4095 movups (%rcx),%xmm15 4096 movups (%r8),%xmm6 4097 movups 16(%rcx),%xmm4 4098 leaq 112(%rcx),%rcx 4099 4100 pshufd $0x1b,%xmm1,%xmm0 4101 pshufd $0xb1,%xmm1,%xmm1 4102 pshufd $0x1b,%xmm2,%xmm2 4103 movdqa %xmm3,%xmm7 4104.byte 102,15,58,15,202,8 4105 punpcklqdq %xmm0,%xmm2 4106 4107 jmp .Loop_shaext 4108 4109.align 16 4110.Loop_shaext: 4111 movdqu (%r10),%xmm10 4112 movdqu 16(%r10),%xmm11 4113 movdqu 32(%r10),%xmm12 4114.byte 102,68,15,56,0,211 4115 movdqu 48(%r10),%xmm13 4116 4117 movdqa 0-128(%rax),%xmm0 4118 paddd %xmm10,%xmm0 4119.byte 102,68,15,56,0,219 4120 movdqa %xmm2,%xmm9 4121 movdqa %xmm1,%xmm8 4122 movups 0(%rdi),%xmm14 4123 xorps %xmm15,%xmm14 4124 xorps %xmm14,%xmm6 4125 movups -80(%rcx),%xmm5 4126 aesenc %xmm4,%xmm6 4127.byte 15,56,203,209 4128 pshufd $0x0e,%xmm0,%xmm0 4129 movups -64(%rcx),%xmm4 4130 aesenc %xmm5,%xmm6 4131.byte 15,56,203,202 4132 4133 movdqa 32-128(%rax),%xmm0 4134 paddd %xmm11,%xmm0 4135.byte 102,68,15,56,0,227 4136 leaq 64(%r10),%r10 4137 movups -48(%rcx),%xmm5 4138 aesenc %xmm4,%xmm6 4139.byte 15,56,203,209 4140 pshufd $0x0e,%xmm0,%xmm0 4141 movups -32(%rcx),%xmm4 4142 aesenc %xmm5,%xmm6 4143.byte 15,56,203,202 4144 4145 movdqa 64-128(%rax),%xmm0 4146 paddd %xmm12,%xmm0 4147.byte 102,68,15,56,0,235 4148.byte 69,15,56,204,211 4149 movups -16(%rcx),%xmm5 4150 aesenc %xmm4,%xmm6 4151.byte 15,56,203,209 4152 pshufd $0x0e,%xmm0,%xmm0 4153 movdqa %xmm13,%xmm3 4154.byte 102,65,15,58,15,220,4 4155 paddd %xmm3,%xmm10 4156 movups 0(%rcx),%xmm4 4157 aesenc %xmm5,%xmm6 4158.byte 15,56,203,202 4159 4160 movdqa 96-128(%rax),%xmm0 4161 paddd %xmm13,%xmm0 4162.byte 69,15,56,205,213 4163.byte 69,15,56,204,220 4164 movups 16(%rcx),%xmm5 4165 aesenc %xmm4,%xmm6 4166.byte 15,56,203,209 4167 pshufd $0x0e,%xmm0,%xmm0 4168 movups 32(%rcx),%xmm4 4169 aesenc %xmm5,%xmm6 4170 movdqa %xmm10,%xmm3 4171.byte 102,65,15,58,15,221,4 4172 paddd %xmm3,%xmm11 4173.byte 15,56,203,202 4174 movdqa 128-128(%rax),%xmm0 4175 paddd %xmm10,%xmm0 4176.byte 69,15,56,205,218 4177.byte 69,15,56,204,229 4178 movups 48(%rcx),%xmm5 4179 aesenc %xmm4,%xmm6 4180.byte 15,56,203,209 4181 pshufd $0x0e,%xmm0,%xmm0 4182 movdqa %xmm11,%xmm3 4183.byte 102,65,15,58,15,218,4 4184 paddd %xmm3,%xmm12 4185 cmpl $11,%r11d 4186 jb .Laesenclast1 4187 movups 64(%rcx),%xmm4 4188 aesenc %xmm5,%xmm6 4189 movups 80(%rcx),%xmm5 4190 aesenc %xmm4,%xmm6 4191 je .Laesenclast1 4192 movups 96(%rcx),%xmm4 4193 aesenc %xmm5,%xmm6 4194 movups 112(%rcx),%xmm5 4195 aesenc %xmm4,%xmm6 4196.Laesenclast1: 4197 aesenclast %xmm5,%xmm6 4198 movups 16-112(%rcx),%xmm4 4199 nop 4200.byte 15,56,203,202 4201 movups 16(%rdi),%xmm14 4202 xorps %xmm15,%xmm14 4203 movups %xmm6,0(%rsi,%rdi,1) 4204 xorps %xmm14,%xmm6 4205 movups -80(%rcx),%xmm5 4206 aesenc %xmm4,%xmm6 4207 movdqa 160-128(%rax),%xmm0 4208 paddd %xmm11,%xmm0 4209.byte 69,15,56,205,227 4210.byte 69,15,56,204,234 4211 movups -64(%rcx),%xmm4 4212 aesenc %xmm5,%xmm6 4213.byte 15,56,203,209 4214 pshufd $0x0e,%xmm0,%xmm0 4215 movdqa %xmm12,%xmm3 4216.byte 102,65,15,58,15,219,4 4217 paddd %xmm3,%xmm13 4218 movups -48(%rcx),%xmm5 4219 aesenc %xmm4,%xmm6 4220.byte 15,56,203,202 4221 movdqa 192-128(%rax),%xmm0 4222 paddd %xmm12,%xmm0 4223.byte 69,15,56,205,236 4224.byte 69,15,56,204,211 4225 movups -32(%rcx),%xmm4 4226 aesenc %xmm5,%xmm6 4227.byte 15,56,203,209 4228 pshufd $0x0e,%xmm0,%xmm0 4229 movdqa %xmm13,%xmm3 4230.byte 102,65,15,58,15,220,4 4231 paddd %xmm3,%xmm10 4232 movups -16(%rcx),%xmm5 4233 aesenc %xmm4,%xmm6 4234.byte 15,56,203,202 4235 movdqa 224-128(%rax),%xmm0 4236 paddd %xmm13,%xmm0 4237.byte 69,15,56,205,213 4238.byte 69,15,56,204,220 4239 movups 0(%rcx),%xmm4 4240 aesenc %xmm5,%xmm6 4241.byte 15,56,203,209 4242 pshufd $0x0e,%xmm0,%xmm0 4243 movdqa %xmm10,%xmm3 4244.byte 102,65,15,58,15,221,4 4245 paddd %xmm3,%xmm11 4246 movups 16(%rcx),%xmm5 4247 aesenc %xmm4,%xmm6 4248.byte 15,56,203,202 4249 movdqa 256-128(%rax),%xmm0 4250 paddd %xmm10,%xmm0 4251.byte 69,15,56,205,218 4252.byte 69,15,56,204,229 4253 movups 32(%rcx),%xmm4 4254 aesenc %xmm5,%xmm6 4255.byte 15,56,203,209 4256 pshufd $0x0e,%xmm0,%xmm0 4257 movdqa %xmm11,%xmm3 4258.byte 102,65,15,58,15,218,4 4259 paddd %xmm3,%xmm12 4260 movups 48(%rcx),%xmm5 4261 aesenc %xmm4,%xmm6 4262 cmpl $11,%r11d 4263 jb .Laesenclast2 4264 movups 64(%rcx),%xmm4 4265 aesenc %xmm5,%xmm6 4266 movups 80(%rcx),%xmm5 4267 aesenc %xmm4,%xmm6 4268 je .Laesenclast2 4269 movups 96(%rcx),%xmm4 4270 aesenc %xmm5,%xmm6 4271 movups 112(%rcx),%xmm5 4272 aesenc %xmm4,%xmm6 4273.Laesenclast2: 4274 aesenclast %xmm5,%xmm6 4275 movups 16-112(%rcx),%xmm4 4276 nop 4277.byte 15,56,203,202 4278 movups 32(%rdi),%xmm14 4279 xorps %xmm15,%xmm14 4280 movups %xmm6,16(%rsi,%rdi,1) 4281 xorps %xmm14,%xmm6 4282 movups -80(%rcx),%xmm5 4283 aesenc %xmm4,%xmm6 4284 movdqa 288-128(%rax),%xmm0 4285 paddd %xmm11,%xmm0 4286.byte 69,15,56,205,227 4287.byte 69,15,56,204,234 4288 movups -64(%rcx),%xmm4 4289 aesenc %xmm5,%xmm6 4290.byte 15,56,203,209 4291 pshufd $0x0e,%xmm0,%xmm0 4292 movdqa %xmm12,%xmm3 4293.byte 102,65,15,58,15,219,4 4294 paddd %xmm3,%xmm13 4295 movups -48(%rcx),%xmm5 4296 aesenc %xmm4,%xmm6 4297.byte 15,56,203,202 4298 movdqa 320-128(%rax),%xmm0 4299 paddd %xmm12,%xmm0 4300.byte 69,15,56,205,236 4301.byte 69,15,56,204,211 4302 movups -32(%rcx),%xmm4 4303 aesenc %xmm5,%xmm6 4304.byte 15,56,203,209 4305 pshufd $0x0e,%xmm0,%xmm0 4306 movdqa %xmm13,%xmm3 4307.byte 102,65,15,58,15,220,4 4308 paddd %xmm3,%xmm10 4309 movups -16(%rcx),%xmm5 4310 aesenc %xmm4,%xmm6 4311.byte 15,56,203,202 4312 movdqa 352-128(%rax),%xmm0 4313 paddd %xmm13,%xmm0 4314.byte 69,15,56,205,213 4315.byte 69,15,56,204,220 4316 movups 0(%rcx),%xmm4 4317 aesenc %xmm5,%xmm6 4318.byte 15,56,203,209 4319 pshufd $0x0e,%xmm0,%xmm0 4320 movdqa %xmm10,%xmm3 4321.byte 102,65,15,58,15,221,4 4322 paddd %xmm3,%xmm11 4323 movups 16(%rcx),%xmm5 4324 aesenc %xmm4,%xmm6 4325.byte 15,56,203,202 4326 movdqa 384-128(%rax),%xmm0 4327 paddd %xmm10,%xmm0 4328.byte 69,15,56,205,218 4329.byte 69,15,56,204,229 4330 movups 32(%rcx),%xmm4 4331 aesenc %xmm5,%xmm6 4332.byte 15,56,203,209 4333 pshufd $0x0e,%xmm0,%xmm0 4334 movdqa %xmm11,%xmm3 4335.byte 102,65,15,58,15,218,4 4336 paddd %xmm3,%xmm12 4337 movups 48(%rcx),%xmm5 4338 aesenc %xmm4,%xmm6 4339.byte 15,56,203,202 4340 movdqa 416-128(%rax),%xmm0 4341 paddd %xmm11,%xmm0 4342.byte 69,15,56,205,227 4343.byte 69,15,56,204,234 4344 cmpl $11,%r11d 4345 jb .Laesenclast3 4346 movups 64(%rcx),%xmm4 4347 aesenc %xmm5,%xmm6 4348 movups 80(%rcx),%xmm5 4349 aesenc %xmm4,%xmm6 4350 je .Laesenclast3 4351 movups 96(%rcx),%xmm4 4352 aesenc %xmm5,%xmm6 4353 movups 112(%rcx),%xmm5 4354 aesenc %xmm4,%xmm6 4355.Laesenclast3: 4356 aesenclast %xmm5,%xmm6 4357 movups 16-112(%rcx),%xmm4 4358 nop 4359.byte 15,56,203,209 4360 pshufd $0x0e,%xmm0,%xmm0 4361 movdqa %xmm12,%xmm3 4362.byte 102,65,15,58,15,219,4 4363 paddd %xmm3,%xmm13 4364 movups 48(%rdi),%xmm14 4365 xorps %xmm15,%xmm14 4366 movups %xmm6,32(%rsi,%rdi,1) 4367 xorps %xmm14,%xmm6 4368 movups -80(%rcx),%xmm5 4369 aesenc %xmm4,%xmm6 4370 movups -64(%rcx),%xmm4 4371 aesenc %xmm5,%xmm6 4372.byte 15,56,203,202 4373 4374 movdqa 448-128(%rax),%xmm0 4375 paddd %xmm12,%xmm0 4376.byte 69,15,56,205,236 4377 movdqa %xmm7,%xmm3 4378 movups -48(%rcx),%xmm5 4379 aesenc %xmm4,%xmm6 4380.byte 15,56,203,209 4381 pshufd $0x0e,%xmm0,%xmm0 4382 movups -32(%rcx),%xmm4 4383 aesenc %xmm5,%xmm6 4384.byte 15,56,203,202 4385 4386 movdqa 480-128(%rax),%xmm0 4387 paddd %xmm13,%xmm0 4388 movups -16(%rcx),%xmm5 4389 aesenc %xmm4,%xmm6 4390 movups 0(%rcx),%xmm4 4391 aesenc %xmm5,%xmm6 4392.byte 15,56,203,209 4393 pshufd $0x0e,%xmm0,%xmm0 4394 movups 16(%rcx),%xmm5 4395 aesenc %xmm4,%xmm6 4396.byte 15,56,203,202 4397 4398 movups 32(%rcx),%xmm4 4399 aesenc %xmm5,%xmm6 4400 movups 48(%rcx),%xmm5 4401 aesenc %xmm4,%xmm6 4402 cmpl $11,%r11d 4403 jb .Laesenclast4 4404 movups 64(%rcx),%xmm4 4405 aesenc %xmm5,%xmm6 4406 movups 80(%rcx),%xmm5 4407 aesenc %xmm4,%xmm6 4408 je .Laesenclast4 4409 movups 96(%rcx),%xmm4 4410 aesenc %xmm5,%xmm6 4411 movups 112(%rcx),%xmm5 4412 aesenc %xmm4,%xmm6 4413.Laesenclast4: 4414 aesenclast %xmm5,%xmm6 4415 movups 16-112(%rcx),%xmm4 4416 nop 4417 4418 paddd %xmm9,%xmm2 4419 paddd %xmm8,%xmm1 4420 4421 decq %rdx 4422 movups %xmm6,48(%rsi,%rdi,1) 4423 leaq 64(%rdi),%rdi 4424 jnz .Loop_shaext 4425 4426 pshufd $0xb1,%xmm2,%xmm2 4427 pshufd $0x1b,%xmm1,%xmm3 4428 pshufd $0xb1,%xmm1,%xmm1 4429 punpckhqdq %xmm2,%xmm1 4430.byte 102,15,58,15,211,8 4431 4432 movups %xmm6,(%r8) 4433 movdqu %xmm1,(%r9) 4434 movdqu %xmm2,16(%r9) 4435 .byte 0xf3,0xc3 4436.cfi_endproc 4437.size aesni_cbc_sha256_enc_shaext,.-aesni_cbc_sha256_enc_shaext 4438