/* Do not modify. This file is auto-generated from e_padlock-x86_64.pl. */ .text .globl padlock_capability .type padlock_capability,@function .align 16 padlock_capability: movq %rbx,%r8 xorl %eax,%eax cpuid xorl %eax,%eax cmpl $0x746e6543,%ebx jne .Lzhaoxin cmpl $0x48727561,%edx jne .Lnoluck cmpl $0x736c7561,%ecx jne .Lnoluck jmp .LzhaoxinEnd .Lzhaoxin: cmpl $0x68532020,%ebx jne .Lnoluck cmpl $0x68676e61,%edx jne .Lnoluck cmpl $0x20206961,%ecx jne .Lnoluck .LzhaoxinEnd: movl $0xC0000000,%eax cpuid movl %eax,%edx xorl %eax,%eax cmpl $0xC0000001,%edx jb .Lnoluck movl $0xC0000001,%eax cpuid movl %edx,%eax andl $0xffffffef,%eax orl $0x10,%eax .Lnoluck: movq %r8,%rbx .byte 0xf3,0xc3 .size padlock_capability,.-padlock_capability .globl padlock_key_bswap .type padlock_key_bswap,@function .align 16 padlock_key_bswap: movl 240(%rdi),%edx incl %edx shll $2,%edx .Lbswap_loop: movl (%rdi),%eax bswapl %eax movl %eax,(%rdi) leaq 4(%rdi),%rdi subl $1,%edx jnz .Lbswap_loop .byte 0xf3,0xc3 .size padlock_key_bswap,.-padlock_key_bswap .globl padlock_verify_context .type padlock_verify_context,@function .align 16 padlock_verify_context: movq %rdi,%rdx pushf leaq .Lpadlock_saved_context(%rip),%rax call _padlock_verify_ctx leaq 8(%rsp),%rsp .byte 0xf3,0xc3 .size padlock_verify_context,.-padlock_verify_context .type _padlock_verify_ctx,@function .align 16 _padlock_verify_ctx: movq 8(%rsp),%r8 btq $30,%r8 jnc .Lverified cmpq (%rax),%rdx je .Lverified pushf popf .Lverified: movq %rdx,(%rax) .byte 0xf3,0xc3 .size _padlock_verify_ctx,.-_padlock_verify_ctx .globl padlock_reload_key .type padlock_reload_key,@function .align 16 padlock_reload_key: pushf popf .byte 0xf3,0xc3 .size padlock_reload_key,.-padlock_reload_key .globl padlock_aes_block .type padlock_aes_block,@function .align 16 padlock_aes_block: movq %rbx,%r8 movq $1,%rcx leaq 32(%rdx),%rbx leaq 16(%rdx),%rdx .byte 0xf3,0x0f,0xa7,0xc8 movq %r8,%rbx .byte 0xf3,0xc3 .size padlock_aes_block,.-padlock_aes_block .globl padlock_xstore .type padlock_xstore,@function .align 16 padlock_xstore: movl %esi,%edx .byte 0x0f,0xa7,0xc0 .byte 0xf3,0xc3 .size padlock_xstore,.-padlock_xstore .globl padlock_sha1_oneshot .type padlock_sha1_oneshot,@function .align 16 padlock_sha1_oneshot: movq %rdx,%rcx movq %rdi,%rdx movups (%rdi),%xmm0 subq $128+8,%rsp movl 16(%rdi),%eax movaps %xmm0,(%rsp) movq %rsp,%rdi movl %eax,16(%rsp) xorq %rax,%rax .byte 0xf3,0x0f,0xa6,0xc8 movaps (%rsp),%xmm0 movl 16(%rsp),%eax addq $128+8,%rsp movups %xmm0,(%rdx) movl %eax,16(%rdx) .byte 0xf3,0xc3 .size padlock_sha1_oneshot,.-padlock_sha1_oneshot .globl padlock_sha1_blocks .type padlock_sha1_blocks,@function .align 16 padlock_sha1_blocks: movq %rdx,%rcx movq %rdi,%rdx movups (%rdi),%xmm0 subq $128+8,%rsp movl 16(%rdi),%eax movaps %xmm0,(%rsp) movq %rsp,%rdi movl %eax,16(%rsp) movq $-1,%rax .byte 0xf3,0x0f,0xa6,0xc8 movaps (%rsp),%xmm0 movl 16(%rsp),%eax addq $128+8,%rsp movups %xmm0,(%rdx) movl %eax,16(%rdx) .byte 0xf3,0xc3 .size padlock_sha1_blocks,.-padlock_sha1_blocks .globl padlock_sha256_oneshot .type padlock_sha256_oneshot,@function .align 16 padlock_sha256_oneshot: movq %rdx,%rcx movq %rdi,%rdx movups (%rdi),%xmm0 subq $128+8,%rsp movups 16(%rdi),%xmm1 movaps %xmm0,(%rsp) movq %rsp,%rdi movaps %xmm1,16(%rsp) xorq %rax,%rax .byte 0xf3,0x0f,0xa6,0xd0 movaps (%rsp),%xmm0 movaps 16(%rsp),%xmm1 addq $128+8,%rsp movups %xmm0,(%rdx) movups %xmm1,16(%rdx) .byte 0xf3,0xc3 .size padlock_sha256_oneshot,.-padlock_sha256_oneshot .globl padlock_sha256_blocks .type padlock_sha256_blocks,@function .align 16 padlock_sha256_blocks: movq %rdx,%rcx movq %rdi,%rdx movups (%rdi),%xmm0 subq $128+8,%rsp movups 16(%rdi),%xmm1 movaps %xmm0,(%rsp) movq %rsp,%rdi movaps %xmm1,16(%rsp) movq $-1,%rax .byte 0xf3,0x0f,0xa6,0xd0 movaps (%rsp),%xmm0 movaps 16(%rsp),%xmm1 addq $128+8,%rsp movups %xmm0,(%rdx) movups %xmm1,16(%rdx) .byte 0xf3,0xc3 .size padlock_sha256_blocks,.-padlock_sha256_blocks .globl padlock_sha512_blocks .type padlock_sha512_blocks,@function .align 16 padlock_sha512_blocks: movq %rdx,%rcx movq %rdi,%rdx movups (%rdi),%xmm0 subq $128+8,%rsp movups 16(%rdi),%xmm1 movups 32(%rdi),%xmm2 movups 48(%rdi),%xmm3 movaps %xmm0,(%rsp) movq %rsp,%rdi movaps %xmm1,16(%rsp) movaps %xmm2,32(%rsp) movaps %xmm3,48(%rsp) .byte 0xf3,0x0f,0xa6,0xe0 movaps (%rsp),%xmm0 movaps 16(%rsp),%xmm1 movaps 32(%rsp),%xmm2 movaps 48(%rsp),%xmm3 addq $128+8,%rsp movups %xmm0,(%rdx) movups %xmm1,16(%rdx) movups %xmm2,32(%rdx) movups %xmm3,48(%rdx) .byte 0xf3,0xc3 .size padlock_sha512_blocks,.-padlock_sha512_blocks .globl padlock_ecb_encrypt .type padlock_ecb_encrypt,@function .align 16 padlock_ecb_encrypt: pushq %rbp pushq %rbx xorl %eax,%eax testq $15,%rdx jnz .Lecb_abort testq $15,%rcx jnz .Lecb_abort leaq .Lpadlock_saved_context(%rip),%rax pushf cld call _padlock_verify_ctx leaq 16(%rdx),%rdx xorl %eax,%eax xorl %ebx,%ebx testl $32,(%rdx) jnz .Lecb_aligned testq $0x0f,%rdi setz %al testq $0x0f,%rsi setz %bl testl %ebx,%eax jnz .Lecb_aligned negq %rax movq $512,%rbx notq %rax leaq (%rsp),%rbp cmpq %rbx,%rcx cmovcq %rcx,%rbx andq %rbx,%rax movq %rcx,%rbx negq %rax andq $512-1,%rbx leaq (%rax,%rbp,1),%rsp movq $512,%rax cmovzq %rax,%rbx cmpq %rbx,%rcx ja .Lecb_loop movq %rsi,%rax cmpq %rsp,%rbp cmoveq %rdi,%rax addq %rcx,%rax negq %rax andq $0xfff,%rax cmpq $128,%rax movq $-128,%rax cmovaeq %rbx,%rax andq %rax,%rbx jz .Lecb_unaligned_tail jmp .Lecb_loop .align 16 .Lecb_loop: cmpq %rcx,%rbx cmovaq %rcx,%rbx movq %rdi,%r8 movq %rsi,%r9 movq %rcx,%r10 movq %rbx,%rcx movq %rbx,%r11 testq $0x0f,%rdi cmovnzq %rsp,%rdi testq $0x0f,%rsi jz .Lecb_inp_aligned shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi movq %rbx,%rcx movq %rdi,%rsi .Lecb_inp_aligned: leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,200 movq %r8,%rdi movq %r11,%rbx testq $0x0f,%rdi jz .Lecb_out_aligned movq %rbx,%rcx leaq (%rsp),%rsi shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi .Lecb_out_aligned: movq %r9,%rsi movq %r10,%rcx addq %rbx,%rdi addq %rbx,%rsi subq %rbx,%rcx movq $512,%rbx jz .Lecb_break cmpq %rbx,%rcx jae .Lecb_loop .Lecb_unaligned_tail: xorl %eax,%eax cmpq %rsp,%rbp cmoveq %rcx,%rax movq %rdi,%r8 movq %rcx,%rbx subq %rax,%rsp shrq $3,%rcx leaq (%rsp),%rdi .byte 0xf3,0x48,0xa5 movq %rsp,%rsi movq %r8,%rdi movq %rbx,%rcx jmp .Lecb_loop .align 16 .Lecb_break: cmpq %rbp,%rsp je .Lecb_done pxor %xmm0,%xmm0 leaq (%rsp),%rax .Lecb_bzero: movaps %xmm0,(%rax) leaq 16(%rax),%rax cmpq %rax,%rbp ja .Lecb_bzero .Lecb_done: leaq (%rbp),%rsp jmp .Lecb_exit .align 16 .Lecb_aligned: leaq (%rsi,%rcx,1),%rbp negq %rbp andq $0xfff,%rbp xorl %eax,%eax cmpq $128,%rbp movq $128-1,%rbp cmovaeq %rax,%rbp andq %rcx,%rbp subq %rbp,%rcx jz .Lecb_aligned_tail leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,200 testq %rbp,%rbp jz .Lecb_exit .Lecb_aligned_tail: movq %rdi,%r8 movq %rbp,%rbx movq %rbp,%rcx leaq (%rsp),%rbp subq %rcx,%rsp shrq $3,%rcx leaq (%rsp),%rdi .byte 0xf3,0x48,0xa5 leaq (%r8),%rdi leaq (%rsp),%rsi movq %rbx,%rcx jmp .Lecb_loop .Lecb_exit: movl $1,%eax leaq 8(%rsp),%rsp .Lecb_abort: popq %rbx popq %rbp .byte 0xf3,0xc3 .size padlock_ecb_encrypt,.-padlock_ecb_encrypt .globl padlock_cbc_encrypt .type padlock_cbc_encrypt,@function .align 16 padlock_cbc_encrypt: pushq %rbp pushq %rbx xorl %eax,%eax testq $15,%rdx jnz .Lcbc_abort testq $15,%rcx jnz .Lcbc_abort leaq .Lpadlock_saved_context(%rip),%rax pushf cld call _padlock_verify_ctx leaq 16(%rdx),%rdx xorl %eax,%eax xorl %ebx,%ebx testl $32,(%rdx) jnz .Lcbc_aligned testq $0x0f,%rdi setz %al testq $0x0f,%rsi setz %bl testl %ebx,%eax jnz .Lcbc_aligned negq %rax movq $512,%rbx notq %rax leaq (%rsp),%rbp cmpq %rbx,%rcx cmovcq %rcx,%rbx andq %rbx,%rax movq %rcx,%rbx negq %rax andq $512-1,%rbx leaq (%rax,%rbp,1),%rsp movq $512,%rax cmovzq %rax,%rbx cmpq %rbx,%rcx ja .Lcbc_loop movq %rsi,%rax cmpq %rsp,%rbp cmoveq %rdi,%rax addq %rcx,%rax negq %rax andq $0xfff,%rax cmpq $64,%rax movq $-64,%rax cmovaeq %rbx,%rax andq %rax,%rbx jz .Lcbc_unaligned_tail jmp .Lcbc_loop .align 16 .Lcbc_loop: cmpq %rcx,%rbx cmovaq %rcx,%rbx movq %rdi,%r8 movq %rsi,%r9 movq %rcx,%r10 movq %rbx,%rcx movq %rbx,%r11 testq $0x0f,%rdi cmovnzq %rsp,%rdi testq $0x0f,%rsi jz .Lcbc_inp_aligned shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi movq %rbx,%rcx movq %rdi,%rsi .Lcbc_inp_aligned: leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,208 movdqa (%rax),%xmm0 movdqa %xmm0,-16(%rdx) movq %r8,%rdi movq %r11,%rbx testq $0x0f,%rdi jz .Lcbc_out_aligned movq %rbx,%rcx leaq (%rsp),%rsi shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi .Lcbc_out_aligned: movq %r9,%rsi movq %r10,%rcx addq %rbx,%rdi addq %rbx,%rsi subq %rbx,%rcx movq $512,%rbx jz .Lcbc_break cmpq %rbx,%rcx jae .Lcbc_loop .Lcbc_unaligned_tail: xorl %eax,%eax cmpq %rsp,%rbp cmoveq %rcx,%rax movq %rdi,%r8 movq %rcx,%rbx subq %rax,%rsp shrq $3,%rcx leaq (%rsp),%rdi .byte 0xf3,0x48,0xa5 movq %rsp,%rsi movq %r8,%rdi movq %rbx,%rcx jmp .Lcbc_loop .align 16 .Lcbc_break: cmpq %rbp,%rsp je .Lcbc_done pxor %xmm0,%xmm0 leaq (%rsp),%rax .Lcbc_bzero: movaps %xmm0,(%rax) leaq 16(%rax),%rax cmpq %rax,%rbp ja .Lcbc_bzero .Lcbc_done: leaq (%rbp),%rsp jmp .Lcbc_exit .align 16 .Lcbc_aligned: leaq (%rsi,%rcx,1),%rbp negq %rbp andq $0xfff,%rbp xorl %eax,%eax cmpq $64,%rbp movq $64-1,%rbp cmovaeq %rax,%rbp andq %rcx,%rbp subq %rbp,%rcx jz .Lcbc_aligned_tail leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,208 movdqa (%rax),%xmm0 movdqa %xmm0,-16(%rdx) testq %rbp,%rbp jz .Lcbc_exit .Lcbc_aligned_tail: movq %rdi,%r8 movq %rbp,%rbx movq %rbp,%rcx leaq (%rsp),%rbp subq %rcx,%rsp shrq $3,%rcx leaq (%rsp),%rdi .byte 0xf3,0x48,0xa5 leaq (%r8),%rdi leaq (%rsp),%rsi movq %rbx,%rcx jmp .Lcbc_loop .Lcbc_exit: movl $1,%eax leaq 8(%rsp),%rsp .Lcbc_abort: popq %rbx popq %rbp .byte 0xf3,0xc3 .size padlock_cbc_encrypt,.-padlock_cbc_encrypt .globl padlock_cfb_encrypt .type padlock_cfb_encrypt,@function .align 16 padlock_cfb_encrypt: pushq %rbp pushq %rbx xorl %eax,%eax testq $15,%rdx jnz .Lcfb_abort testq $15,%rcx jnz .Lcfb_abort leaq .Lpadlock_saved_context(%rip),%rax pushf cld call _padlock_verify_ctx leaq 16(%rdx),%rdx xorl %eax,%eax xorl %ebx,%ebx testl $32,(%rdx) jnz .Lcfb_aligned testq $0x0f,%rdi setz %al testq $0x0f,%rsi setz %bl testl %ebx,%eax jnz .Lcfb_aligned negq %rax movq $512,%rbx notq %rax leaq (%rsp),%rbp cmpq %rbx,%rcx cmovcq %rcx,%rbx andq %rbx,%rax movq %rcx,%rbx negq %rax andq $512-1,%rbx leaq (%rax,%rbp,1),%rsp movq $512,%rax cmovzq %rax,%rbx jmp .Lcfb_loop .align 16 .Lcfb_loop: cmpq %rcx,%rbx cmovaq %rcx,%rbx movq %rdi,%r8 movq %rsi,%r9 movq %rcx,%r10 movq %rbx,%rcx movq %rbx,%r11 testq $0x0f,%rdi cmovnzq %rsp,%rdi testq $0x0f,%rsi jz .Lcfb_inp_aligned shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi movq %rbx,%rcx movq %rdi,%rsi .Lcfb_inp_aligned: leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,224 movdqa (%rax),%xmm0 movdqa %xmm0,-16(%rdx) movq %r8,%rdi movq %r11,%rbx testq $0x0f,%rdi jz .Lcfb_out_aligned movq %rbx,%rcx leaq (%rsp),%rsi shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi .Lcfb_out_aligned: movq %r9,%rsi movq %r10,%rcx addq %rbx,%rdi addq %rbx,%rsi subq %rbx,%rcx movq $512,%rbx jnz .Lcfb_loop cmpq %rbp,%rsp je .Lcfb_done pxor %xmm0,%xmm0 leaq (%rsp),%rax .Lcfb_bzero: movaps %xmm0,(%rax) leaq 16(%rax),%rax cmpq %rax,%rbp ja .Lcfb_bzero .Lcfb_done: leaq (%rbp),%rsp jmp .Lcfb_exit .align 16 .Lcfb_aligned: leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,224 movdqa (%rax),%xmm0 movdqa %xmm0,-16(%rdx) .Lcfb_exit: movl $1,%eax leaq 8(%rsp),%rsp .Lcfb_abort: popq %rbx popq %rbp .byte 0xf3,0xc3 .size padlock_cfb_encrypt,.-padlock_cfb_encrypt .globl padlock_ofb_encrypt .type padlock_ofb_encrypt,@function .align 16 padlock_ofb_encrypt: pushq %rbp pushq %rbx xorl %eax,%eax testq $15,%rdx jnz .Lofb_abort testq $15,%rcx jnz .Lofb_abort leaq .Lpadlock_saved_context(%rip),%rax pushf cld call _padlock_verify_ctx leaq 16(%rdx),%rdx xorl %eax,%eax xorl %ebx,%ebx testl $32,(%rdx) jnz .Lofb_aligned testq $0x0f,%rdi setz %al testq $0x0f,%rsi setz %bl testl %ebx,%eax jnz .Lofb_aligned negq %rax movq $512,%rbx notq %rax leaq (%rsp),%rbp cmpq %rbx,%rcx cmovcq %rcx,%rbx andq %rbx,%rax movq %rcx,%rbx negq %rax andq $512-1,%rbx leaq (%rax,%rbp,1),%rsp movq $512,%rax cmovzq %rax,%rbx jmp .Lofb_loop .align 16 .Lofb_loop: cmpq %rcx,%rbx cmovaq %rcx,%rbx movq %rdi,%r8 movq %rsi,%r9 movq %rcx,%r10 movq %rbx,%rcx movq %rbx,%r11 testq $0x0f,%rdi cmovnzq %rsp,%rdi testq $0x0f,%rsi jz .Lofb_inp_aligned shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi movq %rbx,%rcx movq %rdi,%rsi .Lofb_inp_aligned: leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,232 movdqa (%rax),%xmm0 movdqa %xmm0,-16(%rdx) movq %r8,%rdi movq %r11,%rbx testq $0x0f,%rdi jz .Lofb_out_aligned movq %rbx,%rcx leaq (%rsp),%rsi shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi .Lofb_out_aligned: movq %r9,%rsi movq %r10,%rcx addq %rbx,%rdi addq %rbx,%rsi subq %rbx,%rcx movq $512,%rbx jnz .Lofb_loop cmpq %rbp,%rsp je .Lofb_done pxor %xmm0,%xmm0 leaq (%rsp),%rax .Lofb_bzero: movaps %xmm0,(%rax) leaq 16(%rax),%rax cmpq %rax,%rbp ja .Lofb_bzero .Lofb_done: leaq (%rbp),%rsp jmp .Lofb_exit .align 16 .Lofb_aligned: leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,232 movdqa (%rax),%xmm0 movdqa %xmm0,-16(%rdx) .Lofb_exit: movl $1,%eax leaq 8(%rsp),%rsp .Lofb_abort: popq %rbx popq %rbp .byte 0xf3,0xc3 .size padlock_ofb_encrypt,.-padlock_ofb_encrypt .globl padlock_ctr32_encrypt .type padlock_ctr32_encrypt,@function .align 16 padlock_ctr32_encrypt: pushq %rbp pushq %rbx xorl %eax,%eax testq $15,%rdx jnz .Lctr32_abort testq $15,%rcx jnz .Lctr32_abort leaq .Lpadlock_saved_context(%rip),%rax pushf cld call _padlock_verify_ctx leaq 16(%rdx),%rdx xorl %eax,%eax xorl %ebx,%ebx testl $32,(%rdx) jnz .Lctr32_aligned testq $0x0f,%rdi setz %al testq $0x0f,%rsi setz %bl testl %ebx,%eax jnz .Lctr32_aligned negq %rax movq $512,%rbx notq %rax leaq (%rsp),%rbp cmpq %rbx,%rcx cmovcq %rcx,%rbx andq %rbx,%rax movq %rcx,%rbx negq %rax andq $512-1,%rbx leaq (%rax,%rbp,1),%rsp movq $512,%rax cmovzq %rax,%rbx .Lctr32_reenter: movl -4(%rdx),%eax bswapl %eax negl %eax andl $31,%eax movq $512,%rbx shll $4,%eax cmovzq %rbx,%rax cmpq %rax,%rcx cmovaq %rax,%rbx cmovbeq %rcx,%rbx cmpq %rbx,%rcx ja .Lctr32_loop movq %rsi,%rax cmpq %rsp,%rbp cmoveq %rdi,%rax addq %rcx,%rax negq %rax andq $0xfff,%rax cmpq $32,%rax movq $-32,%rax cmovaeq %rbx,%rax andq %rax,%rbx jz .Lctr32_unaligned_tail jmp .Lctr32_loop .align 16 .Lctr32_loop: cmpq %rcx,%rbx cmovaq %rcx,%rbx movq %rdi,%r8 movq %rsi,%r9 movq %rcx,%r10 movq %rbx,%rcx movq %rbx,%r11 testq $0x0f,%rdi cmovnzq %rsp,%rdi testq $0x0f,%rsi jz .Lctr32_inp_aligned shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi movq %rbx,%rcx movq %rdi,%rsi .Lctr32_inp_aligned: leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,216 movl -4(%rdx),%eax testl $0xffff0000,%eax jnz .Lctr32_no_carry bswapl %eax addl $0x10000,%eax bswapl %eax movl %eax,-4(%rdx) .Lctr32_no_carry: movq %r8,%rdi movq %r11,%rbx testq $0x0f,%rdi jz .Lctr32_out_aligned movq %rbx,%rcx leaq (%rsp),%rsi shrq $3,%rcx .byte 0xf3,0x48,0xa5 subq %rbx,%rdi .Lctr32_out_aligned: movq %r9,%rsi movq %r10,%rcx addq %rbx,%rdi addq %rbx,%rsi subq %rbx,%rcx movq $512,%rbx jz .Lctr32_break cmpq %rbx,%rcx jae .Lctr32_loop movq %rcx,%rbx movq %rsi,%rax cmpq %rsp,%rbp cmoveq %rdi,%rax addq %rcx,%rax negq %rax andq $0xfff,%rax cmpq $32,%rax movq $-32,%rax cmovaeq %rbx,%rax andq %rax,%rbx jnz .Lctr32_loop .Lctr32_unaligned_tail: xorl %eax,%eax cmpq %rsp,%rbp cmoveq %rcx,%rax movq %rdi,%r8 movq %rcx,%rbx subq %rax,%rsp shrq $3,%rcx leaq (%rsp),%rdi .byte 0xf3,0x48,0xa5 movq %rsp,%rsi movq %r8,%rdi movq %rbx,%rcx jmp .Lctr32_loop .align 16 .Lctr32_break: cmpq %rbp,%rsp je .Lctr32_done pxor %xmm0,%xmm0 leaq (%rsp),%rax .Lctr32_bzero: movaps %xmm0,(%rax) leaq 16(%rax),%rax cmpq %rax,%rbp ja .Lctr32_bzero .Lctr32_done: leaq (%rbp),%rsp jmp .Lctr32_exit .align 16 .Lctr32_aligned: movl -4(%rdx),%eax bswapl %eax negl %eax andl $0xffff,%eax movq $1048576,%rbx shll $4,%eax cmovzq %rbx,%rax cmpq %rax,%rcx cmovaq %rax,%rbx cmovbeq %rcx,%rbx jbe .Lctr32_aligned_skip .Lctr32_aligned_loop: movq %rcx,%r10 movq %rbx,%rcx movq %rbx,%r11 leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,216 movl -4(%rdx),%eax bswapl %eax addl $0x10000,%eax bswapl %eax movl %eax,-4(%rdx) movq %r10,%rcx subq %r11,%rcx movq $1048576,%rbx jz .Lctr32_exit cmpq %rbx,%rcx jae .Lctr32_aligned_loop .Lctr32_aligned_skip: leaq (%rsi,%rcx,1),%rbp negq %rbp andq $0xfff,%rbp xorl %eax,%eax cmpq $32,%rbp movq $32-1,%rbp cmovaeq %rax,%rbp andq %rcx,%rbp subq %rbp,%rcx jz .Lctr32_aligned_tail leaq -16(%rdx),%rax leaq 16(%rdx),%rbx shrq $4,%rcx .byte 0xf3,0x0f,0xa7,216 testq %rbp,%rbp jz .Lctr32_exit .Lctr32_aligned_tail: movq %rdi,%r8 movq %rbp,%rbx movq %rbp,%rcx leaq (%rsp),%rbp subq %rcx,%rsp shrq $3,%rcx leaq (%rsp),%rdi .byte 0xf3,0x48,0xa5 leaq (%r8),%rdi leaq (%rsp),%rsi movq %rbx,%rcx jmp .Lctr32_loop .Lctr32_exit: movl $1,%eax leaq 8(%rsp),%rsp .Lctr32_abort: popq %rbx popq %rbp .byte 0xf3,0xc3 .size padlock_ctr32_encrypt,.-padlock_ctr32_encrypt .byte 86,73,65,32,80,97,100,108,111,99,107,32,120,56,54,95,54,52,32,109,111,100,117,108,101,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 .align 16 .data .align 8 .Lpadlock_saved_context: .quad 0 .section ".note.gnu.property", "a" .p2align 3 .long 1f - 0f .long 4f - 1f .long 5 0: # "GNU" encoded with .byte, since .asciz isn't supported # on Solaris. .byte 0x47 .byte 0x4e .byte 0x55 .byte 0 1: .p2align 3 .long 0xc0000002 .long 3f - 2f 2: .long 3 3: .p2align 3 4: