[edk2-devel] [PATCH v7 3/3] CryptoPkg/OpensslLib: Commit the auto-generated assembly files for X64
Yao, Jiewen
jiewen.yao at intel.com
Wed Jul 21 11:44:47 UTC 2021
Reviewed-by: Jiewen Yao <Jiewen.yao at intel.com>
> -----Original Message-----
> From: christopher.zurcher at outlook.com <christopher.zurcher at outlook.com>
> Sent: Wednesday, July 21, 2021 6:07 AM
> To: devel at edk2.groups.io
> Cc: Yao, Jiewen <jiewen.yao at intel.com>; Wang, Jian J <jian.j.wang at intel.com>;
> Lu, XiaoyuX <xiaoyux.lu at intel.com>; Kinney, Michael D
> <michael.d.kinney at intel.com>; Ard Biesheuvel <ardb at kernel.org>
> Subject: [PATCH v7 3/3] CryptoPkg/OpensslLib: Commit the auto-generated
> assembly files for X64
>
> From: Christopher Zurcher <christopher.zurcher at microsoft.com>
>
> BZ: https://bugzilla.tianocore.org/show_bug.cgi?id=2507
>
> Adding the auto-generated assembly files for X64 architectures.
>
> Cc: Jiewen Yao <jiewen.yao at intel.com>
> Cc: Jian J Wang <jian.j.wang at intel.com>
> Cc: Xiaoyu Lu <xiaoyux.lu at intel.com>
> Cc: Mike Kinney <michael.d.kinney at intel.com>
> Cc: Ard Biesheuvel <ardb at kernel.org>
> Signed-off-by: Christopher Zurcher <christopher.zurcher at microsoft.com>
> ---
> CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-mb-x86_64.nasm | 732
> +++
> CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-sha1-x86_64.nasm |
> 1916 ++++++++
> CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-sha256-x86_64.nasm |
> 78 +
> CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-x86_64.nasm | 5103
> ++++++++++++++++++++
> CryptoPkg/Library/OpensslLib/X64/crypto/aes/vpaes-x86_64.nasm | 1173
> +++++
> CryptoPkg/Library/OpensslLib/X64/crypto/modes/aesni-gcm-x86_64.nasm |
> 34 +
> CryptoPkg/Library/OpensslLib/X64/crypto/modes/ghash-x86_64.nasm |
> 1569 ++++++
> CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha1-mb-x86_64.nasm | 3137
> ++++++++++++
> CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha1-x86_64.nasm | 2884
> +++++++++++
> CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha256-mb-x86_64.nasm |
> 3461 +++++++++++++
> CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha256-x86_64.nasm | 3313
> +++++++++++++
> CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha512-x86_64.nasm | 1938
> ++++++++
> CryptoPkg/Library/OpensslLib/X64/crypto/x86_64cpuid.nasm | 491 ++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-mb-x86_64.S | 552
> +++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-sha1-x86_64.S | 1719
> +++++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-sha256-x86_64.S | 69
> +
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-x86_64.S | 4484
> +++++++++++++++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/vpaes-x86_64.S | 863
> ++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/modes/aesni-gcm-x86_64.S |
> 29 +
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/modes/ghash-x86_64.S | 1386
> ++++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha1-mb-x86_64.S | 2962
> ++++++++++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha1-x86_64.S | 2631
> ++++++++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha256-mb-x86_64.S |
> 3286 +++++++++++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha256-x86_64.S | 3097
> ++++++++++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha512-x86_64.S | 1811
> +++++++
> CryptoPkg/Library/OpensslLib/X64Gcc/crypto/x86_64cpuid.S | 491 ++
> 26 files changed, 49209 insertions(+)
>
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-mb-
> x86_64.nasm b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-mb-
> x86_64.nasm
> new file mode 100644
> index 0000000000..1a3ed1dd35
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-mb-x86_64.nasm
> @@ -0,0 +1,732 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/aes/asm/aesni-mb-x86_64.pl
> +;
> +; Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +
> +EXTERN OPENSSL_ia32cap_P
> +
> +global aesni_multi_cbc_encrypt
> +
> +ALIGN 32
> +aesni_multi_cbc_encrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_multi_cbc_encrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + push r15
> +
> + lea rsp,[((-168))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> + movaps XMMWORD[64+rsp],xmm10
> + movaps XMMWORD[80+rsp],xmm11
> + movaps XMMWORD[96+rsp],xmm12
> + movaps XMMWORD[(-104)+rax],xmm13
> + movaps XMMWORD[(-88)+rax],xmm14
> + movaps XMMWORD[(-72)+rax],xmm15
> +
> +
> +
> +
> +
> +
> + sub rsp,48
> + and rsp,-64
> + mov QWORD[16+rsp],rax
> +
> +
> +$L$enc4x_body:
> + movdqu xmm12,XMMWORD[rsi]
> + lea rsi,[120+rsi]
> + lea rdi,[80+rdi]
> +
> +$L$enc4x_loop_grande:
> + mov DWORD[24+rsp],edx
> + xor edx,edx
> + mov ecx,DWORD[((-64))+rdi]
> + mov r8,QWORD[((-80))+rdi]
> + cmp ecx,edx
> + mov r12,QWORD[((-72))+rdi]
> + cmovg edx,ecx
> + test ecx,ecx
> + movdqu xmm2,XMMWORD[((-56))+rdi]
> + mov DWORD[32+rsp],ecx
> + cmovle r8,rsp
> + mov ecx,DWORD[((-24))+rdi]
> + mov r9,QWORD[((-40))+rdi]
> + cmp ecx,edx
> + mov r13,QWORD[((-32))+rdi]
> + cmovg edx,ecx
> + test ecx,ecx
> + movdqu xmm3,XMMWORD[((-16))+rdi]
> + mov DWORD[36+rsp],ecx
> + cmovle r9,rsp
> + mov ecx,DWORD[16+rdi]
> + mov r10,QWORD[rdi]
> + cmp ecx,edx
> + mov r14,QWORD[8+rdi]
> + cmovg edx,ecx
> + test ecx,ecx
> + movdqu xmm4,XMMWORD[24+rdi]
> + mov DWORD[40+rsp],ecx
> + cmovle r10,rsp
> + mov ecx,DWORD[56+rdi]
> + mov r11,QWORD[40+rdi]
> + cmp ecx,edx
> + mov r15,QWORD[48+rdi]
> + cmovg edx,ecx
> + test ecx,ecx
> + movdqu xmm5,XMMWORD[64+rdi]
> + mov DWORD[44+rsp],ecx
> + cmovle r11,rsp
> + test edx,edx
> + jz NEAR $L$enc4x_done
> +
> + movups xmm1,XMMWORD[((16-120))+rsi]
> + pxor xmm2,xmm12
> + movups xmm0,XMMWORD[((32-120))+rsi]
> + pxor xmm3,xmm12
> + mov eax,DWORD[((240-120))+rsi]
> + pxor xmm4,xmm12
> + movdqu xmm6,XMMWORD[r8]
> + pxor xmm5,xmm12
> + movdqu xmm7,XMMWORD[r9]
> + pxor xmm2,xmm6
> + movdqu xmm8,XMMWORD[r10]
> + pxor xmm3,xmm7
> + movdqu xmm9,XMMWORD[r11]
> + pxor xmm4,xmm8
> + pxor xmm5,xmm9
> + movdqa xmm10,XMMWORD[32+rsp]
> + xor rbx,rbx
> + jmp NEAR $L$oop_enc4x
> +
> +ALIGN 32
> +$L$oop_enc4x:
> + add rbx,16
> + lea rbp,[16+rsp]
> + mov ecx,1
> + sub rbp,rbx
> +
> +DB 102,15,56,220,209
> + prefetcht0 [31+rbx*1+r8]
> + prefetcht0 [31+rbx*1+r9]
> +DB 102,15,56,220,217
> + prefetcht0 [31+rbx*1+r10]
> + prefetcht0 [31+rbx*1+r10]
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[((48-120))+rsi]
> + cmp ecx,DWORD[32+rsp]
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> + cmovge r8,rbp
> + cmovg r12,rbp
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[((-56))+rsi]
> + cmp ecx,DWORD[36+rsp]
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> + cmovge r9,rbp
> + cmovg r13,rbp
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[((-40))+rsi]
> + cmp ecx,DWORD[40+rsp]
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> + cmovge r10,rbp
> + cmovg r14,rbp
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[((-24))+rsi]
> + cmp ecx,DWORD[44+rsp]
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> + cmovge r11,rbp
> + cmovg r15,rbp
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[((-8))+rsi]
> + movdqa xmm11,xmm10
> +DB 102,15,56,220,208
> + prefetcht0 [15+rbx*1+r12]
> + prefetcht0 [15+rbx*1+r13]
> +DB 102,15,56,220,216
> + prefetcht0 [15+rbx*1+r14]
> + prefetcht0 [15+rbx*1+r15]
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[((128-120))+rsi]
> + pxor xmm12,xmm12
> +
> +DB 102,15,56,220,209
> + pcmpgtd xmm11,xmm12
> + movdqu xmm12,XMMWORD[((-120))+rsi]
> +DB 102,15,56,220,217
> + paddd xmm10,xmm11
> + movdqa XMMWORD[32+rsp],xmm10
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[((144-120))+rsi]
> +
> + cmp eax,11
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[((160-120))+rsi]
> +
> + jb NEAR $L$enc4x_tail
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[((176-120))+rsi]
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[((192-120))+rsi]
> +
> + je NEAR $L$enc4x_tail
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[((208-120))+rsi]
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[((224-120))+rsi]
> + jmp NEAR $L$enc4x_tail
> +
> +ALIGN 32
> +$L$enc4x_tail:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movdqu xmm6,XMMWORD[rbx*1+r8]
> + movdqu xmm1,XMMWORD[((16-120))+rsi]
> +
> +DB 102,15,56,221,208
> + movdqu xmm7,XMMWORD[rbx*1+r9]
> + pxor xmm6,xmm12
> +DB 102,15,56,221,216
> + movdqu xmm8,XMMWORD[rbx*1+r10]
> + pxor xmm7,xmm12
> +DB 102,15,56,221,224
> + movdqu xmm9,XMMWORD[rbx*1+r11]
> + pxor xmm8,xmm12
> +DB 102,15,56,221,232
> + movdqu xmm0,XMMWORD[((32-120))+rsi]
> + pxor xmm9,xmm12
> +
> + movups XMMWORD[(-16)+rbx*1+r12],xmm2
> + pxor xmm2,xmm6
> + movups XMMWORD[(-16)+rbx*1+r13],xmm3
> + pxor xmm3,xmm7
> + movups XMMWORD[(-16)+rbx*1+r14],xmm4
> + pxor xmm4,xmm8
> + movups XMMWORD[(-16)+rbx*1+r15],xmm5
> + pxor xmm5,xmm9
> +
> + dec edx
> + jnz NEAR $L$oop_enc4x
> +
> + mov rax,QWORD[16+rsp]
> +
> + mov edx,DWORD[24+rsp]
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> + lea rdi,[160+rdi]
> + dec edx
> + jnz NEAR $L$enc4x_loop_grande
> +
> +$L$enc4x_done:
> + movaps xmm6,XMMWORD[((-216))+rax]
> + movaps xmm7,XMMWORD[((-200))+rax]
> + movaps xmm8,XMMWORD[((-184))+rax]
> + movaps xmm9,XMMWORD[((-168))+rax]
> + movaps xmm10,XMMWORD[((-152))+rax]
> + movaps xmm11,XMMWORD[((-136))+rax]
> + movaps xmm12,XMMWORD[((-120))+rax]
> +
> +
> +
> + mov r15,QWORD[((-48))+rax]
> +
> + mov r14,QWORD[((-40))+rax]
> +
> + mov r13,QWORD[((-32))+rax]
> +
> + mov r12,QWORD[((-24))+rax]
> +
> + mov rbp,QWORD[((-16))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> +
> + lea rsp,[rax]
> +
> +$L$enc4x_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_multi_cbc_encrypt:
> +
> +global aesni_multi_cbc_decrypt
> +
> +ALIGN 32
> +aesni_multi_cbc_decrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_multi_cbc_decrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + push r15
> +
> + lea rsp,[((-168))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> + movaps XMMWORD[64+rsp],xmm10
> + movaps XMMWORD[80+rsp],xmm11
> + movaps XMMWORD[96+rsp],xmm12
> + movaps XMMWORD[(-104)+rax],xmm13
> + movaps XMMWORD[(-88)+rax],xmm14
> + movaps XMMWORD[(-72)+rax],xmm15
> +
> +
> +
> +
> +
> +
> + sub rsp,48
> + and rsp,-64
> + mov QWORD[16+rsp],rax
> +
> +
> +$L$dec4x_body:
> + movdqu xmm12,XMMWORD[rsi]
> + lea rsi,[120+rsi]
> + lea rdi,[80+rdi]
> +
> +$L$dec4x_loop_grande:
> + mov DWORD[24+rsp],edx
> + xor edx,edx
> + mov ecx,DWORD[((-64))+rdi]
> + mov r8,QWORD[((-80))+rdi]
> + cmp ecx,edx
> + mov r12,QWORD[((-72))+rdi]
> + cmovg edx,ecx
> + test ecx,ecx
> + movdqu xmm6,XMMWORD[((-56))+rdi]
> + mov DWORD[32+rsp],ecx
> + cmovle r8,rsp
> + mov ecx,DWORD[((-24))+rdi]
> + mov r9,QWORD[((-40))+rdi]
> + cmp ecx,edx
> + mov r13,QWORD[((-32))+rdi]
> + cmovg edx,ecx
> + test ecx,ecx
> + movdqu xmm7,XMMWORD[((-16))+rdi]
> + mov DWORD[36+rsp],ecx
> + cmovle r9,rsp
> + mov ecx,DWORD[16+rdi]
> + mov r10,QWORD[rdi]
> + cmp ecx,edx
> + mov r14,QWORD[8+rdi]
> + cmovg edx,ecx
> + test ecx,ecx
> + movdqu xmm8,XMMWORD[24+rdi]
> + mov DWORD[40+rsp],ecx
> + cmovle r10,rsp
> + mov ecx,DWORD[56+rdi]
> + mov r11,QWORD[40+rdi]
> + cmp ecx,edx
> + mov r15,QWORD[48+rdi]
> + cmovg edx,ecx
> + test ecx,ecx
> + movdqu xmm9,XMMWORD[64+rdi]
> + mov DWORD[44+rsp],ecx
> + cmovle r11,rsp
> + test edx,edx
> + jz NEAR $L$dec4x_done
> +
> + movups xmm1,XMMWORD[((16-120))+rsi]
> + movups xmm0,XMMWORD[((32-120))+rsi]
> + mov eax,DWORD[((240-120))+rsi]
> + movdqu xmm2,XMMWORD[r8]
> + movdqu xmm3,XMMWORD[r9]
> + pxor xmm2,xmm12
> + movdqu xmm4,XMMWORD[r10]
> + pxor xmm3,xmm12
> + movdqu xmm5,XMMWORD[r11]
> + pxor xmm4,xmm12
> + pxor xmm5,xmm12
> + movdqa xmm10,XMMWORD[32+rsp]
> + xor rbx,rbx
> + jmp NEAR $L$oop_dec4x
> +
> +ALIGN 32
> +$L$oop_dec4x:
> + add rbx,16
> + lea rbp,[16+rsp]
> + mov ecx,1
> + sub rbp,rbx
> +
> +DB 102,15,56,222,209
> + prefetcht0 [31+rbx*1+r8]
> + prefetcht0 [31+rbx*1+r9]
> +DB 102,15,56,222,217
> + prefetcht0 [31+rbx*1+r10]
> + prefetcht0 [31+rbx*1+r11]
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[((48-120))+rsi]
> + cmp ecx,DWORD[32+rsp]
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> + cmovge r8,rbp
> + cmovg r12,rbp
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[((-56))+rsi]
> + cmp ecx,DWORD[36+rsp]
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> + cmovge r9,rbp
> + cmovg r13,rbp
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[((-40))+rsi]
> + cmp ecx,DWORD[40+rsp]
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> + cmovge r10,rbp
> + cmovg r14,rbp
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[((-24))+rsi]
> + cmp ecx,DWORD[44+rsp]
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> + cmovge r11,rbp
> + cmovg r15,rbp
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[((-8))+rsi]
> + movdqa xmm11,xmm10
> +DB 102,15,56,222,208
> + prefetcht0 [15+rbx*1+r12]
> + prefetcht0 [15+rbx*1+r13]
> +DB 102,15,56,222,216
> + prefetcht0 [15+rbx*1+r14]
> + prefetcht0 [15+rbx*1+r15]
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[((128-120))+rsi]
> + pxor xmm12,xmm12
> +
> +DB 102,15,56,222,209
> + pcmpgtd xmm11,xmm12
> + movdqu xmm12,XMMWORD[((-120))+rsi]
> +DB 102,15,56,222,217
> + paddd xmm10,xmm11
> + movdqa XMMWORD[32+rsp],xmm10
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[((144-120))+rsi]
> +
> + cmp eax,11
> +
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[((160-120))+rsi]
> +
> + jb NEAR $L$dec4x_tail
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[((176-120))+rsi]
> +
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[((192-120))+rsi]
> +
> + je NEAR $L$dec4x_tail
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[((208-120))+rsi]
> +
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[((224-120))+rsi]
> + jmp NEAR $L$dec4x_tail
> +
> +ALIGN 32
> +$L$dec4x_tail:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> + pxor xmm6,xmm0
> + pxor xmm7,xmm0
> +DB 102,15,56,222,233
> + movdqu xmm1,XMMWORD[((16-120))+rsi]
> + pxor xmm8,xmm0
> + pxor xmm9,xmm0
> + movdqu xmm0,XMMWORD[((32-120))+rsi]
> +
> +DB 102,15,56,223,214
> +DB 102,15,56,223,223
> + movdqu xmm6,XMMWORD[((-16))+rbx*1+r8]
> + movdqu xmm7,XMMWORD[((-16))+rbx*1+r9]
> +DB 102,65,15,56,223,224
> +DB 102,65,15,56,223,233
> + movdqu xmm8,XMMWORD[((-16))+rbx*1+r10]
> + movdqu xmm9,XMMWORD[((-16))+rbx*1+r11]
> +
> + movups XMMWORD[(-16)+rbx*1+r12],xmm2
> + movdqu xmm2,XMMWORD[rbx*1+r8]
> + movups XMMWORD[(-16)+rbx*1+r13],xmm3
> + movdqu xmm3,XMMWORD[rbx*1+r9]
> + pxor xmm2,xmm12
> + movups XMMWORD[(-16)+rbx*1+r14],xmm4
> + movdqu xmm4,XMMWORD[rbx*1+r10]
> + pxor xmm3,xmm12
> + movups XMMWORD[(-16)+rbx*1+r15],xmm5
> + movdqu xmm5,XMMWORD[rbx*1+r11]
> + pxor xmm4,xmm12
> + pxor xmm5,xmm12
> +
> + dec edx
> + jnz NEAR $L$oop_dec4x
> +
> + mov rax,QWORD[16+rsp]
> +
> + mov edx,DWORD[24+rsp]
> +
> + lea rdi,[160+rdi]
> + dec edx
> + jnz NEAR $L$dec4x_loop_grande
> +
> +$L$dec4x_done:
> + movaps xmm6,XMMWORD[((-216))+rax]
> + movaps xmm7,XMMWORD[((-200))+rax]
> + movaps xmm8,XMMWORD[((-184))+rax]
> + movaps xmm9,XMMWORD[((-168))+rax]
> + movaps xmm10,XMMWORD[((-152))+rax]
> + movaps xmm11,XMMWORD[((-136))+rax]
> + movaps xmm12,XMMWORD[((-120))+rax]
> +
> +
> +
> + mov r15,QWORD[((-48))+rax]
> +
> + mov r14,QWORD[((-40))+rax]
> +
> + mov r13,QWORD[((-32))+rax]
> +
> + mov r12,QWORD[((-24))+rax]
> +
> + mov rbp,QWORD[((-16))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> +
> + lea rsp,[rax]
> +
> +$L$dec4x_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_multi_cbc_decrypt:
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$in_prologue
> +
> + mov rax,QWORD[16+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov r12,QWORD[((-24))+rax]
> + mov r13,QWORD[((-32))+rax]
> + mov r14,QWORD[((-40))+rax]
> + mov r15,QWORD[((-48))+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> + mov QWORD[216+r8],r12
> + mov QWORD[224+r8],r13
> + mov QWORD[232+r8],r14
> + mov QWORD[240+r8],r15
> +
> + lea rsi,[((-56-160))+rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> +
> +$L$in_prologue:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_aesni_multi_cbc_encrypt wrt ..imagebase
> + DD $L$SEH_end_aesni_multi_cbc_encrypt wrt ..imagebase
> + DD $L$SEH_info_aesni_multi_cbc_encrypt wrt ..imagebase
> + DD $L$SEH_begin_aesni_multi_cbc_decrypt wrt ..imagebase
> + DD $L$SEH_end_aesni_multi_cbc_decrypt wrt ..imagebase
> + DD $L$SEH_info_aesni_multi_cbc_decrypt wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_aesni_multi_cbc_encrypt:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$enc4x_body wrt ..imagebase,$L$enc4x_epilogue
> wrt ..imagebase
> +$L$SEH_info_aesni_multi_cbc_decrypt:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$dec4x_body wrt ..imagebase,$L$dec4x_epilogue
> wrt ..imagebase
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-sha1-
> x86_64.nasm b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-sha1-
> x86_64.nasm
> new file mode 100644
> index 0000000000..f4fd9ca50d
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-sha1-x86_64.nasm
> @@ -0,0 +1,1916 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/aes/asm/aesni-sha1-x86_64.pl
> +;
> +; Copyright 2011-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +EXTERN OPENSSL_ia32cap_P
> +
> +global aesni_cbc_sha1_enc
> +
> +ALIGN 32
> +aesni_cbc_sha1_enc:
> +
> +
> + mov r10d,DWORD[((OPENSSL_ia32cap_P+0))]
> + mov r11,QWORD[((OPENSSL_ia32cap_P+4))]
> + bt r11,61
> + jc NEAR aesni_cbc_sha1_enc_shaext
> + jmp NEAR aesni_cbc_sha1_enc_ssse3
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 32
> +aesni_cbc_sha1_enc_ssse3:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_cbc_sha1_enc_ssse3:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + mov r10,QWORD[56+rsp]
> +
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + push r15
> +
> + lea rsp,[((-264))+rsp]
> +
> +
> +
> + movaps XMMWORD[(96+0)+rsp],xmm6
> + movaps XMMWORD[(96+16)+rsp],xmm7
> + movaps XMMWORD[(96+32)+rsp],xmm8
> + movaps XMMWORD[(96+48)+rsp],xmm9
> + movaps XMMWORD[(96+64)+rsp],xmm10
> + movaps XMMWORD[(96+80)+rsp],xmm11
> + movaps XMMWORD[(96+96)+rsp],xmm12
> + movaps XMMWORD[(96+112)+rsp],xmm13
> + movaps XMMWORD[(96+128)+rsp],xmm14
> + movaps XMMWORD[(96+144)+rsp],xmm15
> +$L$prologue_ssse3:
> + mov r12,rdi
> + mov r13,rsi
> + mov r14,rdx
> + lea r15,[112+rcx]
> + movdqu xmm2,XMMWORD[r8]
> + mov QWORD[88+rsp],r8
> + shl r14,6
> + sub r13,r12
> + mov r8d,DWORD[((240-112))+r15]
> + add r14,r10
> +
> + lea r11,[K_XX_XX]
> + mov eax,DWORD[r9]
> + mov ebx,DWORD[4+r9]
> + mov ecx,DWORD[8+r9]
> + mov edx,DWORD[12+r9]
> + mov esi,ebx
> + mov ebp,DWORD[16+r9]
> + mov edi,ecx
> + xor edi,edx
> + and esi,edi
> +
> + movdqa xmm3,XMMWORD[64+r11]
> + movdqa xmm13,XMMWORD[r11]
> + movdqu xmm4,XMMWORD[r10]
> + movdqu xmm5,XMMWORD[16+r10]
> + movdqu xmm6,XMMWORD[32+r10]
> + movdqu xmm7,XMMWORD[48+r10]
> +DB 102,15,56,0,227
> +DB 102,15,56,0,235
> +DB 102,15,56,0,243
> + add r10,64
> + paddd xmm4,xmm13
> +DB 102,15,56,0,251
> + paddd xmm5,xmm13
> + paddd xmm6,xmm13
> + movdqa XMMWORD[rsp],xmm4
> + psubd xmm4,xmm13
> + movdqa XMMWORD[16+rsp],xmm5
> + psubd xmm5,xmm13
> + movdqa XMMWORD[32+rsp],xmm6
> + psubd xmm6,xmm13
> + movups xmm15,XMMWORD[((-112))+r15]
> + movups xmm0,XMMWORD[((16-112))+r15]
> + jmp NEAR $L$oop_ssse3
> +ALIGN 32
> +$L$oop_ssse3:
> + ror ebx,2
> + movups xmm14,XMMWORD[r12]
> + xorps xmm14,xmm15
> + xorps xmm2,xmm14
> + movups xmm1,XMMWORD[((-80))+r15]
> +DB 102,15,56,220,208
> + pshufd xmm8,xmm4,238
> + xor esi,edx
> + movdqa xmm12,xmm7
> + paddd xmm13,xmm7
> + mov edi,eax
> + add ebp,DWORD[rsp]
> + punpcklqdq xmm8,xmm5
> + xor ebx,ecx
> + rol eax,5
> + add ebp,esi
> + psrldq xmm12,4
> + and edi,ebx
> + xor ebx,ecx
> + pxor xmm8,xmm4
> + add ebp,eax
> + ror eax,7
> + pxor xmm12,xmm6
> + xor edi,ecx
> + mov esi,ebp
> + add edx,DWORD[4+rsp]
> + pxor xmm8,xmm12
> + xor eax,ebx
> + rol ebp,5
> + movdqa XMMWORD[48+rsp],xmm13
> + add edx,edi
> + movups xmm0,XMMWORD[((-64))+r15]
> +DB 102,15,56,220,209
> + and esi,eax
> + movdqa xmm3,xmm8
> + xor eax,ebx
> + add edx,ebp
> + ror ebp,7
> + movdqa xmm12,xmm8
> + xor esi,ebx
> + pslldq xmm3,12
> + paddd xmm8,xmm8
> + mov edi,edx
> + add ecx,DWORD[8+rsp]
> + psrld xmm12,31
> + xor ebp,eax
> + rol edx,5
> + add ecx,esi
> + movdqa xmm13,xmm3
> + and edi,ebp
> + xor ebp,eax
> + psrld xmm3,30
> + add ecx,edx
> + ror edx,7
> + por xmm8,xmm12
> + xor edi,eax
> + mov esi,ecx
> + add ebx,DWORD[12+rsp]
> + movups xmm1,XMMWORD[((-48))+r15]
> +DB 102,15,56,220,208
> + pslld xmm13,2
> + pxor xmm8,xmm3
> + xor edx,ebp
> + movdqa xmm3,XMMWORD[r11]
> + rol ecx,5
> + add ebx,edi
> + and esi,edx
> + pxor xmm8,xmm13
> + xor edx,ebp
> + add ebx,ecx
> + ror ecx,7
> + pshufd xmm9,xmm5,238
> + xor esi,ebp
> + movdqa xmm13,xmm8
> + paddd xmm3,xmm8
> + mov edi,ebx
> + add eax,DWORD[16+rsp]
> + punpcklqdq xmm9,xmm6
> + xor ecx,edx
> + rol ebx,5
> + add eax,esi
> + psrldq xmm13,4
> + and edi,ecx
> + xor ecx,edx
> + pxor xmm9,xmm5
> + add eax,ebx
> + ror ebx,7
> + movups xmm0,XMMWORD[((-32))+r15]
> +DB 102,15,56,220,209
> + pxor xmm13,xmm7
> + xor edi,edx
> + mov esi,eax
> + add ebp,DWORD[20+rsp]
> + pxor xmm9,xmm13
> + xor ebx,ecx
> + rol eax,5
> + movdqa XMMWORD[rsp],xmm3
> + add ebp,edi
> + and esi,ebx
> + movdqa xmm12,xmm9
> + xor ebx,ecx
> + add ebp,eax
> + ror eax,7
> + movdqa xmm13,xmm9
> + xor esi,ecx
> + pslldq xmm12,12
> + paddd xmm9,xmm9
> + mov edi,ebp
> + add edx,DWORD[24+rsp]
> + psrld xmm13,31
> + xor eax,ebx
> + rol ebp,5
> + add edx,esi
> + movups xmm1,XMMWORD[((-16))+r15]
> +DB 102,15,56,220,208
> + movdqa xmm3,xmm12
> + and edi,eax
> + xor eax,ebx
> + psrld xmm12,30
> + add edx,ebp
> + ror ebp,7
> + por xmm9,xmm13
> + xor edi,ebx
> + mov esi,edx
> + add ecx,DWORD[28+rsp]
> + pslld xmm3,2
> + pxor xmm9,xmm12
> + xor ebp,eax
> + movdqa xmm12,XMMWORD[16+r11]
> + rol edx,5
> + add ecx,edi
> + and esi,ebp
> + pxor xmm9,xmm3
> + xor ebp,eax
> + add ecx,edx
> + ror edx,7
> + pshufd xmm10,xmm6,238
> + xor esi,eax
> + movdqa xmm3,xmm9
> + paddd xmm12,xmm9
> + mov edi,ecx
> + add ebx,DWORD[32+rsp]
> + movups xmm0,XMMWORD[r15]
> +DB 102,15,56,220,209
> + punpcklqdq xmm10,xmm7
> + xor edx,ebp
> + rol ecx,5
> + add ebx,esi
> + psrldq xmm3,4
> + and edi,edx
> + xor edx,ebp
> + pxor xmm10,xmm6
> + add ebx,ecx
> + ror ecx,7
> + pxor xmm3,xmm8
> + xor edi,ebp
> + mov esi,ebx
> + add eax,DWORD[36+rsp]
> + pxor xmm10,xmm3
> + xor ecx,edx
> + rol ebx,5
> + movdqa XMMWORD[16+rsp],xmm12
> + add eax,edi
> + and esi,ecx
> + movdqa xmm13,xmm10
> + xor ecx,edx
> + add eax,ebx
> + ror ebx,7
> + movups xmm1,XMMWORD[16+r15]
> +DB 102,15,56,220,208
> + movdqa xmm3,xmm10
> + xor esi,edx
> + pslldq xmm13,12
> + paddd xmm10,xmm10
> + mov edi,eax
> + add ebp,DWORD[40+rsp]
> + psrld xmm3,31
> + xor ebx,ecx
> + rol eax,5
> + add ebp,esi
> + movdqa xmm12,xmm13
> + and edi,ebx
> + xor ebx,ecx
> + psrld xmm13,30
> + add ebp,eax
> + ror eax,7
> + por xmm10,xmm3
> + xor edi,ecx
> + mov esi,ebp
> + add edx,DWORD[44+rsp]
> + pslld xmm12,2
> + pxor xmm10,xmm13
> + xor eax,ebx
> + movdqa xmm13,XMMWORD[16+r11]
> + rol ebp,5
> + add edx,edi
> + movups xmm0,XMMWORD[32+r15]
> +DB 102,15,56,220,209
> + and esi,eax
> + pxor xmm10,xmm12
> + xor eax,ebx
> + add edx,ebp
> + ror ebp,7
> + pshufd xmm11,xmm7,238
> + xor esi,ebx
> + movdqa xmm12,xmm10
> + paddd xmm13,xmm10
> + mov edi,edx
> + add ecx,DWORD[48+rsp]
> + punpcklqdq xmm11,xmm8
> + xor ebp,eax
> + rol edx,5
> + add ecx,esi
> + psrldq xmm12,4
> + and edi,ebp
> + xor ebp,eax
> + pxor xmm11,xmm7
> + add ecx,edx
> + ror edx,7
> + pxor xmm12,xmm9
> + xor edi,eax
> + mov esi,ecx
> + add ebx,DWORD[52+rsp]
> + movups xmm1,XMMWORD[48+r15]
> +DB 102,15,56,220,208
> + pxor xmm11,xmm12
> + xor edx,ebp
> + rol ecx,5
> + movdqa XMMWORD[32+rsp],xmm13
> + add ebx,edi
> + and esi,edx
> + movdqa xmm3,xmm11
> + xor edx,ebp
> + add ebx,ecx
> + ror ecx,7
> + movdqa xmm12,xmm11
> + xor esi,ebp
> + pslldq xmm3,12
> + paddd xmm11,xmm11
> + mov edi,ebx
> + add eax,DWORD[56+rsp]
> + psrld xmm12,31
> + xor ecx,edx
> + rol ebx,5
> + add eax,esi
> + movdqa xmm13,xmm3
> + and edi,ecx
> + xor ecx,edx
> + psrld xmm3,30
> + add eax,ebx
> + ror ebx,7
> + cmp r8d,11
> + jb NEAR $L$aesenclast1
> + movups xmm0,XMMWORD[64+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+r15]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast1
> + movups xmm0,XMMWORD[96+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+r15]
> +DB 102,15,56,220,208
> +$L$aesenclast1:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+r15]
> + por xmm11,xmm12
> + xor edi,edx
> + mov esi,eax
> + add ebp,DWORD[60+rsp]
> + pslld xmm13,2
> + pxor xmm11,xmm3
> + xor ebx,ecx
> + movdqa xmm3,XMMWORD[16+r11]
> + rol eax,5
> + add ebp,edi
> + and esi,ebx
> + pxor xmm11,xmm13
> + pshufd xmm13,xmm10,238
> + xor ebx,ecx
> + add ebp,eax
> + ror eax,7
> + pxor xmm4,xmm8
> + xor esi,ecx
> + mov edi,ebp
> + add edx,DWORD[rsp]
> + punpcklqdq xmm13,xmm11
> + xor eax,ebx
> + rol ebp,5
> + pxor xmm4,xmm5
> + add edx,esi
> + movups xmm14,XMMWORD[16+r12]
> + xorps xmm14,xmm15
> + movups XMMWORD[r13*1+r12],xmm2
> + xorps xmm2,xmm14
> + movups xmm1,XMMWORD[((-80))+r15]
> +DB 102,15,56,220,208
> + and edi,eax
> + movdqa xmm12,xmm3
> + xor eax,ebx
> + paddd xmm3,xmm11
> + add edx,ebp
> + pxor xmm4,xmm13
> + ror ebp,7
> + xor edi,ebx
> + mov esi,edx
> + add ecx,DWORD[4+rsp]
> + movdqa xmm13,xmm4
> + xor ebp,eax
> + rol edx,5
> + movdqa XMMWORD[48+rsp],xmm3
> + add ecx,edi
> + and esi,ebp
> + xor ebp,eax
> + pslld xmm4,2
> + add ecx,edx
> + ror edx,7
> + psrld xmm13,30
> + xor esi,eax
> + mov edi,ecx
> + add ebx,DWORD[8+rsp]
> + movups xmm0,XMMWORD[((-64))+r15]
> +DB 102,15,56,220,209
> + por xmm4,xmm13
> + xor edx,ebp
> + rol ecx,5
> + pshufd xmm3,xmm11,238
> + add ebx,esi
> + and edi,edx
> + xor edx,ebp
> + add ebx,ecx
> + add eax,DWORD[12+rsp]
> + xor edi,ebp
> + mov esi,ebx
> + rol ebx,5
> + add eax,edi
> + xor esi,edx
> + ror ecx,7
> + add eax,ebx
> + pxor xmm5,xmm9
> + add ebp,DWORD[16+rsp]
> + movups xmm1,XMMWORD[((-48))+r15]
> +DB 102,15,56,220,208
> + xor esi,ecx
> + punpcklqdq xmm3,xmm4
> + mov edi,eax
> + rol eax,5
> + pxor xmm5,xmm6
> + add ebp,esi
> + xor edi,ecx
> + movdqa xmm13,xmm12
> + ror ebx,7
> + paddd xmm12,xmm4
> + add ebp,eax
> + pxor xmm5,xmm3
> + add edx,DWORD[20+rsp]
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + movdqa xmm3,xmm5
> + add edx,edi
> + xor esi,ebx
> + movdqa XMMWORD[rsp],xmm12
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[24+rsp]
> + pslld xmm5,2
> + xor esi,eax
> + mov edi,edx
> + psrld xmm3,30
> + rol edx,5
> + add ecx,esi
> + movups xmm0,XMMWORD[((-32))+r15]
> +DB 102,15,56,220,209
> + xor edi,eax
> + ror ebp,7
> + por xmm5,xmm3
> + add ecx,edx
> + add ebx,DWORD[28+rsp]
> + pshufd xmm12,xmm4,238
> + xor edi,ebp
> + mov esi,ecx
> + rol ecx,5
> + add ebx,edi
> + xor esi,ebp
> + ror edx,7
> + add ebx,ecx
> + pxor xmm6,xmm10
> + add eax,DWORD[32+rsp]
> + xor esi,edx
> + punpcklqdq xmm12,xmm5
> + mov edi,ebx
> + rol ebx,5
> + pxor xmm6,xmm7
> + add eax,esi
> + xor edi,edx
> + movdqa xmm3,XMMWORD[32+r11]
> + ror ecx,7
> + paddd xmm13,xmm5
> + add eax,ebx
> + pxor xmm6,xmm12
> + add ebp,DWORD[36+rsp]
> + movups xmm1,XMMWORD[((-16))+r15]
> +DB 102,15,56,220,208
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + movdqa xmm12,xmm6
> + add ebp,edi
> + xor esi,ecx
> + movdqa XMMWORD[16+rsp],xmm13
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[40+rsp]
> + pslld xmm6,2
> + xor esi,ebx
> + mov edi,ebp
> + psrld xmm12,30
> + rol ebp,5
> + add edx,esi
> + xor edi,ebx
> + ror eax,7
> + por xmm6,xmm12
> + add edx,ebp
> + add ecx,DWORD[44+rsp]
> + pshufd xmm13,xmm5,238
> + xor edi,eax
> + mov esi,edx
> + rol edx,5
> + add ecx,edi
> + movups xmm0,XMMWORD[r15]
> +DB 102,15,56,220,209
> + xor esi,eax
> + ror ebp,7
> + add ecx,edx
> + pxor xmm7,xmm11
> + add ebx,DWORD[48+rsp]
> + xor esi,ebp
> + punpcklqdq xmm13,xmm6
> + mov edi,ecx
> + rol ecx,5
> + pxor xmm7,xmm8
> + add ebx,esi
> + xor edi,ebp
> + movdqa xmm12,xmm3
> + ror edx,7
> + paddd xmm3,xmm6
> + add ebx,ecx
> + pxor xmm7,xmm13
> + add eax,DWORD[52+rsp]
> + xor edi,edx
> + mov esi,ebx
> + rol ebx,5
> + movdqa xmm13,xmm7
> + add eax,edi
> + xor esi,edx
> + movdqa XMMWORD[32+rsp],xmm3
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[56+rsp]
> + movups xmm1,XMMWORD[16+r15]
> +DB 102,15,56,220,208
> + pslld xmm7,2
> + xor esi,ecx
> + mov edi,eax
> + psrld xmm13,30
> + rol eax,5
> + add ebp,esi
> + xor edi,ecx
> + ror ebx,7
> + por xmm7,xmm13
> + add ebp,eax
> + add edx,DWORD[60+rsp]
> + pshufd xmm3,xmm6,238
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + add edx,edi
> + xor esi,ebx
> + ror eax,7
> + add edx,ebp
> + pxor xmm8,xmm4
> + add ecx,DWORD[rsp]
> + xor esi,eax
> + punpcklqdq xmm3,xmm7
> + mov edi,edx
> + rol edx,5
> + pxor xmm8,xmm9
> + add ecx,esi
> + movups xmm0,XMMWORD[32+r15]
> +DB 102,15,56,220,209
> + xor edi,eax
> + movdqa xmm13,xmm12
> + ror ebp,7
> + paddd xmm12,xmm7
> + add ecx,edx
> + pxor xmm8,xmm3
> + add ebx,DWORD[4+rsp]
> + xor edi,ebp
> + mov esi,ecx
> + rol ecx,5
> + movdqa xmm3,xmm8
> + add ebx,edi
> + xor esi,ebp
> + movdqa XMMWORD[48+rsp],xmm12
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[8+rsp]
> + pslld xmm8,2
> + xor esi,edx
> + mov edi,ebx
> + psrld xmm3,30
> + rol ebx,5
> + add eax,esi
> + xor edi,edx
> + ror ecx,7
> + por xmm8,xmm3
> + add eax,ebx
> + add ebp,DWORD[12+rsp]
> + movups xmm1,XMMWORD[48+r15]
> +DB 102,15,56,220,208
> + pshufd xmm12,xmm7,238
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + add ebp,edi
> + xor esi,ecx
> + ror ebx,7
> + add ebp,eax
> + pxor xmm9,xmm5
> + add edx,DWORD[16+rsp]
> + xor esi,ebx
> + punpcklqdq xmm12,xmm8
> + mov edi,ebp
> + rol ebp,5
> + pxor xmm9,xmm10
> + add edx,esi
> + xor edi,ebx
> + movdqa xmm3,xmm13
> + ror eax,7
> + paddd xmm13,xmm8
> + add edx,ebp
> + pxor xmm9,xmm12
> + add ecx,DWORD[20+rsp]
> + xor edi,eax
> + mov esi,edx
> + rol edx,5
> + movdqa xmm12,xmm9
> + add ecx,edi
> + cmp r8d,11
> + jb NEAR $L$aesenclast2
> + movups xmm0,XMMWORD[64+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+r15]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast2
> + movups xmm0,XMMWORD[96+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+r15]
> +DB 102,15,56,220,208
> +$L$aesenclast2:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+r15]
> + xor esi,eax
> + movdqa XMMWORD[rsp],xmm13
> + ror ebp,7
> + add ecx,edx
> + add ebx,DWORD[24+rsp]
> + pslld xmm9,2
> + xor esi,ebp
> + mov edi,ecx
> + psrld xmm12,30
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + por xmm9,xmm12
> + add ebx,ecx
> + add eax,DWORD[28+rsp]
> + pshufd xmm13,xmm8,238
> + ror ecx,7
> + mov esi,ebx
> + xor edi,edx
> + rol ebx,5
> + add eax,edi
> + xor esi,ecx
> + xor ecx,edx
> + add eax,ebx
> + pxor xmm10,xmm6
> + add ebp,DWORD[32+rsp]
> + movups xmm14,XMMWORD[32+r12]
> + xorps xmm14,xmm15
> + movups XMMWORD[16+r12*1+r13],xmm2
> + xorps xmm2,xmm14
> + movups xmm1,XMMWORD[((-80))+r15]
> +DB 102,15,56,220,208
> + and esi,ecx
> + xor ecx,edx
> + ror ebx,7
> + punpcklqdq xmm13,xmm9
> + mov edi,eax
> + xor esi,ecx
> + pxor xmm10,xmm11
> + rol eax,5
> + add ebp,esi
> + movdqa xmm12,xmm3
> + xor edi,ebx
> + paddd xmm3,xmm9
> + xor ebx,ecx
> + pxor xmm10,xmm13
> + add ebp,eax
> + add edx,DWORD[36+rsp]
> + and edi,ebx
> + xor ebx,ecx
> + ror eax,7
> + movdqa xmm13,xmm10
> + mov esi,ebp
> + xor edi,ebx
> + movdqa XMMWORD[16+rsp],xmm3
> + rol ebp,5
> + add edx,edi
> + movups xmm0,XMMWORD[((-64))+r15]
> +DB 102,15,56,220,209
> + xor esi,eax
> + pslld xmm10,2
> + xor eax,ebx
> + add edx,ebp
> + psrld xmm13,30
> + add ecx,DWORD[40+rsp]
> + and esi,eax
> + xor eax,ebx
> + por xmm10,xmm13
> + ror ebp,7
> + mov edi,edx
> + xor esi,eax
> + rol edx,5
> + pshufd xmm3,xmm9,238
> + add ecx,esi
> + xor edi,ebp
> + xor ebp,eax
> + add ecx,edx
> + add ebx,DWORD[44+rsp]
> + and edi,ebp
> + xor ebp,eax
> + ror edx,7
> + movups xmm1,XMMWORD[((-48))+r15]
> +DB 102,15,56,220,208
> + mov esi,ecx
> + xor edi,ebp
> + rol ecx,5
> + add ebx,edi
> + xor esi,edx
> + xor edx,ebp
> + add ebx,ecx
> + pxor xmm11,xmm7
> + add eax,DWORD[48+rsp]
> + and esi,edx
> + xor edx,ebp
> + ror ecx,7
> + punpcklqdq xmm3,xmm10
> + mov edi,ebx
> + xor esi,edx
> + pxor xmm11,xmm4
> + rol ebx,5
> + add eax,esi
> + movdqa xmm13,XMMWORD[48+r11]
> + xor edi,ecx
> + paddd xmm12,xmm10
> + xor ecx,edx
> + pxor xmm11,xmm3
> + add eax,ebx
> + add ebp,DWORD[52+rsp]
> + movups xmm0,XMMWORD[((-32))+r15]
> +DB 102,15,56,220,209
> + and edi,ecx
> + xor ecx,edx
> + ror ebx,7
> + movdqa xmm3,xmm11
> + mov esi,eax
> + xor edi,ecx
> + movdqa XMMWORD[32+rsp],xmm12
> + rol eax,5
> + add ebp,edi
> + xor esi,ebx
> + pslld xmm11,2
> + xor ebx,ecx
> + add ebp,eax
> + psrld xmm3,30
> + add edx,DWORD[56+rsp]
> + and esi,ebx
> + xor ebx,ecx
> + por xmm11,xmm3
> + ror eax,7
> + mov edi,ebp
> + xor esi,ebx
> + rol ebp,5
> + pshufd xmm12,xmm10,238
> + add edx,esi
> + movups xmm1,XMMWORD[((-16))+r15]
> +DB 102,15,56,220,208
> + xor edi,eax
> + xor eax,ebx
> + add edx,ebp
> + add ecx,DWORD[60+rsp]
> + and edi,eax
> + xor eax,ebx
> + ror ebp,7
> + mov esi,edx
> + xor edi,eax
> + rol edx,5
> + add ecx,edi
> + xor esi,ebp
> + xor ebp,eax
> + add ecx,edx
> + pxor xmm4,xmm8
> + add ebx,DWORD[rsp]
> + and esi,ebp
> + xor ebp,eax
> + ror edx,7
> + movups xmm0,XMMWORD[r15]
> +DB 102,15,56,220,209
> + punpcklqdq xmm12,xmm11
> + mov edi,ecx
> + xor esi,ebp
> + pxor xmm4,xmm5
> + rol ecx,5
> + add ebx,esi
> + movdqa xmm3,xmm13
> + xor edi,edx
> + paddd xmm13,xmm11
> + xor edx,ebp
> + pxor xmm4,xmm12
> + add ebx,ecx
> + add eax,DWORD[4+rsp]
> + and edi,edx
> + xor edx,ebp
> + ror ecx,7
> + movdqa xmm12,xmm4
> + mov esi,ebx
> + xor edi,edx
> + movdqa XMMWORD[48+rsp],xmm13
> + rol ebx,5
> + add eax,edi
> + xor esi,ecx
> + pslld xmm4,2
> + xor ecx,edx
> + add eax,ebx
> + psrld xmm12,30
> + add ebp,DWORD[8+rsp]
> + movups xmm1,XMMWORD[16+r15]
> +DB 102,15,56,220,208
> + and esi,ecx
> + xor ecx,edx
> + por xmm4,xmm12
> + ror ebx,7
> + mov edi,eax
> + xor esi,ecx
> + rol eax,5
> + pshufd xmm13,xmm11,238
> + add ebp,esi
> + xor edi,ebx
> + xor ebx,ecx
> + add ebp,eax
> + add edx,DWORD[12+rsp]
> + and edi,ebx
> + xor ebx,ecx
> + ror eax,7
> + mov esi,ebp
> + xor edi,ebx
> + rol ebp,5
> + add edx,edi
> + movups xmm0,XMMWORD[32+r15]
> +DB 102,15,56,220,209
> + xor esi,eax
> + xor eax,ebx
> + add edx,ebp
> + pxor xmm5,xmm9
> + add ecx,DWORD[16+rsp]
> + and esi,eax
> + xor eax,ebx
> + ror ebp,7
> + punpcklqdq xmm13,xmm4
> + mov edi,edx
> + xor esi,eax
> + pxor xmm5,xmm6
> + rol edx,5
> + add ecx,esi
> + movdqa xmm12,xmm3
> + xor edi,ebp
> + paddd xmm3,xmm4
> + xor ebp,eax
> + pxor xmm5,xmm13
> + add ecx,edx
> + add ebx,DWORD[20+rsp]
> + and edi,ebp
> + xor ebp,eax
> + ror edx,7
> + movups xmm1,XMMWORD[48+r15]
> +DB 102,15,56,220,208
> + movdqa xmm13,xmm5
> + mov esi,ecx
> + xor edi,ebp
> + movdqa XMMWORD[rsp],xmm3
> + rol ecx,5
> + add ebx,edi
> + xor esi,edx
> + pslld xmm5,2
> + xor edx,ebp
> + add ebx,ecx
> + psrld xmm13,30
> + add eax,DWORD[24+rsp]
> + and esi,edx
> + xor edx,ebp
> + por xmm5,xmm13
> + ror ecx,7
> + mov edi,ebx
> + xor esi,edx
> + rol ebx,5
> + pshufd xmm3,xmm4,238
> + add eax,esi
> + xor edi,ecx
> + xor ecx,edx
> + add eax,ebx
> + add ebp,DWORD[28+rsp]
> + cmp r8d,11
> + jb NEAR $L$aesenclast3
> + movups xmm0,XMMWORD[64+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+r15]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast3
> + movups xmm0,XMMWORD[96+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+r15]
> +DB 102,15,56,220,208
> +$L$aesenclast3:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+r15]
> + and edi,ecx
> + xor ecx,edx
> + ror ebx,7
> + mov esi,eax
> + xor edi,ecx
> + rol eax,5
> + add ebp,edi
> + xor esi,ebx
> + xor ebx,ecx
> + add ebp,eax
> + pxor xmm6,xmm10
> + add edx,DWORD[32+rsp]
> + and esi,ebx
> + xor ebx,ecx
> + ror eax,7
> + punpcklqdq xmm3,xmm5
> + mov edi,ebp
> + xor esi,ebx
> + pxor xmm6,xmm7
> + rol ebp,5
> + add edx,esi
> + movups xmm14,XMMWORD[48+r12]
> + xorps xmm14,xmm15
> + movups XMMWORD[32+r12*1+r13],xmm2
> + xorps xmm2,xmm14
> + movups xmm1,XMMWORD[((-80))+r15]
> +DB 102,15,56,220,208
> + movdqa xmm13,xmm12
> + xor edi,eax
> + paddd xmm12,xmm5
> + xor eax,ebx
> + pxor xmm6,xmm3
> + add edx,ebp
> + add ecx,DWORD[36+rsp]
> + and edi,eax
> + xor eax,ebx
> + ror ebp,7
> + movdqa xmm3,xmm6
> + mov esi,edx
> + xor edi,eax
> + movdqa XMMWORD[16+rsp],xmm12
> + rol edx,5
> + add ecx,edi
> + xor esi,ebp
> + pslld xmm6,2
> + xor ebp,eax
> + add ecx,edx
> + psrld xmm3,30
> + add ebx,DWORD[40+rsp]
> + and esi,ebp
> + xor ebp,eax
> + por xmm6,xmm3
> + ror edx,7
> + movups xmm0,XMMWORD[((-64))+r15]
> +DB 102,15,56,220,209
> + mov edi,ecx
> + xor esi,ebp
> + rol ecx,5
> + pshufd xmm12,xmm5,238
> + add ebx,esi
> + xor edi,edx
> + xor edx,ebp
> + add ebx,ecx
> + add eax,DWORD[44+rsp]
> + and edi,edx
> + xor edx,ebp
> + ror ecx,7
> + mov esi,ebx
> + xor edi,edx
> + rol ebx,5
> + add eax,edi
> + xor esi,edx
> + add eax,ebx
> + pxor xmm7,xmm11
> + add ebp,DWORD[48+rsp]
> + movups xmm1,XMMWORD[((-48))+r15]
> +DB 102,15,56,220,208
> + xor esi,ecx
> + punpcklqdq xmm12,xmm6
> + mov edi,eax
> + rol eax,5
> + pxor xmm7,xmm8
> + add ebp,esi
> + xor edi,ecx
> + movdqa xmm3,xmm13
> + ror ebx,7
> + paddd xmm13,xmm6
> + add ebp,eax
> + pxor xmm7,xmm12
> + add edx,DWORD[52+rsp]
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + movdqa xmm12,xmm7
> + add edx,edi
> + xor esi,ebx
> + movdqa XMMWORD[32+rsp],xmm13
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[56+rsp]
> + pslld xmm7,2
> + xor esi,eax
> + mov edi,edx
> + psrld xmm12,30
> + rol edx,5
> + add ecx,esi
> + movups xmm0,XMMWORD[((-32))+r15]
> +DB 102,15,56,220,209
> + xor edi,eax
> + ror ebp,7
> + por xmm7,xmm12
> + add ecx,edx
> + add ebx,DWORD[60+rsp]
> + xor edi,ebp
> + mov esi,ecx
> + rol ecx,5
> + add ebx,edi
> + xor esi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[rsp]
> + xor esi,edx
> + mov edi,ebx
> + rol ebx,5
> + paddd xmm3,xmm7
> + add eax,esi
> + xor edi,edx
> + movdqa XMMWORD[48+rsp],xmm3
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[4+rsp]
> + movups xmm1,XMMWORD[((-16))+r15]
> +DB 102,15,56,220,208
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + add ebp,edi
> + xor esi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[8+rsp]
> + xor esi,ebx
> + mov edi,ebp
> + rol ebp,5
> + add edx,esi
> + xor edi,ebx
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[12+rsp]
> + xor edi,eax
> + mov esi,edx
> + rol edx,5
> + add ecx,edi
> + movups xmm0,XMMWORD[r15]
> +DB 102,15,56,220,209
> + xor esi,eax
> + ror ebp,7
> + add ecx,edx
> + cmp r10,r14
> + je NEAR $L$done_ssse3
> + movdqa xmm3,XMMWORD[64+r11]
> + movdqa xmm13,XMMWORD[r11]
> + movdqu xmm4,XMMWORD[r10]
> + movdqu xmm5,XMMWORD[16+r10]
> + movdqu xmm6,XMMWORD[32+r10]
> + movdqu xmm7,XMMWORD[48+r10]
> +DB 102,15,56,0,227
> + add r10,64
> + add ebx,DWORD[16+rsp]
> + xor esi,ebp
> + mov edi,ecx
> +DB 102,15,56,0,235
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + paddd xmm4,xmm13
> + add ebx,ecx
> + add eax,DWORD[20+rsp]
> + xor edi,edx
> + mov esi,ebx
> + movdqa XMMWORD[rsp],xmm4
> + rol ebx,5
> + add eax,edi
> + xor esi,edx
> + ror ecx,7
> + psubd xmm4,xmm13
> + add eax,ebx
> + add ebp,DWORD[24+rsp]
> + movups xmm1,XMMWORD[16+r15]
> +DB 102,15,56,220,208
> + xor esi,ecx
> + mov edi,eax
> + rol eax,5
> + add ebp,esi
> + xor edi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[28+rsp]
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + add edx,edi
> + xor esi,ebx
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[32+rsp]
> + xor esi,eax
> + mov edi,edx
> +DB 102,15,56,0,243
> + rol edx,5
> + add ecx,esi
> + movups xmm0,XMMWORD[32+r15]
> +DB 102,15,56,220,209
> + xor edi,eax
> + ror ebp,7
> + paddd xmm5,xmm13
> + add ecx,edx
> + add ebx,DWORD[36+rsp]
> + xor edi,ebp
> + mov esi,ecx
> + movdqa XMMWORD[16+rsp],xmm5
> + rol ecx,5
> + add ebx,edi
> + xor esi,ebp
> + ror edx,7
> + psubd xmm5,xmm13
> + add ebx,ecx
> + add eax,DWORD[40+rsp]
> + xor esi,edx
> + mov edi,ebx
> + rol ebx,5
> + add eax,esi
> + xor edi,edx
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[44+rsp]
> + movups xmm1,XMMWORD[48+r15]
> +DB 102,15,56,220,208
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + add ebp,edi
> + xor esi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[48+rsp]
> + xor esi,ebx
> + mov edi,ebp
> +DB 102,15,56,0,251
> + rol ebp,5
> + add edx,esi
> + xor edi,ebx
> + ror eax,7
> + paddd xmm6,xmm13
> + add edx,ebp
> + add ecx,DWORD[52+rsp]
> + xor edi,eax
> + mov esi,edx
> + movdqa XMMWORD[32+rsp],xmm6
> + rol edx,5
> + add ecx,edi
> + cmp r8d,11
> + jb NEAR $L$aesenclast4
> + movups xmm0,XMMWORD[64+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+r15]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast4
> + movups xmm0,XMMWORD[96+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+r15]
> +DB 102,15,56,220,208
> +$L$aesenclast4:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+r15]
> + xor esi,eax
> + ror ebp,7
> + psubd xmm6,xmm13
> + add ecx,edx
> + add ebx,DWORD[56+rsp]
> + xor esi,ebp
> + mov edi,ecx
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[60+rsp]
> + xor edi,edx
> + mov esi,ebx
> + rol ebx,5
> + add eax,edi
> + ror ecx,7
> + add eax,ebx
> + movups XMMWORD[48+r12*1+r13],xmm2
> + lea r12,[64+r12]
> +
> + add eax,DWORD[r9]
> + add esi,DWORD[4+r9]
> + add ecx,DWORD[8+r9]
> + add edx,DWORD[12+r9]
> + mov DWORD[r9],eax
> + add ebp,DWORD[16+r9]
> + mov DWORD[4+r9],esi
> + mov ebx,esi
> + mov DWORD[8+r9],ecx
> + mov edi,ecx
> + mov DWORD[12+r9],edx
> + xor edi,edx
> + mov DWORD[16+r9],ebp
> + and esi,edi
> + jmp NEAR $L$oop_ssse3
> +
> +$L$done_ssse3:
> + add ebx,DWORD[16+rsp]
> + xor esi,ebp
> + mov edi,ecx
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[20+rsp]
> + xor edi,edx
> + mov esi,ebx
> + rol ebx,5
> + add eax,edi
> + xor esi,edx
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[24+rsp]
> + movups xmm1,XMMWORD[16+r15]
> +DB 102,15,56,220,208
> + xor esi,ecx
> + mov edi,eax
> + rol eax,5
> + add ebp,esi
> + xor edi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[28+rsp]
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + add edx,edi
> + xor esi,ebx
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[32+rsp]
> + xor esi,eax
> + mov edi,edx
> + rol edx,5
> + add ecx,esi
> + movups xmm0,XMMWORD[32+r15]
> +DB 102,15,56,220,209
> + xor edi,eax
> + ror ebp,7
> + add ecx,edx
> + add ebx,DWORD[36+rsp]
> + xor edi,ebp
> + mov esi,ecx
> + rol ecx,5
> + add ebx,edi
> + xor esi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[40+rsp]
> + xor esi,edx
> + mov edi,ebx
> + rol ebx,5
> + add eax,esi
> + xor edi,edx
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[44+rsp]
> + movups xmm1,XMMWORD[48+r15]
> +DB 102,15,56,220,208
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + add ebp,edi
> + xor esi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[48+rsp]
> + xor esi,ebx
> + mov edi,ebp
> + rol ebp,5
> + add edx,esi
> + xor edi,ebx
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[52+rsp]
> + xor edi,eax
> + mov esi,edx
> + rol edx,5
> + add ecx,edi
> + cmp r8d,11
> + jb NEAR $L$aesenclast5
> + movups xmm0,XMMWORD[64+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+r15]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast5
> + movups xmm0,XMMWORD[96+r15]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+r15]
> +DB 102,15,56,220,208
> +$L$aesenclast5:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+r15]
> + xor esi,eax
> + ror ebp,7
> + add ecx,edx
> + add ebx,DWORD[56+rsp]
> + xor esi,ebp
> + mov edi,ecx
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[60+rsp]
> + xor edi,edx
> + mov esi,ebx
> + rol ebx,5
> + add eax,edi
> + ror ecx,7
> + add eax,ebx
> + movups XMMWORD[48+r12*1+r13],xmm2
> + mov r8,QWORD[88+rsp]
> +
> + add eax,DWORD[r9]
> + add esi,DWORD[4+r9]
> + add ecx,DWORD[8+r9]
> + mov DWORD[r9],eax
> + add edx,DWORD[12+r9]
> + mov DWORD[4+r9],esi
> + add ebp,DWORD[16+r9]
> + mov DWORD[8+r9],ecx
> + mov DWORD[12+r9],edx
> + mov DWORD[16+r9],ebp
> + movups XMMWORD[r8],xmm2
> + movaps xmm6,XMMWORD[((96+0))+rsp]
> + movaps xmm7,XMMWORD[((96+16))+rsp]
> + movaps xmm8,XMMWORD[((96+32))+rsp]
> + movaps xmm9,XMMWORD[((96+48))+rsp]
> + movaps xmm10,XMMWORD[((96+64))+rsp]
> + movaps xmm11,XMMWORD[((96+80))+rsp]
> + movaps xmm12,XMMWORD[((96+96))+rsp]
> + movaps xmm13,XMMWORD[((96+112))+rsp]
> + movaps xmm14,XMMWORD[((96+128))+rsp]
> + movaps xmm15,XMMWORD[((96+144))+rsp]
> + lea rsi,[264+rsp]
> +
> + mov r15,QWORD[rsi]
> +
> + mov r14,QWORD[8+rsi]
> +
> + mov r13,QWORD[16+rsi]
> +
> + mov r12,QWORD[24+rsi]
> +
> + mov rbp,QWORD[32+rsi]
> +
> + mov rbx,QWORD[40+rsi]
> +
> + lea rsp,[48+rsi]
> +
> +$L$epilogue_ssse3:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_cbc_sha1_enc_ssse3:
> +ALIGN 64
> +K_XX_XX:
> + DD 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> + DD 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> + DD 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> + DD 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +DB 0xf,0xe,0xd,0xc,0xb,0xa,0x9,0x8,0x7,0x6,0x5,0x4,0x3,0x2,0x1,0x0
> +
> +DB 65,69,83,78,73,45,67,66,67,43,83,72,65,49,32,115
> +DB 116,105,116,99,104,32,102,111,114,32,120,56,54,95,54,52
> +DB 44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32
> +DB 60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111
> +DB 114,103,62,0
> +ALIGN 64
> +
> +ALIGN 32
> +aesni_cbc_sha1_enc_shaext:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_cbc_sha1_enc_shaext:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + mov r10,QWORD[56+rsp]
> + lea rsp,[((-168))+rsp]
> + movaps XMMWORD[(-8-160)+rax],xmm6
> + movaps XMMWORD[(-8-144)+rax],xmm7
> + movaps XMMWORD[(-8-128)+rax],xmm8
> + movaps XMMWORD[(-8-112)+rax],xmm9
> + movaps XMMWORD[(-8-96)+rax],xmm10
> + movaps XMMWORD[(-8-80)+rax],xmm11
> + movaps XMMWORD[(-8-64)+rax],xmm12
> + movaps XMMWORD[(-8-48)+rax],xmm13
> + movaps XMMWORD[(-8-32)+rax],xmm14
> + movaps XMMWORD[(-8-16)+rax],xmm15
> +$L$prologue_shaext:
> + movdqu xmm8,XMMWORD[r9]
> + movd xmm9,DWORD[16+r9]
> + movdqa xmm7,XMMWORD[((K_XX_XX+80))]
> +
> + mov r11d,DWORD[240+rcx]
> + sub rsi,rdi
> + movups xmm15,XMMWORD[rcx]
> + movups xmm2,XMMWORD[r8]
> + movups xmm0,XMMWORD[16+rcx]
> + lea rcx,[112+rcx]
> +
> + pshufd xmm8,xmm8,27
> + pshufd xmm9,xmm9,27
> + jmp NEAR $L$oop_shaext
> +
> +ALIGN 16
> +$L$oop_shaext:
> + movups xmm14,XMMWORD[rdi]
> + xorps xmm14,xmm15
> + xorps xmm2,xmm14
> + movups xmm1,XMMWORD[((-80))+rcx]
> +DB 102,15,56,220,208
> + movdqu xmm3,XMMWORD[r10]
> + movdqa xmm12,xmm9
> +DB 102,15,56,0,223
> + movdqu xmm4,XMMWORD[16+r10]
> + movdqa xmm11,xmm8
> + movups xmm0,XMMWORD[((-64))+rcx]
> +DB 102,15,56,220,209
> +DB 102,15,56,0,231
> +
> + paddd xmm9,xmm3
> + movdqu xmm5,XMMWORD[32+r10]
> + lea r10,[64+r10]
> + pxor xmm3,xmm12
> + movups xmm1,XMMWORD[((-48))+rcx]
> +DB 102,15,56,220,208
> + pxor xmm3,xmm12
> + movdqa xmm10,xmm8
> +DB 102,15,56,0,239
> +DB 69,15,58,204,193,0
> +DB 68,15,56,200,212
> + movups xmm0,XMMWORD[((-32))+rcx]
> +DB 102,15,56,220,209
> +DB 15,56,201,220
> + movdqu xmm6,XMMWORD[((-16))+r10]
> + movdqa xmm9,xmm8
> +DB 102,15,56,0,247
> + movups xmm1,XMMWORD[((-16))+rcx]
> +DB 102,15,56,220,208
> +DB 69,15,58,204,194,0
> +DB 68,15,56,200,205
> + pxor xmm3,xmm5
> +DB 15,56,201,229
> + movups xmm0,XMMWORD[rcx]
> +DB 102,15,56,220,209
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,0
> +DB 68,15,56,200,214
> + movups xmm1,XMMWORD[16+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,222
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> + movups xmm0,XMMWORD[32+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,0
> +DB 68,15,56,200,203
> + movups xmm1,XMMWORD[48+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,227
> + pxor xmm5,xmm3
> +DB 15,56,201,243
> + cmp r11d,11
> + jb NEAR $L$aesenclast6
> + movups xmm0,XMMWORD[64+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+rcx]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast6
> + movups xmm0,XMMWORD[96+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+rcx]
> +DB 102,15,56,220,208
> +$L$aesenclast6:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+rcx]
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,0
> +DB 68,15,56,200,212
> + movups xmm14,XMMWORD[16+rdi]
> + xorps xmm14,xmm15
> + movups XMMWORD[rdi*1+rsi],xmm2
> + xorps xmm2,xmm14
> + movups xmm1,XMMWORD[((-80))+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,236
> + pxor xmm6,xmm4
> +DB 15,56,201,220
> + movups xmm0,XMMWORD[((-64))+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,1
> +DB 68,15,56,200,205
> + movups xmm1,XMMWORD[((-48))+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,245
> + pxor xmm3,xmm5
> +DB 15,56,201,229
> + movups xmm0,XMMWORD[((-32))+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,1
> +DB 68,15,56,200,214
> + movups xmm1,XMMWORD[((-16))+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,222
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> + movups xmm0,XMMWORD[rcx]
> +DB 102,15,56,220,209
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,1
> +DB 68,15,56,200,203
> + movups xmm1,XMMWORD[16+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,227
> + pxor xmm5,xmm3
> +DB 15,56,201,243
> + movups xmm0,XMMWORD[32+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,1
> +DB 68,15,56,200,212
> + movups xmm1,XMMWORD[48+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,236
> + pxor xmm6,xmm4
> +DB 15,56,201,220
> + cmp r11d,11
> + jb NEAR $L$aesenclast7
> + movups xmm0,XMMWORD[64+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+rcx]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast7
> + movups xmm0,XMMWORD[96+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+rcx]
> +DB 102,15,56,220,208
> +$L$aesenclast7:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+rcx]
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,1
> +DB 68,15,56,200,205
> + movups xmm14,XMMWORD[32+rdi]
> + xorps xmm14,xmm15
> + movups XMMWORD[16+rdi*1+rsi],xmm2
> + xorps xmm2,xmm14
> + movups xmm1,XMMWORD[((-80))+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,245
> + pxor xmm3,xmm5
> +DB 15,56,201,229
> + movups xmm0,XMMWORD[((-64))+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,2
> +DB 68,15,56,200,214
> + movups xmm1,XMMWORD[((-48))+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,222
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> + movups xmm0,XMMWORD[((-32))+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,2
> +DB 68,15,56,200,203
> + movups xmm1,XMMWORD[((-16))+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,227
> + pxor xmm5,xmm3
> +DB 15,56,201,243
> + movups xmm0,XMMWORD[rcx]
> +DB 102,15,56,220,209
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,2
> +DB 68,15,56,200,212
> + movups xmm1,XMMWORD[16+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,236
> + pxor xmm6,xmm4
> +DB 15,56,201,220
> + movups xmm0,XMMWORD[32+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,2
> +DB 68,15,56,200,205
> + movups xmm1,XMMWORD[48+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,245
> + pxor xmm3,xmm5
> +DB 15,56,201,229
> + cmp r11d,11
> + jb NEAR $L$aesenclast8
> + movups xmm0,XMMWORD[64+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+rcx]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast8
> + movups xmm0,XMMWORD[96+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+rcx]
> +DB 102,15,56,220,208
> +$L$aesenclast8:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+rcx]
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,2
> +DB 68,15,56,200,214
> + movups xmm14,XMMWORD[48+rdi]
> + xorps xmm14,xmm15
> + movups XMMWORD[32+rdi*1+rsi],xmm2
> + xorps xmm2,xmm14
> + movups xmm1,XMMWORD[((-80))+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,222
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> + movups xmm0,XMMWORD[((-64))+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,3
> +DB 68,15,56,200,203
> + movups xmm1,XMMWORD[((-48))+rcx]
> +DB 102,15,56,220,208
> +DB 15,56,202,227
> + pxor xmm5,xmm3
> +DB 15,56,201,243
> + movups xmm0,XMMWORD[((-32))+rcx]
> +DB 102,15,56,220,209
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,3
> +DB 68,15,56,200,212
> +DB 15,56,202,236
> + pxor xmm6,xmm4
> + movups xmm1,XMMWORD[((-16))+rcx]
> +DB 102,15,56,220,208
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,3
> +DB 68,15,56,200,205
> +DB 15,56,202,245
> + movups xmm0,XMMWORD[rcx]
> +DB 102,15,56,220,209
> + movdqa xmm5,xmm12
> + movdqa xmm10,xmm8
> +DB 69,15,58,204,193,3
> +DB 68,15,56,200,214
> + movups xmm1,XMMWORD[16+rcx]
> +DB 102,15,56,220,208
> + movdqa xmm9,xmm8
> +DB 69,15,58,204,194,3
> +DB 68,15,56,200,205
> + movups xmm0,XMMWORD[32+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[48+rcx]
> +DB 102,15,56,220,208
> + cmp r11d,11
> + jb NEAR $L$aesenclast9
> + movups xmm0,XMMWORD[64+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[80+rcx]
> +DB 102,15,56,220,208
> + je NEAR $L$aesenclast9
> + movups xmm0,XMMWORD[96+rcx]
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[112+rcx]
> +DB 102,15,56,220,208
> +$L$aesenclast9:
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[((16-112))+rcx]
> + dec rdx
> +
> + paddd xmm8,xmm11
> + movups XMMWORD[48+rdi*1+rsi],xmm2
> + lea rdi,[64+rdi]
> + jnz NEAR $L$oop_shaext
> +
> + pshufd xmm8,xmm8,27
> + pshufd xmm9,xmm9,27
> + movups XMMWORD[r8],xmm2
> + movdqu XMMWORD[r9],xmm8
> + movd DWORD[16+r9],xmm9
> + movaps xmm6,XMMWORD[((-8-160))+rax]
> + movaps xmm7,XMMWORD[((-8-144))+rax]
> + movaps xmm8,XMMWORD[((-8-128))+rax]
> + movaps xmm9,XMMWORD[((-8-112))+rax]
> + movaps xmm10,XMMWORD[((-8-96))+rax]
> + movaps xmm11,XMMWORD[((-8-80))+rax]
> + movaps xmm12,XMMWORD[((-8-64))+rax]
> + movaps xmm13,XMMWORD[((-8-48))+rax]
> + movaps xmm14,XMMWORD[((-8-32))+rax]
> + movaps xmm15,XMMWORD[((-8-16))+rax]
> + mov rsp,rax
> +$L$epilogue_shaext:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_cbc_sha1_enc_shaext:
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +ssse3_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$common_seh_tail
> + lea r10,[aesni_cbc_sha1_enc_shaext]
> + cmp rbx,r10
> + jb NEAR $L$seh_no_shaext
> +
> + lea rsi,[rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> + lea rax,[168+rax]
> + jmp NEAR $L$common_seh_tail
> +$L$seh_no_shaext:
> + lea rsi,[96+rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> + lea rax,[264+rax]
> +
> + mov r15,QWORD[rax]
> + mov r14,QWORD[8+rax]
> + mov r13,QWORD[16+rax]
> + mov r12,QWORD[24+rax]
> + mov rbp,QWORD[32+rax]
> + mov rbx,QWORD[40+rax]
> + lea rax,[48+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> + mov QWORD[216+r8],r12
> + mov QWORD[224+r8],r13
> + mov QWORD[232+r8],r14
> + mov QWORD[240+r8],r15
> +
> +$L$common_seh_tail:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_aesni_cbc_sha1_enc_ssse3 wrt ..imagebase
> + DD $L$SEH_end_aesni_cbc_sha1_enc_ssse3 wrt ..imagebase
> + DD $L$SEH_info_aesni_cbc_sha1_enc_ssse3 wrt ..imagebase
> + DD $L$SEH_begin_aesni_cbc_sha1_enc_shaext wrt ..imagebase
> + DD $L$SEH_end_aesni_cbc_sha1_enc_shaext wrt ..imagebase
> + DD $L$SEH_info_aesni_cbc_sha1_enc_shaext wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_aesni_cbc_sha1_enc_ssse3:
> +DB 9,0,0,0
> + DD ssse3_handler wrt ..imagebase
> + DD $L$prologue_ssse3 wrt ..imagebase,$L$epilogue_ssse3
> wrt ..imagebase
> +$L$SEH_info_aesni_cbc_sha1_enc_shaext:
> +DB 9,0,0,0
> + DD ssse3_handler wrt ..imagebase
> + DD $L$prologue_shaext wrt ..imagebase,$L$epilogue_shaext
> wrt ..imagebase
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-sha256-
> x86_64.nasm b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-sha256-
> x86_64.nasm
> new file mode 100644
> index 0000000000..f5c250b904
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-sha256-
> x86_64.nasm
> @@ -0,0 +1,78 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/aes/asm/aesni-sha256-x86_64.pl
> +;
> +; Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +
> +EXTERN OPENSSL_ia32cap_P
> +global aesni_cbc_sha256_enc
> +
> +ALIGN 16
> +aesni_cbc_sha256_enc:
> +
> + xor eax,eax
> + cmp rcx,0
> + je NEAR $L$probe
> + ud2
> +$L$probe:
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 64
> +
> +K256:
> + DD 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> + DD 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> + DD 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> + DD 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> + DD 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> + DD 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> + DD 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> + DD 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> + DD 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> + DD 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> + DD 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> + DD 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> + DD 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> + DD 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> + DD 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> + DD 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> + DD 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> + DD 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> + DD 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> + DD 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> + DD 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> + DD 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> + DD 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> + DD 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> + DD 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> + DD 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> + DD 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> + DD 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> + DD 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> + DD 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> + DD 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> + DD 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> +
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> + DD 0,0,0,0,0,0,0,0,-1,-1,-1,-1
> + DD 0,0,0,0,0,0,0,0
> +DB 65,69,83,78,73,45,67,66,67,43,83,72,65,50,53,54
> +DB 32,115,116,105,116,99,104,32,102,111,114,32,120,56,54,95
> +DB 54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98
> +DB 121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108
> +DB 46,111,114,103,62,0
> +ALIGN 64
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-x86_64.nasm
> b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-x86_64.nasm
> new file mode 100644
> index 0000000000..57ee23ea8c
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/aesni-x86_64.nasm
> @@ -0,0 +1,5103 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/aes/asm/aesni-x86_64.pl
> +;
> +; Copyright 2009-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +EXTERN OPENSSL_ia32cap_P
> +global aesni_encrypt
> +
> +ALIGN 16
> +aesni_encrypt:
> +
> + movups xmm2,XMMWORD[rcx]
> + mov eax,DWORD[240+r8]
> + movups xmm0,XMMWORD[r8]
> + movups xmm1,XMMWORD[16+r8]
> + lea r8,[32+r8]
> + xorps xmm2,xmm0
> +$L$oop_enc1_1:
> +DB 102,15,56,220,209
> + dec eax
> + movups xmm1,XMMWORD[r8]
> + lea r8,[16+r8]
> + jnz NEAR $L$oop_enc1_1
> +DB 102,15,56,221,209
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + movups XMMWORD[rdx],xmm2
> + pxor xmm2,xmm2
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +global aesni_decrypt
> +
> +ALIGN 16
> +aesni_decrypt:
> +
> + movups xmm2,XMMWORD[rcx]
> + mov eax,DWORD[240+r8]
> + movups xmm0,XMMWORD[r8]
> + movups xmm1,XMMWORD[16+r8]
> + lea r8,[32+r8]
> + xorps xmm2,xmm0
> +$L$oop_dec1_2:
> +DB 102,15,56,222,209
> + dec eax
> + movups xmm1,XMMWORD[r8]
> + lea r8,[16+r8]
> + jnz NEAR $L$oop_dec1_2
> +DB 102,15,56,223,209
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + movups XMMWORD[rdx],xmm2
> + pxor xmm2,xmm2
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_encrypt2:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + xorps xmm3,xmm0
> + movups xmm0,XMMWORD[32+rcx]
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> + add rax,16
> +
> +$L$enc_loop2:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$enc_loop2
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,221,208
> +DB 102,15,56,221,216
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_decrypt2:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + xorps xmm3,xmm0
> + movups xmm0,XMMWORD[32+rcx]
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> + add rax,16
> +
> +$L$dec_loop2:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$dec_loop2
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,223,208
> +DB 102,15,56,223,216
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_encrypt3:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + xorps xmm3,xmm0
> + xorps xmm4,xmm0
> + movups xmm0,XMMWORD[32+rcx]
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> + add rax,16
> +
> +$L$enc_loop3:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$enc_loop3
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,221,208
> +DB 102,15,56,221,216
> +DB 102,15,56,221,224
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_decrypt3:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + xorps xmm3,xmm0
> + xorps xmm4,xmm0
> + movups xmm0,XMMWORD[32+rcx]
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> + add rax,16
> +
> +$L$dec_loop3:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$dec_loop3
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,223,208
> +DB 102,15,56,223,216
> +DB 102,15,56,223,224
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_encrypt4:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + xorps xmm3,xmm0
> + xorps xmm4,xmm0
> + xorps xmm5,xmm0
> + movups xmm0,XMMWORD[32+rcx]
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> +DB 0x0f,0x1f,0x00
> + add rax,16
> +
> +$L$enc_loop4:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$enc_loop4
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,221,208
> +DB 102,15,56,221,216
> +DB 102,15,56,221,224
> +DB 102,15,56,221,232
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_decrypt4:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + xorps xmm3,xmm0
> + xorps xmm4,xmm0
> + xorps xmm5,xmm0
> + movups xmm0,XMMWORD[32+rcx]
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> +DB 0x0f,0x1f,0x00
> + add rax,16
> +
> +$L$dec_loop4:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$dec_loop4
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,223,208
> +DB 102,15,56,223,216
> +DB 102,15,56,223,224
> +DB 102,15,56,223,232
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_encrypt6:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + pxor xmm3,xmm0
> + pxor xmm4,xmm0
> +DB 102,15,56,220,209
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> +DB 102,15,56,220,217
> + pxor xmm5,xmm0
> + pxor xmm6,xmm0
> +DB 102,15,56,220,225
> + pxor xmm7,xmm0
> + movups xmm0,XMMWORD[rax*1+rcx]
> + add rax,16
> + jmp NEAR $L$enc_loop6_enter
> +ALIGN 16
> +$L$enc_loop6:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +$L$enc_loop6_enter:
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$enc_loop6
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,15,56,221,208
> +DB 102,15,56,221,216
> +DB 102,15,56,221,224
> +DB 102,15,56,221,232
> +DB 102,15,56,221,240
> +DB 102,15,56,221,248
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_decrypt6:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + pxor xmm3,xmm0
> + pxor xmm4,xmm0
> +DB 102,15,56,222,209
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> +DB 102,15,56,222,217
> + pxor xmm5,xmm0
> + pxor xmm6,xmm0
> +DB 102,15,56,222,225
> + pxor xmm7,xmm0
> + movups xmm0,XMMWORD[rax*1+rcx]
> + add rax,16
> + jmp NEAR $L$dec_loop6_enter
> +ALIGN 16
> +$L$dec_loop6:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +$L$dec_loop6_enter:
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$dec_loop6
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,15,56,223,208
> +DB 102,15,56,223,216
> +DB 102,15,56,223,224
> +DB 102,15,56,223,232
> +DB 102,15,56,223,240
> +DB 102,15,56,223,248
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_encrypt8:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + xorps xmm3,xmm0
> + pxor xmm4,xmm0
> + pxor xmm5,xmm0
> + pxor xmm6,xmm0
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> +DB 102,15,56,220,209
> + pxor xmm7,xmm0
> + pxor xmm8,xmm0
> +DB 102,15,56,220,217
> + pxor xmm9,xmm0
> + movups xmm0,XMMWORD[rax*1+rcx]
> + add rax,16
> + jmp NEAR $L$enc_loop8_inner
> +ALIGN 16
> +$L$enc_loop8:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +$L$enc_loop8_inner:
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> +$L$enc_loop8_enter:
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> +DB 102,68,15,56,220,192
> +DB 102,68,15,56,220,200
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$enc_loop8
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> +DB 102,15,56,221,208
> +DB 102,15,56,221,216
> +DB 102,15,56,221,224
> +DB 102,15,56,221,232
> +DB 102,15,56,221,240
> +DB 102,15,56,221,248
> +DB 102,68,15,56,221,192
> +DB 102,68,15,56,221,200
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 16
> +_aesni_decrypt8:
> +
> + movups xmm0,XMMWORD[rcx]
> + shl eax,4
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm2,xmm0
> + xorps xmm3,xmm0
> + pxor xmm4,xmm0
> + pxor xmm5,xmm0
> + pxor xmm6,xmm0
> + lea rcx,[32+rax*1+rcx]
> + neg rax
> +DB 102,15,56,222,209
> + pxor xmm7,xmm0
> + pxor xmm8,xmm0
> +DB 102,15,56,222,217
> + pxor xmm9,xmm0
> + movups xmm0,XMMWORD[rax*1+rcx]
> + add rax,16
> + jmp NEAR $L$dec_loop8_inner
> +ALIGN 16
> +$L$dec_loop8:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +$L$dec_loop8_inner:
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,68,15,56,222,193
> +DB 102,68,15,56,222,201
> +$L$dec_loop8_enter:
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> +DB 102,68,15,56,222,192
> +DB 102,68,15,56,222,200
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$dec_loop8
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,68,15,56,222,193
> +DB 102,68,15,56,222,201
> +DB 102,15,56,223,208
> +DB 102,15,56,223,216
> +DB 102,15,56,223,224
> +DB 102,15,56,223,232
> +DB 102,15,56,223,240
> +DB 102,15,56,223,248
> +DB 102,68,15,56,223,192
> +DB 102,68,15,56,223,200
> + DB 0F3h,0C3h ;repret
> +
> +
> +global aesni_ecb_encrypt
> +
> +ALIGN 16
> +aesni_ecb_encrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_ecb_encrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> +
> +
> +
> + lea rsp,[((-88))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> +$L$ecb_enc_body:
> + and rdx,-16
> + jz NEAR $L$ecb_ret
> +
> + mov eax,DWORD[240+rcx]
> + movups xmm0,XMMWORD[rcx]
> + mov r11,rcx
> + mov r10d,eax
> + test r8d,r8d
> + jz NEAR $L$ecb_decrypt
> +
> + cmp rdx,0x80
> + jb NEAR $L$ecb_enc_tail
> +
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[16+rdi]
> + movdqu xmm4,XMMWORD[32+rdi]
> + movdqu xmm5,XMMWORD[48+rdi]
> + movdqu xmm6,XMMWORD[64+rdi]
> + movdqu xmm7,XMMWORD[80+rdi]
> + movdqu xmm8,XMMWORD[96+rdi]
> + movdqu xmm9,XMMWORD[112+rdi]
> + lea rdi,[128+rdi]
> + sub rdx,0x80
> + jmp NEAR $L$ecb_enc_loop8_enter
> +ALIGN 16
> +$L$ecb_enc_loop8:
> + movups XMMWORD[rsi],xmm2
> + mov rcx,r11
> + movdqu xmm2,XMMWORD[rdi]
> + mov eax,r10d
> + movups XMMWORD[16+rsi],xmm3
> + movdqu xmm3,XMMWORD[16+rdi]
> + movups XMMWORD[32+rsi],xmm4
> + movdqu xmm4,XMMWORD[32+rdi]
> + movups XMMWORD[48+rsi],xmm5
> + movdqu xmm5,XMMWORD[48+rdi]
> + movups XMMWORD[64+rsi],xmm6
> + movdqu xmm6,XMMWORD[64+rdi]
> + movups XMMWORD[80+rsi],xmm7
> + movdqu xmm7,XMMWORD[80+rdi]
> + movups XMMWORD[96+rsi],xmm8
> + movdqu xmm8,XMMWORD[96+rdi]
> + movups XMMWORD[112+rsi],xmm9
> + lea rsi,[128+rsi]
> + movdqu xmm9,XMMWORD[112+rdi]
> + lea rdi,[128+rdi]
> +$L$ecb_enc_loop8_enter:
> +
> + call _aesni_encrypt8
> +
> + sub rdx,0x80
> + jnc NEAR $L$ecb_enc_loop8
> +
> + movups XMMWORD[rsi],xmm2
> + mov rcx,r11
> + movups XMMWORD[16+rsi],xmm3
> + mov eax,r10d
> + movups XMMWORD[32+rsi],xmm4
> + movups XMMWORD[48+rsi],xmm5
> + movups XMMWORD[64+rsi],xmm6
> + movups XMMWORD[80+rsi],xmm7
> + movups XMMWORD[96+rsi],xmm8
> + movups XMMWORD[112+rsi],xmm9
> + lea rsi,[128+rsi]
> + add rdx,0x80
> + jz NEAR $L$ecb_ret
> +
> +$L$ecb_enc_tail:
> + movups xmm2,XMMWORD[rdi]
> + cmp rdx,0x20
> + jb NEAR $L$ecb_enc_one
> + movups xmm3,XMMWORD[16+rdi]
> + je NEAR $L$ecb_enc_two
> + movups xmm4,XMMWORD[32+rdi]
> + cmp rdx,0x40
> + jb NEAR $L$ecb_enc_three
> + movups xmm5,XMMWORD[48+rdi]
> + je NEAR $L$ecb_enc_four
> + movups xmm6,XMMWORD[64+rdi]
> + cmp rdx,0x60
> + jb NEAR $L$ecb_enc_five
> + movups xmm7,XMMWORD[80+rdi]
> + je NEAR $L$ecb_enc_six
> + movdqu xmm8,XMMWORD[96+rdi]
> + xorps xmm9,xmm9
> + call _aesni_encrypt8
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + movups XMMWORD[48+rsi],xmm5
> + movups XMMWORD[64+rsi],xmm6
> + movups XMMWORD[80+rsi],xmm7
> + movups XMMWORD[96+rsi],xmm8
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_enc_one:
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_enc1_3:
> +DB 102,15,56,220,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_enc1_3
> +DB 102,15,56,221,209
> + movups XMMWORD[rsi],xmm2
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_enc_two:
> + call _aesni_encrypt2
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_enc_three:
> + call _aesni_encrypt3
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_enc_four:
> + call _aesni_encrypt4
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + movups XMMWORD[48+rsi],xmm5
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_enc_five:
> + xorps xmm7,xmm7
> + call _aesni_encrypt6
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + movups XMMWORD[48+rsi],xmm5
> + movups XMMWORD[64+rsi],xmm6
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_enc_six:
> + call _aesni_encrypt6
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + movups XMMWORD[48+rsi],xmm5
> + movups XMMWORD[64+rsi],xmm6
> + movups XMMWORD[80+rsi],xmm7
> + jmp NEAR $L$ecb_ret
> +
> +ALIGN 16
> +$L$ecb_decrypt:
> + cmp rdx,0x80
> + jb NEAR $L$ecb_dec_tail
> +
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[16+rdi]
> + movdqu xmm4,XMMWORD[32+rdi]
> + movdqu xmm5,XMMWORD[48+rdi]
> + movdqu xmm6,XMMWORD[64+rdi]
> + movdqu xmm7,XMMWORD[80+rdi]
> + movdqu xmm8,XMMWORD[96+rdi]
> + movdqu xmm9,XMMWORD[112+rdi]
> + lea rdi,[128+rdi]
> + sub rdx,0x80
> + jmp NEAR $L$ecb_dec_loop8_enter
> +ALIGN 16
> +$L$ecb_dec_loop8:
> + movups XMMWORD[rsi],xmm2
> + mov rcx,r11
> + movdqu xmm2,XMMWORD[rdi]
> + mov eax,r10d
> + movups XMMWORD[16+rsi],xmm3
> + movdqu xmm3,XMMWORD[16+rdi]
> + movups XMMWORD[32+rsi],xmm4
> + movdqu xmm4,XMMWORD[32+rdi]
> + movups XMMWORD[48+rsi],xmm5
> + movdqu xmm5,XMMWORD[48+rdi]
> + movups XMMWORD[64+rsi],xmm6
> + movdqu xmm6,XMMWORD[64+rdi]
> + movups XMMWORD[80+rsi],xmm7
> + movdqu xmm7,XMMWORD[80+rdi]
> + movups XMMWORD[96+rsi],xmm8
> + movdqu xmm8,XMMWORD[96+rdi]
> + movups XMMWORD[112+rsi],xmm9
> + lea rsi,[128+rsi]
> + movdqu xmm9,XMMWORD[112+rdi]
> + lea rdi,[128+rdi]
> +$L$ecb_dec_loop8_enter:
> +
> + call _aesni_decrypt8
> +
> + movups xmm0,XMMWORD[r11]
> + sub rdx,0x80
> + jnc NEAR $L$ecb_dec_loop8
> +
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + mov rcx,r11
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + mov eax,r10d
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + movups XMMWORD[48+rsi],xmm5
> + pxor xmm5,xmm5
> + movups XMMWORD[64+rsi],xmm6
> + pxor xmm6,xmm6
> + movups XMMWORD[80+rsi],xmm7
> + pxor xmm7,xmm7
> + movups XMMWORD[96+rsi],xmm8
> + pxor xmm8,xmm8
> + movups XMMWORD[112+rsi],xmm9
> + pxor xmm9,xmm9
> + lea rsi,[128+rsi]
> + add rdx,0x80
> + jz NEAR $L$ecb_ret
> +
> +$L$ecb_dec_tail:
> + movups xmm2,XMMWORD[rdi]
> + cmp rdx,0x20
> + jb NEAR $L$ecb_dec_one
> + movups xmm3,XMMWORD[16+rdi]
> + je NEAR $L$ecb_dec_two
> + movups xmm4,XMMWORD[32+rdi]
> + cmp rdx,0x40
> + jb NEAR $L$ecb_dec_three
> + movups xmm5,XMMWORD[48+rdi]
> + je NEAR $L$ecb_dec_four
> + movups xmm6,XMMWORD[64+rdi]
> + cmp rdx,0x60
> + jb NEAR $L$ecb_dec_five
> + movups xmm7,XMMWORD[80+rdi]
> + je NEAR $L$ecb_dec_six
> + movups xmm8,XMMWORD[96+rdi]
> + movups xmm0,XMMWORD[rcx]
> + xorps xmm9,xmm9
> + call _aesni_decrypt8
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + movups XMMWORD[48+rsi],xmm5
> + pxor xmm5,xmm5
> + movups XMMWORD[64+rsi],xmm6
> + pxor xmm6,xmm6
> + movups XMMWORD[80+rsi],xmm7
> + pxor xmm7,xmm7
> + movups XMMWORD[96+rsi],xmm8
> + pxor xmm8,xmm8
> + pxor xmm9,xmm9
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_dec_one:
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_dec1_4:
> +DB 102,15,56,222,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_dec1_4
> +DB 102,15,56,223,209
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_dec_two:
> + call _aesni_decrypt2
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_dec_three:
> + call _aesni_decrypt3
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_dec_four:
> + call _aesni_decrypt4
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + movups XMMWORD[48+rsi],xmm5
> + pxor xmm5,xmm5
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_dec_five:
> + xorps xmm7,xmm7
> + call _aesni_decrypt6
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + movups XMMWORD[48+rsi],xmm5
> + pxor xmm5,xmm5
> + movups XMMWORD[64+rsi],xmm6
> + pxor xmm6,xmm6
> + pxor xmm7,xmm7
> + jmp NEAR $L$ecb_ret
> +ALIGN 16
> +$L$ecb_dec_six:
> + call _aesni_decrypt6
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + movups XMMWORD[48+rsi],xmm5
> + pxor xmm5,xmm5
> + movups XMMWORD[64+rsi],xmm6
> + pxor xmm6,xmm6
> + movups XMMWORD[80+rsi],xmm7
> + pxor xmm7,xmm7
> +
> +$L$ecb_ret:
> + xorps xmm0,xmm0
> + pxor xmm1,xmm1
> + movaps xmm6,XMMWORD[rsp]
> + movaps XMMWORD[rsp],xmm0
> + movaps xmm7,XMMWORD[16+rsp]
> + movaps XMMWORD[16+rsp],xmm0
> + movaps xmm8,XMMWORD[32+rsp]
> + movaps XMMWORD[32+rsp],xmm0
> + movaps xmm9,XMMWORD[48+rsp]
> + movaps XMMWORD[48+rsp],xmm0
> + lea rsp,[88+rsp]
> +$L$ecb_enc_ret:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_ecb_encrypt:
> +global aesni_ccm64_encrypt_blocks
> +
> +ALIGN 16
> +aesni_ccm64_encrypt_blocks:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_ccm64_encrypt_blocks:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + lea rsp,[((-88))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> +$L$ccm64_enc_body:
> + mov eax,DWORD[240+rcx]
> + movdqu xmm6,XMMWORD[r8]
> + movdqa xmm9,XMMWORD[$L$increment64]
> + movdqa xmm7,XMMWORD[$L$bswap_mask]
> +
> + shl eax,4
> + mov r10d,16
> + lea r11,[rcx]
> + movdqu xmm3,XMMWORD[r9]
> + movdqa xmm2,xmm6
> + lea rcx,[32+rax*1+rcx]
> +DB 102,15,56,0,247
> + sub r10,rax
> + jmp NEAR $L$ccm64_enc_outer
> +ALIGN 16
> +$L$ccm64_enc_outer:
> + movups xmm0,XMMWORD[r11]
> + mov rax,r10
> + movups xmm8,XMMWORD[rdi]
> +
> + xorps xmm2,xmm0
> + movups xmm1,XMMWORD[16+r11]
> + xorps xmm0,xmm8
> + xorps xmm3,xmm0
> + movups xmm0,XMMWORD[32+r11]
> +
> +$L$ccm64_enc2_loop:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$ccm64_enc2_loop
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> + paddq xmm6,xmm9
> + dec rdx
> +DB 102,15,56,221,208
> +DB 102,15,56,221,216
> +
> + lea rdi,[16+rdi]
> + xorps xmm8,xmm2
> + movdqa xmm2,xmm6
> + movups XMMWORD[rsi],xmm8
> +DB 102,15,56,0,215
> + lea rsi,[16+rsi]
> + jnz NEAR $L$ccm64_enc_outer
> +
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + movups XMMWORD[r9],xmm3
> + pxor xmm3,xmm3
> + pxor xmm8,xmm8
> + pxor xmm6,xmm6
> + movaps xmm6,XMMWORD[rsp]
> + movaps XMMWORD[rsp],xmm0
> + movaps xmm7,XMMWORD[16+rsp]
> + movaps XMMWORD[16+rsp],xmm0
> + movaps xmm8,XMMWORD[32+rsp]
> + movaps XMMWORD[32+rsp],xmm0
> + movaps xmm9,XMMWORD[48+rsp]
> + movaps XMMWORD[48+rsp],xmm0
> + lea rsp,[88+rsp]
> +$L$ccm64_enc_ret:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_ccm64_encrypt_blocks:
> +global aesni_ccm64_decrypt_blocks
> +
> +ALIGN 16
> +aesni_ccm64_decrypt_blocks:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_ccm64_decrypt_blocks:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + lea rsp,[((-88))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> +$L$ccm64_dec_body:
> + mov eax,DWORD[240+rcx]
> + movups xmm6,XMMWORD[r8]
> + movdqu xmm3,XMMWORD[r9]
> + movdqa xmm9,XMMWORD[$L$increment64]
> + movdqa xmm7,XMMWORD[$L$bswap_mask]
> +
> + movaps xmm2,xmm6
> + mov r10d,eax
> + mov r11,rcx
> +DB 102,15,56,0,247
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_enc1_5:
> +DB 102,15,56,220,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_enc1_5
> +DB 102,15,56,221,209
> + shl r10d,4
> + mov eax,16
> + movups xmm8,XMMWORD[rdi]
> + paddq xmm6,xmm9
> + lea rdi,[16+rdi]
> + sub rax,r10
> + lea rcx,[32+r10*1+r11]
> + mov r10,rax
> + jmp NEAR $L$ccm64_dec_outer
> +ALIGN 16
> +$L$ccm64_dec_outer:
> + xorps xmm8,xmm2
> + movdqa xmm2,xmm6
> + movups XMMWORD[rsi],xmm8
> + lea rsi,[16+rsi]
> +DB 102,15,56,0,215
> +
> + sub rdx,1
> + jz NEAR $L$ccm64_dec_break
> +
> + movups xmm0,XMMWORD[r11]
> + mov rax,r10
> + movups xmm1,XMMWORD[16+r11]
> + xorps xmm8,xmm0
> + xorps xmm2,xmm0
> + xorps xmm3,xmm8
> + movups xmm0,XMMWORD[32+r11]
> + jmp NEAR $L$ccm64_dec2_loop
> +ALIGN 16
> +$L$ccm64_dec2_loop:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$ccm64_dec2_loop
> + movups xmm8,XMMWORD[rdi]
> + paddq xmm6,xmm9
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,221,208
> +DB 102,15,56,221,216
> + lea rdi,[16+rdi]
> + jmp NEAR $L$ccm64_dec_outer
> +
> +ALIGN 16
> +$L$ccm64_dec_break:
> +
> + mov eax,DWORD[240+r11]
> + movups xmm0,XMMWORD[r11]
> + movups xmm1,XMMWORD[16+r11]
> + xorps xmm8,xmm0
> + lea r11,[32+r11]
> + xorps xmm3,xmm8
> +$L$oop_enc1_6:
> +DB 102,15,56,220,217
> + dec eax
> + movups xmm1,XMMWORD[r11]
> + lea r11,[16+r11]
> + jnz NEAR $L$oop_enc1_6
> +DB 102,15,56,221,217
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + movups XMMWORD[r9],xmm3
> + pxor xmm3,xmm3
> + pxor xmm8,xmm8
> + pxor xmm6,xmm6
> + movaps xmm6,XMMWORD[rsp]
> + movaps XMMWORD[rsp],xmm0
> + movaps xmm7,XMMWORD[16+rsp]
> + movaps XMMWORD[16+rsp],xmm0
> + movaps xmm8,XMMWORD[32+rsp]
> + movaps XMMWORD[32+rsp],xmm0
> + movaps xmm9,XMMWORD[48+rsp]
> + movaps XMMWORD[48+rsp],xmm0
> + lea rsp,[88+rsp]
> +$L$ccm64_dec_ret:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_ccm64_decrypt_blocks:
> +global aesni_ctr32_encrypt_blocks
> +
> +ALIGN 16
> +aesni_ctr32_encrypt_blocks:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_ctr32_encrypt_blocks:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> +
> +
> +
> + cmp rdx,1
> + jne NEAR $L$ctr32_bulk
> +
> +
> +
> + movups xmm2,XMMWORD[r8]
> + movups xmm3,XMMWORD[rdi]
> + mov edx,DWORD[240+rcx]
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_enc1_7:
> +DB 102,15,56,220,209
> + dec edx
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_enc1_7
> +DB 102,15,56,221,209
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + xorps xmm2,xmm3
> + pxor xmm3,xmm3
> + movups XMMWORD[rsi],xmm2
> + xorps xmm2,xmm2
> + jmp NEAR $L$ctr32_epilogue
> +
> +ALIGN 16
> +$L$ctr32_bulk:
> + lea r11,[rsp]
> +
> + push rbp
> +
> + sub rsp,288
> + and rsp,-16
> + movaps XMMWORD[(-168)+r11],xmm6
> + movaps XMMWORD[(-152)+r11],xmm7
> + movaps XMMWORD[(-136)+r11],xmm8
> + movaps XMMWORD[(-120)+r11],xmm9
> + movaps XMMWORD[(-104)+r11],xmm10
> + movaps XMMWORD[(-88)+r11],xmm11
> + movaps XMMWORD[(-72)+r11],xmm12
> + movaps XMMWORD[(-56)+r11],xmm13
> + movaps XMMWORD[(-40)+r11],xmm14
> + movaps XMMWORD[(-24)+r11],xmm15
> +$L$ctr32_body:
> +
> +
> +
> +
> + movdqu xmm2,XMMWORD[r8]
> + movdqu xmm0,XMMWORD[rcx]
> + mov r8d,DWORD[12+r8]
> + pxor xmm2,xmm0
> + mov ebp,DWORD[12+rcx]
> + movdqa XMMWORD[rsp],xmm2
> + bswap r8d
> + movdqa xmm3,xmm2
> + movdqa xmm4,xmm2
> + movdqa xmm5,xmm2
> + movdqa XMMWORD[64+rsp],xmm2
> + movdqa XMMWORD[80+rsp],xmm2
> + movdqa XMMWORD[96+rsp],xmm2
> + mov r10,rdx
> + movdqa XMMWORD[112+rsp],xmm2
> +
> + lea rax,[1+r8]
> + lea rdx,[2+r8]
> + bswap eax
> + bswap edx
> + xor eax,ebp
> + xor edx,ebp
> +DB 102,15,58,34,216,3
> + lea rax,[3+r8]
> + movdqa XMMWORD[16+rsp],xmm3
> +DB 102,15,58,34,226,3
> + bswap eax
> + mov rdx,r10
> + lea r10,[4+r8]
> + movdqa XMMWORD[32+rsp],xmm4
> + xor eax,ebp
> + bswap r10d
> +DB 102,15,58,34,232,3
> + xor r10d,ebp
> + movdqa XMMWORD[48+rsp],xmm5
> + lea r9,[5+r8]
> + mov DWORD[((64+12))+rsp],r10d
> + bswap r9d
> + lea r10,[6+r8]
> + mov eax,DWORD[240+rcx]
> + xor r9d,ebp
> + bswap r10d
> + mov DWORD[((80+12))+rsp],r9d
> + xor r10d,ebp
> + lea r9,[7+r8]
> + mov DWORD[((96+12))+rsp],r10d
> + bswap r9d
> + mov r10d,DWORD[((OPENSSL_ia32cap_P+4))]
> + xor r9d,ebp
> + and r10d,71303168
> + mov DWORD[((112+12))+rsp],r9d
> +
> + movups xmm1,XMMWORD[16+rcx]
> +
> + movdqa xmm6,XMMWORD[64+rsp]
> + movdqa xmm7,XMMWORD[80+rsp]
> +
> + cmp rdx,8
> + jb NEAR $L$ctr32_tail
> +
> + sub rdx,6
> + cmp r10d,4194304
> + je NEAR $L$ctr32_6x
> +
> + lea rcx,[128+rcx]
> + sub rdx,2
> + jmp NEAR $L$ctr32_loop8
> +
> +ALIGN 16
> +$L$ctr32_6x:
> + shl eax,4
> + mov r10d,48
> + bswap ebp
> + lea rcx,[32+rax*1+rcx]
> + sub r10,rax
> + jmp NEAR $L$ctr32_loop6
> +
> +ALIGN 16
> +$L$ctr32_loop6:
> + add r8d,6
> + movups xmm0,XMMWORD[((-48))+r10*1+rcx]
> +DB 102,15,56,220,209
> + mov eax,r8d
> + xor eax,ebp
> +DB 102,15,56,220,217
> +DB 0x0f,0x38,0xf1,0x44,0x24,12
> + lea eax,[1+r8]
> +DB 102,15,56,220,225
> + xor eax,ebp
> +DB 0x0f,0x38,0xf1,0x44,0x24,28
> +DB 102,15,56,220,233
> + lea eax,[2+r8]
> + xor eax,ebp
> +DB 102,15,56,220,241
> +DB 0x0f,0x38,0xf1,0x44,0x24,44
> + lea eax,[3+r8]
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[((-32))+r10*1+rcx]
> + xor eax,ebp
> +
> +DB 102,15,56,220,208
> +DB 0x0f,0x38,0xf1,0x44,0x24,60
> + lea eax,[4+r8]
> +DB 102,15,56,220,216
> + xor eax,ebp
> +DB 0x0f,0x38,0xf1,0x44,0x24,76
> +DB 102,15,56,220,224
> + lea eax,[5+r8]
> + xor eax,ebp
> +DB 102,15,56,220,232
> +DB 0x0f,0x38,0xf1,0x44,0x24,92
> + mov rax,r10
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> + movups xmm0,XMMWORD[((-16))+r10*1+rcx]
> +
> + call $L$enc_loop6
> +
> + movdqu xmm8,XMMWORD[rdi]
> + movdqu xmm9,XMMWORD[16+rdi]
> + movdqu xmm10,XMMWORD[32+rdi]
> + movdqu xmm11,XMMWORD[48+rdi]
> + movdqu xmm12,XMMWORD[64+rdi]
> + movdqu xmm13,XMMWORD[80+rdi]
> + lea rdi,[96+rdi]
> + movups xmm1,XMMWORD[((-64))+r10*1+rcx]
> + pxor xmm8,xmm2
> + movaps xmm2,XMMWORD[rsp]
> + pxor xmm9,xmm3
> + movaps xmm3,XMMWORD[16+rsp]
> + pxor xmm10,xmm4
> + movaps xmm4,XMMWORD[32+rsp]
> + pxor xmm11,xmm5
> + movaps xmm5,XMMWORD[48+rsp]
> + pxor xmm12,xmm6
> + movaps xmm6,XMMWORD[64+rsp]
> + pxor xmm13,xmm7
> + movaps xmm7,XMMWORD[80+rsp]
> + movdqu XMMWORD[rsi],xmm8
> + movdqu XMMWORD[16+rsi],xmm9
> + movdqu XMMWORD[32+rsi],xmm10
> + movdqu XMMWORD[48+rsi],xmm11
> + movdqu XMMWORD[64+rsi],xmm12
> + movdqu XMMWORD[80+rsi],xmm13
> + lea rsi,[96+rsi]
> +
> + sub rdx,6
> + jnc NEAR $L$ctr32_loop6
> +
> + add rdx,6
> + jz NEAR $L$ctr32_done
> +
> + lea eax,[((-48))+r10]
> + lea rcx,[((-80))+r10*1+rcx]
> + neg eax
> + shr eax,4
> + jmp NEAR $L$ctr32_tail
> +
> +ALIGN 32
> +$L$ctr32_loop8:
> + add r8d,8
> + movdqa xmm8,XMMWORD[96+rsp]
> +DB 102,15,56,220,209
> + mov r9d,r8d
> + movdqa xmm9,XMMWORD[112+rsp]
> +DB 102,15,56,220,217
> + bswap r9d
> + movups xmm0,XMMWORD[((32-128))+rcx]
> +DB 102,15,56,220,225
> + xor r9d,ebp
> + nop
> +DB 102,15,56,220,233
> + mov DWORD[((0+12))+rsp],r9d
> + lea r9,[1+r8]
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> + movups xmm1,XMMWORD[((48-128))+rcx]
> + bswap r9d
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> + xor r9d,ebp
> +DB 0x66,0x90
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + mov DWORD[((16+12))+rsp],r9d
> + lea r9,[2+r8]
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> +DB 102,68,15,56,220,192
> +DB 102,68,15,56,220,200
> + movups xmm0,XMMWORD[((64-128))+rcx]
> + bswap r9d
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> + xor r9d,ebp
> +DB 0x66,0x90
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + mov DWORD[((32+12))+rsp],r9d
> + lea r9,[3+r8]
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> + movups xmm1,XMMWORD[((80-128))+rcx]
> + bswap r9d
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> + xor r9d,ebp
> +DB 0x66,0x90
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + mov DWORD[((48+12))+rsp],r9d
> + lea r9,[4+r8]
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> +DB 102,68,15,56,220,192
> +DB 102,68,15,56,220,200
> + movups xmm0,XMMWORD[((96-128))+rcx]
> + bswap r9d
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> + xor r9d,ebp
> +DB 0x66,0x90
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + mov DWORD[((64+12))+rsp],r9d
> + lea r9,[5+r8]
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> + movups xmm1,XMMWORD[((112-128))+rcx]
> + bswap r9d
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> + xor r9d,ebp
> +DB 0x66,0x90
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + mov DWORD[((80+12))+rsp],r9d
> + lea r9,[6+r8]
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> +DB 102,68,15,56,220,192
> +DB 102,68,15,56,220,200
> + movups xmm0,XMMWORD[((128-128))+rcx]
> + bswap r9d
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> + xor r9d,ebp
> +DB 0x66,0x90
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + mov DWORD[((96+12))+rsp],r9d
> + lea r9,[7+r8]
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> + movups xmm1,XMMWORD[((144-128))+rcx]
> + bswap r9d
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> + xor r9d,ebp
> + movdqu xmm10,XMMWORD[rdi]
> +DB 102,15,56,220,232
> + mov DWORD[((112+12))+rsp],r9d
> + cmp eax,11
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> +DB 102,68,15,56,220,192
> +DB 102,68,15,56,220,200
> + movups xmm0,XMMWORD[((160-128))+rcx]
> +
> + jb NEAR $L$ctr32_enc_done
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> + movups xmm1,XMMWORD[((176-128))+rcx]
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> +DB 102,68,15,56,220,192
> +DB 102,68,15,56,220,200
> + movups xmm0,XMMWORD[((192-128))+rcx]
> + je NEAR $L$ctr32_enc_done
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> + movups xmm1,XMMWORD[((208-128))+rcx]
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> +DB 102,68,15,56,220,192
> +DB 102,68,15,56,220,200
> + movups xmm0,XMMWORD[((224-128))+rcx]
> + jmp NEAR $L$ctr32_enc_done
> +
> +ALIGN 16
> +$L$ctr32_enc_done:
> + movdqu xmm11,XMMWORD[16+rdi]
> + pxor xmm10,xmm0
> + movdqu xmm12,XMMWORD[32+rdi]
> + pxor xmm11,xmm0
> + movdqu xmm13,XMMWORD[48+rdi]
> + pxor xmm12,xmm0
> + movdqu xmm14,XMMWORD[64+rdi]
> + pxor xmm13,xmm0
> + movdqu xmm15,XMMWORD[80+rdi]
> + pxor xmm14,xmm0
> + pxor xmm15,xmm0
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +DB 102,68,15,56,220,201
> + movdqu xmm1,XMMWORD[96+rdi]
> + lea rdi,[128+rdi]
> +
> +DB 102,65,15,56,221,210
> + pxor xmm1,xmm0
> + movdqu xmm10,XMMWORD[((112-128))+rdi]
> +DB 102,65,15,56,221,219
> + pxor xmm10,xmm0
> + movdqa xmm11,XMMWORD[rsp]
> +DB 102,65,15,56,221,228
> +DB 102,65,15,56,221,237
> + movdqa xmm12,XMMWORD[16+rsp]
> + movdqa xmm13,XMMWORD[32+rsp]
> +DB 102,65,15,56,221,246
> +DB 102,65,15,56,221,255
> + movdqa xmm14,XMMWORD[48+rsp]
> + movdqa xmm15,XMMWORD[64+rsp]
> +DB 102,68,15,56,221,193
> + movdqa xmm0,XMMWORD[80+rsp]
> + movups xmm1,XMMWORD[((16-128))+rcx]
> +DB 102,69,15,56,221,202
> +
> + movups XMMWORD[rsi],xmm2
> + movdqa xmm2,xmm11
> + movups XMMWORD[16+rsi],xmm3
> + movdqa xmm3,xmm12
> + movups XMMWORD[32+rsi],xmm4
> + movdqa xmm4,xmm13
> + movups XMMWORD[48+rsi],xmm5
> + movdqa xmm5,xmm14
> + movups XMMWORD[64+rsi],xmm6
> + movdqa xmm6,xmm15
> + movups XMMWORD[80+rsi],xmm7
> + movdqa xmm7,xmm0
> + movups XMMWORD[96+rsi],xmm8
> + movups XMMWORD[112+rsi],xmm9
> + lea rsi,[128+rsi]
> +
> + sub rdx,8
> + jnc NEAR $L$ctr32_loop8
> +
> + add rdx,8
> + jz NEAR $L$ctr32_done
> + lea rcx,[((-128))+rcx]
> +
> +$L$ctr32_tail:
> +
> +
> + lea rcx,[16+rcx]
> + cmp rdx,4
> + jb NEAR $L$ctr32_loop3
> + je NEAR $L$ctr32_loop4
> +
> +
> + shl eax,4
> + movdqa xmm8,XMMWORD[96+rsp]
> + pxor xmm9,xmm9
> +
> + movups xmm0,XMMWORD[16+rcx]
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> + lea rcx,[((32-16))+rax*1+rcx]
> + neg rax
> +DB 102,15,56,220,225
> + add rax,16
> + movups xmm10,XMMWORD[rdi]
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> + movups xmm11,XMMWORD[16+rdi]
> + movups xmm12,XMMWORD[32+rdi]
> +DB 102,15,56,220,249
> +DB 102,68,15,56,220,193
> +
> + call $L$enc_loop8_enter
> +
> + movdqu xmm13,XMMWORD[48+rdi]
> + pxor xmm2,xmm10
> + movdqu xmm10,XMMWORD[64+rdi]
> + pxor xmm3,xmm11
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm4,xmm12
> + movdqu XMMWORD[16+rsi],xmm3
> + pxor xmm5,xmm13
> + movdqu XMMWORD[32+rsi],xmm4
> + pxor xmm6,xmm10
> + movdqu XMMWORD[48+rsi],xmm5
> + movdqu XMMWORD[64+rsi],xmm6
> + cmp rdx,6
> + jb NEAR $L$ctr32_done
> +
> + movups xmm11,XMMWORD[80+rdi]
> + xorps xmm7,xmm11
> + movups XMMWORD[80+rsi],xmm7
> + je NEAR $L$ctr32_done
> +
> + movups xmm12,XMMWORD[96+rdi]
> + xorps xmm8,xmm12
> + movups XMMWORD[96+rsi],xmm8
> + jmp NEAR $L$ctr32_done
> +
> +ALIGN 32
> +$L$ctr32_loop4:
> +DB 102,15,56,220,209
> + lea rcx,[16+rcx]
> + dec eax
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[rcx]
> + jnz NEAR $L$ctr32_loop4
> +DB 102,15,56,221,209
> +DB 102,15,56,221,217
> + movups xmm10,XMMWORD[rdi]
> + movups xmm11,XMMWORD[16+rdi]
> +DB 102,15,56,221,225
> +DB 102,15,56,221,233
> + movups xmm12,XMMWORD[32+rdi]
> + movups xmm13,XMMWORD[48+rdi]
> +
> + xorps xmm2,xmm10
> + movups XMMWORD[rsi],xmm2
> + xorps xmm3,xmm11
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm4,xmm12
> + movdqu XMMWORD[32+rsi],xmm4
> + pxor xmm5,xmm13
> + movdqu XMMWORD[48+rsi],xmm5
> + jmp NEAR $L$ctr32_done
> +
> +ALIGN 32
> +$L$ctr32_loop3:
> +DB 102,15,56,220,209
> + lea rcx,[16+rcx]
> + dec eax
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> + movups xmm1,XMMWORD[rcx]
> + jnz NEAR $L$ctr32_loop3
> +DB 102,15,56,221,209
> +DB 102,15,56,221,217
> +DB 102,15,56,221,225
> +
> + movups xmm10,XMMWORD[rdi]
> + xorps xmm2,xmm10
> + movups XMMWORD[rsi],xmm2
> + cmp rdx,2
> + jb NEAR $L$ctr32_done
> +
> + movups xmm11,XMMWORD[16+rdi]
> + xorps xmm3,xmm11
> + movups XMMWORD[16+rsi],xmm3
> + je NEAR $L$ctr32_done
> +
> + movups xmm12,XMMWORD[32+rdi]
> + xorps xmm4,xmm12
> + movups XMMWORD[32+rsi],xmm4
> +
> +$L$ctr32_done:
> + xorps xmm0,xmm0
> + xor ebp,ebp
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> + movaps xmm6,XMMWORD[((-168))+r11]
> + movaps XMMWORD[(-168)+r11],xmm0
> + movaps xmm7,XMMWORD[((-152))+r11]
> + movaps XMMWORD[(-152)+r11],xmm0
> + movaps xmm8,XMMWORD[((-136))+r11]
> + movaps XMMWORD[(-136)+r11],xmm0
> + movaps xmm9,XMMWORD[((-120))+r11]
> + movaps XMMWORD[(-120)+r11],xmm0
> + movaps xmm10,XMMWORD[((-104))+r11]
> + movaps XMMWORD[(-104)+r11],xmm0
> + movaps xmm11,XMMWORD[((-88))+r11]
> + movaps XMMWORD[(-88)+r11],xmm0
> + movaps xmm12,XMMWORD[((-72))+r11]
> + movaps XMMWORD[(-72)+r11],xmm0
> + movaps xmm13,XMMWORD[((-56))+r11]
> + movaps XMMWORD[(-56)+r11],xmm0
> + movaps xmm14,XMMWORD[((-40))+r11]
> + movaps XMMWORD[(-40)+r11],xmm0
> + movaps xmm15,XMMWORD[((-24))+r11]
> + movaps XMMWORD[(-24)+r11],xmm0
> + movaps XMMWORD[rsp],xmm0
> + movaps XMMWORD[16+rsp],xmm0
> + movaps XMMWORD[32+rsp],xmm0
> + movaps XMMWORD[48+rsp],xmm0
> + movaps XMMWORD[64+rsp],xmm0
> + movaps XMMWORD[80+rsp],xmm0
> + movaps XMMWORD[96+rsp],xmm0
> + movaps XMMWORD[112+rsp],xmm0
> + mov rbp,QWORD[((-8))+r11]
> +
> + lea rsp,[r11]
> +
> +$L$ctr32_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_ctr32_encrypt_blocks:
> +global aesni_xts_encrypt
> +
> +ALIGN 16
> +aesni_xts_encrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_xts_encrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + lea r11,[rsp]
> +
> + push rbp
> +
> + sub rsp,272
> + and rsp,-16
> + movaps XMMWORD[(-168)+r11],xmm6
> + movaps XMMWORD[(-152)+r11],xmm7
> + movaps XMMWORD[(-136)+r11],xmm8
> + movaps XMMWORD[(-120)+r11],xmm9
> + movaps XMMWORD[(-104)+r11],xmm10
> + movaps XMMWORD[(-88)+r11],xmm11
> + movaps XMMWORD[(-72)+r11],xmm12
> + movaps XMMWORD[(-56)+r11],xmm13
> + movaps XMMWORD[(-40)+r11],xmm14
> + movaps XMMWORD[(-24)+r11],xmm15
> +$L$xts_enc_body:
> + movups xmm2,XMMWORD[r9]
> + mov eax,DWORD[240+r8]
> + mov r10d,DWORD[240+rcx]
> + movups xmm0,XMMWORD[r8]
> + movups xmm1,XMMWORD[16+r8]
> + lea r8,[32+r8]
> + xorps xmm2,xmm0
> +$L$oop_enc1_8:
> +DB 102,15,56,220,209
> + dec eax
> + movups xmm1,XMMWORD[r8]
> + lea r8,[16+r8]
> + jnz NEAR $L$oop_enc1_8
> +DB 102,15,56,221,209
> + movups xmm0,XMMWORD[rcx]
> + mov rbp,rcx
> + mov eax,r10d
> + shl r10d,4
> + mov r9,rdx
> + and rdx,-16
> +
> + movups xmm1,XMMWORD[16+r10*1+rcx]
> +
> + movdqa xmm8,XMMWORD[$L$xts_magic]
> + movdqa xmm15,xmm2
> + pshufd xmm9,xmm2,0x5f
> + pxor xmm1,xmm0
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> + movdqa xmm10,xmm15
> + psrad xmm14,31
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> + pxor xmm10,xmm0
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> + movdqa xmm11,xmm15
> + psrad xmm14,31
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> + pxor xmm11,xmm0
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> + movdqa xmm12,xmm15
> + psrad xmm14,31
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> + pxor xmm12,xmm0
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> + movdqa xmm13,xmm15
> + psrad xmm14,31
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> + pxor xmm13,xmm0
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm15
> + psrad xmm9,31
> + paddq xmm15,xmm15
> + pand xmm9,xmm8
> + pxor xmm14,xmm0
> + pxor xmm15,xmm9
> + movaps XMMWORD[96+rsp],xmm1
> +
> + sub rdx,16*6
> + jc NEAR $L$xts_enc_short
> +
> + mov eax,16+96
> + lea rcx,[32+r10*1+rbp]
> + sub rax,r10
> + movups xmm1,XMMWORD[16+rbp]
> + mov r10,rax
> + lea r8,[$L$xts_magic]
> + jmp NEAR $L$xts_enc_grandloop
> +
> +ALIGN 32
> +$L$xts_enc_grandloop:
> + movdqu xmm2,XMMWORD[rdi]
> + movdqa xmm8,xmm0
> + movdqu xmm3,XMMWORD[16+rdi]
> + pxor xmm2,xmm10
> + movdqu xmm4,XMMWORD[32+rdi]
> + pxor xmm3,xmm11
> +DB 102,15,56,220,209
> + movdqu xmm5,XMMWORD[48+rdi]
> + pxor xmm4,xmm12
> +DB 102,15,56,220,217
> + movdqu xmm6,XMMWORD[64+rdi]
> + pxor xmm5,xmm13
> +DB 102,15,56,220,225
> + movdqu xmm7,XMMWORD[80+rdi]
> + pxor xmm8,xmm15
> + movdqa xmm9,XMMWORD[96+rsp]
> + pxor xmm6,xmm14
> +DB 102,15,56,220,233
> + movups xmm0,XMMWORD[32+rbp]
> + lea rdi,[96+rdi]
> + pxor xmm7,xmm8
> +
> + pxor xmm10,xmm9
> +DB 102,15,56,220,241
> + pxor xmm11,xmm9
> + movdqa XMMWORD[rsp],xmm10
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[48+rbp]
> + pxor xmm12,xmm9
> +
> +DB 102,15,56,220,208
> + pxor xmm13,xmm9
> + movdqa XMMWORD[16+rsp],xmm11
> +DB 102,15,56,220,216
> + pxor xmm14,xmm9
> + movdqa XMMWORD[32+rsp],xmm12
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + pxor xmm8,xmm9
> + movdqa XMMWORD[64+rsp],xmm14
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> + movups xmm0,XMMWORD[64+rbp]
> + movdqa XMMWORD[80+rsp],xmm8
> + pshufd xmm9,xmm15,0x5f
> + jmp NEAR $L$xts_enc_loop6
> +ALIGN 32
> +$L$xts_enc_loop6:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[((-64))+rax*1+rcx]
> + add rax,32
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> + movups xmm0,XMMWORD[((-80))+rax*1+rcx]
> + jnz NEAR $L$xts_enc_loop6
> +
> + movdqa xmm8,XMMWORD[r8]
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> +DB 102,15,56,220,209
> + paddq xmm15,xmm15
> + psrad xmm14,31
> +DB 102,15,56,220,217
> + pand xmm14,xmm8
> + movups xmm10,XMMWORD[rbp]
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> + pxor xmm15,xmm14
> + movaps xmm11,xmm10
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[((-64))+rcx]
> +
> + movdqa xmm14,xmm9
> +DB 102,15,56,220,208
> + paddd xmm9,xmm9
> + pxor xmm10,xmm15
> +DB 102,15,56,220,216
> + psrad xmm14,31
> + paddq xmm15,xmm15
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + pand xmm14,xmm8
> + movaps xmm12,xmm11
> +DB 102,15,56,220,240
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm9
> +DB 102,15,56,220,248
> + movups xmm0,XMMWORD[((-48))+rcx]
> +
> + paddd xmm9,xmm9
> +DB 102,15,56,220,209
> + pxor xmm11,xmm15
> + psrad xmm14,31
> +DB 102,15,56,220,217
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movdqa XMMWORD[48+rsp],xmm13
> + pxor xmm15,xmm14
> +DB 102,15,56,220,241
> + movaps xmm13,xmm12
> + movdqa xmm14,xmm9
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[((-32))+rcx]
> +
> + paddd xmm9,xmm9
> +DB 102,15,56,220,208
> + pxor xmm12,xmm15
> + psrad xmm14,31
> +DB 102,15,56,220,216
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> +DB 102,15,56,220,240
> + pxor xmm15,xmm14
> + movaps xmm14,xmm13
> +DB 102,15,56,220,248
> +
> + movdqa xmm0,xmm9
> + paddd xmm9,xmm9
> +DB 102,15,56,220,209
> + pxor xmm13,xmm15
> + psrad xmm0,31
> +DB 102,15,56,220,217
> + paddq xmm15,xmm15
> + pand xmm0,xmm8
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + pxor xmm15,xmm0
> + movups xmm0,XMMWORD[rbp]
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[16+rbp]
> +
> + pxor xmm14,xmm15
> +DB 102,15,56,221,84,36,0
> + psrad xmm9,31
> + paddq xmm15,xmm15
> +DB 102,15,56,221,92,36,16
> +DB 102,15,56,221,100,36,32
> + pand xmm9,xmm8
> + mov rax,r10
> +DB 102,15,56,221,108,36,48
> +DB 102,15,56,221,116,36,64
> +DB 102,15,56,221,124,36,80
> + pxor xmm15,xmm9
> +
> + lea rsi,[96+rsi]
> + movups XMMWORD[(-96)+rsi],xmm2
> + movups XMMWORD[(-80)+rsi],xmm3
> + movups XMMWORD[(-64)+rsi],xmm4
> + movups XMMWORD[(-48)+rsi],xmm5
> + movups XMMWORD[(-32)+rsi],xmm6
> + movups XMMWORD[(-16)+rsi],xmm7
> + sub rdx,16*6
> + jnc NEAR $L$xts_enc_grandloop
> +
> + mov eax,16+96
> + sub eax,r10d
> + mov rcx,rbp
> + shr eax,4
> +
> +$L$xts_enc_short:
> +
> + mov r10d,eax
> + pxor xmm10,xmm0
> + add rdx,16*6
> + jz NEAR $L$xts_enc_done
> +
> + pxor xmm11,xmm0
> + cmp rdx,0x20
> + jb NEAR $L$xts_enc_one
> + pxor xmm12,xmm0
> + je NEAR $L$xts_enc_two
> +
> + pxor xmm13,xmm0
> + cmp rdx,0x40
> + jb NEAR $L$xts_enc_three
> + pxor xmm14,xmm0
> + je NEAR $L$xts_enc_four
> +
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[16+rdi]
> + movdqu xmm4,XMMWORD[32+rdi]
> + pxor xmm2,xmm10
> + movdqu xmm5,XMMWORD[48+rdi]
> + pxor xmm3,xmm11
> + movdqu xmm6,XMMWORD[64+rdi]
> + lea rdi,[80+rdi]
> + pxor xmm4,xmm12
> + pxor xmm5,xmm13
> + pxor xmm6,xmm14
> + pxor xmm7,xmm7
> +
> + call _aesni_encrypt6
> +
> + xorps xmm2,xmm10
> + movdqa xmm10,xmm15
> + xorps xmm3,xmm11
> + xorps xmm4,xmm12
> + movdqu XMMWORD[rsi],xmm2
> + xorps xmm5,xmm13
> + movdqu XMMWORD[16+rsi],xmm3
> + xorps xmm6,xmm14
> + movdqu XMMWORD[32+rsi],xmm4
> + movdqu XMMWORD[48+rsi],xmm5
> + movdqu XMMWORD[64+rsi],xmm6
> + lea rsi,[80+rsi]
> + jmp NEAR $L$xts_enc_done
> +
> +ALIGN 16
> +$L$xts_enc_one:
> + movups xmm2,XMMWORD[rdi]
> + lea rdi,[16+rdi]
> + xorps xmm2,xmm10
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_enc1_9:
> +DB 102,15,56,220,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_enc1_9
> +DB 102,15,56,221,209
> + xorps xmm2,xmm10
> + movdqa xmm10,xmm11
> + movups XMMWORD[rsi],xmm2
> + lea rsi,[16+rsi]
> + jmp NEAR $L$xts_enc_done
> +
> +ALIGN 16
> +$L$xts_enc_two:
> + movups xmm2,XMMWORD[rdi]
> + movups xmm3,XMMWORD[16+rdi]
> + lea rdi,[32+rdi]
> + xorps xmm2,xmm10
> + xorps xmm3,xmm11
> +
> + call _aesni_encrypt2
> +
> + xorps xmm2,xmm10
> + movdqa xmm10,xmm12
> + xorps xmm3,xmm11
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + lea rsi,[32+rsi]
> + jmp NEAR $L$xts_enc_done
> +
> +ALIGN 16
> +$L$xts_enc_three:
> + movups xmm2,XMMWORD[rdi]
> + movups xmm3,XMMWORD[16+rdi]
> + movups xmm4,XMMWORD[32+rdi]
> + lea rdi,[48+rdi]
> + xorps xmm2,xmm10
> + xorps xmm3,xmm11
> + xorps xmm4,xmm12
> +
> + call _aesni_encrypt3
> +
> + xorps xmm2,xmm10
> + movdqa xmm10,xmm13
> + xorps xmm3,xmm11
> + xorps xmm4,xmm12
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + lea rsi,[48+rsi]
> + jmp NEAR $L$xts_enc_done
> +
> +ALIGN 16
> +$L$xts_enc_four:
> + movups xmm2,XMMWORD[rdi]
> + movups xmm3,XMMWORD[16+rdi]
> + movups xmm4,XMMWORD[32+rdi]
> + xorps xmm2,xmm10
> + movups xmm5,XMMWORD[48+rdi]
> + lea rdi,[64+rdi]
> + xorps xmm3,xmm11
> + xorps xmm4,xmm12
> + xorps xmm5,xmm13
> +
> + call _aesni_encrypt4
> +
> + pxor xmm2,xmm10
> + movdqa xmm10,xmm14
> + pxor xmm3,xmm11
> + pxor xmm4,xmm12
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm5,xmm13
> + movdqu XMMWORD[16+rsi],xmm3
> + movdqu XMMWORD[32+rsi],xmm4
> + movdqu XMMWORD[48+rsi],xmm5
> + lea rsi,[64+rsi]
> + jmp NEAR $L$xts_enc_done
> +
> +ALIGN 16
> +$L$xts_enc_done:
> + and r9,15
> + jz NEAR $L$xts_enc_ret
> + mov rdx,r9
> +
> +$L$xts_enc_steal:
> + movzx eax,BYTE[rdi]
> + movzx ecx,BYTE[((-16))+rsi]
> + lea rdi,[1+rdi]
> + mov BYTE[((-16))+rsi],al
> + mov BYTE[rsi],cl
> + lea rsi,[1+rsi]
> + sub rdx,1
> + jnz NEAR $L$xts_enc_steal
> +
> + sub rsi,r9
> + mov rcx,rbp
> + mov eax,r10d
> +
> + movups xmm2,XMMWORD[((-16))+rsi]
> + xorps xmm2,xmm10
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_enc1_10:
> +DB 102,15,56,220,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_enc1_10
> +DB 102,15,56,221,209
> + xorps xmm2,xmm10
> + movups XMMWORD[(-16)+rsi],xmm2
> +
> +$L$xts_enc_ret:
> + xorps xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> + movaps xmm6,XMMWORD[((-168))+r11]
> + movaps XMMWORD[(-168)+r11],xmm0
> + movaps xmm7,XMMWORD[((-152))+r11]
> + movaps XMMWORD[(-152)+r11],xmm0
> + movaps xmm8,XMMWORD[((-136))+r11]
> + movaps XMMWORD[(-136)+r11],xmm0
> + movaps xmm9,XMMWORD[((-120))+r11]
> + movaps XMMWORD[(-120)+r11],xmm0
> + movaps xmm10,XMMWORD[((-104))+r11]
> + movaps XMMWORD[(-104)+r11],xmm0
> + movaps xmm11,XMMWORD[((-88))+r11]
> + movaps XMMWORD[(-88)+r11],xmm0
> + movaps xmm12,XMMWORD[((-72))+r11]
> + movaps XMMWORD[(-72)+r11],xmm0
> + movaps xmm13,XMMWORD[((-56))+r11]
> + movaps XMMWORD[(-56)+r11],xmm0
> + movaps xmm14,XMMWORD[((-40))+r11]
> + movaps XMMWORD[(-40)+r11],xmm0
> + movaps xmm15,XMMWORD[((-24))+r11]
> + movaps XMMWORD[(-24)+r11],xmm0
> + movaps XMMWORD[rsp],xmm0
> + movaps XMMWORD[16+rsp],xmm0
> + movaps XMMWORD[32+rsp],xmm0
> + movaps XMMWORD[48+rsp],xmm0
> + movaps XMMWORD[64+rsp],xmm0
> + movaps XMMWORD[80+rsp],xmm0
> + movaps XMMWORD[96+rsp],xmm0
> + mov rbp,QWORD[((-8))+r11]
> +
> + lea rsp,[r11]
> +
> +$L$xts_enc_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_xts_encrypt:
> +global aesni_xts_decrypt
> +
> +ALIGN 16
> +aesni_xts_decrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_xts_decrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + lea r11,[rsp]
> +
> + push rbp
> +
> + sub rsp,272
> + and rsp,-16
> + movaps XMMWORD[(-168)+r11],xmm6
> + movaps XMMWORD[(-152)+r11],xmm7
> + movaps XMMWORD[(-136)+r11],xmm8
> + movaps XMMWORD[(-120)+r11],xmm9
> + movaps XMMWORD[(-104)+r11],xmm10
> + movaps XMMWORD[(-88)+r11],xmm11
> + movaps XMMWORD[(-72)+r11],xmm12
> + movaps XMMWORD[(-56)+r11],xmm13
> + movaps XMMWORD[(-40)+r11],xmm14
> + movaps XMMWORD[(-24)+r11],xmm15
> +$L$xts_dec_body:
> + movups xmm2,XMMWORD[r9]
> + mov eax,DWORD[240+r8]
> + mov r10d,DWORD[240+rcx]
> + movups xmm0,XMMWORD[r8]
> + movups xmm1,XMMWORD[16+r8]
> + lea r8,[32+r8]
> + xorps xmm2,xmm0
> +$L$oop_enc1_11:
> +DB 102,15,56,220,209
> + dec eax
> + movups xmm1,XMMWORD[r8]
> + lea r8,[16+r8]
> + jnz NEAR $L$oop_enc1_11
> +DB 102,15,56,221,209
> + xor eax,eax
> + test rdx,15
> + setnz al
> + shl rax,4
> + sub rdx,rax
> +
> + movups xmm0,XMMWORD[rcx]
> + mov rbp,rcx
> + mov eax,r10d
> + shl r10d,4
> + mov r9,rdx
> + and rdx,-16
> +
> + movups xmm1,XMMWORD[16+r10*1+rcx]
> +
> + movdqa xmm8,XMMWORD[$L$xts_magic]
> + movdqa xmm15,xmm2
> + pshufd xmm9,xmm2,0x5f
> + pxor xmm1,xmm0
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> + movdqa xmm10,xmm15
> + psrad xmm14,31
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> + pxor xmm10,xmm0
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> + movdqa xmm11,xmm15
> + psrad xmm14,31
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> + pxor xmm11,xmm0
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> + movdqa xmm12,xmm15
> + psrad xmm14,31
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> + pxor xmm12,xmm0
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> + movdqa xmm13,xmm15
> + psrad xmm14,31
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> + pxor xmm13,xmm0
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm15
> + psrad xmm9,31
> + paddq xmm15,xmm15
> + pand xmm9,xmm8
> + pxor xmm14,xmm0
> + pxor xmm15,xmm9
> + movaps XMMWORD[96+rsp],xmm1
> +
> + sub rdx,16*6
> + jc NEAR $L$xts_dec_short
> +
> + mov eax,16+96
> + lea rcx,[32+r10*1+rbp]
> + sub rax,r10
> + movups xmm1,XMMWORD[16+rbp]
> + mov r10,rax
> + lea r8,[$L$xts_magic]
> + jmp NEAR $L$xts_dec_grandloop
> +
> +ALIGN 32
> +$L$xts_dec_grandloop:
> + movdqu xmm2,XMMWORD[rdi]
> + movdqa xmm8,xmm0
> + movdqu xmm3,XMMWORD[16+rdi]
> + pxor xmm2,xmm10
> + movdqu xmm4,XMMWORD[32+rdi]
> + pxor xmm3,xmm11
> +DB 102,15,56,222,209
> + movdqu xmm5,XMMWORD[48+rdi]
> + pxor xmm4,xmm12
> +DB 102,15,56,222,217
> + movdqu xmm6,XMMWORD[64+rdi]
> + pxor xmm5,xmm13
> +DB 102,15,56,222,225
> + movdqu xmm7,XMMWORD[80+rdi]
> + pxor xmm8,xmm15
> + movdqa xmm9,XMMWORD[96+rsp]
> + pxor xmm6,xmm14
> +DB 102,15,56,222,233
> + movups xmm0,XMMWORD[32+rbp]
> + lea rdi,[96+rdi]
> + pxor xmm7,xmm8
> +
> + pxor xmm10,xmm9
> +DB 102,15,56,222,241
> + pxor xmm11,xmm9
> + movdqa XMMWORD[rsp],xmm10
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[48+rbp]
> + pxor xmm12,xmm9
> +
> +DB 102,15,56,222,208
> + pxor xmm13,xmm9
> + movdqa XMMWORD[16+rsp],xmm11
> +DB 102,15,56,222,216
> + pxor xmm14,xmm9
> + movdqa XMMWORD[32+rsp],xmm12
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + pxor xmm8,xmm9
> + movdqa XMMWORD[64+rsp],xmm14
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> + movups xmm0,XMMWORD[64+rbp]
> + movdqa XMMWORD[80+rsp],xmm8
> + pshufd xmm9,xmm15,0x5f
> + jmp NEAR $L$xts_dec_loop6
> +ALIGN 32
> +$L$xts_dec_loop6:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[((-64))+rax*1+rcx]
> + add rax,32
> +
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> + movups xmm0,XMMWORD[((-80))+rax*1+rcx]
> + jnz NEAR $L$xts_dec_loop6
> +
> + movdqa xmm8,XMMWORD[r8]
> + movdqa xmm14,xmm9
> + paddd xmm9,xmm9
> +DB 102,15,56,222,209
> + paddq xmm15,xmm15
> + psrad xmm14,31
> +DB 102,15,56,222,217
> + pand xmm14,xmm8
> + movups xmm10,XMMWORD[rbp]
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> + pxor xmm15,xmm14
> + movaps xmm11,xmm10
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[((-64))+rcx]
> +
> + movdqa xmm14,xmm9
> +DB 102,15,56,222,208
> + paddd xmm9,xmm9
> + pxor xmm10,xmm15
> +DB 102,15,56,222,216
> + psrad xmm14,31
> + paddq xmm15,xmm15
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + pand xmm14,xmm8
> + movaps xmm12,xmm11
> +DB 102,15,56,222,240
> + pxor xmm15,xmm14
> + movdqa xmm14,xmm9
> +DB 102,15,56,222,248
> + movups xmm0,XMMWORD[((-48))+rcx]
> +
> + paddd xmm9,xmm9
> +DB 102,15,56,222,209
> + pxor xmm11,xmm15
> + psrad xmm14,31
> +DB 102,15,56,222,217
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movdqa XMMWORD[48+rsp],xmm13
> + pxor xmm15,xmm14
> +DB 102,15,56,222,241
> + movaps xmm13,xmm12
> + movdqa xmm14,xmm9
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[((-32))+rcx]
> +
> + paddd xmm9,xmm9
> +DB 102,15,56,222,208
> + pxor xmm12,xmm15
> + psrad xmm14,31
> +DB 102,15,56,222,216
> + paddq xmm15,xmm15
> + pand xmm14,xmm8
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> + pxor xmm15,xmm14
> + movaps xmm14,xmm13
> +DB 102,15,56,222,248
> +
> + movdqa xmm0,xmm9
> + paddd xmm9,xmm9
> +DB 102,15,56,222,209
> + pxor xmm13,xmm15
> + psrad xmm0,31
> +DB 102,15,56,222,217
> + paddq xmm15,xmm15
> + pand xmm0,xmm8
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + pxor xmm15,xmm0
> + movups xmm0,XMMWORD[rbp]
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[16+rbp]
> +
> + pxor xmm14,xmm15
> +DB 102,15,56,223,84,36,0
> + psrad xmm9,31
> + paddq xmm15,xmm15
> +DB 102,15,56,223,92,36,16
> +DB 102,15,56,223,100,36,32
> + pand xmm9,xmm8
> + mov rax,r10
> +DB 102,15,56,223,108,36,48
> +DB 102,15,56,223,116,36,64
> +DB 102,15,56,223,124,36,80
> + pxor xmm15,xmm9
> +
> + lea rsi,[96+rsi]
> + movups XMMWORD[(-96)+rsi],xmm2
> + movups XMMWORD[(-80)+rsi],xmm3
> + movups XMMWORD[(-64)+rsi],xmm4
> + movups XMMWORD[(-48)+rsi],xmm5
> + movups XMMWORD[(-32)+rsi],xmm6
> + movups XMMWORD[(-16)+rsi],xmm7
> + sub rdx,16*6
> + jnc NEAR $L$xts_dec_grandloop
> +
> + mov eax,16+96
> + sub eax,r10d
> + mov rcx,rbp
> + shr eax,4
> +
> +$L$xts_dec_short:
> +
> + mov r10d,eax
> + pxor xmm10,xmm0
> + pxor xmm11,xmm0
> + add rdx,16*6
> + jz NEAR $L$xts_dec_done
> +
> + pxor xmm12,xmm0
> + cmp rdx,0x20
> + jb NEAR $L$xts_dec_one
> + pxor xmm13,xmm0
> + je NEAR $L$xts_dec_two
> +
> + pxor xmm14,xmm0
> + cmp rdx,0x40
> + jb NEAR $L$xts_dec_three
> + je NEAR $L$xts_dec_four
> +
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[16+rdi]
> + movdqu xmm4,XMMWORD[32+rdi]
> + pxor xmm2,xmm10
> + movdqu xmm5,XMMWORD[48+rdi]
> + pxor xmm3,xmm11
> + movdqu xmm6,XMMWORD[64+rdi]
> + lea rdi,[80+rdi]
> + pxor xmm4,xmm12
> + pxor xmm5,xmm13
> + pxor xmm6,xmm14
> +
> + call _aesni_decrypt6
> +
> + xorps xmm2,xmm10
> + xorps xmm3,xmm11
> + xorps xmm4,xmm12
> + movdqu XMMWORD[rsi],xmm2
> + xorps xmm5,xmm13
> + movdqu XMMWORD[16+rsi],xmm3
> + xorps xmm6,xmm14
> + movdqu XMMWORD[32+rsi],xmm4
> + pxor xmm14,xmm14
> + movdqu XMMWORD[48+rsi],xmm5
> + pcmpgtd xmm14,xmm15
> + movdqu XMMWORD[64+rsi],xmm6
> + lea rsi,[80+rsi]
> + pshufd xmm11,xmm14,0x13
> + and r9,15
> + jz NEAR $L$xts_dec_ret
> +
> + movdqa xmm10,xmm15
> + paddq xmm15,xmm15
> + pand xmm11,xmm8
> + pxor xmm11,xmm15
> + jmp NEAR $L$xts_dec_done2
> +
> +ALIGN 16
> +$L$xts_dec_one:
> + movups xmm2,XMMWORD[rdi]
> + lea rdi,[16+rdi]
> + xorps xmm2,xmm10
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_dec1_12:
> +DB 102,15,56,222,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_dec1_12
> +DB 102,15,56,223,209
> + xorps xmm2,xmm10
> + movdqa xmm10,xmm11
> + movups XMMWORD[rsi],xmm2
> + movdqa xmm11,xmm12
> + lea rsi,[16+rsi]
> + jmp NEAR $L$xts_dec_done
> +
> +ALIGN 16
> +$L$xts_dec_two:
> + movups xmm2,XMMWORD[rdi]
> + movups xmm3,XMMWORD[16+rdi]
> + lea rdi,[32+rdi]
> + xorps xmm2,xmm10
> + xorps xmm3,xmm11
> +
> + call _aesni_decrypt2
> +
> + xorps xmm2,xmm10
> + movdqa xmm10,xmm12
> + xorps xmm3,xmm11
> + movdqa xmm11,xmm13
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + lea rsi,[32+rsi]
> + jmp NEAR $L$xts_dec_done
> +
> +ALIGN 16
> +$L$xts_dec_three:
> + movups xmm2,XMMWORD[rdi]
> + movups xmm3,XMMWORD[16+rdi]
> + movups xmm4,XMMWORD[32+rdi]
> + lea rdi,[48+rdi]
> + xorps xmm2,xmm10
> + xorps xmm3,xmm11
> + xorps xmm4,xmm12
> +
> + call _aesni_decrypt3
> +
> + xorps xmm2,xmm10
> + movdqa xmm10,xmm13
> + xorps xmm3,xmm11
> + movdqa xmm11,xmm14
> + xorps xmm4,xmm12
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + lea rsi,[48+rsi]
> + jmp NEAR $L$xts_dec_done
> +
> +ALIGN 16
> +$L$xts_dec_four:
> + movups xmm2,XMMWORD[rdi]
> + movups xmm3,XMMWORD[16+rdi]
> + movups xmm4,XMMWORD[32+rdi]
> + xorps xmm2,xmm10
> + movups xmm5,XMMWORD[48+rdi]
> + lea rdi,[64+rdi]
> + xorps xmm3,xmm11
> + xorps xmm4,xmm12
> + xorps xmm5,xmm13
> +
> + call _aesni_decrypt4
> +
> + pxor xmm2,xmm10
> + movdqa xmm10,xmm14
> + pxor xmm3,xmm11
> + movdqa xmm11,xmm15
> + pxor xmm4,xmm12
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm5,xmm13
> + movdqu XMMWORD[16+rsi],xmm3
> + movdqu XMMWORD[32+rsi],xmm4
> + movdqu XMMWORD[48+rsi],xmm5
> + lea rsi,[64+rsi]
> + jmp NEAR $L$xts_dec_done
> +
> +ALIGN 16
> +$L$xts_dec_done:
> + and r9,15
> + jz NEAR $L$xts_dec_ret
> +$L$xts_dec_done2:
> + mov rdx,r9
> + mov rcx,rbp
> + mov eax,r10d
> +
> + movups xmm2,XMMWORD[rdi]
> + xorps xmm2,xmm11
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_dec1_13:
> +DB 102,15,56,222,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_dec1_13
> +DB 102,15,56,223,209
> + xorps xmm2,xmm11
> + movups XMMWORD[rsi],xmm2
> +
> +$L$xts_dec_steal:
> + movzx eax,BYTE[16+rdi]
> + movzx ecx,BYTE[rsi]
> + lea rdi,[1+rdi]
> + mov BYTE[rsi],al
> + mov BYTE[16+rsi],cl
> + lea rsi,[1+rsi]
> + sub rdx,1
> + jnz NEAR $L$xts_dec_steal
> +
> + sub rsi,r9
> + mov rcx,rbp
> + mov eax,r10d
> +
> + movups xmm2,XMMWORD[rsi]
> + xorps xmm2,xmm10
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_dec1_14:
> +DB 102,15,56,222,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_dec1_14
> +DB 102,15,56,223,209
> + xorps xmm2,xmm10
> + movups XMMWORD[rsi],xmm2
> +
> +$L$xts_dec_ret:
> + xorps xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> + movaps xmm6,XMMWORD[((-168))+r11]
> + movaps XMMWORD[(-168)+r11],xmm0
> + movaps xmm7,XMMWORD[((-152))+r11]
> + movaps XMMWORD[(-152)+r11],xmm0
> + movaps xmm8,XMMWORD[((-136))+r11]
> + movaps XMMWORD[(-136)+r11],xmm0
> + movaps xmm9,XMMWORD[((-120))+r11]
> + movaps XMMWORD[(-120)+r11],xmm0
> + movaps xmm10,XMMWORD[((-104))+r11]
> + movaps XMMWORD[(-104)+r11],xmm0
> + movaps xmm11,XMMWORD[((-88))+r11]
> + movaps XMMWORD[(-88)+r11],xmm0
> + movaps xmm12,XMMWORD[((-72))+r11]
> + movaps XMMWORD[(-72)+r11],xmm0
> + movaps xmm13,XMMWORD[((-56))+r11]
> + movaps XMMWORD[(-56)+r11],xmm0
> + movaps xmm14,XMMWORD[((-40))+r11]
> + movaps XMMWORD[(-40)+r11],xmm0
> + movaps xmm15,XMMWORD[((-24))+r11]
> + movaps XMMWORD[(-24)+r11],xmm0
> + movaps XMMWORD[rsp],xmm0
> + movaps XMMWORD[16+rsp],xmm0
> + movaps XMMWORD[32+rsp],xmm0
> + movaps XMMWORD[48+rsp],xmm0
> + movaps XMMWORD[64+rsp],xmm0
> + movaps XMMWORD[80+rsp],xmm0
> + movaps XMMWORD[96+rsp],xmm0
> + mov rbp,QWORD[((-8))+r11]
> +
> + lea rsp,[r11]
> +
> +$L$xts_dec_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_xts_decrypt:
> +global aesni_ocb_encrypt
> +
> +ALIGN 32
> +aesni_ocb_encrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_ocb_encrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + lea rax,[rsp]
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + lea rsp,[((-160))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> + movaps XMMWORD[64+rsp],xmm10
> + movaps XMMWORD[80+rsp],xmm11
> + movaps XMMWORD[96+rsp],xmm12
> + movaps XMMWORD[112+rsp],xmm13
> + movaps XMMWORD[128+rsp],xmm14
> + movaps XMMWORD[144+rsp],xmm15
> +$L$ocb_enc_body:
> + mov rbx,QWORD[56+rax]
> + mov rbp,QWORD[((56+8))+rax]
> +
> + mov r10d,DWORD[240+rcx]
> + mov r11,rcx
> + shl r10d,4
> + movups xmm9,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+r10*1+rcx]
> +
> + movdqu xmm15,XMMWORD[r9]
> + pxor xmm9,xmm1
> + pxor xmm15,xmm1
> +
> + mov eax,16+32
> + lea rcx,[32+r10*1+r11]
> + movups xmm1,XMMWORD[16+r11]
> + sub rax,r10
> + mov r10,rax
> +
> + movdqu xmm10,XMMWORD[rbx]
> + movdqu xmm8,XMMWORD[rbp]
> +
> + test r8,1
> + jnz NEAR $L$ocb_enc_odd
> +
> + bsf r12,r8
> + add r8,1
> + shl r12,4
> + movdqu xmm7,XMMWORD[r12*1+rbx]
> + movdqu xmm2,XMMWORD[rdi]
> + lea rdi,[16+rdi]
> +
> + call __ocb_encrypt1
> +
> + movdqa xmm15,xmm7
> + movups XMMWORD[rsi],xmm2
> + lea rsi,[16+rsi]
> + sub rdx,1
> + jz NEAR $L$ocb_enc_done
> +
> +$L$ocb_enc_odd:
> + lea r12,[1+r8]
> + lea r13,[3+r8]
> + lea r14,[5+r8]
> + lea r8,[6+r8]
> + bsf r12,r12
> + bsf r13,r13
> + bsf r14,r14
> + shl r12,4
> + shl r13,4
> + shl r14,4
> +
> + sub rdx,6
> + jc NEAR $L$ocb_enc_short
> + jmp NEAR $L$ocb_enc_grandloop
> +
> +ALIGN 32
> +$L$ocb_enc_grandloop:
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[16+rdi]
> + movdqu xmm4,XMMWORD[32+rdi]
> + movdqu xmm5,XMMWORD[48+rdi]
> + movdqu xmm6,XMMWORD[64+rdi]
> + movdqu xmm7,XMMWORD[80+rdi]
> + lea rdi,[96+rdi]
> +
> + call __ocb_encrypt6
> +
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + movups XMMWORD[48+rsi],xmm5
> + movups XMMWORD[64+rsi],xmm6
> + movups XMMWORD[80+rsi],xmm7
> + lea rsi,[96+rsi]
> + sub rdx,6
> + jnc NEAR $L$ocb_enc_grandloop
> +
> +$L$ocb_enc_short:
> + add rdx,6
> + jz NEAR $L$ocb_enc_done
> +
> + movdqu xmm2,XMMWORD[rdi]
> + cmp rdx,2
> + jb NEAR $L$ocb_enc_one
> + movdqu xmm3,XMMWORD[16+rdi]
> + je NEAR $L$ocb_enc_two
> +
> + movdqu xmm4,XMMWORD[32+rdi]
> + cmp rdx,4
> + jb NEAR $L$ocb_enc_three
> + movdqu xmm5,XMMWORD[48+rdi]
> + je NEAR $L$ocb_enc_four
> +
> + movdqu xmm6,XMMWORD[64+rdi]
> + pxor xmm7,xmm7
> +
> + call __ocb_encrypt6
> +
> + movdqa xmm15,xmm14
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + movups XMMWORD[48+rsi],xmm5
> + movups XMMWORD[64+rsi],xmm6
> +
> + jmp NEAR $L$ocb_enc_done
> +
> +ALIGN 16
> +$L$ocb_enc_one:
> + movdqa xmm7,xmm10
> +
> + call __ocb_encrypt1
> +
> + movdqa xmm15,xmm7
> + movups XMMWORD[rsi],xmm2
> + jmp NEAR $L$ocb_enc_done
> +
> +ALIGN 16
> +$L$ocb_enc_two:
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> +
> + call __ocb_encrypt4
> +
> + movdqa xmm15,xmm11
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> +
> + jmp NEAR $L$ocb_enc_done
> +
> +ALIGN 16
> +$L$ocb_enc_three:
> + pxor xmm5,xmm5
> +
> + call __ocb_encrypt4
> +
> + movdqa xmm15,xmm12
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> +
> + jmp NEAR $L$ocb_enc_done
> +
> +ALIGN 16
> +$L$ocb_enc_four:
> + call __ocb_encrypt4
> +
> + movdqa xmm15,xmm13
> + movups XMMWORD[rsi],xmm2
> + movups XMMWORD[16+rsi],xmm3
> + movups XMMWORD[32+rsi],xmm4
> + movups XMMWORD[48+rsi],xmm5
> +
> +$L$ocb_enc_done:
> + pxor xmm15,xmm0
> + movdqu XMMWORD[rbp],xmm8
> + movdqu XMMWORD[r9],xmm15
> +
> + xorps xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> + movaps xmm6,XMMWORD[rsp]
> + movaps XMMWORD[rsp],xmm0
> + movaps xmm7,XMMWORD[16+rsp]
> + movaps XMMWORD[16+rsp],xmm0
> + movaps xmm8,XMMWORD[32+rsp]
> + movaps XMMWORD[32+rsp],xmm0
> + movaps xmm9,XMMWORD[48+rsp]
> + movaps XMMWORD[48+rsp],xmm0
> + movaps xmm10,XMMWORD[64+rsp]
> + movaps XMMWORD[64+rsp],xmm0
> + movaps xmm11,XMMWORD[80+rsp]
> + movaps XMMWORD[80+rsp],xmm0
> + movaps xmm12,XMMWORD[96+rsp]
> + movaps XMMWORD[96+rsp],xmm0
> + movaps xmm13,XMMWORD[112+rsp]
> + movaps XMMWORD[112+rsp],xmm0
> + movaps xmm14,XMMWORD[128+rsp]
> + movaps XMMWORD[128+rsp],xmm0
> + movaps xmm15,XMMWORD[144+rsp]
> + movaps XMMWORD[144+rsp],xmm0
> + lea rax,[((160+40))+rsp]
> +$L$ocb_enc_pop:
> + mov r14,QWORD[((-40))+rax]
> +
> + mov r13,QWORD[((-32))+rax]
> +
> + mov r12,QWORD[((-24))+rax]
> +
> + mov rbp,QWORD[((-16))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> +
> + lea rsp,[rax]
> +
> +$L$ocb_enc_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_ocb_encrypt:
> +
> +
> +ALIGN 32
> +__ocb_encrypt6:
> +
> + pxor xmm15,xmm9
> + movdqu xmm11,XMMWORD[r12*1+rbx]
> + movdqa xmm12,xmm10
> + movdqu xmm13,XMMWORD[r13*1+rbx]
> + movdqa xmm14,xmm10
> + pxor xmm10,xmm15
> + movdqu xmm15,XMMWORD[r14*1+rbx]
> + pxor xmm11,xmm10
> + pxor xmm8,xmm2
> + pxor xmm2,xmm10
> + pxor xmm12,xmm11
> + pxor xmm8,xmm3
> + pxor xmm3,xmm11
> + pxor xmm13,xmm12
> + pxor xmm8,xmm4
> + pxor xmm4,xmm12
> + pxor xmm14,xmm13
> + pxor xmm8,xmm5
> + pxor xmm5,xmm13
> + pxor xmm15,xmm14
> + pxor xmm8,xmm6
> + pxor xmm6,xmm14
> + pxor xmm8,xmm7
> + pxor xmm7,xmm15
> + movups xmm0,XMMWORD[32+r11]
> +
> + lea r12,[1+r8]
> + lea r13,[3+r8]
> + lea r14,[5+r8]
> + add r8,6
> + pxor xmm10,xmm9
> + bsf r12,r12
> + bsf r13,r13
> + bsf r14,r14
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + pxor xmm11,xmm9
> + pxor xmm12,xmm9
> +DB 102,15,56,220,241
> + pxor xmm13,xmm9
> + pxor xmm14,xmm9
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[48+r11]
> + pxor xmm15,xmm9
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> + movups xmm0,XMMWORD[64+r11]
> + shl r12,4
> + shl r13,4
> + jmp NEAR $L$ocb_enc_loop6
> +
> +ALIGN 32
> +$L$ocb_enc_loop6:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> +DB 102,15,56,220,240
> +DB 102,15,56,220,248
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$ocb_enc_loop6
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> +DB 102,15,56,220,241
> +DB 102,15,56,220,249
> + movups xmm1,XMMWORD[16+r11]
> + shl r14,4
> +
> +DB 102,65,15,56,221,210
> + movdqu xmm10,XMMWORD[rbx]
> + mov rax,r10
> +DB 102,65,15,56,221,219
> +DB 102,65,15,56,221,228
> +DB 102,65,15,56,221,237
> +DB 102,65,15,56,221,246
> +DB 102,65,15,56,221,255
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +ALIGN 32
> +__ocb_encrypt4:
> +
> + pxor xmm15,xmm9
> + movdqu xmm11,XMMWORD[r12*1+rbx]
> + movdqa xmm12,xmm10
> + movdqu xmm13,XMMWORD[r13*1+rbx]
> + pxor xmm10,xmm15
> + pxor xmm11,xmm10
> + pxor xmm8,xmm2
> + pxor xmm2,xmm10
> + pxor xmm12,xmm11
> + pxor xmm8,xmm3
> + pxor xmm3,xmm11
> + pxor xmm13,xmm12
> + pxor xmm8,xmm4
> + pxor xmm4,xmm12
> + pxor xmm8,xmm5
> + pxor xmm5,xmm13
> + movups xmm0,XMMWORD[32+r11]
> +
> + pxor xmm10,xmm9
> + pxor xmm11,xmm9
> + pxor xmm12,xmm9
> + pxor xmm13,xmm9
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[48+r11]
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[64+r11]
> + jmp NEAR $L$ocb_enc_loop4
> +
> +ALIGN 32
> +$L$ocb_enc_loop4:
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +
> +DB 102,15,56,220,208
> +DB 102,15,56,220,216
> +DB 102,15,56,220,224
> +DB 102,15,56,220,232
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$ocb_enc_loop4
> +
> +DB 102,15,56,220,209
> +DB 102,15,56,220,217
> +DB 102,15,56,220,225
> +DB 102,15,56,220,233
> + movups xmm1,XMMWORD[16+r11]
> + mov rax,r10
> +
> +DB 102,65,15,56,221,210
> +DB 102,65,15,56,221,219
> +DB 102,65,15,56,221,228
> +DB 102,65,15,56,221,237
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +ALIGN 32
> +__ocb_encrypt1:
> +
> + pxor xmm7,xmm15
> + pxor xmm7,xmm9
> + pxor xmm8,xmm2
> + pxor xmm2,xmm7
> + movups xmm0,XMMWORD[32+r11]
> +
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[48+r11]
> + pxor xmm7,xmm9
> +
> +DB 102,15,56,220,208
> + movups xmm0,XMMWORD[64+r11]
> + jmp NEAR $L$ocb_enc_loop1
> +
> +ALIGN 32
> +$L$ocb_enc_loop1:
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +
> +DB 102,15,56,220,208
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$ocb_enc_loop1
> +
> +DB 102,15,56,220,209
> + movups xmm1,XMMWORD[16+r11]
> + mov rax,r10
> +
> +DB 102,15,56,221,215
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +global aesni_ocb_decrypt
> +
> +ALIGN 32
> +aesni_ocb_decrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_ocb_decrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + lea rax,[rsp]
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + lea rsp,[((-160))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> + movaps XMMWORD[64+rsp],xmm10
> + movaps XMMWORD[80+rsp],xmm11
> + movaps XMMWORD[96+rsp],xmm12
> + movaps XMMWORD[112+rsp],xmm13
> + movaps XMMWORD[128+rsp],xmm14
> + movaps XMMWORD[144+rsp],xmm15
> +$L$ocb_dec_body:
> + mov rbx,QWORD[56+rax]
> + mov rbp,QWORD[((56+8))+rax]
> +
> + mov r10d,DWORD[240+rcx]
> + mov r11,rcx
> + shl r10d,4
> + movups xmm9,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+r10*1+rcx]
> +
> + movdqu xmm15,XMMWORD[r9]
> + pxor xmm9,xmm1
> + pxor xmm15,xmm1
> +
> + mov eax,16+32
> + lea rcx,[32+r10*1+r11]
> + movups xmm1,XMMWORD[16+r11]
> + sub rax,r10
> + mov r10,rax
> +
> + movdqu xmm10,XMMWORD[rbx]
> + movdqu xmm8,XMMWORD[rbp]
> +
> + test r8,1
> + jnz NEAR $L$ocb_dec_odd
> +
> + bsf r12,r8
> + add r8,1
> + shl r12,4
> + movdqu xmm7,XMMWORD[r12*1+rbx]
> + movdqu xmm2,XMMWORD[rdi]
> + lea rdi,[16+rdi]
> +
> + call __ocb_decrypt1
> +
> + movdqa xmm15,xmm7
> + movups XMMWORD[rsi],xmm2
> + xorps xmm8,xmm2
> + lea rsi,[16+rsi]
> + sub rdx,1
> + jz NEAR $L$ocb_dec_done
> +
> +$L$ocb_dec_odd:
> + lea r12,[1+r8]
> + lea r13,[3+r8]
> + lea r14,[5+r8]
> + lea r8,[6+r8]
> + bsf r12,r12
> + bsf r13,r13
> + bsf r14,r14
> + shl r12,4
> + shl r13,4
> + shl r14,4
> +
> + sub rdx,6
> + jc NEAR $L$ocb_dec_short
> + jmp NEAR $L$ocb_dec_grandloop
> +
> +ALIGN 32
> +$L$ocb_dec_grandloop:
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[16+rdi]
> + movdqu xmm4,XMMWORD[32+rdi]
> + movdqu xmm5,XMMWORD[48+rdi]
> + movdqu xmm6,XMMWORD[64+rdi]
> + movdqu xmm7,XMMWORD[80+rdi]
> + lea rdi,[96+rdi]
> +
> + call __ocb_decrypt6
> +
> + movups XMMWORD[rsi],xmm2
> + pxor xmm8,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm8,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm8,xmm4
> + movups XMMWORD[48+rsi],xmm5
> + pxor xmm8,xmm5
> + movups XMMWORD[64+rsi],xmm6
> + pxor xmm8,xmm6
> + movups XMMWORD[80+rsi],xmm7
> + pxor xmm8,xmm7
> + lea rsi,[96+rsi]
> + sub rdx,6
> + jnc NEAR $L$ocb_dec_grandloop
> +
> +$L$ocb_dec_short:
> + add rdx,6
> + jz NEAR $L$ocb_dec_done
> +
> + movdqu xmm2,XMMWORD[rdi]
> + cmp rdx,2
> + jb NEAR $L$ocb_dec_one
> + movdqu xmm3,XMMWORD[16+rdi]
> + je NEAR $L$ocb_dec_two
> +
> + movdqu xmm4,XMMWORD[32+rdi]
> + cmp rdx,4
> + jb NEAR $L$ocb_dec_three
> + movdqu xmm5,XMMWORD[48+rdi]
> + je NEAR $L$ocb_dec_four
> +
> + movdqu xmm6,XMMWORD[64+rdi]
> + pxor xmm7,xmm7
> +
> + call __ocb_decrypt6
> +
> + movdqa xmm15,xmm14
> + movups XMMWORD[rsi],xmm2
> + pxor xmm8,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm8,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm8,xmm4
> + movups XMMWORD[48+rsi],xmm5
> + pxor xmm8,xmm5
> + movups XMMWORD[64+rsi],xmm6
> + pxor xmm8,xmm6
> +
> + jmp NEAR $L$ocb_dec_done
> +
> +ALIGN 16
> +$L$ocb_dec_one:
> + movdqa xmm7,xmm10
> +
> + call __ocb_decrypt1
> +
> + movdqa xmm15,xmm7
> + movups XMMWORD[rsi],xmm2
> + xorps xmm8,xmm2
> + jmp NEAR $L$ocb_dec_done
> +
> +ALIGN 16
> +$L$ocb_dec_two:
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> +
> + call __ocb_decrypt4
> +
> + movdqa xmm15,xmm11
> + movups XMMWORD[rsi],xmm2
> + xorps xmm8,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + xorps xmm8,xmm3
> +
> + jmp NEAR $L$ocb_dec_done
> +
> +ALIGN 16
> +$L$ocb_dec_three:
> + pxor xmm5,xmm5
> +
> + call __ocb_decrypt4
> +
> + movdqa xmm15,xmm12
> + movups XMMWORD[rsi],xmm2
> + xorps xmm8,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + xorps xmm8,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + xorps xmm8,xmm4
> +
> + jmp NEAR $L$ocb_dec_done
> +
> +ALIGN 16
> +$L$ocb_dec_four:
> + call __ocb_decrypt4
> +
> + movdqa xmm15,xmm13
> + movups XMMWORD[rsi],xmm2
> + pxor xmm8,xmm2
> + movups XMMWORD[16+rsi],xmm3
> + pxor xmm8,xmm3
> + movups XMMWORD[32+rsi],xmm4
> + pxor xmm8,xmm4
> + movups XMMWORD[48+rsi],xmm5
> + pxor xmm8,xmm5
> +
> +$L$ocb_dec_done:
> + pxor xmm15,xmm0
> + movdqu XMMWORD[rbp],xmm8
> + movdqu XMMWORD[r9],xmm15
> +
> + xorps xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> + movaps xmm6,XMMWORD[rsp]
> + movaps XMMWORD[rsp],xmm0
> + movaps xmm7,XMMWORD[16+rsp]
> + movaps XMMWORD[16+rsp],xmm0
> + movaps xmm8,XMMWORD[32+rsp]
> + movaps XMMWORD[32+rsp],xmm0
> + movaps xmm9,XMMWORD[48+rsp]
> + movaps XMMWORD[48+rsp],xmm0
> + movaps xmm10,XMMWORD[64+rsp]
> + movaps XMMWORD[64+rsp],xmm0
> + movaps xmm11,XMMWORD[80+rsp]
> + movaps XMMWORD[80+rsp],xmm0
> + movaps xmm12,XMMWORD[96+rsp]
> + movaps XMMWORD[96+rsp],xmm0
> + movaps xmm13,XMMWORD[112+rsp]
> + movaps XMMWORD[112+rsp],xmm0
> + movaps xmm14,XMMWORD[128+rsp]
> + movaps XMMWORD[128+rsp],xmm0
> + movaps xmm15,XMMWORD[144+rsp]
> + movaps XMMWORD[144+rsp],xmm0
> + lea rax,[((160+40))+rsp]
> +$L$ocb_dec_pop:
> + mov r14,QWORD[((-40))+rax]
> +
> + mov r13,QWORD[((-32))+rax]
> +
> + mov r12,QWORD[((-24))+rax]
> +
> + mov rbp,QWORD[((-16))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> +
> + lea rsp,[rax]
> +
> +$L$ocb_dec_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_ocb_decrypt:
> +
> +
> +ALIGN 32
> +__ocb_decrypt6:
> +
> + pxor xmm15,xmm9
> + movdqu xmm11,XMMWORD[r12*1+rbx]
> + movdqa xmm12,xmm10
> + movdqu xmm13,XMMWORD[r13*1+rbx]
> + movdqa xmm14,xmm10
> + pxor xmm10,xmm15
> + movdqu xmm15,XMMWORD[r14*1+rbx]
> + pxor xmm11,xmm10
> + pxor xmm2,xmm10
> + pxor xmm12,xmm11
> + pxor xmm3,xmm11
> + pxor xmm13,xmm12
> + pxor xmm4,xmm12
> + pxor xmm14,xmm13
> + pxor xmm5,xmm13
> + pxor xmm15,xmm14
> + pxor xmm6,xmm14
> + pxor xmm7,xmm15
> + movups xmm0,XMMWORD[32+r11]
> +
> + lea r12,[1+r8]
> + lea r13,[3+r8]
> + lea r14,[5+r8]
> + add r8,6
> + pxor xmm10,xmm9
> + bsf r12,r12
> + bsf r13,r13
> + bsf r14,r14
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + pxor xmm11,xmm9
> + pxor xmm12,xmm9
> +DB 102,15,56,222,241
> + pxor xmm13,xmm9
> + pxor xmm14,xmm9
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[48+r11]
> + pxor xmm15,xmm9
> +
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> + movups xmm0,XMMWORD[64+r11]
> + shl r12,4
> + shl r13,4
> + jmp NEAR $L$ocb_dec_loop6
> +
> +ALIGN 32
> +$L$ocb_dec_loop6:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$ocb_dec_loop6
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> + movups xmm1,XMMWORD[16+r11]
> + shl r14,4
> +
> +DB 102,65,15,56,223,210
> + movdqu xmm10,XMMWORD[rbx]
> + mov rax,r10
> +DB 102,65,15,56,223,219
> +DB 102,65,15,56,223,228
> +DB 102,65,15,56,223,237
> +DB 102,65,15,56,223,246
> +DB 102,65,15,56,223,255
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +ALIGN 32
> +__ocb_decrypt4:
> +
> + pxor xmm15,xmm9
> + movdqu xmm11,XMMWORD[r12*1+rbx]
> + movdqa xmm12,xmm10
> + movdqu xmm13,XMMWORD[r13*1+rbx]
> + pxor xmm10,xmm15
> + pxor xmm11,xmm10
> + pxor xmm2,xmm10
> + pxor xmm12,xmm11
> + pxor xmm3,xmm11
> + pxor xmm13,xmm12
> + pxor xmm4,xmm12
> + pxor xmm5,xmm13
> + movups xmm0,XMMWORD[32+r11]
> +
> + pxor xmm10,xmm9
> + pxor xmm11,xmm9
> + pxor xmm12,xmm9
> + pxor xmm13,xmm9
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[48+r11]
> +
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[64+r11]
> + jmp NEAR $L$ocb_dec_loop4
> +
> +ALIGN 32
> +$L$ocb_dec_loop4:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$ocb_dec_loop4
> +
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + movups xmm1,XMMWORD[16+r11]
> + mov rax,r10
> +
> +DB 102,65,15,56,223,210
> +DB 102,65,15,56,223,219
> +DB 102,65,15,56,223,228
> +DB 102,65,15,56,223,237
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +ALIGN 32
> +__ocb_decrypt1:
> +
> + pxor xmm7,xmm15
> + pxor xmm7,xmm9
> + pxor xmm2,xmm7
> + movups xmm0,XMMWORD[32+r11]
> +
> +DB 102,15,56,222,209
> + movups xmm1,XMMWORD[48+r11]
> + pxor xmm7,xmm9
> +
> +DB 102,15,56,222,208
> + movups xmm0,XMMWORD[64+r11]
> + jmp NEAR $L$ocb_dec_loop1
> +
> +ALIGN 32
> +$L$ocb_dec_loop1:
> +DB 102,15,56,222,209
> + movups xmm1,XMMWORD[rax*1+rcx]
> + add rax,32
> +
> +DB 102,15,56,222,208
> + movups xmm0,XMMWORD[((-16))+rax*1+rcx]
> + jnz NEAR $L$ocb_dec_loop1
> +
> +DB 102,15,56,222,209
> + movups xmm1,XMMWORD[16+r11]
> + mov rax,r10
> +
> +DB 102,15,56,223,215
> + DB 0F3h,0C3h ;repret
> +
> +
> +global aesni_cbc_encrypt
> +
> +ALIGN 16
> +aesni_cbc_encrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_aesni_cbc_encrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + test rdx,rdx
> + jz NEAR $L$cbc_ret
> +
> + mov r10d,DWORD[240+rcx]
> + mov r11,rcx
> + test r9d,r9d
> + jz NEAR $L$cbc_decrypt
> +
> + movups xmm2,XMMWORD[r8]
> + mov eax,r10d
> + cmp rdx,16
> + jb NEAR $L$cbc_enc_tail
> + sub rdx,16
> + jmp NEAR $L$cbc_enc_loop
> +ALIGN 16
> +$L$cbc_enc_loop:
> + movups xmm3,XMMWORD[rdi]
> + lea rdi,[16+rdi]
> +
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + xorps xmm3,xmm0
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm3
> +$L$oop_enc1_15:
> +DB 102,15,56,220,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_enc1_15
> +DB 102,15,56,221,209
> + mov eax,r10d
> + mov rcx,r11
> + movups XMMWORD[rsi],xmm2
> + lea rsi,[16+rsi]
> + sub rdx,16
> + jnc NEAR $L$cbc_enc_loop
> + add rdx,16
> + jnz NEAR $L$cbc_enc_tail
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + movups XMMWORD[r8],xmm2
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + jmp NEAR $L$cbc_ret
> +
> +$L$cbc_enc_tail:
> + mov rcx,rdx
> + xchg rsi,rdi
> + DD 0x9066A4F3
> + mov ecx,16
> + sub rcx,rdx
> + xor eax,eax
> + DD 0x9066AAF3
> + lea rdi,[((-16))+rdi]
> + mov eax,r10d
> + mov rsi,rdi
> + mov rcx,r11
> + xor rdx,rdx
> + jmp NEAR $L$cbc_enc_loop
> +
> +ALIGN 16
> +$L$cbc_decrypt:
> + cmp rdx,16
> + jne NEAR $L$cbc_decrypt_bulk
> +
> +
> +
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[r8]
> + movdqa xmm4,xmm2
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_dec1_16:
> +DB 102,15,56,222,209
> + dec r10d
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_dec1_16
> +DB 102,15,56,223,209
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + movdqu XMMWORD[r8],xmm4
> + xorps xmm2,xmm3
> + pxor xmm3,xmm3
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + jmp NEAR $L$cbc_ret
> +ALIGN 16
> +$L$cbc_decrypt_bulk:
> + lea r11,[rsp]
> +
> + push rbp
> +
> + sub rsp,176
> + and rsp,-16
> + movaps XMMWORD[16+rsp],xmm6
> + movaps XMMWORD[32+rsp],xmm7
> + movaps XMMWORD[48+rsp],xmm8
> + movaps XMMWORD[64+rsp],xmm9
> + movaps XMMWORD[80+rsp],xmm10
> + movaps XMMWORD[96+rsp],xmm11
> + movaps XMMWORD[112+rsp],xmm12
> + movaps XMMWORD[128+rsp],xmm13
> + movaps XMMWORD[144+rsp],xmm14
> + movaps XMMWORD[160+rsp],xmm15
> +$L$cbc_decrypt_body:
> + mov rbp,rcx
> + movups xmm10,XMMWORD[r8]
> + mov eax,r10d
> + cmp rdx,0x50
> + jbe NEAR $L$cbc_dec_tail
> +
> + movups xmm0,XMMWORD[rcx]
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[16+rdi]
> + movdqa xmm11,xmm2
> + movdqu xmm4,XMMWORD[32+rdi]
> + movdqa xmm12,xmm3
> + movdqu xmm5,XMMWORD[48+rdi]
> + movdqa xmm13,xmm4
> + movdqu xmm6,XMMWORD[64+rdi]
> + movdqa xmm14,xmm5
> + movdqu xmm7,XMMWORD[80+rdi]
> + movdqa xmm15,xmm6
> + mov r9d,DWORD[((OPENSSL_ia32cap_P+4))]
> + cmp rdx,0x70
> + jbe NEAR $L$cbc_dec_six_or_seven
> +
> + and r9d,71303168
> + sub rdx,0x50
> + cmp r9d,4194304
> + je NEAR $L$cbc_dec_loop6_enter
> + sub rdx,0x20
> + lea rcx,[112+rcx]
> + jmp NEAR $L$cbc_dec_loop8_enter
> +ALIGN 16
> +$L$cbc_dec_loop8:
> + movups XMMWORD[rsi],xmm9
> + lea rsi,[16+rsi]
> +$L$cbc_dec_loop8_enter:
> + movdqu xmm8,XMMWORD[96+rdi]
> + pxor xmm2,xmm0
> + movdqu xmm9,XMMWORD[112+rdi]
> + pxor xmm3,xmm0
> + movups xmm1,XMMWORD[((16-112))+rcx]
> + pxor xmm4,xmm0
> + mov rbp,-1
> + cmp rdx,0x70
> + pxor xmm5,xmm0
> + pxor xmm6,xmm0
> + pxor xmm7,xmm0
> + pxor xmm8,xmm0
> +
> +DB 102,15,56,222,209
> + pxor xmm9,xmm0
> + movups xmm0,XMMWORD[((32-112))+rcx]
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,68,15,56,222,193
> + adc rbp,0
> + and rbp,128
> +DB 102,68,15,56,222,201
> + add rbp,rdi
> + movups xmm1,XMMWORD[((48-112))+rcx]
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> +DB 102,68,15,56,222,192
> +DB 102,68,15,56,222,200
> + movups xmm0,XMMWORD[((64-112))+rcx]
> + nop
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,68,15,56,222,193
> +DB 102,68,15,56,222,201
> + movups xmm1,XMMWORD[((80-112))+rcx]
> + nop
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> +DB 102,68,15,56,222,192
> +DB 102,68,15,56,222,200
> + movups xmm0,XMMWORD[((96-112))+rcx]
> + nop
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,68,15,56,222,193
> +DB 102,68,15,56,222,201
> + movups xmm1,XMMWORD[((112-112))+rcx]
> + nop
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> +DB 102,68,15,56,222,192
> +DB 102,68,15,56,222,200
> + movups xmm0,XMMWORD[((128-112))+rcx]
> + nop
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,68,15,56,222,193
> +DB 102,68,15,56,222,201
> + movups xmm1,XMMWORD[((144-112))+rcx]
> + cmp eax,11
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> +DB 102,68,15,56,222,192
> +DB 102,68,15,56,222,200
> + movups xmm0,XMMWORD[((160-112))+rcx]
> + jb NEAR $L$cbc_dec_done
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,68,15,56,222,193
> +DB 102,68,15,56,222,201
> + movups xmm1,XMMWORD[((176-112))+rcx]
> + nop
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> +DB 102,68,15,56,222,192
> +DB 102,68,15,56,222,200
> + movups xmm0,XMMWORD[((192-112))+rcx]
> + je NEAR $L$cbc_dec_done
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> +DB 102,68,15,56,222,193
> +DB 102,68,15,56,222,201
> + movups xmm1,XMMWORD[((208-112))+rcx]
> + nop
> +DB 102,15,56,222,208
> +DB 102,15,56,222,216
> +DB 102,15,56,222,224
> +DB 102,15,56,222,232
> +DB 102,15,56,222,240
> +DB 102,15,56,222,248
> +DB 102,68,15,56,222,192
> +DB 102,68,15,56,222,200
> + movups xmm0,XMMWORD[((224-112))+rcx]
> + jmp NEAR $L$cbc_dec_done
> +ALIGN 16
> +$L$cbc_dec_done:
> +DB 102,15,56,222,209
> +DB 102,15,56,222,217
> + pxor xmm10,xmm0
> + pxor xmm11,xmm0
> +DB 102,15,56,222,225
> +DB 102,15,56,222,233
> + pxor xmm12,xmm0
> + pxor xmm13,xmm0
> +DB 102,15,56,222,241
> +DB 102,15,56,222,249
> + pxor xmm14,xmm0
> + pxor xmm15,xmm0
> +DB 102,68,15,56,222,193
> +DB 102,68,15,56,222,201
> + movdqu xmm1,XMMWORD[80+rdi]
> +
> +DB 102,65,15,56,223,210
> + movdqu xmm10,XMMWORD[96+rdi]
> + pxor xmm1,xmm0
> +DB 102,65,15,56,223,219
> + pxor xmm10,xmm0
> + movdqu xmm0,XMMWORD[112+rdi]
> +DB 102,65,15,56,223,228
> + lea rdi,[128+rdi]
> + movdqu xmm11,XMMWORD[rbp]
> +DB 102,65,15,56,223,237
> +DB 102,65,15,56,223,246
> + movdqu xmm12,XMMWORD[16+rbp]
> + movdqu xmm13,XMMWORD[32+rbp]
> +DB 102,65,15,56,223,255
> +DB 102,68,15,56,223,193
> + movdqu xmm14,XMMWORD[48+rbp]
> + movdqu xmm15,XMMWORD[64+rbp]
> +DB 102,69,15,56,223,202
> + movdqa xmm10,xmm0
> + movdqu xmm1,XMMWORD[80+rbp]
> + movups xmm0,XMMWORD[((-112))+rcx]
> +
> + movups XMMWORD[rsi],xmm2
> + movdqa xmm2,xmm11
> + movups XMMWORD[16+rsi],xmm3
> + movdqa xmm3,xmm12
> + movups XMMWORD[32+rsi],xmm4
> + movdqa xmm4,xmm13
> + movups XMMWORD[48+rsi],xmm5
> + movdqa xmm5,xmm14
> + movups XMMWORD[64+rsi],xmm6
> + movdqa xmm6,xmm15
> + movups XMMWORD[80+rsi],xmm7
> + movdqa xmm7,xmm1
> + movups XMMWORD[96+rsi],xmm8
> + lea rsi,[112+rsi]
> +
> + sub rdx,0x80
> + ja NEAR $L$cbc_dec_loop8
> +
> + movaps xmm2,xmm9
> + lea rcx,[((-112))+rcx]
> + add rdx,0x70
> + jle NEAR $L$cbc_dec_clear_tail_collected
> + movups XMMWORD[rsi],xmm9
> + lea rsi,[16+rsi]
> + cmp rdx,0x50
> + jbe NEAR $L$cbc_dec_tail
> +
> + movaps xmm2,xmm11
> +$L$cbc_dec_six_or_seven:
> + cmp rdx,0x60
> + ja NEAR $L$cbc_dec_seven
> +
> + movaps xmm8,xmm7
> + call _aesni_decrypt6
> + pxor xmm2,xmm10
> + movaps xmm10,xmm8
> + pxor xmm3,xmm11
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm4,xmm12
> + movdqu XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + pxor xmm5,xmm13
> + movdqu XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + pxor xmm6,xmm14
> + movdqu XMMWORD[48+rsi],xmm5
> + pxor xmm5,xmm5
> + pxor xmm7,xmm15
> + movdqu XMMWORD[64+rsi],xmm6
> + pxor xmm6,xmm6
> + lea rsi,[80+rsi]
> + movdqa xmm2,xmm7
> + pxor xmm7,xmm7
> + jmp NEAR $L$cbc_dec_tail_collected
> +
> +ALIGN 16
> +$L$cbc_dec_seven:
> + movups xmm8,XMMWORD[96+rdi]
> + xorps xmm9,xmm9
> + call _aesni_decrypt8
> + movups xmm9,XMMWORD[80+rdi]
> + pxor xmm2,xmm10
> + movups xmm10,XMMWORD[96+rdi]
> + pxor xmm3,xmm11
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm4,xmm12
> + movdqu XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + pxor xmm5,xmm13
> + movdqu XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + pxor xmm6,xmm14
> + movdqu XMMWORD[48+rsi],xmm5
> + pxor xmm5,xmm5
> + pxor xmm7,xmm15
> + movdqu XMMWORD[64+rsi],xmm6
> + pxor xmm6,xmm6
> + pxor xmm8,xmm9
> + movdqu XMMWORD[80+rsi],xmm7
> + pxor xmm7,xmm7
> + lea rsi,[96+rsi]
> + movdqa xmm2,xmm8
> + pxor xmm8,xmm8
> + pxor xmm9,xmm9
> + jmp NEAR $L$cbc_dec_tail_collected
> +
> +ALIGN 16
> +$L$cbc_dec_loop6:
> + movups XMMWORD[rsi],xmm7
> + lea rsi,[16+rsi]
> + movdqu xmm2,XMMWORD[rdi]
> + movdqu xmm3,XMMWORD[16+rdi]
> + movdqa xmm11,xmm2
> + movdqu xmm4,XMMWORD[32+rdi]
> + movdqa xmm12,xmm3
> + movdqu xmm5,XMMWORD[48+rdi]
> + movdqa xmm13,xmm4
> + movdqu xmm6,XMMWORD[64+rdi]
> + movdqa xmm14,xmm5
> + movdqu xmm7,XMMWORD[80+rdi]
> + movdqa xmm15,xmm6
> +$L$cbc_dec_loop6_enter:
> + lea rdi,[96+rdi]
> + movdqa xmm8,xmm7
> +
> + call _aesni_decrypt6
> +
> + pxor xmm2,xmm10
> + movdqa xmm10,xmm8
> + pxor xmm3,xmm11
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm4,xmm12
> + movdqu XMMWORD[16+rsi],xmm3
> + pxor xmm5,xmm13
> + movdqu XMMWORD[32+rsi],xmm4
> + pxor xmm6,xmm14
> + mov rcx,rbp
> + movdqu XMMWORD[48+rsi],xmm5
> + pxor xmm7,xmm15
> + mov eax,r10d
> + movdqu XMMWORD[64+rsi],xmm6
> + lea rsi,[80+rsi]
> + sub rdx,0x60
> + ja NEAR $L$cbc_dec_loop6
> +
> + movdqa xmm2,xmm7
> + add rdx,0x50
> + jle NEAR $L$cbc_dec_clear_tail_collected
> + movups XMMWORD[rsi],xmm7
> + lea rsi,[16+rsi]
> +
> +$L$cbc_dec_tail:
> + movups xmm2,XMMWORD[rdi]
> + sub rdx,0x10
> + jbe NEAR $L$cbc_dec_one
> +
> + movups xmm3,XMMWORD[16+rdi]
> + movaps xmm11,xmm2
> + sub rdx,0x10
> + jbe NEAR $L$cbc_dec_two
> +
> + movups xmm4,XMMWORD[32+rdi]
> + movaps xmm12,xmm3
> + sub rdx,0x10
> + jbe NEAR $L$cbc_dec_three
> +
> + movups xmm5,XMMWORD[48+rdi]
> + movaps xmm13,xmm4
> + sub rdx,0x10
> + jbe NEAR $L$cbc_dec_four
> +
> + movups xmm6,XMMWORD[64+rdi]
> + movaps xmm14,xmm5
> + movaps xmm15,xmm6
> + xorps xmm7,xmm7
> + call _aesni_decrypt6
> + pxor xmm2,xmm10
> + movaps xmm10,xmm15
> + pxor xmm3,xmm11
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm4,xmm12
> + movdqu XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + pxor xmm5,xmm13
> + movdqu XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + pxor xmm6,xmm14
> + movdqu XMMWORD[48+rsi],xmm5
> + pxor xmm5,xmm5
> + lea rsi,[64+rsi]
> + movdqa xmm2,xmm6
> + pxor xmm6,xmm6
> + pxor xmm7,xmm7
> + sub rdx,0x10
> + jmp NEAR $L$cbc_dec_tail_collected
> +
> +ALIGN 16
> +$L$cbc_dec_one:
> + movaps xmm11,xmm2
> + movups xmm0,XMMWORD[rcx]
> + movups xmm1,XMMWORD[16+rcx]
> + lea rcx,[32+rcx]
> + xorps xmm2,xmm0
> +$L$oop_dec1_17:
> +DB 102,15,56,222,209
> + dec eax
> + movups xmm1,XMMWORD[rcx]
> + lea rcx,[16+rcx]
> + jnz NEAR $L$oop_dec1_17
> +DB 102,15,56,223,209
> + xorps xmm2,xmm10
> + movaps xmm10,xmm11
> + jmp NEAR $L$cbc_dec_tail_collected
> +ALIGN 16
> +$L$cbc_dec_two:
> + movaps xmm12,xmm3
> + call _aesni_decrypt2
> + pxor xmm2,xmm10
> + movaps xmm10,xmm12
> + pxor xmm3,xmm11
> + movdqu XMMWORD[rsi],xmm2
> + movdqa xmm2,xmm3
> + pxor xmm3,xmm3
> + lea rsi,[16+rsi]
> + jmp NEAR $L$cbc_dec_tail_collected
> +ALIGN 16
> +$L$cbc_dec_three:
> + movaps xmm13,xmm4
> + call _aesni_decrypt3
> + pxor xmm2,xmm10
> + movaps xmm10,xmm13
> + pxor xmm3,xmm11
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm4,xmm12
> + movdqu XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + movdqa xmm2,xmm4
> + pxor xmm4,xmm4
> + lea rsi,[32+rsi]
> + jmp NEAR $L$cbc_dec_tail_collected
> +ALIGN 16
> +$L$cbc_dec_four:
> + movaps xmm14,xmm5
> + call _aesni_decrypt4
> + pxor xmm2,xmm10
> + movaps xmm10,xmm14
> + pxor xmm3,xmm11
> + movdqu XMMWORD[rsi],xmm2
> + pxor xmm4,xmm12
> + movdqu XMMWORD[16+rsi],xmm3
> + pxor xmm3,xmm3
> + pxor xmm5,xmm13
> + movdqu XMMWORD[32+rsi],xmm4
> + pxor xmm4,xmm4
> + movdqa xmm2,xmm5
> + pxor xmm5,xmm5
> + lea rsi,[48+rsi]
> + jmp NEAR $L$cbc_dec_tail_collected
> +
> +ALIGN 16
> +$L$cbc_dec_clear_tail_collected:
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> +$L$cbc_dec_tail_collected:
> + movups XMMWORD[r8],xmm10
> + and rdx,15
> + jnz NEAR $L$cbc_dec_tail_partial
> + movups XMMWORD[rsi],xmm2
> + pxor xmm2,xmm2
> + jmp NEAR $L$cbc_dec_ret
> +ALIGN 16
> +$L$cbc_dec_tail_partial:
> + movaps XMMWORD[rsp],xmm2
> + pxor xmm2,xmm2
> + mov rcx,16
> + mov rdi,rsi
> + sub rcx,rdx
> + lea rsi,[rsp]
> + DD 0x9066A4F3
> + movdqa XMMWORD[rsp],xmm2
> +
> +$L$cbc_dec_ret:
> + xorps xmm0,xmm0
> + pxor xmm1,xmm1
> + movaps xmm6,XMMWORD[16+rsp]
> + movaps XMMWORD[16+rsp],xmm0
> + movaps xmm7,XMMWORD[32+rsp]
> + movaps XMMWORD[32+rsp],xmm0
> + movaps xmm8,XMMWORD[48+rsp]
> + movaps XMMWORD[48+rsp],xmm0
> + movaps xmm9,XMMWORD[64+rsp]
> + movaps XMMWORD[64+rsp],xmm0
> + movaps xmm10,XMMWORD[80+rsp]
> + movaps XMMWORD[80+rsp],xmm0
> + movaps xmm11,XMMWORD[96+rsp]
> + movaps XMMWORD[96+rsp],xmm0
> + movaps xmm12,XMMWORD[112+rsp]
> + movaps XMMWORD[112+rsp],xmm0
> + movaps xmm13,XMMWORD[128+rsp]
> + movaps XMMWORD[128+rsp],xmm0
> + movaps xmm14,XMMWORD[144+rsp]
> + movaps XMMWORD[144+rsp],xmm0
> + movaps xmm15,XMMWORD[160+rsp]
> + movaps XMMWORD[160+rsp],xmm0
> + mov rbp,QWORD[((-8))+r11]
> +
> + lea rsp,[r11]
> +
> +$L$cbc_ret:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_aesni_cbc_encrypt:
> +global aesni_set_decrypt_key
> +
> +ALIGN 16
> +aesni_set_decrypt_key:
> +
> +DB 0x48,0x83,0xEC,0x08
> +
> + call __aesni_set_encrypt_key
> + shl edx,4
> + test eax,eax
> + jnz NEAR $L$dec_key_ret
> + lea rcx,[16+rdx*1+r8]
> +
> + movups xmm0,XMMWORD[r8]
> + movups xmm1,XMMWORD[rcx]
> + movups XMMWORD[rcx],xmm0
> + movups XMMWORD[r8],xmm1
> + lea r8,[16+r8]
> + lea rcx,[((-16))+rcx]
> +
> +$L$dec_key_inverse:
> + movups xmm0,XMMWORD[r8]
> + movups xmm1,XMMWORD[rcx]
> +DB 102,15,56,219,192
> +DB 102,15,56,219,201
> + lea r8,[16+r8]
> + lea rcx,[((-16))+rcx]
> + movups XMMWORD[16+rcx],xmm0
> + movups XMMWORD[(-16)+r8],xmm1
> + cmp rcx,r8
> + ja NEAR $L$dec_key_inverse
> +
> + movups xmm0,XMMWORD[r8]
> +DB 102,15,56,219,192
> + pxor xmm1,xmm1
> + movups XMMWORD[rcx],xmm0
> + pxor xmm0,xmm0
> +$L$dec_key_ret:
> + add rsp,8
> +
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_set_decrypt_key:
> +
> +global aesni_set_encrypt_key
> +
> +ALIGN 16
> +aesni_set_encrypt_key:
> +__aesni_set_encrypt_key:
> +
> +DB 0x48,0x83,0xEC,0x08
> +
> + mov rax,-1
> + test rcx,rcx
> + jz NEAR $L$enc_key_ret
> + test r8,r8
> + jz NEAR $L$enc_key_ret
> +
> + mov r10d,268437504
> + movups xmm0,XMMWORD[rcx]
> + xorps xmm4,xmm4
> + and r10d,DWORD[((OPENSSL_ia32cap_P+4))]
> + lea rax,[16+r8]
> + cmp edx,256
> + je NEAR $L$14rounds
> + cmp edx,192
> + je NEAR $L$12rounds
> + cmp edx,128
> + jne NEAR $L$bad_keybits
> +
> +$L$10rounds:
> + mov edx,9
> + cmp r10d,268435456
> + je NEAR $L$10rounds_alt
> +
> + movups XMMWORD[r8],xmm0
> +DB 102,15,58,223,200,1
> + call $L$key_expansion_128_cold
> +DB 102,15,58,223,200,2
> + call $L$key_expansion_128
> +DB 102,15,58,223,200,4
> + call $L$key_expansion_128
> +DB 102,15,58,223,200,8
> + call $L$key_expansion_128
> +DB 102,15,58,223,200,16
> + call $L$key_expansion_128
> +DB 102,15,58,223,200,32
> + call $L$key_expansion_128
> +DB 102,15,58,223,200,64
> + call $L$key_expansion_128
> +DB 102,15,58,223,200,128
> + call $L$key_expansion_128
> +DB 102,15,58,223,200,27
> + call $L$key_expansion_128
> +DB 102,15,58,223,200,54
> + call $L$key_expansion_128
> + movups XMMWORD[rax],xmm0
> + mov DWORD[80+rax],edx
> + xor eax,eax
> + jmp NEAR $L$enc_key_ret
> +
> +ALIGN 16
> +$L$10rounds_alt:
> + movdqa xmm5,XMMWORD[$L$key_rotate]
> + mov r10d,8
> + movdqa xmm4,XMMWORD[$L$key_rcon1]
> + movdqa xmm2,xmm0
> + movdqu XMMWORD[r8],xmm0
> + jmp NEAR $L$oop_key128
> +
> +ALIGN 16
> +$L$oop_key128:
> +DB 102,15,56,0,197
> +DB 102,15,56,221,196
> + pslld xmm4,1
> + lea rax,[16+rax]
> +
> + movdqa xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm2,xmm3
> +
> + pxor xmm0,xmm2
> + movdqu XMMWORD[(-16)+rax],xmm0
> + movdqa xmm2,xmm0
> +
> + dec r10d
> + jnz NEAR $L$oop_key128
> +
> + movdqa xmm4,XMMWORD[$L$key_rcon1b]
> +
> +DB 102,15,56,0,197
> +DB 102,15,56,221,196
> + pslld xmm4,1
> +
> + movdqa xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm2,xmm3
> +
> + pxor xmm0,xmm2
> + movdqu XMMWORD[rax],xmm0
> +
> + movdqa xmm2,xmm0
> +DB 102,15,56,0,197
> +DB 102,15,56,221,196
> +
> + movdqa xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm3,xmm2
> + pslldq xmm2,4
> + pxor xmm2,xmm3
> +
> + pxor xmm0,xmm2
> + movdqu XMMWORD[16+rax],xmm0
> +
> + mov DWORD[96+rax],edx
> + xor eax,eax
> + jmp NEAR $L$enc_key_ret
> +
> +ALIGN 16
> +$L$12rounds:
> + movq xmm2,QWORD[16+rcx]
> + mov edx,11
> + cmp r10d,268435456
> + je NEAR $L$12rounds_alt
> +
> + movups XMMWORD[r8],xmm0
> +DB 102,15,58,223,202,1
> + call $L$key_expansion_192a_cold
> +DB 102,15,58,223,202,2
> + call $L$key_expansion_192b
> +DB 102,15,58,223,202,4
> + call $L$key_expansion_192a
> +DB 102,15,58,223,202,8
> + call $L$key_expansion_192b
> +DB 102,15,58,223,202,16
> + call $L$key_expansion_192a
> +DB 102,15,58,223,202,32
> + call $L$key_expansion_192b
> +DB 102,15,58,223,202,64
> + call $L$key_expansion_192a
> +DB 102,15,58,223,202,128
> + call $L$key_expansion_192b
> + movups XMMWORD[rax],xmm0
> + mov DWORD[48+rax],edx
> + xor rax,rax
> + jmp NEAR $L$enc_key_ret
> +
> +ALIGN 16
> +$L$12rounds_alt:
> + movdqa xmm5,XMMWORD[$L$key_rotate192]
> + movdqa xmm4,XMMWORD[$L$key_rcon1]
> + mov r10d,8
> + movdqu XMMWORD[r8],xmm0
> + jmp NEAR $L$oop_key192
> +
> +ALIGN 16
> +$L$oop_key192:
> + movq QWORD[rax],xmm2
> + movdqa xmm1,xmm2
> +DB 102,15,56,0,213
> +DB 102,15,56,221,212
> + pslld xmm4,1
> + lea rax,[24+rax]
> +
> + movdqa xmm3,xmm0
> + pslldq xmm0,4
> + pxor xmm3,xmm0
> + pslldq xmm0,4
> + pxor xmm3,xmm0
> + pslldq xmm0,4
> + pxor xmm0,xmm3
> +
> + pshufd xmm3,xmm0,0xff
> + pxor xmm3,xmm1
> + pslldq xmm1,4
> + pxor xmm3,xmm1
> +
> + pxor xmm0,xmm2
> + pxor xmm2,xmm3
> + movdqu XMMWORD[(-16)+rax],xmm0
> +
> + dec r10d
> + jnz NEAR $L$oop_key192
> +
> + mov DWORD[32+rax],edx
> + xor eax,eax
> + jmp NEAR $L$enc_key_ret
> +
> +ALIGN 16
> +$L$14rounds:
> + movups xmm2,XMMWORD[16+rcx]
> + mov edx,13
> + lea rax,[16+rax]
> + cmp r10d,268435456
> + je NEAR $L$14rounds_alt
> +
> + movups XMMWORD[r8],xmm0
> + movups XMMWORD[16+r8],xmm2
> +DB 102,15,58,223,202,1
> + call $L$key_expansion_256a_cold
> +DB 102,15,58,223,200,1
> + call $L$key_expansion_256b
> +DB 102,15,58,223,202,2
> + call $L$key_expansion_256a
> +DB 102,15,58,223,200,2
> + call $L$key_expansion_256b
> +DB 102,15,58,223,202,4
> + call $L$key_expansion_256a
> +DB 102,15,58,223,200,4
> + call $L$key_expansion_256b
> +DB 102,15,58,223,202,8
> + call $L$key_expansion_256a
> +DB 102,15,58,223,200,8
> + call $L$key_expansion_256b
> +DB 102,15,58,223,202,16
> + call $L$key_expansion_256a
> +DB 102,15,58,223,200,16
> + call $L$key_expansion_256b
> +DB 102,15,58,223,202,32
> + call $L$key_expansion_256a
> +DB 102,15,58,223,200,32
> + call $L$key_expansion_256b
> +DB 102,15,58,223,202,64
> + call $L$key_expansion_256a
> + movups XMMWORD[rax],xmm0
> + mov DWORD[16+rax],edx
> + xor rax,rax
> + jmp NEAR $L$enc_key_ret
> +
> +ALIGN 16
> +$L$14rounds_alt:
> + movdqa xmm5,XMMWORD[$L$key_rotate]
> + movdqa xmm4,XMMWORD[$L$key_rcon1]
> + mov r10d,7
> + movdqu XMMWORD[r8],xmm0
> + movdqa xmm1,xmm2
> + movdqu XMMWORD[16+r8],xmm2
> + jmp NEAR $L$oop_key256
> +
> +ALIGN 16
> +$L$oop_key256:
> +DB 102,15,56,0,213
> +DB 102,15,56,221,212
> +
> + movdqa xmm3,xmm0
> + pslldq xmm0,4
> + pxor xmm3,xmm0
> + pslldq xmm0,4
> + pxor xmm3,xmm0
> + pslldq xmm0,4
> + pxor xmm0,xmm3
> + pslld xmm4,1
> +
> + pxor xmm0,xmm2
> + movdqu XMMWORD[rax],xmm0
> +
> + dec r10d
> + jz NEAR $L$done_key256
> +
> + pshufd xmm2,xmm0,0xff
> + pxor xmm3,xmm3
> +DB 102,15,56,221,211
> +
> + movdqa xmm3,xmm1
> + pslldq xmm1,4
> + pxor xmm3,xmm1
> + pslldq xmm1,4
> + pxor xmm3,xmm1
> + pslldq xmm1,4
> + pxor xmm1,xmm3
> +
> + pxor xmm2,xmm1
> + movdqu XMMWORD[16+rax],xmm2
> + lea rax,[32+rax]
> + movdqa xmm1,xmm2
> +
> + jmp NEAR $L$oop_key256
> +
> +$L$done_key256:
> + mov DWORD[16+rax],edx
> + xor eax,eax
> + jmp NEAR $L$enc_key_ret
> +
> +ALIGN 16
> +$L$bad_keybits:
> + mov rax,-2
> +$L$enc_key_ret:
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> + add rsp,8
> +
> + DB 0F3h,0C3h ;repret
> +$L$SEH_end_set_encrypt_key:
> +
> +ALIGN 16
> +$L$key_expansion_128:
> + movups XMMWORD[rax],xmm0
> + lea rax,[16+rax]
> +$L$key_expansion_128_cold:
> + shufps xmm4,xmm0,16
> + xorps xmm0,xmm4
> + shufps xmm4,xmm0,140
> + xorps xmm0,xmm4
> + shufps xmm1,xmm1,255
> + xorps xmm0,xmm1
> + DB 0F3h,0C3h ;repret
> +
> +ALIGN 16
> +$L$key_expansion_192a:
> + movups XMMWORD[rax],xmm0
> + lea rax,[16+rax]
> +$L$key_expansion_192a_cold:
> + movaps xmm5,xmm2
> +$L$key_expansion_192b_warm:
> + shufps xmm4,xmm0,16
> + movdqa xmm3,xmm2
> + xorps xmm0,xmm4
> + shufps xmm4,xmm0,140
> + pslldq xmm3,4
> + xorps xmm0,xmm4
> + pshufd xmm1,xmm1,85
> + pxor xmm2,xmm3
> + pxor xmm0,xmm1
> + pshufd xmm3,xmm0,255
> + pxor xmm2,xmm3
> + DB 0F3h,0C3h ;repret
> +
> +ALIGN 16
> +$L$key_expansion_192b:
> + movaps xmm3,xmm0
> + shufps xmm5,xmm0,68
> + movups XMMWORD[rax],xmm5
> + shufps xmm3,xmm2,78
> + movups XMMWORD[16+rax],xmm3
> + lea rax,[32+rax]
> + jmp NEAR $L$key_expansion_192b_warm
> +
> +ALIGN 16
> +$L$key_expansion_256a:
> + movups XMMWORD[rax],xmm2
> + lea rax,[16+rax]
> +$L$key_expansion_256a_cold:
> + shufps xmm4,xmm0,16
> + xorps xmm0,xmm4
> + shufps xmm4,xmm0,140
> + xorps xmm0,xmm4
> + shufps xmm1,xmm1,255
> + xorps xmm0,xmm1
> + DB 0F3h,0C3h ;repret
> +
> +ALIGN 16
> +$L$key_expansion_256b:
> + movups XMMWORD[rax],xmm0
> + lea rax,[16+rax]
> +
> + shufps xmm4,xmm2,16
> + xorps xmm2,xmm4
> + shufps xmm4,xmm2,140
> + xorps xmm2,xmm4
> + shufps xmm1,xmm1,170
> + xorps xmm2,xmm1
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +ALIGN 64
> +$L$bswap_mask:
> +DB 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
> +$L$increment32:
> + DD 6,6,6,0
> +$L$increment64:
> + DD 1,0,0,0
> +$L$xts_magic:
> + DD 0x87,0,1,0
> +$L$increment1:
> +DB 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1
> +$L$key_rotate:
> + DD 0x0c0f0e0d,0x0c0f0e0d,0x0c0f0e0d,0x0c0f0e0d
> +$L$key_rotate192:
> + DD 0x04070605,0x04070605,0x04070605,0x04070605
> +$L$key_rcon1:
> + DD 1,1,1,1
> +$L$key_rcon1b:
> + DD 0x1b,0x1b,0x1b,0x1b
> +
> +DB 65,69,83,32,102,111,114,32,73,110,116,101,108,32,65,69
> +DB 83,45,78,73,44,32,67,82,89,80,84,79,71,65,77,83
> +DB 32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115
> +DB 115,108,46,111,114,103,62,0
> +ALIGN 64
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +ecb_ccm64_se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$common_seh_tail
> +
> + lea rsi,[rax]
> + lea rdi,[512+r8]
> + mov ecx,8
> + DD 0xa548f3fc
> + lea rax,[88+rax]
> +
> + jmp NEAR $L$common_seh_tail
> +
> +
> +
> +ALIGN 16
> +ctr_xts_se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[208+r8]
> +
> + lea rsi,[((-168))+rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> +
> + mov rbp,QWORD[((-8))+rax]
> + mov QWORD[160+r8],rbp
> + jmp NEAR $L$common_seh_tail
> +
> +
> +
> +ALIGN 16
> +ocb_se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$common_seh_tail
> +
> + mov r10d,DWORD[8+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$ocb_no_xmm
> +
> + mov rax,QWORD[152+r8]
> +
> + lea rsi,[rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> + lea rax,[((160+40))+rax]
> +
> +$L$ocb_no_xmm:
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov r12,QWORD[((-24))+rax]
> + mov r13,QWORD[((-32))+rax]
> + mov r14,QWORD[((-40))+rax]
> +
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> + mov QWORD[216+r8],r12
> + mov QWORD[224+r8],r13
> + mov QWORD[232+r8],r14
> +
> + jmp NEAR $L$common_seh_tail
> +
> +
> +ALIGN 16
> +cbc_se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[152+r8]
> + mov rbx,QWORD[248+r8]
> +
> + lea r10,[$L$cbc_decrypt_bulk]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[120+r8]
> +
> + lea r10,[$L$cbc_decrypt_body]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[152+r8]
> +
> + lea r10,[$L$cbc_ret]
> + cmp rbx,r10
> + jae NEAR $L$common_seh_tail
> +
> + lea rsi,[16+rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> +
> + mov rax,QWORD[208+r8]
> +
> + mov rbp,QWORD[((-8))+rax]
> + mov QWORD[160+r8],rbp
> +
> +$L$common_seh_tail:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_aesni_ecb_encrypt wrt ..imagebase
> + DD $L$SEH_end_aesni_ecb_encrypt wrt ..imagebase
> + DD $L$SEH_info_ecb wrt ..imagebase
> +
> + DD $L$SEH_begin_aesni_ccm64_encrypt_blocks wrt ..imagebase
> + DD $L$SEH_end_aesni_ccm64_encrypt_blocks wrt ..imagebase
> + DD $L$SEH_info_ccm64_enc wrt ..imagebase
> +
> + DD $L$SEH_begin_aesni_ccm64_decrypt_blocks wrt ..imagebase
> + DD $L$SEH_end_aesni_ccm64_decrypt_blocks wrt ..imagebase
> + DD $L$SEH_info_ccm64_dec wrt ..imagebase
> +
> + DD $L$SEH_begin_aesni_ctr32_encrypt_blocks wrt ..imagebase
> + DD $L$SEH_end_aesni_ctr32_encrypt_blocks wrt ..imagebase
> + DD $L$SEH_info_ctr32 wrt ..imagebase
> +
> + DD $L$SEH_begin_aesni_xts_encrypt wrt ..imagebase
> + DD $L$SEH_end_aesni_xts_encrypt wrt ..imagebase
> + DD $L$SEH_info_xts_enc wrt ..imagebase
> +
> + DD $L$SEH_begin_aesni_xts_decrypt wrt ..imagebase
> + DD $L$SEH_end_aesni_xts_decrypt wrt ..imagebase
> + DD $L$SEH_info_xts_dec wrt ..imagebase
> +
> + DD $L$SEH_begin_aesni_ocb_encrypt wrt ..imagebase
> + DD $L$SEH_end_aesni_ocb_encrypt wrt ..imagebase
> + DD $L$SEH_info_ocb_enc wrt ..imagebase
> +
> + DD $L$SEH_begin_aesni_ocb_decrypt wrt ..imagebase
> + DD $L$SEH_end_aesni_ocb_decrypt wrt ..imagebase
> + DD $L$SEH_info_ocb_dec wrt ..imagebase
> + DD $L$SEH_begin_aesni_cbc_encrypt wrt ..imagebase
> + DD $L$SEH_end_aesni_cbc_encrypt wrt ..imagebase
> + DD $L$SEH_info_cbc wrt ..imagebase
> +
> + DD aesni_set_decrypt_key wrt ..imagebase
> + DD $L$SEH_end_set_decrypt_key wrt ..imagebase
> + DD $L$SEH_info_key wrt ..imagebase
> +
> + DD aesni_set_encrypt_key wrt ..imagebase
> + DD $L$SEH_end_set_encrypt_key wrt ..imagebase
> + DD $L$SEH_info_key wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_ecb:
> +DB 9,0,0,0
> + DD ecb_ccm64_se_handler wrt ..imagebase
> + DD $L$ecb_enc_body wrt ..imagebase,$L$ecb_enc_ret wrt ..imagebase
> +$L$SEH_info_ccm64_enc:
> +DB 9,0,0,0
> + DD ecb_ccm64_se_handler wrt ..imagebase
> + DD $L$ccm64_enc_body wrt ..imagebase,$L$ccm64_enc_ret
> wrt ..imagebase
> +$L$SEH_info_ccm64_dec:
> +DB 9,0,0,0
> + DD ecb_ccm64_se_handler wrt ..imagebase
> + DD $L$ccm64_dec_body wrt ..imagebase,$L$ccm64_dec_ret
> wrt ..imagebase
> +$L$SEH_info_ctr32:
> +DB 9,0,0,0
> + DD ctr_xts_se_handler wrt ..imagebase
> + DD $L$ctr32_body wrt ..imagebase,$L$ctr32_epilogue wrt ..imagebase
> +$L$SEH_info_xts_enc:
> +DB 9,0,0,0
> + DD ctr_xts_se_handler wrt ..imagebase
> + DD $L$xts_enc_body wrt ..imagebase,$L$xts_enc_epilogue
> wrt ..imagebase
> +$L$SEH_info_xts_dec:
> +DB 9,0,0,0
> + DD ctr_xts_se_handler wrt ..imagebase
> + DD $L$xts_dec_body wrt ..imagebase,$L$xts_dec_epilogue
> wrt ..imagebase
> +$L$SEH_info_ocb_enc:
> +DB 9,0,0,0
> + DD ocb_se_handler wrt ..imagebase
> + DD $L$ocb_enc_body wrt ..imagebase,$L$ocb_enc_epilogue
> wrt ..imagebase
> + DD $L$ocb_enc_pop wrt ..imagebase
> + DD 0
> +$L$SEH_info_ocb_dec:
> +DB 9,0,0,0
> + DD ocb_se_handler wrt ..imagebase
> + DD $L$ocb_dec_body wrt ..imagebase,$L$ocb_dec_epilogue
> wrt ..imagebase
> + DD $L$ocb_dec_pop wrt ..imagebase
> + DD 0
> +$L$SEH_info_cbc:
> +DB 9,0,0,0
> + DD cbc_se_handler wrt ..imagebase
> +$L$SEH_info_key:
> +DB 0x01,0x04,0x01,0x00
> +DB 0x04,0x02,0x00,0x00
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/aes/vpaes-x86_64.nasm
> b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/vpaes-x86_64.nasm
> new file mode 100644
> index 0000000000..1c911fa294
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/aes/vpaes-x86_64.nasm
> @@ -0,0 +1,1173 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/aes/asm/vpaes-x86_64.pl
> +;
> +; Copyright 2011-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +_vpaes_encrypt_core:
> +
> + mov r9,rdx
> + mov r11,16
> + mov eax,DWORD[240+rdx]
> + movdqa xmm1,xmm9
> + movdqa xmm2,XMMWORD[$L$k_ipt]
> + pandn xmm1,xmm0
> + movdqu xmm5,XMMWORD[r9]
> + psrld xmm1,4
> + pand xmm0,xmm9
> +DB 102,15,56,0,208
> + movdqa xmm0,XMMWORD[(($L$k_ipt+16))]
> +DB 102,15,56,0,193
> + pxor xmm2,xmm5
> + add r9,16
> + pxor xmm0,xmm2
> + lea r10,[$L$k_mc_backward]
> + jmp NEAR $L$enc_entry
> +
> +ALIGN 16
> +$L$enc_loop:
> +
> + movdqa xmm4,xmm13
> + movdqa xmm0,xmm12
> +DB 102,15,56,0,226
> +DB 102,15,56,0,195
> + pxor xmm4,xmm5
> + movdqa xmm5,xmm15
> + pxor xmm0,xmm4
> + movdqa xmm1,XMMWORD[((-64))+r10*1+r11]
> +DB 102,15,56,0,234
> + movdqa xmm4,XMMWORD[r10*1+r11]
> + movdqa xmm2,xmm14
> +DB 102,15,56,0,211
> + movdqa xmm3,xmm0
> + pxor xmm2,xmm5
> +DB 102,15,56,0,193
> + add r9,16
> + pxor xmm0,xmm2
> +DB 102,15,56,0,220
> + add r11,16
> + pxor xmm3,xmm0
> +DB 102,15,56,0,193
> + and r11,0x30
> + sub rax,1
> + pxor xmm0,xmm3
> +
> +$L$enc_entry:
> +
> + movdqa xmm1,xmm9
> + movdqa xmm5,xmm11
> + pandn xmm1,xmm0
> + psrld xmm1,4
> + pand xmm0,xmm9
> +DB 102,15,56,0,232
> + movdqa xmm3,xmm10
> + pxor xmm0,xmm1
> +DB 102,15,56,0,217
> + movdqa xmm4,xmm10
> + pxor xmm3,xmm5
> +DB 102,15,56,0,224
> + movdqa xmm2,xmm10
> + pxor xmm4,xmm5
> +DB 102,15,56,0,211
> + movdqa xmm3,xmm10
> + pxor xmm2,xmm0
> +DB 102,15,56,0,220
> + movdqu xmm5,XMMWORD[r9]
> + pxor xmm3,xmm1
> + jnz NEAR $L$enc_loop
> +
> +
> + movdqa xmm4,XMMWORD[((-96))+r10]
> + movdqa xmm0,XMMWORD[((-80))+r10]
> +DB 102,15,56,0,226
> + pxor xmm4,xmm5
> +DB 102,15,56,0,195
> + movdqa xmm1,XMMWORD[64+r10*1+r11]
> + pxor xmm0,xmm4
> +DB 102,15,56,0,193
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +_vpaes_decrypt_core:
> +
> + mov r9,rdx
> + mov eax,DWORD[240+rdx]
> + movdqa xmm1,xmm9
> + movdqa xmm2,XMMWORD[$L$k_dipt]
> + pandn xmm1,xmm0
> + mov r11,rax
> + psrld xmm1,4
> + movdqu xmm5,XMMWORD[r9]
> + shl r11,4
> + pand xmm0,xmm9
> +DB 102,15,56,0,208
> + movdqa xmm0,XMMWORD[(($L$k_dipt+16))]
> + xor r11,0x30
> + lea r10,[$L$k_dsbd]
> +DB 102,15,56,0,193
> + and r11,0x30
> + pxor xmm2,xmm5
> + movdqa xmm5,XMMWORD[(($L$k_mc_forward+48))]
> + pxor xmm0,xmm2
> + add r9,16
> + add r11,r10
> + jmp NEAR $L$dec_entry
> +
> +ALIGN 16
> +$L$dec_loop:
> +
> +
> +
> + movdqa xmm4,XMMWORD[((-32))+r10]
> + movdqa xmm1,XMMWORD[((-16))+r10]
> +DB 102,15,56,0,226
> +DB 102,15,56,0,203
> + pxor xmm0,xmm4
> + movdqa xmm4,XMMWORD[r10]
> + pxor xmm0,xmm1
> + movdqa xmm1,XMMWORD[16+r10]
> +
> +DB 102,15,56,0,226
> +DB 102,15,56,0,197
> +DB 102,15,56,0,203
> + pxor xmm0,xmm4
> + movdqa xmm4,XMMWORD[32+r10]
> + pxor xmm0,xmm1
> + movdqa xmm1,XMMWORD[48+r10]
> +
> +DB 102,15,56,0,226
> +DB 102,15,56,0,197
> +DB 102,15,56,0,203
> + pxor xmm0,xmm4
> + movdqa xmm4,XMMWORD[64+r10]
> + pxor xmm0,xmm1
> + movdqa xmm1,XMMWORD[80+r10]
> +
> +DB 102,15,56,0,226
> +DB 102,15,56,0,197
> +DB 102,15,56,0,203
> + pxor xmm0,xmm4
> + add r9,16
> +DB 102,15,58,15,237,12
> + pxor xmm0,xmm1
> + sub rax,1
> +
> +$L$dec_entry:
> +
> + movdqa xmm1,xmm9
> + pandn xmm1,xmm0
> + movdqa xmm2,xmm11
> + psrld xmm1,4
> + pand xmm0,xmm9
> +DB 102,15,56,0,208
> + movdqa xmm3,xmm10
> + pxor xmm0,xmm1
> +DB 102,15,56,0,217
> + movdqa xmm4,xmm10
> + pxor xmm3,xmm2
> +DB 102,15,56,0,224
> + pxor xmm4,xmm2
> + movdqa xmm2,xmm10
> +DB 102,15,56,0,211
> + movdqa xmm3,xmm10
> + pxor xmm2,xmm0
> +DB 102,15,56,0,220
> + movdqu xmm0,XMMWORD[r9]
> + pxor xmm3,xmm1
> + jnz NEAR $L$dec_loop
> +
> +
> + movdqa xmm4,XMMWORD[96+r10]
> +DB 102,15,56,0,226
> + pxor xmm4,xmm0
> + movdqa xmm0,XMMWORD[112+r10]
> + movdqa xmm2,XMMWORD[((-352))+r11]
> +DB 102,15,56,0,195
> + pxor xmm0,xmm4
> +DB 102,15,56,0,194
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +_vpaes_schedule_core:
> +
> +
> +
> +
> +
> +
> + call _vpaes_preheat
> + movdqa xmm8,XMMWORD[$L$k_rcon]
> + movdqu xmm0,XMMWORD[rdi]
> +
> +
> + movdqa xmm3,xmm0
> + lea r11,[$L$k_ipt]
> + call _vpaes_schedule_transform
> + movdqa xmm7,xmm0
> +
> + lea r10,[$L$k_sr]
> + test rcx,rcx
> + jnz NEAR $L$schedule_am_decrypting
> +
> +
> + movdqu XMMWORD[rdx],xmm0
> + jmp NEAR $L$schedule_go
> +
> +$L$schedule_am_decrypting:
> +
> + movdqa xmm1,XMMWORD[r10*1+r8]
> +DB 102,15,56,0,217
> + movdqu XMMWORD[rdx],xmm3
> + xor r8,0x30
> +
> +$L$schedule_go:
> + cmp esi,192
> + ja NEAR $L$schedule_256
> + je NEAR $L$schedule_192
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +$L$schedule_128:
> + mov esi,10
> +
> +$L$oop_schedule_128:
> + call _vpaes_schedule_round
> + dec rsi
> + jz NEAR $L$schedule_mangle_last
> + call _vpaes_schedule_mangle
> + jmp NEAR $L$oop_schedule_128
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +$L$schedule_192:
> + movdqu xmm0,XMMWORD[8+rdi]
> + call _vpaes_schedule_transform
> + movdqa xmm6,xmm0
> + pxor xmm4,xmm4
> + movhlps xmm6,xmm4
> + mov esi,4
> +
> +$L$oop_schedule_192:
> + call _vpaes_schedule_round
> +DB 102,15,58,15,198,8
> + call _vpaes_schedule_mangle
> + call _vpaes_schedule_192_smear
> + call _vpaes_schedule_mangle
> + call _vpaes_schedule_round
> + dec rsi
> + jz NEAR $L$schedule_mangle_last
> + call _vpaes_schedule_mangle
> + call _vpaes_schedule_192_smear
> + jmp NEAR $L$oop_schedule_192
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +$L$schedule_256:
> + movdqu xmm0,XMMWORD[16+rdi]
> + call _vpaes_schedule_transform
> + mov esi,7
> +
> +$L$oop_schedule_256:
> + call _vpaes_schedule_mangle
> + movdqa xmm6,xmm0
> +
> +
> + call _vpaes_schedule_round
> + dec rsi
> + jz NEAR $L$schedule_mangle_last
> + call _vpaes_schedule_mangle
> +
> +
> + pshufd xmm0,xmm0,0xFF
> + movdqa xmm5,xmm7
> + movdqa xmm7,xmm6
> + call _vpaes_schedule_low_round
> + movdqa xmm7,xmm5
> +
> + jmp NEAR $L$oop_schedule_256
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +$L$schedule_mangle_last:
> +
> + lea r11,[$L$k_deskew]
> + test rcx,rcx
> + jnz NEAR $L$schedule_mangle_last_dec
> +
> +
> + movdqa xmm1,XMMWORD[r10*1+r8]
> +DB 102,15,56,0,193
> + lea r11,[$L$k_opt]
> + add rdx,32
> +
> +$L$schedule_mangle_last_dec:
> + add rdx,-16
> + pxor xmm0,XMMWORD[$L$k_s63]
> + call _vpaes_schedule_transform
> + movdqu XMMWORD[rdx],xmm0
> +
> +
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> + pxor xmm6,xmm6
> + pxor xmm7,xmm7
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +_vpaes_schedule_192_smear:
> +
> + pshufd xmm1,xmm6,0x80
> + pshufd xmm0,xmm7,0xFE
> + pxor xmm6,xmm1
> + pxor xmm1,xmm1
> + pxor xmm6,xmm0
> + movdqa xmm0,xmm6
> + movhlps xmm6,xmm1
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +_vpaes_schedule_round:
> +
> +
> + pxor xmm1,xmm1
> +DB 102,65,15,58,15,200,15
> +DB 102,69,15,58,15,192,15
> + pxor xmm7,xmm1
> +
> +
> + pshufd xmm0,xmm0,0xFF
> +DB 102,15,58,15,192,1
> +
> +
> +
> +
> +_vpaes_schedule_low_round:
> +
> + movdqa xmm1,xmm7
> + pslldq xmm7,4
> + pxor xmm7,xmm1
> + movdqa xmm1,xmm7
> + pslldq xmm7,8
> + pxor xmm7,xmm1
> + pxor xmm7,XMMWORD[$L$k_s63]
> +
> +
> + movdqa xmm1,xmm9
> + pandn xmm1,xmm0
> + psrld xmm1,4
> + pand xmm0,xmm9
> + movdqa xmm2,xmm11
> +DB 102,15,56,0,208
> + pxor xmm0,xmm1
> + movdqa xmm3,xmm10
> +DB 102,15,56,0,217
> + pxor xmm3,xmm2
> + movdqa xmm4,xmm10
> +DB 102,15,56,0,224
> + pxor xmm4,xmm2
> + movdqa xmm2,xmm10
> +DB 102,15,56,0,211
> + pxor xmm2,xmm0
> + movdqa xmm3,xmm10
> +DB 102,15,56,0,220
> + pxor xmm3,xmm1
> + movdqa xmm4,xmm13
> +DB 102,15,56,0,226
> + movdqa xmm0,xmm12
> +DB 102,15,56,0,195
> + pxor xmm0,xmm4
> +
> +
> + pxor xmm0,xmm7
> + movdqa xmm7,xmm0
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +_vpaes_schedule_transform:
> +
> + movdqa xmm1,xmm9
> + pandn xmm1,xmm0
> + psrld xmm1,4
> + pand xmm0,xmm9
> + movdqa xmm2,XMMWORD[r11]
> +DB 102,15,56,0,208
> + movdqa xmm0,XMMWORD[16+r11]
> +DB 102,15,56,0,193
> + pxor xmm0,xmm2
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +_vpaes_schedule_mangle:
> +
> + movdqa xmm4,xmm0
> + movdqa xmm5,XMMWORD[$L$k_mc_forward]
> + test rcx,rcx
> + jnz NEAR $L$schedule_mangle_dec
> +
> +
> + add rdx,16
> + pxor xmm4,XMMWORD[$L$k_s63]
> +DB 102,15,56,0,229
> + movdqa xmm3,xmm4
> +DB 102,15,56,0,229
> + pxor xmm3,xmm4
> +DB 102,15,56,0,229
> + pxor xmm3,xmm4
> +
> + jmp NEAR $L$schedule_mangle_both
> +ALIGN 16
> +$L$schedule_mangle_dec:
> +
> + lea r11,[$L$k_dksd]
> + movdqa xmm1,xmm9
> + pandn xmm1,xmm4
> + psrld xmm1,4
> + pand xmm4,xmm9
> +
> + movdqa xmm2,XMMWORD[r11]
> +DB 102,15,56,0,212
> + movdqa xmm3,XMMWORD[16+r11]
> +DB 102,15,56,0,217
> + pxor xmm3,xmm2
> +DB 102,15,56,0,221
> +
> + movdqa xmm2,XMMWORD[32+r11]
> +DB 102,15,56,0,212
> + pxor xmm2,xmm3
> + movdqa xmm3,XMMWORD[48+r11]
> +DB 102,15,56,0,217
> + pxor xmm3,xmm2
> +DB 102,15,56,0,221
> +
> + movdqa xmm2,XMMWORD[64+r11]
> +DB 102,15,56,0,212
> + pxor xmm2,xmm3
> + movdqa xmm3,XMMWORD[80+r11]
> +DB 102,15,56,0,217
> + pxor xmm3,xmm2
> +DB 102,15,56,0,221
> +
> + movdqa xmm2,XMMWORD[96+r11]
> +DB 102,15,56,0,212
> + pxor xmm2,xmm3
> + movdqa xmm3,XMMWORD[112+r11]
> +DB 102,15,56,0,217
> + pxor xmm3,xmm2
> +
> + add rdx,-16
> +
> +$L$schedule_mangle_both:
> + movdqa xmm1,XMMWORD[r10*1+r8]
> +DB 102,15,56,0,217
> + add r8,-16
> + and r8,0x30
> + movdqu XMMWORD[rdx],xmm3
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +
> +
> +global vpaes_set_encrypt_key
> +
> +ALIGN 16
> +vpaes_set_encrypt_key:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_vpaes_set_encrypt_key:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + lea rsp,[((-184))+rsp]
> + movaps XMMWORD[16+rsp],xmm6
> + movaps XMMWORD[32+rsp],xmm7
> + movaps XMMWORD[48+rsp],xmm8
> + movaps XMMWORD[64+rsp],xmm9
> + movaps XMMWORD[80+rsp],xmm10
> + movaps XMMWORD[96+rsp],xmm11
> + movaps XMMWORD[112+rsp],xmm12
> + movaps XMMWORD[128+rsp],xmm13
> + movaps XMMWORD[144+rsp],xmm14
> + movaps XMMWORD[160+rsp],xmm15
> +$L$enc_key_body:
> + mov eax,esi
> + shr eax,5
> + add eax,5
> + mov DWORD[240+rdx],eax
> +
> + mov ecx,0
> + mov r8d,0x30
> + call _vpaes_schedule_core
> + movaps xmm6,XMMWORD[16+rsp]
> + movaps xmm7,XMMWORD[32+rsp]
> + movaps xmm8,XMMWORD[48+rsp]
> + movaps xmm9,XMMWORD[64+rsp]
> + movaps xmm10,XMMWORD[80+rsp]
> + movaps xmm11,XMMWORD[96+rsp]
> + movaps xmm12,XMMWORD[112+rsp]
> + movaps xmm13,XMMWORD[128+rsp]
> + movaps xmm14,XMMWORD[144+rsp]
> + movaps xmm15,XMMWORD[160+rsp]
> + lea rsp,[184+rsp]
> +$L$enc_key_epilogue:
> + xor eax,eax
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_vpaes_set_encrypt_key:
> +
> +global vpaes_set_decrypt_key
> +
> +ALIGN 16
> +vpaes_set_decrypt_key:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_vpaes_set_decrypt_key:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + lea rsp,[((-184))+rsp]
> + movaps XMMWORD[16+rsp],xmm6
> + movaps XMMWORD[32+rsp],xmm7
> + movaps XMMWORD[48+rsp],xmm8
> + movaps XMMWORD[64+rsp],xmm9
> + movaps XMMWORD[80+rsp],xmm10
> + movaps XMMWORD[96+rsp],xmm11
> + movaps XMMWORD[112+rsp],xmm12
> + movaps XMMWORD[128+rsp],xmm13
> + movaps XMMWORD[144+rsp],xmm14
> + movaps XMMWORD[160+rsp],xmm15
> +$L$dec_key_body:
> + mov eax,esi
> + shr eax,5
> + add eax,5
> + mov DWORD[240+rdx],eax
> + shl eax,4
> + lea rdx,[16+rax*1+rdx]
> +
> + mov ecx,1
> + mov r8d,esi
> + shr r8d,1
> + and r8d,32
> + xor r8d,32
> + call _vpaes_schedule_core
> + movaps xmm6,XMMWORD[16+rsp]
> + movaps xmm7,XMMWORD[32+rsp]
> + movaps xmm8,XMMWORD[48+rsp]
> + movaps xmm9,XMMWORD[64+rsp]
> + movaps xmm10,XMMWORD[80+rsp]
> + movaps xmm11,XMMWORD[96+rsp]
> + movaps xmm12,XMMWORD[112+rsp]
> + movaps xmm13,XMMWORD[128+rsp]
> + movaps xmm14,XMMWORD[144+rsp]
> + movaps xmm15,XMMWORD[160+rsp]
> + lea rsp,[184+rsp]
> +$L$dec_key_epilogue:
> + xor eax,eax
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_vpaes_set_decrypt_key:
> +
> +global vpaes_encrypt
> +
> +ALIGN 16
> +vpaes_encrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_vpaes_encrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + lea rsp,[((-184))+rsp]
> + movaps XMMWORD[16+rsp],xmm6
> + movaps XMMWORD[32+rsp],xmm7
> + movaps XMMWORD[48+rsp],xmm8
> + movaps XMMWORD[64+rsp],xmm9
> + movaps XMMWORD[80+rsp],xmm10
> + movaps XMMWORD[96+rsp],xmm11
> + movaps XMMWORD[112+rsp],xmm12
> + movaps XMMWORD[128+rsp],xmm13
> + movaps XMMWORD[144+rsp],xmm14
> + movaps XMMWORD[160+rsp],xmm15
> +$L$enc_body:
> + movdqu xmm0,XMMWORD[rdi]
> + call _vpaes_preheat
> + call _vpaes_encrypt_core
> + movdqu XMMWORD[rsi],xmm0
> + movaps xmm6,XMMWORD[16+rsp]
> + movaps xmm7,XMMWORD[32+rsp]
> + movaps xmm8,XMMWORD[48+rsp]
> + movaps xmm9,XMMWORD[64+rsp]
> + movaps xmm10,XMMWORD[80+rsp]
> + movaps xmm11,XMMWORD[96+rsp]
> + movaps xmm12,XMMWORD[112+rsp]
> + movaps xmm13,XMMWORD[128+rsp]
> + movaps xmm14,XMMWORD[144+rsp]
> + movaps xmm15,XMMWORD[160+rsp]
> + lea rsp,[184+rsp]
> +$L$enc_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_vpaes_encrypt:
> +
> +global vpaes_decrypt
> +
> +ALIGN 16
> +vpaes_decrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_vpaes_decrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + lea rsp,[((-184))+rsp]
> + movaps XMMWORD[16+rsp],xmm6
> + movaps XMMWORD[32+rsp],xmm7
> + movaps XMMWORD[48+rsp],xmm8
> + movaps XMMWORD[64+rsp],xmm9
> + movaps XMMWORD[80+rsp],xmm10
> + movaps XMMWORD[96+rsp],xmm11
> + movaps XMMWORD[112+rsp],xmm12
> + movaps XMMWORD[128+rsp],xmm13
> + movaps XMMWORD[144+rsp],xmm14
> + movaps XMMWORD[160+rsp],xmm15
> +$L$dec_body:
> + movdqu xmm0,XMMWORD[rdi]
> + call _vpaes_preheat
> + call _vpaes_decrypt_core
> + movdqu XMMWORD[rsi],xmm0
> + movaps xmm6,XMMWORD[16+rsp]
> + movaps xmm7,XMMWORD[32+rsp]
> + movaps xmm8,XMMWORD[48+rsp]
> + movaps xmm9,XMMWORD[64+rsp]
> + movaps xmm10,XMMWORD[80+rsp]
> + movaps xmm11,XMMWORD[96+rsp]
> + movaps xmm12,XMMWORD[112+rsp]
> + movaps xmm13,XMMWORD[128+rsp]
> + movaps xmm14,XMMWORD[144+rsp]
> + movaps xmm15,XMMWORD[160+rsp]
> + lea rsp,[184+rsp]
> +$L$dec_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_vpaes_decrypt:
> +global vpaes_cbc_encrypt
> +
> +ALIGN 16
> +vpaes_cbc_encrypt:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_vpaes_cbc_encrypt:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> + mov r8,QWORD[40+rsp]
> + mov r9,QWORD[48+rsp]
> +
> +
> +
> + xchg rdx,rcx
> + sub rcx,16
> + jc NEAR $L$cbc_abort
> + lea rsp,[((-184))+rsp]
> + movaps XMMWORD[16+rsp],xmm6
> + movaps XMMWORD[32+rsp],xmm7
> + movaps XMMWORD[48+rsp],xmm8
> + movaps XMMWORD[64+rsp],xmm9
> + movaps XMMWORD[80+rsp],xmm10
> + movaps XMMWORD[96+rsp],xmm11
> + movaps XMMWORD[112+rsp],xmm12
> + movaps XMMWORD[128+rsp],xmm13
> + movaps XMMWORD[144+rsp],xmm14
> + movaps XMMWORD[160+rsp],xmm15
> +$L$cbc_body:
> + movdqu xmm6,XMMWORD[r8]
> + sub rsi,rdi
> + call _vpaes_preheat
> + cmp r9d,0
> + je NEAR $L$cbc_dec_loop
> + jmp NEAR $L$cbc_enc_loop
> +ALIGN 16
> +$L$cbc_enc_loop:
> + movdqu xmm0,XMMWORD[rdi]
> + pxor xmm0,xmm6
> + call _vpaes_encrypt_core
> + movdqa xmm6,xmm0
> + movdqu XMMWORD[rdi*1+rsi],xmm0
> + lea rdi,[16+rdi]
> + sub rcx,16
> + jnc NEAR $L$cbc_enc_loop
> + jmp NEAR $L$cbc_done
> +ALIGN 16
> +$L$cbc_dec_loop:
> + movdqu xmm0,XMMWORD[rdi]
> + movdqa xmm7,xmm0
> + call _vpaes_decrypt_core
> + pxor xmm0,xmm6
> + movdqa xmm6,xmm7
> + movdqu XMMWORD[rdi*1+rsi],xmm0
> + lea rdi,[16+rdi]
> + sub rcx,16
> + jnc NEAR $L$cbc_dec_loop
> +$L$cbc_done:
> + movdqu XMMWORD[r8],xmm6
> + movaps xmm6,XMMWORD[16+rsp]
> + movaps xmm7,XMMWORD[32+rsp]
> + movaps xmm8,XMMWORD[48+rsp]
> + movaps xmm9,XMMWORD[64+rsp]
> + movaps xmm10,XMMWORD[80+rsp]
> + movaps xmm11,XMMWORD[96+rsp]
> + movaps xmm12,XMMWORD[112+rsp]
> + movaps xmm13,XMMWORD[128+rsp]
> + movaps xmm14,XMMWORD[144+rsp]
> + movaps xmm15,XMMWORD[160+rsp]
> + lea rsp,[184+rsp]
> +$L$cbc_epilogue:
> +$L$cbc_abort:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_vpaes_cbc_encrypt:
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 16
> +_vpaes_preheat:
> +
> + lea r10,[$L$k_s0F]
> + movdqa xmm10,XMMWORD[((-32))+r10]
> + movdqa xmm11,XMMWORD[((-16))+r10]
> + movdqa xmm9,XMMWORD[r10]
> + movdqa xmm13,XMMWORD[48+r10]
> + movdqa xmm12,XMMWORD[64+r10]
> + movdqa xmm15,XMMWORD[80+r10]
> + movdqa xmm14,XMMWORD[96+r10]
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +
> +
> +
> +
> +
> +ALIGN 64
> +_vpaes_consts:
> +$L$k_inv:
> + DQ 0x0E05060F0D080180,0x040703090A0B0C02
> + DQ 0x01040A060F0B0780,0x030D0E0C02050809
> +
> +$L$k_s0F:
> + DQ 0x0F0F0F0F0F0F0F0F,0x0F0F0F0F0F0F0F0F
> +
> +$L$k_ipt:
> + DQ 0xC2B2E8985A2A7000,0xCABAE09052227808
> + DQ 0x4C01307D317C4D00,0xCD80B1FCB0FDCC81
> +
> +$L$k_sb1:
> + DQ 0xB19BE18FCB503E00,0xA5DF7A6E142AF544
> + DQ 0x3618D415FAE22300,0x3BF7CCC10D2ED9EF
> +$L$k_sb2:
> + DQ 0xE27A93C60B712400,0x5EB7E955BC982FCD
> + DQ 0x69EB88400AE12900,0xC2A163C8AB82234A
> +$L$k_sbo:
> + DQ 0xD0D26D176FBDC700,0x15AABF7AC502A878
> + DQ 0xCFE474A55FBB6A00,0x8E1E90D1412B35FA
> +
> +$L$k_mc_forward:
> + DQ 0x0407060500030201,0x0C0F0E0D080B0A09
> + DQ 0x080B0A0904070605,0x000302010C0F0E0D
> + DQ 0x0C0F0E0D080B0A09,0x0407060500030201
> + DQ 0x000302010C0F0E0D,0x080B0A0904070605
> +
> +$L$k_mc_backward:
> + DQ 0x0605040702010003,0x0E0D0C0F0A09080B
> + DQ 0x020100030E0D0C0F,0x0A09080B06050407
> + DQ 0x0E0D0C0F0A09080B,0x0605040702010003
> + DQ 0x0A09080B06050407,0x020100030E0D0C0F
> +
> +$L$k_sr:
> + DQ 0x0706050403020100,0x0F0E0D0C0B0A0908
> + DQ 0x030E09040F0A0500,0x0B06010C07020D08
> + DQ 0x0F060D040B020900,0x070E050C030A0108
> + DQ 0x0B0E0104070A0D00,0x0306090C0F020508
> +
> +$L$k_rcon:
> + DQ 0x1F8391B9AF9DEEB6,0x702A98084D7C7D81
> +
> +$L$k_s63:
> + DQ 0x5B5B5B5B5B5B5B5B,0x5B5B5B5B5B5B5B5B
> +
> +$L$k_opt:
> + DQ 0xFF9F4929D6B66000,0xF7974121DEBE6808
> + DQ 0x01EDBD5150BCEC00,0xE10D5DB1B05C0CE0
> +
> +$L$k_deskew:
> + DQ 0x07E4A34047A4E300,0x1DFEB95A5DBEF91A
> + DQ 0x5F36B5DC83EA6900,0x2841C2ABF49D1E77
> +
> +
> +
> +
> +
> +$L$k_dksd:
> + DQ 0xFEB91A5DA3E44700,0x0740E3A45A1DBEF9
> + DQ 0x41C277F4B5368300,0x5FDC69EAAB289D1E
> +$L$k_dksb:
> + DQ 0x9A4FCA1F8550D500,0x03D653861CC94C99
> + DQ 0x115BEDA7B6FC4A00,0xD993256F7E3482C8
> +$L$k_dkse:
> + DQ 0xD5031CCA1FC9D600,0x53859A4C994F5086
> + DQ 0xA23196054FDC7BE8,0xCD5EF96A20B31487
> +$L$k_dks9:
> + DQ 0xB6116FC87ED9A700,0x4AED933482255BFC
> + DQ 0x4576516227143300,0x8BB89FACE9DAFDCE
> +
> +
> +
> +
> +
> +$L$k_dipt:
> + DQ 0x0F505B040B545F00,0x154A411E114E451A
> + DQ 0x86E383E660056500,0x12771772F491F194
> +
> +$L$k_dsb9:
> + DQ 0x851C03539A86D600,0xCAD51F504F994CC9
> + DQ 0xC03B1789ECD74900,0x725E2C9EB2FBA565
> +$L$k_dsbd:
> + DQ 0x7D57CCDFE6B1A200,0xF56E9B13882A4439
> + DQ 0x3CE2FAF724C6CB00,0x2931180D15DEEFD3
> +$L$k_dsbb:
> + DQ 0xD022649296B44200,0x602646F6B0F2D404
> + DQ 0xC19498A6CD596700,0xF3FF0C3E3255AA6B
> +$L$k_dsbe:
> + DQ 0x46F2929626D4D000,0x2242600464B4F6B0
> + DQ 0x0C55A6CDFFAAC100,0x9467F36B98593E32
> +$L$k_dsbo:
> + DQ 0x1387EA537EF94000,0xC7AA6DB9D4943E2D
> + DQ 0x12D7560F93441D00,0xCA4B8159D8C58E9C
> +DB 86,101,99,116,111,114,32,80,101,114,109,117,116,97,116,105
> +DB 111,110,32,65,69,83,32,102,111,114,32,120,56,54,95,54
> +DB 52,47,83,83,83,69,51,44,32,77,105,107,101,32,72,97
> +DB 109,98,117,114,103,32,40,83,116,97,110,102,111,114,100,32
> +DB 85,110,105,118,101,114,115,105,116,121,41,0
> +ALIGN 64
> +
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$in_prologue
> +
> + lea rsi,[16+rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> + lea rax,[184+rax]
> +
> +$L$in_prologue:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_vpaes_set_encrypt_key wrt ..imagebase
> + DD $L$SEH_end_vpaes_set_encrypt_key wrt ..imagebase
> + DD $L$SEH_info_vpaes_set_encrypt_key wrt ..imagebase
> +
> + DD $L$SEH_begin_vpaes_set_decrypt_key wrt ..imagebase
> + DD $L$SEH_end_vpaes_set_decrypt_key wrt ..imagebase
> + DD $L$SEH_info_vpaes_set_decrypt_key wrt ..imagebase
> +
> + DD $L$SEH_begin_vpaes_encrypt wrt ..imagebase
> + DD $L$SEH_end_vpaes_encrypt wrt ..imagebase
> + DD $L$SEH_info_vpaes_encrypt wrt ..imagebase
> +
> + DD $L$SEH_begin_vpaes_decrypt wrt ..imagebase
> + DD $L$SEH_end_vpaes_decrypt wrt ..imagebase
> + DD $L$SEH_info_vpaes_decrypt wrt ..imagebase
> +
> + DD $L$SEH_begin_vpaes_cbc_encrypt wrt ..imagebase
> + DD $L$SEH_end_vpaes_cbc_encrypt wrt ..imagebase
> + DD $L$SEH_info_vpaes_cbc_encrypt wrt ..imagebase
> +
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_vpaes_set_encrypt_key:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$enc_key_body wrt ..imagebase,$L$enc_key_epilogue
> wrt ..imagebase
> +$L$SEH_info_vpaes_set_decrypt_key:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$dec_key_body wrt ..imagebase,$L$dec_key_epilogue
> wrt ..imagebase
> +$L$SEH_info_vpaes_encrypt:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$enc_body wrt ..imagebase,$L$enc_epilogue wrt ..imagebase
> +$L$SEH_info_vpaes_decrypt:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$dec_body wrt ..imagebase,$L$dec_epilogue wrt ..imagebase
> +$L$SEH_info_vpaes_cbc_encrypt:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$cbc_body wrt ..imagebase,$L$cbc_epilogue wrt ..imagebase
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/modes/aesni-gcm-
> x86_64.nasm b/CryptoPkg/Library/OpensslLib/X64/crypto/modes/aesni-gcm-
> x86_64.nasm
> new file mode 100644
> index 0000000000..9e1a2d0a40
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/modes/aesni-gcm-
> x86_64.nasm
> @@ -0,0 +1,34 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/modes/asm/aesni-gcm-x86_64.pl
> +;
> +; Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +
> +global aesni_gcm_encrypt
> +
> +aesni_gcm_encrypt:
> +
> + xor eax,eax
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +global aesni_gcm_decrypt
> +
> +aesni_gcm_decrypt:
> +
> + xor eax,eax
> + DB 0F3h,0C3h ;repret
> +
> +
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/modes/ghash-
> x86_64.nasm b/CryptoPkg/Library/OpensslLib/X64/crypto/modes/ghash-
> x86_64.nasm
> new file mode 100644
> index 0000000000..60f283d5fb
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/modes/ghash-x86_64.nasm
> @@ -0,0 +1,1569 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/modes/asm/ghash-x86_64.pl
> +;
> +; Copyright 2010-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +EXTERN OPENSSL_ia32cap_P
> +
> +global gcm_gmult_4bit
> +
> +ALIGN 16
> +gcm_gmult_4bit:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_gcm_gmult_4bit:
> + mov rdi,rcx
> + mov rsi,rdx
> +
> +
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + push r15
> +
> + sub rsp,280
> +
> +$L$gmult_prologue:
> +
> + movzx r8,BYTE[15+rdi]
> + lea r11,[$L$rem_4bit]
> + xor rax,rax
> + xor rbx,rbx
> + mov al,r8b
> + mov bl,r8b
> + shl al,4
> + mov rcx,14
> + mov r8,QWORD[8+rax*1+rsi]
> + mov r9,QWORD[rax*1+rsi]
> + and bl,0xf0
> + mov rdx,r8
> + jmp NEAR $L$oop1
> +
> +ALIGN 16
> +$L$oop1:
> + shr r8,4
> + and rdx,0xf
> + mov r10,r9
> + mov al,BYTE[rcx*1+rdi]
> + shr r9,4
> + xor r8,QWORD[8+rbx*1+rsi]
> + shl r10,60
> + xor r9,QWORD[rbx*1+rsi]
> + mov bl,al
> + xor r9,QWORD[rdx*8+r11]
> + mov rdx,r8
> + shl al,4
> + xor r8,r10
> + dec rcx
> + js NEAR $L$break1
> +
> + shr r8,4
> + and rdx,0xf
> + mov r10,r9
> + shr r9,4
> + xor r8,QWORD[8+rax*1+rsi]
> + shl r10,60
> + xor r9,QWORD[rax*1+rsi]
> + and bl,0xf0
> + xor r9,QWORD[rdx*8+r11]
> + mov rdx,r8
> + xor r8,r10
> + jmp NEAR $L$oop1
> +
> +ALIGN 16
> +$L$break1:
> + shr r8,4
> + and rdx,0xf
> + mov r10,r9
> + shr r9,4
> + xor r8,QWORD[8+rax*1+rsi]
> + shl r10,60
> + xor r9,QWORD[rax*1+rsi]
> + and bl,0xf0
> + xor r9,QWORD[rdx*8+r11]
> + mov rdx,r8
> + xor r8,r10
> +
> + shr r8,4
> + and rdx,0xf
> + mov r10,r9
> + shr r9,4
> + xor r8,QWORD[8+rbx*1+rsi]
> + shl r10,60
> + xor r9,QWORD[rbx*1+rsi]
> + xor r8,r10
> + xor r9,QWORD[rdx*8+r11]
> +
> + bswap r8
> + bswap r9
> + mov QWORD[8+rdi],r8
> + mov QWORD[rdi],r9
> +
> + lea rsi,[((280+48))+rsp]
> +
> + mov rbx,QWORD[((-8))+rsi]
> +
> + lea rsp,[rsi]
> +
> +$L$gmult_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_gcm_gmult_4bit:
> +global gcm_ghash_4bit
> +
> +ALIGN 16
> +gcm_ghash_4bit:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_gcm_ghash_4bit:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> + mov rcx,r9
> +
> +
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + push r15
> +
> + sub rsp,280
> +
> +$L$ghash_prologue:
> + mov r14,rdx
> + mov r15,rcx
> + sub rsi,-128
> + lea rbp,[((16+128))+rsp]
> + xor edx,edx
> + mov r8,QWORD[((0+0-128))+rsi]
> + mov rax,QWORD[((0+8-128))+rsi]
> + mov dl,al
> + shr rax,4
> + mov r10,r8
> + shr r8,4
> + mov r9,QWORD[((16+0-128))+rsi]
> + shl dl,4
> + mov rbx,QWORD[((16+8-128))+rsi]
> + shl r10,60
> + mov BYTE[rsp],dl
> + or rax,r10
> + mov dl,bl
> + shr rbx,4
> + mov r10,r9
> + shr r9,4
> + mov QWORD[rbp],r8
> + mov r8,QWORD[((32+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((0-128))+rbp],rax
> + mov rax,QWORD[((32+8-128))+rsi]
> + shl r10,60
> + mov BYTE[1+rsp],dl
> + or rbx,r10
> + mov dl,al
> + shr rax,4
> + mov r10,r8
> + shr r8,4
> + mov QWORD[8+rbp],r9
> + mov r9,QWORD[((48+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((8-128))+rbp],rbx
> + mov rbx,QWORD[((48+8-128))+rsi]
> + shl r10,60
> + mov BYTE[2+rsp],dl
> + or rax,r10
> + mov dl,bl
> + shr rbx,4
> + mov r10,r9
> + shr r9,4
> + mov QWORD[16+rbp],r8
> + mov r8,QWORD[((64+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((16-128))+rbp],rax
> + mov rax,QWORD[((64+8-128))+rsi]
> + shl r10,60
> + mov BYTE[3+rsp],dl
> + or rbx,r10
> + mov dl,al
> + shr rax,4
> + mov r10,r8
> + shr r8,4
> + mov QWORD[24+rbp],r9
> + mov r9,QWORD[((80+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((24-128))+rbp],rbx
> + mov rbx,QWORD[((80+8-128))+rsi]
> + shl r10,60
> + mov BYTE[4+rsp],dl
> + or rax,r10
> + mov dl,bl
> + shr rbx,4
> + mov r10,r9
> + shr r9,4
> + mov QWORD[32+rbp],r8
> + mov r8,QWORD[((96+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((32-128))+rbp],rax
> + mov rax,QWORD[((96+8-128))+rsi]
> + shl r10,60
> + mov BYTE[5+rsp],dl
> + or rbx,r10
> + mov dl,al
> + shr rax,4
> + mov r10,r8
> + shr r8,4
> + mov QWORD[40+rbp],r9
> + mov r9,QWORD[((112+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((40-128))+rbp],rbx
> + mov rbx,QWORD[((112+8-128))+rsi]
> + shl r10,60
> + mov BYTE[6+rsp],dl
> + or rax,r10
> + mov dl,bl
> + shr rbx,4
> + mov r10,r9
> + shr r9,4
> + mov QWORD[48+rbp],r8
> + mov r8,QWORD[((128+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((48-128))+rbp],rax
> + mov rax,QWORD[((128+8-128))+rsi]
> + shl r10,60
> + mov BYTE[7+rsp],dl
> + or rbx,r10
> + mov dl,al
> + shr rax,4
> + mov r10,r8
> + shr r8,4
> + mov QWORD[56+rbp],r9
> + mov r9,QWORD[((144+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((56-128))+rbp],rbx
> + mov rbx,QWORD[((144+8-128))+rsi]
> + shl r10,60
> + mov BYTE[8+rsp],dl
> + or rax,r10
> + mov dl,bl
> + shr rbx,4
> + mov r10,r9
> + shr r9,4
> + mov QWORD[64+rbp],r8
> + mov r8,QWORD[((160+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((64-128))+rbp],rax
> + mov rax,QWORD[((160+8-128))+rsi]
> + shl r10,60
> + mov BYTE[9+rsp],dl
> + or rbx,r10
> + mov dl,al
> + shr rax,4
> + mov r10,r8
> + shr r8,4
> + mov QWORD[72+rbp],r9
> + mov r9,QWORD[((176+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((72-128))+rbp],rbx
> + mov rbx,QWORD[((176+8-128))+rsi]
> + shl r10,60
> + mov BYTE[10+rsp],dl
> + or rax,r10
> + mov dl,bl
> + shr rbx,4
> + mov r10,r9
> + shr r9,4
> + mov QWORD[80+rbp],r8
> + mov r8,QWORD[((192+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((80-128))+rbp],rax
> + mov rax,QWORD[((192+8-128))+rsi]
> + shl r10,60
> + mov BYTE[11+rsp],dl
> + or rbx,r10
> + mov dl,al
> + shr rax,4
> + mov r10,r8
> + shr r8,4
> + mov QWORD[88+rbp],r9
> + mov r9,QWORD[((208+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((88-128))+rbp],rbx
> + mov rbx,QWORD[((208+8-128))+rsi]
> + shl r10,60
> + mov BYTE[12+rsp],dl
> + or rax,r10
> + mov dl,bl
> + shr rbx,4
> + mov r10,r9
> + shr r9,4
> + mov QWORD[96+rbp],r8
> + mov r8,QWORD[((224+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((96-128))+rbp],rax
> + mov rax,QWORD[((224+8-128))+rsi]
> + shl r10,60
> + mov BYTE[13+rsp],dl
> + or rbx,r10
> + mov dl,al
> + shr rax,4
> + mov r10,r8
> + shr r8,4
> + mov QWORD[104+rbp],r9
> + mov r9,QWORD[((240+0-128))+rsi]
> + shl dl,4
> + mov QWORD[((104-128))+rbp],rbx
> + mov rbx,QWORD[((240+8-128))+rsi]
> + shl r10,60
> + mov BYTE[14+rsp],dl
> + or rax,r10
> + mov dl,bl
> + shr rbx,4
> + mov r10,r9
> + shr r9,4
> + mov QWORD[112+rbp],r8
> + shl dl,4
> + mov QWORD[((112-128))+rbp],rax
> + shl r10,60
> + mov BYTE[15+rsp],dl
> + or rbx,r10
> + mov QWORD[120+rbp],r9
> + mov QWORD[((120-128))+rbp],rbx
> + add rsi,-128
> + mov r8,QWORD[8+rdi]
> + mov r9,QWORD[rdi]
> + add r15,r14
> + lea r11,[$L$rem_8bit]
> + jmp NEAR $L$outer_loop
> +ALIGN 16
> +$L$outer_loop:
> + xor r9,QWORD[r14]
> + mov rdx,QWORD[8+r14]
> + lea r14,[16+r14]
> + xor rdx,r8
> + mov QWORD[rdi],r9
> + mov QWORD[8+rdi],rdx
> + shr rdx,32
> + xor rax,rax
> + rol edx,8
> + mov al,dl
> + movzx ebx,dl
> + shl al,4
> + shr ebx,4
> + rol edx,8
> + mov r8,QWORD[8+rax*1+rsi]
> + mov r9,QWORD[rax*1+rsi]
> + mov al,dl
> + movzx ecx,dl
> + shl al,4
> + movzx r12,BYTE[rbx*1+rsp]
> + shr ecx,4
> + xor r12,r8
> + mov r10,r9
> + shr r8,8
> + movzx r12,r12b
> + shr r9,8
> + xor r8,QWORD[((-128))+rbx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rbx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r12,WORD[r12*2+r11]
> + movzx ebx,dl
> + shl al,4
> + movzx r13,BYTE[rcx*1+rsp]
> + shr ebx,4
> + shl r12,48
> + xor r13,r8
> + mov r10,r9
> + xor r9,r12
> + shr r8,8
> + movzx r13,r13b
> + shr r9,8
> + xor r8,QWORD[((-128))+rcx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rcx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r13,WORD[r13*2+r11]
> + movzx ecx,dl
> + shl al,4
> + movzx r12,BYTE[rbx*1+rsp]
> + shr ecx,4
> + shl r13,48
> + xor r12,r8
> + mov r10,r9
> + xor r9,r13
> + shr r8,8
> + movzx r12,r12b
> + mov edx,DWORD[8+rdi]
> + shr r9,8
> + xor r8,QWORD[((-128))+rbx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rbx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r12,WORD[r12*2+r11]
> + movzx ebx,dl
> + shl al,4
> + movzx r13,BYTE[rcx*1+rsp]
> + shr ebx,4
> + shl r12,48
> + xor r13,r8
> + mov r10,r9
> + xor r9,r12
> + shr r8,8
> + movzx r13,r13b
> + shr r9,8
> + xor r8,QWORD[((-128))+rcx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rcx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r13,WORD[r13*2+r11]
> + movzx ecx,dl
> + shl al,4
> + movzx r12,BYTE[rbx*1+rsp]
> + shr ecx,4
> + shl r13,48
> + xor r12,r8
> + mov r10,r9
> + xor r9,r13
> + shr r8,8
> + movzx r12,r12b
> + shr r9,8
> + xor r8,QWORD[((-128))+rbx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rbx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r12,WORD[r12*2+r11]
> + movzx ebx,dl
> + shl al,4
> + movzx r13,BYTE[rcx*1+rsp]
> + shr ebx,4
> + shl r12,48
> + xor r13,r8
> + mov r10,r9
> + xor r9,r12
> + shr r8,8
> + movzx r13,r13b
> + shr r9,8
> + xor r8,QWORD[((-128))+rcx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rcx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r13,WORD[r13*2+r11]
> + movzx ecx,dl
> + shl al,4
> + movzx r12,BYTE[rbx*1+rsp]
> + shr ecx,4
> + shl r13,48
> + xor r12,r8
> + mov r10,r9
> + xor r9,r13
> + shr r8,8
> + movzx r12,r12b
> + mov edx,DWORD[4+rdi]
> + shr r9,8
> + xor r8,QWORD[((-128))+rbx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rbx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r12,WORD[r12*2+r11]
> + movzx ebx,dl
> + shl al,4
> + movzx r13,BYTE[rcx*1+rsp]
> + shr ebx,4
> + shl r12,48
> + xor r13,r8
> + mov r10,r9
> + xor r9,r12
> + shr r8,8
> + movzx r13,r13b
> + shr r9,8
> + xor r8,QWORD[((-128))+rcx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rcx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r13,WORD[r13*2+r11]
> + movzx ecx,dl
> + shl al,4
> + movzx r12,BYTE[rbx*1+rsp]
> + shr ecx,4
> + shl r13,48
> + xor r12,r8
> + mov r10,r9
> + xor r9,r13
> + shr r8,8
> + movzx r12,r12b
> + shr r9,8
> + xor r8,QWORD[((-128))+rbx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rbx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r12,WORD[r12*2+r11]
> + movzx ebx,dl
> + shl al,4
> + movzx r13,BYTE[rcx*1+rsp]
> + shr ebx,4
> + shl r12,48
> + xor r13,r8
> + mov r10,r9
> + xor r9,r12
> + shr r8,8
> + movzx r13,r13b
> + shr r9,8
> + xor r8,QWORD[((-128))+rcx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rcx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r13,WORD[r13*2+r11]
> + movzx ecx,dl
> + shl al,4
> + movzx r12,BYTE[rbx*1+rsp]
> + shr ecx,4
> + shl r13,48
> + xor r12,r8
> + mov r10,r9
> + xor r9,r13
> + shr r8,8
> + movzx r12,r12b
> + mov edx,DWORD[rdi]
> + shr r9,8
> + xor r8,QWORD[((-128))+rbx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rbx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r12,WORD[r12*2+r11]
> + movzx ebx,dl
> + shl al,4
> + movzx r13,BYTE[rcx*1+rsp]
> + shr ebx,4
> + shl r12,48
> + xor r13,r8
> + mov r10,r9
> + xor r9,r12
> + shr r8,8
> + movzx r13,r13b
> + shr r9,8
> + xor r8,QWORD[((-128))+rcx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rcx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r13,WORD[r13*2+r11]
> + movzx ecx,dl
> + shl al,4
> + movzx r12,BYTE[rbx*1+rsp]
> + shr ecx,4
> + shl r13,48
> + xor r12,r8
> + mov r10,r9
> + xor r9,r13
> + shr r8,8
> + movzx r12,r12b
> + shr r9,8
> + xor r8,QWORD[((-128))+rbx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rbx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r12,WORD[r12*2+r11]
> + movzx ebx,dl
> + shl al,4
> + movzx r13,BYTE[rcx*1+rsp]
> + shr ebx,4
> + shl r12,48
> + xor r13,r8
> + mov r10,r9
> + xor r9,r12
> + shr r8,8
> + movzx r13,r13b
> + shr r9,8
> + xor r8,QWORD[((-128))+rcx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rcx*8+rbp]
> + rol edx,8
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + mov al,dl
> + xor r8,r10
> + movzx r13,WORD[r13*2+r11]
> + movzx ecx,dl
> + shl al,4
> + movzx r12,BYTE[rbx*1+rsp]
> + and ecx,240
> + shl r13,48
> + xor r12,r8
> + mov r10,r9
> + xor r9,r13
> + shr r8,8
> + movzx r12,r12b
> + mov edx,DWORD[((-4))+rdi]
> + shr r9,8
> + xor r8,QWORD[((-128))+rbx*8+rbp]
> + shl r10,56
> + xor r9,QWORD[rbx*8+rbp]
> + movzx r12,WORD[r12*2+r11]
> + xor r8,QWORD[8+rax*1+rsi]
> + xor r9,QWORD[rax*1+rsi]
> + shl r12,48
> + xor r8,r10
> + xor r9,r12
> + movzx r13,r8b
> + shr r8,4
> + mov r10,r9
> + shl r13b,4
> + shr r9,4
> + xor r8,QWORD[8+rcx*1+rsi]
> + movzx r13,WORD[r13*2+r11]
> + shl r10,60
> + xor r9,QWORD[rcx*1+rsi]
> + xor r8,r10
> + shl r13,48
> + bswap r8
> + xor r9,r13
> + bswap r9
> + cmp r14,r15
> + jb NEAR $L$outer_loop
> + mov QWORD[8+rdi],r8
> + mov QWORD[rdi],r9
> +
> + lea rsi,[((280+48))+rsp]
> +
> + mov r15,QWORD[((-48))+rsi]
> +
> + mov r14,QWORD[((-40))+rsi]
> +
> + mov r13,QWORD[((-32))+rsi]
> +
> + mov r12,QWORD[((-24))+rsi]
> +
> + mov rbp,QWORD[((-16))+rsi]
> +
> + mov rbx,QWORD[((-8))+rsi]
> +
> + lea rsp,[rsi]
> +
> +$L$ghash_epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_gcm_ghash_4bit:
> +global gcm_init_clmul
> +
> +ALIGN 16
> +gcm_init_clmul:
> +
> +$L$_init_clmul:
> +$L$SEH_begin_gcm_init_clmul:
> +
> +DB 0x48,0x83,0xec,0x18
> +DB 0x0f,0x29,0x34,0x24
> + movdqu xmm2,XMMWORD[rdx]
> + pshufd xmm2,xmm2,78
> +
> +
> + pshufd xmm4,xmm2,255
> + movdqa xmm3,xmm2
> + psllq xmm2,1
> + pxor xmm5,xmm5
> + psrlq xmm3,63
> + pcmpgtd xmm5,xmm4
> + pslldq xmm3,8
> + por xmm2,xmm3
> +
> +
> + pand xmm5,XMMWORD[$L$0x1c2_polynomial]
> + pxor xmm2,xmm5
> +
> +
> + pshufd xmm6,xmm2,78
> + movdqa xmm0,xmm2
> + pxor xmm6,xmm2
> + movdqa xmm1,xmm0
> + pshufd xmm3,xmm0,78
> + pxor xmm3,xmm0
> +DB 102,15,58,68,194,0
> +DB 102,15,58,68,202,17
> +DB 102,15,58,68,222,0
> + pxor xmm3,xmm0
> + pxor xmm3,xmm1
> +
> + movdqa xmm4,xmm3
> + psrldq xmm3,8
> + pslldq xmm4,8
> + pxor xmm1,xmm3
> + pxor xmm0,xmm4
> +
> + movdqa xmm4,xmm0
> + movdqa xmm3,xmm0
> + psllq xmm0,5
> + pxor xmm3,xmm0
> + psllq xmm0,1
> + pxor xmm0,xmm3
> + psllq xmm0,57
> + movdqa xmm3,xmm0
> + pslldq xmm0,8
> + psrldq xmm3,8
> + pxor xmm0,xmm4
> + pxor xmm1,xmm3
> +
> +
> + movdqa xmm4,xmm0
> + psrlq xmm0,1
> + pxor xmm1,xmm4
> + pxor xmm4,xmm0
> + psrlq xmm0,5
> + pxor xmm0,xmm4
> + psrlq xmm0,1
> + pxor xmm0,xmm1
> + pshufd xmm3,xmm2,78
> + pshufd xmm4,xmm0,78
> + pxor xmm3,xmm2
> + movdqu XMMWORD[rcx],xmm2
> + pxor xmm4,xmm0
> + movdqu XMMWORD[16+rcx],xmm0
> +DB 102,15,58,15,227,8
> + movdqu XMMWORD[32+rcx],xmm4
> + movdqa xmm1,xmm0
> + pshufd xmm3,xmm0,78
> + pxor xmm3,xmm0
> +DB 102,15,58,68,194,0
> +DB 102,15,58,68,202,17
> +DB 102,15,58,68,222,0
> + pxor xmm3,xmm0
> + pxor xmm3,xmm1
> +
> + movdqa xmm4,xmm3
> + psrldq xmm3,8
> + pslldq xmm4,8
> + pxor xmm1,xmm3
> + pxor xmm0,xmm4
> +
> + movdqa xmm4,xmm0
> + movdqa xmm3,xmm0
> + psllq xmm0,5
> + pxor xmm3,xmm0
> + psllq xmm0,1
> + pxor xmm0,xmm3
> + psllq xmm0,57
> + movdqa xmm3,xmm0
> + pslldq xmm0,8
> + psrldq xmm3,8
> + pxor xmm0,xmm4
> + pxor xmm1,xmm3
> +
> +
> + movdqa xmm4,xmm0
> + psrlq xmm0,1
> + pxor xmm1,xmm4
> + pxor xmm4,xmm0
> + psrlq xmm0,5
> + pxor xmm0,xmm4
> + psrlq xmm0,1
> + pxor xmm0,xmm1
> + movdqa xmm5,xmm0
> + movdqa xmm1,xmm0
> + pshufd xmm3,xmm0,78
> + pxor xmm3,xmm0
> +DB 102,15,58,68,194,0
> +DB 102,15,58,68,202,17
> +DB 102,15,58,68,222,0
> + pxor xmm3,xmm0
> + pxor xmm3,xmm1
> +
> + movdqa xmm4,xmm3
> + psrldq xmm3,8
> + pslldq xmm4,8
> + pxor xmm1,xmm3
> + pxor xmm0,xmm4
> +
> + movdqa xmm4,xmm0
> + movdqa xmm3,xmm0
> + psllq xmm0,5
> + pxor xmm3,xmm0
> + psllq xmm0,1
> + pxor xmm0,xmm3
> + psllq xmm0,57
> + movdqa xmm3,xmm0
> + pslldq xmm0,8
> + psrldq xmm3,8
> + pxor xmm0,xmm4
> + pxor xmm1,xmm3
> +
> +
> + movdqa xmm4,xmm0
> + psrlq xmm0,1
> + pxor xmm1,xmm4
> + pxor xmm4,xmm0
> + psrlq xmm0,5
> + pxor xmm0,xmm4
> + psrlq xmm0,1
> + pxor xmm0,xmm1
> + pshufd xmm3,xmm5,78
> + pshufd xmm4,xmm0,78
> + pxor xmm3,xmm5
> + movdqu XMMWORD[48+rcx],xmm5
> + pxor xmm4,xmm0
> + movdqu XMMWORD[64+rcx],xmm0
> +DB 102,15,58,15,227,8
> + movdqu XMMWORD[80+rcx],xmm4
> + movaps xmm6,XMMWORD[rsp]
> + lea rsp,[24+rsp]
> +$L$SEH_end_gcm_init_clmul:
> + DB 0F3h,0C3h ;repret
> +
> +
> +global gcm_gmult_clmul
> +
> +ALIGN 16
> +gcm_gmult_clmul:
> +
> +$L$_gmult_clmul:
> + movdqu xmm0,XMMWORD[rcx]
> + movdqa xmm5,XMMWORD[$L$bswap_mask]
> + movdqu xmm2,XMMWORD[rdx]
> + movdqu xmm4,XMMWORD[32+rdx]
> +DB 102,15,56,0,197
> + movdqa xmm1,xmm0
> + pshufd xmm3,xmm0,78
> + pxor xmm3,xmm0
> +DB 102,15,58,68,194,0
> +DB 102,15,58,68,202,17
> +DB 102,15,58,68,220,0
> + pxor xmm3,xmm0
> + pxor xmm3,xmm1
> +
> + movdqa xmm4,xmm3
> + psrldq xmm3,8
> + pslldq xmm4,8
> + pxor xmm1,xmm3
> + pxor xmm0,xmm4
> +
> + movdqa xmm4,xmm0
> + movdqa xmm3,xmm0
> + psllq xmm0,5
> + pxor xmm3,xmm0
> + psllq xmm0,1
> + pxor xmm0,xmm3
> + psllq xmm0,57
> + movdqa xmm3,xmm0
> + pslldq xmm0,8
> + psrldq xmm3,8
> + pxor xmm0,xmm4
> + pxor xmm1,xmm3
> +
> +
> + movdqa xmm4,xmm0
> + psrlq xmm0,1
> + pxor xmm1,xmm4
> + pxor xmm4,xmm0
> + psrlq xmm0,5
> + pxor xmm0,xmm4
> + psrlq xmm0,1
> + pxor xmm0,xmm1
> +DB 102,15,56,0,197
> + movdqu XMMWORD[rcx],xmm0
> + DB 0F3h,0C3h ;repret
> +
> +
> +global gcm_ghash_clmul
> +
> +ALIGN 32
> +gcm_ghash_clmul:
> +
> +$L$_ghash_clmul:
> + lea rax,[((-136))+rsp]
> +$L$SEH_begin_gcm_ghash_clmul:
> +
> +DB 0x48,0x8d,0x60,0xe0
> +DB 0x0f,0x29,0x70,0xe0
> +DB 0x0f,0x29,0x78,0xf0
> +DB 0x44,0x0f,0x29,0x00
> +DB 0x44,0x0f,0x29,0x48,0x10
> +DB 0x44,0x0f,0x29,0x50,0x20
> +DB 0x44,0x0f,0x29,0x58,0x30
> +DB 0x44,0x0f,0x29,0x60,0x40
> +DB 0x44,0x0f,0x29,0x68,0x50
> +DB 0x44,0x0f,0x29,0x70,0x60
> +DB 0x44,0x0f,0x29,0x78,0x70
> + movdqa xmm10,XMMWORD[$L$bswap_mask]
> +
> + movdqu xmm0,XMMWORD[rcx]
> + movdqu xmm2,XMMWORD[rdx]
> + movdqu xmm7,XMMWORD[32+rdx]
> +DB 102,65,15,56,0,194
> +
> + sub r9,0x10
> + jz NEAR $L$odd_tail
> +
> + movdqu xmm6,XMMWORD[16+rdx]
> + mov eax,DWORD[((OPENSSL_ia32cap_P+4))]
> + cmp r9,0x30
> + jb NEAR $L$skip4x
> +
> + and eax,71303168
> + cmp eax,4194304
> + je NEAR $L$skip4x
> +
> + sub r9,0x30
> + mov rax,0xA040608020C0E000
> + movdqu xmm14,XMMWORD[48+rdx]
> + movdqu xmm15,XMMWORD[64+rdx]
> +
> +
> +
> +
> + movdqu xmm3,XMMWORD[48+r8]
> + movdqu xmm11,XMMWORD[32+r8]
> +DB 102,65,15,56,0,218
> +DB 102,69,15,56,0,218
> + movdqa xmm5,xmm3
> + pshufd xmm4,xmm3,78
> + pxor xmm4,xmm3
> +DB 102,15,58,68,218,0
> +DB 102,15,58,68,234,17
> +DB 102,15,58,68,231,0
> +
> + movdqa xmm13,xmm11
> + pshufd xmm12,xmm11,78
> + pxor xmm12,xmm11
> +DB 102,68,15,58,68,222,0
> +DB 102,68,15,58,68,238,17
> +DB 102,68,15,58,68,231,16
> + xorps xmm3,xmm11
> + xorps xmm5,xmm13
> + movups xmm7,XMMWORD[80+rdx]
> + xorps xmm4,xmm12
> +
> + movdqu xmm11,XMMWORD[16+r8]
> + movdqu xmm8,XMMWORD[r8]
> +DB 102,69,15,56,0,218
> +DB 102,69,15,56,0,194
> + movdqa xmm13,xmm11
> + pshufd xmm12,xmm11,78
> + pxor xmm0,xmm8
> + pxor xmm12,xmm11
> +DB 102,69,15,58,68,222,0
> + movdqa xmm1,xmm0
> + pshufd xmm8,xmm0,78
> + pxor xmm8,xmm0
> +DB 102,69,15,58,68,238,17
> +DB 102,68,15,58,68,231,0
> + xorps xmm3,xmm11
> + xorps xmm5,xmm13
> +
> + lea r8,[64+r8]
> + sub r9,0x40
> + jc NEAR $L$tail4x
> +
> + jmp NEAR $L$mod4_loop
> +ALIGN 32
> +$L$mod4_loop:
> +DB 102,65,15,58,68,199,0
> + xorps xmm4,xmm12
> + movdqu xmm11,XMMWORD[48+r8]
> +DB 102,69,15,56,0,218
> +DB 102,65,15,58,68,207,17
> + xorps xmm0,xmm3
> + movdqu xmm3,XMMWORD[32+r8]
> + movdqa xmm13,xmm11
> +DB 102,68,15,58,68,199,16
> + pshufd xmm12,xmm11,78
> + xorps xmm1,xmm5
> + pxor xmm12,xmm11
> +DB 102,65,15,56,0,218
> + movups xmm7,XMMWORD[32+rdx]
> + xorps xmm8,xmm4
> +DB 102,68,15,58,68,218,0
> + pshufd xmm4,xmm3,78
> +
> + pxor xmm8,xmm0
> + movdqa xmm5,xmm3
> + pxor xmm8,xmm1
> + pxor xmm4,xmm3
> + movdqa xmm9,xmm8
> +DB 102,68,15,58,68,234,17
> + pslldq xmm8,8
> + psrldq xmm9,8
> + pxor xmm0,xmm8
> + movdqa xmm8,XMMWORD[$L$7_mask]
> + pxor xmm1,xmm9
> +DB 102,76,15,110,200
> +
> + pand xmm8,xmm0
> +DB 102,69,15,56,0,200
> + pxor xmm9,xmm0
> +DB 102,68,15,58,68,231,0
> + psllq xmm9,57
> + movdqa xmm8,xmm9
> + pslldq xmm9,8
> +DB 102,15,58,68,222,0
> + psrldq xmm8,8
> + pxor xmm0,xmm9
> + pxor xmm1,xmm8
> + movdqu xmm8,XMMWORD[r8]
> +
> + movdqa xmm9,xmm0
> + psrlq xmm0,1
> +DB 102,15,58,68,238,17
> + xorps xmm3,xmm11
> + movdqu xmm11,XMMWORD[16+r8]
> +DB 102,69,15,56,0,218
> +DB 102,15,58,68,231,16
> + xorps xmm5,xmm13
> + movups xmm7,XMMWORD[80+rdx]
> +DB 102,69,15,56,0,194
> + pxor xmm1,xmm9
> + pxor xmm9,xmm0
> + psrlq xmm0,5
> +
> + movdqa xmm13,xmm11
> + pxor xmm4,xmm12
> + pshufd xmm12,xmm11,78
> + pxor xmm0,xmm9
> + pxor xmm1,xmm8
> + pxor xmm12,xmm11
> +DB 102,69,15,58,68,222,0
> + psrlq xmm0,1
> + pxor xmm0,xmm1
> + movdqa xmm1,xmm0
> +DB 102,69,15,58,68,238,17
> + xorps xmm3,xmm11
> + pshufd xmm8,xmm0,78
> + pxor xmm8,xmm0
> +
> +DB 102,68,15,58,68,231,0
> + xorps xmm5,xmm13
> +
> + lea r8,[64+r8]
> + sub r9,0x40
> + jnc NEAR $L$mod4_loop
> +
> +$L$tail4x:
> +DB 102,65,15,58,68,199,0
> +DB 102,65,15,58,68,207,17
> +DB 102,68,15,58,68,199,16
> + xorps xmm4,xmm12
> + xorps xmm0,xmm3
> + xorps xmm1,xmm5
> + pxor xmm1,xmm0
> + pxor xmm8,xmm4
> +
> + pxor xmm8,xmm1
> + pxor xmm1,xmm0
> +
> + movdqa xmm9,xmm8
> + psrldq xmm8,8
> + pslldq xmm9,8
> + pxor xmm1,xmm8
> + pxor xmm0,xmm9
> +
> + movdqa xmm4,xmm0
> + movdqa xmm3,xmm0
> + psllq xmm0,5
> + pxor xmm3,xmm0
> + psllq xmm0,1
> + pxor xmm0,xmm3
> + psllq xmm0,57
> + movdqa xmm3,xmm0
> + pslldq xmm0,8
> + psrldq xmm3,8
> + pxor xmm0,xmm4
> + pxor xmm1,xmm3
> +
> +
> + movdqa xmm4,xmm0
> + psrlq xmm0,1
> + pxor xmm1,xmm4
> + pxor xmm4,xmm0
> + psrlq xmm0,5
> + pxor xmm0,xmm4
> + psrlq xmm0,1
> + pxor xmm0,xmm1
> + add r9,0x40
> + jz NEAR $L$done
> + movdqu xmm7,XMMWORD[32+rdx]
> + sub r9,0x10
> + jz NEAR $L$odd_tail
> +$L$skip4x:
> +
> +
> +
> +
> +
> + movdqu xmm8,XMMWORD[r8]
> + movdqu xmm3,XMMWORD[16+r8]
> +DB 102,69,15,56,0,194
> +DB 102,65,15,56,0,218
> + pxor xmm0,xmm8
> +
> + movdqa xmm5,xmm3
> + pshufd xmm4,xmm3,78
> + pxor xmm4,xmm3
> +DB 102,15,58,68,218,0
> +DB 102,15,58,68,234,17
> +DB 102,15,58,68,231,0
> +
> + lea r8,[32+r8]
> + nop
> + sub r9,0x20
> + jbe NEAR $L$even_tail
> + nop
> + jmp NEAR $L$mod_loop
> +
> +ALIGN 32
> +$L$mod_loop:
> + movdqa xmm1,xmm0
> + movdqa xmm8,xmm4
> + pshufd xmm4,xmm0,78
> + pxor xmm4,xmm0
> +
> +DB 102,15,58,68,198,0
> +DB 102,15,58,68,206,17
> +DB 102,15,58,68,231,16
> +
> + pxor xmm0,xmm3
> + pxor xmm1,xmm5
> + movdqu xmm9,XMMWORD[r8]
> + pxor xmm8,xmm0
> +DB 102,69,15,56,0,202
> + movdqu xmm3,XMMWORD[16+r8]
> +
> + pxor xmm8,xmm1
> + pxor xmm1,xmm9
> + pxor xmm4,xmm8
> +DB 102,65,15,56,0,218
> + movdqa xmm8,xmm4
> + psrldq xmm8,8
> + pslldq xmm4,8
> + pxor xmm1,xmm8
> + pxor xmm0,xmm4
> +
> + movdqa xmm5,xmm3
> +
> + movdqa xmm9,xmm0
> + movdqa xmm8,xmm0
> + psllq xmm0,5
> + pxor xmm8,xmm0
> +DB 102,15,58,68,218,0
> + psllq xmm0,1
> + pxor xmm0,xmm8
> + psllq xmm0,57
> + movdqa xmm8,xmm0
> + pslldq xmm0,8
> + psrldq xmm8,8
> + pxor xmm0,xmm9
> + pshufd xmm4,xmm5,78
> + pxor xmm1,xmm8
> + pxor xmm4,xmm5
> +
> + movdqa xmm9,xmm0
> + psrlq xmm0,1
> +DB 102,15,58,68,234,17
> + pxor xmm1,xmm9
> + pxor xmm9,xmm0
> + psrlq xmm0,5
> + pxor xmm0,xmm9
> + lea r8,[32+r8]
> + psrlq xmm0,1
> +DB 102,15,58,68,231,0
> + pxor xmm0,xmm1
> +
> + sub r9,0x20
> + ja NEAR $L$mod_loop
> +
> +$L$even_tail:
> + movdqa xmm1,xmm0
> + movdqa xmm8,xmm4
> + pshufd xmm4,xmm0,78
> + pxor xmm4,xmm0
> +
> +DB 102,15,58,68,198,0
> +DB 102,15,58,68,206,17
> +DB 102,15,58,68,231,16
> +
> + pxor xmm0,xmm3
> + pxor xmm1,xmm5
> + pxor xmm8,xmm0
> + pxor xmm8,xmm1
> + pxor xmm4,xmm8
> + movdqa xmm8,xmm4
> + psrldq xmm8,8
> + pslldq xmm4,8
> + pxor xmm1,xmm8
> + pxor xmm0,xmm4
> +
> + movdqa xmm4,xmm0
> + movdqa xmm3,xmm0
> + psllq xmm0,5
> + pxor xmm3,xmm0
> + psllq xmm0,1
> + pxor xmm0,xmm3
> + psllq xmm0,57
> + movdqa xmm3,xmm0
> + pslldq xmm0,8
> + psrldq xmm3,8
> + pxor xmm0,xmm4
> + pxor xmm1,xmm3
> +
> +
> + movdqa xmm4,xmm0
> + psrlq xmm0,1
> + pxor xmm1,xmm4
> + pxor xmm4,xmm0
> + psrlq xmm0,5
> + pxor xmm0,xmm4
> + psrlq xmm0,1
> + pxor xmm0,xmm1
> + test r9,r9
> + jnz NEAR $L$done
> +
> +$L$odd_tail:
> + movdqu xmm8,XMMWORD[r8]
> +DB 102,69,15,56,0,194
> + pxor xmm0,xmm8
> + movdqa xmm1,xmm0
> + pshufd xmm3,xmm0,78
> + pxor xmm3,xmm0
> +DB 102,15,58,68,194,0
> +DB 102,15,58,68,202,17
> +DB 102,15,58,68,223,0
> + pxor xmm3,xmm0
> + pxor xmm3,xmm1
> +
> + movdqa xmm4,xmm3
> + psrldq xmm3,8
> + pslldq xmm4,8
> + pxor xmm1,xmm3
> + pxor xmm0,xmm4
> +
> + movdqa xmm4,xmm0
> + movdqa xmm3,xmm0
> + psllq xmm0,5
> + pxor xmm3,xmm0
> + psllq xmm0,1
> + pxor xmm0,xmm3
> + psllq xmm0,57
> + movdqa xmm3,xmm0
> + pslldq xmm0,8
> + psrldq xmm3,8
> + pxor xmm0,xmm4
> + pxor xmm1,xmm3
> +
> +
> + movdqa xmm4,xmm0
> + psrlq xmm0,1
> + pxor xmm1,xmm4
> + pxor xmm4,xmm0
> + psrlq xmm0,5
> + pxor xmm0,xmm4
> + psrlq xmm0,1
> + pxor xmm0,xmm1
> +$L$done:
> +DB 102,65,15,56,0,194
> + movdqu XMMWORD[rcx],xmm0
> + movaps xmm6,XMMWORD[rsp]
> + movaps xmm7,XMMWORD[16+rsp]
> + movaps xmm8,XMMWORD[32+rsp]
> + movaps xmm9,XMMWORD[48+rsp]
> + movaps xmm10,XMMWORD[64+rsp]
> + movaps xmm11,XMMWORD[80+rsp]
> + movaps xmm12,XMMWORD[96+rsp]
> + movaps xmm13,XMMWORD[112+rsp]
> + movaps xmm14,XMMWORD[128+rsp]
> + movaps xmm15,XMMWORD[144+rsp]
> + lea rsp,[168+rsp]
> +$L$SEH_end_gcm_ghash_clmul:
> + DB 0F3h,0C3h ;repret
> +
> +
> +global gcm_init_avx
> +
> +ALIGN 32
> +gcm_init_avx:
> +
> + jmp NEAR $L$_init_clmul
> +
> +
> +global gcm_gmult_avx
> +
> +ALIGN 32
> +gcm_gmult_avx:
> +
> + jmp NEAR $L$_gmult_clmul
> +
> +
> +global gcm_ghash_avx
> +
> +ALIGN 32
> +gcm_ghash_avx:
> +
> + jmp NEAR $L$_ghash_clmul
> +
> +
> +ALIGN 64
> +$L$bswap_mask:
> +DB 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
> +$L$0x1c2_polynomial:
> +DB 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xc2
> +$L$7_mask:
> + DD 7,0,7,0
> +$L$7_mask_poly:
> + DD 7,0,450,0
> +ALIGN 64
> +
> +$L$rem_4bit:
> + DD 0,0,0,471859200,0,943718400,0,610271232
> + DD 0,1887436800,0,1822425088,0,1220542464,0,1423966208
> + DD 0,3774873600,0,4246732800,0,3644850176,0,3311403008
> + DD 0,2441084928,0,2376073216,0,2847932416,0,3051356160
> +
> +$L$rem_8bit:
> + DW 0x0000,0x01C2,0x0384,0x0246,0x0708,0x06CA,0x048C,0x054E
> + DW 0x0E10,0x0FD2,0x0D94,0x0C56,0x0918,0x08DA,0x0A9C,0x0B5E
> + DW 0x1C20,0x1DE2,0x1FA4,0x1E66,0x1B28,0x1AEA,0x18AC,0x196E
> + DW 0x1230,0x13F2,0x11B4,0x1076,0x1538,0x14FA,0x16BC,0x177E
> + DW 0x3840,0x3982,0x3BC4,0x3A06,0x3F48,0x3E8A,0x3CCC,0x3D0E
> + DW 0x3650,0x3792,0x35D4,0x3416,0x3158,0x309A,0x32DC,0x331E
> + DW 0x2460,0x25A2,0x27E4,0x2626,0x2368,0x22AA,0x20EC,0x212E
> + DW 0x2A70,0x2BB2,0x29F4,0x2836,0x2D78,0x2CBA,0x2EFC,0x2F3E
> + DW 0x7080,0x7142,0x7304,0x72C6,0x7788,0x764A,0x740C,0x75CE
> + DW 0x7E90,0x7F52,0x7D14,0x7CD6,0x7998,0x785A,0x7A1C,0x7BDE
> + DW 0x6CA0,0x6D62,0x6F24,0x6EE6,0x6BA8,0x6A6A,0x682C,0x69EE
> + DW 0x62B0,0x6372,0x6134,0x60F6,0x65B8,0x647A,0x663C,0x67FE
> + DW 0x48C0,0x4902,0x4B44,0x4A86,0x4FC8,0x4E0A,0x4C4C,0x4D8E
> + DW 0x46D0,0x4712,0x4554,0x4496,0x41D8,0x401A,0x425C,0x439E
> + DW 0x54E0,0x5522,0x5764,0x56A6,0x53E8,0x522A,0x506C,0x51AE
> + DW 0x5AF0,0x5B32,0x5974,0x58B6,0x5DF8,0x5C3A,0x5E7C,0x5FBE
> + DW 0xE100,0xE0C2,0xE284,0xE346,0xE608,0xE7CA,0xE58C,0xE44E
> + DW 0xEF10,0xEED2,0xEC94,0xED56,0xE818,0xE9DA,0xEB9C,0xEA5E
> + DW 0xFD20,0xFCE2,0xFEA4,0xFF66,0xFA28,0xFBEA,0xF9AC,0xF86E
> + DW 0xF330,0xF2F2,0xF0B4,0xF176,0xF438,0xF5FA,0xF7BC,0xF67E
> + DW 0xD940,0xD882,0xDAC4,0xDB06,0xDE48,0xDF8A,0xDDCC,0xDC0E
> + DW 0xD750,0xD692,0xD4D4,0xD516,0xD058,0xD19A,0xD3DC,0xD21E
> + DW 0xC560,0xC4A2,0xC6E4,0xC726,0xC268,0xC3AA,0xC1EC,0xC02E
> + DW 0xCB70,0xCAB2,0xC8F4,0xC936,0xCC78,0xCDBA,0xCFFC,0xCE3E
> + DW 0x9180,0x9042,0x9204,0x93C6,0x9688,0x974A,0x950C,0x94CE
> + DW 0x9F90,0x9E52,0x9C14,0x9DD6,0x9898,0x995A,0x9B1C,0x9ADE
> + DW 0x8DA0,0x8C62,0x8E24,0x8FE6,0x8AA8,0x8B6A,0x892C,0x88EE
> + DW 0x83B0,0x8272,0x8034,0x81F6,0x84B8,0x857A,0x873C,0x86FE
> + DW 0xA9C0,0xA802,0xAA44,0xAB86,0xAEC8,0xAF0A,0xAD4C,0xAC8E
> + DW 0xA7D0,0xA612,0xA454,0xA596,0xA0D8,0xA11A,0xA35C,0xA29E
> + DW 0xB5E0,0xB422,0xB664,0xB7A6,0xB2E8,0xB32A,0xB16C,0xB0AE
> + DW 0xBBF0,0xBA32,0xB874,0xB9B6,0xBCF8,0xBD3A,0xBF7C,0xBEBE
> +
> +DB 71,72,65,83,72,32,102,111,114,32,120,56,54,95,54,52
> +DB 44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32
> +DB 60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111
> +DB 114,103,62,0
> +ALIGN 64
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$in_prologue
> +
> + lea rax,[((48+280))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov r12,QWORD[((-24))+rax]
> + mov r13,QWORD[((-32))+rax]
> + mov r14,QWORD[((-40))+rax]
> + mov r15,QWORD[((-48))+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> + mov QWORD[216+r8],r12
> + mov QWORD[224+r8],r13
> + mov QWORD[232+r8],r14
> + mov QWORD[240+r8],r15
> +
> +$L$in_prologue:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_gcm_gmult_4bit wrt ..imagebase
> + DD $L$SEH_end_gcm_gmult_4bit wrt ..imagebase
> + DD $L$SEH_info_gcm_gmult_4bit wrt ..imagebase
> +
> + DD $L$SEH_begin_gcm_ghash_4bit wrt ..imagebase
> + DD $L$SEH_end_gcm_ghash_4bit wrt ..imagebase
> + DD $L$SEH_info_gcm_ghash_4bit wrt ..imagebase
> +
> + DD $L$SEH_begin_gcm_init_clmul wrt ..imagebase
> + DD $L$SEH_end_gcm_init_clmul wrt ..imagebase
> + DD $L$SEH_info_gcm_init_clmul wrt ..imagebase
> +
> + DD $L$SEH_begin_gcm_ghash_clmul wrt ..imagebase
> + DD $L$SEH_end_gcm_ghash_clmul wrt ..imagebase
> + DD $L$SEH_info_gcm_ghash_clmul wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_gcm_gmult_4bit:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$gmult_prologue wrt ..imagebase,$L$gmult_epilogue
> wrt ..imagebase
> +$L$SEH_info_gcm_ghash_4bit:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$ghash_prologue wrt ..imagebase,$L$ghash_epilogue
> wrt ..imagebase
> +$L$SEH_info_gcm_init_clmul:
> +DB 0x01,0x08,0x03,0x00
> +DB 0x08,0x68,0x00,0x00
> +DB 0x04,0x22,0x00,0x00
> +$L$SEH_info_gcm_ghash_clmul:
> +DB 0x01,0x33,0x16,0x00
> +DB 0x33,0xf8,0x09,0x00
> +DB 0x2e,0xe8,0x08,0x00
> +DB 0x29,0xd8,0x07,0x00
> +DB 0x24,0xc8,0x06,0x00
> +DB 0x1f,0xb8,0x05,0x00
> +DB 0x1a,0xa8,0x04,0x00
> +DB 0x15,0x98,0x03,0x00
> +DB 0x10,0x88,0x02,0x00
> +DB 0x0c,0x78,0x01,0x00
> +DB 0x08,0x68,0x00,0x00
> +DB 0x04,0x01,0x15,0x00
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha1-mb-
> x86_64.nasm b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha1-mb-
> x86_64.nasm
> new file mode 100644
> index 0000000000..f3b7b0e35e
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha1-mb-x86_64.nasm
> @@ -0,0 +1,3137 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/sha/asm/sha1-mb-x86_64.pl
> +;
> +; Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +
> +EXTERN OPENSSL_ia32cap_P
> +
> +global sha1_multi_block
> +
> +ALIGN 32
> +sha1_multi_block:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha1_multi_block:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + mov rcx,QWORD[((OPENSSL_ia32cap_P+4))]
> + bt rcx,61
> + jc NEAR _shaext_shortcut
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + lea rsp,[((-168))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> + movaps XMMWORD[(-120)+rax],xmm10
> + movaps XMMWORD[(-104)+rax],xmm11
> + movaps XMMWORD[(-88)+rax],xmm12
> + movaps XMMWORD[(-72)+rax],xmm13
> + movaps XMMWORD[(-56)+rax],xmm14
> + movaps XMMWORD[(-40)+rax],xmm15
> + sub rsp,288
> + and rsp,-256
> + mov QWORD[272+rsp],rax
> +
> +$L$body:
> + lea rbp,[K_XX_XX]
> + lea rbx,[256+rsp]
> +
> +$L$oop_grande:
> + mov DWORD[280+rsp],edx
> + xor edx,edx
> + mov r8,QWORD[rsi]
> + mov ecx,DWORD[8+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[rbx],ecx
> + cmovle r8,rbp
> + mov r9,QWORD[16+rsi]
> + mov ecx,DWORD[24+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[4+rbx],ecx
> + cmovle r9,rbp
> + mov r10,QWORD[32+rsi]
> + mov ecx,DWORD[40+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[8+rbx],ecx
> + cmovle r10,rbp
> + mov r11,QWORD[48+rsi]
> + mov ecx,DWORD[56+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[12+rbx],ecx
> + cmovle r11,rbp
> + test edx,edx
> + jz NEAR $L$done
> +
> + movdqu xmm10,XMMWORD[rdi]
> + lea rax,[128+rsp]
> + movdqu xmm11,XMMWORD[32+rdi]
> + movdqu xmm12,XMMWORD[64+rdi]
> + movdqu xmm13,XMMWORD[96+rdi]
> + movdqu xmm14,XMMWORD[128+rdi]
> + movdqa xmm5,XMMWORD[96+rbp]
> + movdqa xmm15,XMMWORD[((-32))+rbp]
> + jmp NEAR $L$oop
> +
> +ALIGN 32
> +$L$oop:
> + movd xmm0,DWORD[r8]
> + lea r8,[64+r8]
> + movd xmm2,DWORD[r9]
> + lea r9,[64+r9]
> + movd xmm3,DWORD[r10]
> + lea r10,[64+r10]
> + movd xmm4,DWORD[r11]
> + lea r11,[64+r11]
> + punpckldq xmm0,xmm3
> + movd xmm1,DWORD[((-60))+r8]
> + punpckldq xmm2,xmm4
> + movd xmm9,DWORD[((-60))+r9]
> + punpckldq xmm0,xmm2
> + movd xmm8,DWORD[((-60))+r10]
> +DB 102,15,56,0,197
> + movd xmm7,DWORD[((-60))+r11]
> + punpckldq xmm1,xmm8
> + movdqa xmm8,xmm10
> + paddd xmm14,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm11
> + movdqa xmm6,xmm11
> + pslld xmm8,5
> + pandn xmm7,xmm13
> + pand xmm6,xmm12
> + punpckldq xmm1,xmm9
> + movdqa xmm9,xmm10
> +
> + movdqa XMMWORD[(0-128)+rax],xmm0
> + paddd xmm14,xmm0
> + movd xmm2,DWORD[((-56))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm11
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-56))+r9]
> + pslld xmm7,30
> + paddd xmm14,xmm6
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> +DB 102,15,56,0,205
> + movd xmm8,DWORD[((-56))+r10]
> + por xmm11,xmm7
> + movd xmm7,DWORD[((-56))+r11]
> + punpckldq xmm2,xmm8
> + movdqa xmm8,xmm14
> + paddd xmm13,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm10
> + movdqa xmm6,xmm10
> + pslld xmm8,5
> + pandn xmm7,xmm12
> + pand xmm6,xmm11
> + punpckldq xmm2,xmm9
> + movdqa xmm9,xmm14
> +
> + movdqa XMMWORD[(16-128)+rax],xmm1
> + paddd xmm13,xmm1
> + movd xmm3,DWORD[((-52))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm10
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-52))+r9]
> + pslld xmm7,30
> + paddd xmm13,xmm6
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> +DB 102,15,56,0,213
> + movd xmm8,DWORD[((-52))+r10]
> + por xmm10,xmm7
> + movd xmm7,DWORD[((-52))+r11]
> + punpckldq xmm3,xmm8
> + movdqa xmm8,xmm13
> + paddd xmm12,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm14
> + movdqa xmm6,xmm14
> + pslld xmm8,5
> + pandn xmm7,xmm11
> + pand xmm6,xmm10
> + punpckldq xmm3,xmm9
> + movdqa xmm9,xmm13
> +
> + movdqa XMMWORD[(32-128)+rax],xmm2
> + paddd xmm12,xmm2
> + movd xmm4,DWORD[((-48))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm14
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-48))+r9]
> + pslld xmm7,30
> + paddd xmm12,xmm6
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> +DB 102,15,56,0,221
> + movd xmm8,DWORD[((-48))+r10]
> + por xmm14,xmm7
> + movd xmm7,DWORD[((-48))+r11]
> + punpckldq xmm4,xmm8
> + movdqa xmm8,xmm12
> + paddd xmm11,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm13
> + movdqa xmm6,xmm13
> + pslld xmm8,5
> + pandn xmm7,xmm10
> + pand xmm6,xmm14
> + punpckldq xmm4,xmm9
> + movdqa xmm9,xmm12
> +
> + movdqa XMMWORD[(48-128)+rax],xmm3
> + paddd xmm11,xmm3
> + movd xmm0,DWORD[((-44))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm13
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-44))+r9]
> + pslld xmm7,30
> + paddd xmm11,xmm6
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> +DB 102,15,56,0,229
> + movd xmm8,DWORD[((-44))+r10]
> + por xmm13,xmm7
> + movd xmm7,DWORD[((-44))+r11]
> + punpckldq xmm0,xmm8
> + movdqa xmm8,xmm11
> + paddd xmm10,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm12
> + movdqa xmm6,xmm12
> + pslld xmm8,5
> + pandn xmm7,xmm14
> + pand xmm6,xmm13
> + punpckldq xmm0,xmm9
> + movdqa xmm9,xmm11
> +
> + movdqa XMMWORD[(64-128)+rax],xmm4
> + paddd xmm10,xmm4
> + movd xmm1,DWORD[((-40))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm12
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-40))+r9]
> + pslld xmm7,30
> + paddd xmm10,xmm6
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> +DB 102,15,56,0,197
> + movd xmm8,DWORD[((-40))+r10]
> + por xmm12,xmm7
> + movd xmm7,DWORD[((-40))+r11]
> + punpckldq xmm1,xmm8
> + movdqa xmm8,xmm10
> + paddd xmm14,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm11
> + movdqa xmm6,xmm11
> + pslld xmm8,5
> + pandn xmm7,xmm13
> + pand xmm6,xmm12
> + punpckldq xmm1,xmm9
> + movdqa xmm9,xmm10
> +
> + movdqa XMMWORD[(80-128)+rax],xmm0
> + paddd xmm14,xmm0
> + movd xmm2,DWORD[((-36))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm11
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-36))+r9]
> + pslld xmm7,30
> + paddd xmm14,xmm6
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> +DB 102,15,56,0,205
> + movd xmm8,DWORD[((-36))+r10]
> + por xmm11,xmm7
> + movd xmm7,DWORD[((-36))+r11]
> + punpckldq xmm2,xmm8
> + movdqa xmm8,xmm14
> + paddd xmm13,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm10
> + movdqa xmm6,xmm10
> + pslld xmm8,5
> + pandn xmm7,xmm12
> + pand xmm6,xmm11
> + punpckldq xmm2,xmm9
> + movdqa xmm9,xmm14
> +
> + movdqa XMMWORD[(96-128)+rax],xmm1
> + paddd xmm13,xmm1
> + movd xmm3,DWORD[((-32))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm10
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-32))+r9]
> + pslld xmm7,30
> + paddd xmm13,xmm6
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> +DB 102,15,56,0,213
> + movd xmm8,DWORD[((-32))+r10]
> + por xmm10,xmm7
> + movd xmm7,DWORD[((-32))+r11]
> + punpckldq xmm3,xmm8
> + movdqa xmm8,xmm13
> + paddd xmm12,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm14
> + movdqa xmm6,xmm14
> + pslld xmm8,5
> + pandn xmm7,xmm11
> + pand xmm6,xmm10
> + punpckldq xmm3,xmm9
> + movdqa xmm9,xmm13
> +
> + movdqa XMMWORD[(112-128)+rax],xmm2
> + paddd xmm12,xmm2
> + movd xmm4,DWORD[((-28))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm14
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-28))+r9]
> + pslld xmm7,30
> + paddd xmm12,xmm6
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> +DB 102,15,56,0,221
> + movd xmm8,DWORD[((-28))+r10]
> + por xmm14,xmm7
> + movd xmm7,DWORD[((-28))+r11]
> + punpckldq xmm4,xmm8
> + movdqa xmm8,xmm12
> + paddd xmm11,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm13
> + movdqa xmm6,xmm13
> + pslld xmm8,5
> + pandn xmm7,xmm10
> + pand xmm6,xmm14
> + punpckldq xmm4,xmm9
> + movdqa xmm9,xmm12
> +
> + movdqa XMMWORD[(128-128)+rax],xmm3
> + paddd xmm11,xmm3
> + movd xmm0,DWORD[((-24))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm13
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-24))+r9]
> + pslld xmm7,30
> + paddd xmm11,xmm6
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> +DB 102,15,56,0,229
> + movd xmm8,DWORD[((-24))+r10]
> + por xmm13,xmm7
> + movd xmm7,DWORD[((-24))+r11]
> + punpckldq xmm0,xmm8
> + movdqa xmm8,xmm11
> + paddd xmm10,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm12
> + movdqa xmm6,xmm12
> + pslld xmm8,5
> + pandn xmm7,xmm14
> + pand xmm6,xmm13
> + punpckldq xmm0,xmm9
> + movdqa xmm9,xmm11
> +
> + movdqa XMMWORD[(144-128)+rax],xmm4
> + paddd xmm10,xmm4
> + movd xmm1,DWORD[((-20))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm12
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-20))+r9]
> + pslld xmm7,30
> + paddd xmm10,xmm6
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> +DB 102,15,56,0,197
> + movd xmm8,DWORD[((-20))+r10]
> + por xmm12,xmm7
> + movd xmm7,DWORD[((-20))+r11]
> + punpckldq xmm1,xmm8
> + movdqa xmm8,xmm10
> + paddd xmm14,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm11
> + movdqa xmm6,xmm11
> + pslld xmm8,5
> + pandn xmm7,xmm13
> + pand xmm6,xmm12
> + punpckldq xmm1,xmm9
> + movdqa xmm9,xmm10
> +
> + movdqa XMMWORD[(160-128)+rax],xmm0
> + paddd xmm14,xmm0
> + movd xmm2,DWORD[((-16))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm11
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-16))+r9]
> + pslld xmm7,30
> + paddd xmm14,xmm6
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> +DB 102,15,56,0,205
> + movd xmm8,DWORD[((-16))+r10]
> + por xmm11,xmm7
> + movd xmm7,DWORD[((-16))+r11]
> + punpckldq xmm2,xmm8
> + movdqa xmm8,xmm14
> + paddd xmm13,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm10
> + movdqa xmm6,xmm10
> + pslld xmm8,5
> + pandn xmm7,xmm12
> + pand xmm6,xmm11
> + punpckldq xmm2,xmm9
> + movdqa xmm9,xmm14
> +
> + movdqa XMMWORD[(176-128)+rax],xmm1
> + paddd xmm13,xmm1
> + movd xmm3,DWORD[((-12))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm10
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-12))+r9]
> + pslld xmm7,30
> + paddd xmm13,xmm6
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> +DB 102,15,56,0,213
> + movd xmm8,DWORD[((-12))+r10]
> + por xmm10,xmm7
> + movd xmm7,DWORD[((-12))+r11]
> + punpckldq xmm3,xmm8
> + movdqa xmm8,xmm13
> + paddd xmm12,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm14
> + movdqa xmm6,xmm14
> + pslld xmm8,5
> + pandn xmm7,xmm11
> + pand xmm6,xmm10
> + punpckldq xmm3,xmm9
> + movdqa xmm9,xmm13
> +
> + movdqa XMMWORD[(192-128)+rax],xmm2
> + paddd xmm12,xmm2
> + movd xmm4,DWORD[((-8))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm14
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-8))+r9]
> + pslld xmm7,30
> + paddd xmm12,xmm6
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> +DB 102,15,56,0,221
> + movd xmm8,DWORD[((-8))+r10]
> + por xmm14,xmm7
> + movd xmm7,DWORD[((-8))+r11]
> + punpckldq xmm4,xmm8
> + movdqa xmm8,xmm12
> + paddd xmm11,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm13
> + movdqa xmm6,xmm13
> + pslld xmm8,5
> + pandn xmm7,xmm10
> + pand xmm6,xmm14
> + punpckldq xmm4,xmm9
> + movdqa xmm9,xmm12
> +
> + movdqa XMMWORD[(208-128)+rax],xmm3
> + paddd xmm11,xmm3
> + movd xmm0,DWORD[((-4))+r8]
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm13
> +
> + por xmm8,xmm9
> + movd xmm9,DWORD[((-4))+r9]
> + pslld xmm7,30
> + paddd xmm11,xmm6
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> +DB 102,15,56,0,229
> + movd xmm8,DWORD[((-4))+r10]
> + por xmm13,xmm7
> + movdqa xmm1,XMMWORD[((0-128))+rax]
> + movd xmm7,DWORD[((-4))+r11]
> + punpckldq xmm0,xmm8
> + movdqa xmm8,xmm11
> + paddd xmm10,xmm15
> + punpckldq xmm9,xmm7
> + movdqa xmm7,xmm12
> + movdqa xmm6,xmm12
> + pslld xmm8,5
> + prefetcht0 [63+r8]
> + pandn xmm7,xmm14
> + pand xmm6,xmm13
> + punpckldq xmm0,xmm9
> + movdqa xmm9,xmm11
> +
> + movdqa XMMWORD[(224-128)+rax],xmm4
> + paddd xmm10,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm7
> + movdqa xmm7,xmm12
> + prefetcht0 [63+r9]
> +
> + por xmm8,xmm9
> + pslld xmm7,30
> + paddd xmm10,xmm6
> + prefetcht0 [63+r10]
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> +DB 102,15,56,0,197
> + prefetcht0 [63+r11]
> + por xmm12,xmm7
> + movdqa xmm2,XMMWORD[((16-128))+rax]
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((32-128))+rax]
> +
> + movdqa xmm8,xmm10
> + pxor xmm1,XMMWORD[((128-128))+rax]
> + paddd xmm14,xmm15
> + movdqa xmm7,xmm11
> + pslld xmm8,5
> + pxor xmm1,xmm3
> + movdqa xmm6,xmm11
> + pandn xmm7,xmm13
> + movdqa xmm5,xmm1
> + pand xmm6,xmm12
> + movdqa xmm9,xmm10
> + psrld xmm5,31
> + paddd xmm1,xmm1
> +
> + movdqa XMMWORD[(240-128)+rax],xmm0
> + paddd xmm14,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm7
> +
> + movdqa xmm7,xmm11
> + por xmm8,xmm9
> + pslld xmm7,30
> + paddd xmm14,xmm6
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((48-128))+rax]
> +
> + movdqa xmm8,xmm14
> + pxor xmm2,XMMWORD[((144-128))+rax]
> + paddd xmm13,xmm15
> + movdqa xmm7,xmm10
> + pslld xmm8,5
> + pxor xmm2,xmm4
> + movdqa xmm6,xmm10
> + pandn xmm7,xmm12
> + movdqa xmm5,xmm2
> + pand xmm6,xmm11
> + movdqa xmm9,xmm14
> + psrld xmm5,31
> + paddd xmm2,xmm2
> +
> + movdqa XMMWORD[(0-128)+rax],xmm1
> + paddd xmm13,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm7
> +
> + movdqa xmm7,xmm10
> + por xmm8,xmm9
> + pslld xmm7,30
> + paddd xmm13,xmm6
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((64-128))+rax]
> +
> + movdqa xmm8,xmm13
> + pxor xmm3,XMMWORD[((160-128))+rax]
> + paddd xmm12,xmm15
> + movdqa xmm7,xmm14
> + pslld xmm8,5
> + pxor xmm3,xmm0
> + movdqa xmm6,xmm14
> + pandn xmm7,xmm11
> + movdqa xmm5,xmm3
> + pand xmm6,xmm10
> + movdqa xmm9,xmm13
> + psrld xmm5,31
> + paddd xmm3,xmm3
> +
> + movdqa XMMWORD[(16-128)+rax],xmm2
> + paddd xmm12,xmm2
> + psrld xmm9,27
> + pxor xmm6,xmm7
> +
> + movdqa xmm7,xmm14
> + por xmm8,xmm9
> + pslld xmm7,30
> + paddd xmm12,xmm6
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((80-128))+rax]
> +
> + movdqa xmm8,xmm12
> + pxor xmm4,XMMWORD[((176-128))+rax]
> + paddd xmm11,xmm15
> + movdqa xmm7,xmm13
> + pslld xmm8,5
> + pxor xmm4,xmm1
> + movdqa xmm6,xmm13
> + pandn xmm7,xmm10
> + movdqa xmm5,xmm4
> + pand xmm6,xmm14
> + movdqa xmm9,xmm12
> + psrld xmm5,31
> + paddd xmm4,xmm4
> +
> + movdqa XMMWORD[(32-128)+rax],xmm3
> + paddd xmm11,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm7
> +
> + movdqa xmm7,xmm13
> + por xmm8,xmm9
> + pslld xmm7,30
> + paddd xmm11,xmm6
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((96-128))+rax]
> +
> + movdqa xmm8,xmm11
> + pxor xmm0,XMMWORD[((192-128))+rax]
> + paddd xmm10,xmm15
> + movdqa xmm7,xmm12
> + pslld xmm8,5
> + pxor xmm0,xmm2
> + movdqa xmm6,xmm12
> + pandn xmm7,xmm14
> + movdqa xmm5,xmm0
> + pand xmm6,xmm13
> + movdqa xmm9,xmm11
> + psrld xmm5,31
> + paddd xmm0,xmm0
> +
> + movdqa XMMWORD[(48-128)+rax],xmm4
> + paddd xmm10,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm7
> +
> + movdqa xmm7,xmm12
> + por xmm8,xmm9
> + pslld xmm7,30
> + paddd xmm10,xmm6
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + movdqa xmm15,XMMWORD[rbp]
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((112-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm6,xmm13
> + pxor xmm1,XMMWORD[((208-128))+rax]
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm11
> +
> + movdqa xmm9,xmm10
> + movdqa XMMWORD[(64-128)+rax],xmm0
> + paddd xmm14,xmm0
> + pxor xmm1,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm12
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm14,xmm6
> + paddd xmm1,xmm1
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((128-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm6,xmm12
> + pxor xmm2,XMMWORD[((224-128))+rax]
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm10
> +
> + movdqa xmm9,xmm14
> + movdqa XMMWORD[(80-128)+rax],xmm1
> + paddd xmm13,xmm1
> + pxor xmm2,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm11
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm13,xmm6
> + paddd xmm2,xmm2
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((144-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm6,xmm11
> + pxor xmm3,XMMWORD[((240-128))+rax]
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm14
> +
> + movdqa xmm9,xmm13
> + movdqa XMMWORD[(96-128)+rax],xmm2
> + paddd xmm12,xmm2
> + pxor xmm3,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm10
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm12,xmm6
> + paddd xmm3,xmm3
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((160-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm6,xmm10
> + pxor xmm4,XMMWORD[((0-128))+rax]
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm13
> +
> + movdqa xmm9,xmm12
> + movdqa XMMWORD[(112-128)+rax],xmm3
> + paddd xmm11,xmm3
> + pxor xmm4,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm14
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm11,xmm6
> + paddd xmm4,xmm4
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((176-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm6,xmm14
> + pxor xmm0,XMMWORD[((16-128))+rax]
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm12
> +
> + movdqa xmm9,xmm11
> + movdqa XMMWORD[(128-128)+rax],xmm4
> + paddd xmm10,xmm4
> + pxor xmm0,xmm2
> + psrld xmm9,27
> + pxor xmm6,xmm13
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm10,xmm6
> + paddd xmm0,xmm0
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((192-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm6,xmm13
> + pxor xmm1,XMMWORD[((32-128))+rax]
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm11
> +
> + movdqa xmm9,xmm10
> + movdqa XMMWORD[(144-128)+rax],xmm0
> + paddd xmm14,xmm0
> + pxor xmm1,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm12
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm14,xmm6
> + paddd xmm1,xmm1
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((208-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm6,xmm12
> + pxor xmm2,XMMWORD[((48-128))+rax]
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm10
> +
> + movdqa xmm9,xmm14
> + movdqa XMMWORD[(160-128)+rax],xmm1
> + paddd xmm13,xmm1
> + pxor xmm2,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm11
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm13,xmm6
> + paddd xmm2,xmm2
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((224-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm6,xmm11
> + pxor xmm3,XMMWORD[((64-128))+rax]
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm14
> +
> + movdqa xmm9,xmm13
> + movdqa XMMWORD[(176-128)+rax],xmm2
> + paddd xmm12,xmm2
> + pxor xmm3,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm10
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm12,xmm6
> + paddd xmm3,xmm3
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((240-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm6,xmm10
> + pxor xmm4,XMMWORD[((80-128))+rax]
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm13
> +
> + movdqa xmm9,xmm12
> + movdqa XMMWORD[(192-128)+rax],xmm3
> + paddd xmm11,xmm3
> + pxor xmm4,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm14
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm11,xmm6
> + paddd xmm4,xmm4
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((0-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm6,xmm14
> + pxor xmm0,XMMWORD[((96-128))+rax]
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm12
> +
> + movdqa xmm9,xmm11
> + movdqa XMMWORD[(208-128)+rax],xmm4
> + paddd xmm10,xmm4
> + pxor xmm0,xmm2
> + psrld xmm9,27
> + pxor xmm6,xmm13
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm10,xmm6
> + paddd xmm0,xmm0
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((16-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm6,xmm13
> + pxor xmm1,XMMWORD[((112-128))+rax]
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm11
> +
> + movdqa xmm9,xmm10
> + movdqa XMMWORD[(224-128)+rax],xmm0
> + paddd xmm14,xmm0
> + pxor xmm1,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm12
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm14,xmm6
> + paddd xmm1,xmm1
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((32-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm6,xmm12
> + pxor xmm2,XMMWORD[((128-128))+rax]
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm10
> +
> + movdqa xmm9,xmm14
> + movdqa XMMWORD[(240-128)+rax],xmm1
> + paddd xmm13,xmm1
> + pxor xmm2,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm11
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm13,xmm6
> + paddd xmm2,xmm2
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((48-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm6,xmm11
> + pxor xmm3,XMMWORD[((144-128))+rax]
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm14
> +
> + movdqa xmm9,xmm13
> + movdqa XMMWORD[(0-128)+rax],xmm2
> + paddd xmm12,xmm2
> + pxor xmm3,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm10
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm12,xmm6
> + paddd xmm3,xmm3
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((64-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm6,xmm10
> + pxor xmm4,XMMWORD[((160-128))+rax]
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm13
> +
> + movdqa xmm9,xmm12
> + movdqa XMMWORD[(16-128)+rax],xmm3
> + paddd xmm11,xmm3
> + pxor xmm4,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm14
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm11,xmm6
> + paddd xmm4,xmm4
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((80-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm6,xmm14
> + pxor xmm0,XMMWORD[((176-128))+rax]
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm12
> +
> + movdqa xmm9,xmm11
> + movdqa XMMWORD[(32-128)+rax],xmm4
> + paddd xmm10,xmm4
> + pxor xmm0,xmm2
> + psrld xmm9,27
> + pxor xmm6,xmm13
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm10,xmm6
> + paddd xmm0,xmm0
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((96-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm6,xmm13
> + pxor xmm1,XMMWORD[((192-128))+rax]
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm11
> +
> + movdqa xmm9,xmm10
> + movdqa XMMWORD[(48-128)+rax],xmm0
> + paddd xmm14,xmm0
> + pxor xmm1,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm12
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm14,xmm6
> + paddd xmm1,xmm1
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((112-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm6,xmm12
> + pxor xmm2,XMMWORD[((208-128))+rax]
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm10
> +
> + movdqa xmm9,xmm14
> + movdqa XMMWORD[(64-128)+rax],xmm1
> + paddd xmm13,xmm1
> + pxor xmm2,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm11
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm13,xmm6
> + paddd xmm2,xmm2
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((128-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm6,xmm11
> + pxor xmm3,XMMWORD[((224-128))+rax]
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm14
> +
> + movdqa xmm9,xmm13
> + movdqa XMMWORD[(80-128)+rax],xmm2
> + paddd xmm12,xmm2
> + pxor xmm3,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm10
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm12,xmm6
> + paddd xmm3,xmm3
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((144-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm6,xmm10
> + pxor xmm4,XMMWORD[((240-128))+rax]
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm13
> +
> + movdqa xmm9,xmm12
> + movdqa XMMWORD[(96-128)+rax],xmm3
> + paddd xmm11,xmm3
> + pxor xmm4,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm14
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm11,xmm6
> + paddd xmm4,xmm4
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((160-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm6,xmm14
> + pxor xmm0,XMMWORD[((0-128))+rax]
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm12
> +
> + movdqa xmm9,xmm11
> + movdqa XMMWORD[(112-128)+rax],xmm4
> + paddd xmm10,xmm4
> + pxor xmm0,xmm2
> + psrld xmm9,27
> + pxor xmm6,xmm13
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm10,xmm6
> + paddd xmm0,xmm0
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + movdqa xmm15,XMMWORD[32+rbp]
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((176-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm7,xmm13
> + pxor xmm1,XMMWORD[((16-128))+rax]
> + pxor xmm1,xmm3
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm10
> + pand xmm7,xmm12
> +
> + movdqa xmm6,xmm13
> + movdqa xmm5,xmm1
> + psrld xmm9,27
> + paddd xmm14,xmm7
> + pxor xmm6,xmm12
> +
> + movdqa XMMWORD[(128-128)+rax],xmm0
> + paddd xmm14,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm11
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + paddd xmm1,xmm1
> + paddd xmm14,xmm6
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((192-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm7,xmm12
> + pxor xmm2,XMMWORD[((32-128))+rax]
> + pxor xmm2,xmm4
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm14
> + pand xmm7,xmm11
> +
> + movdqa xmm6,xmm12
> + movdqa xmm5,xmm2
> + psrld xmm9,27
> + paddd xmm13,xmm7
> + pxor xmm6,xmm11
> +
> + movdqa XMMWORD[(144-128)+rax],xmm1
> + paddd xmm13,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm10
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + paddd xmm2,xmm2
> + paddd xmm13,xmm6
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((208-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm7,xmm11
> + pxor xmm3,XMMWORD[((48-128))+rax]
> + pxor xmm3,xmm0
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm13
> + pand xmm7,xmm10
> +
> + movdqa xmm6,xmm11
> + movdqa xmm5,xmm3
> + psrld xmm9,27
> + paddd xmm12,xmm7
> + pxor xmm6,xmm10
> +
> + movdqa XMMWORD[(160-128)+rax],xmm2
> + paddd xmm12,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm14
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + paddd xmm3,xmm3
> + paddd xmm12,xmm6
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((224-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm7,xmm10
> + pxor xmm4,XMMWORD[((64-128))+rax]
> + pxor xmm4,xmm1
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm12
> + pand xmm7,xmm14
> +
> + movdqa xmm6,xmm10
> + movdqa xmm5,xmm4
> + psrld xmm9,27
> + paddd xmm11,xmm7
> + pxor xmm6,xmm14
> +
> + movdqa XMMWORD[(176-128)+rax],xmm3
> + paddd xmm11,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm13
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + paddd xmm4,xmm4
> + paddd xmm11,xmm6
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((240-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm7,xmm14
> + pxor xmm0,XMMWORD[((80-128))+rax]
> + pxor xmm0,xmm2
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm11
> + pand xmm7,xmm13
> +
> + movdqa xmm6,xmm14
> + movdqa xmm5,xmm0
> + psrld xmm9,27
> + paddd xmm10,xmm7
> + pxor xmm6,xmm13
> +
> + movdqa XMMWORD[(192-128)+rax],xmm4
> + paddd xmm10,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm12
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + paddd xmm0,xmm0
> + paddd xmm10,xmm6
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((0-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm7,xmm13
> + pxor xmm1,XMMWORD[((96-128))+rax]
> + pxor xmm1,xmm3
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm10
> + pand xmm7,xmm12
> +
> + movdqa xmm6,xmm13
> + movdqa xmm5,xmm1
> + psrld xmm9,27
> + paddd xmm14,xmm7
> + pxor xmm6,xmm12
> +
> + movdqa XMMWORD[(208-128)+rax],xmm0
> + paddd xmm14,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm11
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + paddd xmm1,xmm1
> + paddd xmm14,xmm6
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((16-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm7,xmm12
> + pxor xmm2,XMMWORD[((112-128))+rax]
> + pxor xmm2,xmm4
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm14
> + pand xmm7,xmm11
> +
> + movdqa xmm6,xmm12
> + movdqa xmm5,xmm2
> + psrld xmm9,27
> + paddd xmm13,xmm7
> + pxor xmm6,xmm11
> +
> + movdqa XMMWORD[(224-128)+rax],xmm1
> + paddd xmm13,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm10
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + paddd xmm2,xmm2
> + paddd xmm13,xmm6
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((32-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm7,xmm11
> + pxor xmm3,XMMWORD[((128-128))+rax]
> + pxor xmm3,xmm0
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm13
> + pand xmm7,xmm10
> +
> + movdqa xmm6,xmm11
> + movdqa xmm5,xmm3
> + psrld xmm9,27
> + paddd xmm12,xmm7
> + pxor xmm6,xmm10
> +
> + movdqa XMMWORD[(240-128)+rax],xmm2
> + paddd xmm12,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm14
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + paddd xmm3,xmm3
> + paddd xmm12,xmm6
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((48-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm7,xmm10
> + pxor xmm4,XMMWORD[((144-128))+rax]
> + pxor xmm4,xmm1
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm12
> + pand xmm7,xmm14
> +
> + movdqa xmm6,xmm10
> + movdqa xmm5,xmm4
> + psrld xmm9,27
> + paddd xmm11,xmm7
> + pxor xmm6,xmm14
> +
> + movdqa XMMWORD[(0-128)+rax],xmm3
> + paddd xmm11,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm13
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + paddd xmm4,xmm4
> + paddd xmm11,xmm6
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((64-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm7,xmm14
> + pxor xmm0,XMMWORD[((160-128))+rax]
> + pxor xmm0,xmm2
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm11
> + pand xmm7,xmm13
> +
> + movdqa xmm6,xmm14
> + movdqa xmm5,xmm0
> + psrld xmm9,27
> + paddd xmm10,xmm7
> + pxor xmm6,xmm13
> +
> + movdqa XMMWORD[(16-128)+rax],xmm4
> + paddd xmm10,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm12
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + paddd xmm0,xmm0
> + paddd xmm10,xmm6
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((80-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm7,xmm13
> + pxor xmm1,XMMWORD[((176-128))+rax]
> + pxor xmm1,xmm3
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm10
> + pand xmm7,xmm12
> +
> + movdqa xmm6,xmm13
> + movdqa xmm5,xmm1
> + psrld xmm9,27
> + paddd xmm14,xmm7
> + pxor xmm6,xmm12
> +
> + movdqa XMMWORD[(32-128)+rax],xmm0
> + paddd xmm14,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm11
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + paddd xmm1,xmm1
> + paddd xmm14,xmm6
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((96-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm7,xmm12
> + pxor xmm2,XMMWORD[((192-128))+rax]
> + pxor xmm2,xmm4
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm14
> + pand xmm7,xmm11
> +
> + movdqa xmm6,xmm12
> + movdqa xmm5,xmm2
> + psrld xmm9,27
> + paddd xmm13,xmm7
> + pxor xmm6,xmm11
> +
> + movdqa XMMWORD[(48-128)+rax],xmm1
> + paddd xmm13,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm10
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + paddd xmm2,xmm2
> + paddd xmm13,xmm6
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((112-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm7,xmm11
> + pxor xmm3,XMMWORD[((208-128))+rax]
> + pxor xmm3,xmm0
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm13
> + pand xmm7,xmm10
> +
> + movdqa xmm6,xmm11
> + movdqa xmm5,xmm3
> + psrld xmm9,27
> + paddd xmm12,xmm7
> + pxor xmm6,xmm10
> +
> + movdqa XMMWORD[(64-128)+rax],xmm2
> + paddd xmm12,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm14
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + paddd xmm3,xmm3
> + paddd xmm12,xmm6
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((128-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm7,xmm10
> + pxor xmm4,XMMWORD[((224-128))+rax]
> + pxor xmm4,xmm1
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm12
> + pand xmm7,xmm14
> +
> + movdqa xmm6,xmm10
> + movdqa xmm5,xmm4
> + psrld xmm9,27
> + paddd xmm11,xmm7
> + pxor xmm6,xmm14
> +
> + movdqa XMMWORD[(80-128)+rax],xmm3
> + paddd xmm11,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm13
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + paddd xmm4,xmm4
> + paddd xmm11,xmm6
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((144-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm7,xmm14
> + pxor xmm0,XMMWORD[((240-128))+rax]
> + pxor xmm0,xmm2
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm11
> + pand xmm7,xmm13
> +
> + movdqa xmm6,xmm14
> + movdqa xmm5,xmm0
> + psrld xmm9,27
> + paddd xmm10,xmm7
> + pxor xmm6,xmm13
> +
> + movdqa XMMWORD[(96-128)+rax],xmm4
> + paddd xmm10,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm12
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + paddd xmm0,xmm0
> + paddd xmm10,xmm6
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((160-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm7,xmm13
> + pxor xmm1,XMMWORD[((0-128))+rax]
> + pxor xmm1,xmm3
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm10
> + pand xmm7,xmm12
> +
> + movdqa xmm6,xmm13
> + movdqa xmm5,xmm1
> + psrld xmm9,27
> + paddd xmm14,xmm7
> + pxor xmm6,xmm12
> +
> + movdqa XMMWORD[(112-128)+rax],xmm0
> + paddd xmm14,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm11
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + paddd xmm1,xmm1
> + paddd xmm14,xmm6
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((176-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm7,xmm12
> + pxor xmm2,XMMWORD[((16-128))+rax]
> + pxor xmm2,xmm4
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm14
> + pand xmm7,xmm11
> +
> + movdqa xmm6,xmm12
> + movdqa xmm5,xmm2
> + psrld xmm9,27
> + paddd xmm13,xmm7
> + pxor xmm6,xmm11
> +
> + movdqa XMMWORD[(128-128)+rax],xmm1
> + paddd xmm13,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm10
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + paddd xmm2,xmm2
> + paddd xmm13,xmm6
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((192-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm7,xmm11
> + pxor xmm3,XMMWORD[((32-128))+rax]
> + pxor xmm3,xmm0
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm13
> + pand xmm7,xmm10
> +
> + movdqa xmm6,xmm11
> + movdqa xmm5,xmm3
> + psrld xmm9,27
> + paddd xmm12,xmm7
> + pxor xmm6,xmm10
> +
> + movdqa XMMWORD[(144-128)+rax],xmm2
> + paddd xmm12,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm14
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + paddd xmm3,xmm3
> + paddd xmm12,xmm6
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((208-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm7,xmm10
> + pxor xmm4,XMMWORD[((48-128))+rax]
> + pxor xmm4,xmm1
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm12
> + pand xmm7,xmm14
> +
> + movdqa xmm6,xmm10
> + movdqa xmm5,xmm4
> + psrld xmm9,27
> + paddd xmm11,xmm7
> + pxor xmm6,xmm14
> +
> + movdqa XMMWORD[(160-128)+rax],xmm3
> + paddd xmm11,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm13
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + paddd xmm4,xmm4
> + paddd xmm11,xmm6
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((224-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm7,xmm14
> + pxor xmm0,XMMWORD[((64-128))+rax]
> + pxor xmm0,xmm2
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + movdqa xmm9,xmm11
> + pand xmm7,xmm13
> +
> + movdqa xmm6,xmm14
> + movdqa xmm5,xmm0
> + psrld xmm9,27
> + paddd xmm10,xmm7
> + pxor xmm6,xmm13
> +
> + movdqa XMMWORD[(176-128)+rax],xmm4
> + paddd xmm10,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + pand xmm6,xmm12
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + paddd xmm0,xmm0
> + paddd xmm10,xmm6
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + movdqa xmm15,XMMWORD[64+rbp]
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((240-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm6,xmm13
> + pxor xmm1,XMMWORD[((80-128))+rax]
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm11
> +
> + movdqa xmm9,xmm10
> + movdqa XMMWORD[(192-128)+rax],xmm0
> + paddd xmm14,xmm0
> + pxor xmm1,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm12
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm14,xmm6
> + paddd xmm1,xmm1
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((0-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm6,xmm12
> + pxor xmm2,XMMWORD[((96-128))+rax]
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm10
> +
> + movdqa xmm9,xmm14
> + movdqa XMMWORD[(208-128)+rax],xmm1
> + paddd xmm13,xmm1
> + pxor xmm2,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm11
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm13,xmm6
> + paddd xmm2,xmm2
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((16-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm6,xmm11
> + pxor xmm3,XMMWORD[((112-128))+rax]
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm14
> +
> + movdqa xmm9,xmm13
> + movdqa XMMWORD[(224-128)+rax],xmm2
> + paddd xmm12,xmm2
> + pxor xmm3,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm10
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm12,xmm6
> + paddd xmm3,xmm3
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((32-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm6,xmm10
> + pxor xmm4,XMMWORD[((128-128))+rax]
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm13
> +
> + movdqa xmm9,xmm12
> + movdqa XMMWORD[(240-128)+rax],xmm3
> + paddd xmm11,xmm3
> + pxor xmm4,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm14
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm11,xmm6
> + paddd xmm4,xmm4
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((48-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm6,xmm14
> + pxor xmm0,XMMWORD[((144-128))+rax]
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm12
> +
> + movdqa xmm9,xmm11
> + movdqa XMMWORD[(0-128)+rax],xmm4
> + paddd xmm10,xmm4
> + pxor xmm0,xmm2
> + psrld xmm9,27
> + pxor xmm6,xmm13
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm10,xmm6
> + paddd xmm0,xmm0
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((64-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm6,xmm13
> + pxor xmm1,XMMWORD[((160-128))+rax]
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm11
> +
> + movdqa xmm9,xmm10
> + movdqa XMMWORD[(16-128)+rax],xmm0
> + paddd xmm14,xmm0
> + pxor xmm1,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm12
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm14,xmm6
> + paddd xmm1,xmm1
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((80-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm6,xmm12
> + pxor xmm2,XMMWORD[((176-128))+rax]
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm10
> +
> + movdqa xmm9,xmm14
> + movdqa XMMWORD[(32-128)+rax],xmm1
> + paddd xmm13,xmm1
> + pxor xmm2,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm11
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm13,xmm6
> + paddd xmm2,xmm2
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((96-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm6,xmm11
> + pxor xmm3,XMMWORD[((192-128))+rax]
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm14
> +
> + movdqa xmm9,xmm13
> + movdqa XMMWORD[(48-128)+rax],xmm2
> + paddd xmm12,xmm2
> + pxor xmm3,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm10
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm12,xmm6
> + paddd xmm3,xmm3
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((112-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm6,xmm10
> + pxor xmm4,XMMWORD[((208-128))+rax]
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm13
> +
> + movdqa xmm9,xmm12
> + movdqa XMMWORD[(64-128)+rax],xmm3
> + paddd xmm11,xmm3
> + pxor xmm4,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm14
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm11,xmm6
> + paddd xmm4,xmm4
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((128-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm6,xmm14
> + pxor xmm0,XMMWORD[((224-128))+rax]
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm12
> +
> + movdqa xmm9,xmm11
> + movdqa XMMWORD[(80-128)+rax],xmm4
> + paddd xmm10,xmm4
> + pxor xmm0,xmm2
> + psrld xmm9,27
> + pxor xmm6,xmm13
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm10,xmm6
> + paddd xmm0,xmm0
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((144-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm6,xmm13
> + pxor xmm1,XMMWORD[((240-128))+rax]
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm11
> +
> + movdqa xmm9,xmm10
> + movdqa XMMWORD[(96-128)+rax],xmm0
> + paddd xmm14,xmm0
> + pxor xmm1,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm12
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm14,xmm6
> + paddd xmm1,xmm1
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((160-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm6,xmm12
> + pxor xmm2,XMMWORD[((0-128))+rax]
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm10
> +
> + movdqa xmm9,xmm14
> + movdqa XMMWORD[(112-128)+rax],xmm1
> + paddd xmm13,xmm1
> + pxor xmm2,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm11
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm13,xmm6
> + paddd xmm2,xmm2
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((176-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm6,xmm11
> + pxor xmm3,XMMWORD[((16-128))+rax]
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm14
> +
> + movdqa xmm9,xmm13
> + paddd xmm12,xmm2
> + pxor xmm3,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm10
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm12,xmm6
> + paddd xmm3,xmm3
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((192-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm6,xmm10
> + pxor xmm4,XMMWORD[((32-128))+rax]
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm13
> +
> + movdqa xmm9,xmm12
> + paddd xmm11,xmm3
> + pxor xmm4,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm14
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm11,xmm6
> + paddd xmm4,xmm4
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + pxor xmm0,xmm2
> + movdqa xmm2,XMMWORD[((208-128))+rax]
> +
> + movdqa xmm8,xmm11
> + movdqa xmm6,xmm14
> + pxor xmm0,XMMWORD[((48-128))+rax]
> + paddd xmm10,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm12
> +
> + movdqa xmm9,xmm11
> + paddd xmm10,xmm4
> + pxor xmm0,xmm2
> + psrld xmm9,27
> + pxor xmm6,xmm13
> + movdqa xmm7,xmm12
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm0
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm10,xmm6
> + paddd xmm0,xmm0
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm0,xmm5
> + por xmm12,xmm7
> + pxor xmm1,xmm3
> + movdqa xmm3,XMMWORD[((224-128))+rax]
> +
> + movdqa xmm8,xmm10
> + movdqa xmm6,xmm13
> + pxor xmm1,XMMWORD[((64-128))+rax]
> + paddd xmm14,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm11
> +
> + movdqa xmm9,xmm10
> + paddd xmm14,xmm0
> + pxor xmm1,xmm3
> + psrld xmm9,27
> + pxor xmm6,xmm12
> + movdqa xmm7,xmm11
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm1
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm14,xmm6
> + paddd xmm1,xmm1
> +
> + psrld xmm11,2
> + paddd xmm14,xmm8
> + por xmm1,xmm5
> + por xmm11,xmm7
> + pxor xmm2,xmm4
> + movdqa xmm4,XMMWORD[((240-128))+rax]
> +
> + movdqa xmm8,xmm14
> + movdqa xmm6,xmm12
> + pxor xmm2,XMMWORD[((80-128))+rax]
> + paddd xmm13,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm10
> +
> + movdqa xmm9,xmm14
> + paddd xmm13,xmm1
> + pxor xmm2,xmm4
> + psrld xmm9,27
> + pxor xmm6,xmm11
> + movdqa xmm7,xmm10
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm2
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm13,xmm6
> + paddd xmm2,xmm2
> +
> + psrld xmm10,2
> + paddd xmm13,xmm8
> + por xmm2,xmm5
> + por xmm10,xmm7
> + pxor xmm3,xmm0
> + movdqa xmm0,XMMWORD[((0-128))+rax]
> +
> + movdqa xmm8,xmm13
> + movdqa xmm6,xmm11
> + pxor xmm3,XMMWORD[((96-128))+rax]
> + paddd xmm12,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm14
> +
> + movdqa xmm9,xmm13
> + paddd xmm12,xmm2
> + pxor xmm3,xmm0
> + psrld xmm9,27
> + pxor xmm6,xmm10
> + movdqa xmm7,xmm14
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm3
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm12,xmm6
> + paddd xmm3,xmm3
> +
> + psrld xmm14,2
> + paddd xmm12,xmm8
> + por xmm3,xmm5
> + por xmm14,xmm7
> + pxor xmm4,xmm1
> + movdqa xmm1,XMMWORD[((16-128))+rax]
> +
> + movdqa xmm8,xmm12
> + movdqa xmm6,xmm10
> + pxor xmm4,XMMWORD[((112-128))+rax]
> + paddd xmm11,xmm15
> + pslld xmm8,5
> + pxor xmm6,xmm13
> +
> + movdqa xmm9,xmm12
> + paddd xmm11,xmm3
> + pxor xmm4,xmm1
> + psrld xmm9,27
> + pxor xmm6,xmm14
> + movdqa xmm7,xmm13
> +
> + pslld xmm7,30
> + movdqa xmm5,xmm4
> + por xmm8,xmm9
> + psrld xmm5,31
> + paddd xmm11,xmm6
> + paddd xmm4,xmm4
> +
> + psrld xmm13,2
> + paddd xmm11,xmm8
> + por xmm4,xmm5
> + por xmm13,xmm7
> + movdqa xmm8,xmm11
> + paddd xmm10,xmm15
> + movdqa xmm6,xmm14
> + pslld xmm8,5
> + pxor xmm6,xmm12
> +
> + movdqa xmm9,xmm11
> + paddd xmm10,xmm4
> + psrld xmm9,27
> + movdqa xmm7,xmm12
> + pxor xmm6,xmm13
> +
> + pslld xmm7,30
> + por xmm8,xmm9
> + paddd xmm10,xmm6
> +
> + psrld xmm12,2
> + paddd xmm10,xmm8
> + por xmm12,xmm7
> + movdqa xmm0,XMMWORD[rbx]
> + mov ecx,1
> + cmp ecx,DWORD[rbx]
> + pxor xmm8,xmm8
> + cmovge r8,rbp
> + cmp ecx,DWORD[4+rbx]
> + movdqa xmm1,xmm0
> + cmovge r9,rbp
> + cmp ecx,DWORD[8+rbx]
> + pcmpgtd xmm1,xmm8
> + cmovge r10,rbp
> + cmp ecx,DWORD[12+rbx]
> + paddd xmm0,xmm1
> + cmovge r11,rbp
> +
> + movdqu xmm6,XMMWORD[rdi]
> + pand xmm10,xmm1
> + movdqu xmm7,XMMWORD[32+rdi]
> + pand xmm11,xmm1
> + paddd xmm10,xmm6
> + movdqu xmm8,XMMWORD[64+rdi]
> + pand xmm12,xmm1
> + paddd xmm11,xmm7
> + movdqu xmm9,XMMWORD[96+rdi]
> + pand xmm13,xmm1
> + paddd xmm12,xmm8
> + movdqu xmm5,XMMWORD[128+rdi]
> + pand xmm14,xmm1
> + movdqu XMMWORD[rdi],xmm10
> + paddd xmm13,xmm9
> + movdqu XMMWORD[32+rdi],xmm11
> + paddd xmm14,xmm5
> + movdqu XMMWORD[64+rdi],xmm12
> + movdqu XMMWORD[96+rdi],xmm13
> + movdqu XMMWORD[128+rdi],xmm14
> +
> + movdqa XMMWORD[rbx],xmm0
> + movdqa xmm5,XMMWORD[96+rbp]
> + movdqa xmm15,XMMWORD[((-32))+rbp]
> + dec edx
> + jnz NEAR $L$oop
> +
> + mov edx,DWORD[280+rsp]
> + lea rdi,[16+rdi]
> + lea rsi,[64+rsi]
> + dec edx
> + jnz NEAR $L$oop_grande
> +
> +$L$done:
> + mov rax,QWORD[272+rsp]
> +
> + movaps xmm6,XMMWORD[((-184))+rax]
> + movaps xmm7,XMMWORD[((-168))+rax]
> + movaps xmm8,XMMWORD[((-152))+rax]
> + movaps xmm9,XMMWORD[((-136))+rax]
> + movaps xmm10,XMMWORD[((-120))+rax]
> + movaps xmm11,XMMWORD[((-104))+rax]
> + movaps xmm12,XMMWORD[((-88))+rax]
> + movaps xmm13,XMMWORD[((-72))+rax]
> + movaps xmm14,XMMWORD[((-56))+rax]
> + movaps xmm15,XMMWORD[((-40))+rax]
> + mov rbp,QWORD[((-16))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> +
> + lea rsp,[rax]
> +
> +$L$epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha1_multi_block:
> +
> +ALIGN 32
> +sha1_multi_block_shaext:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha1_multi_block_shaext:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> +_shaext_shortcut:
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + lea rsp,[((-168))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> + movaps XMMWORD[(-120)+rax],xmm10
> + movaps XMMWORD[(-104)+rax],xmm11
> + movaps XMMWORD[(-88)+rax],xmm12
> + movaps XMMWORD[(-72)+rax],xmm13
> + movaps XMMWORD[(-56)+rax],xmm14
> + movaps XMMWORD[(-40)+rax],xmm15
> + sub rsp,288
> + shl edx,1
> + and rsp,-256
> + lea rdi,[64+rdi]
> + mov QWORD[272+rsp],rax
> +$L$body_shaext:
> + lea rbx,[256+rsp]
> + movdqa xmm3,XMMWORD[((K_XX_XX+128))]
> +
> +$L$oop_grande_shaext:
> + mov DWORD[280+rsp],edx
> + xor edx,edx
> + mov r8,QWORD[rsi]
> + mov ecx,DWORD[8+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[rbx],ecx
> + cmovle r8,rsp
> + mov r9,QWORD[16+rsi]
> + mov ecx,DWORD[24+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[4+rbx],ecx
> + cmovle r9,rsp
> + test edx,edx
> + jz NEAR $L$done_shaext
> +
> + movq xmm0,QWORD[((0-64))+rdi]
> + movq xmm4,QWORD[((32-64))+rdi]
> + movq xmm5,QWORD[((64-64))+rdi]
> + movq xmm6,QWORD[((96-64))+rdi]
> + movq xmm7,QWORD[((128-64))+rdi]
> +
> + punpckldq xmm0,xmm4
> + punpckldq xmm5,xmm6
> +
> + movdqa xmm8,xmm0
> + punpcklqdq xmm0,xmm5
> + punpckhqdq xmm8,xmm5
> +
> + pshufd xmm1,xmm7,63
> + pshufd xmm9,xmm7,127
> + pshufd xmm0,xmm0,27
> + pshufd xmm8,xmm8,27
> + jmp NEAR $L$oop_shaext
> +
> +ALIGN 32
> +$L$oop_shaext:
> + movdqu xmm4,XMMWORD[r8]
> + movdqu xmm11,XMMWORD[r9]
> + movdqu xmm5,XMMWORD[16+r8]
> + movdqu xmm12,XMMWORD[16+r9]
> + movdqu xmm6,XMMWORD[32+r8]
> +DB 102,15,56,0,227
> + movdqu xmm13,XMMWORD[32+r9]
> +DB 102,68,15,56,0,219
> + movdqu xmm7,XMMWORD[48+r8]
> + lea r8,[64+r8]
> +DB 102,15,56,0,235
> + movdqu xmm14,XMMWORD[48+r9]
> + lea r9,[64+r9]
> +DB 102,68,15,56,0,227
> +
> + movdqa XMMWORD[80+rsp],xmm1
> + paddd xmm1,xmm4
> + movdqa XMMWORD[112+rsp],xmm9
> + paddd xmm9,xmm11
> + movdqa XMMWORD[64+rsp],xmm0
> + movdqa xmm2,xmm0
> + movdqa XMMWORD[96+rsp],xmm8
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,0
> +DB 15,56,200,213
> +DB 69,15,58,204,193,0
> +DB 69,15,56,200,212
> +DB 102,15,56,0,243
> + prefetcht0 [127+r8]
> +DB 15,56,201,229
> +DB 102,68,15,56,0,235
> + prefetcht0 [127+r9]
> +DB 69,15,56,201,220
> +
> +DB 102,15,56,0,251
> + movdqa xmm1,xmm0
> +DB 102,68,15,56,0,243
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,0
> +DB 15,56,200,206
> +DB 69,15,58,204,194,0
> +DB 69,15,56,200,205
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> + pxor xmm11,xmm13
> +DB 69,15,56,201,229
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,0
> +DB 15,56,200,215
> +DB 69,15,58,204,193,0
> +DB 69,15,56,200,214
> +DB 15,56,202,231
> +DB 69,15,56,202,222
> + pxor xmm5,xmm7
> +DB 15,56,201,247
> + pxor xmm12,xmm14
> +DB 69,15,56,201,238
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,0
> +DB 15,56,200,204
> +DB 69,15,58,204,194,0
> +DB 69,15,56,200,203
> +DB 15,56,202,236
> +DB 69,15,56,202,227
> + pxor xmm6,xmm4
> +DB 15,56,201,252
> + pxor xmm13,xmm11
> +DB 69,15,56,201,243
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,0
> +DB 15,56,200,213
> +DB 69,15,58,204,193,0
> +DB 69,15,56,200,212
> +DB 15,56,202,245
> +DB 69,15,56,202,236
> + pxor xmm7,xmm5
> +DB 15,56,201,229
> + pxor xmm14,xmm12
> +DB 69,15,56,201,220
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,1
> +DB 15,56,200,206
> +DB 69,15,58,204,194,1
> +DB 69,15,56,200,205
> +DB 15,56,202,254
> +DB 69,15,56,202,245
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> + pxor xmm11,xmm13
> +DB 69,15,56,201,229
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,1
> +DB 15,56,200,215
> +DB 69,15,58,204,193,1
> +DB 69,15,56,200,214
> +DB 15,56,202,231
> +DB 69,15,56,202,222
> + pxor xmm5,xmm7
> +DB 15,56,201,247
> + pxor xmm12,xmm14
> +DB 69,15,56,201,238
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,1
> +DB 15,56,200,204
> +DB 69,15,58,204,194,1
> +DB 69,15,56,200,203
> +DB 15,56,202,236
> +DB 69,15,56,202,227
> + pxor xmm6,xmm4
> +DB 15,56,201,252
> + pxor xmm13,xmm11
> +DB 69,15,56,201,243
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,1
> +DB 15,56,200,213
> +DB 69,15,58,204,193,1
> +DB 69,15,56,200,212
> +DB 15,56,202,245
> +DB 69,15,56,202,236
> + pxor xmm7,xmm5
> +DB 15,56,201,229
> + pxor xmm14,xmm12
> +DB 69,15,56,201,220
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,1
> +DB 15,56,200,206
> +DB 69,15,58,204,194,1
> +DB 69,15,56,200,205
> +DB 15,56,202,254
> +DB 69,15,56,202,245
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> + pxor xmm11,xmm13
> +DB 69,15,56,201,229
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,2
> +DB 15,56,200,215
> +DB 69,15,58,204,193,2
> +DB 69,15,56,200,214
> +DB 15,56,202,231
> +DB 69,15,56,202,222
> + pxor xmm5,xmm7
> +DB 15,56,201,247
> + pxor xmm12,xmm14
> +DB 69,15,56,201,238
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,2
> +DB 15,56,200,204
> +DB 69,15,58,204,194,2
> +DB 69,15,56,200,203
> +DB 15,56,202,236
> +DB 69,15,56,202,227
> + pxor xmm6,xmm4
> +DB 15,56,201,252
> + pxor xmm13,xmm11
> +DB 69,15,56,201,243
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,2
> +DB 15,56,200,213
> +DB 69,15,58,204,193,2
> +DB 69,15,56,200,212
> +DB 15,56,202,245
> +DB 69,15,56,202,236
> + pxor xmm7,xmm5
> +DB 15,56,201,229
> + pxor xmm14,xmm12
> +DB 69,15,56,201,220
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,2
> +DB 15,56,200,206
> +DB 69,15,58,204,194,2
> +DB 69,15,56,200,205
> +DB 15,56,202,254
> +DB 69,15,56,202,245
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> + pxor xmm11,xmm13
> +DB 69,15,56,201,229
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,2
> +DB 15,56,200,215
> +DB 69,15,58,204,193,2
> +DB 69,15,56,200,214
> +DB 15,56,202,231
> +DB 69,15,56,202,222
> + pxor xmm5,xmm7
> +DB 15,56,201,247
> + pxor xmm12,xmm14
> +DB 69,15,56,201,238
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,3
> +DB 15,56,200,204
> +DB 69,15,58,204,194,3
> +DB 69,15,56,200,203
> +DB 15,56,202,236
> +DB 69,15,56,202,227
> + pxor xmm6,xmm4
> +DB 15,56,201,252
> + pxor xmm13,xmm11
> +DB 69,15,56,201,243
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,3
> +DB 15,56,200,213
> +DB 69,15,58,204,193,3
> +DB 69,15,56,200,212
> +DB 15,56,202,245
> +DB 69,15,56,202,236
> + pxor xmm7,xmm5
> + pxor xmm14,xmm12
> +
> + mov ecx,1
> + pxor xmm4,xmm4
> + cmp ecx,DWORD[rbx]
> + cmovge r8,rsp
> +
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,3
> +DB 15,56,200,206
> +DB 69,15,58,204,194,3
> +DB 69,15,56,200,205
> +DB 15,56,202,254
> +DB 69,15,56,202,245
> +
> + cmp ecx,DWORD[4+rbx]
> + cmovge r9,rsp
> + movq xmm6,QWORD[rbx]
> +
> + movdqa xmm2,xmm0
> + movdqa xmm10,xmm8
> +DB 15,58,204,193,3
> +DB 15,56,200,215
> +DB 69,15,58,204,193,3
> +DB 69,15,56,200,214
> +
> + pshufd xmm11,xmm6,0x00
> + pshufd xmm12,xmm6,0x55
> + movdqa xmm7,xmm6
> + pcmpgtd xmm11,xmm4
> + pcmpgtd xmm12,xmm4
> +
> + movdqa xmm1,xmm0
> + movdqa xmm9,xmm8
> +DB 15,58,204,194,3
> +DB 15,56,200,204
> +DB 69,15,58,204,194,3
> +DB 68,15,56,200,204
> +
> + pcmpgtd xmm7,xmm4
> + pand xmm0,xmm11
> + pand xmm1,xmm11
> + pand xmm8,xmm12
> + pand xmm9,xmm12
> + paddd xmm6,xmm7
> +
> + paddd xmm0,XMMWORD[64+rsp]
> + paddd xmm1,XMMWORD[80+rsp]
> + paddd xmm8,XMMWORD[96+rsp]
> + paddd xmm9,XMMWORD[112+rsp]
> +
> + movq QWORD[rbx],xmm6
> + dec edx
> + jnz NEAR $L$oop_shaext
> +
> + mov edx,DWORD[280+rsp]
> +
> + pshufd xmm0,xmm0,27
> + pshufd xmm8,xmm8,27
> +
> + movdqa xmm6,xmm0
> + punpckldq xmm0,xmm8
> + punpckhdq xmm6,xmm8
> + punpckhdq xmm1,xmm9
> + movq QWORD[(0-64)+rdi],xmm0
> + psrldq xmm0,8
> + movq QWORD[(64-64)+rdi],xmm6
> + psrldq xmm6,8
> + movq QWORD[(32-64)+rdi],xmm0
> + psrldq xmm1,8
> + movq QWORD[(96-64)+rdi],xmm6
> + movq QWORD[(128-64)+rdi],xmm1
> +
> + lea rdi,[8+rdi]
> + lea rsi,[32+rsi]
> + dec edx
> + jnz NEAR $L$oop_grande_shaext
> +
> +$L$done_shaext:
> +
> + movaps xmm6,XMMWORD[((-184))+rax]
> + movaps xmm7,XMMWORD[((-168))+rax]
> + movaps xmm8,XMMWORD[((-152))+rax]
> + movaps xmm9,XMMWORD[((-136))+rax]
> + movaps xmm10,XMMWORD[((-120))+rax]
> + movaps xmm11,XMMWORD[((-104))+rax]
> + movaps xmm12,XMMWORD[((-88))+rax]
> + movaps xmm13,XMMWORD[((-72))+rax]
> + movaps xmm14,XMMWORD[((-56))+rax]
> + movaps xmm15,XMMWORD[((-40))+rax]
> + mov rbp,QWORD[((-16))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> +
> + lea rsp,[rax]
> +
> +$L$epilogue_shaext:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha1_multi_block_shaext:
> +
> +ALIGN 256
> + DD 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> + DD 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> +K_XX_XX:
> + DD 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> + DD 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> + DD 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> + DD 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> + DD 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> + DD 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +DB 0xf,0xe,0xd,0xc,0xb,0xa,0x9,0x8,0x7,0x6,0x5,0x4,0x3,0x2,0x1,0x0
> +DB 83,72,65,49,32,109,117,108,116,105,45,98,108,111,99,107
> +DB 32,116,114,97,110,115,102,111,114,109,32,102,111,114,32,120
> +DB 56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77
> +DB 83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110
> +DB 115,115,108,46,111,114,103,62,0
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$in_prologue
> +
> + mov rax,QWORD[272+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> +
> + lea rsi,[((-24-160))+rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> +
> +$L$in_prologue:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_sha1_multi_block wrt ..imagebase
> + DD $L$SEH_end_sha1_multi_block wrt ..imagebase
> + DD $L$SEH_info_sha1_multi_block wrt ..imagebase
> + DD $L$SEH_begin_sha1_multi_block_shaext wrt ..imagebase
> + DD $L$SEH_end_sha1_multi_block_shaext wrt ..imagebase
> + DD $L$SEH_info_sha1_multi_block_shaext wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_sha1_multi_block:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$body wrt ..imagebase,$L$epilogue wrt ..imagebase
> +$L$SEH_info_sha1_multi_block_shaext:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$body_shaext wrt ..imagebase,$L$epilogue_shaext
> wrt ..imagebase
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha1-x86_64.nasm
> b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha1-x86_64.nasm
> new file mode 100644
> index 0000000000..c6d68d348f
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha1-x86_64.nasm
> @@ -0,0 +1,2884 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/sha/asm/sha1-x86_64.pl
> +;
> +; Copyright 2006-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +EXTERN OPENSSL_ia32cap_P
> +
> +global sha1_block_data_order
> +
> +ALIGN 16
> +sha1_block_data_order:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha1_block_data_order:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + mov r9d,DWORD[((OPENSSL_ia32cap_P+0))]
> + mov r8d,DWORD[((OPENSSL_ia32cap_P+4))]
> + mov r10d,DWORD[((OPENSSL_ia32cap_P+8))]
> + test r8d,512
> + jz NEAR $L$ialu
> + test r10d,536870912
> + jnz NEAR _shaext_shortcut
> + jmp NEAR _ssse3_shortcut
> +
> +ALIGN 16
> +$L$ialu:
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + mov r8,rdi
> + sub rsp,72
> + mov r9,rsi
> + and rsp,-64
> + mov r10,rdx
> + mov QWORD[64+rsp],rax
> +
> +$L$prologue:
> +
> + mov esi,DWORD[r8]
> + mov edi,DWORD[4+r8]
> + mov r11d,DWORD[8+r8]
> + mov r12d,DWORD[12+r8]
> + mov r13d,DWORD[16+r8]
> + jmp NEAR $L$loop
> +
> +ALIGN 16
> +$L$loop:
> + mov edx,DWORD[r9]
> + bswap edx
> + mov ebp,DWORD[4+r9]
> + mov eax,r12d
> + mov DWORD[rsp],edx
> + mov ecx,esi
> + bswap ebp
> + xor eax,r11d
> + rol ecx,5
> + and eax,edi
> + lea r13d,[1518500249+r13*1+rdx]
> + add r13d,ecx
> + xor eax,r12d
> + rol edi,30
> + add r13d,eax
> + mov r14d,DWORD[8+r9]
> + mov eax,r11d
> + mov DWORD[4+rsp],ebp
> + mov ecx,r13d
> + bswap r14d
> + xor eax,edi
> + rol ecx,5
> + and eax,esi
> + lea r12d,[1518500249+r12*1+rbp]
> + add r12d,ecx
> + xor eax,r11d
> + rol esi,30
> + add r12d,eax
> + mov edx,DWORD[12+r9]
> + mov eax,edi
> + mov DWORD[8+rsp],r14d
> + mov ecx,r12d
> + bswap edx
> + xor eax,esi
> + rol ecx,5
> + and eax,r13d
> + lea r11d,[1518500249+r11*1+r14]
> + add r11d,ecx
> + xor eax,edi
> + rol r13d,30
> + add r11d,eax
> + mov ebp,DWORD[16+r9]
> + mov eax,esi
> + mov DWORD[12+rsp],edx
> + mov ecx,r11d
> + bswap ebp
> + xor eax,r13d
> + rol ecx,5
> + and eax,r12d
> + lea edi,[1518500249+rdi*1+rdx]
> + add edi,ecx
> + xor eax,esi
> + rol r12d,30
> + add edi,eax
> + mov r14d,DWORD[20+r9]
> + mov eax,r13d
> + mov DWORD[16+rsp],ebp
> + mov ecx,edi
> + bswap r14d
> + xor eax,r12d
> + rol ecx,5
> + and eax,r11d
> + lea esi,[1518500249+rsi*1+rbp]
> + add esi,ecx
> + xor eax,r13d
> + rol r11d,30
> + add esi,eax
> + mov edx,DWORD[24+r9]
> + mov eax,r12d
> + mov DWORD[20+rsp],r14d
> + mov ecx,esi
> + bswap edx
> + xor eax,r11d
> + rol ecx,5
> + and eax,edi
> + lea r13d,[1518500249+r13*1+r14]
> + add r13d,ecx
> + xor eax,r12d
> + rol edi,30
> + add r13d,eax
> + mov ebp,DWORD[28+r9]
> + mov eax,r11d
> + mov DWORD[24+rsp],edx
> + mov ecx,r13d
> + bswap ebp
> + xor eax,edi
> + rol ecx,5
> + and eax,esi
> + lea r12d,[1518500249+r12*1+rdx]
> + add r12d,ecx
> + xor eax,r11d
> + rol esi,30
> + add r12d,eax
> + mov r14d,DWORD[32+r9]
> + mov eax,edi
> + mov DWORD[28+rsp],ebp
> + mov ecx,r12d
> + bswap r14d
> + xor eax,esi
> + rol ecx,5
> + and eax,r13d
> + lea r11d,[1518500249+r11*1+rbp]
> + add r11d,ecx
> + xor eax,edi
> + rol r13d,30
> + add r11d,eax
> + mov edx,DWORD[36+r9]
> + mov eax,esi
> + mov DWORD[32+rsp],r14d
> + mov ecx,r11d
> + bswap edx
> + xor eax,r13d
> + rol ecx,5
> + and eax,r12d
> + lea edi,[1518500249+rdi*1+r14]
> + add edi,ecx
> + xor eax,esi
> + rol r12d,30
> + add edi,eax
> + mov ebp,DWORD[40+r9]
> + mov eax,r13d
> + mov DWORD[36+rsp],edx
> + mov ecx,edi
> + bswap ebp
> + xor eax,r12d
> + rol ecx,5
> + and eax,r11d
> + lea esi,[1518500249+rsi*1+rdx]
> + add esi,ecx
> + xor eax,r13d
> + rol r11d,30
> + add esi,eax
> + mov r14d,DWORD[44+r9]
> + mov eax,r12d
> + mov DWORD[40+rsp],ebp
> + mov ecx,esi
> + bswap r14d
> + xor eax,r11d
> + rol ecx,5
> + and eax,edi
> + lea r13d,[1518500249+r13*1+rbp]
> + add r13d,ecx
> + xor eax,r12d
> + rol edi,30
> + add r13d,eax
> + mov edx,DWORD[48+r9]
> + mov eax,r11d
> + mov DWORD[44+rsp],r14d
> + mov ecx,r13d
> + bswap edx
> + xor eax,edi
> + rol ecx,5
> + and eax,esi
> + lea r12d,[1518500249+r12*1+r14]
> + add r12d,ecx
> + xor eax,r11d
> + rol esi,30
> + add r12d,eax
> + mov ebp,DWORD[52+r9]
> + mov eax,edi
> + mov DWORD[48+rsp],edx
> + mov ecx,r12d
> + bswap ebp
> + xor eax,esi
> + rol ecx,5
> + and eax,r13d
> + lea r11d,[1518500249+r11*1+rdx]
> + add r11d,ecx
> + xor eax,edi
> + rol r13d,30
> + add r11d,eax
> + mov r14d,DWORD[56+r9]
> + mov eax,esi
> + mov DWORD[52+rsp],ebp
> + mov ecx,r11d
> + bswap r14d
> + xor eax,r13d
> + rol ecx,5
> + and eax,r12d
> + lea edi,[1518500249+rdi*1+rbp]
> + add edi,ecx
> + xor eax,esi
> + rol r12d,30
> + add edi,eax
> + mov edx,DWORD[60+r9]
> + mov eax,r13d
> + mov DWORD[56+rsp],r14d
> + mov ecx,edi
> + bswap edx
> + xor eax,r12d
> + rol ecx,5
> + and eax,r11d
> + lea esi,[1518500249+rsi*1+r14]
> + add esi,ecx
> + xor eax,r13d
> + rol r11d,30
> + add esi,eax
> + xor ebp,DWORD[rsp]
> + mov eax,r12d
> + mov DWORD[60+rsp],edx
> + mov ecx,esi
> + xor ebp,DWORD[8+rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor ebp,DWORD[32+rsp]
> + and eax,edi
> + lea r13d,[1518500249+r13*1+rdx]
> + rol edi,30
> + xor eax,r12d
> + add r13d,ecx
> + rol ebp,1
> + add r13d,eax
> + xor r14d,DWORD[4+rsp]
> + mov eax,r11d
> + mov DWORD[rsp],ebp
> + mov ecx,r13d
> + xor r14d,DWORD[12+rsp]
> + xor eax,edi
> + rol ecx,5
> + xor r14d,DWORD[36+rsp]
> + and eax,esi
> + lea r12d,[1518500249+r12*1+rbp]
> + rol esi,30
> + xor eax,r11d
> + add r12d,ecx
> + rol r14d,1
> + add r12d,eax
> + xor edx,DWORD[8+rsp]
> + mov eax,edi
> + mov DWORD[4+rsp],r14d
> + mov ecx,r12d
> + xor edx,DWORD[16+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor edx,DWORD[40+rsp]
> + and eax,r13d
> + lea r11d,[1518500249+r11*1+r14]
> + rol r13d,30
> + xor eax,edi
> + add r11d,ecx
> + rol edx,1
> + add r11d,eax
> + xor ebp,DWORD[12+rsp]
> + mov eax,esi
> + mov DWORD[8+rsp],edx
> + mov ecx,r11d
> + xor ebp,DWORD[20+rsp]
> + xor eax,r13d
> + rol ecx,5
> + xor ebp,DWORD[44+rsp]
> + and eax,r12d
> + lea edi,[1518500249+rdi*1+rdx]
> + rol r12d,30
> + xor eax,esi
> + add edi,ecx
> + rol ebp,1
> + add edi,eax
> + xor r14d,DWORD[16+rsp]
> + mov eax,r13d
> + mov DWORD[12+rsp],ebp
> + mov ecx,edi
> + xor r14d,DWORD[24+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor r14d,DWORD[48+rsp]
> + and eax,r11d
> + lea esi,[1518500249+rsi*1+rbp]
> + rol r11d,30
> + xor eax,r13d
> + add esi,ecx
> + rol r14d,1
> + add esi,eax
> + xor edx,DWORD[20+rsp]
> + mov eax,edi
> + mov DWORD[16+rsp],r14d
> + mov ecx,esi
> + xor edx,DWORD[28+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor edx,DWORD[52+rsp]
> + lea r13d,[1859775393+r13*1+r14]
> + xor eax,r11d
> + add r13d,ecx
> + rol edi,30
> + add r13d,eax
> + rol edx,1
> + xor ebp,DWORD[24+rsp]
> + mov eax,esi
> + mov DWORD[20+rsp],edx
> + mov ecx,r13d
> + xor ebp,DWORD[32+rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor ebp,DWORD[56+rsp]
> + lea r12d,[1859775393+r12*1+rdx]
> + xor eax,edi
> + add r12d,ecx
> + rol esi,30
> + add r12d,eax
> + rol ebp,1
> + xor r14d,DWORD[28+rsp]
> + mov eax,r13d
> + mov DWORD[24+rsp],ebp
> + mov ecx,r12d
> + xor r14d,DWORD[36+rsp]
> + xor eax,edi
> + rol ecx,5
> + xor r14d,DWORD[60+rsp]
> + lea r11d,[1859775393+r11*1+rbp]
> + xor eax,esi
> + add r11d,ecx
> + rol r13d,30
> + add r11d,eax
> + rol r14d,1
> + xor edx,DWORD[32+rsp]
> + mov eax,r12d
> + mov DWORD[28+rsp],r14d
> + mov ecx,r11d
> + xor edx,DWORD[40+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor edx,DWORD[rsp]
> + lea edi,[1859775393+rdi*1+r14]
> + xor eax,r13d
> + add edi,ecx
> + rol r12d,30
> + add edi,eax
> + rol edx,1
> + xor ebp,DWORD[36+rsp]
> + mov eax,r11d
> + mov DWORD[32+rsp],edx
> + mov ecx,edi
> + xor ebp,DWORD[44+rsp]
> + xor eax,r13d
> + rol ecx,5
> + xor ebp,DWORD[4+rsp]
> + lea esi,[1859775393+rsi*1+rdx]
> + xor eax,r12d
> + add esi,ecx
> + rol r11d,30
> + add esi,eax
> + rol ebp,1
> + xor r14d,DWORD[40+rsp]
> + mov eax,edi
> + mov DWORD[36+rsp],ebp
> + mov ecx,esi
> + xor r14d,DWORD[48+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor r14d,DWORD[8+rsp]
> + lea r13d,[1859775393+r13*1+rbp]
> + xor eax,r11d
> + add r13d,ecx
> + rol edi,30
> + add r13d,eax
> + rol r14d,1
> + xor edx,DWORD[44+rsp]
> + mov eax,esi
> + mov DWORD[40+rsp],r14d
> + mov ecx,r13d
> + xor edx,DWORD[52+rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor edx,DWORD[12+rsp]
> + lea r12d,[1859775393+r12*1+r14]
> + xor eax,edi
> + add r12d,ecx
> + rol esi,30
> + add r12d,eax
> + rol edx,1
> + xor ebp,DWORD[48+rsp]
> + mov eax,r13d
> + mov DWORD[44+rsp],edx
> + mov ecx,r12d
> + xor ebp,DWORD[56+rsp]
> + xor eax,edi
> + rol ecx,5
> + xor ebp,DWORD[16+rsp]
> + lea r11d,[1859775393+r11*1+rdx]
> + xor eax,esi
> + add r11d,ecx
> + rol r13d,30
> + add r11d,eax
> + rol ebp,1
> + xor r14d,DWORD[52+rsp]
> + mov eax,r12d
> + mov DWORD[48+rsp],ebp
> + mov ecx,r11d
> + xor r14d,DWORD[60+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor r14d,DWORD[20+rsp]
> + lea edi,[1859775393+rdi*1+rbp]
> + xor eax,r13d
> + add edi,ecx
> + rol r12d,30
> + add edi,eax
> + rol r14d,1
> + xor edx,DWORD[56+rsp]
> + mov eax,r11d
> + mov DWORD[52+rsp],r14d
> + mov ecx,edi
> + xor edx,DWORD[rsp]
> + xor eax,r13d
> + rol ecx,5
> + xor edx,DWORD[24+rsp]
> + lea esi,[1859775393+rsi*1+r14]
> + xor eax,r12d
> + add esi,ecx
> + rol r11d,30
> + add esi,eax
> + rol edx,1
> + xor ebp,DWORD[60+rsp]
> + mov eax,edi
> + mov DWORD[56+rsp],edx
> + mov ecx,esi
> + xor ebp,DWORD[4+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor ebp,DWORD[28+rsp]
> + lea r13d,[1859775393+r13*1+rdx]
> + xor eax,r11d
> + add r13d,ecx
> + rol edi,30
> + add r13d,eax
> + rol ebp,1
> + xor r14d,DWORD[rsp]
> + mov eax,esi
> + mov DWORD[60+rsp],ebp
> + mov ecx,r13d
> + xor r14d,DWORD[8+rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor r14d,DWORD[32+rsp]
> + lea r12d,[1859775393+r12*1+rbp]
> + xor eax,edi
> + add r12d,ecx
> + rol esi,30
> + add r12d,eax
> + rol r14d,1
> + xor edx,DWORD[4+rsp]
> + mov eax,r13d
> + mov DWORD[rsp],r14d
> + mov ecx,r12d
> + xor edx,DWORD[12+rsp]
> + xor eax,edi
> + rol ecx,5
> + xor edx,DWORD[36+rsp]
> + lea r11d,[1859775393+r11*1+r14]
> + xor eax,esi
> + add r11d,ecx
> + rol r13d,30
> + add r11d,eax
> + rol edx,1
> + xor ebp,DWORD[8+rsp]
> + mov eax,r12d
> + mov DWORD[4+rsp],edx
> + mov ecx,r11d
> + xor ebp,DWORD[16+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor ebp,DWORD[40+rsp]
> + lea edi,[1859775393+rdi*1+rdx]
> + xor eax,r13d
> + add edi,ecx
> + rol r12d,30
> + add edi,eax
> + rol ebp,1
> + xor r14d,DWORD[12+rsp]
> + mov eax,r11d
> + mov DWORD[8+rsp],ebp
> + mov ecx,edi
> + xor r14d,DWORD[20+rsp]
> + xor eax,r13d
> + rol ecx,5
> + xor r14d,DWORD[44+rsp]
> + lea esi,[1859775393+rsi*1+rbp]
> + xor eax,r12d
> + add esi,ecx
> + rol r11d,30
> + add esi,eax
> + rol r14d,1
> + xor edx,DWORD[16+rsp]
> + mov eax,edi
> + mov DWORD[12+rsp],r14d
> + mov ecx,esi
> + xor edx,DWORD[24+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor edx,DWORD[48+rsp]
> + lea r13d,[1859775393+r13*1+r14]
> + xor eax,r11d
> + add r13d,ecx
> + rol edi,30
> + add r13d,eax
> + rol edx,1
> + xor ebp,DWORD[20+rsp]
> + mov eax,esi
> + mov DWORD[16+rsp],edx
> + mov ecx,r13d
> + xor ebp,DWORD[28+rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor ebp,DWORD[52+rsp]
> + lea r12d,[1859775393+r12*1+rdx]
> + xor eax,edi
> + add r12d,ecx
> + rol esi,30
> + add r12d,eax
> + rol ebp,1
> + xor r14d,DWORD[24+rsp]
> + mov eax,r13d
> + mov DWORD[20+rsp],ebp
> + mov ecx,r12d
> + xor r14d,DWORD[32+rsp]
> + xor eax,edi
> + rol ecx,5
> + xor r14d,DWORD[56+rsp]
> + lea r11d,[1859775393+r11*1+rbp]
> + xor eax,esi
> + add r11d,ecx
> + rol r13d,30
> + add r11d,eax
> + rol r14d,1
> + xor edx,DWORD[28+rsp]
> + mov eax,r12d
> + mov DWORD[24+rsp],r14d
> + mov ecx,r11d
> + xor edx,DWORD[36+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor edx,DWORD[60+rsp]
> + lea edi,[1859775393+rdi*1+r14]
> + xor eax,r13d
> + add edi,ecx
> + rol r12d,30
> + add edi,eax
> + rol edx,1
> + xor ebp,DWORD[32+rsp]
> + mov eax,r11d
> + mov DWORD[28+rsp],edx
> + mov ecx,edi
> + xor ebp,DWORD[40+rsp]
> + xor eax,r13d
> + rol ecx,5
> + xor ebp,DWORD[rsp]
> + lea esi,[1859775393+rsi*1+rdx]
> + xor eax,r12d
> + add esi,ecx
> + rol r11d,30
> + add esi,eax
> + rol ebp,1
> + xor r14d,DWORD[36+rsp]
> + mov eax,r12d
> + mov DWORD[32+rsp],ebp
> + mov ebx,r12d
> + xor r14d,DWORD[44+rsp]
> + and eax,r11d
> + mov ecx,esi
> + xor r14d,DWORD[4+rsp]
> + lea r13d,[((-1894007588))+r13*1+rbp]
> + xor ebx,r11d
> + rol ecx,5
> + add r13d,eax
> + rol r14d,1
> + and ebx,edi
> + add r13d,ecx
> + rol edi,30
> + add r13d,ebx
> + xor edx,DWORD[40+rsp]
> + mov eax,r11d
> + mov DWORD[36+rsp],r14d
> + mov ebx,r11d
> + xor edx,DWORD[48+rsp]
> + and eax,edi
> + mov ecx,r13d
> + xor edx,DWORD[8+rsp]
> + lea r12d,[((-1894007588))+r12*1+r14]
> + xor ebx,edi
> + rol ecx,5
> + add r12d,eax
> + rol edx,1
> + and ebx,esi
> + add r12d,ecx
> + rol esi,30
> + add r12d,ebx
> + xor ebp,DWORD[44+rsp]
> + mov eax,edi
> + mov DWORD[40+rsp],edx
> + mov ebx,edi
> + xor ebp,DWORD[52+rsp]
> + and eax,esi
> + mov ecx,r12d
> + xor ebp,DWORD[12+rsp]
> + lea r11d,[((-1894007588))+r11*1+rdx]
> + xor ebx,esi
> + rol ecx,5
> + add r11d,eax
> + rol ebp,1
> + and ebx,r13d
> + add r11d,ecx
> + rol r13d,30
> + add r11d,ebx
> + xor r14d,DWORD[48+rsp]
> + mov eax,esi
> + mov DWORD[44+rsp],ebp
> + mov ebx,esi
> + xor r14d,DWORD[56+rsp]
> + and eax,r13d
> + mov ecx,r11d
> + xor r14d,DWORD[16+rsp]
> + lea edi,[((-1894007588))+rdi*1+rbp]
> + xor ebx,r13d
> + rol ecx,5
> + add edi,eax
> + rol r14d,1
> + and ebx,r12d
> + add edi,ecx
> + rol r12d,30
> + add edi,ebx
> + xor edx,DWORD[52+rsp]
> + mov eax,r13d
> + mov DWORD[48+rsp],r14d
> + mov ebx,r13d
> + xor edx,DWORD[60+rsp]
> + and eax,r12d
> + mov ecx,edi
> + xor edx,DWORD[20+rsp]
> + lea esi,[((-1894007588))+rsi*1+r14]
> + xor ebx,r12d
> + rol ecx,5
> + add esi,eax
> + rol edx,1
> + and ebx,r11d
> + add esi,ecx
> + rol r11d,30
> + add esi,ebx
> + xor ebp,DWORD[56+rsp]
> + mov eax,r12d
> + mov DWORD[52+rsp],edx
> + mov ebx,r12d
> + xor ebp,DWORD[rsp]
> + and eax,r11d
> + mov ecx,esi
> + xor ebp,DWORD[24+rsp]
> + lea r13d,[((-1894007588))+r13*1+rdx]
> + xor ebx,r11d
> + rol ecx,5
> + add r13d,eax
> + rol ebp,1
> + and ebx,edi
> + add r13d,ecx
> + rol edi,30
> + add r13d,ebx
> + xor r14d,DWORD[60+rsp]
> + mov eax,r11d
> + mov DWORD[56+rsp],ebp
> + mov ebx,r11d
> + xor r14d,DWORD[4+rsp]
> + and eax,edi
> + mov ecx,r13d
> + xor r14d,DWORD[28+rsp]
> + lea r12d,[((-1894007588))+r12*1+rbp]
> + xor ebx,edi
> + rol ecx,5
> + add r12d,eax
> + rol r14d,1
> + and ebx,esi
> + add r12d,ecx
> + rol esi,30
> + add r12d,ebx
> + xor edx,DWORD[rsp]
> + mov eax,edi
> + mov DWORD[60+rsp],r14d
> + mov ebx,edi
> + xor edx,DWORD[8+rsp]
> + and eax,esi
> + mov ecx,r12d
> + xor edx,DWORD[32+rsp]
> + lea r11d,[((-1894007588))+r11*1+r14]
> + xor ebx,esi
> + rol ecx,5
> + add r11d,eax
> + rol edx,1
> + and ebx,r13d
> + add r11d,ecx
> + rol r13d,30
> + add r11d,ebx
> + xor ebp,DWORD[4+rsp]
> + mov eax,esi
> + mov DWORD[rsp],edx
> + mov ebx,esi
> + xor ebp,DWORD[12+rsp]
> + and eax,r13d
> + mov ecx,r11d
> + xor ebp,DWORD[36+rsp]
> + lea edi,[((-1894007588))+rdi*1+rdx]
> + xor ebx,r13d
> + rol ecx,5
> + add edi,eax
> + rol ebp,1
> + and ebx,r12d
> + add edi,ecx
> + rol r12d,30
> + add edi,ebx
> + xor r14d,DWORD[8+rsp]
> + mov eax,r13d
> + mov DWORD[4+rsp],ebp
> + mov ebx,r13d
> + xor r14d,DWORD[16+rsp]
> + and eax,r12d
> + mov ecx,edi
> + xor r14d,DWORD[40+rsp]
> + lea esi,[((-1894007588))+rsi*1+rbp]
> + xor ebx,r12d
> + rol ecx,5
> + add esi,eax
> + rol r14d,1
> + and ebx,r11d
> + add esi,ecx
> + rol r11d,30
> + add esi,ebx
> + xor edx,DWORD[12+rsp]
> + mov eax,r12d
> + mov DWORD[8+rsp],r14d
> + mov ebx,r12d
> + xor edx,DWORD[20+rsp]
> + and eax,r11d
> + mov ecx,esi
> + xor edx,DWORD[44+rsp]
> + lea r13d,[((-1894007588))+r13*1+r14]
> + xor ebx,r11d
> + rol ecx,5
> + add r13d,eax
> + rol edx,1
> + and ebx,edi
> + add r13d,ecx
> + rol edi,30
> + add r13d,ebx
> + xor ebp,DWORD[16+rsp]
> + mov eax,r11d
> + mov DWORD[12+rsp],edx
> + mov ebx,r11d
> + xor ebp,DWORD[24+rsp]
> + and eax,edi
> + mov ecx,r13d
> + xor ebp,DWORD[48+rsp]
> + lea r12d,[((-1894007588))+r12*1+rdx]
> + xor ebx,edi
> + rol ecx,5
> + add r12d,eax
> + rol ebp,1
> + and ebx,esi
> + add r12d,ecx
> + rol esi,30
> + add r12d,ebx
> + xor r14d,DWORD[20+rsp]
> + mov eax,edi
> + mov DWORD[16+rsp],ebp
> + mov ebx,edi
> + xor r14d,DWORD[28+rsp]
> + and eax,esi
> + mov ecx,r12d
> + xor r14d,DWORD[52+rsp]
> + lea r11d,[((-1894007588))+r11*1+rbp]
> + xor ebx,esi
> + rol ecx,5
> + add r11d,eax
> + rol r14d,1
> + and ebx,r13d
> + add r11d,ecx
> + rol r13d,30
> + add r11d,ebx
> + xor edx,DWORD[24+rsp]
> + mov eax,esi
> + mov DWORD[20+rsp],r14d
> + mov ebx,esi
> + xor edx,DWORD[32+rsp]
> + and eax,r13d
> + mov ecx,r11d
> + xor edx,DWORD[56+rsp]
> + lea edi,[((-1894007588))+rdi*1+r14]
> + xor ebx,r13d
> + rol ecx,5
> + add edi,eax
> + rol edx,1
> + and ebx,r12d
> + add edi,ecx
> + rol r12d,30
> + add edi,ebx
> + xor ebp,DWORD[28+rsp]
> + mov eax,r13d
> + mov DWORD[24+rsp],edx
> + mov ebx,r13d
> + xor ebp,DWORD[36+rsp]
> + and eax,r12d
> + mov ecx,edi
> + xor ebp,DWORD[60+rsp]
> + lea esi,[((-1894007588))+rsi*1+rdx]
> + xor ebx,r12d
> + rol ecx,5
> + add esi,eax
> + rol ebp,1
> + and ebx,r11d
> + add esi,ecx
> + rol r11d,30
> + add esi,ebx
> + xor r14d,DWORD[32+rsp]
> + mov eax,r12d
> + mov DWORD[28+rsp],ebp
> + mov ebx,r12d
> + xor r14d,DWORD[40+rsp]
> + and eax,r11d
> + mov ecx,esi
> + xor r14d,DWORD[rsp]
> + lea r13d,[((-1894007588))+r13*1+rbp]
> + xor ebx,r11d
> + rol ecx,5
> + add r13d,eax
> + rol r14d,1
> + and ebx,edi
> + add r13d,ecx
> + rol edi,30
> + add r13d,ebx
> + xor edx,DWORD[36+rsp]
> + mov eax,r11d
> + mov DWORD[32+rsp],r14d
> + mov ebx,r11d
> + xor edx,DWORD[44+rsp]
> + and eax,edi
> + mov ecx,r13d
> + xor edx,DWORD[4+rsp]
> + lea r12d,[((-1894007588))+r12*1+r14]
> + xor ebx,edi
> + rol ecx,5
> + add r12d,eax
> + rol edx,1
> + and ebx,esi
> + add r12d,ecx
> + rol esi,30
> + add r12d,ebx
> + xor ebp,DWORD[40+rsp]
> + mov eax,edi
> + mov DWORD[36+rsp],edx
> + mov ebx,edi
> + xor ebp,DWORD[48+rsp]
> + and eax,esi
> + mov ecx,r12d
> + xor ebp,DWORD[8+rsp]
> + lea r11d,[((-1894007588))+r11*1+rdx]
> + xor ebx,esi
> + rol ecx,5
> + add r11d,eax
> + rol ebp,1
> + and ebx,r13d
> + add r11d,ecx
> + rol r13d,30
> + add r11d,ebx
> + xor r14d,DWORD[44+rsp]
> + mov eax,esi
> + mov DWORD[40+rsp],ebp
> + mov ebx,esi
> + xor r14d,DWORD[52+rsp]
> + and eax,r13d
> + mov ecx,r11d
> + xor r14d,DWORD[12+rsp]
> + lea edi,[((-1894007588))+rdi*1+rbp]
> + xor ebx,r13d
> + rol ecx,5
> + add edi,eax
> + rol r14d,1
> + and ebx,r12d
> + add edi,ecx
> + rol r12d,30
> + add edi,ebx
> + xor edx,DWORD[48+rsp]
> + mov eax,r13d
> + mov DWORD[44+rsp],r14d
> + mov ebx,r13d
> + xor edx,DWORD[56+rsp]
> + and eax,r12d
> + mov ecx,edi
> + xor edx,DWORD[16+rsp]
> + lea esi,[((-1894007588))+rsi*1+r14]
> + xor ebx,r12d
> + rol ecx,5
> + add esi,eax
> + rol edx,1
> + and ebx,r11d
> + add esi,ecx
> + rol r11d,30
> + add esi,ebx
> + xor ebp,DWORD[52+rsp]
> + mov eax,edi
> + mov DWORD[48+rsp],edx
> + mov ecx,esi
> + xor ebp,DWORD[60+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor ebp,DWORD[20+rsp]
> + lea r13d,[((-899497514))+r13*1+rdx]
> + xor eax,r11d
> + add r13d,ecx
> + rol edi,30
> + add r13d,eax
> + rol ebp,1
> + xor r14d,DWORD[56+rsp]
> + mov eax,esi
> + mov DWORD[52+rsp],ebp
> + mov ecx,r13d
> + xor r14d,DWORD[rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor r14d,DWORD[24+rsp]
> + lea r12d,[((-899497514))+r12*1+rbp]
> + xor eax,edi
> + add r12d,ecx
> + rol esi,30
> + add r12d,eax
> + rol r14d,1
> + xor edx,DWORD[60+rsp]
> + mov eax,r13d
> + mov DWORD[56+rsp],r14d
> + mov ecx,r12d
> + xor edx,DWORD[4+rsp]
> + xor eax,edi
> + rol ecx,5
> + xor edx,DWORD[28+rsp]
> + lea r11d,[((-899497514))+r11*1+r14]
> + xor eax,esi
> + add r11d,ecx
> + rol r13d,30
> + add r11d,eax
> + rol edx,1
> + xor ebp,DWORD[rsp]
> + mov eax,r12d
> + mov DWORD[60+rsp],edx
> + mov ecx,r11d
> + xor ebp,DWORD[8+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor ebp,DWORD[32+rsp]
> + lea edi,[((-899497514))+rdi*1+rdx]
> + xor eax,r13d
> + add edi,ecx
> + rol r12d,30
> + add edi,eax
> + rol ebp,1
> + xor r14d,DWORD[4+rsp]
> + mov eax,r11d
> + mov DWORD[rsp],ebp
> + mov ecx,edi
> + xor r14d,DWORD[12+rsp]
> + xor eax,r13d
> + rol ecx,5
> + xor r14d,DWORD[36+rsp]
> + lea esi,[((-899497514))+rsi*1+rbp]
> + xor eax,r12d
> + add esi,ecx
> + rol r11d,30
> + add esi,eax
> + rol r14d,1
> + xor edx,DWORD[8+rsp]
> + mov eax,edi
> + mov DWORD[4+rsp],r14d
> + mov ecx,esi
> + xor edx,DWORD[16+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor edx,DWORD[40+rsp]
> + lea r13d,[((-899497514))+r13*1+r14]
> + xor eax,r11d
> + add r13d,ecx
> + rol edi,30
> + add r13d,eax
> + rol edx,1
> + xor ebp,DWORD[12+rsp]
> + mov eax,esi
> + mov DWORD[8+rsp],edx
> + mov ecx,r13d
> + xor ebp,DWORD[20+rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor ebp,DWORD[44+rsp]
> + lea r12d,[((-899497514))+r12*1+rdx]
> + xor eax,edi
> + add r12d,ecx
> + rol esi,30
> + add r12d,eax
> + rol ebp,1
> + xor r14d,DWORD[16+rsp]
> + mov eax,r13d
> + mov DWORD[12+rsp],ebp
> + mov ecx,r12d
> + xor r14d,DWORD[24+rsp]
> + xor eax,edi
> + rol ecx,5
> + xor r14d,DWORD[48+rsp]
> + lea r11d,[((-899497514))+r11*1+rbp]
> + xor eax,esi
> + add r11d,ecx
> + rol r13d,30
> + add r11d,eax
> + rol r14d,1
> + xor edx,DWORD[20+rsp]
> + mov eax,r12d
> + mov DWORD[16+rsp],r14d
> + mov ecx,r11d
> + xor edx,DWORD[28+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor edx,DWORD[52+rsp]
> + lea edi,[((-899497514))+rdi*1+r14]
> + xor eax,r13d
> + add edi,ecx
> + rol r12d,30
> + add edi,eax
> + rol edx,1
> + xor ebp,DWORD[24+rsp]
> + mov eax,r11d
> + mov DWORD[20+rsp],edx
> + mov ecx,edi
> + xor ebp,DWORD[32+rsp]
> + xor eax,r13d
> + rol ecx,5
> + xor ebp,DWORD[56+rsp]
> + lea esi,[((-899497514))+rsi*1+rdx]
> + xor eax,r12d
> + add esi,ecx
> + rol r11d,30
> + add esi,eax
> + rol ebp,1
> + xor r14d,DWORD[28+rsp]
> + mov eax,edi
> + mov DWORD[24+rsp],ebp
> + mov ecx,esi
> + xor r14d,DWORD[36+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor r14d,DWORD[60+rsp]
> + lea r13d,[((-899497514))+r13*1+rbp]
> + xor eax,r11d
> + add r13d,ecx
> + rol edi,30
> + add r13d,eax
> + rol r14d,1
> + xor edx,DWORD[32+rsp]
> + mov eax,esi
> + mov DWORD[28+rsp],r14d
> + mov ecx,r13d
> + xor edx,DWORD[40+rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor edx,DWORD[rsp]
> + lea r12d,[((-899497514))+r12*1+r14]
> + xor eax,edi
> + add r12d,ecx
> + rol esi,30
> + add r12d,eax
> + rol edx,1
> + xor ebp,DWORD[36+rsp]
> + mov eax,r13d
> +
> + mov ecx,r12d
> + xor ebp,DWORD[44+rsp]
> + xor eax,edi
> + rol ecx,5
> + xor ebp,DWORD[4+rsp]
> + lea r11d,[((-899497514))+r11*1+rdx]
> + xor eax,esi
> + add r11d,ecx
> + rol r13d,30
> + add r11d,eax
> + rol ebp,1
> + xor r14d,DWORD[40+rsp]
> + mov eax,r12d
> +
> + mov ecx,r11d
> + xor r14d,DWORD[48+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor r14d,DWORD[8+rsp]
> + lea edi,[((-899497514))+rdi*1+rbp]
> + xor eax,r13d
> + add edi,ecx
> + rol r12d,30
> + add edi,eax
> + rol r14d,1
> + xor edx,DWORD[44+rsp]
> + mov eax,r11d
> +
> + mov ecx,edi
> + xor edx,DWORD[52+rsp]
> + xor eax,r13d
> + rol ecx,5
> + xor edx,DWORD[12+rsp]
> + lea esi,[((-899497514))+rsi*1+r14]
> + xor eax,r12d
> + add esi,ecx
> + rol r11d,30
> + add esi,eax
> + rol edx,1
> + xor ebp,DWORD[48+rsp]
> + mov eax,edi
> +
> + mov ecx,esi
> + xor ebp,DWORD[56+rsp]
> + xor eax,r12d
> + rol ecx,5
> + xor ebp,DWORD[16+rsp]
> + lea r13d,[((-899497514))+r13*1+rdx]
> + xor eax,r11d
> + add r13d,ecx
> + rol edi,30
> + add r13d,eax
> + rol ebp,1
> + xor r14d,DWORD[52+rsp]
> + mov eax,esi
> +
> + mov ecx,r13d
> + xor r14d,DWORD[60+rsp]
> + xor eax,r11d
> + rol ecx,5
> + xor r14d,DWORD[20+rsp]
> + lea r12d,[((-899497514))+r12*1+rbp]
> + xor eax,edi
> + add r12d,ecx
> + rol esi,30
> + add r12d,eax
> + rol r14d,1
> + xor edx,DWORD[56+rsp]
> + mov eax,r13d
> +
> + mov ecx,r12d
> + xor edx,DWORD[rsp]
> + xor eax,edi
> + rol ecx,5
> + xor edx,DWORD[24+rsp]
> + lea r11d,[((-899497514))+r11*1+r14]
> + xor eax,esi
> + add r11d,ecx
> + rol r13d,30
> + add r11d,eax
> + rol edx,1
> + xor ebp,DWORD[60+rsp]
> + mov eax,r12d
> +
> + mov ecx,r11d
> + xor ebp,DWORD[4+rsp]
> + xor eax,esi
> + rol ecx,5
> + xor ebp,DWORD[28+rsp]
> + lea edi,[((-899497514))+rdi*1+rdx]
> + xor eax,r13d
> + add edi,ecx
> + rol r12d,30
> + add edi,eax
> + rol ebp,1
> + mov eax,r11d
> + mov ecx,edi
> + xor eax,r13d
> + lea esi,[((-899497514))+rsi*1+rbp]
> + rol ecx,5
> + xor eax,r12d
> + add esi,ecx
> + rol r11d,30
> + add esi,eax
> + add esi,DWORD[r8]
> + add edi,DWORD[4+r8]
> + add r11d,DWORD[8+r8]
> + add r12d,DWORD[12+r8]
> + add r13d,DWORD[16+r8]
> + mov DWORD[r8],esi
> + mov DWORD[4+r8],edi
> + mov DWORD[8+r8],r11d
> + mov DWORD[12+r8],r12d
> + mov DWORD[16+r8],r13d
> +
> + sub r10,1
> + lea r9,[64+r9]
> + jnz NEAR $L$loop
> +
> + mov rsi,QWORD[64+rsp]
> +
> + mov r14,QWORD[((-40))+rsi]
> +
> + mov r13,QWORD[((-32))+rsi]
> +
> + mov r12,QWORD[((-24))+rsi]
> +
> + mov rbp,QWORD[((-16))+rsi]
> +
> + mov rbx,QWORD[((-8))+rsi]
> +
> + lea rsp,[rsi]
> +
> +$L$epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha1_block_data_order:
> +
> +ALIGN 32
> +sha1_block_data_order_shaext:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha1_block_data_order_shaext:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +_shaext_shortcut:
> +
> + lea rsp,[((-72))+rsp]
> + movaps XMMWORD[(-8-64)+rax],xmm6
> + movaps XMMWORD[(-8-48)+rax],xmm7
> + movaps XMMWORD[(-8-32)+rax],xmm8
> + movaps XMMWORD[(-8-16)+rax],xmm9
> +$L$prologue_shaext:
> + movdqu xmm0,XMMWORD[rdi]
> + movd xmm1,DWORD[16+rdi]
> + movdqa xmm3,XMMWORD[((K_XX_XX+160))]
> +
> + movdqu xmm4,XMMWORD[rsi]
> + pshufd xmm0,xmm0,27
> + movdqu xmm5,XMMWORD[16+rsi]
> + pshufd xmm1,xmm1,27
> + movdqu xmm6,XMMWORD[32+rsi]
> +DB 102,15,56,0,227
> + movdqu xmm7,XMMWORD[48+rsi]
> +DB 102,15,56,0,235
> +DB 102,15,56,0,243
> + movdqa xmm9,xmm1
> +DB 102,15,56,0,251
> + jmp NEAR $L$oop_shaext
> +
> +ALIGN 16
> +$L$oop_shaext:
> + dec rdx
> + lea r8,[64+rsi]
> + paddd xmm1,xmm4
> + cmovne rsi,r8
> + movdqa xmm8,xmm0
> +DB 15,56,201,229
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,0
> +DB 15,56,200,213
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> +DB 15,56,202,231
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,0
> +DB 15,56,200,206
> + pxor xmm5,xmm7
> +DB 15,56,202,236
> +DB 15,56,201,247
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,0
> +DB 15,56,200,215
> + pxor xmm6,xmm4
> +DB 15,56,201,252
> +DB 15,56,202,245
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,0
> +DB 15,56,200,204
> + pxor xmm7,xmm5
> +DB 15,56,202,254
> +DB 15,56,201,229
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,0
> +DB 15,56,200,213
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> +DB 15,56,202,231
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,1
> +DB 15,56,200,206
> + pxor xmm5,xmm7
> +DB 15,56,202,236
> +DB 15,56,201,247
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,1
> +DB 15,56,200,215
> + pxor xmm6,xmm4
> +DB 15,56,201,252
> +DB 15,56,202,245
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,1
> +DB 15,56,200,204
> + pxor xmm7,xmm5
> +DB 15,56,202,254
> +DB 15,56,201,229
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,1
> +DB 15,56,200,213
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> +DB 15,56,202,231
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,1
> +DB 15,56,200,206
> + pxor xmm5,xmm7
> +DB 15,56,202,236
> +DB 15,56,201,247
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,2
> +DB 15,56,200,215
> + pxor xmm6,xmm4
> +DB 15,56,201,252
> +DB 15,56,202,245
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,2
> +DB 15,56,200,204
> + pxor xmm7,xmm5
> +DB 15,56,202,254
> +DB 15,56,201,229
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,2
> +DB 15,56,200,213
> + pxor xmm4,xmm6
> +DB 15,56,201,238
> +DB 15,56,202,231
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,2
> +DB 15,56,200,206
> + pxor xmm5,xmm7
> +DB 15,56,202,236
> +DB 15,56,201,247
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,2
> +DB 15,56,200,215
> + pxor xmm6,xmm4
> +DB 15,56,201,252
> +DB 15,56,202,245
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,3
> +DB 15,56,200,204
> + pxor xmm7,xmm5
> +DB 15,56,202,254
> + movdqu xmm4,XMMWORD[rsi]
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,3
> +DB 15,56,200,213
> + movdqu xmm5,XMMWORD[16+rsi]
> +DB 102,15,56,0,227
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,3
> +DB 15,56,200,206
> + movdqu xmm6,XMMWORD[32+rsi]
> +DB 102,15,56,0,235
> +
> + movdqa xmm2,xmm0
> +DB 15,58,204,193,3
> +DB 15,56,200,215
> + movdqu xmm7,XMMWORD[48+rsi]
> +DB 102,15,56,0,243
> +
> + movdqa xmm1,xmm0
> +DB 15,58,204,194,3
> +DB 65,15,56,200,201
> +DB 102,15,56,0,251
> +
> + paddd xmm0,xmm8
> + movdqa xmm9,xmm1
> +
> + jnz NEAR $L$oop_shaext
> +
> + pshufd xmm0,xmm0,27
> + pshufd xmm1,xmm1,27
> + movdqu XMMWORD[rdi],xmm0
> + movd DWORD[16+rdi],xmm1
> + movaps xmm6,XMMWORD[((-8-64))+rax]
> + movaps xmm7,XMMWORD[((-8-48))+rax]
> + movaps xmm8,XMMWORD[((-8-32))+rax]
> + movaps xmm9,XMMWORD[((-8-16))+rax]
> + mov rsp,rax
> +$L$epilogue_shaext:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha1_block_data_order_shaext:
> +
> +ALIGN 16
> +sha1_block_data_order_ssse3:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha1_block_data_order_ssse3:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +_ssse3_shortcut:
> +
> + mov r11,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + lea rsp,[((-160))+rsp]
> + movaps XMMWORD[(-40-96)+r11],xmm6
> + movaps XMMWORD[(-40-80)+r11],xmm7
> + movaps XMMWORD[(-40-64)+r11],xmm8
> + movaps XMMWORD[(-40-48)+r11],xmm9
> + movaps XMMWORD[(-40-32)+r11],xmm10
> + movaps XMMWORD[(-40-16)+r11],xmm11
> +$L$prologue_ssse3:
> + and rsp,-64
> + mov r8,rdi
> + mov r9,rsi
> + mov r10,rdx
> +
> + shl r10,6
> + add r10,r9
> + lea r14,[((K_XX_XX+64))]
> +
> + mov eax,DWORD[r8]
> + mov ebx,DWORD[4+r8]
> + mov ecx,DWORD[8+r8]
> + mov edx,DWORD[12+r8]
> + mov esi,ebx
> + mov ebp,DWORD[16+r8]
> + mov edi,ecx
> + xor edi,edx
> + and esi,edi
> +
> + movdqa xmm6,XMMWORD[64+r14]
> + movdqa xmm9,XMMWORD[((-64))+r14]
> + movdqu xmm0,XMMWORD[r9]
> + movdqu xmm1,XMMWORD[16+r9]
> + movdqu xmm2,XMMWORD[32+r9]
> + movdqu xmm3,XMMWORD[48+r9]
> +DB 102,15,56,0,198
> +DB 102,15,56,0,206
> +DB 102,15,56,0,214
> + add r9,64
> + paddd xmm0,xmm9
> +DB 102,15,56,0,222
> + paddd xmm1,xmm9
> + paddd xmm2,xmm9
> + movdqa XMMWORD[rsp],xmm0
> + psubd xmm0,xmm9
> + movdqa XMMWORD[16+rsp],xmm1
> + psubd xmm1,xmm9
> + movdqa XMMWORD[32+rsp],xmm2
> + psubd xmm2,xmm9
> + jmp NEAR $L$oop_ssse3
> +ALIGN 16
> +$L$oop_ssse3:
> + ror ebx,2
> + pshufd xmm4,xmm0,238
> + xor esi,edx
> + movdqa xmm8,xmm3
> + paddd xmm9,xmm3
> + mov edi,eax
> + add ebp,DWORD[rsp]
> + punpcklqdq xmm4,xmm1
> + xor ebx,ecx
> + rol eax,5
> + add ebp,esi
> + psrldq xmm8,4
> + and edi,ebx
> + xor ebx,ecx
> + pxor xmm4,xmm0
> + add ebp,eax
> + ror eax,7
> + pxor xmm8,xmm2
> + xor edi,ecx
> + mov esi,ebp
> + add edx,DWORD[4+rsp]
> + pxor xmm4,xmm8
> + xor eax,ebx
> + rol ebp,5
> + movdqa XMMWORD[48+rsp],xmm9
> + add edx,edi
> + and esi,eax
> + movdqa xmm10,xmm4
> + xor eax,ebx
> + add edx,ebp
> + ror ebp,7
> + movdqa xmm8,xmm4
> + xor esi,ebx
> + pslldq xmm10,12
> + paddd xmm4,xmm4
> + mov edi,edx
> + add ecx,DWORD[8+rsp]
> + psrld xmm8,31
> + xor ebp,eax
> + rol edx,5
> + add ecx,esi
> + movdqa xmm9,xmm10
> + and edi,ebp
> + xor ebp,eax
> + psrld xmm10,30
> + add ecx,edx
> + ror edx,7
> + por xmm4,xmm8
> + xor edi,eax
> + mov esi,ecx
> + add ebx,DWORD[12+rsp]
> + pslld xmm9,2
> + pxor xmm4,xmm10
> + xor edx,ebp
> + movdqa xmm10,XMMWORD[((-64))+r14]
> + rol ecx,5
> + add ebx,edi
> + and esi,edx
> + pxor xmm4,xmm9
> + xor edx,ebp
> + add ebx,ecx
> + ror ecx,7
> + pshufd xmm5,xmm1,238
> + xor esi,ebp
> + movdqa xmm9,xmm4
> + paddd xmm10,xmm4
> + mov edi,ebx
> + add eax,DWORD[16+rsp]
> + punpcklqdq xmm5,xmm2
> + xor ecx,edx
> + rol ebx,5
> + add eax,esi
> + psrldq xmm9,4
> + and edi,ecx
> + xor ecx,edx
> + pxor xmm5,xmm1
> + add eax,ebx
> + ror ebx,7
> + pxor xmm9,xmm3
> + xor edi,edx
> + mov esi,eax
> + add ebp,DWORD[20+rsp]
> + pxor xmm5,xmm9
> + xor ebx,ecx
> + rol eax,5
> + movdqa XMMWORD[rsp],xmm10
> + add ebp,edi
> + and esi,ebx
> + movdqa xmm8,xmm5
> + xor ebx,ecx
> + add ebp,eax
> + ror eax,7
> + movdqa xmm9,xmm5
> + xor esi,ecx
> + pslldq xmm8,12
> + paddd xmm5,xmm5
> + mov edi,ebp
> + add edx,DWORD[24+rsp]
> + psrld xmm9,31
> + xor eax,ebx
> + rol ebp,5
> + add edx,esi
> + movdqa xmm10,xmm8
> + and edi,eax
> + xor eax,ebx
> + psrld xmm8,30
> + add edx,ebp
> + ror ebp,7
> + por xmm5,xmm9
> + xor edi,ebx
> + mov esi,edx
> + add ecx,DWORD[28+rsp]
> + pslld xmm10,2
> + pxor xmm5,xmm8
> + xor ebp,eax
> + movdqa xmm8,XMMWORD[((-32))+r14]
> + rol edx,5
> + add ecx,edi
> + and esi,ebp
> + pxor xmm5,xmm10
> + xor ebp,eax
> + add ecx,edx
> + ror edx,7
> + pshufd xmm6,xmm2,238
> + xor esi,eax
> + movdqa xmm10,xmm5
> + paddd xmm8,xmm5
> + mov edi,ecx
> + add ebx,DWORD[32+rsp]
> + punpcklqdq xmm6,xmm3
> + xor edx,ebp
> + rol ecx,5
> + add ebx,esi
> + psrldq xmm10,4
> + and edi,edx
> + xor edx,ebp
> + pxor xmm6,xmm2
> + add ebx,ecx
> + ror ecx,7
> + pxor xmm10,xmm4
> + xor edi,ebp
> + mov esi,ebx
> + add eax,DWORD[36+rsp]
> + pxor xmm6,xmm10
> + xor ecx,edx
> + rol ebx,5
> + movdqa XMMWORD[16+rsp],xmm8
> + add eax,edi
> + and esi,ecx
> + movdqa xmm9,xmm6
> + xor ecx,edx
> + add eax,ebx
> + ror ebx,7
> + movdqa xmm10,xmm6
> + xor esi,edx
> + pslldq xmm9,12
> + paddd xmm6,xmm6
> + mov edi,eax
> + add ebp,DWORD[40+rsp]
> + psrld xmm10,31
> + xor ebx,ecx
> + rol eax,5
> + add ebp,esi
> + movdqa xmm8,xmm9
> + and edi,ebx
> + xor ebx,ecx
> + psrld xmm9,30
> + add ebp,eax
> + ror eax,7
> + por xmm6,xmm10
> + xor edi,ecx
> + mov esi,ebp
> + add edx,DWORD[44+rsp]
> + pslld xmm8,2
> + pxor xmm6,xmm9
> + xor eax,ebx
> + movdqa xmm9,XMMWORD[((-32))+r14]
> + rol ebp,5
> + add edx,edi
> + and esi,eax
> + pxor xmm6,xmm8
> + xor eax,ebx
> + add edx,ebp
> + ror ebp,7
> + pshufd xmm7,xmm3,238
> + xor esi,ebx
> + movdqa xmm8,xmm6
> + paddd xmm9,xmm6
> + mov edi,edx
> + add ecx,DWORD[48+rsp]
> + punpcklqdq xmm7,xmm4
> + xor ebp,eax
> + rol edx,5
> + add ecx,esi
> + psrldq xmm8,4
> + and edi,ebp
> + xor ebp,eax
> + pxor xmm7,xmm3
> + add ecx,edx
> + ror edx,7
> + pxor xmm8,xmm5
> + xor edi,eax
> + mov esi,ecx
> + add ebx,DWORD[52+rsp]
> + pxor xmm7,xmm8
> + xor edx,ebp
> + rol ecx,5
> + movdqa XMMWORD[32+rsp],xmm9
> + add ebx,edi
> + and esi,edx
> + movdqa xmm10,xmm7
> + xor edx,ebp
> + add ebx,ecx
> + ror ecx,7
> + movdqa xmm8,xmm7
> + xor esi,ebp
> + pslldq xmm10,12
> + paddd xmm7,xmm7
> + mov edi,ebx
> + add eax,DWORD[56+rsp]
> + psrld xmm8,31
> + xor ecx,edx
> + rol ebx,5
> + add eax,esi
> + movdqa xmm9,xmm10
> + and edi,ecx
> + xor ecx,edx
> + psrld xmm10,30
> + add eax,ebx
> + ror ebx,7
> + por xmm7,xmm8
> + xor edi,edx
> + mov esi,eax
> + add ebp,DWORD[60+rsp]
> + pslld xmm9,2
> + pxor xmm7,xmm10
> + xor ebx,ecx
> + movdqa xmm10,XMMWORD[((-32))+r14]
> + rol eax,5
> + add ebp,edi
> + and esi,ebx
> + pxor xmm7,xmm9
> + pshufd xmm9,xmm6,238
> + xor ebx,ecx
> + add ebp,eax
> + ror eax,7
> + pxor xmm0,xmm4
> + xor esi,ecx
> + mov edi,ebp
> + add edx,DWORD[rsp]
> + punpcklqdq xmm9,xmm7
> + xor eax,ebx
> + rol ebp,5
> + pxor xmm0,xmm1
> + add edx,esi
> + and edi,eax
> + movdqa xmm8,xmm10
> + xor eax,ebx
> + paddd xmm10,xmm7
> + add edx,ebp
> + pxor xmm0,xmm9
> + ror ebp,7
> + xor edi,ebx
> + mov esi,edx
> + add ecx,DWORD[4+rsp]
> + movdqa xmm9,xmm0
> + xor ebp,eax
> + rol edx,5
> + movdqa XMMWORD[48+rsp],xmm10
> + add ecx,edi
> + and esi,ebp
> + xor ebp,eax
> + pslld xmm0,2
> + add ecx,edx
> + ror edx,7
> + psrld xmm9,30
> + xor esi,eax
> + mov edi,ecx
> + add ebx,DWORD[8+rsp]
> + por xmm0,xmm9
> + xor edx,ebp
> + rol ecx,5
> + pshufd xmm10,xmm7,238
> + add ebx,esi
> + and edi,edx
> + xor edx,ebp
> + add ebx,ecx
> + add eax,DWORD[12+rsp]
> + xor edi,ebp
> + mov esi,ebx
> + rol ebx,5
> + add eax,edi
> + xor esi,edx
> + ror ecx,7
> + add eax,ebx
> + pxor xmm1,xmm5
> + add ebp,DWORD[16+rsp]
> + xor esi,ecx
> + punpcklqdq xmm10,xmm0
> + mov edi,eax
> + rol eax,5
> + pxor xmm1,xmm2
> + add ebp,esi
> + xor edi,ecx
> + movdqa xmm9,xmm8
> + ror ebx,7
> + paddd xmm8,xmm0
> + add ebp,eax
> + pxor xmm1,xmm10
> + add edx,DWORD[20+rsp]
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + movdqa xmm10,xmm1
> + add edx,edi
> + xor esi,ebx
> + movdqa XMMWORD[rsp],xmm8
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[24+rsp]
> + pslld xmm1,2
> + xor esi,eax
> + mov edi,edx
> + psrld xmm10,30
> + rol edx,5
> + add ecx,esi
> + xor edi,eax
> + ror ebp,7
> + por xmm1,xmm10
> + add ecx,edx
> + add ebx,DWORD[28+rsp]
> + pshufd xmm8,xmm0,238
> + xor edi,ebp
> + mov esi,ecx
> + rol ecx,5
> + add ebx,edi
> + xor esi,ebp
> + ror edx,7
> + add ebx,ecx
> + pxor xmm2,xmm6
> + add eax,DWORD[32+rsp]
> + xor esi,edx
> + punpcklqdq xmm8,xmm1
> + mov edi,ebx
> + rol ebx,5
> + pxor xmm2,xmm3
> + add eax,esi
> + xor edi,edx
> + movdqa xmm10,XMMWORD[r14]
> + ror ecx,7
> + paddd xmm9,xmm1
> + add eax,ebx
> + pxor xmm2,xmm8
> + add ebp,DWORD[36+rsp]
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + movdqa xmm8,xmm2
> + add ebp,edi
> + xor esi,ecx
> + movdqa XMMWORD[16+rsp],xmm9
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[40+rsp]
> + pslld xmm2,2
> + xor esi,ebx
> + mov edi,ebp
> + psrld xmm8,30
> + rol ebp,5
> + add edx,esi
> + xor edi,ebx
> + ror eax,7
> + por xmm2,xmm8
> + add edx,ebp
> + add ecx,DWORD[44+rsp]
> + pshufd xmm9,xmm1,238
> + xor edi,eax
> + mov esi,edx
> + rol edx,5
> + add ecx,edi
> + xor esi,eax
> + ror ebp,7
> + add ecx,edx
> + pxor xmm3,xmm7
> + add ebx,DWORD[48+rsp]
> + xor esi,ebp
> + punpcklqdq xmm9,xmm2
> + mov edi,ecx
> + rol ecx,5
> + pxor xmm3,xmm4
> + add ebx,esi
> + xor edi,ebp
> + movdqa xmm8,xmm10
> + ror edx,7
> + paddd xmm10,xmm2
> + add ebx,ecx
> + pxor xmm3,xmm9
> + add eax,DWORD[52+rsp]
> + xor edi,edx
> + mov esi,ebx
> + rol ebx,5
> + movdqa xmm9,xmm3
> + add eax,edi
> + xor esi,edx
> + movdqa XMMWORD[32+rsp],xmm10
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[56+rsp]
> + pslld xmm3,2
> + xor esi,ecx
> + mov edi,eax
> + psrld xmm9,30
> + rol eax,5
> + add ebp,esi
> + xor edi,ecx
> + ror ebx,7
> + por xmm3,xmm9
> + add ebp,eax
> + add edx,DWORD[60+rsp]
> + pshufd xmm10,xmm2,238
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + add edx,edi
> + xor esi,ebx
> + ror eax,7
> + add edx,ebp
> + pxor xmm4,xmm0
> + add ecx,DWORD[rsp]
> + xor esi,eax
> + punpcklqdq xmm10,xmm3
> + mov edi,edx
> + rol edx,5
> + pxor xmm4,xmm5
> + add ecx,esi
> + xor edi,eax
> + movdqa xmm9,xmm8
> + ror ebp,7
> + paddd xmm8,xmm3
> + add ecx,edx
> + pxor xmm4,xmm10
> + add ebx,DWORD[4+rsp]
> + xor edi,ebp
> + mov esi,ecx
> + rol ecx,5
> + movdqa xmm10,xmm4
> + add ebx,edi
> + xor esi,ebp
> + movdqa XMMWORD[48+rsp],xmm8
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[8+rsp]
> + pslld xmm4,2
> + xor esi,edx
> + mov edi,ebx
> + psrld xmm10,30
> + rol ebx,5
> + add eax,esi
> + xor edi,edx
> + ror ecx,7
> + por xmm4,xmm10
> + add eax,ebx
> + add ebp,DWORD[12+rsp]
> + pshufd xmm8,xmm3,238
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + add ebp,edi
> + xor esi,ecx
> + ror ebx,7
> + add ebp,eax
> + pxor xmm5,xmm1
> + add edx,DWORD[16+rsp]
> + xor esi,ebx
> + punpcklqdq xmm8,xmm4
> + mov edi,ebp
> + rol ebp,5
> + pxor xmm5,xmm6
> + add edx,esi
> + xor edi,ebx
> + movdqa xmm10,xmm9
> + ror eax,7
> + paddd xmm9,xmm4
> + add edx,ebp
> + pxor xmm5,xmm8
> + add ecx,DWORD[20+rsp]
> + xor edi,eax
> + mov esi,edx
> + rol edx,5
> + movdqa xmm8,xmm5
> + add ecx,edi
> + xor esi,eax
> + movdqa XMMWORD[rsp],xmm9
> + ror ebp,7
> + add ecx,edx
> + add ebx,DWORD[24+rsp]
> + pslld xmm5,2
> + xor esi,ebp
> + mov edi,ecx
> + psrld xmm8,30
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + por xmm5,xmm8
> + add ebx,ecx
> + add eax,DWORD[28+rsp]
> + pshufd xmm9,xmm4,238
> + ror ecx,7
> + mov esi,ebx
> + xor edi,edx
> + rol ebx,5
> + add eax,edi
> + xor esi,ecx
> + xor ecx,edx
> + add eax,ebx
> + pxor xmm6,xmm2
> + add ebp,DWORD[32+rsp]
> + and esi,ecx
> + xor ecx,edx
> + ror ebx,7
> + punpcklqdq xmm9,xmm5
> + mov edi,eax
> + xor esi,ecx
> + pxor xmm6,xmm7
> + rol eax,5
> + add ebp,esi
> + movdqa xmm8,xmm10
> + xor edi,ebx
> + paddd xmm10,xmm5
> + xor ebx,ecx
> + pxor xmm6,xmm9
> + add ebp,eax
> + add edx,DWORD[36+rsp]
> + and edi,ebx
> + xor ebx,ecx
> + ror eax,7
> + movdqa xmm9,xmm6
> + mov esi,ebp
> + xor edi,ebx
> + movdqa XMMWORD[16+rsp],xmm10
> + rol ebp,5
> + add edx,edi
> + xor esi,eax
> + pslld xmm6,2
> + xor eax,ebx
> + add edx,ebp
> + psrld xmm9,30
> + add ecx,DWORD[40+rsp]
> + and esi,eax
> + xor eax,ebx
> + por xmm6,xmm9
> + ror ebp,7
> + mov edi,edx
> + xor esi,eax
> + rol edx,5
> + pshufd xmm10,xmm5,238
> + add ecx,esi
> + xor edi,ebp
> + xor ebp,eax
> + add ecx,edx
> + add ebx,DWORD[44+rsp]
> + and edi,ebp
> + xor ebp,eax
> + ror edx,7
> + mov esi,ecx
> + xor edi,ebp
> + rol ecx,5
> + add ebx,edi
> + xor esi,edx
> + xor edx,ebp
> + add ebx,ecx
> + pxor xmm7,xmm3
> + add eax,DWORD[48+rsp]
> + and esi,edx
> + xor edx,ebp
> + ror ecx,7
> + punpcklqdq xmm10,xmm6
> + mov edi,ebx
> + xor esi,edx
> + pxor xmm7,xmm0
> + rol ebx,5
> + add eax,esi
> + movdqa xmm9,XMMWORD[32+r14]
> + xor edi,ecx
> + paddd xmm8,xmm6
> + xor ecx,edx
> + pxor xmm7,xmm10
> + add eax,ebx
> + add ebp,DWORD[52+rsp]
> + and edi,ecx
> + xor ecx,edx
> + ror ebx,7
> + movdqa xmm10,xmm7
> + mov esi,eax
> + xor edi,ecx
> + movdqa XMMWORD[32+rsp],xmm8
> + rol eax,5
> + add ebp,edi
> + xor esi,ebx
> + pslld xmm7,2
> + xor ebx,ecx
> + add ebp,eax
> + psrld xmm10,30
> + add edx,DWORD[56+rsp]
> + and esi,ebx
> + xor ebx,ecx
> + por xmm7,xmm10
> + ror eax,7
> + mov edi,ebp
> + xor esi,ebx
> + rol ebp,5
> + pshufd xmm8,xmm6,238
> + add edx,esi
> + xor edi,eax
> + xor eax,ebx
> + add edx,ebp
> + add ecx,DWORD[60+rsp]
> + and edi,eax
> + xor eax,ebx
> + ror ebp,7
> + mov esi,edx
> + xor edi,eax
> + rol edx,5
> + add ecx,edi
> + xor esi,ebp
> + xor ebp,eax
> + add ecx,edx
> + pxor xmm0,xmm4
> + add ebx,DWORD[rsp]
> + and esi,ebp
> + xor ebp,eax
> + ror edx,7
> + punpcklqdq xmm8,xmm7
> + mov edi,ecx
> + xor esi,ebp
> + pxor xmm0,xmm1
> + rol ecx,5
> + add ebx,esi
> + movdqa xmm10,xmm9
> + xor edi,edx
> + paddd xmm9,xmm7
> + xor edx,ebp
> + pxor xmm0,xmm8
> + add ebx,ecx
> + add eax,DWORD[4+rsp]
> + and edi,edx
> + xor edx,ebp
> + ror ecx,7
> + movdqa xmm8,xmm0
> + mov esi,ebx
> + xor edi,edx
> + movdqa XMMWORD[48+rsp],xmm9
> + rol ebx,5
> + add eax,edi
> + xor esi,ecx
> + pslld xmm0,2
> + xor ecx,edx
> + add eax,ebx
> + psrld xmm8,30
> + add ebp,DWORD[8+rsp]
> + and esi,ecx
> + xor ecx,edx
> + por xmm0,xmm8
> + ror ebx,7
> + mov edi,eax
> + xor esi,ecx
> + rol eax,5
> + pshufd xmm9,xmm7,238
> + add ebp,esi
> + xor edi,ebx
> + xor ebx,ecx
> + add ebp,eax
> + add edx,DWORD[12+rsp]
> + and edi,ebx
> + xor ebx,ecx
> + ror eax,7
> + mov esi,ebp
> + xor edi,ebx
> + rol ebp,5
> + add edx,edi
> + xor esi,eax
> + xor eax,ebx
> + add edx,ebp
> + pxor xmm1,xmm5
> + add ecx,DWORD[16+rsp]
> + and esi,eax
> + xor eax,ebx
> + ror ebp,7
> + punpcklqdq xmm9,xmm0
> + mov edi,edx
> + xor esi,eax
> + pxor xmm1,xmm2
> + rol edx,5
> + add ecx,esi
> + movdqa xmm8,xmm10
> + xor edi,ebp
> + paddd xmm10,xmm0
> + xor ebp,eax
> + pxor xmm1,xmm9
> + add ecx,edx
> + add ebx,DWORD[20+rsp]
> + and edi,ebp
> + xor ebp,eax
> + ror edx,7
> + movdqa xmm9,xmm1
> + mov esi,ecx
> + xor edi,ebp
> + movdqa XMMWORD[rsp],xmm10
> + rol ecx,5
> + add ebx,edi
> + xor esi,edx
> + pslld xmm1,2
> + xor edx,ebp
> + add ebx,ecx
> + psrld xmm9,30
> + add eax,DWORD[24+rsp]
> + and esi,edx
> + xor edx,ebp
> + por xmm1,xmm9
> + ror ecx,7
> + mov edi,ebx
> + xor esi,edx
> + rol ebx,5
> + pshufd xmm10,xmm0,238
> + add eax,esi
> + xor edi,ecx
> + xor ecx,edx
> + add eax,ebx
> + add ebp,DWORD[28+rsp]
> + and edi,ecx
> + xor ecx,edx
> + ror ebx,7
> + mov esi,eax
> + xor edi,ecx
> + rol eax,5
> + add ebp,edi
> + xor esi,ebx
> + xor ebx,ecx
> + add ebp,eax
> + pxor xmm2,xmm6
> + add edx,DWORD[32+rsp]
> + and esi,ebx
> + xor ebx,ecx
> + ror eax,7
> + punpcklqdq xmm10,xmm1
> + mov edi,ebp
> + xor esi,ebx
> + pxor xmm2,xmm3
> + rol ebp,5
> + add edx,esi
> + movdqa xmm9,xmm8
> + xor edi,eax
> + paddd xmm8,xmm1
> + xor eax,ebx
> + pxor xmm2,xmm10
> + add edx,ebp
> + add ecx,DWORD[36+rsp]
> + and edi,eax
> + xor eax,ebx
> + ror ebp,7
> + movdqa xmm10,xmm2
> + mov esi,edx
> + xor edi,eax
> + movdqa XMMWORD[16+rsp],xmm8
> + rol edx,5
> + add ecx,edi
> + xor esi,ebp
> + pslld xmm2,2
> + xor ebp,eax
> + add ecx,edx
> + psrld xmm10,30
> + add ebx,DWORD[40+rsp]
> + and esi,ebp
> + xor ebp,eax
> + por xmm2,xmm10
> + ror edx,7
> + mov edi,ecx
> + xor esi,ebp
> + rol ecx,5
> + pshufd xmm8,xmm1,238
> + add ebx,esi
> + xor edi,edx
> + xor edx,ebp
> + add ebx,ecx
> + add eax,DWORD[44+rsp]
> + and edi,edx
> + xor edx,ebp
> + ror ecx,7
> + mov esi,ebx
> + xor edi,edx
> + rol ebx,5
> + add eax,edi
> + xor esi,edx
> + add eax,ebx
> + pxor xmm3,xmm7
> + add ebp,DWORD[48+rsp]
> + xor esi,ecx
> + punpcklqdq xmm8,xmm2
> + mov edi,eax
> + rol eax,5
> + pxor xmm3,xmm4
> + add ebp,esi
> + xor edi,ecx
> + movdqa xmm10,xmm9
> + ror ebx,7
> + paddd xmm9,xmm2
> + add ebp,eax
> + pxor xmm3,xmm8
> + add edx,DWORD[52+rsp]
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + movdqa xmm8,xmm3
> + add edx,edi
> + xor esi,ebx
> + movdqa XMMWORD[32+rsp],xmm9
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[56+rsp]
> + pslld xmm3,2
> + xor esi,eax
> + mov edi,edx
> + psrld xmm8,30
> + rol edx,5
> + add ecx,esi
> + xor edi,eax
> + ror ebp,7
> + por xmm3,xmm8
> + add ecx,edx
> + add ebx,DWORD[60+rsp]
> + xor edi,ebp
> + mov esi,ecx
> + rol ecx,5
> + add ebx,edi
> + xor esi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[rsp]
> + xor esi,edx
> + mov edi,ebx
> + rol ebx,5
> + paddd xmm10,xmm3
> + add eax,esi
> + xor edi,edx
> + movdqa XMMWORD[48+rsp],xmm10
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[4+rsp]
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + add ebp,edi
> + xor esi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[8+rsp]
> + xor esi,ebx
> + mov edi,ebp
> + rol ebp,5
> + add edx,esi
> + xor edi,ebx
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[12+rsp]
> + xor edi,eax
> + mov esi,edx
> + rol edx,5
> + add ecx,edi
> + xor esi,eax
> + ror ebp,7
> + add ecx,edx
> + cmp r9,r10
> + je NEAR $L$done_ssse3
> + movdqa xmm6,XMMWORD[64+r14]
> + movdqa xmm9,XMMWORD[((-64))+r14]
> + movdqu xmm0,XMMWORD[r9]
> + movdqu xmm1,XMMWORD[16+r9]
> + movdqu xmm2,XMMWORD[32+r9]
> + movdqu xmm3,XMMWORD[48+r9]
> +DB 102,15,56,0,198
> + add r9,64
> + add ebx,DWORD[16+rsp]
> + xor esi,ebp
> + mov edi,ecx
> +DB 102,15,56,0,206
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + paddd xmm0,xmm9
> + add ebx,ecx
> + add eax,DWORD[20+rsp]
> + xor edi,edx
> + mov esi,ebx
> + movdqa XMMWORD[rsp],xmm0
> + rol ebx,5
> + add eax,edi
> + xor esi,edx
> + ror ecx,7
> + psubd xmm0,xmm9
> + add eax,ebx
> + add ebp,DWORD[24+rsp]
> + xor esi,ecx
> + mov edi,eax
> + rol eax,5
> + add ebp,esi
> + xor edi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[28+rsp]
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + add edx,edi
> + xor esi,ebx
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[32+rsp]
> + xor esi,eax
> + mov edi,edx
> +DB 102,15,56,0,214
> + rol edx,5
> + add ecx,esi
> + xor edi,eax
> + ror ebp,7
> + paddd xmm1,xmm9
> + add ecx,edx
> + add ebx,DWORD[36+rsp]
> + xor edi,ebp
> + mov esi,ecx
> + movdqa XMMWORD[16+rsp],xmm1
> + rol ecx,5
> + add ebx,edi
> + xor esi,ebp
> + ror edx,7
> + psubd xmm1,xmm9
> + add ebx,ecx
> + add eax,DWORD[40+rsp]
> + xor esi,edx
> + mov edi,ebx
> + rol ebx,5
> + add eax,esi
> + xor edi,edx
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[44+rsp]
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + add ebp,edi
> + xor esi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[48+rsp]
> + xor esi,ebx
> + mov edi,ebp
> +DB 102,15,56,0,222
> + rol ebp,5
> + add edx,esi
> + xor edi,ebx
> + ror eax,7
> + paddd xmm2,xmm9
> + add edx,ebp
> + add ecx,DWORD[52+rsp]
> + xor edi,eax
> + mov esi,edx
> + movdqa XMMWORD[32+rsp],xmm2
> + rol edx,5
> + add ecx,edi
> + xor esi,eax
> + ror ebp,7
> + psubd xmm2,xmm9
> + add ecx,edx
> + add ebx,DWORD[56+rsp]
> + xor esi,ebp
> + mov edi,ecx
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[60+rsp]
> + xor edi,edx
> + mov esi,ebx
> + rol ebx,5
> + add eax,edi
> + ror ecx,7
> + add eax,ebx
> + add eax,DWORD[r8]
> + add esi,DWORD[4+r8]
> + add ecx,DWORD[8+r8]
> + add edx,DWORD[12+r8]
> + mov DWORD[r8],eax
> + add ebp,DWORD[16+r8]
> + mov DWORD[4+r8],esi
> + mov ebx,esi
> + mov DWORD[8+r8],ecx
> + mov edi,ecx
> + mov DWORD[12+r8],edx
> + xor edi,edx
> + mov DWORD[16+r8],ebp
> + and esi,edi
> + jmp NEAR $L$oop_ssse3
> +
> +ALIGN 16
> +$L$done_ssse3:
> + add ebx,DWORD[16+rsp]
> + xor esi,ebp
> + mov edi,ecx
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[20+rsp]
> + xor edi,edx
> + mov esi,ebx
> + rol ebx,5
> + add eax,edi
> + xor esi,edx
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[24+rsp]
> + xor esi,ecx
> + mov edi,eax
> + rol eax,5
> + add ebp,esi
> + xor edi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[28+rsp]
> + xor edi,ebx
> + mov esi,ebp
> + rol ebp,5
> + add edx,edi
> + xor esi,ebx
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[32+rsp]
> + xor esi,eax
> + mov edi,edx
> + rol edx,5
> + add ecx,esi
> + xor edi,eax
> + ror ebp,7
> + add ecx,edx
> + add ebx,DWORD[36+rsp]
> + xor edi,ebp
> + mov esi,ecx
> + rol ecx,5
> + add ebx,edi
> + xor esi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[40+rsp]
> + xor esi,edx
> + mov edi,ebx
> + rol ebx,5
> + add eax,esi
> + xor edi,edx
> + ror ecx,7
> + add eax,ebx
> + add ebp,DWORD[44+rsp]
> + xor edi,ecx
> + mov esi,eax
> + rol eax,5
> + add ebp,edi
> + xor esi,ecx
> + ror ebx,7
> + add ebp,eax
> + add edx,DWORD[48+rsp]
> + xor esi,ebx
> + mov edi,ebp
> + rol ebp,5
> + add edx,esi
> + xor edi,ebx
> + ror eax,7
> + add edx,ebp
> + add ecx,DWORD[52+rsp]
> + xor edi,eax
> + mov esi,edx
> + rol edx,5
> + add ecx,edi
> + xor esi,eax
> + ror ebp,7
> + add ecx,edx
> + add ebx,DWORD[56+rsp]
> + xor esi,ebp
> + mov edi,ecx
> + rol ecx,5
> + add ebx,esi
> + xor edi,ebp
> + ror edx,7
> + add ebx,ecx
> + add eax,DWORD[60+rsp]
> + xor edi,edx
> + mov esi,ebx
> + rol ebx,5
> + add eax,edi
> + ror ecx,7
> + add eax,ebx
> + add eax,DWORD[r8]
> + add esi,DWORD[4+r8]
> + add ecx,DWORD[8+r8]
> + mov DWORD[r8],eax
> + add edx,DWORD[12+r8]
> + mov DWORD[4+r8],esi
> + add ebp,DWORD[16+r8]
> + mov DWORD[8+r8],ecx
> + mov DWORD[12+r8],edx
> + mov DWORD[16+r8],ebp
> + movaps xmm6,XMMWORD[((-40-96))+r11]
> + movaps xmm7,XMMWORD[((-40-80))+r11]
> + movaps xmm8,XMMWORD[((-40-64))+r11]
> + movaps xmm9,XMMWORD[((-40-48))+r11]
> + movaps xmm10,XMMWORD[((-40-32))+r11]
> + movaps xmm11,XMMWORD[((-40-16))+r11]
> + mov r14,QWORD[((-40))+r11]
> +
> + mov r13,QWORD[((-32))+r11]
> +
> + mov r12,QWORD[((-24))+r11]
> +
> + mov rbp,QWORD[((-16))+r11]
> +
> + mov rbx,QWORD[((-8))+r11]
> +
> + lea rsp,[r11]
> +
> +$L$epilogue_ssse3:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha1_block_data_order_ssse3:
> +ALIGN 64
> +K_XX_XX:
> + DD 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> + DD 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> + DD 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> + DD 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> + DD 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> + DD 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> + DD 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> + DD 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +DB 0xf,0xe,0xd,0xc,0xb,0xa,0x9,0x8,0x7,0x6,0x5,0x4,0x3,0x2,0x1,0x0
> +DB 83,72,65,49,32,98,108,111,99,107,32,116,114,97,110,115
> +DB 102,111,114,109,32,102,111,114,32,120,56,54,95,54,52,44
> +DB 32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60
> +DB 97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114
> +DB 103,62,0
> +ALIGN 64
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + lea r10,[$L$prologue]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[152+r8]
> +
> + lea r10,[$L$epilogue]
> + cmp rbx,r10
> + jae NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[64+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov r12,QWORD[((-24))+rax]
> + mov r13,QWORD[((-32))+rax]
> + mov r14,QWORD[((-40))+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> + mov QWORD[216+r8],r12
> + mov QWORD[224+r8],r13
> + mov QWORD[232+r8],r14
> +
> + jmp NEAR $L$common_seh_tail
> +
> +
> +ALIGN 16
> +shaext_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + lea r10,[$L$prologue_shaext]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + lea r10,[$L$epilogue_shaext]
> + cmp rbx,r10
> + jae NEAR $L$common_seh_tail
> +
> + lea rsi,[((-8-64))+rax]
> + lea rdi,[512+r8]
> + mov ecx,8
> + DD 0xa548f3fc
> +
> + jmp NEAR $L$common_seh_tail
> +
> +
> +ALIGN 16
> +ssse3_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$common_seh_tail
> +
> + mov rax,QWORD[208+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$common_seh_tail
> +
> + lea rsi,[((-40-96))+rax]
> + lea rdi,[512+r8]
> + mov ecx,12
> + DD 0xa548f3fc
> +
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov r12,QWORD[((-24))+rax]
> + mov r13,QWORD[((-32))+rax]
> + mov r14,QWORD[((-40))+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> + mov QWORD[216+r8],r12
> + mov QWORD[224+r8],r13
> + mov QWORD[232+r8],r14
> +
> +$L$common_seh_tail:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_sha1_block_data_order wrt ..imagebase
> + DD $L$SEH_end_sha1_block_data_order wrt ..imagebase
> + DD $L$SEH_info_sha1_block_data_order wrt ..imagebase
> + DD $L$SEH_begin_sha1_block_data_order_shaext wrt ..imagebase
> + DD $L$SEH_end_sha1_block_data_order_shaext wrt ..imagebase
> + DD $L$SEH_info_sha1_block_data_order_shaext wrt ..imagebase
> + DD $L$SEH_begin_sha1_block_data_order_ssse3 wrt ..imagebase
> + DD $L$SEH_end_sha1_block_data_order_ssse3 wrt ..imagebase
> + DD $L$SEH_info_sha1_block_data_order_ssse3 wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_sha1_block_data_order:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> +$L$SEH_info_sha1_block_data_order_shaext:
> +DB 9,0,0,0
> + DD shaext_handler wrt ..imagebase
> +$L$SEH_info_sha1_block_data_order_ssse3:
> +DB 9,0,0,0
> + DD ssse3_handler wrt ..imagebase
> + DD $L$prologue_ssse3 wrt ..imagebase,$L$epilogue_ssse3
> wrt ..imagebase
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha256-mb-
> x86_64.nasm b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha256-mb-
> x86_64.nasm
> new file mode 100644
> index 0000000000..7cd5eae85c
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha256-mb-x86_64.nasm
> @@ -0,0 +1,3461 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/sha/asm/sha256-mb-x86_64.pl
> +;
> +; Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +
> +EXTERN OPENSSL_ia32cap_P
> +
> +global sha256_multi_block
> +
> +ALIGN 32
> +sha256_multi_block:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha256_multi_block:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + mov rcx,QWORD[((OPENSSL_ia32cap_P+4))]
> + bt rcx,61
> + jc NEAR _shaext_shortcut
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + lea rsp,[((-168))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> + movaps XMMWORD[(-120)+rax],xmm10
> + movaps XMMWORD[(-104)+rax],xmm11
> + movaps XMMWORD[(-88)+rax],xmm12
> + movaps XMMWORD[(-72)+rax],xmm13
> + movaps XMMWORD[(-56)+rax],xmm14
> + movaps XMMWORD[(-40)+rax],xmm15
> + sub rsp,288
> + and rsp,-256
> + mov QWORD[272+rsp],rax
> +
> +$L$body:
> + lea rbp,[((K256+128))]
> + lea rbx,[256+rsp]
> + lea rdi,[128+rdi]
> +
> +$L$oop_grande:
> + mov DWORD[280+rsp],edx
> + xor edx,edx
> + mov r8,QWORD[rsi]
> + mov ecx,DWORD[8+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[rbx],ecx
> + cmovle r8,rbp
> + mov r9,QWORD[16+rsi]
> + mov ecx,DWORD[24+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[4+rbx],ecx
> + cmovle r9,rbp
> + mov r10,QWORD[32+rsi]
> + mov ecx,DWORD[40+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[8+rbx],ecx
> + cmovle r10,rbp
> + mov r11,QWORD[48+rsi]
> + mov ecx,DWORD[56+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[12+rbx],ecx
> + cmovle r11,rbp
> + test edx,edx
> + jz NEAR $L$done
> +
> + movdqu xmm8,XMMWORD[((0-128))+rdi]
> + lea rax,[128+rsp]
> + movdqu xmm9,XMMWORD[((32-128))+rdi]
> + movdqu xmm10,XMMWORD[((64-128))+rdi]
> + movdqu xmm11,XMMWORD[((96-128))+rdi]
> + movdqu xmm12,XMMWORD[((128-128))+rdi]
> + movdqu xmm13,XMMWORD[((160-128))+rdi]
> + movdqu xmm14,XMMWORD[((192-128))+rdi]
> + movdqu xmm15,XMMWORD[((224-128))+rdi]
> + movdqu xmm6,XMMWORD[$L$pbswap]
> + jmp NEAR $L$oop
> +
> +ALIGN 32
> +$L$oop:
> + movdqa xmm4,xmm10
> + pxor xmm4,xmm9
> + movd xmm5,DWORD[r8]
> + movd xmm0,DWORD[r9]
> + movd xmm1,DWORD[r10]
> + movd xmm2,DWORD[r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm12
> +DB 102,15,56,0,238
> + movdqa xmm2,xmm12
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm12
> + pslld xmm2,7
> + movdqa XMMWORD[(0-128)+rax],xmm5
> + paddd xmm5,xmm15
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-128))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm12
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm12
> + pslld xmm2,26-21
> + pandn xmm0,xmm14
> + pand xmm3,xmm13
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm8
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm8
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm9
> + movdqa xmm7,xmm8
> + pslld xmm2,10
> + pxor xmm3,xmm8
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm15,xmm9
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm15,xmm4
> + paddd xmm11,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm15,xmm5
> + paddd xmm15,xmm7
> + movd xmm5,DWORD[4+r8]
> + movd xmm0,DWORD[4+r9]
> + movd xmm1,DWORD[4+r10]
> + movd xmm2,DWORD[4+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm11
> +
> + movdqa xmm2,xmm11
> +DB 102,15,56,0,238
> + psrld xmm7,6
> + movdqa xmm1,xmm11
> + pslld xmm2,7
> + movdqa XMMWORD[(16-128)+rax],xmm5
> + paddd xmm5,xmm14
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-96))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm11
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm11
> + pslld xmm2,26-21
> + pandn xmm0,xmm13
> + pand xmm4,xmm12
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm15
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm15
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm8
> + movdqa xmm7,xmm15
> + pslld xmm2,10
> + pxor xmm4,xmm15
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm14,xmm8
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm14,xmm3
> + paddd xmm10,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm14,xmm5
> + paddd xmm14,xmm7
> + movd xmm5,DWORD[8+r8]
> + movd xmm0,DWORD[8+r9]
> + movd xmm1,DWORD[8+r10]
> + movd xmm2,DWORD[8+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm10
> +DB 102,15,56,0,238
> + movdqa xmm2,xmm10
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm10
> + pslld xmm2,7
> + movdqa XMMWORD[(32-128)+rax],xmm5
> + paddd xmm5,xmm13
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-64))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm10
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm10
> + pslld xmm2,26-21
> + pandn xmm0,xmm12
> + pand xmm3,xmm11
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm14
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm14
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm15
> + movdqa xmm7,xmm14
> + pslld xmm2,10
> + pxor xmm3,xmm14
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm13,xmm15
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm13,xmm4
> + paddd xmm9,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm13,xmm5
> + paddd xmm13,xmm7
> + movd xmm5,DWORD[12+r8]
> + movd xmm0,DWORD[12+r9]
> + movd xmm1,DWORD[12+r10]
> + movd xmm2,DWORD[12+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm9
> +
> + movdqa xmm2,xmm9
> +DB 102,15,56,0,238
> + psrld xmm7,6
> + movdqa xmm1,xmm9
> + pslld xmm2,7
> + movdqa XMMWORD[(48-128)+rax],xmm5
> + paddd xmm5,xmm12
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-32))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm9
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm9
> + pslld xmm2,26-21
> + pandn xmm0,xmm11
> + pand xmm4,xmm10
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm13
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm13
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm14
> + movdqa xmm7,xmm13
> + pslld xmm2,10
> + pxor xmm4,xmm13
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm12,xmm14
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm12,xmm3
> + paddd xmm8,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm12,xmm5
> + paddd xmm12,xmm7
> + movd xmm5,DWORD[16+r8]
> + movd xmm0,DWORD[16+r9]
> + movd xmm1,DWORD[16+r10]
> + movd xmm2,DWORD[16+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm8
> +DB 102,15,56,0,238
> + movdqa xmm2,xmm8
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm8
> + pslld xmm2,7
> + movdqa XMMWORD[(64-128)+rax],xmm5
> + paddd xmm5,xmm11
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm8
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm8
> + pslld xmm2,26-21
> + pandn xmm0,xmm10
> + pand xmm3,xmm9
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm12
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm12
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm13
> + movdqa xmm7,xmm12
> + pslld xmm2,10
> + pxor xmm3,xmm12
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm11,xmm13
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm11,xmm4
> + paddd xmm15,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm11,xmm5
> + paddd xmm11,xmm7
> + movd xmm5,DWORD[20+r8]
> + movd xmm0,DWORD[20+r9]
> + movd xmm1,DWORD[20+r10]
> + movd xmm2,DWORD[20+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm15
> +
> + movdqa xmm2,xmm15
> +DB 102,15,56,0,238
> + psrld xmm7,6
> + movdqa xmm1,xmm15
> + pslld xmm2,7
> + movdqa XMMWORD[(80-128)+rax],xmm5
> + paddd xmm5,xmm10
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[32+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm15
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm15
> + pslld xmm2,26-21
> + pandn xmm0,xmm9
> + pand xmm4,xmm8
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm11
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm11
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm12
> + movdqa xmm7,xmm11
> + pslld xmm2,10
> + pxor xmm4,xmm11
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm10,xmm12
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm10,xmm3
> + paddd xmm14,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm10,xmm5
> + paddd xmm10,xmm7
> + movd xmm5,DWORD[24+r8]
> + movd xmm0,DWORD[24+r9]
> + movd xmm1,DWORD[24+r10]
> + movd xmm2,DWORD[24+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm14
> +DB 102,15,56,0,238
> + movdqa xmm2,xmm14
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm14
> + pslld xmm2,7
> + movdqa XMMWORD[(96-128)+rax],xmm5
> + paddd xmm5,xmm9
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[64+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm14
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm14
> + pslld xmm2,26-21
> + pandn xmm0,xmm8
> + pand xmm3,xmm15
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm10
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm10
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm11
> + movdqa xmm7,xmm10
> + pslld xmm2,10
> + pxor xmm3,xmm10
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm9,xmm11
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm9,xmm4
> + paddd xmm13,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm9,xmm5
> + paddd xmm9,xmm7
> + movd xmm5,DWORD[28+r8]
> + movd xmm0,DWORD[28+r9]
> + movd xmm1,DWORD[28+r10]
> + movd xmm2,DWORD[28+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm13
> +
> + movdqa xmm2,xmm13
> +DB 102,15,56,0,238
> + psrld xmm7,6
> + movdqa xmm1,xmm13
> + pslld xmm2,7
> + movdqa XMMWORD[(112-128)+rax],xmm5
> + paddd xmm5,xmm8
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[96+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm13
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm13
> + pslld xmm2,26-21
> + pandn xmm0,xmm15
> + pand xmm4,xmm14
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm9
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm9
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm10
> + movdqa xmm7,xmm9
> + pslld xmm2,10
> + pxor xmm4,xmm9
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm8,xmm10
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm8,xmm3
> + paddd xmm12,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm8,xmm5
> + paddd xmm8,xmm7
> + lea rbp,[256+rbp]
> + movd xmm5,DWORD[32+r8]
> + movd xmm0,DWORD[32+r9]
> + movd xmm1,DWORD[32+r10]
> + movd xmm2,DWORD[32+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm12
> +DB 102,15,56,0,238
> + movdqa xmm2,xmm12
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm12
> + pslld xmm2,7
> + movdqa XMMWORD[(128-128)+rax],xmm5
> + paddd xmm5,xmm15
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-128))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm12
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm12
> + pslld xmm2,26-21
> + pandn xmm0,xmm14
> + pand xmm3,xmm13
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm8
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm8
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm9
> + movdqa xmm7,xmm8
> + pslld xmm2,10
> + pxor xmm3,xmm8
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm15,xmm9
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm15,xmm4
> + paddd xmm11,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm15,xmm5
> + paddd xmm15,xmm7
> + movd xmm5,DWORD[36+r8]
> + movd xmm0,DWORD[36+r9]
> + movd xmm1,DWORD[36+r10]
> + movd xmm2,DWORD[36+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm11
> +
> + movdqa xmm2,xmm11
> +DB 102,15,56,0,238
> + psrld xmm7,6
> + movdqa xmm1,xmm11
> + pslld xmm2,7
> + movdqa XMMWORD[(144-128)+rax],xmm5
> + paddd xmm5,xmm14
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-96))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm11
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm11
> + pslld xmm2,26-21
> + pandn xmm0,xmm13
> + pand xmm4,xmm12
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm15
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm15
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm8
> + movdqa xmm7,xmm15
> + pslld xmm2,10
> + pxor xmm4,xmm15
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm14,xmm8
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm14,xmm3
> + paddd xmm10,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm14,xmm5
> + paddd xmm14,xmm7
> + movd xmm5,DWORD[40+r8]
> + movd xmm0,DWORD[40+r9]
> + movd xmm1,DWORD[40+r10]
> + movd xmm2,DWORD[40+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm10
> +DB 102,15,56,0,238
> + movdqa xmm2,xmm10
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm10
> + pslld xmm2,7
> + movdqa XMMWORD[(160-128)+rax],xmm5
> + paddd xmm5,xmm13
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-64))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm10
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm10
> + pslld xmm2,26-21
> + pandn xmm0,xmm12
> + pand xmm3,xmm11
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm14
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm14
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm15
> + movdqa xmm7,xmm14
> + pslld xmm2,10
> + pxor xmm3,xmm14
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm13,xmm15
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm13,xmm4
> + paddd xmm9,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm13,xmm5
> + paddd xmm13,xmm7
> + movd xmm5,DWORD[44+r8]
> + movd xmm0,DWORD[44+r9]
> + movd xmm1,DWORD[44+r10]
> + movd xmm2,DWORD[44+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm9
> +
> + movdqa xmm2,xmm9
> +DB 102,15,56,0,238
> + psrld xmm7,6
> + movdqa xmm1,xmm9
> + pslld xmm2,7
> + movdqa XMMWORD[(176-128)+rax],xmm5
> + paddd xmm5,xmm12
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-32))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm9
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm9
> + pslld xmm2,26-21
> + pandn xmm0,xmm11
> + pand xmm4,xmm10
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm13
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm13
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm14
> + movdqa xmm7,xmm13
> + pslld xmm2,10
> + pxor xmm4,xmm13
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm12,xmm14
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm12,xmm3
> + paddd xmm8,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm12,xmm5
> + paddd xmm12,xmm7
> + movd xmm5,DWORD[48+r8]
> + movd xmm0,DWORD[48+r9]
> + movd xmm1,DWORD[48+r10]
> + movd xmm2,DWORD[48+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm8
> +DB 102,15,56,0,238
> + movdqa xmm2,xmm8
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm8
> + pslld xmm2,7
> + movdqa XMMWORD[(192-128)+rax],xmm5
> + paddd xmm5,xmm11
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm8
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm8
> + pslld xmm2,26-21
> + pandn xmm0,xmm10
> + pand xmm3,xmm9
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm12
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm12
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm13
> + movdqa xmm7,xmm12
> + pslld xmm2,10
> + pxor xmm3,xmm12
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm11,xmm13
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm11,xmm4
> + paddd xmm15,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm11,xmm5
> + paddd xmm11,xmm7
> + movd xmm5,DWORD[52+r8]
> + movd xmm0,DWORD[52+r9]
> + movd xmm1,DWORD[52+r10]
> + movd xmm2,DWORD[52+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm15
> +
> + movdqa xmm2,xmm15
> +DB 102,15,56,0,238
> + psrld xmm7,6
> + movdqa xmm1,xmm15
> + pslld xmm2,7
> + movdqa XMMWORD[(208-128)+rax],xmm5
> + paddd xmm5,xmm10
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[32+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm15
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm15
> + pslld xmm2,26-21
> + pandn xmm0,xmm9
> + pand xmm4,xmm8
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm11
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm11
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm12
> + movdqa xmm7,xmm11
> + pslld xmm2,10
> + pxor xmm4,xmm11
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm10,xmm12
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm10,xmm3
> + paddd xmm14,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm10,xmm5
> + paddd xmm10,xmm7
> + movd xmm5,DWORD[56+r8]
> + movd xmm0,DWORD[56+r9]
> + movd xmm1,DWORD[56+r10]
> + movd xmm2,DWORD[56+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm14
> +DB 102,15,56,0,238
> + movdqa xmm2,xmm14
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm14
> + pslld xmm2,7
> + movdqa XMMWORD[(224-128)+rax],xmm5
> + paddd xmm5,xmm9
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[64+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm14
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm14
> + pslld xmm2,26-21
> + pandn xmm0,xmm8
> + pand xmm3,xmm15
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm10
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm10
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm11
> + movdqa xmm7,xmm10
> + pslld xmm2,10
> + pxor xmm3,xmm10
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm9,xmm11
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm9,xmm4
> + paddd xmm13,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm9,xmm5
> + paddd xmm9,xmm7
> + movd xmm5,DWORD[60+r8]
> + lea r8,[64+r8]
> + movd xmm0,DWORD[60+r9]
> + lea r9,[64+r9]
> + movd xmm1,DWORD[60+r10]
> + lea r10,[64+r10]
> + movd xmm2,DWORD[60+r11]
> + lea r11,[64+r11]
> + punpckldq xmm5,xmm1
> + punpckldq xmm0,xmm2
> + punpckldq xmm5,xmm0
> + movdqa xmm7,xmm13
> +
> + movdqa xmm2,xmm13
> +DB 102,15,56,0,238
> + psrld xmm7,6
> + movdqa xmm1,xmm13
> + pslld xmm2,7
> + movdqa XMMWORD[(240-128)+rax],xmm5
> + paddd xmm5,xmm8
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[96+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm13
> + prefetcht0 [63+r8]
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm13
> + pslld xmm2,26-21
> + pandn xmm0,xmm15
> + pand xmm4,xmm14
> + pxor xmm7,xmm1
> +
> + prefetcht0 [63+r9]
> + movdqa xmm1,xmm9
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm9
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm10
> + movdqa xmm7,xmm9
> + pslld xmm2,10
> + pxor xmm4,xmm9
> +
> + prefetcht0 [63+r10]
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> + prefetcht0 [63+r11]
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm8,xmm10
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm8,xmm3
> + paddd xmm12,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm8,xmm5
> + paddd xmm8,xmm7
> + lea rbp,[256+rbp]
> + movdqu xmm5,XMMWORD[((0-128))+rax]
> + mov ecx,3
> + jmp NEAR $L$oop_16_xx
> +ALIGN 32
> +$L$oop_16_xx:
> + movdqa xmm6,XMMWORD[((16-128))+rax]
> + paddd xmm5,XMMWORD[((144-128))+rax]
> +
> + movdqa xmm7,xmm6
> + movdqa xmm1,xmm6
> + psrld xmm7,3
> + movdqa xmm2,xmm6
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((224-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm3,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm3
> +
> + psrld xmm3,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + psrld xmm3,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm3
> + pxor xmm0,xmm1
> + paddd xmm5,xmm0
> + movdqa xmm7,xmm12
> +
> + movdqa xmm2,xmm12
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm12
> + pslld xmm2,7
> + movdqa XMMWORD[(0-128)+rax],xmm5
> + paddd xmm5,xmm15
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-128))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm12
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm12
> + pslld xmm2,26-21
> + pandn xmm0,xmm14
> + pand xmm3,xmm13
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm8
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm8
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm9
> + movdqa xmm7,xmm8
> + pslld xmm2,10
> + pxor xmm3,xmm8
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm15,xmm9
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm15,xmm4
> + paddd xmm11,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm15,xmm5
> + paddd xmm15,xmm7
> + movdqa xmm5,XMMWORD[((32-128))+rax]
> + paddd xmm6,XMMWORD[((160-128))+rax]
> +
> + movdqa xmm7,xmm5
> + movdqa xmm1,xmm5
> + psrld xmm7,3
> + movdqa xmm2,xmm5
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((240-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm4,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm4
> +
> + psrld xmm4,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + psrld xmm4,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm4
> + pxor xmm0,xmm1
> + paddd xmm6,xmm0
> + movdqa xmm7,xmm11
> +
> + movdqa xmm2,xmm11
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm11
> + pslld xmm2,7
> + movdqa XMMWORD[(16-128)+rax],xmm6
> + paddd xmm6,xmm14
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm6,XMMWORD[((-96))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm11
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm11
> + pslld xmm2,26-21
> + pandn xmm0,xmm13
> + pand xmm4,xmm12
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm15
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm15
> + psrld xmm1,2
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm8
> + movdqa xmm7,xmm15
> + pslld xmm2,10
> + pxor xmm4,xmm15
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm6,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm14,xmm8
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm14,xmm3
> + paddd xmm10,xmm6
> + pxor xmm7,xmm2
> +
> + paddd xmm14,xmm6
> + paddd xmm14,xmm7
> + movdqa xmm6,XMMWORD[((48-128))+rax]
> + paddd xmm5,XMMWORD[((176-128))+rax]
> +
> + movdqa xmm7,xmm6
> + movdqa xmm1,xmm6
> + psrld xmm7,3
> + movdqa xmm2,xmm6
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((0-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm3,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm3
> +
> + psrld xmm3,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + psrld xmm3,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm3
> + pxor xmm0,xmm1
> + paddd xmm5,xmm0
> + movdqa xmm7,xmm10
> +
> + movdqa xmm2,xmm10
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm10
> + pslld xmm2,7
> + movdqa XMMWORD[(32-128)+rax],xmm5
> + paddd xmm5,xmm13
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-64))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm10
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm10
> + pslld xmm2,26-21
> + pandn xmm0,xmm12
> + pand xmm3,xmm11
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm14
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm14
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm15
> + movdqa xmm7,xmm14
> + pslld xmm2,10
> + pxor xmm3,xmm14
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm13,xmm15
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm13,xmm4
> + paddd xmm9,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm13,xmm5
> + paddd xmm13,xmm7
> + movdqa xmm5,XMMWORD[((64-128))+rax]
> + paddd xmm6,XMMWORD[((192-128))+rax]
> +
> + movdqa xmm7,xmm5
> + movdqa xmm1,xmm5
> + psrld xmm7,3
> + movdqa xmm2,xmm5
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((16-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm4,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm4
> +
> + psrld xmm4,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + psrld xmm4,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm4
> + pxor xmm0,xmm1
> + paddd xmm6,xmm0
> + movdqa xmm7,xmm9
> +
> + movdqa xmm2,xmm9
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm9
> + pslld xmm2,7
> + movdqa XMMWORD[(48-128)+rax],xmm6
> + paddd xmm6,xmm12
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm6,XMMWORD[((-32))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm9
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm9
> + pslld xmm2,26-21
> + pandn xmm0,xmm11
> + pand xmm4,xmm10
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm13
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm13
> + psrld xmm1,2
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm14
> + movdqa xmm7,xmm13
> + pslld xmm2,10
> + pxor xmm4,xmm13
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm6,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm12,xmm14
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm12,xmm3
> + paddd xmm8,xmm6
> + pxor xmm7,xmm2
> +
> + paddd xmm12,xmm6
> + paddd xmm12,xmm7
> + movdqa xmm6,XMMWORD[((80-128))+rax]
> + paddd xmm5,XMMWORD[((208-128))+rax]
> +
> + movdqa xmm7,xmm6
> + movdqa xmm1,xmm6
> + psrld xmm7,3
> + movdqa xmm2,xmm6
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((32-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm3,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm3
> +
> + psrld xmm3,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + psrld xmm3,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm3
> + pxor xmm0,xmm1
> + paddd xmm5,xmm0
> + movdqa xmm7,xmm8
> +
> + movdqa xmm2,xmm8
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm8
> + pslld xmm2,7
> + movdqa XMMWORD[(64-128)+rax],xmm5
> + paddd xmm5,xmm11
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm8
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm8
> + pslld xmm2,26-21
> + pandn xmm0,xmm10
> + pand xmm3,xmm9
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm12
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm12
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm13
> + movdqa xmm7,xmm12
> + pslld xmm2,10
> + pxor xmm3,xmm12
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm11,xmm13
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm11,xmm4
> + paddd xmm15,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm11,xmm5
> + paddd xmm11,xmm7
> + movdqa xmm5,XMMWORD[((96-128))+rax]
> + paddd xmm6,XMMWORD[((224-128))+rax]
> +
> + movdqa xmm7,xmm5
> + movdqa xmm1,xmm5
> + psrld xmm7,3
> + movdqa xmm2,xmm5
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((48-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm4,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm4
> +
> + psrld xmm4,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + psrld xmm4,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm4
> + pxor xmm0,xmm1
> + paddd xmm6,xmm0
> + movdqa xmm7,xmm15
> +
> + movdqa xmm2,xmm15
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm15
> + pslld xmm2,7
> + movdqa XMMWORD[(80-128)+rax],xmm6
> + paddd xmm6,xmm10
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm6,XMMWORD[32+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm15
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm15
> + pslld xmm2,26-21
> + pandn xmm0,xmm9
> + pand xmm4,xmm8
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm11
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm11
> + psrld xmm1,2
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm12
> + movdqa xmm7,xmm11
> + pslld xmm2,10
> + pxor xmm4,xmm11
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm6,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm10,xmm12
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm10,xmm3
> + paddd xmm14,xmm6
> + pxor xmm7,xmm2
> +
> + paddd xmm10,xmm6
> + paddd xmm10,xmm7
> + movdqa xmm6,XMMWORD[((112-128))+rax]
> + paddd xmm5,XMMWORD[((240-128))+rax]
> +
> + movdqa xmm7,xmm6
> + movdqa xmm1,xmm6
> + psrld xmm7,3
> + movdqa xmm2,xmm6
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((64-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm3,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm3
> +
> + psrld xmm3,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + psrld xmm3,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm3
> + pxor xmm0,xmm1
> + paddd xmm5,xmm0
> + movdqa xmm7,xmm14
> +
> + movdqa xmm2,xmm14
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm14
> + pslld xmm2,7
> + movdqa XMMWORD[(96-128)+rax],xmm5
> + paddd xmm5,xmm9
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[64+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm14
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm14
> + pslld xmm2,26-21
> + pandn xmm0,xmm8
> + pand xmm3,xmm15
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm10
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm10
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm11
> + movdqa xmm7,xmm10
> + pslld xmm2,10
> + pxor xmm3,xmm10
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm9,xmm11
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm9,xmm4
> + paddd xmm13,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm9,xmm5
> + paddd xmm9,xmm7
> + movdqa xmm5,XMMWORD[((128-128))+rax]
> + paddd xmm6,XMMWORD[((0-128))+rax]
> +
> + movdqa xmm7,xmm5
> + movdqa xmm1,xmm5
> + psrld xmm7,3
> + movdqa xmm2,xmm5
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((80-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm4,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm4
> +
> + psrld xmm4,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + psrld xmm4,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm4
> + pxor xmm0,xmm1
> + paddd xmm6,xmm0
> + movdqa xmm7,xmm13
> +
> + movdqa xmm2,xmm13
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm13
> + pslld xmm2,7
> + movdqa XMMWORD[(112-128)+rax],xmm6
> + paddd xmm6,xmm8
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm6,XMMWORD[96+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm13
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm13
> + pslld xmm2,26-21
> + pandn xmm0,xmm15
> + pand xmm4,xmm14
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm9
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm9
> + psrld xmm1,2
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm10
> + movdqa xmm7,xmm9
> + pslld xmm2,10
> + pxor xmm4,xmm9
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm6,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm8,xmm10
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm8,xmm3
> + paddd xmm12,xmm6
> + pxor xmm7,xmm2
> +
> + paddd xmm8,xmm6
> + paddd xmm8,xmm7
> + lea rbp,[256+rbp]
> + movdqa xmm6,XMMWORD[((144-128))+rax]
> + paddd xmm5,XMMWORD[((16-128))+rax]
> +
> + movdqa xmm7,xmm6
> + movdqa xmm1,xmm6
> + psrld xmm7,3
> + movdqa xmm2,xmm6
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((96-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm3,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm3
> +
> + psrld xmm3,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + psrld xmm3,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm3
> + pxor xmm0,xmm1
> + paddd xmm5,xmm0
> + movdqa xmm7,xmm12
> +
> + movdqa xmm2,xmm12
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm12
> + pslld xmm2,7
> + movdqa XMMWORD[(128-128)+rax],xmm5
> + paddd xmm5,xmm15
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-128))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm12
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm12
> + pslld xmm2,26-21
> + pandn xmm0,xmm14
> + pand xmm3,xmm13
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm8
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm8
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm9
> + movdqa xmm7,xmm8
> + pslld xmm2,10
> + pxor xmm3,xmm8
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm15,xmm9
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm15,xmm4
> + paddd xmm11,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm15,xmm5
> + paddd xmm15,xmm7
> + movdqa xmm5,XMMWORD[((160-128))+rax]
> + paddd xmm6,XMMWORD[((32-128))+rax]
> +
> + movdqa xmm7,xmm5
> + movdqa xmm1,xmm5
> + psrld xmm7,3
> + movdqa xmm2,xmm5
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((112-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm4,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm4
> +
> + psrld xmm4,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + psrld xmm4,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm4
> + pxor xmm0,xmm1
> + paddd xmm6,xmm0
> + movdqa xmm7,xmm11
> +
> + movdqa xmm2,xmm11
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm11
> + pslld xmm2,7
> + movdqa XMMWORD[(144-128)+rax],xmm6
> + paddd xmm6,xmm14
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm6,XMMWORD[((-96))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm11
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm11
> + pslld xmm2,26-21
> + pandn xmm0,xmm13
> + pand xmm4,xmm12
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm15
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm15
> + psrld xmm1,2
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm8
> + movdqa xmm7,xmm15
> + pslld xmm2,10
> + pxor xmm4,xmm15
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm6,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm14,xmm8
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm14,xmm3
> + paddd xmm10,xmm6
> + pxor xmm7,xmm2
> +
> + paddd xmm14,xmm6
> + paddd xmm14,xmm7
> + movdqa xmm6,XMMWORD[((176-128))+rax]
> + paddd xmm5,XMMWORD[((48-128))+rax]
> +
> + movdqa xmm7,xmm6
> + movdqa xmm1,xmm6
> + psrld xmm7,3
> + movdqa xmm2,xmm6
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((128-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm3,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm3
> +
> + psrld xmm3,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + psrld xmm3,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm3
> + pxor xmm0,xmm1
> + paddd xmm5,xmm0
> + movdqa xmm7,xmm10
> +
> + movdqa xmm2,xmm10
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm10
> + pslld xmm2,7
> + movdqa XMMWORD[(160-128)+rax],xmm5
> + paddd xmm5,xmm13
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[((-64))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm10
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm10
> + pslld xmm2,26-21
> + pandn xmm0,xmm12
> + pand xmm3,xmm11
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm14
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm14
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm15
> + movdqa xmm7,xmm14
> + pslld xmm2,10
> + pxor xmm3,xmm14
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm13,xmm15
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm13,xmm4
> + paddd xmm9,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm13,xmm5
> + paddd xmm13,xmm7
> + movdqa xmm5,XMMWORD[((192-128))+rax]
> + paddd xmm6,XMMWORD[((64-128))+rax]
> +
> + movdqa xmm7,xmm5
> + movdqa xmm1,xmm5
> + psrld xmm7,3
> + movdqa xmm2,xmm5
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((144-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm4,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm4
> +
> + psrld xmm4,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + psrld xmm4,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm4
> + pxor xmm0,xmm1
> + paddd xmm6,xmm0
> + movdqa xmm7,xmm9
> +
> + movdqa xmm2,xmm9
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm9
> + pslld xmm2,7
> + movdqa XMMWORD[(176-128)+rax],xmm6
> + paddd xmm6,xmm12
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm6,XMMWORD[((-32))+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm9
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm9
> + pslld xmm2,26-21
> + pandn xmm0,xmm11
> + pand xmm4,xmm10
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm13
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm13
> + psrld xmm1,2
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm14
> + movdqa xmm7,xmm13
> + pslld xmm2,10
> + pxor xmm4,xmm13
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm6,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm12,xmm14
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm12,xmm3
> + paddd xmm8,xmm6
> + pxor xmm7,xmm2
> +
> + paddd xmm12,xmm6
> + paddd xmm12,xmm7
> + movdqa xmm6,XMMWORD[((208-128))+rax]
> + paddd xmm5,XMMWORD[((80-128))+rax]
> +
> + movdqa xmm7,xmm6
> + movdqa xmm1,xmm6
> + psrld xmm7,3
> + movdqa xmm2,xmm6
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((160-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm3,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm3
> +
> + psrld xmm3,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + psrld xmm3,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm3
> + pxor xmm0,xmm1
> + paddd xmm5,xmm0
> + movdqa xmm7,xmm8
> +
> + movdqa xmm2,xmm8
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm8
> + pslld xmm2,7
> + movdqa XMMWORD[(192-128)+rax],xmm5
> + paddd xmm5,xmm11
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm8
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm8
> + pslld xmm2,26-21
> + pandn xmm0,xmm10
> + pand xmm3,xmm9
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm12
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm12
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm13
> + movdqa xmm7,xmm12
> + pslld xmm2,10
> + pxor xmm3,xmm12
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm11,xmm13
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm11,xmm4
> + paddd xmm15,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm11,xmm5
> + paddd xmm11,xmm7
> + movdqa xmm5,XMMWORD[((224-128))+rax]
> + paddd xmm6,XMMWORD[((96-128))+rax]
> +
> + movdqa xmm7,xmm5
> + movdqa xmm1,xmm5
> + psrld xmm7,3
> + movdqa xmm2,xmm5
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((176-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm4,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm4
> +
> + psrld xmm4,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + psrld xmm4,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm4
> + pxor xmm0,xmm1
> + paddd xmm6,xmm0
> + movdqa xmm7,xmm15
> +
> + movdqa xmm2,xmm15
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm15
> + pslld xmm2,7
> + movdqa XMMWORD[(208-128)+rax],xmm6
> + paddd xmm6,xmm10
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm6,XMMWORD[32+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm15
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm15
> + pslld xmm2,26-21
> + pandn xmm0,xmm9
> + pand xmm4,xmm8
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm11
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm11
> + psrld xmm1,2
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm12
> + movdqa xmm7,xmm11
> + pslld xmm2,10
> + pxor xmm4,xmm11
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm6,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm10,xmm12
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm10,xmm3
> + paddd xmm14,xmm6
> + pxor xmm7,xmm2
> +
> + paddd xmm10,xmm6
> + paddd xmm10,xmm7
> + movdqa xmm6,XMMWORD[((240-128))+rax]
> + paddd xmm5,XMMWORD[((112-128))+rax]
> +
> + movdqa xmm7,xmm6
> + movdqa xmm1,xmm6
> + psrld xmm7,3
> + movdqa xmm2,xmm6
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((192-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm3,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm3
> +
> + psrld xmm3,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + psrld xmm3,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm3
> + pxor xmm0,xmm1
> + paddd xmm5,xmm0
> + movdqa xmm7,xmm14
> +
> + movdqa xmm2,xmm14
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm14
> + pslld xmm2,7
> + movdqa XMMWORD[(224-128)+rax],xmm5
> + paddd xmm5,xmm9
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm5,XMMWORD[64+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm14
> +
> + pxor xmm7,xmm2
> + movdqa xmm3,xmm14
> + pslld xmm2,26-21
> + pandn xmm0,xmm8
> + pand xmm3,xmm15
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm10
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm10
> + psrld xmm1,2
> + paddd xmm5,xmm7
> + pxor xmm0,xmm3
> + movdqa xmm3,xmm11
> + movdqa xmm7,xmm10
> + pslld xmm2,10
> + pxor xmm3,xmm10
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm5,xmm0
> + pslld xmm2,19-10
> + pand xmm4,xmm3
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm9,xmm11
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm9,xmm4
> + paddd xmm13,xmm5
> + pxor xmm7,xmm2
> +
> + paddd xmm9,xmm5
> + paddd xmm9,xmm7
> + movdqa xmm5,XMMWORD[((0-128))+rax]
> + paddd xmm6,XMMWORD[((128-128))+rax]
> +
> + movdqa xmm7,xmm5
> + movdqa xmm1,xmm5
> + psrld xmm7,3
> + movdqa xmm2,xmm5
> +
> + psrld xmm1,7
> + movdqa xmm0,XMMWORD[((208-128))+rax]
> + pslld xmm2,14
> + pxor xmm7,xmm1
> + psrld xmm1,18-7
> + movdqa xmm4,xmm0
> + pxor xmm7,xmm2
> + pslld xmm2,25-14
> + pxor xmm7,xmm1
> + psrld xmm0,10
> + movdqa xmm1,xmm4
> +
> + psrld xmm4,17
> + pxor xmm7,xmm2
> + pslld xmm1,13
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + psrld xmm4,19-17
> + pxor xmm0,xmm1
> + pslld xmm1,15-13
> + pxor xmm0,xmm4
> + pxor xmm0,xmm1
> + paddd xmm6,xmm0
> + movdqa xmm7,xmm13
> +
> + movdqa xmm2,xmm13
> +
> + psrld xmm7,6
> + movdqa xmm1,xmm13
> + pslld xmm2,7
> + movdqa XMMWORD[(240-128)+rax],xmm6
> + paddd xmm6,xmm8
> +
> + psrld xmm1,11
> + pxor xmm7,xmm2
> + pslld xmm2,21-7
> + paddd xmm6,XMMWORD[96+rbp]
> + pxor xmm7,xmm1
> +
> + psrld xmm1,25-11
> + movdqa xmm0,xmm13
> +
> + pxor xmm7,xmm2
> + movdqa xmm4,xmm13
> + pslld xmm2,26-21
> + pandn xmm0,xmm15
> + pand xmm4,xmm14
> + pxor xmm7,xmm1
> +
> +
> + movdqa xmm1,xmm9
> + pxor xmm7,xmm2
> + movdqa xmm2,xmm9
> + psrld xmm1,2
> + paddd xmm6,xmm7
> + pxor xmm0,xmm4
> + movdqa xmm4,xmm10
> + movdqa xmm7,xmm9
> + pslld xmm2,10
> + pxor xmm4,xmm9
> +
> +
> + psrld xmm7,13
> + pxor xmm1,xmm2
> + paddd xmm6,xmm0
> + pslld xmm2,19-10
> + pand xmm3,xmm4
> + pxor xmm1,xmm7
> +
> +
> + psrld xmm7,22-13
> + pxor xmm1,xmm2
> + movdqa xmm8,xmm10
> + pslld xmm2,30-19
> + pxor xmm7,xmm1
> + pxor xmm8,xmm3
> + paddd xmm12,xmm6
> + pxor xmm7,xmm2
> +
> + paddd xmm8,xmm6
> + paddd xmm8,xmm7
> + lea rbp,[256+rbp]
> + dec ecx
> + jnz NEAR $L$oop_16_xx
> +
> + mov ecx,1
> + lea rbp,[((K256+128))]
> +
> + movdqa xmm7,XMMWORD[rbx]
> + cmp ecx,DWORD[rbx]
> + pxor xmm0,xmm0
> + cmovge r8,rbp
> + cmp ecx,DWORD[4+rbx]
> + movdqa xmm6,xmm7
> + cmovge r9,rbp
> + cmp ecx,DWORD[8+rbx]
> + pcmpgtd xmm6,xmm0
> + cmovge r10,rbp
> + cmp ecx,DWORD[12+rbx]
> + paddd xmm7,xmm6
> + cmovge r11,rbp
> +
> + movdqu xmm0,XMMWORD[((0-128))+rdi]
> + pand xmm8,xmm6
> + movdqu xmm1,XMMWORD[((32-128))+rdi]
> + pand xmm9,xmm6
> + movdqu xmm2,XMMWORD[((64-128))+rdi]
> + pand xmm10,xmm6
> + movdqu xmm5,XMMWORD[((96-128))+rdi]
> + pand xmm11,xmm6
> + paddd xmm8,xmm0
> + movdqu xmm0,XMMWORD[((128-128))+rdi]
> + pand xmm12,xmm6
> + paddd xmm9,xmm1
> + movdqu xmm1,XMMWORD[((160-128))+rdi]
> + pand xmm13,xmm6
> + paddd xmm10,xmm2
> + movdqu xmm2,XMMWORD[((192-128))+rdi]
> + pand xmm14,xmm6
> + paddd xmm11,xmm5
> + movdqu xmm5,XMMWORD[((224-128))+rdi]
> + pand xmm15,xmm6
> + paddd xmm12,xmm0
> + paddd xmm13,xmm1
> + movdqu XMMWORD[(0-128)+rdi],xmm8
> + paddd xmm14,xmm2
> + movdqu XMMWORD[(32-128)+rdi],xmm9
> + paddd xmm15,xmm5
> + movdqu XMMWORD[(64-128)+rdi],xmm10
> + movdqu XMMWORD[(96-128)+rdi],xmm11
> + movdqu XMMWORD[(128-128)+rdi],xmm12
> + movdqu XMMWORD[(160-128)+rdi],xmm13
> + movdqu XMMWORD[(192-128)+rdi],xmm14
> + movdqu XMMWORD[(224-128)+rdi],xmm15
> +
> + movdqa XMMWORD[rbx],xmm7
> + movdqa xmm6,XMMWORD[$L$pbswap]
> + dec edx
> + jnz NEAR $L$oop
> +
> + mov edx,DWORD[280+rsp]
> + lea rdi,[16+rdi]
> + lea rsi,[64+rsi]
> + dec edx
> + jnz NEAR $L$oop_grande
> +
> +$L$done:
> + mov rax,QWORD[272+rsp]
> +
> + movaps xmm6,XMMWORD[((-184))+rax]
> + movaps xmm7,XMMWORD[((-168))+rax]
> + movaps xmm8,XMMWORD[((-152))+rax]
> + movaps xmm9,XMMWORD[((-136))+rax]
> + movaps xmm10,XMMWORD[((-120))+rax]
> + movaps xmm11,XMMWORD[((-104))+rax]
> + movaps xmm12,XMMWORD[((-88))+rax]
> + movaps xmm13,XMMWORD[((-72))+rax]
> + movaps xmm14,XMMWORD[((-56))+rax]
> + movaps xmm15,XMMWORD[((-40))+rax]
> + mov rbp,QWORD[((-16))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> +
> + lea rsp,[rax]
> +
> +$L$epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha256_multi_block:
> +
> +ALIGN 32
> +sha256_multi_block_shaext:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha256_multi_block_shaext:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> +_shaext_shortcut:
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + lea rsp,[((-168))+rsp]
> + movaps XMMWORD[rsp],xmm6
> + movaps XMMWORD[16+rsp],xmm7
> + movaps XMMWORD[32+rsp],xmm8
> + movaps XMMWORD[48+rsp],xmm9
> + movaps XMMWORD[(-120)+rax],xmm10
> + movaps XMMWORD[(-104)+rax],xmm11
> + movaps XMMWORD[(-88)+rax],xmm12
> + movaps XMMWORD[(-72)+rax],xmm13
> + movaps XMMWORD[(-56)+rax],xmm14
> + movaps XMMWORD[(-40)+rax],xmm15
> + sub rsp,288
> + shl edx,1
> + and rsp,-256
> + lea rdi,[128+rdi]
> + mov QWORD[272+rsp],rax
> +$L$body_shaext:
> + lea rbx,[256+rsp]
> + lea rbp,[((K256_shaext+128))]
> +
> +$L$oop_grande_shaext:
> + mov DWORD[280+rsp],edx
> + xor edx,edx
> + mov r8,QWORD[rsi]
> + mov ecx,DWORD[8+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[rbx],ecx
> + cmovle r8,rsp
> + mov r9,QWORD[16+rsi]
> + mov ecx,DWORD[24+rsi]
> + cmp ecx,edx
> + cmovg edx,ecx
> + test ecx,ecx
> + mov DWORD[4+rbx],ecx
> + cmovle r9,rsp
> + test edx,edx
> + jz NEAR $L$done_shaext
> +
> + movq xmm12,QWORD[((0-128))+rdi]
> + movq xmm4,QWORD[((32-128))+rdi]
> + movq xmm13,QWORD[((64-128))+rdi]
> + movq xmm5,QWORD[((96-128))+rdi]
> + movq xmm8,QWORD[((128-128))+rdi]
> + movq xmm9,QWORD[((160-128))+rdi]
> + movq xmm10,QWORD[((192-128))+rdi]
> + movq xmm11,QWORD[((224-128))+rdi]
> +
> + punpckldq xmm12,xmm4
> + punpckldq xmm13,xmm5
> + punpckldq xmm8,xmm9
> + punpckldq xmm10,xmm11
> + movdqa xmm3,XMMWORD[((K256_shaext-16))]
> +
> + movdqa xmm14,xmm12
> + movdqa xmm15,xmm13
> + punpcklqdq xmm12,xmm8
> + punpcklqdq xmm13,xmm10
> + punpckhqdq xmm14,xmm8
> + punpckhqdq xmm15,xmm10
> +
> + pshufd xmm12,xmm12,27
> + pshufd xmm13,xmm13,27
> + pshufd xmm14,xmm14,27
> + pshufd xmm15,xmm15,27
> + jmp NEAR $L$oop_shaext
> +
> +ALIGN 32
> +$L$oop_shaext:
> + movdqu xmm4,XMMWORD[r8]
> + movdqu xmm8,XMMWORD[r9]
> + movdqu xmm5,XMMWORD[16+r8]
> + movdqu xmm9,XMMWORD[16+r9]
> + movdqu xmm6,XMMWORD[32+r8]
> +DB 102,15,56,0,227
> + movdqu xmm10,XMMWORD[32+r9]
> +DB 102,68,15,56,0,195
> + movdqu xmm7,XMMWORD[48+r8]
> + lea r8,[64+r8]
> + movdqu xmm11,XMMWORD[48+r9]
> + lea r9,[64+r9]
> +
> + movdqa xmm0,XMMWORD[((0-128))+rbp]
> +DB 102,15,56,0,235
> + paddd xmm0,xmm4
> + pxor xmm4,xmm12
> + movdqa xmm1,xmm0
> + movdqa xmm2,XMMWORD[((0-128))+rbp]
> +DB 102,68,15,56,0,203
> + paddd xmm2,xmm8
> + movdqa XMMWORD[80+rsp],xmm13
> +DB 69,15,56,203,236
> + pxor xmm8,xmm14
> + movdqa xmm0,xmm2
> + movdqa XMMWORD[112+rsp],xmm15
> +DB 69,15,56,203,254
> + pshufd xmm0,xmm1,0x0e
> + pxor xmm4,xmm12
> + movdqa XMMWORD[64+rsp],xmm12
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + pxor xmm8,xmm14
> + movdqa XMMWORD[96+rsp],xmm14
> + movdqa xmm1,XMMWORD[((16-128))+rbp]
> + paddd xmm1,xmm5
> +DB 102,15,56,0,243
> +DB 69,15,56,203,247
> +
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((16-128))+rbp]
> + paddd xmm2,xmm9
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + prefetcht0 [127+r8]
> +DB 102,15,56,0,251
> +DB 102,68,15,56,0,211
> + prefetcht0 [127+r9]
> +DB 69,15,56,203,254
> + pshufd xmm0,xmm1,0x0e
> +DB 102,68,15,56,0,219
> +DB 15,56,204,229
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((32-128))+rbp]
> + paddd xmm1,xmm6
> +DB 69,15,56,203,247
> +
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((32-128))+rbp]
> + paddd xmm2,xmm10
> +DB 69,15,56,203,236
> +DB 69,15,56,204,193
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm7
> +DB 69,15,56,203,254
> + pshufd xmm0,xmm1,0x0e
> +DB 102,15,58,15,222,4
> + paddd xmm4,xmm3
> + movdqa xmm3,xmm11
> +DB 102,65,15,58,15,218,4
> +DB 15,56,204,238
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((48-128))+rbp]
> + paddd xmm1,xmm7
> +DB 69,15,56,203,247
> +DB 69,15,56,204,202
> +
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((48-128))+rbp]
> + paddd xmm8,xmm3
> + paddd xmm2,xmm11
> +DB 15,56,205,231
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm4
> +DB 102,15,58,15,223,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,195
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm5,xmm3
> + movdqa xmm3,xmm8
> +DB 102,65,15,58,15,219,4
> +DB 15,56,204,247
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((64-128))+rbp]
> + paddd xmm1,xmm4
> +DB 69,15,56,203,247
> +DB 69,15,56,204,211
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((64-128))+rbp]
> + paddd xmm9,xmm3
> + paddd xmm2,xmm8
> +DB 15,56,205,236
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm5
> +DB 102,15,58,15,220,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,200
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm6,xmm3
> + movdqa xmm3,xmm9
> +DB 102,65,15,58,15,216,4
> +DB 15,56,204,252
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((80-128))+rbp]
> + paddd xmm1,xmm5
> +DB 69,15,56,203,247
> +DB 69,15,56,204,216
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((80-128))+rbp]
> + paddd xmm10,xmm3
> + paddd xmm2,xmm9
> +DB 15,56,205,245
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm6
> +DB 102,15,58,15,221,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,209
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm7,xmm3
> + movdqa xmm3,xmm10
> +DB 102,65,15,58,15,217,4
> +DB 15,56,204,229
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((96-128))+rbp]
> + paddd xmm1,xmm6
> +DB 69,15,56,203,247
> +DB 69,15,56,204,193
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((96-128))+rbp]
> + paddd xmm11,xmm3
> + paddd xmm2,xmm10
> +DB 15,56,205,254
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm7
> +DB 102,15,58,15,222,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,218
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm4,xmm3
> + movdqa xmm3,xmm11
> +DB 102,65,15,58,15,218,4
> +DB 15,56,204,238
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((112-128))+rbp]
> + paddd xmm1,xmm7
> +DB 69,15,56,203,247
> +DB 69,15,56,204,202
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((112-128))+rbp]
> + paddd xmm8,xmm3
> + paddd xmm2,xmm11
> +DB 15,56,205,231
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm4
> +DB 102,15,58,15,223,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,195
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm5,xmm3
> + movdqa xmm3,xmm8
> +DB 102,65,15,58,15,219,4
> +DB 15,56,204,247
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((128-128))+rbp]
> + paddd xmm1,xmm4
> +DB 69,15,56,203,247
> +DB 69,15,56,204,211
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((128-128))+rbp]
> + paddd xmm9,xmm3
> + paddd xmm2,xmm8
> +DB 15,56,205,236
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm5
> +DB 102,15,58,15,220,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,200
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm6,xmm3
> + movdqa xmm3,xmm9
> +DB 102,65,15,58,15,216,4
> +DB 15,56,204,252
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((144-128))+rbp]
> + paddd xmm1,xmm5
> +DB 69,15,56,203,247
> +DB 69,15,56,204,216
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((144-128))+rbp]
> + paddd xmm10,xmm3
> + paddd xmm2,xmm9
> +DB 15,56,205,245
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm6
> +DB 102,15,58,15,221,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,209
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm7,xmm3
> + movdqa xmm3,xmm10
> +DB 102,65,15,58,15,217,4
> +DB 15,56,204,229
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((160-128))+rbp]
> + paddd xmm1,xmm6
> +DB 69,15,56,203,247
> +DB 69,15,56,204,193
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((160-128))+rbp]
> + paddd xmm11,xmm3
> + paddd xmm2,xmm10
> +DB 15,56,205,254
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm7
> +DB 102,15,58,15,222,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,218
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm4,xmm3
> + movdqa xmm3,xmm11
> +DB 102,65,15,58,15,218,4
> +DB 15,56,204,238
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((176-128))+rbp]
> + paddd xmm1,xmm7
> +DB 69,15,56,203,247
> +DB 69,15,56,204,202
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((176-128))+rbp]
> + paddd xmm8,xmm3
> + paddd xmm2,xmm11
> +DB 15,56,205,231
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm4
> +DB 102,15,58,15,223,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,195
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm5,xmm3
> + movdqa xmm3,xmm8
> +DB 102,65,15,58,15,219,4
> +DB 15,56,204,247
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((192-128))+rbp]
> + paddd xmm1,xmm4
> +DB 69,15,56,203,247
> +DB 69,15,56,204,211
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((192-128))+rbp]
> + paddd xmm9,xmm3
> + paddd xmm2,xmm8
> +DB 15,56,205,236
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm5
> +DB 102,15,58,15,220,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,200
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm6,xmm3
> + movdqa xmm3,xmm9
> +DB 102,65,15,58,15,216,4
> +DB 15,56,204,252
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((208-128))+rbp]
> + paddd xmm1,xmm5
> +DB 69,15,56,203,247
> +DB 69,15,56,204,216
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((208-128))+rbp]
> + paddd xmm10,xmm3
> + paddd xmm2,xmm9
> +DB 15,56,205,245
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + movdqa xmm3,xmm6
> +DB 102,15,58,15,221,4
> +DB 69,15,56,203,254
> +DB 69,15,56,205,209
> + pshufd xmm0,xmm1,0x0e
> + paddd xmm7,xmm3
> + movdqa xmm3,xmm10
> +DB 102,65,15,58,15,217,4
> + nop
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm1,XMMWORD[((224-128))+rbp]
> + paddd xmm1,xmm6
> +DB 69,15,56,203,247
> +
> + movdqa xmm0,xmm1
> + movdqa xmm2,XMMWORD[((224-128))+rbp]
> + paddd xmm11,xmm3
> + paddd xmm2,xmm10
> +DB 15,56,205,254
> + nop
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + mov ecx,1
> + pxor xmm6,xmm6
> +DB 69,15,56,203,254
> +DB 69,15,56,205,218
> + pshufd xmm0,xmm1,0x0e
> + movdqa xmm1,XMMWORD[((240-128))+rbp]
> + paddd xmm1,xmm7
> + movq xmm7,QWORD[rbx]
> + nop
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + movdqa xmm2,XMMWORD[((240-128))+rbp]
> + paddd xmm2,xmm11
> +DB 69,15,56,203,247
> +
> + movdqa xmm0,xmm1
> + cmp ecx,DWORD[rbx]
> + cmovge r8,rsp
> + cmp ecx,DWORD[4+rbx]
> + cmovge r9,rsp
> + pshufd xmm9,xmm7,0x00
> +DB 69,15,56,203,236
> + movdqa xmm0,xmm2
> + pshufd xmm10,xmm7,0x55
> + movdqa xmm11,xmm7
> +DB 69,15,56,203,254
> + pshufd xmm0,xmm1,0x0e
> + pcmpgtd xmm9,xmm6
> + pcmpgtd xmm10,xmm6
> +DB 69,15,56,203,229
> + pshufd xmm0,xmm2,0x0e
> + pcmpgtd xmm11,xmm6
> + movdqa xmm3,XMMWORD[((K256_shaext-16))]
> +DB 69,15,56,203,247
> +
> + pand xmm13,xmm9
> + pand xmm15,xmm10
> + pand xmm12,xmm9
> + pand xmm14,xmm10
> + paddd xmm11,xmm7
> +
> + paddd xmm13,XMMWORD[80+rsp]
> + paddd xmm15,XMMWORD[112+rsp]
> + paddd xmm12,XMMWORD[64+rsp]
> + paddd xmm14,XMMWORD[96+rsp]
> +
> + movq QWORD[rbx],xmm11
> + dec edx
> + jnz NEAR $L$oop_shaext
> +
> + mov edx,DWORD[280+rsp]
> +
> + pshufd xmm12,xmm12,27
> + pshufd xmm13,xmm13,27
> + pshufd xmm14,xmm14,27
> + pshufd xmm15,xmm15,27
> +
> + movdqa xmm5,xmm12
> + movdqa xmm6,xmm13
> + punpckldq xmm12,xmm14
> + punpckhdq xmm5,xmm14
> + punpckldq xmm13,xmm15
> + punpckhdq xmm6,xmm15
> +
> + movq QWORD[(0-128)+rdi],xmm12
> + psrldq xmm12,8
> + movq QWORD[(128-128)+rdi],xmm5
> + psrldq xmm5,8
> + movq QWORD[(32-128)+rdi],xmm12
> + movq QWORD[(160-128)+rdi],xmm5
> +
> + movq QWORD[(64-128)+rdi],xmm13
> + psrldq xmm13,8
> + movq QWORD[(192-128)+rdi],xmm6
> + psrldq xmm6,8
> + movq QWORD[(96-128)+rdi],xmm13
> + movq QWORD[(224-128)+rdi],xmm6
> +
> + lea rdi,[8+rdi]
> + lea rsi,[32+rsi]
> + dec edx
> + jnz NEAR $L$oop_grande_shaext
> +
> +$L$done_shaext:
> +
> + movaps xmm6,XMMWORD[((-184))+rax]
> + movaps xmm7,XMMWORD[((-168))+rax]
> + movaps xmm8,XMMWORD[((-152))+rax]
> + movaps xmm9,XMMWORD[((-136))+rax]
> + movaps xmm10,XMMWORD[((-120))+rax]
> + movaps xmm11,XMMWORD[((-104))+rax]
> + movaps xmm12,XMMWORD[((-88))+rax]
> + movaps xmm13,XMMWORD[((-72))+rax]
> + movaps xmm14,XMMWORD[((-56))+rax]
> + movaps xmm15,XMMWORD[((-40))+rax]
> + mov rbp,QWORD[((-16))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> +
> + lea rsp,[rax]
> +
> +$L$epilogue_shaext:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha256_multi_block_shaext:
> +ALIGN 256
> +K256:
> + DD 1116352408,1116352408,1116352408,1116352408
> + DD 1116352408,1116352408,1116352408,1116352408
> + DD 1899447441,1899447441,1899447441,1899447441
> + DD 1899447441,1899447441,1899447441,1899447441
> + DD 3049323471,3049323471,3049323471,3049323471
> + DD 3049323471,3049323471,3049323471,3049323471
> + DD 3921009573,3921009573,3921009573,3921009573
> + DD 3921009573,3921009573,3921009573,3921009573
> + DD 961987163,961987163,961987163,961987163
> + DD 961987163,961987163,961987163,961987163
> + DD 1508970993,1508970993,1508970993,1508970993
> + DD 1508970993,1508970993,1508970993,1508970993
> + DD 2453635748,2453635748,2453635748,2453635748
> + DD 2453635748,2453635748,2453635748,2453635748
> + DD 2870763221,2870763221,2870763221,2870763221
> + DD 2870763221,2870763221,2870763221,2870763221
> + DD 3624381080,3624381080,3624381080,3624381080
> + DD 3624381080,3624381080,3624381080,3624381080
> + DD 310598401,310598401,310598401,310598401
> + DD 310598401,310598401,310598401,310598401
> + DD 607225278,607225278,607225278,607225278
> + DD 607225278,607225278,607225278,607225278
> + DD 1426881987,1426881987,1426881987,1426881987
> + DD 1426881987,1426881987,1426881987,1426881987
> + DD 1925078388,1925078388,1925078388,1925078388
> + DD 1925078388,1925078388,1925078388,1925078388
> + DD 2162078206,2162078206,2162078206,2162078206
> + DD 2162078206,2162078206,2162078206,2162078206
> + DD 2614888103,2614888103,2614888103,2614888103
> + DD 2614888103,2614888103,2614888103,2614888103
> + DD 3248222580,3248222580,3248222580,3248222580
> + DD 3248222580,3248222580,3248222580,3248222580
> + DD 3835390401,3835390401,3835390401,3835390401
> + DD 3835390401,3835390401,3835390401,3835390401
> + DD 4022224774,4022224774,4022224774,4022224774
> + DD 4022224774,4022224774,4022224774,4022224774
> + DD 264347078,264347078,264347078,264347078
> + DD 264347078,264347078,264347078,264347078
> + DD 604807628,604807628,604807628,604807628
> + DD 604807628,604807628,604807628,604807628
> + DD 770255983,770255983,770255983,770255983
> + DD 770255983,770255983,770255983,770255983
> + DD 1249150122,1249150122,1249150122,1249150122
> + DD 1249150122,1249150122,1249150122,1249150122
> + DD 1555081692,1555081692,1555081692,1555081692
> + DD 1555081692,1555081692,1555081692,1555081692
> + DD 1996064986,1996064986,1996064986,1996064986
> + DD 1996064986,1996064986,1996064986,1996064986
> + DD 2554220882,2554220882,2554220882,2554220882
> + DD 2554220882,2554220882,2554220882,2554220882
> + DD 2821834349,2821834349,2821834349,2821834349
> + DD 2821834349,2821834349,2821834349,2821834349
> + DD 2952996808,2952996808,2952996808,2952996808
> + DD 2952996808,2952996808,2952996808,2952996808
> + DD 3210313671,3210313671,3210313671,3210313671
> + DD 3210313671,3210313671,3210313671,3210313671
> + DD 3336571891,3336571891,3336571891,3336571891
> + DD 3336571891,3336571891,3336571891,3336571891
> + DD 3584528711,3584528711,3584528711,3584528711
> + DD 3584528711,3584528711,3584528711,3584528711
> + DD 113926993,113926993,113926993,113926993
> + DD 113926993,113926993,113926993,113926993
> + DD 338241895,338241895,338241895,338241895
> + DD 338241895,338241895,338241895,338241895
> + DD 666307205,666307205,666307205,666307205
> + DD 666307205,666307205,666307205,666307205
> + DD 773529912,773529912,773529912,773529912
> + DD 773529912,773529912,773529912,773529912
> + DD 1294757372,1294757372,1294757372,1294757372
> + DD 1294757372,1294757372,1294757372,1294757372
> + DD 1396182291,1396182291,1396182291,1396182291
> + DD 1396182291,1396182291,1396182291,1396182291
> + DD 1695183700,1695183700,1695183700,1695183700
> + DD 1695183700,1695183700,1695183700,1695183700
> + DD 1986661051,1986661051,1986661051,1986661051
> + DD 1986661051,1986661051,1986661051,1986661051
> + DD 2177026350,2177026350,2177026350,2177026350
> + DD 2177026350,2177026350,2177026350,2177026350
> + DD 2456956037,2456956037,2456956037,2456956037
> + DD 2456956037,2456956037,2456956037,2456956037
> + DD 2730485921,2730485921,2730485921,2730485921
> + DD 2730485921,2730485921,2730485921,2730485921
> + DD 2820302411,2820302411,2820302411,2820302411
> + DD 2820302411,2820302411,2820302411,2820302411
> + DD 3259730800,3259730800,3259730800,3259730800
> + DD 3259730800,3259730800,3259730800,3259730800
> + DD 3345764771,3345764771,3345764771,3345764771
> + DD 3345764771,3345764771,3345764771,3345764771
> + DD 3516065817,3516065817,3516065817,3516065817
> + DD 3516065817,3516065817,3516065817,3516065817
> + DD 3600352804,3600352804,3600352804,3600352804
> + DD 3600352804,3600352804,3600352804,3600352804
> + DD 4094571909,4094571909,4094571909,4094571909
> + DD 4094571909,4094571909,4094571909,4094571909
> + DD 275423344,275423344,275423344,275423344
> + DD 275423344,275423344,275423344,275423344
> + DD 430227734,430227734,430227734,430227734
> + DD 430227734,430227734,430227734,430227734
> + DD 506948616,506948616,506948616,506948616
> + DD 506948616,506948616,506948616,506948616
> + DD 659060556,659060556,659060556,659060556
> + DD 659060556,659060556,659060556,659060556
> + DD 883997877,883997877,883997877,883997877
> + DD 883997877,883997877,883997877,883997877
> + DD 958139571,958139571,958139571,958139571
> + DD 958139571,958139571,958139571,958139571
> + DD 1322822218,1322822218,1322822218,1322822218
> + DD 1322822218,1322822218,1322822218,1322822218
> + DD 1537002063,1537002063,1537002063,1537002063
> + DD 1537002063,1537002063,1537002063,1537002063
> + DD 1747873779,1747873779,1747873779,1747873779
> + DD 1747873779,1747873779,1747873779,1747873779
> + DD 1955562222,1955562222,1955562222,1955562222
> + DD 1955562222,1955562222,1955562222,1955562222
> + DD 2024104815,2024104815,2024104815,2024104815
> + DD 2024104815,2024104815,2024104815,2024104815
> + DD 2227730452,2227730452,2227730452,2227730452
> + DD 2227730452,2227730452,2227730452,2227730452
> + DD 2361852424,2361852424,2361852424,2361852424
> + DD 2361852424,2361852424,2361852424,2361852424
> + DD 2428436474,2428436474,2428436474,2428436474
> + DD 2428436474,2428436474,2428436474,2428436474
> + DD 2756734187,2756734187,2756734187,2756734187
> + DD 2756734187,2756734187,2756734187,2756734187
> + DD 3204031479,3204031479,3204031479,3204031479
> + DD 3204031479,3204031479,3204031479,3204031479
> + DD 3329325298,3329325298,3329325298,3329325298
> + DD 3329325298,3329325298,3329325298,3329325298
> +$L$pbswap:
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +K256_shaext:
> + DD 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> + DD 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> + DD 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> + DD 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> + DD 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> + DD 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> + DD 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> + DD 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> + DD 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> + DD 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> + DD 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> + DD 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> + DD 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> + DD 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> + DD 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> + DD 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> +DB 83,72,65,50,53,54,32,109,117,108,116,105,45,98,108,111
> +DB 99,107,32,116,114,97,110,115,102,111,114,109,32,102,111,114
> +DB 32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71
> +DB 65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112
> +DB 101,110,115,115,108,46,111,114,103,62,0
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$in_prologue
> +
> + mov rax,QWORD[272+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> +
> + lea rsi,[((-24-160))+rax]
> + lea rdi,[512+r8]
> + mov ecx,20
> + DD 0xa548f3fc
> +
> +$L$in_prologue:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_sha256_multi_block wrt ..imagebase
> + DD $L$SEH_end_sha256_multi_block wrt ..imagebase
> + DD $L$SEH_info_sha256_multi_block wrt ..imagebase
> + DD $L$SEH_begin_sha256_multi_block_shaext wrt ..imagebase
> + DD $L$SEH_end_sha256_multi_block_shaext wrt ..imagebase
> + DD $L$SEH_info_sha256_multi_block_shaext wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_sha256_multi_block:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$body wrt ..imagebase,$L$epilogue wrt ..imagebase
> +$L$SEH_info_sha256_multi_block_shaext:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$body_shaext wrt ..imagebase,$L$epilogue_shaext
> wrt ..imagebase
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha256-x86_64.nasm
> b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha256-x86_64.nasm
> new file mode 100644
> index 0000000000..70e49862a3
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha256-x86_64.nasm
> @@ -0,0 +1,3313 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/sha/asm/sha512-x86_64.pl
> +;
> +; Copyright 2005-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +
> +EXTERN OPENSSL_ia32cap_P
> +global sha256_block_data_order
> +
> +ALIGN 16
> +sha256_block_data_order:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha256_block_data_order:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + lea r11,[OPENSSL_ia32cap_P]
> + mov r9d,DWORD[r11]
> + mov r10d,DWORD[4+r11]
> + mov r11d,DWORD[8+r11]
> + test r11d,536870912
> + jnz NEAR _shaext_shortcut
> + test r10d,512
> + jnz NEAR $L$ssse3_shortcut
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + push r15
> +
> + shl rdx,4
> + sub rsp,16*4+4*8
> + lea rdx,[rdx*4+rsi]
> + and rsp,-64
> + mov QWORD[((64+0))+rsp],rdi
> + mov QWORD[((64+8))+rsp],rsi
> + mov QWORD[((64+16))+rsp],rdx
> + mov QWORD[88+rsp],rax
> +
> +$L$prologue:
> +
> + mov eax,DWORD[rdi]
> + mov ebx,DWORD[4+rdi]
> + mov ecx,DWORD[8+rdi]
> + mov edx,DWORD[12+rdi]
> + mov r8d,DWORD[16+rdi]
> + mov r9d,DWORD[20+rdi]
> + mov r10d,DWORD[24+rdi]
> + mov r11d,DWORD[28+rdi]
> + jmp NEAR $L$loop
> +
> +ALIGN 16
> +$L$loop:
> + mov edi,ebx
> + lea rbp,[K256]
> + xor edi,ecx
> + mov r12d,DWORD[rsi]
> + mov r13d,r8d
> + mov r14d,eax
> + bswap r12d
> + ror r13d,14
> + mov r15d,r9d
> +
> + xor r13d,r8d
> + ror r14d,9
> + xor r15d,r10d
> +
> + mov DWORD[rsp],r12d
> + xor r14d,eax
> + and r15d,r8d
> +
> + ror r13d,5
> + add r12d,r11d
> + xor r15d,r10d
> +
> + ror r14d,11
> + xor r13d,r8d
> + add r12d,r15d
> +
> + mov r15d,eax
> + add r12d,DWORD[rbp]
> + xor r14d,eax
> +
> + xor r15d,ebx
> + ror r13d,6
> + mov r11d,ebx
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r11d,edi
> + add edx,r12d
> + add r11d,r12d
> +
> + lea rbp,[4+rbp]
> + add r11d,r14d
> + mov r12d,DWORD[4+rsi]
> + mov r13d,edx
> + mov r14d,r11d
> + bswap r12d
> + ror r13d,14
> + mov edi,r8d
> +
> + xor r13d,edx
> + ror r14d,9
> + xor edi,r9d
> +
> + mov DWORD[4+rsp],r12d
> + xor r14d,r11d
> + and edi,edx
> +
> + ror r13d,5
> + add r12d,r10d
> + xor edi,r9d
> +
> + ror r14d,11
> + xor r13d,edx
> + add r12d,edi
> +
> + mov edi,r11d
> + add r12d,DWORD[rbp]
> + xor r14d,r11d
> +
> + xor edi,eax
> + ror r13d,6
> + mov r10d,eax
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r10d,r15d
> + add ecx,r12d
> + add r10d,r12d
> +
> + lea rbp,[4+rbp]
> + add r10d,r14d
> + mov r12d,DWORD[8+rsi]
> + mov r13d,ecx
> + mov r14d,r10d
> + bswap r12d
> + ror r13d,14
> + mov r15d,edx
> +
> + xor r13d,ecx
> + ror r14d,9
> + xor r15d,r8d
> +
> + mov DWORD[8+rsp],r12d
> + xor r14d,r10d
> + and r15d,ecx
> +
> + ror r13d,5
> + add r12d,r9d
> + xor r15d,r8d
> +
> + ror r14d,11
> + xor r13d,ecx
> + add r12d,r15d
> +
> + mov r15d,r10d
> + add r12d,DWORD[rbp]
> + xor r14d,r10d
> +
> + xor r15d,r11d
> + ror r13d,6
> + mov r9d,r11d
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r9d,edi
> + add ebx,r12d
> + add r9d,r12d
> +
> + lea rbp,[4+rbp]
> + add r9d,r14d
> + mov r12d,DWORD[12+rsi]
> + mov r13d,ebx
> + mov r14d,r9d
> + bswap r12d
> + ror r13d,14
> + mov edi,ecx
> +
> + xor r13d,ebx
> + ror r14d,9
> + xor edi,edx
> +
> + mov DWORD[12+rsp],r12d
> + xor r14d,r9d
> + and edi,ebx
> +
> + ror r13d,5
> + add r12d,r8d
> + xor edi,edx
> +
> + ror r14d,11
> + xor r13d,ebx
> + add r12d,edi
> +
> + mov edi,r9d
> + add r12d,DWORD[rbp]
> + xor r14d,r9d
> +
> + xor edi,r10d
> + ror r13d,6
> + mov r8d,r10d
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r8d,r15d
> + add eax,r12d
> + add r8d,r12d
> +
> + lea rbp,[20+rbp]
> + add r8d,r14d
> + mov r12d,DWORD[16+rsi]
> + mov r13d,eax
> + mov r14d,r8d
> + bswap r12d
> + ror r13d,14
> + mov r15d,ebx
> +
> + xor r13d,eax
> + ror r14d,9
> + xor r15d,ecx
> +
> + mov DWORD[16+rsp],r12d
> + xor r14d,r8d
> + and r15d,eax
> +
> + ror r13d,5
> + add r12d,edx
> + xor r15d,ecx
> +
> + ror r14d,11
> + xor r13d,eax
> + add r12d,r15d
> +
> + mov r15d,r8d
> + add r12d,DWORD[rbp]
> + xor r14d,r8d
> +
> + xor r15d,r9d
> + ror r13d,6
> + mov edx,r9d
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor edx,edi
> + add r11d,r12d
> + add edx,r12d
> +
> + lea rbp,[4+rbp]
> + add edx,r14d
> + mov r12d,DWORD[20+rsi]
> + mov r13d,r11d
> + mov r14d,edx
> + bswap r12d
> + ror r13d,14
> + mov edi,eax
> +
> + xor r13d,r11d
> + ror r14d,9
> + xor edi,ebx
> +
> + mov DWORD[20+rsp],r12d
> + xor r14d,edx
> + and edi,r11d
> +
> + ror r13d,5
> + add r12d,ecx
> + xor edi,ebx
> +
> + ror r14d,11
> + xor r13d,r11d
> + add r12d,edi
> +
> + mov edi,edx
> + add r12d,DWORD[rbp]
> + xor r14d,edx
> +
> + xor edi,r8d
> + ror r13d,6
> + mov ecx,r8d
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor ecx,r15d
> + add r10d,r12d
> + add ecx,r12d
> +
> + lea rbp,[4+rbp]
> + add ecx,r14d
> + mov r12d,DWORD[24+rsi]
> + mov r13d,r10d
> + mov r14d,ecx
> + bswap r12d
> + ror r13d,14
> + mov r15d,r11d
> +
> + xor r13d,r10d
> + ror r14d,9
> + xor r15d,eax
> +
> + mov DWORD[24+rsp],r12d
> + xor r14d,ecx
> + and r15d,r10d
> +
> + ror r13d,5
> + add r12d,ebx
> + xor r15d,eax
> +
> + ror r14d,11
> + xor r13d,r10d
> + add r12d,r15d
> +
> + mov r15d,ecx
> + add r12d,DWORD[rbp]
> + xor r14d,ecx
> +
> + xor r15d,edx
> + ror r13d,6
> + mov ebx,edx
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor ebx,edi
> + add r9d,r12d
> + add ebx,r12d
> +
> + lea rbp,[4+rbp]
> + add ebx,r14d
> + mov r12d,DWORD[28+rsi]
> + mov r13d,r9d
> + mov r14d,ebx
> + bswap r12d
> + ror r13d,14
> + mov edi,r10d
> +
> + xor r13d,r9d
> + ror r14d,9
> + xor edi,r11d
> +
> + mov DWORD[28+rsp],r12d
> + xor r14d,ebx
> + and edi,r9d
> +
> + ror r13d,5
> + add r12d,eax
> + xor edi,r11d
> +
> + ror r14d,11
> + xor r13d,r9d
> + add r12d,edi
> +
> + mov edi,ebx
> + add r12d,DWORD[rbp]
> + xor r14d,ebx
> +
> + xor edi,ecx
> + ror r13d,6
> + mov eax,ecx
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor eax,r15d
> + add r8d,r12d
> + add eax,r12d
> +
> + lea rbp,[20+rbp]
> + add eax,r14d
> + mov r12d,DWORD[32+rsi]
> + mov r13d,r8d
> + mov r14d,eax
> + bswap r12d
> + ror r13d,14
> + mov r15d,r9d
> +
> + xor r13d,r8d
> + ror r14d,9
> + xor r15d,r10d
> +
> + mov DWORD[32+rsp],r12d
> + xor r14d,eax
> + and r15d,r8d
> +
> + ror r13d,5
> + add r12d,r11d
> + xor r15d,r10d
> +
> + ror r14d,11
> + xor r13d,r8d
> + add r12d,r15d
> +
> + mov r15d,eax
> + add r12d,DWORD[rbp]
> + xor r14d,eax
> +
> + xor r15d,ebx
> + ror r13d,6
> + mov r11d,ebx
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r11d,edi
> + add edx,r12d
> + add r11d,r12d
> +
> + lea rbp,[4+rbp]
> + add r11d,r14d
> + mov r12d,DWORD[36+rsi]
> + mov r13d,edx
> + mov r14d,r11d
> + bswap r12d
> + ror r13d,14
> + mov edi,r8d
> +
> + xor r13d,edx
> + ror r14d,9
> + xor edi,r9d
> +
> + mov DWORD[36+rsp],r12d
> + xor r14d,r11d
> + and edi,edx
> +
> + ror r13d,5
> + add r12d,r10d
> + xor edi,r9d
> +
> + ror r14d,11
> + xor r13d,edx
> + add r12d,edi
> +
> + mov edi,r11d
> + add r12d,DWORD[rbp]
> + xor r14d,r11d
> +
> + xor edi,eax
> + ror r13d,6
> + mov r10d,eax
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r10d,r15d
> + add ecx,r12d
> + add r10d,r12d
> +
> + lea rbp,[4+rbp]
> + add r10d,r14d
> + mov r12d,DWORD[40+rsi]
> + mov r13d,ecx
> + mov r14d,r10d
> + bswap r12d
> + ror r13d,14
> + mov r15d,edx
> +
> + xor r13d,ecx
> + ror r14d,9
> + xor r15d,r8d
> +
> + mov DWORD[40+rsp],r12d
> + xor r14d,r10d
> + and r15d,ecx
> +
> + ror r13d,5
> + add r12d,r9d
> + xor r15d,r8d
> +
> + ror r14d,11
> + xor r13d,ecx
> + add r12d,r15d
> +
> + mov r15d,r10d
> + add r12d,DWORD[rbp]
> + xor r14d,r10d
> +
> + xor r15d,r11d
> + ror r13d,6
> + mov r9d,r11d
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r9d,edi
> + add ebx,r12d
> + add r9d,r12d
> +
> + lea rbp,[4+rbp]
> + add r9d,r14d
> + mov r12d,DWORD[44+rsi]
> + mov r13d,ebx
> + mov r14d,r9d
> + bswap r12d
> + ror r13d,14
> + mov edi,ecx
> +
> + xor r13d,ebx
> + ror r14d,9
> + xor edi,edx
> +
> + mov DWORD[44+rsp],r12d
> + xor r14d,r9d
> + and edi,ebx
> +
> + ror r13d,5
> + add r12d,r8d
> + xor edi,edx
> +
> + ror r14d,11
> + xor r13d,ebx
> + add r12d,edi
> +
> + mov edi,r9d
> + add r12d,DWORD[rbp]
> + xor r14d,r9d
> +
> + xor edi,r10d
> + ror r13d,6
> + mov r8d,r10d
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r8d,r15d
> + add eax,r12d
> + add r8d,r12d
> +
> + lea rbp,[20+rbp]
> + add r8d,r14d
> + mov r12d,DWORD[48+rsi]
> + mov r13d,eax
> + mov r14d,r8d
> + bswap r12d
> + ror r13d,14
> + mov r15d,ebx
> +
> + xor r13d,eax
> + ror r14d,9
> + xor r15d,ecx
> +
> + mov DWORD[48+rsp],r12d
> + xor r14d,r8d
> + and r15d,eax
> +
> + ror r13d,5
> + add r12d,edx
> + xor r15d,ecx
> +
> + ror r14d,11
> + xor r13d,eax
> + add r12d,r15d
> +
> + mov r15d,r8d
> + add r12d,DWORD[rbp]
> + xor r14d,r8d
> +
> + xor r15d,r9d
> + ror r13d,6
> + mov edx,r9d
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor edx,edi
> + add r11d,r12d
> + add edx,r12d
> +
> + lea rbp,[4+rbp]
> + add edx,r14d
> + mov r12d,DWORD[52+rsi]
> + mov r13d,r11d
> + mov r14d,edx
> + bswap r12d
> + ror r13d,14
> + mov edi,eax
> +
> + xor r13d,r11d
> + ror r14d,9
> + xor edi,ebx
> +
> + mov DWORD[52+rsp],r12d
> + xor r14d,edx
> + and edi,r11d
> +
> + ror r13d,5
> + add r12d,ecx
> + xor edi,ebx
> +
> + ror r14d,11
> + xor r13d,r11d
> + add r12d,edi
> +
> + mov edi,edx
> + add r12d,DWORD[rbp]
> + xor r14d,edx
> +
> + xor edi,r8d
> + ror r13d,6
> + mov ecx,r8d
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor ecx,r15d
> + add r10d,r12d
> + add ecx,r12d
> +
> + lea rbp,[4+rbp]
> + add ecx,r14d
> + mov r12d,DWORD[56+rsi]
> + mov r13d,r10d
> + mov r14d,ecx
> + bswap r12d
> + ror r13d,14
> + mov r15d,r11d
> +
> + xor r13d,r10d
> + ror r14d,9
> + xor r15d,eax
> +
> + mov DWORD[56+rsp],r12d
> + xor r14d,ecx
> + and r15d,r10d
> +
> + ror r13d,5
> + add r12d,ebx
> + xor r15d,eax
> +
> + ror r14d,11
> + xor r13d,r10d
> + add r12d,r15d
> +
> + mov r15d,ecx
> + add r12d,DWORD[rbp]
> + xor r14d,ecx
> +
> + xor r15d,edx
> + ror r13d,6
> + mov ebx,edx
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor ebx,edi
> + add r9d,r12d
> + add ebx,r12d
> +
> + lea rbp,[4+rbp]
> + add ebx,r14d
> + mov r12d,DWORD[60+rsi]
> + mov r13d,r9d
> + mov r14d,ebx
> + bswap r12d
> + ror r13d,14
> + mov edi,r10d
> +
> + xor r13d,r9d
> + ror r14d,9
> + xor edi,r11d
> +
> + mov DWORD[60+rsp],r12d
> + xor r14d,ebx
> + and edi,r9d
> +
> + ror r13d,5
> + add r12d,eax
> + xor edi,r11d
> +
> + ror r14d,11
> + xor r13d,r9d
> + add r12d,edi
> +
> + mov edi,ebx
> + add r12d,DWORD[rbp]
> + xor r14d,ebx
> +
> + xor edi,ecx
> + ror r13d,6
> + mov eax,ecx
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor eax,r15d
> + add r8d,r12d
> + add eax,r12d
> +
> + lea rbp,[20+rbp]
> + jmp NEAR $L$rounds_16_xx
> +ALIGN 16
> +$L$rounds_16_xx:
> + mov r13d,DWORD[4+rsp]
> + mov r15d,DWORD[56+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add eax,r14d
> + mov r14d,r15d
> + ror r15d,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor r15d,r14d
> + shr r14d,10
> +
> + ror r15d,17
> + xor r12d,r13d
> + xor r15d,r14d
> + add r12d,DWORD[36+rsp]
> +
> + add r12d,DWORD[rsp]
> + mov r13d,r8d
> + add r12d,r15d
> + mov r14d,eax
> + ror r13d,14
> + mov r15d,r9d
> +
> + xor r13d,r8d
> + ror r14d,9
> + xor r15d,r10d
> +
> + mov DWORD[rsp],r12d
> + xor r14d,eax
> + and r15d,r8d
> +
> + ror r13d,5
> + add r12d,r11d
> + xor r15d,r10d
> +
> + ror r14d,11
> + xor r13d,r8d
> + add r12d,r15d
> +
> + mov r15d,eax
> + add r12d,DWORD[rbp]
> + xor r14d,eax
> +
> + xor r15d,ebx
> + ror r13d,6
> + mov r11d,ebx
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r11d,edi
> + add edx,r12d
> + add r11d,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[8+rsp]
> + mov edi,DWORD[60+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add r11d,r14d
> + mov r14d,edi
> + ror edi,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor edi,r14d
> + shr r14d,10
> +
> + ror edi,17
> + xor r12d,r13d
> + xor edi,r14d
> + add r12d,DWORD[40+rsp]
> +
> + add r12d,DWORD[4+rsp]
> + mov r13d,edx
> + add r12d,edi
> + mov r14d,r11d
> + ror r13d,14
> + mov edi,r8d
> +
> + xor r13d,edx
> + ror r14d,9
> + xor edi,r9d
> +
> + mov DWORD[4+rsp],r12d
> + xor r14d,r11d
> + and edi,edx
> +
> + ror r13d,5
> + add r12d,r10d
> + xor edi,r9d
> +
> + ror r14d,11
> + xor r13d,edx
> + add r12d,edi
> +
> + mov edi,r11d
> + add r12d,DWORD[rbp]
> + xor r14d,r11d
> +
> + xor edi,eax
> + ror r13d,6
> + mov r10d,eax
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r10d,r15d
> + add ecx,r12d
> + add r10d,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[12+rsp]
> + mov r15d,DWORD[rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add r10d,r14d
> + mov r14d,r15d
> + ror r15d,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor r15d,r14d
> + shr r14d,10
> +
> + ror r15d,17
> + xor r12d,r13d
> + xor r15d,r14d
> + add r12d,DWORD[44+rsp]
> +
> + add r12d,DWORD[8+rsp]
> + mov r13d,ecx
> + add r12d,r15d
> + mov r14d,r10d
> + ror r13d,14
> + mov r15d,edx
> +
> + xor r13d,ecx
> + ror r14d,9
> + xor r15d,r8d
> +
> + mov DWORD[8+rsp],r12d
> + xor r14d,r10d
> + and r15d,ecx
> +
> + ror r13d,5
> + add r12d,r9d
> + xor r15d,r8d
> +
> + ror r14d,11
> + xor r13d,ecx
> + add r12d,r15d
> +
> + mov r15d,r10d
> + add r12d,DWORD[rbp]
> + xor r14d,r10d
> +
> + xor r15d,r11d
> + ror r13d,6
> + mov r9d,r11d
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r9d,edi
> + add ebx,r12d
> + add r9d,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[16+rsp]
> + mov edi,DWORD[4+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add r9d,r14d
> + mov r14d,edi
> + ror edi,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor edi,r14d
> + shr r14d,10
> +
> + ror edi,17
> + xor r12d,r13d
> + xor edi,r14d
> + add r12d,DWORD[48+rsp]
> +
> + add r12d,DWORD[12+rsp]
> + mov r13d,ebx
> + add r12d,edi
> + mov r14d,r9d
> + ror r13d,14
> + mov edi,ecx
> +
> + xor r13d,ebx
> + ror r14d,9
> + xor edi,edx
> +
> + mov DWORD[12+rsp],r12d
> + xor r14d,r9d
> + and edi,ebx
> +
> + ror r13d,5
> + add r12d,r8d
> + xor edi,edx
> +
> + ror r14d,11
> + xor r13d,ebx
> + add r12d,edi
> +
> + mov edi,r9d
> + add r12d,DWORD[rbp]
> + xor r14d,r9d
> +
> + xor edi,r10d
> + ror r13d,6
> + mov r8d,r10d
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r8d,r15d
> + add eax,r12d
> + add r8d,r12d
> +
> + lea rbp,[20+rbp]
> + mov r13d,DWORD[20+rsp]
> + mov r15d,DWORD[8+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add r8d,r14d
> + mov r14d,r15d
> + ror r15d,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor r15d,r14d
> + shr r14d,10
> +
> + ror r15d,17
> + xor r12d,r13d
> + xor r15d,r14d
> + add r12d,DWORD[52+rsp]
> +
> + add r12d,DWORD[16+rsp]
> + mov r13d,eax
> + add r12d,r15d
> + mov r14d,r8d
> + ror r13d,14
> + mov r15d,ebx
> +
> + xor r13d,eax
> + ror r14d,9
> + xor r15d,ecx
> +
> + mov DWORD[16+rsp],r12d
> + xor r14d,r8d
> + and r15d,eax
> +
> + ror r13d,5
> + add r12d,edx
> + xor r15d,ecx
> +
> + ror r14d,11
> + xor r13d,eax
> + add r12d,r15d
> +
> + mov r15d,r8d
> + add r12d,DWORD[rbp]
> + xor r14d,r8d
> +
> + xor r15d,r9d
> + ror r13d,6
> + mov edx,r9d
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor edx,edi
> + add r11d,r12d
> + add edx,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[24+rsp]
> + mov edi,DWORD[12+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add edx,r14d
> + mov r14d,edi
> + ror edi,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor edi,r14d
> + shr r14d,10
> +
> + ror edi,17
> + xor r12d,r13d
> + xor edi,r14d
> + add r12d,DWORD[56+rsp]
> +
> + add r12d,DWORD[20+rsp]
> + mov r13d,r11d
> + add r12d,edi
> + mov r14d,edx
> + ror r13d,14
> + mov edi,eax
> +
> + xor r13d,r11d
> + ror r14d,9
> + xor edi,ebx
> +
> + mov DWORD[20+rsp],r12d
> + xor r14d,edx
> + and edi,r11d
> +
> + ror r13d,5
> + add r12d,ecx
> + xor edi,ebx
> +
> + ror r14d,11
> + xor r13d,r11d
> + add r12d,edi
> +
> + mov edi,edx
> + add r12d,DWORD[rbp]
> + xor r14d,edx
> +
> + xor edi,r8d
> + ror r13d,6
> + mov ecx,r8d
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor ecx,r15d
> + add r10d,r12d
> + add ecx,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[28+rsp]
> + mov r15d,DWORD[16+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add ecx,r14d
> + mov r14d,r15d
> + ror r15d,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor r15d,r14d
> + shr r14d,10
> +
> + ror r15d,17
> + xor r12d,r13d
> + xor r15d,r14d
> + add r12d,DWORD[60+rsp]
> +
> + add r12d,DWORD[24+rsp]
> + mov r13d,r10d
> + add r12d,r15d
> + mov r14d,ecx
> + ror r13d,14
> + mov r15d,r11d
> +
> + xor r13d,r10d
> + ror r14d,9
> + xor r15d,eax
> +
> + mov DWORD[24+rsp],r12d
> + xor r14d,ecx
> + and r15d,r10d
> +
> + ror r13d,5
> + add r12d,ebx
> + xor r15d,eax
> +
> + ror r14d,11
> + xor r13d,r10d
> + add r12d,r15d
> +
> + mov r15d,ecx
> + add r12d,DWORD[rbp]
> + xor r14d,ecx
> +
> + xor r15d,edx
> + ror r13d,6
> + mov ebx,edx
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor ebx,edi
> + add r9d,r12d
> + add ebx,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[32+rsp]
> + mov edi,DWORD[20+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add ebx,r14d
> + mov r14d,edi
> + ror edi,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor edi,r14d
> + shr r14d,10
> +
> + ror edi,17
> + xor r12d,r13d
> + xor edi,r14d
> + add r12d,DWORD[rsp]
> +
> + add r12d,DWORD[28+rsp]
> + mov r13d,r9d
> + add r12d,edi
> + mov r14d,ebx
> + ror r13d,14
> + mov edi,r10d
> +
> + xor r13d,r9d
> + ror r14d,9
> + xor edi,r11d
> +
> + mov DWORD[28+rsp],r12d
> + xor r14d,ebx
> + and edi,r9d
> +
> + ror r13d,5
> + add r12d,eax
> + xor edi,r11d
> +
> + ror r14d,11
> + xor r13d,r9d
> + add r12d,edi
> +
> + mov edi,ebx
> + add r12d,DWORD[rbp]
> + xor r14d,ebx
> +
> + xor edi,ecx
> + ror r13d,6
> + mov eax,ecx
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor eax,r15d
> + add r8d,r12d
> + add eax,r12d
> +
> + lea rbp,[20+rbp]
> + mov r13d,DWORD[36+rsp]
> + mov r15d,DWORD[24+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add eax,r14d
> + mov r14d,r15d
> + ror r15d,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor r15d,r14d
> + shr r14d,10
> +
> + ror r15d,17
> + xor r12d,r13d
> + xor r15d,r14d
> + add r12d,DWORD[4+rsp]
> +
> + add r12d,DWORD[32+rsp]
> + mov r13d,r8d
> + add r12d,r15d
> + mov r14d,eax
> + ror r13d,14
> + mov r15d,r9d
> +
> + xor r13d,r8d
> + ror r14d,9
> + xor r15d,r10d
> +
> + mov DWORD[32+rsp],r12d
> + xor r14d,eax
> + and r15d,r8d
> +
> + ror r13d,5
> + add r12d,r11d
> + xor r15d,r10d
> +
> + ror r14d,11
> + xor r13d,r8d
> + add r12d,r15d
> +
> + mov r15d,eax
> + add r12d,DWORD[rbp]
> + xor r14d,eax
> +
> + xor r15d,ebx
> + ror r13d,6
> + mov r11d,ebx
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r11d,edi
> + add edx,r12d
> + add r11d,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[40+rsp]
> + mov edi,DWORD[28+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add r11d,r14d
> + mov r14d,edi
> + ror edi,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor edi,r14d
> + shr r14d,10
> +
> + ror edi,17
> + xor r12d,r13d
> + xor edi,r14d
> + add r12d,DWORD[8+rsp]
> +
> + add r12d,DWORD[36+rsp]
> + mov r13d,edx
> + add r12d,edi
> + mov r14d,r11d
> + ror r13d,14
> + mov edi,r8d
> +
> + xor r13d,edx
> + ror r14d,9
> + xor edi,r9d
> +
> + mov DWORD[36+rsp],r12d
> + xor r14d,r11d
> + and edi,edx
> +
> + ror r13d,5
> + add r12d,r10d
> + xor edi,r9d
> +
> + ror r14d,11
> + xor r13d,edx
> + add r12d,edi
> +
> + mov edi,r11d
> + add r12d,DWORD[rbp]
> + xor r14d,r11d
> +
> + xor edi,eax
> + ror r13d,6
> + mov r10d,eax
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r10d,r15d
> + add ecx,r12d
> + add r10d,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[44+rsp]
> + mov r15d,DWORD[32+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add r10d,r14d
> + mov r14d,r15d
> + ror r15d,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor r15d,r14d
> + shr r14d,10
> +
> + ror r15d,17
> + xor r12d,r13d
> + xor r15d,r14d
> + add r12d,DWORD[12+rsp]
> +
> + add r12d,DWORD[40+rsp]
> + mov r13d,ecx
> + add r12d,r15d
> + mov r14d,r10d
> + ror r13d,14
> + mov r15d,edx
> +
> + xor r13d,ecx
> + ror r14d,9
> + xor r15d,r8d
> +
> + mov DWORD[40+rsp],r12d
> + xor r14d,r10d
> + and r15d,ecx
> +
> + ror r13d,5
> + add r12d,r9d
> + xor r15d,r8d
> +
> + ror r14d,11
> + xor r13d,ecx
> + add r12d,r15d
> +
> + mov r15d,r10d
> + add r12d,DWORD[rbp]
> + xor r14d,r10d
> +
> + xor r15d,r11d
> + ror r13d,6
> + mov r9d,r11d
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r9d,edi
> + add ebx,r12d
> + add r9d,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[48+rsp]
> + mov edi,DWORD[36+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add r9d,r14d
> + mov r14d,edi
> + ror edi,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor edi,r14d
> + shr r14d,10
> +
> + ror edi,17
> + xor r12d,r13d
> + xor edi,r14d
> + add r12d,DWORD[16+rsp]
> +
> + add r12d,DWORD[44+rsp]
> + mov r13d,ebx
> + add r12d,edi
> + mov r14d,r9d
> + ror r13d,14
> + mov edi,ecx
> +
> + xor r13d,ebx
> + ror r14d,9
> + xor edi,edx
> +
> + mov DWORD[44+rsp],r12d
> + xor r14d,r9d
> + and edi,ebx
> +
> + ror r13d,5
> + add r12d,r8d
> + xor edi,edx
> +
> + ror r14d,11
> + xor r13d,ebx
> + add r12d,edi
> +
> + mov edi,r9d
> + add r12d,DWORD[rbp]
> + xor r14d,r9d
> +
> + xor edi,r10d
> + ror r13d,6
> + mov r8d,r10d
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor r8d,r15d
> + add eax,r12d
> + add r8d,r12d
> +
> + lea rbp,[20+rbp]
> + mov r13d,DWORD[52+rsp]
> + mov r15d,DWORD[40+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add r8d,r14d
> + mov r14d,r15d
> + ror r15d,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor r15d,r14d
> + shr r14d,10
> +
> + ror r15d,17
> + xor r12d,r13d
> + xor r15d,r14d
> + add r12d,DWORD[20+rsp]
> +
> + add r12d,DWORD[48+rsp]
> + mov r13d,eax
> + add r12d,r15d
> + mov r14d,r8d
> + ror r13d,14
> + mov r15d,ebx
> +
> + xor r13d,eax
> + ror r14d,9
> + xor r15d,ecx
> +
> + mov DWORD[48+rsp],r12d
> + xor r14d,r8d
> + and r15d,eax
> +
> + ror r13d,5
> + add r12d,edx
> + xor r15d,ecx
> +
> + ror r14d,11
> + xor r13d,eax
> + add r12d,r15d
> +
> + mov r15d,r8d
> + add r12d,DWORD[rbp]
> + xor r14d,r8d
> +
> + xor r15d,r9d
> + ror r13d,6
> + mov edx,r9d
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor edx,edi
> + add r11d,r12d
> + add edx,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[56+rsp]
> + mov edi,DWORD[44+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add edx,r14d
> + mov r14d,edi
> + ror edi,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor edi,r14d
> + shr r14d,10
> +
> + ror edi,17
> + xor r12d,r13d
> + xor edi,r14d
> + add r12d,DWORD[24+rsp]
> +
> + add r12d,DWORD[52+rsp]
> + mov r13d,r11d
> + add r12d,edi
> + mov r14d,edx
> + ror r13d,14
> + mov edi,eax
> +
> + xor r13d,r11d
> + ror r14d,9
> + xor edi,ebx
> +
> + mov DWORD[52+rsp],r12d
> + xor r14d,edx
> + and edi,r11d
> +
> + ror r13d,5
> + add r12d,ecx
> + xor edi,ebx
> +
> + ror r14d,11
> + xor r13d,r11d
> + add r12d,edi
> +
> + mov edi,edx
> + add r12d,DWORD[rbp]
> + xor r14d,edx
> +
> + xor edi,r8d
> + ror r13d,6
> + mov ecx,r8d
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor ecx,r15d
> + add r10d,r12d
> + add ecx,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[60+rsp]
> + mov r15d,DWORD[48+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add ecx,r14d
> + mov r14d,r15d
> + ror r15d,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor r15d,r14d
> + shr r14d,10
> +
> + ror r15d,17
> + xor r12d,r13d
> + xor r15d,r14d
> + add r12d,DWORD[28+rsp]
> +
> + add r12d,DWORD[56+rsp]
> + mov r13d,r10d
> + add r12d,r15d
> + mov r14d,ecx
> + ror r13d,14
> + mov r15d,r11d
> +
> + xor r13d,r10d
> + ror r14d,9
> + xor r15d,eax
> +
> + mov DWORD[56+rsp],r12d
> + xor r14d,ecx
> + and r15d,r10d
> +
> + ror r13d,5
> + add r12d,ebx
> + xor r15d,eax
> +
> + ror r14d,11
> + xor r13d,r10d
> + add r12d,r15d
> +
> + mov r15d,ecx
> + add r12d,DWORD[rbp]
> + xor r14d,ecx
> +
> + xor r15d,edx
> + ror r13d,6
> + mov ebx,edx
> +
> + and edi,r15d
> + ror r14d,2
> + add r12d,r13d
> +
> + xor ebx,edi
> + add r9d,r12d
> + add ebx,r12d
> +
> + lea rbp,[4+rbp]
> + mov r13d,DWORD[rsp]
> + mov edi,DWORD[52+rsp]
> +
> + mov r12d,r13d
> + ror r13d,11
> + add ebx,r14d
> + mov r14d,edi
> + ror edi,2
> +
> + xor r13d,r12d
> + shr r12d,3
> + ror r13d,7
> + xor edi,r14d
> + shr r14d,10
> +
> + ror edi,17
> + xor r12d,r13d
> + xor edi,r14d
> + add r12d,DWORD[32+rsp]
> +
> + add r12d,DWORD[60+rsp]
> + mov r13d,r9d
> + add r12d,edi
> + mov r14d,ebx
> + ror r13d,14
> + mov edi,r10d
> +
> + xor r13d,r9d
> + ror r14d,9
> + xor edi,r11d
> +
> + mov DWORD[60+rsp],r12d
> + xor r14d,ebx
> + and edi,r9d
> +
> + ror r13d,5
> + add r12d,eax
> + xor edi,r11d
> +
> + ror r14d,11
> + xor r13d,r9d
> + add r12d,edi
> +
> + mov edi,ebx
> + add r12d,DWORD[rbp]
> + xor r14d,ebx
> +
> + xor edi,ecx
> + ror r13d,6
> + mov eax,ecx
> +
> + and r15d,edi
> + ror r14d,2
> + add r12d,r13d
> +
> + xor eax,r15d
> + add r8d,r12d
> + add eax,r12d
> +
> + lea rbp,[20+rbp]
> + cmp BYTE[3+rbp],0
> + jnz NEAR $L$rounds_16_xx
> +
> + mov rdi,QWORD[((64+0))+rsp]
> + add eax,r14d
> + lea rsi,[64+rsi]
> +
> + add eax,DWORD[rdi]
> + add ebx,DWORD[4+rdi]
> + add ecx,DWORD[8+rdi]
> + add edx,DWORD[12+rdi]
> + add r8d,DWORD[16+rdi]
> + add r9d,DWORD[20+rdi]
> + add r10d,DWORD[24+rdi]
> + add r11d,DWORD[28+rdi]
> +
> + cmp rsi,QWORD[((64+16))+rsp]
> +
> + mov DWORD[rdi],eax
> + mov DWORD[4+rdi],ebx
> + mov DWORD[8+rdi],ecx
> + mov DWORD[12+rdi],edx
> + mov DWORD[16+rdi],r8d
> + mov DWORD[20+rdi],r9d
> + mov DWORD[24+rdi],r10d
> + mov DWORD[28+rdi],r11d
> + jb NEAR $L$loop
> +
> + mov rsi,QWORD[88+rsp]
> +
> + mov r15,QWORD[((-48))+rsi]
> +
> + mov r14,QWORD[((-40))+rsi]
> +
> + mov r13,QWORD[((-32))+rsi]
> +
> + mov r12,QWORD[((-24))+rsi]
> +
> + mov rbp,QWORD[((-16))+rsi]
> +
> + mov rbx,QWORD[((-8))+rsi]
> +
> + lea rsp,[rsi]
> +
> +$L$epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha256_block_data_order:
> +ALIGN 64
> +
> +K256:
> + DD 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> + DD 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> + DD 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> + DD 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> + DD 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> + DD 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> + DD 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> + DD 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> + DD 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> + DD 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> + DD 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> + DD 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> + DD 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> + DD 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> + DD 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> + DD 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> + DD 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> + DD 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> + DD 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> + DD 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> + DD 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> + DD 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> + DD 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> + DD 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> + DD 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> + DD 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> + DD 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> + DD 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> + DD 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> + DD 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> + DD 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> + DD 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> +
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> + DD 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> + DD 0x03020100,0x0b0a0908,0xffffffff,0xffffffff
> + DD 0x03020100,0x0b0a0908,0xffffffff,0xffffffff
> + DD 0xffffffff,0xffffffff,0x03020100,0x0b0a0908
> + DD 0xffffffff,0xffffffff,0x03020100,0x0b0a0908
> +DB 83,72,65,50,53,54,32,98,108,111,99,107,32,116,114,97
> +DB 110,115,102,111,114,109,32,102,111,114,32,120,56,54,95,54
> +DB 52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121
> +DB 32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46
> +DB 111,114,103,62,0
> +
> +ALIGN 64
> +sha256_block_data_order_shaext:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha256_block_data_order_shaext:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +_shaext_shortcut:
> +
> + lea rsp,[((-88))+rsp]
> + movaps XMMWORD[(-8-80)+rax],xmm6
> + movaps XMMWORD[(-8-64)+rax],xmm7
> + movaps XMMWORD[(-8-48)+rax],xmm8
> + movaps XMMWORD[(-8-32)+rax],xmm9
> + movaps XMMWORD[(-8-16)+rax],xmm10
> +$L$prologue_shaext:
> + lea rcx,[((K256+128))]
> + movdqu xmm1,XMMWORD[rdi]
> + movdqu xmm2,XMMWORD[16+rdi]
> + movdqa xmm7,XMMWORD[((512-128))+rcx]
> +
> + pshufd xmm0,xmm1,0x1b
> + pshufd xmm1,xmm1,0xb1
> + pshufd xmm2,xmm2,0x1b
> + movdqa xmm8,xmm7
> +DB 102,15,58,15,202,8
> + punpcklqdq xmm2,xmm0
> + jmp NEAR $L$oop_shaext
> +
> +ALIGN 16
> +$L$oop_shaext:
> + movdqu xmm3,XMMWORD[rsi]
> + movdqu xmm4,XMMWORD[16+rsi]
> + movdqu xmm5,XMMWORD[32+rsi]
> +DB 102,15,56,0,223
> + movdqu xmm6,XMMWORD[48+rsi]
> +
> + movdqa xmm0,XMMWORD[((0-128))+rcx]
> + paddd xmm0,xmm3
> +DB 102,15,56,0,231
> + movdqa xmm10,xmm2
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + nop
> + movdqa xmm9,xmm1
> +DB 15,56,203,202
> +
> + movdqa xmm0,XMMWORD[((32-128))+rcx]
> + paddd xmm0,xmm4
> +DB 102,15,56,0,239
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + lea rsi,[64+rsi]
> +DB 15,56,204,220
> +DB 15,56,203,202
> +
> + movdqa xmm0,XMMWORD[((64-128))+rcx]
> + paddd xmm0,xmm5
> +DB 102,15,56,0,247
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm6
> +DB 102,15,58,15,253,4
> + nop
> + paddd xmm3,xmm7
> +DB 15,56,204,229
> +DB 15,56,203,202
> +
> + movdqa xmm0,XMMWORD[((96-128))+rcx]
> + paddd xmm0,xmm6
> +DB 15,56,205,222
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm3
> +DB 102,15,58,15,254,4
> + nop
> + paddd xmm4,xmm7
> +DB 15,56,204,238
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((128-128))+rcx]
> + paddd xmm0,xmm3
> +DB 15,56,205,227
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm4
> +DB 102,15,58,15,251,4
> + nop
> + paddd xmm5,xmm7
> +DB 15,56,204,243
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((160-128))+rcx]
> + paddd xmm0,xmm4
> +DB 15,56,205,236
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm5
> +DB 102,15,58,15,252,4
> + nop
> + paddd xmm6,xmm7
> +DB 15,56,204,220
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((192-128))+rcx]
> + paddd xmm0,xmm5
> +DB 15,56,205,245
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm6
> +DB 102,15,58,15,253,4
> + nop
> + paddd xmm3,xmm7
> +DB 15,56,204,229
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((224-128))+rcx]
> + paddd xmm0,xmm6
> +DB 15,56,205,222
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm3
> +DB 102,15,58,15,254,4
> + nop
> + paddd xmm4,xmm7
> +DB 15,56,204,238
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((256-128))+rcx]
> + paddd xmm0,xmm3
> +DB 15,56,205,227
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm4
> +DB 102,15,58,15,251,4
> + nop
> + paddd xmm5,xmm7
> +DB 15,56,204,243
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((288-128))+rcx]
> + paddd xmm0,xmm4
> +DB 15,56,205,236
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm5
> +DB 102,15,58,15,252,4
> + nop
> + paddd xmm6,xmm7
> +DB 15,56,204,220
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((320-128))+rcx]
> + paddd xmm0,xmm5
> +DB 15,56,205,245
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm6
> +DB 102,15,58,15,253,4
> + nop
> + paddd xmm3,xmm7
> +DB 15,56,204,229
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((352-128))+rcx]
> + paddd xmm0,xmm6
> +DB 15,56,205,222
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm3
> +DB 102,15,58,15,254,4
> + nop
> + paddd xmm4,xmm7
> +DB 15,56,204,238
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((384-128))+rcx]
> + paddd xmm0,xmm3
> +DB 15,56,205,227
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm4
> +DB 102,15,58,15,251,4
> + nop
> + paddd xmm5,xmm7
> +DB 15,56,204,243
> +DB 15,56,203,202
> + movdqa xmm0,XMMWORD[((416-128))+rcx]
> + paddd xmm0,xmm4
> +DB 15,56,205,236
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + movdqa xmm7,xmm5
> +DB 102,15,58,15,252,4
> +DB 15,56,203,202
> + paddd xmm6,xmm7
> +
> + movdqa xmm0,XMMWORD[((448-128))+rcx]
> + paddd xmm0,xmm5
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> +DB 15,56,205,245
> + movdqa xmm7,xmm8
> +DB 15,56,203,202
> +
> + movdqa xmm0,XMMWORD[((480-128))+rcx]
> + paddd xmm0,xmm6
> + nop
> +DB 15,56,203,209
> + pshufd xmm0,xmm0,0x0e
> + dec rdx
> + nop
> +DB 15,56,203,202
> +
> + paddd xmm2,xmm10
> + paddd xmm1,xmm9
> + jnz NEAR $L$oop_shaext
> +
> + pshufd xmm2,xmm2,0xb1
> + pshufd xmm7,xmm1,0x1b
> + pshufd xmm1,xmm1,0xb1
> + punpckhqdq xmm1,xmm2
> +DB 102,15,58,15,215,8
> +
> + movdqu XMMWORD[rdi],xmm1
> + movdqu XMMWORD[16+rdi],xmm2
> + movaps xmm6,XMMWORD[((-8-80))+rax]
> + movaps xmm7,XMMWORD[((-8-64))+rax]
> + movaps xmm8,XMMWORD[((-8-48))+rax]
> + movaps xmm9,XMMWORD[((-8-32))+rax]
> + movaps xmm10,XMMWORD[((-8-16))+rax]
> + mov rsp,rax
> +$L$epilogue_shaext:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha256_block_data_order_shaext:
> +
> +ALIGN 64
> +sha256_block_data_order_ssse3:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha256_block_data_order_ssse3:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> +$L$ssse3_shortcut:
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + push r15
> +
> + shl rdx,4
> + sub rsp,160
> + lea rdx,[rdx*4+rsi]
> + and rsp,-64
> + mov QWORD[((64+0))+rsp],rdi
> + mov QWORD[((64+8))+rsp],rsi
> + mov QWORD[((64+16))+rsp],rdx
> + mov QWORD[88+rsp],rax
> +
> + movaps XMMWORD[(64+32)+rsp],xmm6
> + movaps XMMWORD[(64+48)+rsp],xmm7
> + movaps XMMWORD[(64+64)+rsp],xmm8
> + movaps XMMWORD[(64+80)+rsp],xmm9
> +$L$prologue_ssse3:
> +
> + mov eax,DWORD[rdi]
> + mov ebx,DWORD[4+rdi]
> + mov ecx,DWORD[8+rdi]
> + mov edx,DWORD[12+rdi]
> + mov r8d,DWORD[16+rdi]
> + mov r9d,DWORD[20+rdi]
> + mov r10d,DWORD[24+rdi]
> + mov r11d,DWORD[28+rdi]
> +
> +
> + jmp NEAR $L$loop_ssse3
> +ALIGN 16
> +$L$loop_ssse3:
> + movdqa xmm7,XMMWORD[((K256+512))]
> + movdqu xmm0,XMMWORD[rsi]
> + movdqu xmm1,XMMWORD[16+rsi]
> + movdqu xmm2,XMMWORD[32+rsi]
> +DB 102,15,56,0,199
> + movdqu xmm3,XMMWORD[48+rsi]
> + lea rbp,[K256]
> +DB 102,15,56,0,207
> + movdqa xmm4,XMMWORD[rbp]
> + movdqa xmm5,XMMWORD[32+rbp]
> +DB 102,15,56,0,215
> + paddd xmm4,xmm0
> + movdqa xmm6,XMMWORD[64+rbp]
> +DB 102,15,56,0,223
> + movdqa xmm7,XMMWORD[96+rbp]
> + paddd xmm5,xmm1
> + paddd xmm6,xmm2
> + paddd xmm7,xmm3
> + movdqa XMMWORD[rsp],xmm4
> + mov r14d,eax
> + movdqa XMMWORD[16+rsp],xmm5
> + mov edi,ebx
> + movdqa XMMWORD[32+rsp],xmm6
> + xor edi,ecx
> + movdqa XMMWORD[48+rsp],xmm7
> + mov r13d,r8d
> + jmp NEAR $L$ssse3_00_47
> +
> +ALIGN 16
> +$L$ssse3_00_47:
> + sub rbp,-128
> + ror r13d,14
> + movdqa xmm4,xmm1
> + mov eax,r14d
> + mov r12d,r9d
> + movdqa xmm7,xmm3
> + ror r14d,9
> + xor r13d,r8d
> + xor r12d,r10d
> + ror r13d,5
> + xor r14d,eax
> +DB 102,15,58,15,224,4
> + and r12d,r8d
> + xor r13d,r8d
> +DB 102,15,58,15,250,4
> + add r11d,DWORD[rsp]
> + mov r15d,eax
> + xor r12d,r10d
> + ror r14d,11
> + movdqa xmm5,xmm4
> + xor r15d,ebx
> + add r11d,r12d
> + movdqa xmm6,xmm4
> + ror r13d,6
> + and edi,r15d
> + psrld xmm4,3
> + xor r14d,eax
> + add r11d,r13d
> + xor edi,ebx
> + paddd xmm0,xmm7
> + ror r14d,2
> + add edx,r11d
> + psrld xmm6,7
> + add r11d,edi
> + mov r13d,edx
> + pshufd xmm7,xmm3,250
> + add r14d,r11d
> + ror r13d,14
> + pslld xmm5,14
> + mov r11d,r14d
> + mov r12d,r8d
> + pxor xmm4,xmm6
> + ror r14d,9
> + xor r13d,edx
> + xor r12d,r9d
> + ror r13d,5
> + psrld xmm6,11
> + xor r14d,r11d
> + pxor xmm4,xmm5
> + and r12d,edx
> + xor r13d,edx
> + pslld xmm5,11
> + add r10d,DWORD[4+rsp]
> + mov edi,r11d
> + pxor xmm4,xmm6
> + xor r12d,r9d
> + ror r14d,11
> + movdqa xmm6,xmm7
> + xor edi,eax
> + add r10d,r12d
> + pxor xmm4,xmm5
> + ror r13d,6
> + and r15d,edi
> + xor r14d,r11d
> + psrld xmm7,10
> + add r10d,r13d
> + xor r15d,eax
> + paddd xmm0,xmm4
> + ror r14d,2
> + add ecx,r10d
> + psrlq xmm6,17
> + add r10d,r15d
> + mov r13d,ecx
> + add r14d,r10d
> + pxor xmm7,xmm6
> + ror r13d,14
> + mov r10d,r14d
> + mov r12d,edx
> + ror r14d,9
> + psrlq xmm6,2
> + xor r13d,ecx
> + xor r12d,r8d
> + pxor xmm7,xmm6
> + ror r13d,5
> + xor r14d,r10d
> + and r12d,ecx
> + pshufd xmm7,xmm7,128
> + xor r13d,ecx
> + add r9d,DWORD[8+rsp]
> + mov r15d,r10d
> + psrldq xmm7,8
> + xor r12d,r8d
> + ror r14d,11
> + xor r15d,r11d
> + add r9d,r12d
> + ror r13d,6
> + paddd xmm0,xmm7
> + and edi,r15d
> + xor r14d,r10d
> + add r9d,r13d
> + pshufd xmm7,xmm0,80
> + xor edi,r11d
> + ror r14d,2
> + add ebx,r9d
> + movdqa xmm6,xmm7
> + add r9d,edi
> + mov r13d,ebx
> + psrld xmm7,10
> + add r14d,r9d
> + ror r13d,14
> + psrlq xmm6,17
> + mov r9d,r14d
> + mov r12d,ecx
> + pxor xmm7,xmm6
> + ror r14d,9
> + xor r13d,ebx
> + xor r12d,edx
> + ror r13d,5
> + xor r14d,r9d
> + psrlq xmm6,2
> + and r12d,ebx
> + xor r13d,ebx
> + add r8d,DWORD[12+rsp]
> + pxor xmm7,xmm6
> + mov edi,r9d
> + xor r12d,edx
> + ror r14d,11
> + pshufd xmm7,xmm7,8
> + xor edi,r10d
> + add r8d,r12d
> + movdqa xmm6,XMMWORD[rbp]
> + ror r13d,6
> + and r15d,edi
> + pslldq xmm7,8
> + xor r14d,r9d
> + add r8d,r13d
> + xor r15d,r10d
> + paddd xmm0,xmm7
> + ror r14d,2
> + add eax,r8d
> + add r8d,r15d
> + paddd xmm6,xmm0
> + mov r13d,eax
> + add r14d,r8d
> + movdqa XMMWORD[rsp],xmm6
> + ror r13d,14
> + movdqa xmm4,xmm2
> + mov r8d,r14d
> + mov r12d,ebx
> + movdqa xmm7,xmm0
> + ror r14d,9
> + xor r13d,eax
> + xor r12d,ecx
> + ror r13d,5
> + xor r14d,r8d
> +DB 102,15,58,15,225,4
> + and r12d,eax
> + xor r13d,eax
> +DB 102,15,58,15,251,4
> + add edx,DWORD[16+rsp]
> + mov r15d,r8d
> + xor r12d,ecx
> + ror r14d,11
> + movdqa xmm5,xmm4
> + xor r15d,r9d
> + add edx,r12d
> + movdqa xmm6,xmm4
> + ror r13d,6
> + and edi,r15d
> + psrld xmm4,3
> + xor r14d,r8d
> + add edx,r13d
> + xor edi,r9d
> + paddd xmm1,xmm7
> + ror r14d,2
> + add r11d,edx
> + psrld xmm6,7
> + add edx,edi
> + mov r13d,r11d
> + pshufd xmm7,xmm0,250
> + add r14d,edx
> + ror r13d,14
> + pslld xmm5,14
> + mov edx,r14d
> + mov r12d,eax
> + pxor xmm4,xmm6
> + ror r14d,9
> + xor r13d,r11d
> + xor r12d,ebx
> + ror r13d,5
> + psrld xmm6,11
> + xor r14d,edx
> + pxor xmm4,xmm5
> + and r12d,r11d
> + xor r13d,r11d
> + pslld xmm5,11
> + add ecx,DWORD[20+rsp]
> + mov edi,edx
> + pxor xmm4,xmm6
> + xor r12d,ebx
> + ror r14d,11
> + movdqa xmm6,xmm7
> + xor edi,r8d
> + add ecx,r12d
> + pxor xmm4,xmm5
> + ror r13d,6
> + and r15d,edi
> + xor r14d,edx
> + psrld xmm7,10
> + add ecx,r13d
> + xor r15d,r8d
> + paddd xmm1,xmm4
> + ror r14d,2
> + add r10d,ecx
> + psrlq xmm6,17
> + add ecx,r15d
> + mov r13d,r10d
> + add r14d,ecx
> + pxor xmm7,xmm6
> + ror r13d,14
> + mov ecx,r14d
> + mov r12d,r11d
> + ror r14d,9
> + psrlq xmm6,2
> + xor r13d,r10d
> + xor r12d,eax
> + pxor xmm7,xmm6
> + ror r13d,5
> + xor r14d,ecx
> + and r12d,r10d
> + pshufd xmm7,xmm7,128
> + xor r13d,r10d
> + add ebx,DWORD[24+rsp]
> + mov r15d,ecx
> + psrldq xmm7,8
> + xor r12d,eax
> + ror r14d,11
> + xor r15d,edx
> + add ebx,r12d
> + ror r13d,6
> + paddd xmm1,xmm7
> + and edi,r15d
> + xor r14d,ecx
> + add ebx,r13d
> + pshufd xmm7,xmm1,80
> + xor edi,edx
> + ror r14d,2
> + add r9d,ebx
> + movdqa xmm6,xmm7
> + add ebx,edi
> + mov r13d,r9d
> + psrld xmm7,10
> + add r14d,ebx
> + ror r13d,14
> + psrlq xmm6,17
> + mov ebx,r14d
> + mov r12d,r10d
> + pxor xmm7,xmm6
> + ror r14d,9
> + xor r13d,r9d
> + xor r12d,r11d
> + ror r13d,5
> + xor r14d,ebx
> + psrlq xmm6,2
> + and r12d,r9d
> + xor r13d,r9d
> + add eax,DWORD[28+rsp]
> + pxor xmm7,xmm6
> + mov edi,ebx
> + xor r12d,r11d
> + ror r14d,11
> + pshufd xmm7,xmm7,8
> + xor edi,ecx
> + add eax,r12d
> + movdqa xmm6,XMMWORD[32+rbp]
> + ror r13d,6
> + and r15d,edi
> + pslldq xmm7,8
> + xor r14d,ebx
> + add eax,r13d
> + xor r15d,ecx
> + paddd xmm1,xmm7
> + ror r14d,2
> + add r8d,eax
> + add eax,r15d
> + paddd xmm6,xmm1
> + mov r13d,r8d
> + add r14d,eax
> + movdqa XMMWORD[16+rsp],xmm6
> + ror r13d,14
> + movdqa xmm4,xmm3
> + mov eax,r14d
> + mov r12d,r9d
> + movdqa xmm7,xmm1
> + ror r14d,9
> + xor r13d,r8d
> + xor r12d,r10d
> + ror r13d,5
> + xor r14d,eax
> +DB 102,15,58,15,226,4
> + and r12d,r8d
> + xor r13d,r8d
> +DB 102,15,58,15,248,4
> + add r11d,DWORD[32+rsp]
> + mov r15d,eax
> + xor r12d,r10d
> + ror r14d,11
> + movdqa xmm5,xmm4
> + xor r15d,ebx
> + add r11d,r12d
> + movdqa xmm6,xmm4
> + ror r13d,6
> + and edi,r15d
> + psrld xmm4,3
> + xor r14d,eax
> + add r11d,r13d
> + xor edi,ebx
> + paddd xmm2,xmm7
> + ror r14d,2
> + add edx,r11d
> + psrld xmm6,7
> + add r11d,edi
> + mov r13d,edx
> + pshufd xmm7,xmm1,250
> + add r14d,r11d
> + ror r13d,14
> + pslld xmm5,14
> + mov r11d,r14d
> + mov r12d,r8d
> + pxor xmm4,xmm6
> + ror r14d,9
> + xor r13d,edx
> + xor r12d,r9d
> + ror r13d,5
> + psrld xmm6,11
> + xor r14d,r11d
> + pxor xmm4,xmm5
> + and r12d,edx
> + xor r13d,edx
> + pslld xmm5,11
> + add r10d,DWORD[36+rsp]
> + mov edi,r11d
> + pxor xmm4,xmm6
> + xor r12d,r9d
> + ror r14d,11
> + movdqa xmm6,xmm7
> + xor edi,eax
> + add r10d,r12d
> + pxor xmm4,xmm5
> + ror r13d,6
> + and r15d,edi
> + xor r14d,r11d
> + psrld xmm7,10
> + add r10d,r13d
> + xor r15d,eax
> + paddd xmm2,xmm4
> + ror r14d,2
> + add ecx,r10d
> + psrlq xmm6,17
> + add r10d,r15d
> + mov r13d,ecx
> + add r14d,r10d
> + pxor xmm7,xmm6
> + ror r13d,14
> + mov r10d,r14d
> + mov r12d,edx
> + ror r14d,9
> + psrlq xmm6,2
> + xor r13d,ecx
> + xor r12d,r8d
> + pxor xmm7,xmm6
> + ror r13d,5
> + xor r14d,r10d
> + and r12d,ecx
> + pshufd xmm7,xmm7,128
> + xor r13d,ecx
> + add r9d,DWORD[40+rsp]
> + mov r15d,r10d
> + psrldq xmm7,8
> + xor r12d,r8d
> + ror r14d,11
> + xor r15d,r11d
> + add r9d,r12d
> + ror r13d,6
> + paddd xmm2,xmm7
> + and edi,r15d
> + xor r14d,r10d
> + add r9d,r13d
> + pshufd xmm7,xmm2,80
> + xor edi,r11d
> + ror r14d,2
> + add ebx,r9d
> + movdqa xmm6,xmm7
> + add r9d,edi
> + mov r13d,ebx
> + psrld xmm7,10
> + add r14d,r9d
> + ror r13d,14
> + psrlq xmm6,17
> + mov r9d,r14d
> + mov r12d,ecx
> + pxor xmm7,xmm6
> + ror r14d,9
> + xor r13d,ebx
> + xor r12d,edx
> + ror r13d,5
> + xor r14d,r9d
> + psrlq xmm6,2
> + and r12d,ebx
> + xor r13d,ebx
> + add r8d,DWORD[44+rsp]
> + pxor xmm7,xmm6
> + mov edi,r9d
> + xor r12d,edx
> + ror r14d,11
> + pshufd xmm7,xmm7,8
> + xor edi,r10d
> + add r8d,r12d
> + movdqa xmm6,XMMWORD[64+rbp]
> + ror r13d,6
> + and r15d,edi
> + pslldq xmm7,8
> + xor r14d,r9d
> + add r8d,r13d
> + xor r15d,r10d
> + paddd xmm2,xmm7
> + ror r14d,2
> + add eax,r8d
> + add r8d,r15d
> + paddd xmm6,xmm2
> + mov r13d,eax
> + add r14d,r8d
> + movdqa XMMWORD[32+rsp],xmm6
> + ror r13d,14
> + movdqa xmm4,xmm0
> + mov r8d,r14d
> + mov r12d,ebx
> + movdqa xmm7,xmm2
> + ror r14d,9
> + xor r13d,eax
> + xor r12d,ecx
> + ror r13d,5
> + xor r14d,r8d
> +DB 102,15,58,15,227,4
> + and r12d,eax
> + xor r13d,eax
> +DB 102,15,58,15,249,4
> + add edx,DWORD[48+rsp]
> + mov r15d,r8d
> + xor r12d,ecx
> + ror r14d,11
> + movdqa xmm5,xmm4
> + xor r15d,r9d
> + add edx,r12d
> + movdqa xmm6,xmm4
> + ror r13d,6
> + and edi,r15d
> + psrld xmm4,3
> + xor r14d,r8d
> + add edx,r13d
> + xor edi,r9d
> + paddd xmm3,xmm7
> + ror r14d,2
> + add r11d,edx
> + psrld xmm6,7
> + add edx,edi
> + mov r13d,r11d
> + pshufd xmm7,xmm2,250
> + add r14d,edx
> + ror r13d,14
> + pslld xmm5,14
> + mov edx,r14d
> + mov r12d,eax
> + pxor xmm4,xmm6
> + ror r14d,9
> + xor r13d,r11d
> + xor r12d,ebx
> + ror r13d,5
> + psrld xmm6,11
> + xor r14d,edx
> + pxor xmm4,xmm5
> + and r12d,r11d
> + xor r13d,r11d
> + pslld xmm5,11
> + add ecx,DWORD[52+rsp]
> + mov edi,edx
> + pxor xmm4,xmm6
> + xor r12d,ebx
> + ror r14d,11
> + movdqa xmm6,xmm7
> + xor edi,r8d
> + add ecx,r12d
> + pxor xmm4,xmm5
> + ror r13d,6
> + and r15d,edi
> + xor r14d,edx
> + psrld xmm7,10
> + add ecx,r13d
> + xor r15d,r8d
> + paddd xmm3,xmm4
> + ror r14d,2
> + add r10d,ecx
> + psrlq xmm6,17
> + add ecx,r15d
> + mov r13d,r10d
> + add r14d,ecx
> + pxor xmm7,xmm6
> + ror r13d,14
> + mov ecx,r14d
> + mov r12d,r11d
> + ror r14d,9
> + psrlq xmm6,2
> + xor r13d,r10d
> + xor r12d,eax
> + pxor xmm7,xmm6
> + ror r13d,5
> + xor r14d,ecx
> + and r12d,r10d
> + pshufd xmm7,xmm7,128
> + xor r13d,r10d
> + add ebx,DWORD[56+rsp]
> + mov r15d,ecx
> + psrldq xmm7,8
> + xor r12d,eax
> + ror r14d,11
> + xor r15d,edx
> + add ebx,r12d
> + ror r13d,6
> + paddd xmm3,xmm7
> + and edi,r15d
> + xor r14d,ecx
> + add ebx,r13d
> + pshufd xmm7,xmm3,80
> + xor edi,edx
> + ror r14d,2
> + add r9d,ebx
> + movdqa xmm6,xmm7
> + add ebx,edi
> + mov r13d,r9d
> + psrld xmm7,10
> + add r14d,ebx
> + ror r13d,14
> + psrlq xmm6,17
> + mov ebx,r14d
> + mov r12d,r10d
> + pxor xmm7,xmm6
> + ror r14d,9
> + xor r13d,r9d
> + xor r12d,r11d
> + ror r13d,5
> + xor r14d,ebx
> + psrlq xmm6,2
> + and r12d,r9d
> + xor r13d,r9d
> + add eax,DWORD[60+rsp]
> + pxor xmm7,xmm6
> + mov edi,ebx
> + xor r12d,r11d
> + ror r14d,11
> + pshufd xmm7,xmm7,8
> + xor edi,ecx
> + add eax,r12d
> + movdqa xmm6,XMMWORD[96+rbp]
> + ror r13d,6
> + and r15d,edi
> + pslldq xmm7,8
> + xor r14d,ebx
> + add eax,r13d
> + xor r15d,ecx
> + paddd xmm3,xmm7
> + ror r14d,2
> + add r8d,eax
> + add eax,r15d
> + paddd xmm6,xmm3
> + mov r13d,r8d
> + add r14d,eax
> + movdqa XMMWORD[48+rsp],xmm6
> + cmp BYTE[131+rbp],0
> + jne NEAR $L$ssse3_00_47
> + ror r13d,14
> + mov eax,r14d
> + mov r12d,r9d
> + ror r14d,9
> + xor r13d,r8d
> + xor r12d,r10d
> + ror r13d,5
> + xor r14d,eax
> + and r12d,r8d
> + xor r13d,r8d
> + add r11d,DWORD[rsp]
> + mov r15d,eax
> + xor r12d,r10d
> + ror r14d,11
> + xor r15d,ebx
> + add r11d,r12d
> + ror r13d,6
> + and edi,r15d
> + xor r14d,eax
> + add r11d,r13d
> + xor edi,ebx
> + ror r14d,2
> + add edx,r11d
> + add r11d,edi
> + mov r13d,edx
> + add r14d,r11d
> + ror r13d,14
> + mov r11d,r14d
> + mov r12d,r8d
> + ror r14d,9
> + xor r13d,edx
> + xor r12d,r9d
> + ror r13d,5
> + xor r14d,r11d
> + and r12d,edx
> + xor r13d,edx
> + add r10d,DWORD[4+rsp]
> + mov edi,r11d
> + xor r12d,r9d
> + ror r14d,11
> + xor edi,eax
> + add r10d,r12d
> + ror r13d,6
> + and r15d,edi
> + xor r14d,r11d
> + add r10d,r13d
> + xor r15d,eax
> + ror r14d,2
> + add ecx,r10d
> + add r10d,r15d
> + mov r13d,ecx
> + add r14d,r10d
> + ror r13d,14
> + mov r10d,r14d
> + mov r12d,edx
> + ror r14d,9
> + xor r13d,ecx
> + xor r12d,r8d
> + ror r13d,5
> + xor r14d,r10d
> + and r12d,ecx
> + xor r13d,ecx
> + add r9d,DWORD[8+rsp]
> + mov r15d,r10d
> + xor r12d,r8d
> + ror r14d,11
> + xor r15d,r11d
> + add r9d,r12d
> + ror r13d,6
> + and edi,r15d
> + xor r14d,r10d
> + add r9d,r13d
> + xor edi,r11d
> + ror r14d,2
> + add ebx,r9d
> + add r9d,edi
> + mov r13d,ebx
> + add r14d,r9d
> + ror r13d,14
> + mov r9d,r14d
> + mov r12d,ecx
> + ror r14d,9
> + xor r13d,ebx
> + xor r12d,edx
> + ror r13d,5
> + xor r14d,r9d
> + and r12d,ebx
> + xor r13d,ebx
> + add r8d,DWORD[12+rsp]
> + mov edi,r9d
> + xor r12d,edx
> + ror r14d,11
> + xor edi,r10d
> + add r8d,r12d
> + ror r13d,6
> + and r15d,edi
> + xor r14d,r9d
> + add r8d,r13d
> + xor r15d,r10d
> + ror r14d,2
> + add eax,r8d
> + add r8d,r15d
> + mov r13d,eax
> + add r14d,r8d
> + ror r13d,14
> + mov r8d,r14d
> + mov r12d,ebx
> + ror r14d,9
> + xor r13d,eax
> + xor r12d,ecx
> + ror r13d,5
> + xor r14d,r8d
> + and r12d,eax
> + xor r13d,eax
> + add edx,DWORD[16+rsp]
> + mov r15d,r8d
> + xor r12d,ecx
> + ror r14d,11
> + xor r15d,r9d
> + add edx,r12d
> + ror r13d,6
> + and edi,r15d
> + xor r14d,r8d
> + add edx,r13d
> + xor edi,r9d
> + ror r14d,2
> + add r11d,edx
> + add edx,edi
> + mov r13d,r11d
> + add r14d,edx
> + ror r13d,14
> + mov edx,r14d
> + mov r12d,eax
> + ror r14d,9
> + xor r13d,r11d
> + xor r12d,ebx
> + ror r13d,5
> + xor r14d,edx
> + and r12d,r11d
> + xor r13d,r11d
> + add ecx,DWORD[20+rsp]
> + mov edi,edx
> + xor r12d,ebx
> + ror r14d,11
> + xor edi,r8d
> + add ecx,r12d
> + ror r13d,6
> + and r15d,edi
> + xor r14d,edx
> + add ecx,r13d
> + xor r15d,r8d
> + ror r14d,2
> + add r10d,ecx
> + add ecx,r15d
> + mov r13d,r10d
> + add r14d,ecx
> + ror r13d,14
> + mov ecx,r14d
> + mov r12d,r11d
> + ror r14d,9
> + xor r13d,r10d
> + xor r12d,eax
> + ror r13d,5
> + xor r14d,ecx
> + and r12d,r10d
> + xor r13d,r10d
> + add ebx,DWORD[24+rsp]
> + mov r15d,ecx
> + xor r12d,eax
> + ror r14d,11
> + xor r15d,edx
> + add ebx,r12d
> + ror r13d,6
> + and edi,r15d
> + xor r14d,ecx
> + add ebx,r13d
> + xor edi,edx
> + ror r14d,2
> + add r9d,ebx
> + add ebx,edi
> + mov r13d,r9d
> + add r14d,ebx
> + ror r13d,14
> + mov ebx,r14d
> + mov r12d,r10d
> + ror r14d,9
> + xor r13d,r9d
> + xor r12d,r11d
> + ror r13d,5
> + xor r14d,ebx
> + and r12d,r9d
> + xor r13d,r9d
> + add eax,DWORD[28+rsp]
> + mov edi,ebx
> + xor r12d,r11d
> + ror r14d,11
> + xor edi,ecx
> + add eax,r12d
> + ror r13d,6
> + and r15d,edi
> + xor r14d,ebx
> + add eax,r13d
> + xor r15d,ecx
> + ror r14d,2
> + add r8d,eax
> + add eax,r15d
> + mov r13d,r8d
> + add r14d,eax
> + ror r13d,14
> + mov eax,r14d
> + mov r12d,r9d
> + ror r14d,9
> + xor r13d,r8d
> + xor r12d,r10d
> + ror r13d,5
> + xor r14d,eax
> + and r12d,r8d
> + xor r13d,r8d
> + add r11d,DWORD[32+rsp]
> + mov r15d,eax
> + xor r12d,r10d
> + ror r14d,11
> + xor r15d,ebx
> + add r11d,r12d
> + ror r13d,6
> + and edi,r15d
> + xor r14d,eax
> + add r11d,r13d
> + xor edi,ebx
> + ror r14d,2
> + add edx,r11d
> + add r11d,edi
> + mov r13d,edx
> + add r14d,r11d
> + ror r13d,14
> + mov r11d,r14d
> + mov r12d,r8d
> + ror r14d,9
> + xor r13d,edx
> + xor r12d,r9d
> + ror r13d,5
> + xor r14d,r11d
> + and r12d,edx
> + xor r13d,edx
> + add r10d,DWORD[36+rsp]
> + mov edi,r11d
> + xor r12d,r9d
> + ror r14d,11
> + xor edi,eax
> + add r10d,r12d
> + ror r13d,6
> + and r15d,edi
> + xor r14d,r11d
> + add r10d,r13d
> + xor r15d,eax
> + ror r14d,2
> + add ecx,r10d
> + add r10d,r15d
> + mov r13d,ecx
> + add r14d,r10d
> + ror r13d,14
> + mov r10d,r14d
> + mov r12d,edx
> + ror r14d,9
> + xor r13d,ecx
> + xor r12d,r8d
> + ror r13d,5
> + xor r14d,r10d
> + and r12d,ecx
> + xor r13d,ecx
> + add r9d,DWORD[40+rsp]
> + mov r15d,r10d
> + xor r12d,r8d
> + ror r14d,11
> + xor r15d,r11d
> + add r9d,r12d
> + ror r13d,6
> + and edi,r15d
> + xor r14d,r10d
> + add r9d,r13d
> + xor edi,r11d
> + ror r14d,2
> + add ebx,r9d
> + add r9d,edi
> + mov r13d,ebx
> + add r14d,r9d
> + ror r13d,14
> + mov r9d,r14d
> + mov r12d,ecx
> + ror r14d,9
> + xor r13d,ebx
> + xor r12d,edx
> + ror r13d,5
> + xor r14d,r9d
> + and r12d,ebx
> + xor r13d,ebx
> + add r8d,DWORD[44+rsp]
> + mov edi,r9d
> + xor r12d,edx
> + ror r14d,11
> + xor edi,r10d
> + add r8d,r12d
> + ror r13d,6
> + and r15d,edi
> + xor r14d,r9d
> + add r8d,r13d
> + xor r15d,r10d
> + ror r14d,2
> + add eax,r8d
> + add r8d,r15d
> + mov r13d,eax
> + add r14d,r8d
> + ror r13d,14
> + mov r8d,r14d
> + mov r12d,ebx
> + ror r14d,9
> + xor r13d,eax
> + xor r12d,ecx
> + ror r13d,5
> + xor r14d,r8d
> + and r12d,eax
> + xor r13d,eax
> + add edx,DWORD[48+rsp]
> + mov r15d,r8d
> + xor r12d,ecx
> + ror r14d,11
> + xor r15d,r9d
> + add edx,r12d
> + ror r13d,6
> + and edi,r15d
> + xor r14d,r8d
> + add edx,r13d
> + xor edi,r9d
> + ror r14d,2
> + add r11d,edx
> + add edx,edi
> + mov r13d,r11d
> + add r14d,edx
> + ror r13d,14
> + mov edx,r14d
> + mov r12d,eax
> + ror r14d,9
> + xor r13d,r11d
> + xor r12d,ebx
> + ror r13d,5
> + xor r14d,edx
> + and r12d,r11d
> + xor r13d,r11d
> + add ecx,DWORD[52+rsp]
> + mov edi,edx
> + xor r12d,ebx
> + ror r14d,11
> + xor edi,r8d
> + add ecx,r12d
> + ror r13d,6
> + and r15d,edi
> + xor r14d,edx
> + add ecx,r13d
> + xor r15d,r8d
> + ror r14d,2
> + add r10d,ecx
> + add ecx,r15d
> + mov r13d,r10d
> + add r14d,ecx
> + ror r13d,14
> + mov ecx,r14d
> + mov r12d,r11d
> + ror r14d,9
> + xor r13d,r10d
> + xor r12d,eax
> + ror r13d,5
> + xor r14d,ecx
> + and r12d,r10d
> + xor r13d,r10d
> + add ebx,DWORD[56+rsp]
> + mov r15d,ecx
> + xor r12d,eax
> + ror r14d,11
> + xor r15d,edx
> + add ebx,r12d
> + ror r13d,6
> + and edi,r15d
> + xor r14d,ecx
> + add ebx,r13d
> + xor edi,edx
> + ror r14d,2
> + add r9d,ebx
> + add ebx,edi
> + mov r13d,r9d
> + add r14d,ebx
> + ror r13d,14
> + mov ebx,r14d
> + mov r12d,r10d
> + ror r14d,9
> + xor r13d,r9d
> + xor r12d,r11d
> + ror r13d,5
> + xor r14d,ebx
> + and r12d,r9d
> + xor r13d,r9d
> + add eax,DWORD[60+rsp]
> + mov edi,ebx
> + xor r12d,r11d
> + ror r14d,11
> + xor edi,ecx
> + add eax,r12d
> + ror r13d,6
> + and r15d,edi
> + xor r14d,ebx
> + add eax,r13d
> + xor r15d,ecx
> + ror r14d,2
> + add r8d,eax
> + add eax,r15d
> + mov r13d,r8d
> + add r14d,eax
> + mov rdi,QWORD[((64+0))+rsp]
> + mov eax,r14d
> +
> + add eax,DWORD[rdi]
> + lea rsi,[64+rsi]
> + add ebx,DWORD[4+rdi]
> + add ecx,DWORD[8+rdi]
> + add edx,DWORD[12+rdi]
> + add r8d,DWORD[16+rdi]
> + add r9d,DWORD[20+rdi]
> + add r10d,DWORD[24+rdi]
> + add r11d,DWORD[28+rdi]
> +
> + cmp rsi,QWORD[((64+16))+rsp]
> +
> + mov DWORD[rdi],eax
> + mov DWORD[4+rdi],ebx
> + mov DWORD[8+rdi],ecx
> + mov DWORD[12+rdi],edx
> + mov DWORD[16+rdi],r8d
> + mov DWORD[20+rdi],r9d
> + mov DWORD[24+rdi],r10d
> + mov DWORD[28+rdi],r11d
> + jb NEAR $L$loop_ssse3
> +
> + mov rsi,QWORD[88+rsp]
> +
> + movaps xmm6,XMMWORD[((64+32))+rsp]
> + movaps xmm7,XMMWORD[((64+48))+rsp]
> + movaps xmm8,XMMWORD[((64+64))+rsp]
> + movaps xmm9,XMMWORD[((64+80))+rsp]
> + mov r15,QWORD[((-48))+rsi]
> +
> + mov r14,QWORD[((-40))+rsi]
> +
> + mov r13,QWORD[((-32))+rsi]
> +
> + mov r12,QWORD[((-24))+rsi]
> +
> + mov rbp,QWORD[((-16))+rsi]
> +
> + mov rbx,QWORD[((-8))+rsi]
> +
> + lea rsp,[rsi]
> +
> +$L$epilogue_ssse3:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha256_block_data_order_ssse3:
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$in_prologue
> + mov rsi,rax
> + mov rax,QWORD[((64+24))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov r12,QWORD[((-24))+rax]
> + mov r13,QWORD[((-32))+rax]
> + mov r14,QWORD[((-40))+rax]
> + mov r15,QWORD[((-48))+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> + mov QWORD[216+r8],r12
> + mov QWORD[224+r8],r13
> + mov QWORD[232+r8],r14
> + mov QWORD[240+r8],r15
> +
> + lea r10,[$L$epilogue]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + lea rsi,[((64+32))+rsi]
> + lea rdi,[512+r8]
> + mov ecx,8
> + DD 0xa548f3fc
> +
> +$L$in_prologue:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +
> +ALIGN 16
> +shaext_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + lea r10,[$L$prologue_shaext]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + lea r10,[$L$epilogue_shaext]
> + cmp rbx,r10
> + jae NEAR $L$in_prologue
> +
> + lea rsi,[((-8-80))+rax]
> + lea rdi,[512+r8]
> + mov ecx,10
> + DD 0xa548f3fc
> +
> + jmp NEAR $L$in_prologue
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_sha256_block_data_order wrt ..imagebase
> + DD $L$SEH_end_sha256_block_data_order wrt ..imagebase
> + DD $L$SEH_info_sha256_block_data_order wrt ..imagebase
> + DD $L$SEH_begin_sha256_block_data_order_shaext wrt ..imagebase
> + DD $L$SEH_end_sha256_block_data_order_shaext wrt ..imagebase
> + DD $L$SEH_info_sha256_block_data_order_shaext wrt ..imagebase
> + DD $L$SEH_begin_sha256_block_data_order_ssse3 wrt ..imagebase
> + DD $L$SEH_end_sha256_block_data_order_ssse3 wrt ..imagebase
> + DD $L$SEH_info_sha256_block_data_order_ssse3 wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_sha256_block_data_order:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$prologue wrt ..imagebase,$L$epilogue wrt ..imagebase
> +$L$SEH_info_sha256_block_data_order_shaext:
> +DB 9,0,0,0
> + DD shaext_handler wrt ..imagebase
> +$L$SEH_info_sha256_block_data_order_ssse3:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$prologue_ssse3 wrt ..imagebase,$L$epilogue_ssse3
> wrt ..imagebase
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha512-x86_64.nasm
> b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha512-x86_64.nasm
> new file mode 100644
> index 0000000000..c6397d4393
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/sha/sha512-x86_64.nasm
> @@ -0,0 +1,1938 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/sha/asm/sha512-x86_64.pl
> +;
> +; Copyright 2005-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +section .text code align=64
> +
> +
> +EXTERN OPENSSL_ia32cap_P
> +global sha512_block_data_order
> +
> +ALIGN 16
> +sha512_block_data_order:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_sha512_block_data_order:
> + mov rdi,rcx
> + mov rsi,rdx
> + mov rdx,r8
> +
> +
> +
> + mov rax,rsp
> +
> + push rbx
> +
> + push rbp
> +
> + push r12
> +
> + push r13
> +
> + push r14
> +
> + push r15
> +
> + shl rdx,4
> + sub rsp,16*8+4*8
> + lea rdx,[rdx*8+rsi]
> + and rsp,-64
> + mov QWORD[((128+0))+rsp],rdi
> + mov QWORD[((128+8))+rsp],rsi
> + mov QWORD[((128+16))+rsp],rdx
> + mov QWORD[152+rsp],rax
> +
> +$L$prologue:
> +
> + mov rax,QWORD[rdi]
> + mov rbx,QWORD[8+rdi]
> + mov rcx,QWORD[16+rdi]
> + mov rdx,QWORD[24+rdi]
> + mov r8,QWORD[32+rdi]
> + mov r9,QWORD[40+rdi]
> + mov r10,QWORD[48+rdi]
> + mov r11,QWORD[56+rdi]
> + jmp NEAR $L$loop
> +
> +ALIGN 16
> +$L$loop:
> + mov rdi,rbx
> + lea rbp,[K512]
> + xor rdi,rcx
> + mov r12,QWORD[rsi]
> + mov r13,r8
> + mov r14,rax
> + bswap r12
> + ror r13,23
> + mov r15,r9
> +
> + xor r13,r8
> + ror r14,5
> + xor r15,r10
> +
> + mov QWORD[rsp],r12
> + xor r14,rax
> + and r15,r8
> +
> + ror r13,4
> + add r12,r11
> + xor r15,r10
> +
> + ror r14,6
> + xor r13,r8
> + add r12,r15
> +
> + mov r15,rax
> + add r12,QWORD[rbp]
> + xor r14,rax
> +
> + xor r15,rbx
> + ror r13,14
> + mov r11,rbx
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor r11,rdi
> + add rdx,r12
> + add r11,r12
> +
> + lea rbp,[8+rbp]
> + add r11,r14
> + mov r12,QWORD[8+rsi]
> + mov r13,rdx
> + mov r14,r11
> + bswap r12
> + ror r13,23
> + mov rdi,r8
> +
> + xor r13,rdx
> + ror r14,5
> + xor rdi,r9
> +
> + mov QWORD[8+rsp],r12
> + xor r14,r11
> + and rdi,rdx
> +
> + ror r13,4
> + add r12,r10
> + xor rdi,r9
> +
> + ror r14,6
> + xor r13,rdx
> + add r12,rdi
> +
> + mov rdi,r11
> + add r12,QWORD[rbp]
> + xor r14,r11
> +
> + xor rdi,rax
> + ror r13,14
> + mov r10,rax
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor r10,r15
> + add rcx,r12
> + add r10,r12
> +
> + lea rbp,[24+rbp]
> + add r10,r14
> + mov r12,QWORD[16+rsi]
> + mov r13,rcx
> + mov r14,r10
> + bswap r12
> + ror r13,23
> + mov r15,rdx
> +
> + xor r13,rcx
> + ror r14,5
> + xor r15,r8
> +
> + mov QWORD[16+rsp],r12
> + xor r14,r10
> + and r15,rcx
> +
> + ror r13,4
> + add r12,r9
> + xor r15,r8
> +
> + ror r14,6
> + xor r13,rcx
> + add r12,r15
> +
> + mov r15,r10
> + add r12,QWORD[rbp]
> + xor r14,r10
> +
> + xor r15,r11
> + ror r13,14
> + mov r9,r11
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor r9,rdi
> + add rbx,r12
> + add r9,r12
> +
> + lea rbp,[8+rbp]
> + add r9,r14
> + mov r12,QWORD[24+rsi]
> + mov r13,rbx
> + mov r14,r9
> + bswap r12
> + ror r13,23
> + mov rdi,rcx
> +
> + xor r13,rbx
> + ror r14,5
> + xor rdi,rdx
> +
> + mov QWORD[24+rsp],r12
> + xor r14,r9
> + and rdi,rbx
> +
> + ror r13,4
> + add r12,r8
> + xor rdi,rdx
> +
> + ror r14,6
> + xor r13,rbx
> + add r12,rdi
> +
> + mov rdi,r9
> + add r12,QWORD[rbp]
> + xor r14,r9
> +
> + xor rdi,r10
> + ror r13,14
> + mov r8,r10
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor r8,r15
> + add rax,r12
> + add r8,r12
> +
> + lea rbp,[24+rbp]
> + add r8,r14
> + mov r12,QWORD[32+rsi]
> + mov r13,rax
> + mov r14,r8
> + bswap r12
> + ror r13,23
> + mov r15,rbx
> +
> + xor r13,rax
> + ror r14,5
> + xor r15,rcx
> +
> + mov QWORD[32+rsp],r12
> + xor r14,r8
> + and r15,rax
> +
> + ror r13,4
> + add r12,rdx
> + xor r15,rcx
> +
> + ror r14,6
> + xor r13,rax
> + add r12,r15
> +
> + mov r15,r8
> + add r12,QWORD[rbp]
> + xor r14,r8
> +
> + xor r15,r9
> + ror r13,14
> + mov rdx,r9
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor rdx,rdi
> + add r11,r12
> + add rdx,r12
> +
> + lea rbp,[8+rbp]
> + add rdx,r14
> + mov r12,QWORD[40+rsi]
> + mov r13,r11
> + mov r14,rdx
> + bswap r12
> + ror r13,23
> + mov rdi,rax
> +
> + xor r13,r11
> + ror r14,5
> + xor rdi,rbx
> +
> + mov QWORD[40+rsp],r12
> + xor r14,rdx
> + and rdi,r11
> +
> + ror r13,4
> + add r12,rcx
> + xor rdi,rbx
> +
> + ror r14,6
> + xor r13,r11
> + add r12,rdi
> +
> + mov rdi,rdx
> + add r12,QWORD[rbp]
> + xor r14,rdx
> +
> + xor rdi,r8
> + ror r13,14
> + mov rcx,r8
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor rcx,r15
> + add r10,r12
> + add rcx,r12
> +
> + lea rbp,[24+rbp]
> + add rcx,r14
> + mov r12,QWORD[48+rsi]
> + mov r13,r10
> + mov r14,rcx
> + bswap r12
> + ror r13,23
> + mov r15,r11
> +
> + xor r13,r10
> + ror r14,5
> + xor r15,rax
> +
> + mov QWORD[48+rsp],r12
> + xor r14,rcx
> + and r15,r10
> +
> + ror r13,4
> + add r12,rbx
> + xor r15,rax
> +
> + ror r14,6
> + xor r13,r10
> + add r12,r15
> +
> + mov r15,rcx
> + add r12,QWORD[rbp]
> + xor r14,rcx
> +
> + xor r15,rdx
> + ror r13,14
> + mov rbx,rdx
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor rbx,rdi
> + add r9,r12
> + add rbx,r12
> +
> + lea rbp,[8+rbp]
> + add rbx,r14
> + mov r12,QWORD[56+rsi]
> + mov r13,r9
> + mov r14,rbx
> + bswap r12
> + ror r13,23
> + mov rdi,r10
> +
> + xor r13,r9
> + ror r14,5
> + xor rdi,r11
> +
> + mov QWORD[56+rsp],r12
> + xor r14,rbx
> + and rdi,r9
> +
> + ror r13,4
> + add r12,rax
> + xor rdi,r11
> +
> + ror r14,6
> + xor r13,r9
> + add r12,rdi
> +
> + mov rdi,rbx
> + add r12,QWORD[rbp]
> + xor r14,rbx
> +
> + xor rdi,rcx
> + ror r13,14
> + mov rax,rcx
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor rax,r15
> + add r8,r12
> + add rax,r12
> +
> + lea rbp,[24+rbp]
> + add rax,r14
> + mov r12,QWORD[64+rsi]
> + mov r13,r8
> + mov r14,rax
> + bswap r12
> + ror r13,23
> + mov r15,r9
> +
> + xor r13,r8
> + ror r14,5
> + xor r15,r10
> +
> + mov QWORD[64+rsp],r12
> + xor r14,rax
> + and r15,r8
> +
> + ror r13,4
> + add r12,r11
> + xor r15,r10
> +
> + ror r14,6
> + xor r13,r8
> + add r12,r15
> +
> + mov r15,rax
> + add r12,QWORD[rbp]
> + xor r14,rax
> +
> + xor r15,rbx
> + ror r13,14
> + mov r11,rbx
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor r11,rdi
> + add rdx,r12
> + add r11,r12
> +
> + lea rbp,[8+rbp]
> + add r11,r14
> + mov r12,QWORD[72+rsi]
> + mov r13,rdx
> + mov r14,r11
> + bswap r12
> + ror r13,23
> + mov rdi,r8
> +
> + xor r13,rdx
> + ror r14,5
> + xor rdi,r9
> +
> + mov QWORD[72+rsp],r12
> + xor r14,r11
> + and rdi,rdx
> +
> + ror r13,4
> + add r12,r10
> + xor rdi,r9
> +
> + ror r14,6
> + xor r13,rdx
> + add r12,rdi
> +
> + mov rdi,r11
> + add r12,QWORD[rbp]
> + xor r14,r11
> +
> + xor rdi,rax
> + ror r13,14
> + mov r10,rax
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor r10,r15
> + add rcx,r12
> + add r10,r12
> +
> + lea rbp,[24+rbp]
> + add r10,r14
> + mov r12,QWORD[80+rsi]
> + mov r13,rcx
> + mov r14,r10
> + bswap r12
> + ror r13,23
> + mov r15,rdx
> +
> + xor r13,rcx
> + ror r14,5
> + xor r15,r8
> +
> + mov QWORD[80+rsp],r12
> + xor r14,r10
> + and r15,rcx
> +
> + ror r13,4
> + add r12,r9
> + xor r15,r8
> +
> + ror r14,6
> + xor r13,rcx
> + add r12,r15
> +
> + mov r15,r10
> + add r12,QWORD[rbp]
> + xor r14,r10
> +
> + xor r15,r11
> + ror r13,14
> + mov r9,r11
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor r9,rdi
> + add rbx,r12
> + add r9,r12
> +
> + lea rbp,[8+rbp]
> + add r9,r14
> + mov r12,QWORD[88+rsi]
> + mov r13,rbx
> + mov r14,r9
> + bswap r12
> + ror r13,23
> + mov rdi,rcx
> +
> + xor r13,rbx
> + ror r14,5
> + xor rdi,rdx
> +
> + mov QWORD[88+rsp],r12
> + xor r14,r9
> + and rdi,rbx
> +
> + ror r13,4
> + add r12,r8
> + xor rdi,rdx
> +
> + ror r14,6
> + xor r13,rbx
> + add r12,rdi
> +
> + mov rdi,r9
> + add r12,QWORD[rbp]
> + xor r14,r9
> +
> + xor rdi,r10
> + ror r13,14
> + mov r8,r10
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor r8,r15
> + add rax,r12
> + add r8,r12
> +
> + lea rbp,[24+rbp]
> + add r8,r14
> + mov r12,QWORD[96+rsi]
> + mov r13,rax
> + mov r14,r8
> + bswap r12
> + ror r13,23
> + mov r15,rbx
> +
> + xor r13,rax
> + ror r14,5
> + xor r15,rcx
> +
> + mov QWORD[96+rsp],r12
> + xor r14,r8
> + and r15,rax
> +
> + ror r13,4
> + add r12,rdx
> + xor r15,rcx
> +
> + ror r14,6
> + xor r13,rax
> + add r12,r15
> +
> + mov r15,r8
> + add r12,QWORD[rbp]
> + xor r14,r8
> +
> + xor r15,r9
> + ror r13,14
> + mov rdx,r9
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor rdx,rdi
> + add r11,r12
> + add rdx,r12
> +
> + lea rbp,[8+rbp]
> + add rdx,r14
> + mov r12,QWORD[104+rsi]
> + mov r13,r11
> + mov r14,rdx
> + bswap r12
> + ror r13,23
> + mov rdi,rax
> +
> + xor r13,r11
> + ror r14,5
> + xor rdi,rbx
> +
> + mov QWORD[104+rsp],r12
> + xor r14,rdx
> + and rdi,r11
> +
> + ror r13,4
> + add r12,rcx
> + xor rdi,rbx
> +
> + ror r14,6
> + xor r13,r11
> + add r12,rdi
> +
> + mov rdi,rdx
> + add r12,QWORD[rbp]
> + xor r14,rdx
> +
> + xor rdi,r8
> + ror r13,14
> + mov rcx,r8
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor rcx,r15
> + add r10,r12
> + add rcx,r12
> +
> + lea rbp,[24+rbp]
> + add rcx,r14
> + mov r12,QWORD[112+rsi]
> + mov r13,r10
> + mov r14,rcx
> + bswap r12
> + ror r13,23
> + mov r15,r11
> +
> + xor r13,r10
> + ror r14,5
> + xor r15,rax
> +
> + mov QWORD[112+rsp],r12
> + xor r14,rcx
> + and r15,r10
> +
> + ror r13,4
> + add r12,rbx
> + xor r15,rax
> +
> + ror r14,6
> + xor r13,r10
> + add r12,r15
> +
> + mov r15,rcx
> + add r12,QWORD[rbp]
> + xor r14,rcx
> +
> + xor r15,rdx
> + ror r13,14
> + mov rbx,rdx
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor rbx,rdi
> + add r9,r12
> + add rbx,r12
> +
> + lea rbp,[8+rbp]
> + add rbx,r14
> + mov r12,QWORD[120+rsi]
> + mov r13,r9
> + mov r14,rbx
> + bswap r12
> + ror r13,23
> + mov rdi,r10
> +
> + xor r13,r9
> + ror r14,5
> + xor rdi,r11
> +
> + mov QWORD[120+rsp],r12
> + xor r14,rbx
> + and rdi,r9
> +
> + ror r13,4
> + add r12,rax
> + xor rdi,r11
> +
> + ror r14,6
> + xor r13,r9
> + add r12,rdi
> +
> + mov rdi,rbx
> + add r12,QWORD[rbp]
> + xor r14,rbx
> +
> + xor rdi,rcx
> + ror r13,14
> + mov rax,rcx
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor rax,r15
> + add r8,r12
> + add rax,r12
> +
> + lea rbp,[24+rbp]
> + jmp NEAR $L$rounds_16_xx
> +ALIGN 16
> +$L$rounds_16_xx:
> + mov r13,QWORD[8+rsp]
> + mov r15,QWORD[112+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add rax,r14
> + mov r14,r15
> + ror r15,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor r15,r14
> + shr r14,6
> +
> + ror r15,19
> + xor r12,r13
> + xor r15,r14
> + add r12,QWORD[72+rsp]
> +
> + add r12,QWORD[rsp]
> + mov r13,r8
> + add r12,r15
> + mov r14,rax
> + ror r13,23
> + mov r15,r9
> +
> + xor r13,r8
> + ror r14,5
> + xor r15,r10
> +
> + mov QWORD[rsp],r12
> + xor r14,rax
> + and r15,r8
> +
> + ror r13,4
> + add r12,r11
> + xor r15,r10
> +
> + ror r14,6
> + xor r13,r8
> + add r12,r15
> +
> + mov r15,rax
> + add r12,QWORD[rbp]
> + xor r14,rax
> +
> + xor r15,rbx
> + ror r13,14
> + mov r11,rbx
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor r11,rdi
> + add rdx,r12
> + add r11,r12
> +
> + lea rbp,[8+rbp]
> + mov r13,QWORD[16+rsp]
> + mov rdi,QWORD[120+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add r11,r14
> + mov r14,rdi
> + ror rdi,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor rdi,r14
> + shr r14,6
> +
> + ror rdi,19
> + xor r12,r13
> + xor rdi,r14
> + add r12,QWORD[80+rsp]
> +
> + add r12,QWORD[8+rsp]
> + mov r13,rdx
> + add r12,rdi
> + mov r14,r11
> + ror r13,23
> + mov rdi,r8
> +
> + xor r13,rdx
> + ror r14,5
> + xor rdi,r9
> +
> + mov QWORD[8+rsp],r12
> + xor r14,r11
> + and rdi,rdx
> +
> + ror r13,4
> + add r12,r10
> + xor rdi,r9
> +
> + ror r14,6
> + xor r13,rdx
> + add r12,rdi
> +
> + mov rdi,r11
> + add r12,QWORD[rbp]
> + xor r14,r11
> +
> + xor rdi,rax
> + ror r13,14
> + mov r10,rax
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor r10,r15
> + add rcx,r12
> + add r10,r12
> +
> + lea rbp,[24+rbp]
> + mov r13,QWORD[24+rsp]
> + mov r15,QWORD[rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add r10,r14
> + mov r14,r15
> + ror r15,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor r15,r14
> + shr r14,6
> +
> + ror r15,19
> + xor r12,r13
> + xor r15,r14
> + add r12,QWORD[88+rsp]
> +
> + add r12,QWORD[16+rsp]
> + mov r13,rcx
> + add r12,r15
> + mov r14,r10
> + ror r13,23
> + mov r15,rdx
> +
> + xor r13,rcx
> + ror r14,5
> + xor r15,r8
> +
> + mov QWORD[16+rsp],r12
> + xor r14,r10
> + and r15,rcx
> +
> + ror r13,4
> + add r12,r9
> + xor r15,r8
> +
> + ror r14,6
> + xor r13,rcx
> + add r12,r15
> +
> + mov r15,r10
> + add r12,QWORD[rbp]
> + xor r14,r10
> +
> + xor r15,r11
> + ror r13,14
> + mov r9,r11
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor r9,rdi
> + add rbx,r12
> + add r9,r12
> +
> + lea rbp,[8+rbp]
> + mov r13,QWORD[32+rsp]
> + mov rdi,QWORD[8+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add r9,r14
> + mov r14,rdi
> + ror rdi,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor rdi,r14
> + shr r14,6
> +
> + ror rdi,19
> + xor r12,r13
> + xor rdi,r14
> + add r12,QWORD[96+rsp]
> +
> + add r12,QWORD[24+rsp]
> + mov r13,rbx
> + add r12,rdi
> + mov r14,r9
> + ror r13,23
> + mov rdi,rcx
> +
> + xor r13,rbx
> + ror r14,5
> + xor rdi,rdx
> +
> + mov QWORD[24+rsp],r12
> + xor r14,r9
> + and rdi,rbx
> +
> + ror r13,4
> + add r12,r8
> + xor rdi,rdx
> +
> + ror r14,6
> + xor r13,rbx
> + add r12,rdi
> +
> + mov rdi,r9
> + add r12,QWORD[rbp]
> + xor r14,r9
> +
> + xor rdi,r10
> + ror r13,14
> + mov r8,r10
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor r8,r15
> + add rax,r12
> + add r8,r12
> +
> + lea rbp,[24+rbp]
> + mov r13,QWORD[40+rsp]
> + mov r15,QWORD[16+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add r8,r14
> + mov r14,r15
> + ror r15,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor r15,r14
> + shr r14,6
> +
> + ror r15,19
> + xor r12,r13
> + xor r15,r14
> + add r12,QWORD[104+rsp]
> +
> + add r12,QWORD[32+rsp]
> + mov r13,rax
> + add r12,r15
> + mov r14,r8
> + ror r13,23
> + mov r15,rbx
> +
> + xor r13,rax
> + ror r14,5
> + xor r15,rcx
> +
> + mov QWORD[32+rsp],r12
> + xor r14,r8
> + and r15,rax
> +
> + ror r13,4
> + add r12,rdx
> + xor r15,rcx
> +
> + ror r14,6
> + xor r13,rax
> + add r12,r15
> +
> + mov r15,r8
> + add r12,QWORD[rbp]
> + xor r14,r8
> +
> + xor r15,r9
> + ror r13,14
> + mov rdx,r9
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor rdx,rdi
> + add r11,r12
> + add rdx,r12
> +
> + lea rbp,[8+rbp]
> + mov r13,QWORD[48+rsp]
> + mov rdi,QWORD[24+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add rdx,r14
> + mov r14,rdi
> + ror rdi,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor rdi,r14
> + shr r14,6
> +
> + ror rdi,19
> + xor r12,r13
> + xor rdi,r14
> + add r12,QWORD[112+rsp]
> +
> + add r12,QWORD[40+rsp]
> + mov r13,r11
> + add r12,rdi
> + mov r14,rdx
> + ror r13,23
> + mov rdi,rax
> +
> + xor r13,r11
> + ror r14,5
> + xor rdi,rbx
> +
> + mov QWORD[40+rsp],r12
> + xor r14,rdx
> + and rdi,r11
> +
> + ror r13,4
> + add r12,rcx
> + xor rdi,rbx
> +
> + ror r14,6
> + xor r13,r11
> + add r12,rdi
> +
> + mov rdi,rdx
> + add r12,QWORD[rbp]
> + xor r14,rdx
> +
> + xor rdi,r8
> + ror r13,14
> + mov rcx,r8
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor rcx,r15
> + add r10,r12
> + add rcx,r12
> +
> + lea rbp,[24+rbp]
> + mov r13,QWORD[56+rsp]
> + mov r15,QWORD[32+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add rcx,r14
> + mov r14,r15
> + ror r15,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor r15,r14
> + shr r14,6
> +
> + ror r15,19
> + xor r12,r13
> + xor r15,r14
> + add r12,QWORD[120+rsp]
> +
> + add r12,QWORD[48+rsp]
> + mov r13,r10
> + add r12,r15
> + mov r14,rcx
> + ror r13,23
> + mov r15,r11
> +
> + xor r13,r10
> + ror r14,5
> + xor r15,rax
> +
> + mov QWORD[48+rsp],r12
> + xor r14,rcx
> + and r15,r10
> +
> + ror r13,4
> + add r12,rbx
> + xor r15,rax
> +
> + ror r14,6
> + xor r13,r10
> + add r12,r15
> +
> + mov r15,rcx
> + add r12,QWORD[rbp]
> + xor r14,rcx
> +
> + xor r15,rdx
> + ror r13,14
> + mov rbx,rdx
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor rbx,rdi
> + add r9,r12
> + add rbx,r12
> +
> + lea rbp,[8+rbp]
> + mov r13,QWORD[64+rsp]
> + mov rdi,QWORD[40+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add rbx,r14
> + mov r14,rdi
> + ror rdi,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor rdi,r14
> + shr r14,6
> +
> + ror rdi,19
> + xor r12,r13
> + xor rdi,r14
> + add r12,QWORD[rsp]
> +
> + add r12,QWORD[56+rsp]
> + mov r13,r9
> + add r12,rdi
> + mov r14,rbx
> + ror r13,23
> + mov rdi,r10
> +
> + xor r13,r9
> + ror r14,5
> + xor rdi,r11
> +
> + mov QWORD[56+rsp],r12
> + xor r14,rbx
> + and rdi,r9
> +
> + ror r13,4
> + add r12,rax
> + xor rdi,r11
> +
> + ror r14,6
> + xor r13,r9
> + add r12,rdi
> +
> + mov rdi,rbx
> + add r12,QWORD[rbp]
> + xor r14,rbx
> +
> + xor rdi,rcx
> + ror r13,14
> + mov rax,rcx
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor rax,r15
> + add r8,r12
> + add rax,r12
> +
> + lea rbp,[24+rbp]
> + mov r13,QWORD[72+rsp]
> + mov r15,QWORD[48+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add rax,r14
> + mov r14,r15
> + ror r15,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor r15,r14
> + shr r14,6
> +
> + ror r15,19
> + xor r12,r13
> + xor r15,r14
> + add r12,QWORD[8+rsp]
> +
> + add r12,QWORD[64+rsp]
> + mov r13,r8
> + add r12,r15
> + mov r14,rax
> + ror r13,23
> + mov r15,r9
> +
> + xor r13,r8
> + ror r14,5
> + xor r15,r10
> +
> + mov QWORD[64+rsp],r12
> + xor r14,rax
> + and r15,r8
> +
> + ror r13,4
> + add r12,r11
> + xor r15,r10
> +
> + ror r14,6
> + xor r13,r8
> + add r12,r15
> +
> + mov r15,rax
> + add r12,QWORD[rbp]
> + xor r14,rax
> +
> + xor r15,rbx
> + ror r13,14
> + mov r11,rbx
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor r11,rdi
> + add rdx,r12
> + add r11,r12
> +
> + lea rbp,[8+rbp]
> + mov r13,QWORD[80+rsp]
> + mov rdi,QWORD[56+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add r11,r14
> + mov r14,rdi
> + ror rdi,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor rdi,r14
> + shr r14,6
> +
> + ror rdi,19
> + xor r12,r13
> + xor rdi,r14
> + add r12,QWORD[16+rsp]
> +
> + add r12,QWORD[72+rsp]
> + mov r13,rdx
> + add r12,rdi
> + mov r14,r11
> + ror r13,23
> + mov rdi,r8
> +
> + xor r13,rdx
> + ror r14,5
> + xor rdi,r9
> +
> + mov QWORD[72+rsp],r12
> + xor r14,r11
> + and rdi,rdx
> +
> + ror r13,4
> + add r12,r10
> + xor rdi,r9
> +
> + ror r14,6
> + xor r13,rdx
> + add r12,rdi
> +
> + mov rdi,r11
> + add r12,QWORD[rbp]
> + xor r14,r11
> +
> + xor rdi,rax
> + ror r13,14
> + mov r10,rax
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor r10,r15
> + add rcx,r12
> + add r10,r12
> +
> + lea rbp,[24+rbp]
> + mov r13,QWORD[88+rsp]
> + mov r15,QWORD[64+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add r10,r14
> + mov r14,r15
> + ror r15,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor r15,r14
> + shr r14,6
> +
> + ror r15,19
> + xor r12,r13
> + xor r15,r14
> + add r12,QWORD[24+rsp]
> +
> + add r12,QWORD[80+rsp]
> + mov r13,rcx
> + add r12,r15
> + mov r14,r10
> + ror r13,23
> + mov r15,rdx
> +
> + xor r13,rcx
> + ror r14,5
> + xor r15,r8
> +
> + mov QWORD[80+rsp],r12
> + xor r14,r10
> + and r15,rcx
> +
> + ror r13,4
> + add r12,r9
> + xor r15,r8
> +
> + ror r14,6
> + xor r13,rcx
> + add r12,r15
> +
> + mov r15,r10
> + add r12,QWORD[rbp]
> + xor r14,r10
> +
> + xor r15,r11
> + ror r13,14
> + mov r9,r11
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor r9,rdi
> + add rbx,r12
> + add r9,r12
> +
> + lea rbp,[8+rbp]
> + mov r13,QWORD[96+rsp]
> + mov rdi,QWORD[72+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add r9,r14
> + mov r14,rdi
> + ror rdi,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor rdi,r14
> + shr r14,6
> +
> + ror rdi,19
> + xor r12,r13
> + xor rdi,r14
> + add r12,QWORD[32+rsp]
> +
> + add r12,QWORD[88+rsp]
> + mov r13,rbx
> + add r12,rdi
> + mov r14,r9
> + ror r13,23
> + mov rdi,rcx
> +
> + xor r13,rbx
> + ror r14,5
> + xor rdi,rdx
> +
> + mov QWORD[88+rsp],r12
> + xor r14,r9
> + and rdi,rbx
> +
> + ror r13,4
> + add r12,r8
> + xor rdi,rdx
> +
> + ror r14,6
> + xor r13,rbx
> + add r12,rdi
> +
> + mov rdi,r9
> + add r12,QWORD[rbp]
> + xor r14,r9
> +
> + xor rdi,r10
> + ror r13,14
> + mov r8,r10
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor r8,r15
> + add rax,r12
> + add r8,r12
> +
> + lea rbp,[24+rbp]
> + mov r13,QWORD[104+rsp]
> + mov r15,QWORD[80+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add r8,r14
> + mov r14,r15
> + ror r15,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor r15,r14
> + shr r14,6
> +
> + ror r15,19
> + xor r12,r13
> + xor r15,r14
> + add r12,QWORD[40+rsp]
> +
> + add r12,QWORD[96+rsp]
> + mov r13,rax
> + add r12,r15
> + mov r14,r8
> + ror r13,23
> + mov r15,rbx
> +
> + xor r13,rax
> + ror r14,5
> + xor r15,rcx
> +
> + mov QWORD[96+rsp],r12
> + xor r14,r8
> + and r15,rax
> +
> + ror r13,4
> + add r12,rdx
> + xor r15,rcx
> +
> + ror r14,6
> + xor r13,rax
> + add r12,r15
> +
> + mov r15,r8
> + add r12,QWORD[rbp]
> + xor r14,r8
> +
> + xor r15,r9
> + ror r13,14
> + mov rdx,r9
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor rdx,rdi
> + add r11,r12
> + add rdx,r12
> +
> + lea rbp,[8+rbp]
> + mov r13,QWORD[112+rsp]
> + mov rdi,QWORD[88+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add rdx,r14
> + mov r14,rdi
> + ror rdi,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor rdi,r14
> + shr r14,6
> +
> + ror rdi,19
> + xor r12,r13
> + xor rdi,r14
> + add r12,QWORD[48+rsp]
> +
> + add r12,QWORD[104+rsp]
> + mov r13,r11
> + add r12,rdi
> + mov r14,rdx
> + ror r13,23
> + mov rdi,rax
> +
> + xor r13,r11
> + ror r14,5
> + xor rdi,rbx
> +
> + mov QWORD[104+rsp],r12
> + xor r14,rdx
> + and rdi,r11
> +
> + ror r13,4
> + add r12,rcx
> + xor rdi,rbx
> +
> + ror r14,6
> + xor r13,r11
> + add r12,rdi
> +
> + mov rdi,rdx
> + add r12,QWORD[rbp]
> + xor r14,rdx
> +
> + xor rdi,r8
> + ror r13,14
> + mov rcx,r8
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor rcx,r15
> + add r10,r12
> + add rcx,r12
> +
> + lea rbp,[24+rbp]
> + mov r13,QWORD[120+rsp]
> + mov r15,QWORD[96+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add rcx,r14
> + mov r14,r15
> + ror r15,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor r15,r14
> + shr r14,6
> +
> + ror r15,19
> + xor r12,r13
> + xor r15,r14
> + add r12,QWORD[56+rsp]
> +
> + add r12,QWORD[112+rsp]
> + mov r13,r10
> + add r12,r15
> + mov r14,rcx
> + ror r13,23
> + mov r15,r11
> +
> + xor r13,r10
> + ror r14,5
> + xor r15,rax
> +
> + mov QWORD[112+rsp],r12
> + xor r14,rcx
> + and r15,r10
> +
> + ror r13,4
> + add r12,rbx
> + xor r15,rax
> +
> + ror r14,6
> + xor r13,r10
> + add r12,r15
> +
> + mov r15,rcx
> + add r12,QWORD[rbp]
> + xor r14,rcx
> +
> + xor r15,rdx
> + ror r13,14
> + mov rbx,rdx
> +
> + and rdi,r15
> + ror r14,28
> + add r12,r13
> +
> + xor rbx,rdi
> + add r9,r12
> + add rbx,r12
> +
> + lea rbp,[8+rbp]
> + mov r13,QWORD[rsp]
> + mov rdi,QWORD[104+rsp]
> +
> + mov r12,r13
> + ror r13,7
> + add rbx,r14
> + mov r14,rdi
> + ror rdi,42
> +
> + xor r13,r12
> + shr r12,7
> + ror r13,1
> + xor rdi,r14
> + shr r14,6
> +
> + ror rdi,19
> + xor r12,r13
> + xor rdi,r14
> + add r12,QWORD[64+rsp]
> +
> + add r12,QWORD[120+rsp]
> + mov r13,r9
> + add r12,rdi
> + mov r14,rbx
> + ror r13,23
> + mov rdi,r10
> +
> + xor r13,r9
> + ror r14,5
> + xor rdi,r11
> +
> + mov QWORD[120+rsp],r12
> + xor r14,rbx
> + and rdi,r9
> +
> + ror r13,4
> + add r12,rax
> + xor rdi,r11
> +
> + ror r14,6
> + xor r13,r9
> + add r12,rdi
> +
> + mov rdi,rbx
> + add r12,QWORD[rbp]
> + xor r14,rbx
> +
> + xor rdi,rcx
> + ror r13,14
> + mov rax,rcx
> +
> + and r15,rdi
> + ror r14,28
> + add r12,r13
> +
> + xor rax,r15
> + add r8,r12
> + add rax,r12
> +
> + lea rbp,[24+rbp]
> + cmp BYTE[7+rbp],0
> + jnz NEAR $L$rounds_16_xx
> +
> + mov rdi,QWORD[((128+0))+rsp]
> + add rax,r14
> + lea rsi,[128+rsi]
> +
> + add rax,QWORD[rdi]
> + add rbx,QWORD[8+rdi]
> + add rcx,QWORD[16+rdi]
> + add rdx,QWORD[24+rdi]
> + add r8,QWORD[32+rdi]
> + add r9,QWORD[40+rdi]
> + add r10,QWORD[48+rdi]
> + add r11,QWORD[56+rdi]
> +
> + cmp rsi,QWORD[((128+16))+rsp]
> +
> + mov QWORD[rdi],rax
> + mov QWORD[8+rdi],rbx
> + mov QWORD[16+rdi],rcx
> + mov QWORD[24+rdi],rdx
> + mov QWORD[32+rdi],r8
> + mov QWORD[40+rdi],r9
> + mov QWORD[48+rdi],r10
> + mov QWORD[56+rdi],r11
> + jb NEAR $L$loop
> +
> + mov rsi,QWORD[152+rsp]
> +
> + mov r15,QWORD[((-48))+rsi]
> +
> + mov r14,QWORD[((-40))+rsi]
> +
> + mov r13,QWORD[((-32))+rsi]
> +
> + mov r12,QWORD[((-24))+rsi]
> +
> + mov rbp,QWORD[((-16))+rsi]
> +
> + mov rbx,QWORD[((-8))+rsi]
> +
> + lea rsp,[rsi]
> +
> +$L$epilogue:
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_sha512_block_data_order:
> +ALIGN 64
> +
> +K512:
> + DQ 0x428a2f98d728ae22,0x7137449123ef65cd
> + DQ 0x428a2f98d728ae22,0x7137449123ef65cd
> + DQ 0xb5c0fbcfec4d3b2f,0xe9b5dba58189dbbc
> + DQ 0xb5c0fbcfec4d3b2f,0xe9b5dba58189dbbc
> + DQ 0x3956c25bf348b538,0x59f111f1b605d019
> + DQ 0x3956c25bf348b538,0x59f111f1b605d019
> + DQ 0x923f82a4af194f9b,0xab1c5ed5da6d8118
> + DQ 0x923f82a4af194f9b,0xab1c5ed5da6d8118
> + DQ 0xd807aa98a3030242,0x12835b0145706fbe
> + DQ 0xd807aa98a3030242,0x12835b0145706fbe
> + DQ 0x243185be4ee4b28c,0x550c7dc3d5ffb4e2
> + DQ 0x243185be4ee4b28c,0x550c7dc3d5ffb4e2
> + DQ 0x72be5d74f27b896f,0x80deb1fe3b1696b1
> + DQ 0x72be5d74f27b896f,0x80deb1fe3b1696b1
> + DQ 0x9bdc06a725c71235,0xc19bf174cf692694
> + DQ 0x9bdc06a725c71235,0xc19bf174cf692694
> + DQ 0xe49b69c19ef14ad2,0xefbe4786384f25e3
> + DQ 0xe49b69c19ef14ad2,0xefbe4786384f25e3
> + DQ 0x0fc19dc68b8cd5b5,0x240ca1cc77ac9c65
> + DQ 0x0fc19dc68b8cd5b5,0x240ca1cc77ac9c65
> + DQ 0x2de92c6f592b0275,0x4a7484aa6ea6e483
> + DQ 0x2de92c6f592b0275,0x4a7484aa6ea6e483
> + DQ 0x5cb0a9dcbd41fbd4,0x76f988da831153b5
> + DQ 0x5cb0a9dcbd41fbd4,0x76f988da831153b5
> + DQ 0x983e5152ee66dfab,0xa831c66d2db43210
> + DQ 0x983e5152ee66dfab,0xa831c66d2db43210
> + DQ 0xb00327c898fb213f,0xbf597fc7beef0ee4
> + DQ 0xb00327c898fb213f,0xbf597fc7beef0ee4
> + DQ 0xc6e00bf33da88fc2,0xd5a79147930aa725
> + DQ 0xc6e00bf33da88fc2,0xd5a79147930aa725
> + DQ 0x06ca6351e003826f,0x142929670a0e6e70
> + DQ 0x06ca6351e003826f,0x142929670a0e6e70
> + DQ 0x27b70a8546d22ffc,0x2e1b21385c26c926
> + DQ 0x27b70a8546d22ffc,0x2e1b21385c26c926
> + DQ 0x4d2c6dfc5ac42aed,0x53380d139d95b3df
> + DQ 0x4d2c6dfc5ac42aed,0x53380d139d95b3df
> + DQ 0x650a73548baf63de,0x766a0abb3c77b2a8
> + DQ 0x650a73548baf63de,0x766a0abb3c77b2a8
> + DQ 0x81c2c92e47edaee6,0x92722c851482353b
> + DQ 0x81c2c92e47edaee6,0x92722c851482353b
> + DQ 0xa2bfe8a14cf10364,0xa81a664bbc423001
> + DQ 0xa2bfe8a14cf10364,0xa81a664bbc423001
> + DQ 0xc24b8b70d0f89791,0xc76c51a30654be30
> + DQ 0xc24b8b70d0f89791,0xc76c51a30654be30
> + DQ 0xd192e819d6ef5218,0xd69906245565a910
> + DQ 0xd192e819d6ef5218,0xd69906245565a910
> + DQ 0xf40e35855771202a,0x106aa07032bbd1b8
> + DQ 0xf40e35855771202a,0x106aa07032bbd1b8
> + DQ 0x19a4c116b8d2d0c8,0x1e376c085141ab53
> + DQ 0x19a4c116b8d2d0c8,0x1e376c085141ab53
> + DQ 0x2748774cdf8eeb99,0x34b0bcb5e19b48a8
> + DQ 0x2748774cdf8eeb99,0x34b0bcb5e19b48a8
> + DQ 0x391c0cb3c5c95a63,0x4ed8aa4ae3418acb
> + DQ 0x391c0cb3c5c95a63,0x4ed8aa4ae3418acb
> + DQ 0x5b9cca4f7763e373,0x682e6ff3d6b2b8a3
> + DQ 0x5b9cca4f7763e373,0x682e6ff3d6b2b8a3
> + DQ 0x748f82ee5defb2fc,0x78a5636f43172f60
> + DQ 0x748f82ee5defb2fc,0x78a5636f43172f60
> + DQ 0x84c87814a1f0ab72,0x8cc702081a6439ec
> + DQ 0x84c87814a1f0ab72,0x8cc702081a6439ec
> + DQ 0x90befffa23631e28,0xa4506cebde82bde9
> + DQ 0x90befffa23631e28,0xa4506cebde82bde9
> + DQ 0xbef9a3f7b2c67915,0xc67178f2e372532b
> + DQ 0xbef9a3f7b2c67915,0xc67178f2e372532b
> + DQ 0xca273eceea26619c,0xd186b8c721c0c207
> + DQ 0xca273eceea26619c,0xd186b8c721c0c207
> + DQ 0xeada7dd6cde0eb1e,0xf57d4f7fee6ed178
> + DQ 0xeada7dd6cde0eb1e,0xf57d4f7fee6ed178
> + DQ 0x06f067aa72176fba,0x0a637dc5a2c898a6
> + DQ 0x06f067aa72176fba,0x0a637dc5a2c898a6
> + DQ 0x113f9804bef90dae,0x1b710b35131c471b
> + DQ 0x113f9804bef90dae,0x1b710b35131c471b
> + DQ 0x28db77f523047d84,0x32caab7b40c72493
> + DQ 0x28db77f523047d84,0x32caab7b40c72493
> + DQ 0x3c9ebe0a15c9bebc,0x431d67c49c100d4c
> + DQ 0x3c9ebe0a15c9bebc,0x431d67c49c100d4c
> + DQ 0x4cc5d4becb3e42b6,0x597f299cfc657e2a
> + DQ 0x4cc5d4becb3e42b6,0x597f299cfc657e2a
> + DQ 0x5fcb6fab3ad6faec,0x6c44198c4a475817
> + DQ 0x5fcb6fab3ad6faec,0x6c44198c4a475817
> +
> + DQ 0x0001020304050607,0x08090a0b0c0d0e0f
> + DQ 0x0001020304050607,0x08090a0b0c0d0e0f
> +DB 83,72,65,53,49,50,32,98,108,111,99,107,32,116,114,97
> +DB 110,115,102,111,114,109,32,102,111,114,32,120,56,54,95,54
> +DB 52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121
> +DB 32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46
> +DB 111,114,103,62,0
> +EXTERN __imp_RtlVirtualUnwind
> +
> +ALIGN 16
> +se_handler:
> + push rsi
> + push rdi
> + push rbx
> + push rbp
> + push r12
> + push r13
> + push r14
> + push r15
> + pushfq
> + sub rsp,64
> +
> + mov rax,QWORD[120+r8]
> + mov rbx,QWORD[248+r8]
> +
> + mov rsi,QWORD[8+r9]
> + mov r11,QWORD[56+r9]
> +
> + mov r10d,DWORD[r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + mov rax,QWORD[152+r8]
> +
> + mov r10d,DWORD[4+r11]
> + lea r10,[r10*1+rsi]
> + cmp rbx,r10
> + jae NEAR $L$in_prologue
> + mov rsi,rax
> + mov rax,QWORD[((128+24))+rax]
> +
> + mov rbx,QWORD[((-8))+rax]
> + mov rbp,QWORD[((-16))+rax]
> + mov r12,QWORD[((-24))+rax]
> + mov r13,QWORD[((-32))+rax]
> + mov r14,QWORD[((-40))+rax]
> + mov r15,QWORD[((-48))+rax]
> + mov QWORD[144+r8],rbx
> + mov QWORD[160+r8],rbp
> + mov QWORD[216+r8],r12
> + mov QWORD[224+r8],r13
> + mov QWORD[232+r8],r14
> + mov QWORD[240+r8],r15
> +
> + lea r10,[$L$epilogue]
> + cmp rbx,r10
> + jb NEAR $L$in_prologue
> +
> + lea rsi,[((128+32))+rsi]
> + lea rdi,[512+r8]
> + mov ecx,12
> + DD 0xa548f3fc
> +
> +$L$in_prologue:
> + mov rdi,QWORD[8+rax]
> + mov rsi,QWORD[16+rax]
> + mov QWORD[152+r8],rax
> + mov QWORD[168+r8],rsi
> + mov QWORD[176+r8],rdi
> +
> + mov rdi,QWORD[40+r9]
> + mov rsi,r8
> + mov ecx,154
> + DD 0xa548f3fc
> +
> + mov rsi,r9
> + xor rcx,rcx
> + mov rdx,QWORD[8+rsi]
> + mov r8,QWORD[rsi]
> + mov r9,QWORD[16+rsi]
> + mov r10,QWORD[40+rsi]
> + lea r11,[56+rsi]
> + lea r12,[24+rsi]
> + mov QWORD[32+rsp],r10
> + mov QWORD[40+rsp],r11
> + mov QWORD[48+rsp],r12
> + mov QWORD[56+rsp],rcx
> + call QWORD[__imp_RtlVirtualUnwind]
> +
> + mov eax,1
> + add rsp,64
> + popfq
> + pop r15
> + pop r14
> + pop r13
> + pop r12
> + pop rbp
> + pop rbx
> + pop rdi
> + pop rsi
> + DB 0F3h,0C3h ;repret
> +
> +section .pdata rdata align=4
> +ALIGN 4
> + DD $L$SEH_begin_sha512_block_data_order wrt ..imagebase
> + DD $L$SEH_end_sha512_block_data_order wrt ..imagebase
> + DD $L$SEH_info_sha512_block_data_order wrt ..imagebase
> +section .xdata rdata align=8
> +ALIGN 8
> +$L$SEH_info_sha512_block_data_order:
> +DB 9,0,0,0
> + DD se_handler wrt ..imagebase
> + DD $L$prologue wrt ..imagebase,$L$epilogue wrt ..imagebase
> diff --git a/CryptoPkg/Library/OpensslLib/X64/crypto/x86_64cpuid.nasm
> b/CryptoPkg/Library/OpensslLib/X64/crypto/x86_64cpuid.nasm
> new file mode 100644
> index 0000000000..2a3d5bcf72
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64/crypto/x86_64cpuid.nasm
> @@ -0,0 +1,491 @@
> +; WARNING: do not edit!
> +; Generated from openssl/crypto/x86_64cpuid.pl
> +;
> +; Copyright 2005-2020 The OpenSSL Project Authors. All Rights Reserved.
> +;
> +; Licensed under the OpenSSL license (the "License"). You may not use
> +; this file except in compliance with the License. You can obtain a copy
> +; in the file LICENSE in the source distribution or at
> +; https://www.openssl.org/source/license.html
> +
> +default rel
> +%define XMMWORD
> +%define YMMWORD
> +%define ZMMWORD
> +EXTERN OPENSSL_cpuid_setup
> +
> +section .CRT$XCU rdata align=8
> + DQ OPENSSL_cpuid_setup
> +
> +
> +common OPENSSL_ia32cap_P 16
> +
> +section .text code align=64
> +
> +
> +global OPENSSL_atomic_add
> +
> +ALIGN 16
> +OPENSSL_atomic_add:
> +
> + mov eax,DWORD[rcx]
> +$L$spin: lea r8,[rax*1+rdx]
> +DB 0xf0
> + cmpxchg DWORD[rcx],r8d
> + jne NEAR $L$spin
> + mov eax,r8d
> +DB 0x48,0x98
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +global OPENSSL_rdtsc
> +
> +ALIGN 16
> +OPENSSL_rdtsc:
> +
> + rdtsc
> + shl rdx,32
> + or rax,rdx
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +global OPENSSL_ia32_cpuid
> +
> +ALIGN 16
> +OPENSSL_ia32_cpuid:
> + mov QWORD[8+rsp],rdi ;WIN64 prologue
> + mov QWORD[16+rsp],rsi
> + mov rax,rsp
> +$L$SEH_begin_OPENSSL_ia32_cpuid:
> + mov rdi,rcx
> +
> +
> +
> + mov r8,rbx
> +
> +
> + xor eax,eax
> + mov QWORD[8+rdi],rax
> + cpuid
> + mov r11d,eax
> +
> + xor eax,eax
> + cmp ebx,0x756e6547
> + setne al
> + mov r9d,eax
> + cmp edx,0x49656e69
> + setne al
> + or r9d,eax
> + cmp ecx,0x6c65746e
> + setne al
> + or r9d,eax
> + jz NEAR $L$intel
> +
> + cmp ebx,0x68747541
> + setne al
> + mov r10d,eax
> + cmp edx,0x69746E65
> + setne al
> + or r10d,eax
> + cmp ecx,0x444D4163
> + setne al
> + or r10d,eax
> + jnz NEAR $L$intel
> +
> +
> + mov eax,0x80000000
> + cpuid
> + cmp eax,0x80000001
> + jb NEAR $L$intel
> + mov r10d,eax
> + mov eax,0x80000001
> + cpuid
> + or r9d,ecx
> + and r9d,0x00000801
> +
> + cmp r10d,0x80000008
> + jb NEAR $L$intel
> +
> + mov eax,0x80000008
> + cpuid
> + movzx r10,cl
> + inc r10
> +
> + mov eax,1
> + cpuid
> + bt edx,28
> + jnc NEAR $L$generic
> + shr ebx,16
> + cmp bl,r10b
> + ja NEAR $L$generic
> + and edx,0xefffffff
> + jmp NEAR $L$generic
> +
> +$L$intel:
> + cmp r11d,4
> + mov r10d,-1
> + jb NEAR $L$nocacheinfo
> +
> + mov eax,4
> + mov ecx,0
> + cpuid
> + mov r10d,eax
> + shr r10d,14
> + and r10d,0xfff
> +
> +$L$nocacheinfo:
> + mov eax,1
> + cpuid
> + movd xmm0,eax
> + and edx,0xbfefffff
> + cmp r9d,0
> + jne NEAR $L$notintel
> + or edx,0x40000000
> + and ah,15
> + cmp ah,15
> + jne NEAR $L$notP4
> + or edx,0x00100000
> +$L$notP4:
> + cmp ah,6
> + jne NEAR $L$notintel
> + and eax,0x0fff0ff0
> + cmp eax,0x00050670
> + je NEAR $L$knights
> + cmp eax,0x00080650
> + jne NEAR $L$notintel
> +$L$knights:
> + and ecx,0xfbffffff
> +
> +$L$notintel:
> + bt edx,28
> + jnc NEAR $L$generic
> + and edx,0xefffffff
> + cmp r10d,0
> + je NEAR $L$generic
> +
> + or edx,0x10000000
> + shr ebx,16
> + cmp bl,1
> + ja NEAR $L$generic
> + and edx,0xefffffff
> +$L$generic:
> + and r9d,0x00000800
> + and ecx,0xfffff7ff
> + or r9d,ecx
> +
> + mov r10d,edx
> +
> + cmp r11d,7
> + jb NEAR $L$no_extended_info
> + mov eax,7
> + xor ecx,ecx
> + cpuid
> + bt r9d,26
> + jc NEAR $L$notknights
> + and ebx,0xfff7ffff
> +$L$notknights:
> + movd eax,xmm0
> + and eax,0x0fff0ff0
> + cmp eax,0x00050650
> + jne NEAR $L$notskylakex
> + and ebx,0xfffeffff
> +
> +$L$notskylakex:
> + mov DWORD[8+rdi],ebx
> + mov DWORD[12+rdi],ecx
> +$L$no_extended_info:
> +
> + bt r9d,27
> + jnc NEAR $L$clear_avx
> + xor ecx,ecx
> +DB 0x0f,0x01,0xd0
> + and eax,0xe6
> + cmp eax,0xe6
> + je NEAR $L$done
> + and DWORD[8+rdi],0x3fdeffff
> +
> +
> +
> +
> + and eax,6
> + cmp eax,6
> + je NEAR $L$done
> +$L$clear_avx:
> + mov eax,0xefffe7ff
> + and r9d,eax
> + mov eax,0x3fdeffdf
> + and DWORD[8+rdi],eax
> +$L$done:
> + shl r9,32
> + mov eax,r10d
> + mov rbx,r8
> +
> + or rax,r9
> + mov rdi,QWORD[8+rsp] ;WIN64 epilogue
> + mov rsi,QWORD[16+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +$L$SEH_end_OPENSSL_ia32_cpuid:
> +
> +global OPENSSL_cleanse
> +
> +ALIGN 16
> +OPENSSL_cleanse:
> +
> + xor rax,rax
> + cmp rdx,15
> + jae NEAR $L$ot
> + cmp rdx,0
> + je NEAR $L$ret
> +$L$ittle:
> + mov BYTE[rcx],al
> + sub rdx,1
> + lea rcx,[1+rcx]
> + jnz NEAR $L$ittle
> +$L$ret:
> + DB 0F3h,0C3h ;repret
> +ALIGN 16
> +$L$ot:
> + test rcx,7
> + jz NEAR $L$aligned
> + mov BYTE[rcx],al
> + lea rdx,[((-1))+rdx]
> + lea rcx,[1+rcx]
> + jmp NEAR $L$ot
> +$L$aligned:
> + mov QWORD[rcx],rax
> + lea rdx,[((-8))+rdx]
> + test rdx,-8
> + lea rcx,[8+rcx]
> + jnz NEAR $L$aligned
> + cmp rdx,0
> + jne NEAR $L$ittle
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +global CRYPTO_memcmp
> +
> +ALIGN 16
> +CRYPTO_memcmp:
> +
> + xor rax,rax
> + xor r10,r10
> + cmp r8,0
> + je NEAR $L$no_data
> + cmp r8,16
> + jne NEAR $L$oop_cmp
> + mov r10,QWORD[rcx]
> + mov r11,QWORD[8+rcx]
> + mov r8,1
> + xor r10,QWORD[rdx]
> + xor r11,QWORD[8+rdx]
> + or r10,r11
> + cmovnz rax,r8
> + DB 0F3h,0C3h ;repret
> +
> +ALIGN 16
> +$L$oop_cmp:
> + mov r10b,BYTE[rcx]
> + lea rcx,[1+rcx]
> + xor r10b,BYTE[rdx]
> + lea rdx,[1+rdx]
> + or al,r10b
> + dec r8
> + jnz NEAR $L$oop_cmp
> + neg rax
> + shr rax,63
> +$L$no_data:
> + DB 0F3h,0C3h ;repret
> +
> +
> +global OPENSSL_wipe_cpu
> +
> +ALIGN 16
> +OPENSSL_wipe_cpu:
> + pxor xmm0,xmm0
> + pxor xmm1,xmm1
> + pxor xmm2,xmm2
> + pxor xmm3,xmm3
> + pxor xmm4,xmm4
> + pxor xmm5,xmm5
> + xor rcx,rcx
> + xor rdx,rdx
> + xor r8,r8
> + xor r9,r9
> + xor r10,r10
> + xor r11,r11
> + lea rax,[8+rsp]
> + DB 0F3h,0C3h ;repret
> +
> +global OPENSSL_instrument_bus
> +
> +ALIGN 16
> +OPENSSL_instrument_bus:
> +
> + mov r10,rcx
> + mov rcx,rdx
> + mov r11,rdx
> +
> + rdtsc
> + mov r8d,eax
> + mov r9d,0
> + clflush [r10]
> +DB 0xf0
> + add DWORD[r10],r9d
> + jmp NEAR $L$oop
> +ALIGN 16
> +$L$oop: rdtsc
> + mov edx,eax
> + sub eax,r8d
> + mov r8d,edx
> + mov r9d,eax
> + clflush [r10]
> +DB 0xf0
> + add DWORD[r10],eax
> + lea r10,[4+r10]
> + sub rcx,1
> + jnz NEAR $L$oop
> +
> + mov rax,r11
> + DB 0F3h,0C3h ;repret
> +
> +
> +
> +global OPENSSL_instrument_bus2
> +
> +ALIGN 16
> +OPENSSL_instrument_bus2:
> +
> + mov r10,rcx
> + mov rcx,rdx
> + mov r11,r8
> + mov QWORD[8+rsp],rcx
> +
> + rdtsc
> + mov r8d,eax
> + mov r9d,0
> +
> + clflush [r10]
> +DB 0xf0
> + add DWORD[r10],r9d
> +
> + rdtsc
> + mov edx,eax
> + sub eax,r8d
> + mov r8d,edx
> + mov r9d,eax
> +$L$oop2:
> + clflush [r10]
> +DB 0xf0
> + add DWORD[r10],eax
> +
> + sub r11,1
> + jz NEAR $L$done2
> +
> + rdtsc
> + mov edx,eax
> + sub eax,r8d
> + mov r8d,edx
> + cmp eax,r9d
> + mov r9d,eax
> + mov edx,0
> + setne dl
> + sub rcx,rdx
> + lea r10,[rdx*4+r10]
> + jnz NEAR $L$oop2
> +
> +$L$done2:
> + mov rax,QWORD[8+rsp]
> + sub rax,rcx
> + DB 0F3h,0C3h ;repret
> +
> +
> +global OPENSSL_ia32_rdrand_bytes
> +
> +ALIGN 16
> +OPENSSL_ia32_rdrand_bytes:
> +
> + xor rax,rax
> + cmp rdx,0
> + je NEAR $L$done_rdrand_bytes
> +
> + mov r11,8
> +$L$oop_rdrand_bytes:
> +DB 73,15,199,242
> + jc NEAR $L$break_rdrand_bytes
> + dec r11
> + jnz NEAR $L$oop_rdrand_bytes
> + jmp NEAR $L$done_rdrand_bytes
> +
> +ALIGN 16
> +$L$break_rdrand_bytes:
> + cmp rdx,8
> + jb NEAR $L$tail_rdrand_bytes
> + mov QWORD[rcx],r10
> + lea rcx,[8+rcx]
> + add rax,8
> + sub rdx,8
> + jz NEAR $L$done_rdrand_bytes
> + mov r11,8
> + jmp NEAR $L$oop_rdrand_bytes
> +
> +ALIGN 16
> +$L$tail_rdrand_bytes:
> + mov BYTE[rcx],r10b
> + lea rcx,[1+rcx]
> + inc rax
> + shr r10,8
> + dec rdx
> + jnz NEAR $L$tail_rdrand_bytes
> +
> +$L$done_rdrand_bytes:
> + xor r10,r10
> + DB 0F3h,0C3h ;repret
> +
> +
> +global OPENSSL_ia32_rdseed_bytes
> +
> +ALIGN 16
> +OPENSSL_ia32_rdseed_bytes:
> +
> + xor rax,rax
> + cmp rdx,0
> + je NEAR $L$done_rdseed_bytes
> +
> + mov r11,8
> +$L$oop_rdseed_bytes:
> +DB 73,15,199,250
> + jc NEAR $L$break_rdseed_bytes
> + dec r11
> + jnz NEAR $L$oop_rdseed_bytes
> + jmp NEAR $L$done_rdseed_bytes
> +
> +ALIGN 16
> +$L$break_rdseed_bytes:
> + cmp rdx,8
> + jb NEAR $L$tail_rdseed_bytes
> + mov QWORD[rcx],r10
> + lea rcx,[8+rcx]
> + add rax,8
> + sub rdx,8
> + jz NEAR $L$done_rdseed_bytes
> + mov r11,8
> + jmp NEAR $L$oop_rdseed_bytes
> +
> +ALIGN 16
> +$L$tail_rdseed_bytes:
> + mov BYTE[rcx],r10b
> + lea rcx,[1+rcx]
> + inc rax
> + shr r10,8
> + dec rdx
> + jnz NEAR $L$tail_rdseed_bytes
> +
> +$L$done_rdseed_bytes:
> + xor r10,r10
> + DB 0F3h,0C3h ;repret
> +
> +
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-mb-
> x86_64.S b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-mb-
> x86_64.S
> new file mode 100644
> index 0000000000..7749fd685a
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-mb-x86_64.S
> @@ -0,0 +1,552 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/aes/asm/aesni-mb-x86_64.pl
> +#
> +# Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +
> +.globl aesni_multi_cbc_encrypt
> +.type aesni_multi_cbc_encrypt, at function
> +.align 32
> +aesni_multi_cbc_encrypt:
> +.cfi_startproc
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_offset %r14,-48
> + pushq %r15
> +.cfi_offset %r15,-56
> +
> +
> +
> +
> +
> +
> + subq $48,%rsp
> + andq $-64,%rsp
> + movq %rax,16(%rsp)
> +.cfi_escape 0x0f,0x05,0x77,0x10,0x06,0x23,0x08
> +
> +.Lenc4x_body:
> + movdqu (%rsi),%xmm12
> + leaq 120(%rsi),%rsi
> + leaq 80(%rdi),%rdi
> +
> +.Lenc4x_loop_grande:
> + movl %edx,24(%rsp)
> + xorl %edx,%edx
> + movl -64(%rdi),%ecx
> + movq -80(%rdi),%r8
> + cmpl %edx,%ecx
> + movq -72(%rdi),%r12
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movdqu -56(%rdi),%xmm2
> + movl %ecx,32(%rsp)
> + cmovleq %rsp,%r8
> + movl -24(%rdi),%ecx
> + movq -40(%rdi),%r9
> + cmpl %edx,%ecx
> + movq -32(%rdi),%r13
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movdqu -16(%rdi),%xmm3
> + movl %ecx,36(%rsp)
> + cmovleq %rsp,%r9
> + movl 16(%rdi),%ecx
> + movq 0(%rdi),%r10
> + cmpl %edx,%ecx
> + movq 8(%rdi),%r14
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movdqu 24(%rdi),%xmm4
> + movl %ecx,40(%rsp)
> + cmovleq %rsp,%r10
> + movl 56(%rdi),%ecx
> + movq 40(%rdi),%r11
> + cmpl %edx,%ecx
> + movq 48(%rdi),%r15
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movdqu 64(%rdi),%xmm5
> + movl %ecx,44(%rsp)
> + cmovleq %rsp,%r11
> + testl %edx,%edx
> + jz .Lenc4x_done
> +
> + movups 16-120(%rsi),%xmm1
> + pxor %xmm12,%xmm2
> + movups 32-120(%rsi),%xmm0
> + pxor %xmm12,%xmm3
> + movl 240-120(%rsi),%eax
> + pxor %xmm12,%xmm4
> + movdqu (%r8),%xmm6
> + pxor %xmm12,%xmm5
> + movdqu (%r9),%xmm7
> + pxor %xmm6,%xmm2
> + movdqu (%r10),%xmm8
> + pxor %xmm7,%xmm3
> + movdqu (%r11),%xmm9
> + pxor %xmm8,%xmm4
> + pxor %xmm9,%xmm5
> + movdqa 32(%rsp),%xmm10
> + xorq %rbx,%rbx
> + jmp .Loop_enc4x
> +
> +.align 32
> +.Loop_enc4x:
> + addq $16,%rbx
> + leaq 16(%rsp),%rbp
> + movl $1,%ecx
> + subq %rbx,%rbp
> +
> +.byte 102,15,56,220,209
> + prefetcht0 31(%r8,%rbx,1)
> + prefetcht0 31(%r9,%rbx,1)
> +.byte 102,15,56,220,217
> + prefetcht0 31(%r10,%rbx,1)
> + prefetcht0 31(%r10,%rbx,1)
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups 48-120(%rsi),%xmm1
> + cmpl 32(%rsp),%ecx
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> + cmovgeq %rbp,%r8
> + cmovgq %rbp,%r12
> +.byte 102,15,56,220,232
> + movups -56(%rsi),%xmm0
> + cmpl 36(%rsp),%ecx
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> + cmovgeq %rbp,%r9
> + cmovgq %rbp,%r13
> +.byte 102,15,56,220,233
> + movups -40(%rsi),%xmm1
> + cmpl 40(%rsp),%ecx
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> + cmovgeq %rbp,%r10
> + cmovgq %rbp,%r14
> +.byte 102,15,56,220,232
> + movups -24(%rsi),%xmm0
> + cmpl 44(%rsp),%ecx
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> + cmovgeq %rbp,%r11
> + cmovgq %rbp,%r15
> +.byte 102,15,56,220,233
> + movups -8(%rsi),%xmm1
> + movdqa %xmm10,%xmm11
> +.byte 102,15,56,220,208
> + prefetcht0 15(%r12,%rbx,1)
> + prefetcht0 15(%r13,%rbx,1)
> +.byte 102,15,56,220,216
> + prefetcht0 15(%r14,%rbx,1)
> + prefetcht0 15(%r15,%rbx,1)
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movups 128-120(%rsi),%xmm0
> + pxor %xmm12,%xmm12
> +
> +.byte 102,15,56,220,209
> + pcmpgtd %xmm12,%xmm11
> + movdqu -120(%rsi),%xmm12
> +.byte 102,15,56,220,217
> + paddd %xmm11,%xmm10
> + movdqa %xmm10,32(%rsp)
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups 144-120(%rsi),%xmm1
> +
> + cmpl $11,%eax
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movups 160-120(%rsi),%xmm0
> +
> + jb .Lenc4x_tail
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups 176-120(%rsi),%xmm1
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movups 192-120(%rsi),%xmm0
> +
> + je .Lenc4x_tail
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups 208-120(%rsi),%xmm1
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movups 224-120(%rsi),%xmm0
> + jmp .Lenc4x_tail
> +
> +.align 32
> +.Lenc4x_tail:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movdqu (%r8,%rbx,1),%xmm6
> + movdqu 16-120(%rsi),%xmm1
> +
> +.byte 102,15,56,221,208
> + movdqu (%r9,%rbx,1),%xmm7
> + pxor %xmm12,%xmm6
> +.byte 102,15,56,221,216
> + movdqu (%r10,%rbx,1),%xmm8
> + pxor %xmm12,%xmm7
> +.byte 102,15,56,221,224
> + movdqu (%r11,%rbx,1),%xmm9
> + pxor %xmm12,%xmm8
> +.byte 102,15,56,221,232
> + movdqu 32-120(%rsi),%xmm0
> + pxor %xmm12,%xmm9
> +
> + movups %xmm2,-16(%r12,%rbx,1)
> + pxor %xmm6,%xmm2
> + movups %xmm3,-16(%r13,%rbx,1)
> + pxor %xmm7,%xmm3
> + movups %xmm4,-16(%r14,%rbx,1)
> + pxor %xmm8,%xmm4
> + movups %xmm5,-16(%r15,%rbx,1)
> + pxor %xmm9,%xmm5
> +
> + decl %edx
> + jnz .Loop_enc4x
> +
> + movq 16(%rsp),%rax
> +.cfi_def_cfa %rax,8
> + movl 24(%rsp),%edx
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> + leaq 160(%rdi),%rdi
> + decl %edx
> + jnz .Lenc4x_loop_grande
> +
> +.Lenc4x_done:
> + movq -48(%rax),%r15
> +.cfi_restore %r15
> + movq -40(%rax),%r14
> +.cfi_restore %r14
> + movq -32(%rax),%r13
> +.cfi_restore %r13
> + movq -24(%rax),%r12
> +.cfi_restore %r12
> + movq -16(%rax),%rbp
> +.cfi_restore %rbp
> + movq -8(%rax),%rbx
> +.cfi_restore %rbx
> + leaq (%rax),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lenc4x_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_multi_cbc_encrypt,.-aesni_multi_cbc_encrypt
> +
> +.globl aesni_multi_cbc_decrypt
> +.type aesni_multi_cbc_decrypt, at function
> +.align 32
> +aesni_multi_cbc_decrypt:
> +.cfi_startproc
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_offset %r14,-48
> + pushq %r15
> +.cfi_offset %r15,-56
> +
> +
> +
> +
> +
> +
> + subq $48,%rsp
> + andq $-64,%rsp
> + movq %rax,16(%rsp)
> +.cfi_escape 0x0f,0x05,0x77,0x10,0x06,0x23,0x08
> +
> +.Ldec4x_body:
> + movdqu (%rsi),%xmm12
> + leaq 120(%rsi),%rsi
> + leaq 80(%rdi),%rdi
> +
> +.Ldec4x_loop_grande:
> + movl %edx,24(%rsp)
> + xorl %edx,%edx
> + movl -64(%rdi),%ecx
> + movq -80(%rdi),%r8
> + cmpl %edx,%ecx
> + movq -72(%rdi),%r12
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movdqu -56(%rdi),%xmm6
> + movl %ecx,32(%rsp)
> + cmovleq %rsp,%r8
> + movl -24(%rdi),%ecx
> + movq -40(%rdi),%r9
> + cmpl %edx,%ecx
> + movq -32(%rdi),%r13
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movdqu -16(%rdi),%xmm7
> + movl %ecx,36(%rsp)
> + cmovleq %rsp,%r9
> + movl 16(%rdi),%ecx
> + movq 0(%rdi),%r10
> + cmpl %edx,%ecx
> + movq 8(%rdi),%r14
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movdqu 24(%rdi),%xmm8
> + movl %ecx,40(%rsp)
> + cmovleq %rsp,%r10
> + movl 56(%rdi),%ecx
> + movq 40(%rdi),%r11
> + cmpl %edx,%ecx
> + movq 48(%rdi),%r15
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movdqu 64(%rdi),%xmm9
> + movl %ecx,44(%rsp)
> + cmovleq %rsp,%r11
> + testl %edx,%edx
> + jz .Ldec4x_done
> +
> + movups 16-120(%rsi),%xmm1
> + movups 32-120(%rsi),%xmm0
> + movl 240-120(%rsi),%eax
> + movdqu (%r8),%xmm2
> + movdqu (%r9),%xmm3
> + pxor %xmm12,%xmm2
> + movdqu (%r10),%xmm4
> + pxor %xmm12,%xmm3
> + movdqu (%r11),%xmm5
> + pxor %xmm12,%xmm4
> + pxor %xmm12,%xmm5
> + movdqa 32(%rsp),%xmm10
> + xorq %rbx,%rbx
> + jmp .Loop_dec4x
> +
> +.align 32
> +.Loop_dec4x:
> + addq $16,%rbx
> + leaq 16(%rsp),%rbp
> + movl $1,%ecx
> + subq %rbx,%rbp
> +
> +.byte 102,15,56,222,209
> + prefetcht0 31(%r8,%rbx,1)
> + prefetcht0 31(%r9,%rbx,1)
> +.byte 102,15,56,222,217
> + prefetcht0 31(%r10,%rbx,1)
> + prefetcht0 31(%r11,%rbx,1)
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movups 48-120(%rsi),%xmm1
> + cmpl 32(%rsp),%ecx
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> + cmovgeq %rbp,%r8
> + cmovgq %rbp,%r12
> +.byte 102,15,56,222,232
> + movups -56(%rsi),%xmm0
> + cmpl 36(%rsp),%ecx
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> + cmovgeq %rbp,%r9
> + cmovgq %rbp,%r13
> +.byte 102,15,56,222,233
> + movups -40(%rsi),%xmm1
> + cmpl 40(%rsp),%ecx
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> + cmovgeq %rbp,%r10
> + cmovgq %rbp,%r14
> +.byte 102,15,56,222,232
> + movups -24(%rsi),%xmm0
> + cmpl 44(%rsp),%ecx
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> + cmovgeq %rbp,%r11
> + cmovgq %rbp,%r15
> +.byte 102,15,56,222,233
> + movups -8(%rsi),%xmm1
> + movdqa %xmm10,%xmm11
> +.byte 102,15,56,222,208
> + prefetcht0 15(%r12,%rbx,1)
> + prefetcht0 15(%r13,%rbx,1)
> +.byte 102,15,56,222,216
> + prefetcht0 15(%r14,%rbx,1)
> + prefetcht0 15(%r15,%rbx,1)
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + movups 128-120(%rsi),%xmm0
> + pxor %xmm12,%xmm12
> +
> +.byte 102,15,56,222,209
> + pcmpgtd %xmm12,%xmm11
> + movdqu -120(%rsi),%xmm12
> +.byte 102,15,56,222,217
> + paddd %xmm11,%xmm10
> + movdqa %xmm10,32(%rsp)
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movups 144-120(%rsi),%xmm1
> +
> + cmpl $11,%eax
> +
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + movups 160-120(%rsi),%xmm0
> +
> + jb .Ldec4x_tail
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movups 176-120(%rsi),%xmm1
> +
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + movups 192-120(%rsi),%xmm0
> +
> + je .Ldec4x_tail
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movups 208-120(%rsi),%xmm1
> +
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + movups 224-120(%rsi),%xmm0
> + jmp .Ldec4x_tail
> +
> +.align 32
> +.Ldec4x_tail:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> + pxor %xmm0,%xmm6
> + pxor %xmm0,%xmm7
> +.byte 102,15,56,222,233
> + movdqu 16-120(%rsi),%xmm1
> + pxor %xmm0,%xmm8
> + pxor %xmm0,%xmm9
> + movdqu 32-120(%rsi),%xmm0
> +
> +.byte 102,15,56,223,214
> +.byte 102,15,56,223,223
> + movdqu -16(%r8,%rbx,1),%xmm6
> + movdqu -16(%r9,%rbx,1),%xmm7
> +.byte 102,65,15,56,223,224
> +.byte 102,65,15,56,223,233
> + movdqu -16(%r10,%rbx,1),%xmm8
> + movdqu -16(%r11,%rbx,1),%xmm9
> +
> + movups %xmm2,-16(%r12,%rbx,1)
> + movdqu (%r8,%rbx,1),%xmm2
> + movups %xmm3,-16(%r13,%rbx,1)
> + movdqu (%r9,%rbx,1),%xmm3
> + pxor %xmm12,%xmm2
> + movups %xmm4,-16(%r14,%rbx,1)
> + movdqu (%r10,%rbx,1),%xmm4
> + pxor %xmm12,%xmm3
> + movups %xmm5,-16(%r15,%rbx,1)
> + movdqu (%r11,%rbx,1),%xmm5
> + pxor %xmm12,%xmm4
> + pxor %xmm12,%xmm5
> +
> + decl %edx
> + jnz .Loop_dec4x
> +
> + movq 16(%rsp),%rax
> +.cfi_def_cfa %rax,8
> + movl 24(%rsp),%edx
> +
> + leaq 160(%rdi),%rdi
> + decl %edx
> + jnz .Ldec4x_loop_grande
> +
> +.Ldec4x_done:
> + movq -48(%rax),%r15
> +.cfi_restore %r15
> + movq -40(%rax),%r14
> +.cfi_restore %r14
> + movq -32(%rax),%r13
> +.cfi_restore %r13
> + movq -24(%rax),%r12
> +.cfi_restore %r12
> + movq -16(%rax),%rbp
> +.cfi_restore %rbp
> + movq -8(%rax),%rbx
> +.cfi_restore %rbx
> + leaq (%rax),%rsp
> +.cfi_def_cfa_register %rsp
> +.Ldec4x_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_multi_cbc_decrypt,.-aesni_multi_cbc_decrypt
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-sha1-
> x86_64.S b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-sha1-
> x86_64.S
> new file mode 100644
> index 0000000000..ab763a2eec
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-sha1-x86_64.S
> @@ -0,0 +1,1719 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/aes/asm/aesni-sha1-x86_64.pl
> +#
> +# Copyright 2011-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +.globl aesni_cbc_sha1_enc
> +.type aesni_cbc_sha1_enc, at function
> +.align 32
> +aesni_cbc_sha1_enc:
> +.cfi_startproc
> +
> + movl OPENSSL_ia32cap_P+0(%rip),%r10d
> + movq OPENSSL_ia32cap_P+4(%rip),%r11
> + btq $61,%r11
> + jc aesni_cbc_sha1_enc_shaext
> + jmp aesni_cbc_sha1_enc_ssse3
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_cbc_sha1_enc,.-aesni_cbc_sha1_enc
> +.type aesni_cbc_sha1_enc_ssse3, at function
> +.align 32
> +aesni_cbc_sha1_enc_ssse3:
> +.cfi_startproc
> + movq 8(%rsp),%r10
> +
> +
> + pushq %rbx
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r14,-48
> + pushq %r15
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r15,-56
> + leaq -104(%rsp),%rsp
> +.cfi_adjust_cfa_offset 104
> +
> +
> + movq %rdi,%r12
> + movq %rsi,%r13
> + movq %rdx,%r14
> + leaq 112(%rcx),%r15
> + movdqu (%r8),%xmm2
> + movq %r8,88(%rsp)
> + shlq $6,%r14
> + subq %r12,%r13
> + movl 240-112(%r15),%r8d
> + addq %r10,%r14
> +
> + leaq K_XX_XX(%rip),%r11
> + movl 0(%r9),%eax
> + movl 4(%r9),%ebx
> + movl 8(%r9),%ecx
> + movl 12(%r9),%edx
> + movl %ebx,%esi
> + movl 16(%r9),%ebp
> + movl %ecx,%edi
> + xorl %edx,%edi
> + andl %edi,%esi
> +
> + movdqa 64(%r11),%xmm3
> + movdqa 0(%r11),%xmm13
> + movdqu 0(%r10),%xmm4
> + movdqu 16(%r10),%xmm5
> + movdqu 32(%r10),%xmm6
> + movdqu 48(%r10),%xmm7
> +.byte 102,15,56,0,227
> +.byte 102,15,56,0,235
> +.byte 102,15,56,0,243
> + addq $64,%r10
> + paddd %xmm13,%xmm4
> +.byte 102,15,56,0,251
> + paddd %xmm13,%xmm5
> + paddd %xmm13,%xmm6
> + movdqa %xmm4,0(%rsp)
> + psubd %xmm13,%xmm4
> + movdqa %xmm5,16(%rsp)
> + psubd %xmm13,%xmm5
> + movdqa %xmm6,32(%rsp)
> + psubd %xmm13,%xmm6
> + movups -112(%r15),%xmm15
> + movups 16-112(%r15),%xmm0
> + jmp .Loop_ssse3
> +.align 32
> +.Loop_ssse3:
> + rorl $2,%ebx
> + movups 0(%r12),%xmm14
> + xorps %xmm15,%xmm14
> + xorps %xmm14,%xmm2
> + movups -80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + pshufd $238,%xmm4,%xmm8
> + xorl %edx,%esi
> + movdqa %xmm7,%xmm12
> + paddd %xmm7,%xmm13
> + movl %eax,%edi
> + addl 0(%rsp),%ebp
> + punpcklqdq %xmm5,%xmm8
> + xorl %ecx,%ebx
> + roll $5,%eax
> + addl %esi,%ebp
> + psrldq $4,%xmm12
> + andl %ebx,%edi
> + xorl %ecx,%ebx
> + pxor %xmm4,%xmm8
> + addl %eax,%ebp
> + rorl $7,%eax
> + pxor %xmm6,%xmm12
> + xorl %ecx,%edi
> + movl %ebp,%esi
> + addl 4(%rsp),%edx
> + pxor %xmm12,%xmm8
> + xorl %ebx,%eax
> + roll $5,%ebp
> + movdqa %xmm13,48(%rsp)
> + addl %edi,%edx
> + movups -64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + andl %eax,%esi
> + movdqa %xmm8,%xmm3
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + rorl $7,%ebp
> + movdqa %xmm8,%xmm12
> + xorl %ebx,%esi
> + pslldq $12,%xmm3
> + paddd %xmm8,%xmm8
> + movl %edx,%edi
> + addl 8(%rsp),%ecx
> + psrld $31,%xmm12
> + xorl %eax,%ebp
> + roll $5,%edx
> + addl %esi,%ecx
> + movdqa %xmm3,%xmm13
> + andl %ebp,%edi
> + xorl %eax,%ebp
> + psrld $30,%xmm3
> + addl %edx,%ecx
> + rorl $7,%edx
> + por %xmm12,%xmm8
> + xorl %eax,%edi
> + movl %ecx,%esi
> + addl 12(%rsp),%ebx
> + movups -48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + pslld $2,%xmm13
> + pxor %xmm3,%xmm8
> + xorl %ebp,%edx
> + movdqa 0(%r11),%xmm3
> + roll $5,%ecx
> + addl %edi,%ebx
> + andl %edx,%esi
> + pxor %xmm13,%xmm8
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + rorl $7,%ecx
> + pshufd $238,%xmm5,%xmm9
> + xorl %ebp,%esi
> + movdqa %xmm8,%xmm13
> + paddd %xmm8,%xmm3
> + movl %ebx,%edi
> + addl 16(%rsp),%eax
> + punpcklqdq %xmm6,%xmm9
> + xorl %edx,%ecx
> + roll $5,%ebx
> + addl %esi,%eax
> + psrldq $4,%xmm13
> + andl %ecx,%edi
> + xorl %edx,%ecx
> + pxor %xmm5,%xmm9
> + addl %ebx,%eax
> + rorl $7,%ebx
> + movups -32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + pxor %xmm7,%xmm13
> + xorl %edx,%edi
> + movl %eax,%esi
> + addl 20(%rsp),%ebp
> + pxor %xmm13,%xmm9
> + xorl %ecx,%ebx
> + roll $5,%eax
> + movdqa %xmm3,0(%rsp)
> + addl %edi,%ebp
> + andl %ebx,%esi
> + movdqa %xmm9,%xmm12
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + rorl $7,%eax
> + movdqa %xmm9,%xmm13
> + xorl %ecx,%esi
> + pslldq $12,%xmm12
> + paddd %xmm9,%xmm9
> + movl %ebp,%edi
> + addl 24(%rsp),%edx
> + psrld $31,%xmm13
> + xorl %ebx,%eax
> + roll $5,%ebp
> + addl %esi,%edx
> + movups -16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + movdqa %xmm12,%xmm3
> + andl %eax,%edi
> + xorl %ebx,%eax
> + psrld $30,%xmm12
> + addl %ebp,%edx
> + rorl $7,%ebp
> + por %xmm13,%xmm9
> + xorl %ebx,%edi
> + movl %edx,%esi
> + addl 28(%rsp),%ecx
> + pslld $2,%xmm3
> + pxor %xmm12,%xmm9
> + xorl %eax,%ebp
> + movdqa 16(%r11),%xmm12
> + roll $5,%edx
> + addl %edi,%ecx
> + andl %ebp,%esi
> + pxor %xmm3,%xmm9
> + xorl %eax,%ebp
> + addl %edx,%ecx
> + rorl $7,%edx
> + pshufd $238,%xmm6,%xmm10
> + xorl %eax,%esi
> + movdqa %xmm9,%xmm3
> + paddd %xmm9,%xmm12
> + movl %ecx,%edi
> + addl 32(%rsp),%ebx
> + movups 0(%r15),%xmm0
> +.byte 102,15,56,220,209
> + punpcklqdq %xmm7,%xmm10
> + xorl %ebp,%edx
> + roll $5,%ecx
> + addl %esi,%ebx
> + psrldq $4,%xmm3
> + andl %edx,%edi
> + xorl %ebp,%edx
> + pxor %xmm6,%xmm10
> + addl %ecx,%ebx
> + rorl $7,%ecx
> + pxor %xmm8,%xmm3
> + xorl %ebp,%edi
> + movl %ebx,%esi
> + addl 36(%rsp),%eax
> + pxor %xmm3,%xmm10
> + xorl %edx,%ecx
> + roll $5,%ebx
> + movdqa %xmm12,16(%rsp)
> + addl %edi,%eax
> + andl %ecx,%esi
> + movdqa %xmm10,%xmm13
> + xorl %edx,%ecx
> + addl %ebx,%eax
> + rorl $7,%ebx
> + movups 16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + movdqa %xmm10,%xmm3
> + xorl %edx,%esi
> + pslldq $12,%xmm13
> + paddd %xmm10,%xmm10
> + movl %eax,%edi
> + addl 40(%rsp),%ebp
> + psrld $31,%xmm3
> + xorl %ecx,%ebx
> + roll $5,%eax
> + addl %esi,%ebp
> + movdqa %xmm13,%xmm12
> + andl %ebx,%edi
> + xorl %ecx,%ebx
> + psrld $30,%xmm13
> + addl %eax,%ebp
> + rorl $7,%eax
> + por %xmm3,%xmm10
> + xorl %ecx,%edi
> + movl %ebp,%esi
> + addl 44(%rsp),%edx
> + pslld $2,%xmm12
> + pxor %xmm13,%xmm10
> + xorl %ebx,%eax
> + movdqa 16(%r11),%xmm13
> + roll $5,%ebp
> + addl %edi,%edx
> + movups 32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + andl %eax,%esi
> + pxor %xmm12,%xmm10
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + rorl $7,%ebp
> + pshufd $238,%xmm7,%xmm11
> + xorl %ebx,%esi
> + movdqa %xmm10,%xmm12
> + paddd %xmm10,%xmm13
> + movl %edx,%edi
> + addl 48(%rsp),%ecx
> + punpcklqdq %xmm8,%xmm11
> + xorl %eax,%ebp
> + roll $5,%edx
> + addl %esi,%ecx
> + psrldq $4,%xmm12
> + andl %ebp,%edi
> + xorl %eax,%ebp
> + pxor %xmm7,%xmm11
> + addl %edx,%ecx
> + rorl $7,%edx
> + pxor %xmm9,%xmm12
> + xorl %eax,%edi
> + movl %ecx,%esi
> + addl 52(%rsp),%ebx
> + movups 48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + pxor %xmm12,%xmm11
> + xorl %ebp,%edx
> + roll $5,%ecx
> + movdqa %xmm13,32(%rsp)
> + addl %edi,%ebx
> + andl %edx,%esi
> + movdqa %xmm11,%xmm3
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + rorl $7,%ecx
> + movdqa %xmm11,%xmm12
> + xorl %ebp,%esi
> + pslldq $12,%xmm3
> + paddd %xmm11,%xmm11
> + movl %ebx,%edi
> + addl 56(%rsp),%eax
> + psrld $31,%xmm12
> + xorl %edx,%ecx
> + roll $5,%ebx
> + addl %esi,%eax
> + movdqa %xmm3,%xmm13
> + andl %ecx,%edi
> + xorl %edx,%ecx
> + psrld $30,%xmm3
> + addl %ebx,%eax
> + rorl $7,%ebx
> + cmpl $11,%r8d
> + jb .Laesenclast1
> + movups 64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast1
> + movups 96(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%r15),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast1:
> +.byte 102,15,56,221,209
> + movups 16-112(%r15),%xmm0
> + por %xmm12,%xmm11
> + xorl %edx,%edi
> + movl %eax,%esi
> + addl 60(%rsp),%ebp
> + pslld $2,%xmm13
> + pxor %xmm3,%xmm11
> + xorl %ecx,%ebx
> + movdqa 16(%r11),%xmm3
> + roll $5,%eax
> + addl %edi,%ebp
> + andl %ebx,%esi
> + pxor %xmm13,%xmm11
> + pshufd $238,%xmm10,%xmm13
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + rorl $7,%eax
> + pxor %xmm8,%xmm4
> + xorl %ecx,%esi
> + movl %ebp,%edi
> + addl 0(%rsp),%edx
> + punpcklqdq %xmm11,%xmm13
> + xorl %ebx,%eax
> + roll $5,%ebp
> + pxor %xmm5,%xmm4
> + addl %esi,%edx
> + movups 16(%r12),%xmm14
> + xorps %xmm15,%xmm14
> + movups %xmm2,0(%r12,%r13,1)
> + xorps %xmm14,%xmm2
> + movups -80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + andl %eax,%edi
> + movdqa %xmm3,%xmm12
> + xorl %ebx,%eax
> + paddd %xmm11,%xmm3
> + addl %ebp,%edx
> + pxor %xmm13,%xmm4
> + rorl $7,%ebp
> + xorl %ebx,%edi
> + movl %edx,%esi
> + addl 4(%rsp),%ecx
> + movdqa %xmm4,%xmm13
> + xorl %eax,%ebp
> + roll $5,%edx
> + movdqa %xmm3,48(%rsp)
> + addl %edi,%ecx
> + andl %ebp,%esi
> + xorl %eax,%ebp
> + pslld $2,%xmm4
> + addl %edx,%ecx
> + rorl $7,%edx
> + psrld $30,%xmm13
> + xorl %eax,%esi
> + movl %ecx,%edi
> + addl 8(%rsp),%ebx
> + movups -64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + por %xmm13,%xmm4
> + xorl %ebp,%edx
> + roll $5,%ecx
> + pshufd $238,%xmm11,%xmm3
> + addl %esi,%ebx
> + andl %edx,%edi
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + addl 12(%rsp),%eax
> + xorl %ebp,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %edx,%esi
> + rorl $7,%ecx
> + addl %ebx,%eax
> + pxor %xmm9,%xmm5
> + addl 16(%rsp),%ebp
> + movups -48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %ecx,%esi
> + punpcklqdq %xmm4,%xmm3
> + movl %eax,%edi
> + roll $5,%eax
> + pxor %xmm6,%xmm5
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + movdqa %xmm12,%xmm13
> + rorl $7,%ebx
> + paddd %xmm4,%xmm12
> + addl %eax,%ebp
> + pxor %xmm3,%xmm5
> + addl 20(%rsp),%edx
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + movdqa %xmm5,%xmm3
> + addl %edi,%edx
> + xorl %ebx,%esi
> + movdqa %xmm12,0(%rsp)
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 24(%rsp),%ecx
> + pslld $2,%xmm5
> + xorl %eax,%esi
> + movl %edx,%edi
> + psrld $30,%xmm3
> + roll $5,%edx
> + addl %esi,%ecx
> + movups -32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%edi
> + rorl $7,%ebp
> + por %xmm3,%xmm5
> + addl %edx,%ecx
> + addl 28(%rsp),%ebx
> + pshufd $238,%xmm4,%xmm12
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + pxor %xmm10,%xmm6
> + addl 32(%rsp),%eax
> + xorl %edx,%esi
> + punpcklqdq %xmm5,%xmm12
> + movl %ebx,%edi
> + roll $5,%ebx
> + pxor %xmm7,%xmm6
> + addl %esi,%eax
> + xorl %edx,%edi
> + movdqa 32(%r11),%xmm3
> + rorl $7,%ecx
> + paddd %xmm5,%xmm13
> + addl %ebx,%eax
> + pxor %xmm12,%xmm6
> + addl 36(%rsp),%ebp
> + movups -16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + movdqa %xmm6,%xmm12
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + movdqa %xmm13,16(%rsp)
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 40(%rsp),%edx
> + pslld $2,%xmm6
> + xorl %ebx,%esi
> + movl %ebp,%edi
> + psrld $30,%xmm12
> + roll $5,%ebp
> + addl %esi,%edx
> + xorl %ebx,%edi
> + rorl $7,%eax
> + por %xmm12,%xmm6
> + addl %ebp,%edx
> + addl 44(%rsp),%ecx
> + pshufd $238,%xmm5,%xmm13
> + xorl %eax,%edi
> + movl %edx,%esi
> + roll $5,%edx
> + addl %edi,%ecx
> + movups 0(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%esi
> + rorl $7,%ebp
> + addl %edx,%ecx
> + pxor %xmm11,%xmm7
> + addl 48(%rsp),%ebx
> + xorl %ebp,%esi
> + punpcklqdq %xmm6,%xmm13
> + movl %ecx,%edi
> + roll $5,%ecx
> + pxor %xmm8,%xmm7
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + movdqa %xmm3,%xmm12
> + rorl $7,%edx
> + paddd %xmm6,%xmm3
> + addl %ecx,%ebx
> + pxor %xmm13,%xmm7
> + addl 52(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + movdqa %xmm7,%xmm13
> + addl %edi,%eax
> + xorl %edx,%esi
> + movdqa %xmm3,32(%rsp)
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 56(%rsp),%ebp
> + movups 16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + pslld $2,%xmm7
> + xorl %ecx,%esi
> + movl %eax,%edi
> + psrld $30,%xmm13
> + roll $5,%eax
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + rorl $7,%ebx
> + por %xmm13,%xmm7
> + addl %eax,%ebp
> + addl 60(%rsp),%edx
> + pshufd $238,%xmm6,%xmm3
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + addl %edi,%edx
> + xorl %ebx,%esi
> + rorl $7,%eax
> + addl %ebp,%edx
> + pxor %xmm4,%xmm8
> + addl 0(%rsp),%ecx
> + xorl %eax,%esi
> + punpcklqdq %xmm7,%xmm3
> + movl %edx,%edi
> + roll $5,%edx
> + pxor %xmm9,%xmm8
> + addl %esi,%ecx
> + movups 32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%edi
> + movdqa %xmm12,%xmm13
> + rorl $7,%ebp
> + paddd %xmm7,%xmm12
> + addl %edx,%ecx
> + pxor %xmm3,%xmm8
> + addl 4(%rsp),%ebx
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + roll $5,%ecx
> + movdqa %xmm8,%xmm3
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + movdqa %xmm12,48(%rsp)
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 8(%rsp),%eax
> + pslld $2,%xmm8
> + xorl %edx,%esi
> + movl %ebx,%edi
> + psrld $30,%xmm3
> + roll $5,%ebx
> + addl %esi,%eax
> + xorl %edx,%edi
> + rorl $7,%ecx
> + por %xmm3,%xmm8
> + addl %ebx,%eax
> + addl 12(%rsp),%ebp
> + movups 48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + pshufd $238,%xmm7,%xmm12
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + pxor %xmm5,%xmm9
> + addl 16(%rsp),%edx
> + xorl %ebx,%esi
> + punpcklqdq %xmm8,%xmm12
> + movl %ebp,%edi
> + roll $5,%ebp
> + pxor %xmm10,%xmm9
> + addl %esi,%edx
> + xorl %ebx,%edi
> + movdqa %xmm13,%xmm3
> + rorl $7,%eax
> + paddd %xmm8,%xmm13
> + addl %ebp,%edx
> + pxor %xmm12,%xmm9
> + addl 20(%rsp),%ecx
> + xorl %eax,%edi
> + movl %edx,%esi
> + roll $5,%edx
> + movdqa %xmm9,%xmm12
> + addl %edi,%ecx
> + cmpl $11,%r8d
> + jb .Laesenclast2
> + movups 64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast2
> + movups 96(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%r15),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast2:
> +.byte 102,15,56,221,209
> + movups 16-112(%r15),%xmm0
> + xorl %eax,%esi
> + movdqa %xmm13,0(%rsp)
> + rorl $7,%ebp
> + addl %edx,%ecx
> + addl 24(%rsp),%ebx
> + pslld $2,%xmm9
> + xorl %ebp,%esi
> + movl %ecx,%edi
> + psrld $30,%xmm12
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + por %xmm12,%xmm9
> + addl %ecx,%ebx
> + addl 28(%rsp),%eax
> + pshufd $238,%xmm8,%xmm13
> + rorl $7,%ecx
> + movl %ebx,%esi
> + xorl %edx,%edi
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %ecx,%esi
> + xorl %edx,%ecx
> + addl %ebx,%eax
> + pxor %xmm6,%xmm10
> + addl 32(%rsp),%ebp
> + movups 32(%r12),%xmm14
> + xorps %xmm15,%xmm14
> + movups %xmm2,16(%r13,%r12,1)
> + xorps %xmm14,%xmm2
> + movups -80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + andl %ecx,%esi
> + xorl %edx,%ecx
> + rorl $7,%ebx
> + punpcklqdq %xmm9,%xmm13
> + movl %eax,%edi
> + xorl %ecx,%esi
> + pxor %xmm11,%xmm10
> + roll $5,%eax
> + addl %esi,%ebp
> + movdqa %xmm3,%xmm12
> + xorl %ebx,%edi
> + paddd %xmm9,%xmm3
> + xorl %ecx,%ebx
> + pxor %xmm13,%xmm10
> + addl %eax,%ebp
> + addl 36(%rsp),%edx
> + andl %ebx,%edi
> + xorl %ecx,%ebx
> + rorl $7,%eax
> + movdqa %xmm10,%xmm13
> + movl %ebp,%esi
> + xorl %ebx,%edi
> + movdqa %xmm3,16(%rsp)
> + roll $5,%ebp
> + addl %edi,%edx
> + movups -64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%esi
> + pslld $2,%xmm10
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + psrld $30,%xmm13
> + addl 40(%rsp),%ecx
> + andl %eax,%esi
> + xorl %ebx,%eax
> + por %xmm13,%xmm10
> + rorl $7,%ebp
> + movl %edx,%edi
> + xorl %eax,%esi
> + roll $5,%edx
> + pshufd $238,%xmm9,%xmm3
> + addl %esi,%ecx
> + xorl %ebp,%edi
> + xorl %eax,%ebp
> + addl %edx,%ecx
> + addl 44(%rsp),%ebx
> + andl %ebp,%edi
> + xorl %eax,%ebp
> + rorl $7,%edx
> + movups -48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + movl %ecx,%esi
> + xorl %ebp,%edi
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %edx,%esi
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + pxor %xmm7,%xmm11
> + addl 48(%rsp),%eax
> + andl %edx,%esi
> + xorl %ebp,%edx
> + rorl $7,%ecx
> + punpcklqdq %xmm10,%xmm3
> + movl %ebx,%edi
> + xorl %edx,%esi
> + pxor %xmm4,%xmm11
> + roll $5,%ebx
> + addl %esi,%eax
> + movdqa 48(%r11),%xmm13
> + xorl %ecx,%edi
> + paddd %xmm10,%xmm12
> + xorl %edx,%ecx
> + pxor %xmm3,%xmm11
> + addl %ebx,%eax
> + addl 52(%rsp),%ebp
> + movups -32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + andl %ecx,%edi
> + xorl %edx,%ecx
> + rorl $7,%ebx
> + movdqa %xmm11,%xmm3
> + movl %eax,%esi
> + xorl %ecx,%edi
> + movdqa %xmm12,32(%rsp)
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ebx,%esi
> + pslld $2,%xmm11
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + psrld $30,%xmm3
> + addl 56(%rsp),%edx
> + andl %ebx,%esi
> + xorl %ecx,%ebx
> + por %xmm3,%xmm11
> + rorl $7,%eax
> + movl %ebp,%edi
> + xorl %ebx,%esi
> + roll $5,%ebp
> + pshufd $238,%xmm10,%xmm12
> + addl %esi,%edx
> + movups -16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %eax,%edi
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + addl 60(%rsp),%ecx
> + andl %eax,%edi
> + xorl %ebx,%eax
> + rorl $7,%ebp
> + movl %edx,%esi
> + xorl %eax,%edi
> + roll $5,%edx
> + addl %edi,%ecx
> + xorl %ebp,%esi
> + xorl %eax,%ebp
> + addl %edx,%ecx
> + pxor %xmm8,%xmm4
> + addl 0(%rsp),%ebx
> + andl %ebp,%esi
> + xorl %eax,%ebp
> + rorl $7,%edx
> + movups 0(%r15),%xmm0
> +.byte 102,15,56,220,209
> + punpcklqdq %xmm11,%xmm12
> + movl %ecx,%edi
> + xorl %ebp,%esi
> + pxor %xmm5,%xmm4
> + roll $5,%ecx
> + addl %esi,%ebx
> + movdqa %xmm13,%xmm3
> + xorl %edx,%edi
> + paddd %xmm11,%xmm13
> + xorl %ebp,%edx
> + pxor %xmm12,%xmm4
> + addl %ecx,%ebx
> + addl 4(%rsp),%eax
> + andl %edx,%edi
> + xorl %ebp,%edx
> + rorl $7,%ecx
> + movdqa %xmm4,%xmm12
> + movl %ebx,%esi
> + xorl %edx,%edi
> + movdqa %xmm13,48(%rsp)
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %ecx,%esi
> + pslld $2,%xmm4
> + xorl %edx,%ecx
> + addl %ebx,%eax
> + psrld $30,%xmm12
> + addl 8(%rsp),%ebp
> + movups 16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + andl %ecx,%esi
> + xorl %edx,%ecx
> + por %xmm12,%xmm4
> + rorl $7,%ebx
> + movl %eax,%edi
> + xorl %ecx,%esi
> + roll $5,%eax
> + pshufd $238,%xmm11,%xmm13
> + addl %esi,%ebp
> + xorl %ebx,%edi
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + addl 12(%rsp),%edx
> + andl %ebx,%edi
> + xorl %ecx,%ebx
> + rorl $7,%eax
> + movl %ebp,%esi
> + xorl %ebx,%edi
> + roll $5,%ebp
> + addl %edi,%edx
> + movups 32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%esi
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + pxor %xmm9,%xmm5
> + addl 16(%rsp),%ecx
> + andl %eax,%esi
> + xorl %ebx,%eax
> + rorl $7,%ebp
> + punpcklqdq %xmm4,%xmm13
> + movl %edx,%edi
> + xorl %eax,%esi
> + pxor %xmm6,%xmm5
> + roll $5,%edx
> + addl %esi,%ecx
> + movdqa %xmm3,%xmm12
> + xorl %ebp,%edi
> + paddd %xmm4,%xmm3
> + xorl %eax,%ebp
> + pxor %xmm13,%xmm5
> + addl %edx,%ecx
> + addl 20(%rsp),%ebx
> + andl %ebp,%edi
> + xorl %eax,%ebp
> + rorl $7,%edx
> + movups 48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + movdqa %xmm5,%xmm13
> + movl %ecx,%esi
> + xorl %ebp,%edi
> + movdqa %xmm3,0(%rsp)
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %edx,%esi
> + pslld $2,%xmm5
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + psrld $30,%xmm13
> + addl 24(%rsp),%eax
> + andl %edx,%esi
> + xorl %ebp,%edx
> + por %xmm13,%xmm5
> + rorl $7,%ecx
> + movl %ebx,%edi
> + xorl %edx,%esi
> + roll $5,%ebx
> + pshufd $238,%xmm4,%xmm3
> + addl %esi,%eax
> + xorl %ecx,%edi
> + xorl %edx,%ecx
> + addl %ebx,%eax
> + addl 28(%rsp),%ebp
> + cmpl $11,%r8d
> + jb .Laesenclast3
> + movups 64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast3
> + movups 96(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%r15),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast3:
> +.byte 102,15,56,221,209
> + movups 16-112(%r15),%xmm0
> + andl %ecx,%edi
> + xorl %edx,%ecx
> + rorl $7,%ebx
> + movl %eax,%esi
> + xorl %ecx,%edi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ebx,%esi
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + pxor %xmm10,%xmm6
> + addl 32(%rsp),%edx
> + andl %ebx,%esi
> + xorl %ecx,%ebx
> + rorl $7,%eax
> + punpcklqdq %xmm5,%xmm3
> + movl %ebp,%edi
> + xorl %ebx,%esi
> + pxor %xmm7,%xmm6
> + roll $5,%ebp
> + addl %esi,%edx
> + movups 48(%r12),%xmm14
> + xorps %xmm15,%xmm14
> + movups %xmm2,32(%r13,%r12,1)
> + xorps %xmm14,%xmm2
> + movups -80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + movdqa %xmm12,%xmm13
> + xorl %eax,%edi
> + paddd %xmm5,%xmm12
> + xorl %ebx,%eax
> + pxor %xmm3,%xmm6
> + addl %ebp,%edx
> + addl 36(%rsp),%ecx
> + andl %eax,%edi
> + xorl %ebx,%eax
> + rorl $7,%ebp
> + movdqa %xmm6,%xmm3
> + movl %edx,%esi
> + xorl %eax,%edi
> + movdqa %xmm12,16(%rsp)
> + roll $5,%edx
> + addl %edi,%ecx
> + xorl %ebp,%esi
> + pslld $2,%xmm6
> + xorl %eax,%ebp
> + addl %edx,%ecx
> + psrld $30,%xmm3
> + addl 40(%rsp),%ebx
> + andl %ebp,%esi
> + xorl %eax,%ebp
> + por %xmm3,%xmm6
> + rorl $7,%edx
> + movups -64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movl %ecx,%edi
> + xorl %ebp,%esi
> + roll $5,%ecx
> + pshufd $238,%xmm5,%xmm12
> + addl %esi,%ebx
> + xorl %edx,%edi
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + addl 44(%rsp),%eax
> + andl %edx,%edi
> + xorl %ebp,%edx
> + rorl $7,%ecx
> + movl %ebx,%esi
> + xorl %edx,%edi
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %edx,%esi
> + addl %ebx,%eax
> + pxor %xmm11,%xmm7
> + addl 48(%rsp),%ebp
> + movups -48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %ecx,%esi
> + punpcklqdq %xmm6,%xmm12
> + movl %eax,%edi
> + roll $5,%eax
> + pxor %xmm8,%xmm7
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + movdqa %xmm13,%xmm3
> + rorl $7,%ebx
> + paddd %xmm6,%xmm13
> + addl %eax,%ebp
> + pxor %xmm12,%xmm7
> + addl 52(%rsp),%edx
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + movdqa %xmm7,%xmm12
> + addl %edi,%edx
> + xorl %ebx,%esi
> + movdqa %xmm13,32(%rsp)
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 56(%rsp),%ecx
> + pslld $2,%xmm7
> + xorl %eax,%esi
> + movl %edx,%edi
> + psrld $30,%xmm12
> + roll $5,%edx
> + addl %esi,%ecx
> + movups -32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%edi
> + rorl $7,%ebp
> + por %xmm12,%xmm7
> + addl %edx,%ecx
> + addl 60(%rsp),%ebx
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 0(%rsp),%eax
> + xorl %edx,%esi
> + movl %ebx,%edi
> + roll $5,%ebx
> + paddd %xmm7,%xmm3
> + addl %esi,%eax
> + xorl %edx,%edi
> + movdqa %xmm3,48(%rsp)
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 4(%rsp),%ebp
> + movups -16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 8(%rsp),%edx
> + xorl %ebx,%esi
> + movl %ebp,%edi
> + roll $5,%ebp
> + addl %esi,%edx
> + xorl %ebx,%edi
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 12(%rsp),%ecx
> + xorl %eax,%edi
> + movl %edx,%esi
> + roll $5,%edx
> + addl %edi,%ecx
> + movups 0(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%esi
> + rorl $7,%ebp
> + addl %edx,%ecx
> + cmpq %r14,%r10
> + je .Ldone_ssse3
> + movdqa 64(%r11),%xmm3
> + movdqa 0(%r11),%xmm13
> + movdqu 0(%r10),%xmm4
> + movdqu 16(%r10),%xmm5
> + movdqu 32(%r10),%xmm6
> + movdqu 48(%r10),%xmm7
> +.byte 102,15,56,0,227
> + addq $64,%r10
> + addl 16(%rsp),%ebx
> + xorl %ebp,%esi
> + movl %ecx,%edi
> +.byte 102,15,56,0,235
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + paddd %xmm13,%xmm4
> + addl %ecx,%ebx
> + addl 20(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + movdqa %xmm4,0(%rsp)
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %edx,%esi
> + rorl $7,%ecx
> + psubd %xmm13,%xmm4
> + addl %ebx,%eax
> + addl 24(%rsp),%ebp
> + movups 16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %ecx,%esi
> + movl %eax,%edi
> + roll $5,%eax
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 28(%rsp),%edx
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + addl %edi,%edx
> + xorl %ebx,%esi
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 32(%rsp),%ecx
> + xorl %eax,%esi
> + movl %edx,%edi
> +.byte 102,15,56,0,243
> + roll $5,%edx
> + addl %esi,%ecx
> + movups 32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%edi
> + rorl $7,%ebp
> + paddd %xmm13,%xmm5
> + addl %edx,%ecx
> + addl 36(%rsp),%ebx
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + movdqa %xmm5,16(%rsp)
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + rorl $7,%edx
> + psubd %xmm13,%xmm5
> + addl %ecx,%ebx
> + addl 40(%rsp),%eax
> + xorl %edx,%esi
> + movl %ebx,%edi
> + roll $5,%ebx
> + addl %esi,%eax
> + xorl %edx,%edi
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 44(%rsp),%ebp
> + movups 48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 48(%rsp),%edx
> + xorl %ebx,%esi
> + movl %ebp,%edi
> +.byte 102,15,56,0,251
> + roll $5,%ebp
> + addl %esi,%edx
> + xorl %ebx,%edi
> + rorl $7,%eax
> + paddd %xmm13,%xmm6
> + addl %ebp,%edx
> + addl 52(%rsp),%ecx
> + xorl %eax,%edi
> + movl %edx,%esi
> + movdqa %xmm6,32(%rsp)
> + roll $5,%edx
> + addl %edi,%ecx
> + cmpl $11,%r8d
> + jb .Laesenclast4
> + movups 64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast4
> + movups 96(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%r15),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast4:
> +.byte 102,15,56,221,209
> + movups 16-112(%r15),%xmm0
> + xorl %eax,%esi
> + rorl $7,%ebp
> + psubd %xmm13,%xmm6
> + addl %edx,%ecx
> + addl 56(%rsp),%ebx
> + xorl %ebp,%esi
> + movl %ecx,%edi
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 60(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + addl %edi,%eax
> + rorl $7,%ecx
> + addl %ebx,%eax
> + movups %xmm2,48(%r13,%r12,1)
> + leaq 64(%r12),%r12
> +
> + addl 0(%r9),%eax
> + addl 4(%r9),%esi
> + addl 8(%r9),%ecx
> + addl 12(%r9),%edx
> + movl %eax,0(%r9)
> + addl 16(%r9),%ebp
> + movl %esi,4(%r9)
> + movl %esi,%ebx
> + movl %ecx,8(%r9)
> + movl %ecx,%edi
> + movl %edx,12(%r9)
> + xorl %edx,%edi
> + movl %ebp,16(%r9)
> + andl %edi,%esi
> + jmp .Loop_ssse3
> +
> +.Ldone_ssse3:
> + addl 16(%rsp),%ebx
> + xorl %ebp,%esi
> + movl %ecx,%edi
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 20(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %edx,%esi
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 24(%rsp),%ebp
> + movups 16(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %ecx,%esi
> + movl %eax,%edi
> + roll $5,%eax
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 28(%rsp),%edx
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + addl %edi,%edx
> + xorl %ebx,%esi
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 32(%rsp),%ecx
> + xorl %eax,%esi
> + movl %edx,%edi
> + roll $5,%edx
> + addl %esi,%ecx
> + movups 32(%r15),%xmm0
> +.byte 102,15,56,220,209
> + xorl %eax,%edi
> + rorl $7,%ebp
> + addl %edx,%ecx
> + addl 36(%rsp),%ebx
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 40(%rsp),%eax
> + xorl %edx,%esi
> + movl %ebx,%edi
> + roll $5,%ebx
> + addl %esi,%eax
> + xorl %edx,%edi
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 44(%rsp),%ebp
> + movups 48(%r15),%xmm1
> +.byte 102,15,56,220,208
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 48(%rsp),%edx
> + xorl %ebx,%esi
> + movl %ebp,%edi
> + roll $5,%ebp
> + addl %esi,%edx
> + xorl %ebx,%edi
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 52(%rsp),%ecx
> + xorl %eax,%edi
> + movl %edx,%esi
> + roll $5,%edx
> + addl %edi,%ecx
> + cmpl $11,%r8d
> + jb .Laesenclast5
> + movups 64(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%r15),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast5
> + movups 96(%r15),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%r15),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast5:
> +.byte 102,15,56,221,209
> + movups 16-112(%r15),%xmm0
> + xorl %eax,%esi
> + rorl $7,%ebp
> + addl %edx,%ecx
> + addl 56(%rsp),%ebx
> + xorl %ebp,%esi
> + movl %ecx,%edi
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 60(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + addl %edi,%eax
> + rorl $7,%ecx
> + addl %ebx,%eax
> + movups %xmm2,48(%r13,%r12,1)
> + movq 88(%rsp),%r8
> +
> + addl 0(%r9),%eax
> + addl 4(%r9),%esi
> + addl 8(%r9),%ecx
> + movl %eax,0(%r9)
> + addl 12(%r9),%edx
> + movl %esi,4(%r9)
> + addl 16(%r9),%ebp
> + movl %ecx,8(%r9)
> + movl %edx,12(%r9)
> + movl %ebp,16(%r9)
> + movups %xmm2,(%r8)
> + leaq 104(%rsp),%rsi
> +.cfi_def_cfa %rsi,56
> + movq 0(%rsi),%r15
> +.cfi_restore %r15
> + movq 8(%rsi),%r14
> +.cfi_restore %r14
> + movq 16(%rsi),%r13
> +.cfi_restore %r13
> + movq 24(%rsi),%r12
> +.cfi_restore %r12
> + movq 32(%rsi),%rbp
> +.cfi_restore %rbp
> + movq 40(%rsi),%rbx
> +.cfi_restore %rbx
> + leaq 48(%rsi),%rsp
> +.cfi_def_cfa %rsp,8
> +.Lepilogue_ssse3:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_cbc_sha1_enc_ssse3,.-aesni_cbc_sha1_enc_ssse3
> +.align 64
> +K_XX_XX:
> +.long 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> +.long 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> +.long 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> +.long 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.byte 0xf,0xe,0xd,0xc,0xb,0xa,0x9,0x8,0x7,0x6,0x5,0x4,0x3,0x2,0x1,0x0
> +
> +.byte
> 65,69,83,78,73,45,67,66,67,43,83,72,65,49,32,115,116,105,116,99,104,32,102,
> 111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,
> 32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,
> 0
> +.align 64
> +.type aesni_cbc_sha1_enc_shaext, at function
> +.align 32
> +aesni_cbc_sha1_enc_shaext:
> +.cfi_startproc
> + movq 8(%rsp),%r10
> + movdqu (%r9),%xmm8
> + movd 16(%r9),%xmm9
> + movdqa K_XX_XX+80(%rip),%xmm7
> +
> + movl 240(%rcx),%r11d
> + subq %rdi,%rsi
> + movups (%rcx),%xmm15
> + movups (%r8),%xmm2
> + movups 16(%rcx),%xmm0
> + leaq 112(%rcx),%rcx
> +
> + pshufd $27,%xmm8,%xmm8
> + pshufd $27,%xmm9,%xmm9
> + jmp .Loop_shaext
> +
> +.align 16
> +.Loop_shaext:
> + movups 0(%rdi),%xmm14
> + xorps %xmm15,%xmm14
> + xorps %xmm14,%xmm2
> + movups -80(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + movdqu (%r10),%xmm3
> + movdqa %xmm9,%xmm12
> +.byte 102,15,56,0,223
> + movdqu 16(%r10),%xmm4
> + movdqa %xmm8,%xmm11
> + movups -64(%rcx),%xmm0
> +.byte 102,15,56,220,209
> +.byte 102,15,56,0,231
> +
> + paddd %xmm3,%xmm9
> + movdqu 32(%r10),%xmm5
> + leaq 64(%r10),%r10
> + pxor %xmm12,%xmm3
> + movups -48(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + pxor %xmm12,%xmm3
> + movdqa %xmm8,%xmm10
> +.byte 102,15,56,0,239
> +.byte 69,15,58,204,193,0
> +.byte 68,15,56,200,212
> + movups -32(%rcx),%xmm0
> +.byte 102,15,56,220,209
> +.byte 15,56,201,220
> + movdqu -16(%r10),%xmm6
> + movdqa %xmm8,%xmm9
> +.byte 102,15,56,0,247
> + movups -16(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 69,15,58,204,194,0
> +.byte 68,15,56,200,205
> + pxor %xmm5,%xmm3
> +.byte 15,56,201,229
> + movups 0(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,0
> +.byte 68,15,56,200,214
> + movups 16(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,222
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> + movups 32(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,0
> +.byte 68,15,56,200,203
> + movups 48(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,227
> + pxor %xmm3,%xmm5
> +.byte 15,56,201,243
> + cmpl $11,%r11d
> + jb .Laesenclast6
> + movups 64(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast6
> + movups 96(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast6:
> +.byte 102,15,56,221,209
> + movups 16-112(%rcx),%xmm0
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,0
> +.byte 68,15,56,200,212
> + movups 16(%rdi),%xmm14
> + xorps %xmm15,%xmm14
> + movups %xmm2,0(%rsi,%rdi,1)
> + xorps %xmm14,%xmm2
> + movups -80(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,236
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,220
> + movups -64(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,1
> +.byte 68,15,56,200,205
> + movups -48(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,245
> + pxor %xmm5,%xmm3
> +.byte 15,56,201,229
> + movups -32(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,1
> +.byte 68,15,56,200,214
> + movups -16(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,222
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> + movups 0(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,1
> +.byte 68,15,56,200,203
> + movups 16(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,227
> + pxor %xmm3,%xmm5
> +.byte 15,56,201,243
> + movups 32(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,1
> +.byte 68,15,56,200,212
> + movups 48(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,236
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,220
> + cmpl $11,%r11d
> + jb .Laesenclast7
> + movups 64(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast7
> + movups 96(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast7:
> +.byte 102,15,56,221,209
> + movups 16-112(%rcx),%xmm0
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,1
> +.byte 68,15,56,200,205
> + movups 32(%rdi),%xmm14
> + xorps %xmm15,%xmm14
> + movups %xmm2,16(%rsi,%rdi,1)
> + xorps %xmm14,%xmm2
> + movups -80(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,245
> + pxor %xmm5,%xmm3
> +.byte 15,56,201,229
> + movups -64(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,2
> +.byte 68,15,56,200,214
> + movups -48(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,222
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> + movups -32(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,2
> +.byte 68,15,56,200,203
> + movups -16(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,227
> + pxor %xmm3,%xmm5
> +.byte 15,56,201,243
> + movups 0(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,2
> +.byte 68,15,56,200,212
> + movups 16(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,236
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,220
> + movups 32(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,2
> +.byte 68,15,56,200,205
> + movups 48(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,245
> + pxor %xmm5,%xmm3
> +.byte 15,56,201,229
> + cmpl $11,%r11d
> + jb .Laesenclast8
> + movups 64(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast8
> + movups 96(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast8:
> +.byte 102,15,56,221,209
> + movups 16-112(%rcx),%xmm0
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,2
> +.byte 68,15,56,200,214
> + movups 48(%rdi),%xmm14
> + xorps %xmm15,%xmm14
> + movups %xmm2,32(%rsi,%rdi,1)
> + xorps %xmm14,%xmm2
> + movups -80(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,222
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> + movups -64(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,3
> +.byte 68,15,56,200,203
> + movups -48(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.byte 15,56,202,227
> + pxor %xmm3,%xmm5
> +.byte 15,56,201,243
> + movups -32(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,3
> +.byte 68,15,56,200,212
> +.byte 15,56,202,236
> + pxor %xmm4,%xmm6
> + movups -16(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,3
> +.byte 68,15,56,200,205
> +.byte 15,56,202,245
> + movups 0(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movdqa %xmm12,%xmm5
> + movdqa %xmm8,%xmm10
> +.byte 69,15,58,204,193,3
> +.byte 68,15,56,200,214
> + movups 16(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + movdqa %xmm8,%xmm9
> +.byte 69,15,58,204,194,3
> +.byte 68,15,56,200,205
> + movups 32(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 48(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + cmpl $11,%r11d
> + jb .Laesenclast9
> + movups 64(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 80(%rcx),%xmm1
> +.byte 102,15,56,220,208
> + je .Laesenclast9
> + movups 96(%rcx),%xmm0
> +.byte 102,15,56,220,209
> + movups 112(%rcx),%xmm1
> +.byte 102,15,56,220,208
> +.Laesenclast9:
> +.byte 102,15,56,221,209
> + movups 16-112(%rcx),%xmm0
> + decq %rdx
> +
> + paddd %xmm11,%xmm8
> + movups %xmm2,48(%rsi,%rdi,1)
> + leaq 64(%rdi),%rdi
> + jnz .Loop_shaext
> +
> + pshufd $27,%xmm8,%xmm8
> + pshufd $27,%xmm9,%xmm9
> + movups %xmm2,(%r8)
> + movdqu %xmm8,(%r9)
> + movd %xmm9,16(%r9)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_cbc_sha1_enc_shaext,.-aesni_cbc_sha1_enc_shaext
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-sha256-
> x86_64.S b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-sha256-
> x86_64.S
> new file mode 100644
> index 0000000000..e257169287
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-sha256-x86_64.S
> @@ -0,0 +1,69 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/aes/asm/aesni-sha256-x86_64.pl
> +#
> +# Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +.globl aesni_cbc_sha256_enc
> +.type aesni_cbc_sha256_enc, at function
> +.align 16
> +aesni_cbc_sha256_enc:
> +.cfi_startproc
> + xorl %eax,%eax
> + cmpq $0,%rdi
> + je .Lprobe
> + ud2
> +.Lprobe:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_cbc_sha256_enc,.-aesni_cbc_sha256_enc
> +
> +.align 64
> +.type K256, at object
> +K256:
> +.long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> +.long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> +.long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> +.long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> +.long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> +.long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> +.long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> +.long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> +.long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> +.long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> +.long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> +.long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> +.long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> +.long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> +.long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> +.long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> +.long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> +.long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> +.long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> +.long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> +.long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> +.long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> +.long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> +.long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> +.long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> +.long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> +.long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> +.long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> +.long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> +.long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> +.long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> +.long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> +
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.long 0,0,0,0, 0,0,0,0, -1,-1,-1,-1
> +.long 0,0,0,0, 0,0,0,0
> +.byte
> 65,69,83,78,73,45,67,66,67,43,83,72,65,50,53,54,32,115,116,105,116,99,104,3
> 2,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,9
> 8,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,1
> 03,62,0
> +.align 64
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-x86_64.S
> b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-x86_64.S
> new file mode 100644
> index 0000000000..2bdb5cf251
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/aesni-x86_64.S
> @@ -0,0 +1,4484 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/aes/asm/aesni-x86_64.pl
> +#
> +# Copyright 2009-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +.globl aesni_encrypt
> +.type aesni_encrypt, at function
> +.align 16
> +aesni_encrypt:
> +.cfi_startproc
> + movups (%rdi),%xmm2
> + movl 240(%rdx),%eax
> + movups (%rdx),%xmm0
> + movups 16(%rdx),%xmm1
> + leaq 32(%rdx),%rdx
> + xorps %xmm0,%xmm2
> +.Loop_enc1_1:
> +.byte 102,15,56,220,209
> + decl %eax
> + movups (%rdx),%xmm1
> + leaq 16(%rdx),%rdx
> + jnz .Loop_enc1_1
> +.byte 102,15,56,221,209
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_encrypt,.-aesni_encrypt
> +
> +.globl aesni_decrypt
> +.type aesni_decrypt, at function
> +.align 16
> +aesni_decrypt:
> +.cfi_startproc
> + movups (%rdi),%xmm2
> + movl 240(%rdx),%eax
> + movups (%rdx),%xmm0
> + movups 16(%rdx),%xmm1
> + leaq 32(%rdx),%rdx
> + xorps %xmm0,%xmm2
> +.Loop_dec1_2:
> +.byte 102,15,56,222,209
> + decl %eax
> + movups (%rdx),%xmm1
> + leaq 16(%rdx),%rdx
> + jnz .Loop_dec1_2
> +.byte 102,15,56,223,209
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_decrypt, .-aesni_decrypt
> +.type _aesni_encrypt2, at function
> +.align 16
> +_aesni_encrypt2:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + xorps %xmm0,%xmm3
> + movups 32(%rcx),%xmm0
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> + addq $16,%rax
> +
> +.Lenc_loop2:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Lenc_loop2
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,221,208
> +.byte 102,15,56,221,216
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_encrypt2,.-_aesni_encrypt2
> +.type _aesni_decrypt2, at function
> +.align 16
> +_aesni_decrypt2:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + xorps %xmm0,%xmm3
> + movups 32(%rcx),%xmm0
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> + addq $16,%rax
> +
> +.Ldec_loop2:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Ldec_loop2
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,223,208
> +.byte 102,15,56,223,216
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_decrypt2,.-_aesni_decrypt2
> +.type _aesni_encrypt3, at function
> +.align 16
> +_aesni_encrypt3:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + xorps %xmm0,%xmm3
> + xorps %xmm0,%xmm4
> + movups 32(%rcx),%xmm0
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> + addq $16,%rax
> +
> +.Lenc_loop3:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Lenc_loop3
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,221,208
> +.byte 102,15,56,221,216
> +.byte 102,15,56,221,224
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_encrypt3,.-_aesni_encrypt3
> +.type _aesni_decrypt3, at function
> +.align 16
> +_aesni_decrypt3:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + xorps %xmm0,%xmm3
> + xorps %xmm0,%xmm4
> + movups 32(%rcx),%xmm0
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> + addq $16,%rax
> +
> +.Ldec_loop3:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Ldec_loop3
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,223,208
> +.byte 102,15,56,223,216
> +.byte 102,15,56,223,224
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_decrypt3,.-_aesni_decrypt3
> +.type _aesni_encrypt4, at function
> +.align 16
> +_aesni_encrypt4:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + xorps %xmm0,%xmm3
> + xorps %xmm0,%xmm4
> + xorps %xmm0,%xmm5
> + movups 32(%rcx),%xmm0
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> +.byte 0x0f,0x1f,0x00
> + addq $16,%rax
> +
> +.Lenc_loop4:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Lenc_loop4
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,221,208
> +.byte 102,15,56,221,216
> +.byte 102,15,56,221,224
> +.byte 102,15,56,221,232
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_encrypt4,.-_aesni_encrypt4
> +.type _aesni_decrypt4, at function
> +.align 16
> +_aesni_decrypt4:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + xorps %xmm0,%xmm3
> + xorps %xmm0,%xmm4
> + xorps %xmm0,%xmm5
> + movups 32(%rcx),%xmm0
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> +.byte 0x0f,0x1f,0x00
> + addq $16,%rax
> +
> +.Ldec_loop4:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Ldec_loop4
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,223,208
> +.byte 102,15,56,223,216
> +.byte 102,15,56,223,224
> +.byte 102,15,56,223,232
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_decrypt4,.-_aesni_decrypt4
> +.type _aesni_encrypt6, at function
> +.align 16
> +_aesni_encrypt6:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + pxor %xmm0,%xmm3
> + pxor %xmm0,%xmm4
> +.byte 102,15,56,220,209
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> +.byte 102,15,56,220,217
> + pxor %xmm0,%xmm5
> + pxor %xmm0,%xmm6
> +.byte 102,15,56,220,225
> + pxor %xmm0,%xmm7
> + movups (%rcx,%rax,1),%xmm0
> + addq $16,%rax
> + jmp .Lenc_loop6_enter
> +.align 16
> +.Lenc_loop6:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.Lenc_loop6_enter:
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Lenc_loop6
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,15,56,221,208
> +.byte 102,15,56,221,216
> +.byte 102,15,56,221,224
> +.byte 102,15,56,221,232
> +.byte 102,15,56,221,240
> +.byte 102,15,56,221,248
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_encrypt6,.-_aesni_encrypt6
> +.type _aesni_decrypt6, at function
> +.align 16
> +_aesni_decrypt6:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + pxor %xmm0,%xmm3
> + pxor %xmm0,%xmm4
> +.byte 102,15,56,222,209
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> +.byte 102,15,56,222,217
> + pxor %xmm0,%xmm5
> + pxor %xmm0,%xmm6
> +.byte 102,15,56,222,225
> + pxor %xmm0,%xmm7
> + movups (%rcx,%rax,1),%xmm0
> + addq $16,%rax
> + jmp .Ldec_loop6_enter
> +.align 16
> +.Ldec_loop6:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.Ldec_loop6_enter:
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Ldec_loop6
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,15,56,223,208
> +.byte 102,15,56,223,216
> +.byte 102,15,56,223,224
> +.byte 102,15,56,223,232
> +.byte 102,15,56,223,240
> +.byte 102,15,56,223,248
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_decrypt6,.-_aesni_decrypt6
> +.type _aesni_encrypt8, at function
> +.align 16
> +_aesni_encrypt8:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + xorps %xmm0,%xmm3
> + pxor %xmm0,%xmm4
> + pxor %xmm0,%xmm5
> + pxor %xmm0,%xmm6
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> +.byte 102,15,56,220,209
> + pxor %xmm0,%xmm7
> + pxor %xmm0,%xmm8
> +.byte 102,15,56,220,217
> + pxor %xmm0,%xmm9
> + movups (%rcx,%rax,1),%xmm0
> + addq $16,%rax
> + jmp .Lenc_loop8_inner
> +.align 16
> +.Lenc_loop8:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.Lenc_loop8_inner:
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> +.Lenc_loop8_enter:
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> +.byte 102,68,15,56,220,192
> +.byte 102,68,15,56,220,200
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Lenc_loop8
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> +.byte 102,15,56,221,208
> +.byte 102,15,56,221,216
> +.byte 102,15,56,221,224
> +.byte 102,15,56,221,232
> +.byte 102,15,56,221,240
> +.byte 102,15,56,221,248
> +.byte 102,68,15,56,221,192
> +.byte 102,68,15,56,221,200
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_encrypt8,.-_aesni_encrypt8
> +.type _aesni_decrypt8, at function
> +.align 16
> +_aesni_decrypt8:
> +.cfi_startproc
> + movups (%rcx),%xmm0
> + shll $4,%eax
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm2
> + xorps %xmm0,%xmm3
> + pxor %xmm0,%xmm4
> + pxor %xmm0,%xmm5
> + pxor %xmm0,%xmm6
> + leaq 32(%rcx,%rax,1),%rcx
> + negq %rax
> +.byte 102,15,56,222,209
> + pxor %xmm0,%xmm7
> + pxor %xmm0,%xmm8
> +.byte 102,15,56,222,217
> + pxor %xmm0,%xmm9
> + movups (%rcx,%rax,1),%xmm0
> + addq $16,%rax
> + jmp .Ldec_loop8_inner
> +.align 16
> +.Ldec_loop8:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.Ldec_loop8_inner:
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,68,15,56,222,193
> +.byte 102,68,15,56,222,201
> +.Ldec_loop8_enter:
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> +.byte 102,68,15,56,222,192
> +.byte 102,68,15,56,222,200
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Ldec_loop8
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,68,15,56,222,193
> +.byte 102,68,15,56,222,201
> +.byte 102,15,56,223,208
> +.byte 102,15,56,223,216
> +.byte 102,15,56,223,224
> +.byte 102,15,56,223,232
> +.byte 102,15,56,223,240
> +.byte 102,15,56,223,248
> +.byte 102,68,15,56,223,192
> +.byte 102,68,15,56,223,200
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _aesni_decrypt8,.-_aesni_decrypt8
> +.globl aesni_ecb_encrypt
> +.type aesni_ecb_encrypt, at function
> +.align 16
> +aesni_ecb_encrypt:
> +.cfi_startproc
> + andq $-16,%rdx
> + jz .Lecb_ret
> +
> + movl 240(%rcx),%eax
> + movups (%rcx),%xmm0
> + movq %rcx,%r11
> + movl %eax,%r10d
> + testl %r8d,%r8d
> + jz .Lecb_decrypt
> +
> + cmpq $0x80,%rdx
> + jb .Lecb_enc_tail
> +
> + movdqu (%rdi),%xmm2
> + movdqu 16(%rdi),%xmm3
> + movdqu 32(%rdi),%xmm4
> + movdqu 48(%rdi),%xmm5
> + movdqu 64(%rdi),%xmm6
> + movdqu 80(%rdi),%xmm7
> + movdqu 96(%rdi),%xmm8
> + movdqu 112(%rdi),%xmm9
> + leaq 128(%rdi),%rdi
> + subq $0x80,%rdx
> + jmp .Lecb_enc_loop8_enter
> +.align 16
> +.Lecb_enc_loop8:
> + movups %xmm2,(%rsi)
> + movq %r11,%rcx
> + movdqu (%rdi),%xmm2
> + movl %r10d,%eax
> + movups %xmm3,16(%rsi)
> + movdqu 16(%rdi),%xmm3
> + movups %xmm4,32(%rsi)
> + movdqu 32(%rdi),%xmm4
> + movups %xmm5,48(%rsi)
> + movdqu 48(%rdi),%xmm5
> + movups %xmm6,64(%rsi)
> + movdqu 64(%rdi),%xmm6
> + movups %xmm7,80(%rsi)
> + movdqu 80(%rdi),%xmm7
> + movups %xmm8,96(%rsi)
> + movdqu 96(%rdi),%xmm8
> + movups %xmm9,112(%rsi)
> + leaq 128(%rsi),%rsi
> + movdqu 112(%rdi),%xmm9
> + leaq 128(%rdi),%rdi
> +.Lecb_enc_loop8_enter:
> +
> + call _aesni_encrypt8
> +
> + subq $0x80,%rdx
> + jnc .Lecb_enc_loop8
> +
> + movups %xmm2,(%rsi)
> + movq %r11,%rcx
> + movups %xmm3,16(%rsi)
> + movl %r10d,%eax
> + movups %xmm4,32(%rsi)
> + movups %xmm5,48(%rsi)
> + movups %xmm6,64(%rsi)
> + movups %xmm7,80(%rsi)
> + movups %xmm8,96(%rsi)
> + movups %xmm9,112(%rsi)
> + leaq 128(%rsi),%rsi
> + addq $0x80,%rdx
> + jz .Lecb_ret
> +
> +.Lecb_enc_tail:
> + movups (%rdi),%xmm2
> + cmpq $0x20,%rdx
> + jb .Lecb_enc_one
> + movups 16(%rdi),%xmm3
> + je .Lecb_enc_two
> + movups 32(%rdi),%xmm4
> + cmpq $0x40,%rdx
> + jb .Lecb_enc_three
> + movups 48(%rdi),%xmm5
> + je .Lecb_enc_four
> + movups 64(%rdi),%xmm6
> + cmpq $0x60,%rdx
> + jb .Lecb_enc_five
> + movups 80(%rdi),%xmm7
> + je .Lecb_enc_six
> + movdqu 96(%rdi),%xmm8
> + xorps %xmm9,%xmm9
> + call _aesni_encrypt8
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + movups %xmm5,48(%rsi)
> + movups %xmm6,64(%rsi)
> + movups %xmm7,80(%rsi)
> + movups %xmm8,96(%rsi)
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_enc_one:
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_enc1_3:
> +.byte 102,15,56,220,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_enc1_3
> +.byte 102,15,56,221,209
> + movups %xmm2,(%rsi)
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_enc_two:
> + call _aesni_encrypt2
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_enc_three:
> + call _aesni_encrypt3
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_enc_four:
> + call _aesni_encrypt4
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + movups %xmm5,48(%rsi)
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_enc_five:
> + xorps %xmm7,%xmm7
> + call _aesni_encrypt6
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + movups %xmm5,48(%rsi)
> + movups %xmm6,64(%rsi)
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_enc_six:
> + call _aesni_encrypt6
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + movups %xmm5,48(%rsi)
> + movups %xmm6,64(%rsi)
> + movups %xmm7,80(%rsi)
> + jmp .Lecb_ret
> +
> +.align 16
> +.Lecb_decrypt:
> + cmpq $0x80,%rdx
> + jb .Lecb_dec_tail
> +
> + movdqu (%rdi),%xmm2
> + movdqu 16(%rdi),%xmm3
> + movdqu 32(%rdi),%xmm4
> + movdqu 48(%rdi),%xmm5
> + movdqu 64(%rdi),%xmm6
> + movdqu 80(%rdi),%xmm7
> + movdqu 96(%rdi),%xmm8
> + movdqu 112(%rdi),%xmm9
> + leaq 128(%rdi),%rdi
> + subq $0x80,%rdx
> + jmp .Lecb_dec_loop8_enter
> +.align 16
> +.Lecb_dec_loop8:
> + movups %xmm2,(%rsi)
> + movq %r11,%rcx
> + movdqu (%rdi),%xmm2
> + movl %r10d,%eax
> + movups %xmm3,16(%rsi)
> + movdqu 16(%rdi),%xmm3
> + movups %xmm4,32(%rsi)
> + movdqu 32(%rdi),%xmm4
> + movups %xmm5,48(%rsi)
> + movdqu 48(%rdi),%xmm5
> + movups %xmm6,64(%rsi)
> + movdqu 64(%rdi),%xmm6
> + movups %xmm7,80(%rsi)
> + movdqu 80(%rdi),%xmm7
> + movups %xmm8,96(%rsi)
> + movdqu 96(%rdi),%xmm8
> + movups %xmm9,112(%rsi)
> + leaq 128(%rsi),%rsi
> + movdqu 112(%rdi),%xmm9
> + leaq 128(%rdi),%rdi
> +.Lecb_dec_loop8_enter:
> +
> + call _aesni_decrypt8
> +
> + movups (%r11),%xmm0
> + subq $0x80,%rdx
> + jnc .Lecb_dec_loop8
> +
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + movq %r11,%rcx
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + movl %r10d,%eax
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + movups %xmm5,48(%rsi)
> + pxor %xmm5,%xmm5
> + movups %xmm6,64(%rsi)
> + pxor %xmm6,%xmm6
> + movups %xmm7,80(%rsi)
> + pxor %xmm7,%xmm7
> + movups %xmm8,96(%rsi)
> + pxor %xmm8,%xmm8
> + movups %xmm9,112(%rsi)
> + pxor %xmm9,%xmm9
> + leaq 128(%rsi),%rsi
> + addq $0x80,%rdx
> + jz .Lecb_ret
> +
> +.Lecb_dec_tail:
> + movups (%rdi),%xmm2
> + cmpq $0x20,%rdx
> + jb .Lecb_dec_one
> + movups 16(%rdi),%xmm3
> + je .Lecb_dec_two
> + movups 32(%rdi),%xmm4
> + cmpq $0x40,%rdx
> + jb .Lecb_dec_three
> + movups 48(%rdi),%xmm5
> + je .Lecb_dec_four
> + movups 64(%rdi),%xmm6
> + cmpq $0x60,%rdx
> + jb .Lecb_dec_five
> + movups 80(%rdi),%xmm7
> + je .Lecb_dec_six
> + movups 96(%rdi),%xmm8
> + movups (%rcx),%xmm0
> + xorps %xmm9,%xmm9
> + call _aesni_decrypt8
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + movups %xmm5,48(%rsi)
> + pxor %xmm5,%xmm5
> + movups %xmm6,64(%rsi)
> + pxor %xmm6,%xmm6
> + movups %xmm7,80(%rsi)
> + pxor %xmm7,%xmm7
> + movups %xmm8,96(%rsi)
> + pxor %xmm8,%xmm8
> + pxor %xmm9,%xmm9
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_dec_one:
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_dec1_4:
> +.byte 102,15,56,222,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_dec1_4
> +.byte 102,15,56,223,209
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_dec_two:
> + call _aesni_decrypt2
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_dec_three:
> + call _aesni_decrypt3
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_dec_four:
> + call _aesni_decrypt4
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + movups %xmm5,48(%rsi)
> + pxor %xmm5,%xmm5
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_dec_five:
> + xorps %xmm7,%xmm7
> + call _aesni_decrypt6
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + movups %xmm5,48(%rsi)
> + pxor %xmm5,%xmm5
> + movups %xmm6,64(%rsi)
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + jmp .Lecb_ret
> +.align 16
> +.Lecb_dec_six:
> + call _aesni_decrypt6
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + movups %xmm5,48(%rsi)
> + pxor %xmm5,%xmm5
> + movups %xmm6,64(%rsi)
> + pxor %xmm6,%xmm6
> + movups %xmm7,80(%rsi)
> + pxor %xmm7,%xmm7
> +
> +.Lecb_ret:
> + xorps %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_ecb_encrypt,.-aesni_ecb_encrypt
> +.globl aesni_ccm64_encrypt_blocks
> +.type aesni_ccm64_encrypt_blocks, at function
> +.align 16
> +aesni_ccm64_encrypt_blocks:
> +.cfi_startproc
> + movl 240(%rcx),%eax
> + movdqu (%r8),%xmm6
> + movdqa .Lincrement64(%rip),%xmm9
> + movdqa .Lbswap_mask(%rip),%xmm7
> +
> + shll $4,%eax
> + movl $16,%r10d
> + leaq 0(%rcx),%r11
> + movdqu (%r9),%xmm3
> + movdqa %xmm6,%xmm2
> + leaq 32(%rcx,%rax,1),%rcx
> +.byte 102,15,56,0,247
> + subq %rax,%r10
> + jmp .Lccm64_enc_outer
> +.align 16
> +.Lccm64_enc_outer:
> + movups (%r11),%xmm0
> + movq %r10,%rax
> + movups (%rdi),%xmm8
> +
> + xorps %xmm0,%xmm2
> + movups 16(%r11),%xmm1
> + xorps %xmm8,%xmm0
> + xorps %xmm0,%xmm3
> + movups 32(%r11),%xmm0
> +
> +.Lccm64_enc2_loop:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Lccm64_enc2_loop
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> + paddq %xmm9,%xmm6
> + decq %rdx
> +.byte 102,15,56,221,208
> +.byte 102,15,56,221,216
> +
> + leaq 16(%rdi),%rdi
> + xorps %xmm2,%xmm8
> + movdqa %xmm6,%xmm2
> + movups %xmm8,(%rsi)
> +.byte 102,15,56,0,215
> + leaq 16(%rsi),%rsi
> + jnz .Lccm64_enc_outer
> +
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + movups %xmm3,(%r9)
> + pxor %xmm3,%xmm3
> + pxor %xmm8,%xmm8
> + pxor %xmm6,%xmm6
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_ccm64_encrypt_blocks,.-aesni_ccm64_encrypt_blocks
> +.globl aesni_ccm64_decrypt_blocks
> +.type aesni_ccm64_decrypt_blocks, at function
> +.align 16
> +aesni_ccm64_decrypt_blocks:
> +.cfi_startproc
> + movl 240(%rcx),%eax
> + movups (%r8),%xmm6
> + movdqu (%r9),%xmm3
> + movdqa .Lincrement64(%rip),%xmm9
> + movdqa .Lbswap_mask(%rip),%xmm7
> +
> + movaps %xmm6,%xmm2
> + movl %eax,%r10d
> + movq %rcx,%r11
> +.byte 102,15,56,0,247
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_enc1_5:
> +.byte 102,15,56,220,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_enc1_5
> +.byte 102,15,56,221,209
> + shll $4,%r10d
> + movl $16,%eax
> + movups (%rdi),%xmm8
> + paddq %xmm9,%xmm6
> + leaq 16(%rdi),%rdi
> + subq %r10,%rax
> + leaq 32(%r11,%r10,1),%rcx
> + movq %rax,%r10
> + jmp .Lccm64_dec_outer
> +.align 16
> +.Lccm64_dec_outer:
> + xorps %xmm2,%xmm8
> + movdqa %xmm6,%xmm2
> + movups %xmm8,(%rsi)
> + leaq 16(%rsi),%rsi
> +.byte 102,15,56,0,215
> +
> + subq $1,%rdx
> + jz .Lccm64_dec_break
> +
> + movups (%r11),%xmm0
> + movq %r10,%rax
> + movups 16(%r11),%xmm1
> + xorps %xmm0,%xmm8
> + xorps %xmm0,%xmm2
> + xorps %xmm8,%xmm3
> + movups 32(%r11),%xmm0
> + jmp .Lccm64_dec2_loop
> +.align 16
> +.Lccm64_dec2_loop:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Lccm64_dec2_loop
> + movups (%rdi),%xmm8
> + paddq %xmm9,%xmm6
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,221,208
> +.byte 102,15,56,221,216
> + leaq 16(%rdi),%rdi
> + jmp .Lccm64_dec_outer
> +
> +.align 16
> +.Lccm64_dec_break:
> +
> + movl 240(%r11),%eax
> + movups (%r11),%xmm0
> + movups 16(%r11),%xmm1
> + xorps %xmm0,%xmm8
> + leaq 32(%r11),%r11
> + xorps %xmm8,%xmm3
> +.Loop_enc1_6:
> +.byte 102,15,56,220,217
> + decl %eax
> + movups (%r11),%xmm1
> + leaq 16(%r11),%r11
> + jnz .Loop_enc1_6
> +.byte 102,15,56,221,217
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + movups %xmm3,(%r9)
> + pxor %xmm3,%xmm3
> + pxor %xmm8,%xmm8
> + pxor %xmm6,%xmm6
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_ccm64_decrypt_blocks,.-aesni_ccm64_decrypt_blocks
> +.globl aesni_ctr32_encrypt_blocks
> +.type aesni_ctr32_encrypt_blocks, at function
> +.align 16
> +aesni_ctr32_encrypt_blocks:
> +.cfi_startproc
> + cmpq $1,%rdx
> + jne .Lctr32_bulk
> +
> +
> +
> + movups (%r8),%xmm2
> + movups (%rdi),%xmm3
> + movl 240(%rcx),%edx
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_enc1_7:
> +.byte 102,15,56,220,209
> + decl %edx
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_enc1_7
> +.byte 102,15,56,221,209
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + xorps %xmm3,%xmm2
> + pxor %xmm3,%xmm3
> + movups %xmm2,(%rsi)
> + xorps %xmm2,%xmm2
> + jmp .Lctr32_epilogue
> +
> +.align 16
> +.Lctr32_bulk:
> + leaq (%rsp),%r11
> +.cfi_def_cfa_register %r11
> + pushq %rbp
> +.cfi_offset %rbp,-16
> + subq $128,%rsp
> + andq $-16,%rsp
> +
> +
> +
> +
> + movdqu (%r8),%xmm2
> + movdqu (%rcx),%xmm0
> + movl 12(%r8),%r8d
> + pxor %xmm0,%xmm2
> + movl 12(%rcx),%ebp
> + movdqa %xmm2,0(%rsp)
> + bswapl %r8d
> + movdqa %xmm2,%xmm3
> + movdqa %xmm2,%xmm4
> + movdqa %xmm2,%xmm5
> + movdqa %xmm2,64(%rsp)
> + movdqa %xmm2,80(%rsp)
> + movdqa %xmm2,96(%rsp)
> + movq %rdx,%r10
> + movdqa %xmm2,112(%rsp)
> +
> + leaq 1(%r8),%rax
> + leaq 2(%r8),%rdx
> + bswapl %eax
> + bswapl %edx
> + xorl %ebp,%eax
> + xorl %ebp,%edx
> +.byte 102,15,58,34,216,3
> + leaq 3(%r8),%rax
> + movdqa %xmm3,16(%rsp)
> +.byte 102,15,58,34,226,3
> + bswapl %eax
> + movq %r10,%rdx
> + leaq 4(%r8),%r10
> + movdqa %xmm4,32(%rsp)
> + xorl %ebp,%eax
> + bswapl %r10d
> +.byte 102,15,58,34,232,3
> + xorl %ebp,%r10d
> + movdqa %xmm5,48(%rsp)
> + leaq 5(%r8),%r9
> + movl %r10d,64+12(%rsp)
> + bswapl %r9d
> + leaq 6(%r8),%r10
> + movl 240(%rcx),%eax
> + xorl %ebp,%r9d
> + bswapl %r10d
> + movl %r9d,80+12(%rsp)
> + xorl %ebp,%r10d
> + leaq 7(%r8),%r9
> + movl %r10d,96+12(%rsp)
> + bswapl %r9d
> + movl OPENSSL_ia32cap_P+4(%rip),%r10d
> + xorl %ebp,%r9d
> + andl $71303168,%r10d
> + movl %r9d,112+12(%rsp)
> +
> + movups 16(%rcx),%xmm1
> +
> + movdqa 64(%rsp),%xmm6
> + movdqa 80(%rsp),%xmm7
> +
> + cmpq $8,%rdx
> + jb .Lctr32_tail
> +
> + subq $6,%rdx
> + cmpl $4194304,%r10d
> + je .Lctr32_6x
> +
> + leaq 128(%rcx),%rcx
> + subq $2,%rdx
> + jmp .Lctr32_loop8
> +
> +.align 16
> +.Lctr32_6x:
> + shll $4,%eax
> + movl $48,%r10d
> + bswapl %ebp
> + leaq 32(%rcx,%rax,1),%rcx
> + subq %rax,%r10
> + jmp .Lctr32_loop6
> +
> +.align 16
> +.Lctr32_loop6:
> + addl $6,%r8d
> + movups -48(%rcx,%r10,1),%xmm0
> +.byte 102,15,56,220,209
> + movl %r8d,%eax
> + xorl %ebp,%eax
> +.byte 102,15,56,220,217
> +.byte 0x0f,0x38,0xf1,0x44,0x24,12
> + leal 1(%r8),%eax
> +.byte 102,15,56,220,225
> + xorl %ebp,%eax
> +.byte 0x0f,0x38,0xf1,0x44,0x24,28
> +.byte 102,15,56,220,233
> + leal 2(%r8),%eax
> + xorl %ebp,%eax
> +.byte 102,15,56,220,241
> +.byte 0x0f,0x38,0xf1,0x44,0x24,44
> + leal 3(%r8),%eax
> +.byte 102,15,56,220,249
> + movups -32(%rcx,%r10,1),%xmm1
> + xorl %ebp,%eax
> +
> +.byte 102,15,56,220,208
> +.byte 0x0f,0x38,0xf1,0x44,0x24,60
> + leal 4(%r8),%eax
> +.byte 102,15,56,220,216
> + xorl %ebp,%eax
> +.byte 0x0f,0x38,0xf1,0x44,0x24,76
> +.byte 102,15,56,220,224
> + leal 5(%r8),%eax
> + xorl %ebp,%eax
> +.byte 102,15,56,220,232
> +.byte 0x0f,0x38,0xf1,0x44,0x24,92
> + movq %r10,%rax
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> + movups -16(%rcx,%r10,1),%xmm0
> +
> + call .Lenc_loop6
> +
> + movdqu (%rdi),%xmm8
> + movdqu 16(%rdi),%xmm9
> + movdqu 32(%rdi),%xmm10
> + movdqu 48(%rdi),%xmm11
> + movdqu 64(%rdi),%xmm12
> + movdqu 80(%rdi),%xmm13
> + leaq 96(%rdi),%rdi
> + movups -64(%rcx,%r10,1),%xmm1
> + pxor %xmm2,%xmm8
> + movaps 0(%rsp),%xmm2
> + pxor %xmm3,%xmm9
> + movaps 16(%rsp),%xmm3
> + pxor %xmm4,%xmm10
> + movaps 32(%rsp),%xmm4
> + pxor %xmm5,%xmm11
> + movaps 48(%rsp),%xmm5
> + pxor %xmm6,%xmm12
> + movaps 64(%rsp),%xmm6
> + pxor %xmm7,%xmm13
> + movaps 80(%rsp),%xmm7
> + movdqu %xmm8,(%rsi)
> + movdqu %xmm9,16(%rsi)
> + movdqu %xmm10,32(%rsi)
> + movdqu %xmm11,48(%rsi)
> + movdqu %xmm12,64(%rsi)
> + movdqu %xmm13,80(%rsi)
> + leaq 96(%rsi),%rsi
> +
> + subq $6,%rdx
> + jnc .Lctr32_loop6
> +
> + addq $6,%rdx
> + jz .Lctr32_done
> +
> + leal -48(%r10),%eax
> + leaq -80(%rcx,%r10,1),%rcx
> + negl %eax
> + shrl $4,%eax
> + jmp .Lctr32_tail
> +
> +.align 32
> +.Lctr32_loop8:
> + addl $8,%r8d
> + movdqa 96(%rsp),%xmm8
> +.byte 102,15,56,220,209
> + movl %r8d,%r9d
> + movdqa 112(%rsp),%xmm9
> +.byte 102,15,56,220,217
> + bswapl %r9d
> + movups 32-128(%rcx),%xmm0
> +.byte 102,15,56,220,225
> + xorl %ebp,%r9d
> + nop
> +.byte 102,15,56,220,233
> + movl %r9d,0+12(%rsp)
> + leaq 1(%r8),%r9
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> + movups 48-128(%rcx),%xmm1
> + bswapl %r9d
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> + xorl %ebp,%r9d
> +.byte 0x66,0x90
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movl %r9d,16+12(%rsp)
> + leaq 2(%r8),%r9
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> +.byte 102,68,15,56,220,192
> +.byte 102,68,15,56,220,200
> + movups 64-128(%rcx),%xmm0
> + bswapl %r9d
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> + xorl %ebp,%r9d
> +.byte 0x66,0x90
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movl %r9d,32+12(%rsp)
> + leaq 3(%r8),%r9
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> + movups 80-128(%rcx),%xmm1
> + bswapl %r9d
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> + xorl %ebp,%r9d
> +.byte 0x66,0x90
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movl %r9d,48+12(%rsp)
> + leaq 4(%r8),%r9
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> +.byte 102,68,15,56,220,192
> +.byte 102,68,15,56,220,200
> + movups 96-128(%rcx),%xmm0
> + bswapl %r9d
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> + xorl %ebp,%r9d
> +.byte 0x66,0x90
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movl %r9d,64+12(%rsp)
> + leaq 5(%r8),%r9
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> + movups 112-128(%rcx),%xmm1
> + bswapl %r9d
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> + xorl %ebp,%r9d
> +.byte 0x66,0x90
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movl %r9d,80+12(%rsp)
> + leaq 6(%r8),%r9
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> +.byte 102,68,15,56,220,192
> +.byte 102,68,15,56,220,200
> + movups 128-128(%rcx),%xmm0
> + bswapl %r9d
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> + xorl %ebp,%r9d
> +.byte 0x66,0x90
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movl %r9d,96+12(%rsp)
> + leaq 7(%r8),%r9
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> + movups 144-128(%rcx),%xmm1
> + bswapl %r9d
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> + xorl %ebp,%r9d
> + movdqu 0(%rdi),%xmm10
> +.byte 102,15,56,220,232
> + movl %r9d,112+12(%rsp)
> + cmpl $11,%eax
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> +.byte 102,68,15,56,220,192
> +.byte 102,68,15,56,220,200
> + movups 160-128(%rcx),%xmm0
> +
> + jb .Lctr32_enc_done
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> + movups 176-128(%rcx),%xmm1
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> +.byte 102,68,15,56,220,192
> +.byte 102,68,15,56,220,200
> + movups 192-128(%rcx),%xmm0
> + je .Lctr32_enc_done
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> + movups 208-128(%rcx),%xmm1
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> +.byte 102,68,15,56,220,192
> +.byte 102,68,15,56,220,200
> + movups 224-128(%rcx),%xmm0
> + jmp .Lctr32_enc_done
> +
> +.align 16
> +.Lctr32_enc_done:
> + movdqu 16(%rdi),%xmm11
> + pxor %xmm0,%xmm10
> + movdqu 32(%rdi),%xmm12
> + pxor %xmm0,%xmm11
> + movdqu 48(%rdi),%xmm13
> + pxor %xmm0,%xmm12
> + movdqu 64(%rdi),%xmm14
> + pxor %xmm0,%xmm13
> + movdqu 80(%rdi),%xmm15
> + pxor %xmm0,%xmm14
> + pxor %xmm0,%xmm15
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +.byte 102,68,15,56,220,201
> + movdqu 96(%rdi),%xmm1
> + leaq 128(%rdi),%rdi
> +
> +.byte 102,65,15,56,221,210
> + pxor %xmm0,%xmm1
> + movdqu 112-128(%rdi),%xmm10
> +.byte 102,65,15,56,221,219
> + pxor %xmm0,%xmm10
> + movdqa 0(%rsp),%xmm11
> +.byte 102,65,15,56,221,228
> +.byte 102,65,15,56,221,237
> + movdqa 16(%rsp),%xmm12
> + movdqa 32(%rsp),%xmm13
> +.byte 102,65,15,56,221,246
> +.byte 102,65,15,56,221,255
> + movdqa 48(%rsp),%xmm14
> + movdqa 64(%rsp),%xmm15
> +.byte 102,68,15,56,221,193
> + movdqa 80(%rsp),%xmm0
> + movups 16-128(%rcx),%xmm1
> +.byte 102,69,15,56,221,202
> +
> + movups %xmm2,(%rsi)
> + movdqa %xmm11,%xmm2
> + movups %xmm3,16(%rsi)
> + movdqa %xmm12,%xmm3
> + movups %xmm4,32(%rsi)
> + movdqa %xmm13,%xmm4
> + movups %xmm5,48(%rsi)
> + movdqa %xmm14,%xmm5
> + movups %xmm6,64(%rsi)
> + movdqa %xmm15,%xmm6
> + movups %xmm7,80(%rsi)
> + movdqa %xmm0,%xmm7
> + movups %xmm8,96(%rsi)
> + movups %xmm9,112(%rsi)
> + leaq 128(%rsi),%rsi
> +
> + subq $8,%rdx
> + jnc .Lctr32_loop8
> +
> + addq $8,%rdx
> + jz .Lctr32_done
> + leaq -128(%rcx),%rcx
> +
> +.Lctr32_tail:
> +
> +
> + leaq 16(%rcx),%rcx
> + cmpq $4,%rdx
> + jb .Lctr32_loop3
> + je .Lctr32_loop4
> +
> +
> + shll $4,%eax
> + movdqa 96(%rsp),%xmm8
> + pxor %xmm9,%xmm9
> +
> + movups 16(%rcx),%xmm0
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> + leaq 32-16(%rcx,%rax,1),%rcx
> + negq %rax
> +.byte 102,15,56,220,225
> + addq $16,%rax
> + movups (%rdi),%xmm10
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> + movups 16(%rdi),%xmm11
> + movups 32(%rdi),%xmm12
> +.byte 102,15,56,220,249
> +.byte 102,68,15,56,220,193
> +
> + call .Lenc_loop8_enter
> +
> + movdqu 48(%rdi),%xmm13
> + pxor %xmm10,%xmm2
> + movdqu 64(%rdi),%xmm10
> + pxor %xmm11,%xmm3
> + movdqu %xmm2,(%rsi)
> + pxor %xmm12,%xmm4
> + movdqu %xmm3,16(%rsi)
> + pxor %xmm13,%xmm5
> + movdqu %xmm4,32(%rsi)
> + pxor %xmm10,%xmm6
> + movdqu %xmm5,48(%rsi)
> + movdqu %xmm6,64(%rsi)
> + cmpq $6,%rdx
> + jb .Lctr32_done
> +
> + movups 80(%rdi),%xmm11
> + xorps %xmm11,%xmm7
> + movups %xmm7,80(%rsi)
> + je .Lctr32_done
> +
> + movups 96(%rdi),%xmm12
> + xorps %xmm12,%xmm8
> + movups %xmm8,96(%rsi)
> + jmp .Lctr32_done
> +
> +.align 32
> +.Lctr32_loop4:
> +.byte 102,15,56,220,209
> + leaq 16(%rcx),%rcx
> + decl %eax
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups (%rcx),%xmm1
> + jnz .Lctr32_loop4
> +.byte 102,15,56,221,209
> +.byte 102,15,56,221,217
> + movups (%rdi),%xmm10
> + movups 16(%rdi),%xmm11
> +.byte 102,15,56,221,225
> +.byte 102,15,56,221,233
> + movups 32(%rdi),%xmm12
> + movups 48(%rdi),%xmm13
> +
> + xorps %xmm10,%xmm2
> + movups %xmm2,(%rsi)
> + xorps %xmm11,%xmm3
> + movups %xmm3,16(%rsi)
> + pxor %xmm12,%xmm4
> + movdqu %xmm4,32(%rsi)
> + pxor %xmm13,%xmm5
> + movdqu %xmm5,48(%rsi)
> + jmp .Lctr32_done
> +
> +.align 32
> +.Lctr32_loop3:
> +.byte 102,15,56,220,209
> + leaq 16(%rcx),%rcx
> + decl %eax
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> + movups (%rcx),%xmm1
> + jnz .Lctr32_loop3
> +.byte 102,15,56,221,209
> +.byte 102,15,56,221,217
> +.byte 102,15,56,221,225
> +
> + movups (%rdi),%xmm10
> + xorps %xmm10,%xmm2
> + movups %xmm2,(%rsi)
> + cmpq $2,%rdx
> + jb .Lctr32_done
> +
> + movups 16(%rdi),%xmm11
> + xorps %xmm11,%xmm3
> + movups %xmm3,16(%rsi)
> + je .Lctr32_done
> +
> + movups 32(%rdi),%xmm12
> + xorps %xmm12,%xmm4
> + movups %xmm4,32(%rsi)
> +
> +.Lctr32_done:
> + xorps %xmm0,%xmm0
> + xorl %ebp,%ebp
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + movaps %xmm0,0(%rsp)
> + pxor %xmm8,%xmm8
> + movaps %xmm0,16(%rsp)
> + pxor %xmm9,%xmm9
> + movaps %xmm0,32(%rsp)
> + pxor %xmm10,%xmm10
> + movaps %xmm0,48(%rsp)
> + pxor %xmm11,%xmm11
> + movaps %xmm0,64(%rsp)
> + pxor %xmm12,%xmm12
> + movaps %xmm0,80(%rsp)
> + pxor %xmm13,%xmm13
> + movaps %xmm0,96(%rsp)
> + pxor %xmm14,%xmm14
> + movaps %xmm0,112(%rsp)
> + pxor %xmm15,%xmm15
> + movq -8(%r11),%rbp
> +.cfi_restore %rbp
> + leaq (%r11),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lctr32_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_ctr32_encrypt_blocks,.-aesni_ctr32_encrypt_blocks
> +.globl aesni_xts_encrypt
> +.type aesni_xts_encrypt, at function
> +.align 16
> +aesni_xts_encrypt:
> +.cfi_startproc
> + leaq (%rsp),%r11
> +.cfi_def_cfa_register %r11
> + pushq %rbp
> +.cfi_offset %rbp,-16
> + subq $112,%rsp
> + andq $-16,%rsp
> + movups (%r9),%xmm2
> + movl 240(%r8),%eax
> + movl 240(%rcx),%r10d
> + movups (%r8),%xmm0
> + movups 16(%r8),%xmm1
> + leaq 32(%r8),%r8
> + xorps %xmm0,%xmm2
> +.Loop_enc1_8:
> +.byte 102,15,56,220,209
> + decl %eax
> + movups (%r8),%xmm1
> + leaq 16(%r8),%r8
> + jnz .Loop_enc1_8
> +.byte 102,15,56,221,209
> + movups (%rcx),%xmm0
> + movq %rcx,%rbp
> + movl %r10d,%eax
> + shll $4,%r10d
> + movq %rdx,%r9
> + andq $-16,%rdx
> +
> + movups 16(%rcx,%r10,1),%xmm1
> +
> + movdqa .Lxts_magic(%rip),%xmm8
> + movdqa %xmm2,%xmm15
> + pshufd $0x5f,%xmm2,%xmm9
> + pxor %xmm0,%xmm1
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> + movdqa %xmm15,%xmm10
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> + pxor %xmm0,%xmm10
> + pxor %xmm14,%xmm15
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> + movdqa %xmm15,%xmm11
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> + pxor %xmm0,%xmm11
> + pxor %xmm14,%xmm15
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> + movdqa %xmm15,%xmm12
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> + pxor %xmm0,%xmm12
> + pxor %xmm14,%xmm15
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> + movdqa %xmm15,%xmm13
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> + pxor %xmm0,%xmm13
> + pxor %xmm14,%xmm15
> + movdqa %xmm15,%xmm14
> + psrad $31,%xmm9
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm9
> + pxor %xmm0,%xmm14
> + pxor %xmm9,%xmm15
> + movaps %xmm1,96(%rsp)
> +
> + subq $96,%rdx
> + jc .Lxts_enc_short
> +
> + movl $16+96,%eax
> + leaq 32(%rbp,%r10,1),%rcx
> + subq %r10,%rax
> + movups 16(%rbp),%xmm1
> + movq %rax,%r10
> + leaq .Lxts_magic(%rip),%r8
> + jmp .Lxts_enc_grandloop
> +
> +.align 32
> +.Lxts_enc_grandloop:
> + movdqu 0(%rdi),%xmm2
> + movdqa %xmm0,%xmm8
> + movdqu 16(%rdi),%xmm3
> + pxor %xmm10,%xmm2
> + movdqu 32(%rdi),%xmm4
> + pxor %xmm11,%xmm3
> +.byte 102,15,56,220,209
> + movdqu 48(%rdi),%xmm5
> + pxor %xmm12,%xmm4
> +.byte 102,15,56,220,217
> + movdqu 64(%rdi),%xmm6
> + pxor %xmm13,%xmm5
> +.byte 102,15,56,220,225
> + movdqu 80(%rdi),%xmm7
> + pxor %xmm15,%xmm8
> + movdqa 96(%rsp),%xmm9
> + pxor %xmm14,%xmm6
> +.byte 102,15,56,220,233
> + movups 32(%rbp),%xmm0
> + leaq 96(%rdi),%rdi
> + pxor %xmm8,%xmm7
> +
> + pxor %xmm9,%xmm10
> +.byte 102,15,56,220,241
> + pxor %xmm9,%xmm11
> + movdqa %xmm10,0(%rsp)
> +.byte 102,15,56,220,249
> + movups 48(%rbp),%xmm1
> + pxor %xmm9,%xmm12
> +
> +.byte 102,15,56,220,208
> + pxor %xmm9,%xmm13
> + movdqa %xmm11,16(%rsp)
> +.byte 102,15,56,220,216
> + pxor %xmm9,%xmm14
> + movdqa %xmm12,32(%rsp)
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + pxor %xmm9,%xmm8
> + movdqa %xmm14,64(%rsp)
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> + movups 64(%rbp),%xmm0
> + movdqa %xmm8,80(%rsp)
> + pshufd $0x5f,%xmm15,%xmm9
> + jmp .Lxts_enc_loop6
> +.align 32
> +.Lxts_enc_loop6:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> + movups -64(%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> + movups -80(%rcx,%rax,1),%xmm0
> + jnz .Lxts_enc_loop6
> +
> + movdqa (%r8),%xmm8
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> +.byte 102,15,56,220,209
> + paddq %xmm15,%xmm15
> + psrad $31,%xmm14
> +.byte 102,15,56,220,217
> + pand %xmm8,%xmm14
> + movups (%rbp),%xmm10
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> + pxor %xmm14,%xmm15
> + movaps %xmm10,%xmm11
> +.byte 102,15,56,220,249
> + movups -64(%rcx),%xmm1
> +
> + movdqa %xmm9,%xmm14
> +.byte 102,15,56,220,208
> + paddd %xmm9,%xmm9
> + pxor %xmm15,%xmm10
> +.byte 102,15,56,220,216
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + pand %xmm8,%xmm14
> + movaps %xmm11,%xmm12
> +.byte 102,15,56,220,240
> + pxor %xmm14,%xmm15
> + movdqa %xmm9,%xmm14
> +.byte 102,15,56,220,248
> + movups -48(%rcx),%xmm0
> +
> + paddd %xmm9,%xmm9
> +.byte 102,15,56,220,209
> + pxor %xmm15,%xmm11
> + psrad $31,%xmm14
> +.byte 102,15,56,220,217
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movdqa %xmm13,48(%rsp)
> + pxor %xmm14,%xmm15
> +.byte 102,15,56,220,241
> + movaps %xmm12,%xmm13
> + movdqa %xmm9,%xmm14
> +.byte 102,15,56,220,249
> + movups -32(%rcx),%xmm1
> +
> + paddd %xmm9,%xmm9
> +.byte 102,15,56,220,208
> + pxor %xmm15,%xmm12
> + psrad $31,%xmm14
> +.byte 102,15,56,220,216
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> +.byte 102,15,56,220,240
> + pxor %xmm14,%xmm15
> + movaps %xmm13,%xmm14
> +.byte 102,15,56,220,248
> +
> + movdqa %xmm9,%xmm0
> + paddd %xmm9,%xmm9
> +.byte 102,15,56,220,209
> + pxor %xmm15,%xmm13
> + psrad $31,%xmm0
> +.byte 102,15,56,220,217
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm0
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + pxor %xmm0,%xmm15
> + movups (%rbp),%xmm0
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> + movups 16(%rbp),%xmm1
> +
> + pxor %xmm15,%xmm14
> +.byte 102,15,56,221,84,36,0
> + psrad $31,%xmm9
> + paddq %xmm15,%xmm15
> +.byte 102,15,56,221,92,36,16
> +.byte 102,15,56,221,100,36,32
> + pand %xmm8,%xmm9
> + movq %r10,%rax
> +.byte 102,15,56,221,108,36,48
> +.byte 102,15,56,221,116,36,64
> +.byte 102,15,56,221,124,36,80
> + pxor %xmm9,%xmm15
> +
> + leaq 96(%rsi),%rsi
> + movups %xmm2,-96(%rsi)
> + movups %xmm3,-80(%rsi)
> + movups %xmm4,-64(%rsi)
> + movups %xmm5,-48(%rsi)
> + movups %xmm6,-32(%rsi)
> + movups %xmm7,-16(%rsi)
> + subq $96,%rdx
> + jnc .Lxts_enc_grandloop
> +
> + movl $16+96,%eax
> + subl %r10d,%eax
> + movq %rbp,%rcx
> + shrl $4,%eax
> +
> +.Lxts_enc_short:
> +
> + movl %eax,%r10d
> + pxor %xmm0,%xmm10
> + addq $96,%rdx
> + jz .Lxts_enc_done
> +
> + pxor %xmm0,%xmm11
> + cmpq $0x20,%rdx
> + jb .Lxts_enc_one
> + pxor %xmm0,%xmm12
> + je .Lxts_enc_two
> +
> + pxor %xmm0,%xmm13
> + cmpq $0x40,%rdx
> + jb .Lxts_enc_three
> + pxor %xmm0,%xmm14
> + je .Lxts_enc_four
> +
> + movdqu (%rdi),%xmm2
> + movdqu 16(%rdi),%xmm3
> + movdqu 32(%rdi),%xmm4
> + pxor %xmm10,%xmm2
> + movdqu 48(%rdi),%xmm5
> + pxor %xmm11,%xmm3
> + movdqu 64(%rdi),%xmm6
> + leaq 80(%rdi),%rdi
> + pxor %xmm12,%xmm4
> + pxor %xmm13,%xmm5
> + pxor %xmm14,%xmm6
> + pxor %xmm7,%xmm7
> +
> + call _aesni_encrypt6
> +
> + xorps %xmm10,%xmm2
> + movdqa %xmm15,%xmm10
> + xorps %xmm11,%xmm3
> + xorps %xmm12,%xmm4
> + movdqu %xmm2,(%rsi)
> + xorps %xmm13,%xmm5
> + movdqu %xmm3,16(%rsi)
> + xorps %xmm14,%xmm6
> + movdqu %xmm4,32(%rsi)
> + movdqu %xmm5,48(%rsi)
> + movdqu %xmm6,64(%rsi)
> + leaq 80(%rsi),%rsi
> + jmp .Lxts_enc_done
> +
> +.align 16
> +.Lxts_enc_one:
> + movups (%rdi),%xmm2
> + leaq 16(%rdi),%rdi
> + xorps %xmm10,%xmm2
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_enc1_9:
> +.byte 102,15,56,220,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_enc1_9
> +.byte 102,15,56,221,209
> + xorps %xmm10,%xmm2
> + movdqa %xmm11,%xmm10
> + movups %xmm2,(%rsi)
> + leaq 16(%rsi),%rsi
> + jmp .Lxts_enc_done
> +
> +.align 16
> +.Lxts_enc_two:
> + movups (%rdi),%xmm2
> + movups 16(%rdi),%xmm3
> + leaq 32(%rdi),%rdi
> + xorps %xmm10,%xmm2
> + xorps %xmm11,%xmm3
> +
> + call _aesni_encrypt2
> +
> + xorps %xmm10,%xmm2
> + movdqa %xmm12,%xmm10
> + xorps %xmm11,%xmm3
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + leaq 32(%rsi),%rsi
> + jmp .Lxts_enc_done
> +
> +.align 16
> +.Lxts_enc_three:
> + movups (%rdi),%xmm2
> + movups 16(%rdi),%xmm3
> + movups 32(%rdi),%xmm4
> + leaq 48(%rdi),%rdi
> + xorps %xmm10,%xmm2
> + xorps %xmm11,%xmm3
> + xorps %xmm12,%xmm4
> +
> + call _aesni_encrypt3
> +
> + xorps %xmm10,%xmm2
> + movdqa %xmm13,%xmm10
> + xorps %xmm11,%xmm3
> + xorps %xmm12,%xmm4
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + leaq 48(%rsi),%rsi
> + jmp .Lxts_enc_done
> +
> +.align 16
> +.Lxts_enc_four:
> + movups (%rdi),%xmm2
> + movups 16(%rdi),%xmm3
> + movups 32(%rdi),%xmm4
> + xorps %xmm10,%xmm2
> + movups 48(%rdi),%xmm5
> + leaq 64(%rdi),%rdi
> + xorps %xmm11,%xmm3
> + xorps %xmm12,%xmm4
> + xorps %xmm13,%xmm5
> +
> + call _aesni_encrypt4
> +
> + pxor %xmm10,%xmm2
> + movdqa %xmm14,%xmm10
> + pxor %xmm11,%xmm3
> + pxor %xmm12,%xmm4
> + movdqu %xmm2,(%rsi)
> + pxor %xmm13,%xmm5
> + movdqu %xmm3,16(%rsi)
> + movdqu %xmm4,32(%rsi)
> + movdqu %xmm5,48(%rsi)
> + leaq 64(%rsi),%rsi
> + jmp .Lxts_enc_done
> +
> +.align 16
> +.Lxts_enc_done:
> + andq $15,%r9
> + jz .Lxts_enc_ret
> + movq %r9,%rdx
> +
> +.Lxts_enc_steal:
> + movzbl (%rdi),%eax
> + movzbl -16(%rsi),%ecx
> + leaq 1(%rdi),%rdi
> + movb %al,-16(%rsi)
> + movb %cl,0(%rsi)
> + leaq 1(%rsi),%rsi
> + subq $1,%rdx
> + jnz .Lxts_enc_steal
> +
> + subq %r9,%rsi
> + movq %rbp,%rcx
> + movl %r10d,%eax
> +
> + movups -16(%rsi),%xmm2
> + xorps %xmm10,%xmm2
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_enc1_10:
> +.byte 102,15,56,220,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_enc1_10
> +.byte 102,15,56,221,209
> + xorps %xmm10,%xmm2
> + movups %xmm2,-16(%rsi)
> +
> +.Lxts_enc_ret:
> + xorps %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + movaps %xmm0,0(%rsp)
> + pxor %xmm8,%xmm8
> + movaps %xmm0,16(%rsp)
> + pxor %xmm9,%xmm9
> + movaps %xmm0,32(%rsp)
> + pxor %xmm10,%xmm10
> + movaps %xmm0,48(%rsp)
> + pxor %xmm11,%xmm11
> + movaps %xmm0,64(%rsp)
> + pxor %xmm12,%xmm12
> + movaps %xmm0,80(%rsp)
> + pxor %xmm13,%xmm13
> + movaps %xmm0,96(%rsp)
> + pxor %xmm14,%xmm14
> + pxor %xmm15,%xmm15
> + movq -8(%r11),%rbp
> +.cfi_restore %rbp
> + leaq (%r11),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lxts_enc_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_xts_encrypt,.-aesni_xts_encrypt
> +.globl aesni_xts_decrypt
> +.type aesni_xts_decrypt, at function
> +.align 16
> +aesni_xts_decrypt:
> +.cfi_startproc
> + leaq (%rsp),%r11
> +.cfi_def_cfa_register %r11
> + pushq %rbp
> +.cfi_offset %rbp,-16
> + subq $112,%rsp
> + andq $-16,%rsp
> + movups (%r9),%xmm2
> + movl 240(%r8),%eax
> + movl 240(%rcx),%r10d
> + movups (%r8),%xmm0
> + movups 16(%r8),%xmm1
> + leaq 32(%r8),%r8
> + xorps %xmm0,%xmm2
> +.Loop_enc1_11:
> +.byte 102,15,56,220,209
> + decl %eax
> + movups (%r8),%xmm1
> + leaq 16(%r8),%r8
> + jnz .Loop_enc1_11
> +.byte 102,15,56,221,209
> + xorl %eax,%eax
> + testq $15,%rdx
> + setnz %al
> + shlq $4,%rax
> + subq %rax,%rdx
> +
> + movups (%rcx),%xmm0
> + movq %rcx,%rbp
> + movl %r10d,%eax
> + shll $4,%r10d
> + movq %rdx,%r9
> + andq $-16,%rdx
> +
> + movups 16(%rcx,%r10,1),%xmm1
> +
> + movdqa .Lxts_magic(%rip),%xmm8
> + movdqa %xmm2,%xmm15
> + pshufd $0x5f,%xmm2,%xmm9
> + pxor %xmm0,%xmm1
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> + movdqa %xmm15,%xmm10
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> + pxor %xmm0,%xmm10
> + pxor %xmm14,%xmm15
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> + movdqa %xmm15,%xmm11
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> + pxor %xmm0,%xmm11
> + pxor %xmm14,%xmm15
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> + movdqa %xmm15,%xmm12
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> + pxor %xmm0,%xmm12
> + pxor %xmm14,%xmm15
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> + movdqa %xmm15,%xmm13
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> + pxor %xmm0,%xmm13
> + pxor %xmm14,%xmm15
> + movdqa %xmm15,%xmm14
> + psrad $31,%xmm9
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm9
> + pxor %xmm0,%xmm14
> + pxor %xmm9,%xmm15
> + movaps %xmm1,96(%rsp)
> +
> + subq $96,%rdx
> + jc .Lxts_dec_short
> +
> + movl $16+96,%eax
> + leaq 32(%rbp,%r10,1),%rcx
> + subq %r10,%rax
> + movups 16(%rbp),%xmm1
> + movq %rax,%r10
> + leaq .Lxts_magic(%rip),%r8
> + jmp .Lxts_dec_grandloop
> +
> +.align 32
> +.Lxts_dec_grandloop:
> + movdqu 0(%rdi),%xmm2
> + movdqa %xmm0,%xmm8
> + movdqu 16(%rdi),%xmm3
> + pxor %xmm10,%xmm2
> + movdqu 32(%rdi),%xmm4
> + pxor %xmm11,%xmm3
> +.byte 102,15,56,222,209
> + movdqu 48(%rdi),%xmm5
> + pxor %xmm12,%xmm4
> +.byte 102,15,56,222,217
> + movdqu 64(%rdi),%xmm6
> + pxor %xmm13,%xmm5
> +.byte 102,15,56,222,225
> + movdqu 80(%rdi),%xmm7
> + pxor %xmm15,%xmm8
> + movdqa 96(%rsp),%xmm9
> + pxor %xmm14,%xmm6
> +.byte 102,15,56,222,233
> + movups 32(%rbp),%xmm0
> + leaq 96(%rdi),%rdi
> + pxor %xmm8,%xmm7
> +
> + pxor %xmm9,%xmm10
> +.byte 102,15,56,222,241
> + pxor %xmm9,%xmm11
> + movdqa %xmm10,0(%rsp)
> +.byte 102,15,56,222,249
> + movups 48(%rbp),%xmm1
> + pxor %xmm9,%xmm12
> +
> +.byte 102,15,56,222,208
> + pxor %xmm9,%xmm13
> + movdqa %xmm11,16(%rsp)
> +.byte 102,15,56,222,216
> + pxor %xmm9,%xmm14
> + movdqa %xmm12,32(%rsp)
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + pxor %xmm9,%xmm8
> + movdqa %xmm14,64(%rsp)
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> + movups 64(%rbp),%xmm0
> + movdqa %xmm8,80(%rsp)
> + pshufd $0x5f,%xmm15,%xmm9
> + jmp .Lxts_dec_loop6
> +.align 32
> +.Lxts_dec_loop6:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> + movups -64(%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> + movups -80(%rcx,%rax,1),%xmm0
> + jnz .Lxts_dec_loop6
> +
> + movdqa (%r8),%xmm8
> + movdqa %xmm9,%xmm14
> + paddd %xmm9,%xmm9
> +.byte 102,15,56,222,209
> + paddq %xmm15,%xmm15
> + psrad $31,%xmm14
> +.byte 102,15,56,222,217
> + pand %xmm8,%xmm14
> + movups (%rbp),%xmm10
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> + pxor %xmm14,%xmm15
> + movaps %xmm10,%xmm11
> +.byte 102,15,56,222,249
> + movups -64(%rcx),%xmm1
> +
> + movdqa %xmm9,%xmm14
> +.byte 102,15,56,222,208
> + paddd %xmm9,%xmm9
> + pxor %xmm15,%xmm10
> +.byte 102,15,56,222,216
> + psrad $31,%xmm14
> + paddq %xmm15,%xmm15
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + pand %xmm8,%xmm14
> + movaps %xmm11,%xmm12
> +.byte 102,15,56,222,240
> + pxor %xmm14,%xmm15
> + movdqa %xmm9,%xmm14
> +.byte 102,15,56,222,248
> + movups -48(%rcx),%xmm0
> +
> + paddd %xmm9,%xmm9
> +.byte 102,15,56,222,209
> + pxor %xmm15,%xmm11
> + psrad $31,%xmm14
> +.byte 102,15,56,222,217
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movdqa %xmm13,48(%rsp)
> + pxor %xmm14,%xmm15
> +.byte 102,15,56,222,241
> + movaps %xmm12,%xmm13
> + movdqa %xmm9,%xmm14
> +.byte 102,15,56,222,249
> + movups -32(%rcx),%xmm1
> +
> + paddd %xmm9,%xmm9
> +.byte 102,15,56,222,208
> + pxor %xmm15,%xmm12
> + psrad $31,%xmm14
> +.byte 102,15,56,222,216
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm14
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> + pxor %xmm14,%xmm15
> + movaps %xmm13,%xmm14
> +.byte 102,15,56,222,248
> +
> + movdqa %xmm9,%xmm0
> + paddd %xmm9,%xmm9
> +.byte 102,15,56,222,209
> + pxor %xmm15,%xmm13
> + psrad $31,%xmm0
> +.byte 102,15,56,222,217
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm0
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + pxor %xmm0,%xmm15
> + movups (%rbp),%xmm0
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> + movups 16(%rbp),%xmm1
> +
> + pxor %xmm15,%xmm14
> +.byte 102,15,56,223,84,36,0
> + psrad $31,%xmm9
> + paddq %xmm15,%xmm15
> +.byte 102,15,56,223,92,36,16
> +.byte 102,15,56,223,100,36,32
> + pand %xmm8,%xmm9
> + movq %r10,%rax
> +.byte 102,15,56,223,108,36,48
> +.byte 102,15,56,223,116,36,64
> +.byte 102,15,56,223,124,36,80
> + pxor %xmm9,%xmm15
> +
> + leaq 96(%rsi),%rsi
> + movups %xmm2,-96(%rsi)
> + movups %xmm3,-80(%rsi)
> + movups %xmm4,-64(%rsi)
> + movups %xmm5,-48(%rsi)
> + movups %xmm6,-32(%rsi)
> + movups %xmm7,-16(%rsi)
> + subq $96,%rdx
> + jnc .Lxts_dec_grandloop
> +
> + movl $16+96,%eax
> + subl %r10d,%eax
> + movq %rbp,%rcx
> + shrl $4,%eax
> +
> +.Lxts_dec_short:
> +
> + movl %eax,%r10d
> + pxor %xmm0,%xmm10
> + pxor %xmm0,%xmm11
> + addq $96,%rdx
> + jz .Lxts_dec_done
> +
> + pxor %xmm0,%xmm12
> + cmpq $0x20,%rdx
> + jb .Lxts_dec_one
> + pxor %xmm0,%xmm13
> + je .Lxts_dec_two
> +
> + pxor %xmm0,%xmm14
> + cmpq $0x40,%rdx
> + jb .Lxts_dec_three
> + je .Lxts_dec_four
> +
> + movdqu (%rdi),%xmm2
> + movdqu 16(%rdi),%xmm3
> + movdqu 32(%rdi),%xmm4
> + pxor %xmm10,%xmm2
> + movdqu 48(%rdi),%xmm5
> + pxor %xmm11,%xmm3
> + movdqu 64(%rdi),%xmm6
> + leaq 80(%rdi),%rdi
> + pxor %xmm12,%xmm4
> + pxor %xmm13,%xmm5
> + pxor %xmm14,%xmm6
> +
> + call _aesni_decrypt6
> +
> + xorps %xmm10,%xmm2
> + xorps %xmm11,%xmm3
> + xorps %xmm12,%xmm4
> + movdqu %xmm2,(%rsi)
> + xorps %xmm13,%xmm5
> + movdqu %xmm3,16(%rsi)
> + xorps %xmm14,%xmm6
> + movdqu %xmm4,32(%rsi)
> + pxor %xmm14,%xmm14
> + movdqu %xmm5,48(%rsi)
> + pcmpgtd %xmm15,%xmm14
> + movdqu %xmm6,64(%rsi)
> + leaq 80(%rsi),%rsi
> + pshufd $0x13,%xmm14,%xmm11
> + andq $15,%r9
> + jz .Lxts_dec_ret
> +
> + movdqa %xmm15,%xmm10
> + paddq %xmm15,%xmm15
> + pand %xmm8,%xmm11
> + pxor %xmm15,%xmm11
> + jmp .Lxts_dec_done2
> +
> +.align 16
> +.Lxts_dec_one:
> + movups (%rdi),%xmm2
> + leaq 16(%rdi),%rdi
> + xorps %xmm10,%xmm2
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_dec1_12:
> +.byte 102,15,56,222,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_dec1_12
> +.byte 102,15,56,223,209
> + xorps %xmm10,%xmm2
> + movdqa %xmm11,%xmm10
> + movups %xmm2,(%rsi)
> + movdqa %xmm12,%xmm11
> + leaq 16(%rsi),%rsi
> + jmp .Lxts_dec_done
> +
> +.align 16
> +.Lxts_dec_two:
> + movups (%rdi),%xmm2
> + movups 16(%rdi),%xmm3
> + leaq 32(%rdi),%rdi
> + xorps %xmm10,%xmm2
> + xorps %xmm11,%xmm3
> +
> + call _aesni_decrypt2
> +
> + xorps %xmm10,%xmm2
> + movdqa %xmm12,%xmm10
> + xorps %xmm11,%xmm3
> + movdqa %xmm13,%xmm11
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + leaq 32(%rsi),%rsi
> + jmp .Lxts_dec_done
> +
> +.align 16
> +.Lxts_dec_three:
> + movups (%rdi),%xmm2
> + movups 16(%rdi),%xmm3
> + movups 32(%rdi),%xmm4
> + leaq 48(%rdi),%rdi
> + xorps %xmm10,%xmm2
> + xorps %xmm11,%xmm3
> + xorps %xmm12,%xmm4
> +
> + call _aesni_decrypt3
> +
> + xorps %xmm10,%xmm2
> + movdqa %xmm13,%xmm10
> + xorps %xmm11,%xmm3
> + movdqa %xmm14,%xmm11
> + xorps %xmm12,%xmm4
> + movups %xmm2,(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + leaq 48(%rsi),%rsi
> + jmp .Lxts_dec_done
> +
> +.align 16
> +.Lxts_dec_four:
> + movups (%rdi),%xmm2
> + movups 16(%rdi),%xmm3
> + movups 32(%rdi),%xmm4
> + xorps %xmm10,%xmm2
> + movups 48(%rdi),%xmm5
> + leaq 64(%rdi),%rdi
> + xorps %xmm11,%xmm3
> + xorps %xmm12,%xmm4
> + xorps %xmm13,%xmm5
> +
> + call _aesni_decrypt4
> +
> + pxor %xmm10,%xmm2
> + movdqa %xmm14,%xmm10
> + pxor %xmm11,%xmm3
> + movdqa %xmm15,%xmm11
> + pxor %xmm12,%xmm4
> + movdqu %xmm2,(%rsi)
> + pxor %xmm13,%xmm5
> + movdqu %xmm3,16(%rsi)
> + movdqu %xmm4,32(%rsi)
> + movdqu %xmm5,48(%rsi)
> + leaq 64(%rsi),%rsi
> + jmp .Lxts_dec_done
> +
> +.align 16
> +.Lxts_dec_done:
> + andq $15,%r9
> + jz .Lxts_dec_ret
> +.Lxts_dec_done2:
> + movq %r9,%rdx
> + movq %rbp,%rcx
> + movl %r10d,%eax
> +
> + movups (%rdi),%xmm2
> + xorps %xmm11,%xmm2
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_dec1_13:
> +.byte 102,15,56,222,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_dec1_13
> +.byte 102,15,56,223,209
> + xorps %xmm11,%xmm2
> + movups %xmm2,(%rsi)
> +
> +.Lxts_dec_steal:
> + movzbl 16(%rdi),%eax
> + movzbl (%rsi),%ecx
> + leaq 1(%rdi),%rdi
> + movb %al,(%rsi)
> + movb %cl,16(%rsi)
> + leaq 1(%rsi),%rsi
> + subq $1,%rdx
> + jnz .Lxts_dec_steal
> +
> + subq %r9,%rsi
> + movq %rbp,%rcx
> + movl %r10d,%eax
> +
> + movups (%rsi),%xmm2
> + xorps %xmm10,%xmm2
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_dec1_14:
> +.byte 102,15,56,222,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_dec1_14
> +.byte 102,15,56,223,209
> + xorps %xmm10,%xmm2
> + movups %xmm2,(%rsi)
> +
> +.Lxts_dec_ret:
> + xorps %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + movaps %xmm0,0(%rsp)
> + pxor %xmm8,%xmm8
> + movaps %xmm0,16(%rsp)
> + pxor %xmm9,%xmm9
> + movaps %xmm0,32(%rsp)
> + pxor %xmm10,%xmm10
> + movaps %xmm0,48(%rsp)
> + pxor %xmm11,%xmm11
> + movaps %xmm0,64(%rsp)
> + pxor %xmm12,%xmm12
> + movaps %xmm0,80(%rsp)
> + pxor %xmm13,%xmm13
> + movaps %xmm0,96(%rsp)
> + pxor %xmm14,%xmm14
> + pxor %xmm15,%xmm15
> + movq -8(%r11),%rbp
> +.cfi_restore %rbp
> + leaq (%r11),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lxts_dec_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_xts_decrypt,.-aesni_xts_decrypt
> +.globl aesni_ocb_encrypt
> +.type aesni_ocb_encrypt, at function
> +.align 32
> +aesni_ocb_encrypt:
> +.cfi_startproc
> + leaq (%rsp),%rax
> + pushq %rbx
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r14,-48
> + movq 8(%rax),%rbx
> + movq 8+8(%rax),%rbp
> +
> + movl 240(%rcx),%r10d
> + movq %rcx,%r11
> + shll $4,%r10d
> + movups (%rcx),%xmm9
> + movups 16(%rcx,%r10,1),%xmm1
> +
> + movdqu (%r9),%xmm15
> + pxor %xmm1,%xmm9
> + pxor %xmm1,%xmm15
> +
> + movl $16+32,%eax
> + leaq 32(%r11,%r10,1),%rcx
> + movups 16(%r11),%xmm1
> + subq %r10,%rax
> + movq %rax,%r10
> +
> + movdqu (%rbx),%xmm10
> + movdqu (%rbp),%xmm8
> +
> + testq $1,%r8
> + jnz .Locb_enc_odd
> +
> + bsfq %r8,%r12
> + addq $1,%r8
> + shlq $4,%r12
> + movdqu (%rbx,%r12,1),%xmm7
> + movdqu (%rdi),%xmm2
> + leaq 16(%rdi),%rdi
> +
> + call __ocb_encrypt1
> +
> + movdqa %xmm7,%xmm15
> + movups %xmm2,(%rsi)
> + leaq 16(%rsi),%rsi
> + subq $1,%rdx
> + jz .Locb_enc_done
> +
> +.Locb_enc_odd:
> + leaq 1(%r8),%r12
> + leaq 3(%r8),%r13
> + leaq 5(%r8),%r14
> + leaq 6(%r8),%r8
> + bsfq %r12,%r12
> + bsfq %r13,%r13
> + bsfq %r14,%r14
> + shlq $4,%r12
> + shlq $4,%r13
> + shlq $4,%r14
> +
> + subq $6,%rdx
> + jc .Locb_enc_short
> + jmp .Locb_enc_grandloop
> +
> +.align 32
> +.Locb_enc_grandloop:
> + movdqu 0(%rdi),%xmm2
> + movdqu 16(%rdi),%xmm3
> + movdqu 32(%rdi),%xmm4
> + movdqu 48(%rdi),%xmm5
> + movdqu 64(%rdi),%xmm6
> + movdqu 80(%rdi),%xmm7
> + leaq 96(%rdi),%rdi
> +
> + call __ocb_encrypt6
> +
> + movups %xmm2,0(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + movups %xmm5,48(%rsi)
> + movups %xmm6,64(%rsi)
> + movups %xmm7,80(%rsi)
> + leaq 96(%rsi),%rsi
> + subq $6,%rdx
> + jnc .Locb_enc_grandloop
> +
> +.Locb_enc_short:
> + addq $6,%rdx
> + jz .Locb_enc_done
> +
> + movdqu 0(%rdi),%xmm2
> + cmpq $2,%rdx
> + jb .Locb_enc_one
> + movdqu 16(%rdi),%xmm3
> + je .Locb_enc_two
> +
> + movdqu 32(%rdi),%xmm4
> + cmpq $4,%rdx
> + jb .Locb_enc_three
> + movdqu 48(%rdi),%xmm5
> + je .Locb_enc_four
> +
> + movdqu 64(%rdi),%xmm6
> + pxor %xmm7,%xmm7
> +
> + call __ocb_encrypt6
> +
> + movdqa %xmm14,%xmm15
> + movups %xmm2,0(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + movups %xmm5,48(%rsi)
> + movups %xmm6,64(%rsi)
> +
> + jmp .Locb_enc_done
> +
> +.align 16
> +.Locb_enc_one:
> + movdqa %xmm10,%xmm7
> +
> + call __ocb_encrypt1
> +
> + movdqa %xmm7,%xmm15
> + movups %xmm2,0(%rsi)
> + jmp .Locb_enc_done
> +
> +.align 16
> +.Locb_enc_two:
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> +
> + call __ocb_encrypt4
> +
> + movdqa %xmm11,%xmm15
> + movups %xmm2,0(%rsi)
> + movups %xmm3,16(%rsi)
> +
> + jmp .Locb_enc_done
> +
> +.align 16
> +.Locb_enc_three:
> + pxor %xmm5,%xmm5
> +
> + call __ocb_encrypt4
> +
> + movdqa %xmm12,%xmm15
> + movups %xmm2,0(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> +
> + jmp .Locb_enc_done
> +
> +.align 16
> +.Locb_enc_four:
> + call __ocb_encrypt4
> +
> + movdqa %xmm13,%xmm15
> + movups %xmm2,0(%rsi)
> + movups %xmm3,16(%rsi)
> + movups %xmm4,32(%rsi)
> + movups %xmm5,48(%rsi)
> +
> +.Locb_enc_done:
> + pxor %xmm0,%xmm15
> + movdqu %xmm8,(%rbp)
> + movdqu %xmm15,(%r9)
> +
> + xorps %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + pxor %xmm8,%xmm8
> + pxor %xmm9,%xmm9
> + pxor %xmm10,%xmm10
> + pxor %xmm11,%xmm11
> + pxor %xmm12,%xmm12
> + pxor %xmm13,%xmm13
> + pxor %xmm14,%xmm14
> + pxor %xmm15,%xmm15
> + leaq 40(%rsp),%rax
> +.cfi_def_cfa %rax,8
> + movq -40(%rax),%r14
> +.cfi_restore %r14
> + movq -32(%rax),%r13
> +.cfi_restore %r13
> + movq -24(%rax),%r12
> +.cfi_restore %r12
> + movq -16(%rax),%rbp
> +.cfi_restore %rbp
> + movq -8(%rax),%rbx
> +.cfi_restore %rbx
> + leaq (%rax),%rsp
> +.cfi_def_cfa_register %rsp
> +.Locb_enc_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_ocb_encrypt,.-aesni_ocb_encrypt
> +
> +.type __ocb_encrypt6, at function
> +.align 32
> +__ocb_encrypt6:
> +.cfi_startproc
> + pxor %xmm9,%xmm15
> + movdqu (%rbx,%r12,1),%xmm11
> + movdqa %xmm10,%xmm12
> + movdqu (%rbx,%r13,1),%xmm13
> + movdqa %xmm10,%xmm14
> + pxor %xmm15,%xmm10
> + movdqu (%rbx,%r14,1),%xmm15
> + pxor %xmm10,%xmm11
> + pxor %xmm2,%xmm8
> + pxor %xmm10,%xmm2
> + pxor %xmm11,%xmm12
> + pxor %xmm3,%xmm8
> + pxor %xmm11,%xmm3
> + pxor %xmm12,%xmm13
> + pxor %xmm4,%xmm8
> + pxor %xmm12,%xmm4
> + pxor %xmm13,%xmm14
> + pxor %xmm5,%xmm8
> + pxor %xmm13,%xmm5
> + pxor %xmm14,%xmm15
> + pxor %xmm6,%xmm8
> + pxor %xmm14,%xmm6
> + pxor %xmm7,%xmm8
> + pxor %xmm15,%xmm7
> + movups 32(%r11),%xmm0
> +
> + leaq 1(%r8),%r12
> + leaq 3(%r8),%r13
> + leaq 5(%r8),%r14
> + addq $6,%r8
> + pxor %xmm9,%xmm10
> + bsfq %r12,%r12
> + bsfq %r13,%r13
> + bsfq %r14,%r14
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + pxor %xmm9,%xmm11
> + pxor %xmm9,%xmm12
> +.byte 102,15,56,220,241
> + pxor %xmm9,%xmm13
> + pxor %xmm9,%xmm14
> +.byte 102,15,56,220,249
> + movups 48(%r11),%xmm1
> + pxor %xmm9,%xmm15
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> + movups 64(%r11),%xmm0
> + shlq $4,%r12
> + shlq $4,%r13
> + jmp .Locb_enc_loop6
> +
> +.align 32
> +.Locb_enc_loop6:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> +.byte 102,15,56,220,240
> +.byte 102,15,56,220,248
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Locb_enc_loop6
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> +.byte 102,15,56,220,241
> +.byte 102,15,56,220,249
> + movups 16(%r11),%xmm1
> + shlq $4,%r14
> +
> +.byte 102,65,15,56,221,210
> + movdqu (%rbx),%xmm10
> + movq %r10,%rax
> +.byte 102,65,15,56,221,219
> +.byte 102,65,15,56,221,228
> +.byte 102,65,15,56,221,237
> +.byte 102,65,15,56,221,246
> +.byte 102,65,15,56,221,255
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size __ocb_encrypt6,.-__ocb_encrypt6
> +
> +.type __ocb_encrypt4, at function
> +.align 32
> +__ocb_encrypt4:
> +.cfi_startproc
> + pxor %xmm9,%xmm15
> + movdqu (%rbx,%r12,1),%xmm11
> + movdqa %xmm10,%xmm12
> + movdqu (%rbx,%r13,1),%xmm13
> + pxor %xmm15,%xmm10
> + pxor %xmm10,%xmm11
> + pxor %xmm2,%xmm8
> + pxor %xmm10,%xmm2
> + pxor %xmm11,%xmm12
> + pxor %xmm3,%xmm8
> + pxor %xmm11,%xmm3
> + pxor %xmm12,%xmm13
> + pxor %xmm4,%xmm8
> + pxor %xmm12,%xmm4
> + pxor %xmm5,%xmm8
> + pxor %xmm13,%xmm5
> + movups 32(%r11),%xmm0
> +
> + pxor %xmm9,%xmm10
> + pxor %xmm9,%xmm11
> + pxor %xmm9,%xmm12
> + pxor %xmm9,%xmm13
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups 48(%r11),%xmm1
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movups 64(%r11),%xmm0
> + jmp .Locb_enc_loop4
> +
> +.align 32
> +.Locb_enc_loop4:
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +
> +.byte 102,15,56,220,208
> +.byte 102,15,56,220,216
> +.byte 102,15,56,220,224
> +.byte 102,15,56,220,232
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Locb_enc_loop4
> +
> +.byte 102,15,56,220,209
> +.byte 102,15,56,220,217
> +.byte 102,15,56,220,225
> +.byte 102,15,56,220,233
> + movups 16(%r11),%xmm1
> + movq %r10,%rax
> +
> +.byte 102,65,15,56,221,210
> +.byte 102,65,15,56,221,219
> +.byte 102,65,15,56,221,228
> +.byte 102,65,15,56,221,237
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size __ocb_encrypt4,.-__ocb_encrypt4
> +
> +.type __ocb_encrypt1, at function
> +.align 32
> +__ocb_encrypt1:
> +.cfi_startproc
> + pxor %xmm15,%xmm7
> + pxor %xmm9,%xmm7
> + pxor %xmm2,%xmm8
> + pxor %xmm7,%xmm2
> + movups 32(%r11),%xmm0
> +
> +.byte 102,15,56,220,209
> + movups 48(%r11),%xmm1
> + pxor %xmm9,%xmm7
> +
> +.byte 102,15,56,220,208
> + movups 64(%r11),%xmm0
> + jmp .Locb_enc_loop1
> +
> +.align 32
> +.Locb_enc_loop1:
> +.byte 102,15,56,220,209
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +
> +.byte 102,15,56,220,208
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Locb_enc_loop1
> +
> +.byte 102,15,56,220,209
> + movups 16(%r11),%xmm1
> + movq %r10,%rax
> +
> +.byte 102,15,56,221,215
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size __ocb_encrypt1,.-__ocb_encrypt1
> +
> +.globl aesni_ocb_decrypt
> +.type aesni_ocb_decrypt, at function
> +.align 32
> +aesni_ocb_decrypt:
> +.cfi_startproc
> + leaq (%rsp),%rax
> + pushq %rbx
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r14,-48
> + movq 8(%rax),%rbx
> + movq 8+8(%rax),%rbp
> +
> + movl 240(%rcx),%r10d
> + movq %rcx,%r11
> + shll $4,%r10d
> + movups (%rcx),%xmm9
> + movups 16(%rcx,%r10,1),%xmm1
> +
> + movdqu (%r9),%xmm15
> + pxor %xmm1,%xmm9
> + pxor %xmm1,%xmm15
> +
> + movl $16+32,%eax
> + leaq 32(%r11,%r10,1),%rcx
> + movups 16(%r11),%xmm1
> + subq %r10,%rax
> + movq %rax,%r10
> +
> + movdqu (%rbx),%xmm10
> + movdqu (%rbp),%xmm8
> +
> + testq $1,%r8
> + jnz .Locb_dec_odd
> +
> + bsfq %r8,%r12
> + addq $1,%r8
> + shlq $4,%r12
> + movdqu (%rbx,%r12,1),%xmm7
> + movdqu (%rdi),%xmm2
> + leaq 16(%rdi),%rdi
> +
> + call __ocb_decrypt1
> +
> + movdqa %xmm7,%xmm15
> + movups %xmm2,(%rsi)
> + xorps %xmm2,%xmm8
> + leaq 16(%rsi),%rsi
> + subq $1,%rdx
> + jz .Locb_dec_done
> +
> +.Locb_dec_odd:
> + leaq 1(%r8),%r12
> + leaq 3(%r8),%r13
> + leaq 5(%r8),%r14
> + leaq 6(%r8),%r8
> + bsfq %r12,%r12
> + bsfq %r13,%r13
> + bsfq %r14,%r14
> + shlq $4,%r12
> + shlq $4,%r13
> + shlq $4,%r14
> +
> + subq $6,%rdx
> + jc .Locb_dec_short
> + jmp .Locb_dec_grandloop
> +
> +.align 32
> +.Locb_dec_grandloop:
> + movdqu 0(%rdi),%xmm2
> + movdqu 16(%rdi),%xmm3
> + movdqu 32(%rdi),%xmm4
> + movdqu 48(%rdi),%xmm5
> + movdqu 64(%rdi),%xmm6
> + movdqu 80(%rdi),%xmm7
> + leaq 96(%rdi),%rdi
> +
> + call __ocb_decrypt6
> +
> + movups %xmm2,0(%rsi)
> + pxor %xmm2,%xmm8
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm8
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm8
> + movups %xmm5,48(%rsi)
> + pxor %xmm5,%xmm8
> + movups %xmm6,64(%rsi)
> + pxor %xmm6,%xmm8
> + movups %xmm7,80(%rsi)
> + pxor %xmm7,%xmm8
> + leaq 96(%rsi),%rsi
> + subq $6,%rdx
> + jnc .Locb_dec_grandloop
> +
> +.Locb_dec_short:
> + addq $6,%rdx
> + jz .Locb_dec_done
> +
> + movdqu 0(%rdi),%xmm2
> + cmpq $2,%rdx
> + jb .Locb_dec_one
> + movdqu 16(%rdi),%xmm3
> + je .Locb_dec_two
> +
> + movdqu 32(%rdi),%xmm4
> + cmpq $4,%rdx
> + jb .Locb_dec_three
> + movdqu 48(%rdi),%xmm5
> + je .Locb_dec_four
> +
> + movdqu 64(%rdi),%xmm6
> + pxor %xmm7,%xmm7
> +
> + call __ocb_decrypt6
> +
> + movdqa %xmm14,%xmm15
> + movups %xmm2,0(%rsi)
> + pxor %xmm2,%xmm8
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm8
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm8
> + movups %xmm5,48(%rsi)
> + pxor %xmm5,%xmm8
> + movups %xmm6,64(%rsi)
> + pxor %xmm6,%xmm8
> +
> + jmp .Locb_dec_done
> +
> +.align 16
> +.Locb_dec_one:
> + movdqa %xmm10,%xmm7
> +
> + call __ocb_decrypt1
> +
> + movdqa %xmm7,%xmm15
> + movups %xmm2,0(%rsi)
> + xorps %xmm2,%xmm8
> + jmp .Locb_dec_done
> +
> +.align 16
> +.Locb_dec_two:
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> +
> + call __ocb_decrypt4
> +
> + movdqa %xmm11,%xmm15
> + movups %xmm2,0(%rsi)
> + xorps %xmm2,%xmm8
> + movups %xmm3,16(%rsi)
> + xorps %xmm3,%xmm8
> +
> + jmp .Locb_dec_done
> +
> +.align 16
> +.Locb_dec_three:
> + pxor %xmm5,%xmm5
> +
> + call __ocb_decrypt4
> +
> + movdqa %xmm12,%xmm15
> + movups %xmm2,0(%rsi)
> + xorps %xmm2,%xmm8
> + movups %xmm3,16(%rsi)
> + xorps %xmm3,%xmm8
> + movups %xmm4,32(%rsi)
> + xorps %xmm4,%xmm8
> +
> + jmp .Locb_dec_done
> +
> +.align 16
> +.Locb_dec_four:
> + call __ocb_decrypt4
> +
> + movdqa %xmm13,%xmm15
> + movups %xmm2,0(%rsi)
> + pxor %xmm2,%xmm8
> + movups %xmm3,16(%rsi)
> + pxor %xmm3,%xmm8
> + movups %xmm4,32(%rsi)
> + pxor %xmm4,%xmm8
> + movups %xmm5,48(%rsi)
> + pxor %xmm5,%xmm8
> +
> +.Locb_dec_done:
> + pxor %xmm0,%xmm15
> + movdqu %xmm8,(%rbp)
> + movdqu %xmm15,(%r9)
> +
> + xorps %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + pxor %xmm8,%xmm8
> + pxor %xmm9,%xmm9
> + pxor %xmm10,%xmm10
> + pxor %xmm11,%xmm11
> + pxor %xmm12,%xmm12
> + pxor %xmm13,%xmm13
> + pxor %xmm14,%xmm14
> + pxor %xmm15,%xmm15
> + leaq 40(%rsp),%rax
> +.cfi_def_cfa %rax,8
> + movq -40(%rax),%r14
> +.cfi_restore %r14
> + movq -32(%rax),%r13
> +.cfi_restore %r13
> + movq -24(%rax),%r12
> +.cfi_restore %r12
> + movq -16(%rax),%rbp
> +.cfi_restore %rbp
> + movq -8(%rax),%rbx
> +.cfi_restore %rbx
> + leaq (%rax),%rsp
> +.cfi_def_cfa_register %rsp
> +.Locb_dec_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_ocb_decrypt,.-aesni_ocb_decrypt
> +
> +.type __ocb_decrypt6, at function
> +.align 32
> +__ocb_decrypt6:
> +.cfi_startproc
> + pxor %xmm9,%xmm15
> + movdqu (%rbx,%r12,1),%xmm11
> + movdqa %xmm10,%xmm12
> + movdqu (%rbx,%r13,1),%xmm13
> + movdqa %xmm10,%xmm14
> + pxor %xmm15,%xmm10
> + movdqu (%rbx,%r14,1),%xmm15
> + pxor %xmm10,%xmm11
> + pxor %xmm10,%xmm2
> + pxor %xmm11,%xmm12
> + pxor %xmm11,%xmm3
> + pxor %xmm12,%xmm13
> + pxor %xmm12,%xmm4
> + pxor %xmm13,%xmm14
> + pxor %xmm13,%xmm5
> + pxor %xmm14,%xmm15
> + pxor %xmm14,%xmm6
> + pxor %xmm15,%xmm7
> + movups 32(%r11),%xmm0
> +
> + leaq 1(%r8),%r12
> + leaq 3(%r8),%r13
> + leaq 5(%r8),%r14
> + addq $6,%r8
> + pxor %xmm9,%xmm10
> + bsfq %r12,%r12
> + bsfq %r13,%r13
> + bsfq %r14,%r14
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + pxor %xmm9,%xmm11
> + pxor %xmm9,%xmm12
> +.byte 102,15,56,222,241
> + pxor %xmm9,%xmm13
> + pxor %xmm9,%xmm14
> +.byte 102,15,56,222,249
> + movups 48(%r11),%xmm1
> + pxor %xmm9,%xmm15
> +
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> + movups 64(%r11),%xmm0
> + shlq $4,%r12
> + shlq $4,%r13
> + jmp .Locb_dec_loop6
> +
> +.align 32
> +.Locb_dec_loop6:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Locb_dec_loop6
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> + movups 16(%r11),%xmm1
> + shlq $4,%r14
> +
> +.byte 102,65,15,56,223,210
> + movdqu (%rbx),%xmm10
> + movq %r10,%rax
> +.byte 102,65,15,56,223,219
> +.byte 102,65,15,56,223,228
> +.byte 102,65,15,56,223,237
> +.byte 102,65,15,56,223,246
> +.byte 102,65,15,56,223,255
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size __ocb_decrypt6,.-__ocb_decrypt6
> +
> +.type __ocb_decrypt4, at function
> +.align 32
> +__ocb_decrypt4:
> +.cfi_startproc
> + pxor %xmm9,%xmm15
> + movdqu (%rbx,%r12,1),%xmm11
> + movdqa %xmm10,%xmm12
> + movdqu (%rbx,%r13,1),%xmm13
> + pxor %xmm15,%xmm10
> + pxor %xmm10,%xmm11
> + pxor %xmm10,%xmm2
> + pxor %xmm11,%xmm12
> + pxor %xmm11,%xmm3
> + pxor %xmm12,%xmm13
> + pxor %xmm12,%xmm4
> + pxor %xmm13,%xmm5
> + movups 32(%r11),%xmm0
> +
> + pxor %xmm9,%xmm10
> + pxor %xmm9,%xmm11
> + pxor %xmm9,%xmm12
> + pxor %xmm9,%xmm13
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movups 48(%r11),%xmm1
> +
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + movups 64(%r11),%xmm0
> + jmp .Locb_dec_loop4
> +
> +.align 32
> +.Locb_dec_loop4:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Locb_dec_loop4
> +
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + movups 16(%r11),%xmm1
> + movq %r10,%rax
> +
> +.byte 102,65,15,56,223,210
> +.byte 102,65,15,56,223,219
> +.byte 102,65,15,56,223,228
> +.byte 102,65,15,56,223,237
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size __ocb_decrypt4,.-__ocb_decrypt4
> +
> +.type __ocb_decrypt1, at function
> +.align 32
> +__ocb_decrypt1:
> +.cfi_startproc
> + pxor %xmm15,%xmm7
> + pxor %xmm9,%xmm7
> + pxor %xmm7,%xmm2
> + movups 32(%r11),%xmm0
> +
> +.byte 102,15,56,222,209
> + movups 48(%r11),%xmm1
> + pxor %xmm9,%xmm7
> +
> +.byte 102,15,56,222,208
> + movups 64(%r11),%xmm0
> + jmp .Locb_dec_loop1
> +
> +.align 32
> +.Locb_dec_loop1:
> +.byte 102,15,56,222,209
> + movups (%rcx,%rax,1),%xmm1
> + addq $32,%rax
> +
> +.byte 102,15,56,222,208
> + movups -16(%rcx,%rax,1),%xmm0
> + jnz .Locb_dec_loop1
> +
> +.byte 102,15,56,222,209
> + movups 16(%r11),%xmm1
> + movq %r10,%rax
> +
> +.byte 102,15,56,223,215
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size __ocb_decrypt1,.-__ocb_decrypt1
> +.globl aesni_cbc_encrypt
> +.type aesni_cbc_encrypt, at function
> +.align 16
> +aesni_cbc_encrypt:
> +.cfi_startproc
> + testq %rdx,%rdx
> + jz .Lcbc_ret
> +
> + movl 240(%rcx),%r10d
> + movq %rcx,%r11
> + testl %r9d,%r9d
> + jz .Lcbc_decrypt
> +
> + movups (%r8),%xmm2
> + movl %r10d,%eax
> + cmpq $16,%rdx
> + jb .Lcbc_enc_tail
> + subq $16,%rdx
> + jmp .Lcbc_enc_loop
> +.align 16
> +.Lcbc_enc_loop:
> + movups (%rdi),%xmm3
> + leaq 16(%rdi),%rdi
> +
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + xorps %xmm0,%xmm3
> + leaq 32(%rcx),%rcx
> + xorps %xmm3,%xmm2
> +.Loop_enc1_15:
> +.byte 102,15,56,220,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_enc1_15
> +.byte 102,15,56,221,209
> + movl %r10d,%eax
> + movq %r11,%rcx
> + movups %xmm2,0(%rsi)
> + leaq 16(%rsi),%rsi
> + subq $16,%rdx
> + jnc .Lcbc_enc_loop
> + addq $16,%rdx
> + jnz .Lcbc_enc_tail
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + movups %xmm2,(%r8)
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + jmp .Lcbc_ret
> +
> +.Lcbc_enc_tail:
> + movq %rdx,%rcx
> + xchgq %rdi,%rsi
> +.long 0x9066A4F3
> + movl $16,%ecx
> + subq %rdx,%rcx
> + xorl %eax,%eax
> +.long 0x9066AAF3
> + leaq -16(%rdi),%rdi
> + movl %r10d,%eax
> + movq %rdi,%rsi
> + movq %r11,%rcx
> + xorq %rdx,%rdx
> + jmp .Lcbc_enc_loop
> +
> +.align 16
> +.Lcbc_decrypt:
> + cmpq $16,%rdx
> + jne .Lcbc_decrypt_bulk
> +
> +
> +
> + movdqu (%rdi),%xmm2
> + movdqu (%r8),%xmm3
> + movdqa %xmm2,%xmm4
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_dec1_16:
> +.byte 102,15,56,222,209
> + decl %r10d
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_dec1_16
> +.byte 102,15,56,223,209
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + movdqu %xmm4,(%r8)
> + xorps %xmm3,%xmm2
> + pxor %xmm3,%xmm3
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + jmp .Lcbc_ret
> +.align 16
> +.Lcbc_decrypt_bulk:
> + leaq (%rsp),%r11
> +.cfi_def_cfa_register %r11
> + pushq %rbp
> +.cfi_offset %rbp,-16
> + subq $16,%rsp
> + andq $-16,%rsp
> + movq %rcx,%rbp
> + movups (%r8),%xmm10
> + movl %r10d,%eax
> + cmpq $0x50,%rdx
> + jbe .Lcbc_dec_tail
> +
> + movups (%rcx),%xmm0
> + movdqu 0(%rdi),%xmm2
> + movdqu 16(%rdi),%xmm3
> + movdqa %xmm2,%xmm11
> + movdqu 32(%rdi),%xmm4
> + movdqa %xmm3,%xmm12
> + movdqu 48(%rdi),%xmm5
> + movdqa %xmm4,%xmm13
> + movdqu 64(%rdi),%xmm6
> + movdqa %xmm5,%xmm14
> + movdqu 80(%rdi),%xmm7
> + movdqa %xmm6,%xmm15
> + movl OPENSSL_ia32cap_P+4(%rip),%r9d
> + cmpq $0x70,%rdx
> + jbe .Lcbc_dec_six_or_seven
> +
> + andl $71303168,%r9d
> + subq $0x50,%rdx
> + cmpl $4194304,%r9d
> + je .Lcbc_dec_loop6_enter
> + subq $0x20,%rdx
> + leaq 112(%rcx),%rcx
> + jmp .Lcbc_dec_loop8_enter
> +.align 16
> +.Lcbc_dec_loop8:
> + movups %xmm9,(%rsi)
> + leaq 16(%rsi),%rsi
> +.Lcbc_dec_loop8_enter:
> + movdqu 96(%rdi),%xmm8
> + pxor %xmm0,%xmm2
> + movdqu 112(%rdi),%xmm9
> + pxor %xmm0,%xmm3
> + movups 16-112(%rcx),%xmm1
> + pxor %xmm0,%xmm4
> + movq $-1,%rbp
> + cmpq $0x70,%rdx
> + pxor %xmm0,%xmm5
> + pxor %xmm0,%xmm6
> + pxor %xmm0,%xmm7
> + pxor %xmm0,%xmm8
> +
> +.byte 102,15,56,222,209
> + pxor %xmm0,%xmm9
> + movups 32-112(%rcx),%xmm0
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,68,15,56,222,193
> + adcq $0,%rbp
> + andq $128,%rbp
> +.byte 102,68,15,56,222,201
> + addq %rdi,%rbp
> + movups 48-112(%rcx),%xmm1
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> +.byte 102,68,15,56,222,192
> +.byte 102,68,15,56,222,200
> + movups 64-112(%rcx),%xmm0
> + nop
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,68,15,56,222,193
> +.byte 102,68,15,56,222,201
> + movups 80-112(%rcx),%xmm1
> + nop
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> +.byte 102,68,15,56,222,192
> +.byte 102,68,15,56,222,200
> + movups 96-112(%rcx),%xmm0
> + nop
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,68,15,56,222,193
> +.byte 102,68,15,56,222,201
> + movups 112-112(%rcx),%xmm1
> + nop
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> +.byte 102,68,15,56,222,192
> +.byte 102,68,15,56,222,200
> + movups 128-112(%rcx),%xmm0
> + nop
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,68,15,56,222,193
> +.byte 102,68,15,56,222,201
> + movups 144-112(%rcx),%xmm1
> + cmpl $11,%eax
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> +.byte 102,68,15,56,222,192
> +.byte 102,68,15,56,222,200
> + movups 160-112(%rcx),%xmm0
> + jb .Lcbc_dec_done
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,68,15,56,222,193
> +.byte 102,68,15,56,222,201
> + movups 176-112(%rcx),%xmm1
> + nop
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> +.byte 102,68,15,56,222,192
> +.byte 102,68,15,56,222,200
> + movups 192-112(%rcx),%xmm0
> + je .Lcbc_dec_done
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> +.byte 102,68,15,56,222,193
> +.byte 102,68,15,56,222,201
> + movups 208-112(%rcx),%xmm1
> + nop
> +.byte 102,15,56,222,208
> +.byte 102,15,56,222,216
> +.byte 102,15,56,222,224
> +.byte 102,15,56,222,232
> +.byte 102,15,56,222,240
> +.byte 102,15,56,222,248
> +.byte 102,68,15,56,222,192
> +.byte 102,68,15,56,222,200
> + movups 224-112(%rcx),%xmm0
> + jmp .Lcbc_dec_done
> +.align 16
> +.Lcbc_dec_done:
> +.byte 102,15,56,222,209
> +.byte 102,15,56,222,217
> + pxor %xmm0,%xmm10
> + pxor %xmm0,%xmm11
> +.byte 102,15,56,222,225
> +.byte 102,15,56,222,233
> + pxor %xmm0,%xmm12
> + pxor %xmm0,%xmm13
> +.byte 102,15,56,222,241
> +.byte 102,15,56,222,249
> + pxor %xmm0,%xmm14
> + pxor %xmm0,%xmm15
> +.byte 102,68,15,56,222,193
> +.byte 102,68,15,56,222,201
> + movdqu 80(%rdi),%xmm1
> +
> +.byte 102,65,15,56,223,210
> + movdqu 96(%rdi),%xmm10
> + pxor %xmm0,%xmm1
> +.byte 102,65,15,56,223,219
> + pxor %xmm0,%xmm10
> + movdqu 112(%rdi),%xmm0
> +.byte 102,65,15,56,223,228
> + leaq 128(%rdi),%rdi
> + movdqu 0(%rbp),%xmm11
> +.byte 102,65,15,56,223,237
> +.byte 102,65,15,56,223,246
> + movdqu 16(%rbp),%xmm12
> + movdqu 32(%rbp),%xmm13
> +.byte 102,65,15,56,223,255
> +.byte 102,68,15,56,223,193
> + movdqu 48(%rbp),%xmm14
> + movdqu 64(%rbp),%xmm15
> +.byte 102,69,15,56,223,202
> + movdqa %xmm0,%xmm10
> + movdqu 80(%rbp),%xmm1
> + movups -112(%rcx),%xmm0
> +
> + movups %xmm2,(%rsi)
> + movdqa %xmm11,%xmm2
> + movups %xmm3,16(%rsi)
> + movdqa %xmm12,%xmm3
> + movups %xmm4,32(%rsi)
> + movdqa %xmm13,%xmm4
> + movups %xmm5,48(%rsi)
> + movdqa %xmm14,%xmm5
> + movups %xmm6,64(%rsi)
> + movdqa %xmm15,%xmm6
> + movups %xmm7,80(%rsi)
> + movdqa %xmm1,%xmm7
> + movups %xmm8,96(%rsi)
> + leaq 112(%rsi),%rsi
> +
> + subq $0x80,%rdx
> + ja .Lcbc_dec_loop8
> +
> + movaps %xmm9,%xmm2
> + leaq -112(%rcx),%rcx
> + addq $0x70,%rdx
> + jle .Lcbc_dec_clear_tail_collected
> + movups %xmm9,(%rsi)
> + leaq 16(%rsi),%rsi
> + cmpq $0x50,%rdx
> + jbe .Lcbc_dec_tail
> +
> + movaps %xmm11,%xmm2
> +.Lcbc_dec_six_or_seven:
> + cmpq $0x60,%rdx
> + ja .Lcbc_dec_seven
> +
> + movaps %xmm7,%xmm8
> + call _aesni_decrypt6
> + pxor %xmm10,%xmm2
> + movaps %xmm8,%xmm10
> + pxor %xmm11,%xmm3
> + movdqu %xmm2,(%rsi)
> + pxor %xmm12,%xmm4
> + movdqu %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + pxor %xmm13,%xmm5
> + movdqu %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + pxor %xmm14,%xmm6
> + movdqu %xmm5,48(%rsi)
> + pxor %xmm5,%xmm5
> + pxor %xmm15,%xmm7
> + movdqu %xmm6,64(%rsi)
> + pxor %xmm6,%xmm6
> + leaq 80(%rsi),%rsi
> + movdqa %xmm7,%xmm2
> + pxor %xmm7,%xmm7
> + jmp .Lcbc_dec_tail_collected
> +
> +.align 16
> +.Lcbc_dec_seven:
> + movups 96(%rdi),%xmm8
> + xorps %xmm9,%xmm9
> + call _aesni_decrypt8
> + movups 80(%rdi),%xmm9
> + pxor %xmm10,%xmm2
> + movups 96(%rdi),%xmm10
> + pxor %xmm11,%xmm3
> + movdqu %xmm2,(%rsi)
> + pxor %xmm12,%xmm4
> + movdqu %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + pxor %xmm13,%xmm5
> + movdqu %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + pxor %xmm14,%xmm6
> + movdqu %xmm5,48(%rsi)
> + pxor %xmm5,%xmm5
> + pxor %xmm15,%xmm7
> + movdqu %xmm6,64(%rsi)
> + pxor %xmm6,%xmm6
> + pxor %xmm9,%xmm8
> + movdqu %xmm7,80(%rsi)
> + pxor %xmm7,%xmm7
> + leaq 96(%rsi),%rsi
> + movdqa %xmm8,%xmm2
> + pxor %xmm8,%xmm8
> + pxor %xmm9,%xmm9
> + jmp .Lcbc_dec_tail_collected
> +
> +.align 16
> +.Lcbc_dec_loop6:
> + movups %xmm7,(%rsi)
> + leaq 16(%rsi),%rsi
> + movdqu 0(%rdi),%xmm2
> + movdqu 16(%rdi),%xmm3
> + movdqa %xmm2,%xmm11
> + movdqu 32(%rdi),%xmm4
> + movdqa %xmm3,%xmm12
> + movdqu 48(%rdi),%xmm5
> + movdqa %xmm4,%xmm13
> + movdqu 64(%rdi),%xmm6
> + movdqa %xmm5,%xmm14
> + movdqu 80(%rdi),%xmm7
> + movdqa %xmm6,%xmm15
> +.Lcbc_dec_loop6_enter:
> + leaq 96(%rdi),%rdi
> + movdqa %xmm7,%xmm8
> +
> + call _aesni_decrypt6
> +
> + pxor %xmm10,%xmm2
> + movdqa %xmm8,%xmm10
> + pxor %xmm11,%xmm3
> + movdqu %xmm2,(%rsi)
> + pxor %xmm12,%xmm4
> + movdqu %xmm3,16(%rsi)
> + pxor %xmm13,%xmm5
> + movdqu %xmm4,32(%rsi)
> + pxor %xmm14,%xmm6
> + movq %rbp,%rcx
> + movdqu %xmm5,48(%rsi)
> + pxor %xmm15,%xmm7
> + movl %r10d,%eax
> + movdqu %xmm6,64(%rsi)
> + leaq 80(%rsi),%rsi
> + subq $0x60,%rdx
> + ja .Lcbc_dec_loop6
> +
> + movdqa %xmm7,%xmm2
> + addq $0x50,%rdx
> + jle .Lcbc_dec_clear_tail_collected
> + movups %xmm7,(%rsi)
> + leaq 16(%rsi),%rsi
> +
> +.Lcbc_dec_tail:
> + movups (%rdi),%xmm2
> + subq $0x10,%rdx
> + jbe .Lcbc_dec_one
> +
> + movups 16(%rdi),%xmm3
> + movaps %xmm2,%xmm11
> + subq $0x10,%rdx
> + jbe .Lcbc_dec_two
> +
> + movups 32(%rdi),%xmm4
> + movaps %xmm3,%xmm12
> + subq $0x10,%rdx
> + jbe .Lcbc_dec_three
> +
> + movups 48(%rdi),%xmm5
> + movaps %xmm4,%xmm13
> + subq $0x10,%rdx
> + jbe .Lcbc_dec_four
> +
> + movups 64(%rdi),%xmm6
> + movaps %xmm5,%xmm14
> + movaps %xmm6,%xmm15
> + xorps %xmm7,%xmm7
> + call _aesni_decrypt6
> + pxor %xmm10,%xmm2
> + movaps %xmm15,%xmm10
> + pxor %xmm11,%xmm3
> + movdqu %xmm2,(%rsi)
> + pxor %xmm12,%xmm4
> + movdqu %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + pxor %xmm13,%xmm5
> + movdqu %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + pxor %xmm14,%xmm6
> + movdqu %xmm5,48(%rsi)
> + pxor %xmm5,%xmm5
> + leaq 64(%rsi),%rsi
> + movdqa %xmm6,%xmm2
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + subq $0x10,%rdx
> + jmp .Lcbc_dec_tail_collected
> +
> +.align 16
> +.Lcbc_dec_one:
> + movaps %xmm2,%xmm11
> + movups (%rcx),%xmm0
> + movups 16(%rcx),%xmm1
> + leaq 32(%rcx),%rcx
> + xorps %xmm0,%xmm2
> +.Loop_dec1_17:
> +.byte 102,15,56,222,209
> + decl %eax
> + movups (%rcx),%xmm1
> + leaq 16(%rcx),%rcx
> + jnz .Loop_dec1_17
> +.byte 102,15,56,223,209
> + xorps %xmm10,%xmm2
> + movaps %xmm11,%xmm10
> + jmp .Lcbc_dec_tail_collected
> +.align 16
> +.Lcbc_dec_two:
> + movaps %xmm3,%xmm12
> + call _aesni_decrypt2
> + pxor %xmm10,%xmm2
> + movaps %xmm12,%xmm10
> + pxor %xmm11,%xmm3
> + movdqu %xmm2,(%rsi)
> + movdqa %xmm3,%xmm2
> + pxor %xmm3,%xmm3
> + leaq 16(%rsi),%rsi
> + jmp .Lcbc_dec_tail_collected
> +.align 16
> +.Lcbc_dec_three:
> + movaps %xmm4,%xmm13
> + call _aesni_decrypt3
> + pxor %xmm10,%xmm2
> + movaps %xmm13,%xmm10
> + pxor %xmm11,%xmm3
> + movdqu %xmm2,(%rsi)
> + pxor %xmm12,%xmm4
> + movdqu %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + movdqa %xmm4,%xmm2
> + pxor %xmm4,%xmm4
> + leaq 32(%rsi),%rsi
> + jmp .Lcbc_dec_tail_collected
> +.align 16
> +.Lcbc_dec_four:
> + movaps %xmm5,%xmm14
> + call _aesni_decrypt4
> + pxor %xmm10,%xmm2
> + movaps %xmm14,%xmm10
> + pxor %xmm11,%xmm3
> + movdqu %xmm2,(%rsi)
> + pxor %xmm12,%xmm4
> + movdqu %xmm3,16(%rsi)
> + pxor %xmm3,%xmm3
> + pxor %xmm13,%xmm5
> + movdqu %xmm4,32(%rsi)
> + pxor %xmm4,%xmm4
> + movdqa %xmm5,%xmm2
> + pxor %xmm5,%xmm5
> + leaq 48(%rsi),%rsi
> + jmp .Lcbc_dec_tail_collected
> +
> +.align 16
> +.Lcbc_dec_clear_tail_collected:
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + pxor %xmm8,%xmm8
> + pxor %xmm9,%xmm9
> +.Lcbc_dec_tail_collected:
> + movups %xmm10,(%r8)
> + andq $15,%rdx
> + jnz .Lcbc_dec_tail_partial
> + movups %xmm2,(%rsi)
> + pxor %xmm2,%xmm2
> + jmp .Lcbc_dec_ret
> +.align 16
> +.Lcbc_dec_tail_partial:
> + movaps %xmm2,(%rsp)
> + pxor %xmm2,%xmm2
> + movq $16,%rcx
> + movq %rsi,%rdi
> + subq %rdx,%rcx
> + leaq (%rsp),%rsi
> +.long 0x9066A4F3
> + movdqa %xmm2,(%rsp)
> +
> +.Lcbc_dec_ret:
> + xorps %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + movq -8(%r11),%rbp
> +.cfi_restore %rbp
> + leaq (%r11),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lcbc_ret:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_cbc_encrypt,.-aesni_cbc_encrypt
> +.globl aesni_set_decrypt_key
> +.type aesni_set_decrypt_key, at function
> +.align 16
> +aesni_set_decrypt_key:
> +.cfi_startproc
> +.byte 0x48,0x83,0xEC,0x08
> +.cfi_adjust_cfa_offset 8
> + call __aesni_set_encrypt_key
> + shll $4,%esi
> + testl %eax,%eax
> + jnz .Ldec_key_ret
> + leaq 16(%rdx,%rsi,1),%rdi
> +
> + movups (%rdx),%xmm0
> + movups (%rdi),%xmm1
> + movups %xmm0,(%rdi)
> + movups %xmm1,(%rdx)
> + leaq 16(%rdx),%rdx
> + leaq -16(%rdi),%rdi
> +
> +.Ldec_key_inverse:
> + movups (%rdx),%xmm0
> + movups (%rdi),%xmm1
> +.byte 102,15,56,219,192
> +.byte 102,15,56,219,201
> + leaq 16(%rdx),%rdx
> + leaq -16(%rdi),%rdi
> + movups %xmm0,16(%rdi)
> + movups %xmm1,-16(%rdx)
> + cmpq %rdx,%rdi
> + ja .Ldec_key_inverse
> +
> + movups (%rdx),%xmm0
> +.byte 102,15,56,219,192
> + pxor %xmm1,%xmm1
> + movups %xmm0,(%rdi)
> + pxor %xmm0,%xmm0
> +.Ldec_key_ret:
> + addq $8,%rsp
> +.cfi_adjust_cfa_offset -8
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.LSEH_end_set_decrypt_key:
> +.size aesni_set_decrypt_key,.-aesni_set_decrypt_key
> +.globl aesni_set_encrypt_key
> +.type aesni_set_encrypt_key, at function
> +.align 16
> +aesni_set_encrypt_key:
> +__aesni_set_encrypt_key:
> +.cfi_startproc
> +.byte 0x48,0x83,0xEC,0x08
> +.cfi_adjust_cfa_offset 8
> + movq $-1,%rax
> + testq %rdi,%rdi
> + jz .Lenc_key_ret
> + testq %rdx,%rdx
> + jz .Lenc_key_ret
> +
> + movl $268437504,%r10d
> + movups (%rdi),%xmm0
> + xorps %xmm4,%xmm4
> + andl OPENSSL_ia32cap_P+4(%rip),%r10d
> + leaq 16(%rdx),%rax
> + cmpl $256,%esi
> + je .L14rounds
> + cmpl $192,%esi
> + je .L12rounds
> + cmpl $128,%esi
> + jne .Lbad_keybits
> +
> +.L10rounds:
> + movl $9,%esi
> + cmpl $268435456,%r10d
> + je .L10rounds_alt
> +
> + movups %xmm0,(%rdx)
> +.byte 102,15,58,223,200,1
> + call .Lkey_expansion_128_cold
> +.byte 102,15,58,223,200,2
> + call .Lkey_expansion_128
> +.byte 102,15,58,223,200,4
> + call .Lkey_expansion_128
> +.byte 102,15,58,223,200,8
> + call .Lkey_expansion_128
> +.byte 102,15,58,223,200,16
> + call .Lkey_expansion_128
> +.byte 102,15,58,223,200,32
> + call .Lkey_expansion_128
> +.byte 102,15,58,223,200,64
> + call .Lkey_expansion_128
> +.byte 102,15,58,223,200,128
> + call .Lkey_expansion_128
> +.byte 102,15,58,223,200,27
> + call .Lkey_expansion_128
> +.byte 102,15,58,223,200,54
> + call .Lkey_expansion_128
> + movups %xmm0,(%rax)
> + movl %esi,80(%rax)
> + xorl %eax,%eax
> + jmp .Lenc_key_ret
> +
> +.align 16
> +.L10rounds_alt:
> + movdqa .Lkey_rotate(%rip),%xmm5
> + movl $8,%r10d
> + movdqa .Lkey_rcon1(%rip),%xmm4
> + movdqa %xmm0,%xmm2
> + movdqu %xmm0,(%rdx)
> + jmp .Loop_key128
> +
> +.align 16
> +.Loop_key128:
> +.byte 102,15,56,0,197
> +.byte 102,15,56,221,196
> + pslld $1,%xmm4
> + leaq 16(%rax),%rax
> +
> + movdqa %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm3,%xmm2
> +
> + pxor %xmm2,%xmm0
> + movdqu %xmm0,-16(%rax)
> + movdqa %xmm0,%xmm2
> +
> + decl %r10d
> + jnz .Loop_key128
> +
> + movdqa .Lkey_rcon1b(%rip),%xmm4
> +
> +.byte 102,15,56,0,197
> +.byte 102,15,56,221,196
> + pslld $1,%xmm4
> +
> + movdqa %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm3,%xmm2
> +
> + pxor %xmm2,%xmm0
> + movdqu %xmm0,(%rax)
> +
> + movdqa %xmm0,%xmm2
> +.byte 102,15,56,0,197
> +.byte 102,15,56,221,196
> +
> + movdqa %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm2,%xmm3
> + pslldq $4,%xmm2
> + pxor %xmm3,%xmm2
> +
> + pxor %xmm2,%xmm0
> + movdqu %xmm0,16(%rax)
> +
> + movl %esi,96(%rax)
> + xorl %eax,%eax
> + jmp .Lenc_key_ret
> +
> +.align 16
> +.L12rounds:
> + movq 16(%rdi),%xmm2
> + movl $11,%esi
> + cmpl $268435456,%r10d
> + je .L12rounds_alt
> +
> + movups %xmm0,(%rdx)
> +.byte 102,15,58,223,202,1
> + call .Lkey_expansion_192a_cold
> +.byte 102,15,58,223,202,2
> + call .Lkey_expansion_192b
> +.byte 102,15,58,223,202,4
> + call .Lkey_expansion_192a
> +.byte 102,15,58,223,202,8
> + call .Lkey_expansion_192b
> +.byte 102,15,58,223,202,16
> + call .Lkey_expansion_192a
> +.byte 102,15,58,223,202,32
> + call .Lkey_expansion_192b
> +.byte 102,15,58,223,202,64
> + call .Lkey_expansion_192a
> +.byte 102,15,58,223,202,128
> + call .Lkey_expansion_192b
> + movups %xmm0,(%rax)
> + movl %esi,48(%rax)
> + xorq %rax,%rax
> + jmp .Lenc_key_ret
> +
> +.align 16
> +.L12rounds_alt:
> + movdqa .Lkey_rotate192(%rip),%xmm5
> + movdqa .Lkey_rcon1(%rip),%xmm4
> + movl $8,%r10d
> + movdqu %xmm0,(%rdx)
> + jmp .Loop_key192
> +
> +.align 16
> +.Loop_key192:
> + movq %xmm2,0(%rax)
> + movdqa %xmm2,%xmm1
> +.byte 102,15,56,0,213
> +.byte 102,15,56,221,212
> + pslld $1,%xmm4
> + leaq 24(%rax),%rax
> +
> + movdqa %xmm0,%xmm3
> + pslldq $4,%xmm0
> + pxor %xmm0,%xmm3
> + pslldq $4,%xmm0
> + pxor %xmm0,%xmm3
> + pslldq $4,%xmm0
> + pxor %xmm3,%xmm0
> +
> + pshufd $0xff,%xmm0,%xmm3
> + pxor %xmm1,%xmm3
> + pslldq $4,%xmm1
> + pxor %xmm1,%xmm3
> +
> + pxor %xmm2,%xmm0
> + pxor %xmm3,%xmm2
> + movdqu %xmm0,-16(%rax)
> +
> + decl %r10d
> + jnz .Loop_key192
> +
> + movl %esi,32(%rax)
> + xorl %eax,%eax
> + jmp .Lenc_key_ret
> +
> +.align 16
> +.L14rounds:
> + movups 16(%rdi),%xmm2
> + movl $13,%esi
> + leaq 16(%rax),%rax
> + cmpl $268435456,%r10d
> + je .L14rounds_alt
> +
> + movups %xmm0,(%rdx)
> + movups %xmm2,16(%rdx)
> +.byte 102,15,58,223,202,1
> + call .Lkey_expansion_256a_cold
> +.byte 102,15,58,223,200,1
> + call .Lkey_expansion_256b
> +.byte 102,15,58,223,202,2
> + call .Lkey_expansion_256a
> +.byte 102,15,58,223,200,2
> + call .Lkey_expansion_256b
> +.byte 102,15,58,223,202,4
> + call .Lkey_expansion_256a
> +.byte 102,15,58,223,200,4
> + call .Lkey_expansion_256b
> +.byte 102,15,58,223,202,8
> + call .Lkey_expansion_256a
> +.byte 102,15,58,223,200,8
> + call .Lkey_expansion_256b
> +.byte 102,15,58,223,202,16
> + call .Lkey_expansion_256a
> +.byte 102,15,58,223,200,16
> + call .Lkey_expansion_256b
> +.byte 102,15,58,223,202,32
> + call .Lkey_expansion_256a
> +.byte 102,15,58,223,200,32
> + call .Lkey_expansion_256b
> +.byte 102,15,58,223,202,64
> + call .Lkey_expansion_256a
> + movups %xmm0,(%rax)
> + movl %esi,16(%rax)
> + xorq %rax,%rax
> + jmp .Lenc_key_ret
> +
> +.align 16
> +.L14rounds_alt:
> + movdqa .Lkey_rotate(%rip),%xmm5
> + movdqa .Lkey_rcon1(%rip),%xmm4
> + movl $7,%r10d
> + movdqu %xmm0,0(%rdx)
> + movdqa %xmm2,%xmm1
> + movdqu %xmm2,16(%rdx)
> + jmp .Loop_key256
> +
> +.align 16
> +.Loop_key256:
> +.byte 102,15,56,0,213
> +.byte 102,15,56,221,212
> +
> + movdqa %xmm0,%xmm3
> + pslldq $4,%xmm0
> + pxor %xmm0,%xmm3
> + pslldq $4,%xmm0
> + pxor %xmm0,%xmm3
> + pslldq $4,%xmm0
> + pxor %xmm3,%xmm0
> + pslld $1,%xmm4
> +
> + pxor %xmm2,%xmm0
> + movdqu %xmm0,(%rax)
> +
> + decl %r10d
> + jz .Ldone_key256
> +
> + pshufd $0xff,%xmm0,%xmm2
> + pxor %xmm3,%xmm3
> +.byte 102,15,56,221,211
> +
> + movdqa %xmm1,%xmm3
> + pslldq $4,%xmm1
> + pxor %xmm1,%xmm3
> + pslldq $4,%xmm1
> + pxor %xmm1,%xmm3
> + pslldq $4,%xmm1
> + pxor %xmm3,%xmm1
> +
> + pxor %xmm1,%xmm2
> + movdqu %xmm2,16(%rax)
> + leaq 32(%rax),%rax
> + movdqa %xmm2,%xmm1
> +
> + jmp .Loop_key256
> +
> +.Ldone_key256:
> + movl %esi,16(%rax)
> + xorl %eax,%eax
> + jmp .Lenc_key_ret
> +
> +.align 16
> +.Lbad_keybits:
> + movq $-2,%rax
> +.Lenc_key_ret:
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + addq $8,%rsp
> +.cfi_adjust_cfa_offset -8
> + .byte 0xf3,0xc3
> +.LSEH_end_set_encrypt_key:
> +
> +.align 16
> +.Lkey_expansion_128:
> + movups %xmm0,(%rax)
> + leaq 16(%rax),%rax
> +.Lkey_expansion_128_cold:
> + shufps $16,%xmm0,%xmm4
> + xorps %xmm4,%xmm0
> + shufps $140,%xmm0,%xmm4
> + xorps %xmm4,%xmm0
> + shufps $255,%xmm1,%xmm1
> + xorps %xmm1,%xmm0
> + .byte 0xf3,0xc3
> +
> +.align 16
> +.Lkey_expansion_192a:
> + movups %xmm0,(%rax)
> + leaq 16(%rax),%rax
> +.Lkey_expansion_192a_cold:
> + movaps %xmm2,%xmm5
> +.Lkey_expansion_192b_warm:
> + shufps $16,%xmm0,%xmm4
> + movdqa %xmm2,%xmm3
> + xorps %xmm4,%xmm0
> + shufps $140,%xmm0,%xmm4
> + pslldq $4,%xmm3
> + xorps %xmm4,%xmm0
> + pshufd $85,%xmm1,%xmm1
> + pxor %xmm3,%xmm2
> + pxor %xmm1,%xmm0
> + pshufd $255,%xmm0,%xmm3
> + pxor %xmm3,%xmm2
> + .byte 0xf3,0xc3
> +
> +.align 16
> +.Lkey_expansion_192b:
> + movaps %xmm0,%xmm3
> + shufps $68,%xmm0,%xmm5
> + movups %xmm5,(%rax)
> + shufps $78,%xmm2,%xmm3
> + movups %xmm3,16(%rax)
> + leaq 32(%rax),%rax
> + jmp .Lkey_expansion_192b_warm
> +
> +.align 16
> +.Lkey_expansion_256a:
> + movups %xmm2,(%rax)
> + leaq 16(%rax),%rax
> +.Lkey_expansion_256a_cold:
> + shufps $16,%xmm0,%xmm4
> + xorps %xmm4,%xmm0
> + shufps $140,%xmm0,%xmm4
> + xorps %xmm4,%xmm0
> + shufps $255,%xmm1,%xmm1
> + xorps %xmm1,%xmm0
> + .byte 0xf3,0xc3
> +
> +.align 16
> +.Lkey_expansion_256b:
> + movups %xmm0,(%rax)
> + leaq 16(%rax),%rax
> +
> + shufps $16,%xmm2,%xmm4
> + xorps %xmm4,%xmm2
> + shufps $140,%xmm2,%xmm4
> + xorps %xmm4,%xmm2
> + shufps $170,%xmm1,%xmm1
> + xorps %xmm1,%xmm2
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_set_encrypt_key,.-aesni_set_encrypt_key
> +.size __aesni_set_encrypt_key,.-__aesni_set_encrypt_key
> +.align 64
> +.Lbswap_mask:
> +.byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
> +.Lincrement32:
> +.long 6,6,6,0
> +.Lincrement64:
> +.long 1,0,0,0
> +.Lxts_magic:
> +.long 0x87,0,1,0
> +.Lincrement1:
> +.byte 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1
> +.Lkey_rotate:
> +.long 0x0c0f0e0d,0x0c0f0e0d,0x0c0f0e0d,0x0c0f0e0d
> +.Lkey_rotate192:
> +.long 0x04070605,0x04070605,0x04070605,0x04070605
> +.Lkey_rcon1:
> +.long 1,1,1,1
> +.Lkey_rcon1b:
> +.long 0x1b,0x1b,0x1b,0x1b
> +
> +.byte
> 65,69,83,32,102,111,114,32,73,110,116,101,108,32,65,69,83,45,78,73,44,32,67
> ,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,1
> 01,110,115,115,108,46,111,114,103,62,0
> +.align 64
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/vpaes-x86_64.S
> b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/vpaes-x86_64.S
> new file mode 100644
> index 0000000000..982818f83b
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/aes/vpaes-x86_64.S
> @@ -0,0 +1,863 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/aes/asm/vpaes-x86_64.pl
> +#
> +# Copyright 2011-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.type _vpaes_encrypt_core, at function
> +.align 16
> +_vpaes_encrypt_core:
> +.cfi_startproc
> + movq %rdx,%r9
> + movq $16,%r11
> + movl 240(%rdx),%eax
> + movdqa %xmm9,%xmm1
> + movdqa .Lk_ipt(%rip),%xmm2
> + pandn %xmm0,%xmm1
> + movdqu (%r9),%xmm5
> + psrld $4,%xmm1
> + pand %xmm9,%xmm0
> +.byte 102,15,56,0,208
> + movdqa .Lk_ipt+16(%rip),%xmm0
> +.byte 102,15,56,0,193
> + pxor %xmm5,%xmm2
> + addq $16,%r9
> + pxor %xmm2,%xmm0
> + leaq .Lk_mc_backward(%rip),%r10
> + jmp .Lenc_entry
> +
> +.align 16
> +.Lenc_loop:
> +
> + movdqa %xmm13,%xmm4
> + movdqa %xmm12,%xmm0
> +.byte 102,15,56,0,226
> +.byte 102,15,56,0,195
> + pxor %xmm5,%xmm4
> + movdqa %xmm15,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa -64(%r11,%r10,1),%xmm1
> +.byte 102,15,56,0,234
> + movdqa (%r11,%r10,1),%xmm4
> + movdqa %xmm14,%xmm2
> +.byte 102,15,56,0,211
> + movdqa %xmm0,%xmm3
> + pxor %xmm5,%xmm2
> +.byte 102,15,56,0,193
> + addq $16,%r9
> + pxor %xmm2,%xmm0
> +.byte 102,15,56,0,220
> + addq $16,%r11
> + pxor %xmm0,%xmm3
> +.byte 102,15,56,0,193
> + andq $0x30,%r11
> + subq $1,%rax
> + pxor %xmm3,%xmm0
> +
> +.Lenc_entry:
> +
> + movdqa %xmm9,%xmm1
> + movdqa %xmm11,%xmm5
> + pandn %xmm0,%xmm1
> + psrld $4,%xmm1
> + pand %xmm9,%xmm0
> +.byte 102,15,56,0,232
> + movdqa %xmm10,%xmm3
> + pxor %xmm1,%xmm0
> +.byte 102,15,56,0,217
> + movdqa %xmm10,%xmm4
> + pxor %xmm5,%xmm3
> +.byte 102,15,56,0,224
> + movdqa %xmm10,%xmm2
> + pxor %xmm5,%xmm4
> +.byte 102,15,56,0,211
> + movdqa %xmm10,%xmm3
> + pxor %xmm0,%xmm2
> +.byte 102,15,56,0,220
> + movdqu (%r9),%xmm5
> + pxor %xmm1,%xmm3
> + jnz .Lenc_loop
> +
> +
> + movdqa -96(%r10),%xmm4
> + movdqa -80(%r10),%xmm0
> +.byte 102,15,56,0,226
> + pxor %xmm5,%xmm4
> +.byte 102,15,56,0,195
> + movdqa 64(%r11,%r10,1),%xmm1
> + pxor %xmm4,%xmm0
> +.byte 102,15,56,0,193
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _vpaes_encrypt_core,.-_vpaes_encrypt_core
> +
> +
> +
> +
> +
> +
> +.type _vpaes_decrypt_core, at function
> +.align 16
> +_vpaes_decrypt_core:
> +.cfi_startproc
> + movq %rdx,%r9
> + movl 240(%rdx),%eax
> + movdqa %xmm9,%xmm1
> + movdqa .Lk_dipt(%rip),%xmm2
> + pandn %xmm0,%xmm1
> + movq %rax,%r11
> + psrld $4,%xmm1
> + movdqu (%r9),%xmm5
> + shlq $4,%r11
> + pand %xmm9,%xmm0
> +.byte 102,15,56,0,208
> + movdqa .Lk_dipt+16(%rip),%xmm0
> + xorq $0x30,%r11
> + leaq .Lk_dsbd(%rip),%r10
> +.byte 102,15,56,0,193
> + andq $0x30,%r11
> + pxor %xmm5,%xmm2
> + movdqa .Lk_mc_forward+48(%rip),%xmm5
> + pxor %xmm2,%xmm0
> + addq $16,%r9
> + addq %r10,%r11
> + jmp .Ldec_entry
> +
> +.align 16
> +.Ldec_loop:
> +
> +
> +
> + movdqa -32(%r10),%xmm4
> + movdqa -16(%r10),%xmm1
> +.byte 102,15,56,0,226
> +.byte 102,15,56,0,203
> + pxor %xmm4,%xmm0
> + movdqa 0(%r10),%xmm4
> + pxor %xmm1,%xmm0
> + movdqa 16(%r10),%xmm1
> +
> +.byte 102,15,56,0,226
> +.byte 102,15,56,0,197
> +.byte 102,15,56,0,203
> + pxor %xmm4,%xmm0
> + movdqa 32(%r10),%xmm4
> + pxor %xmm1,%xmm0
> + movdqa 48(%r10),%xmm1
> +
> +.byte 102,15,56,0,226
> +.byte 102,15,56,0,197
> +.byte 102,15,56,0,203
> + pxor %xmm4,%xmm0
> + movdqa 64(%r10),%xmm4
> + pxor %xmm1,%xmm0
> + movdqa 80(%r10),%xmm1
> +
> +.byte 102,15,56,0,226
> +.byte 102,15,56,0,197
> +.byte 102,15,56,0,203
> + pxor %xmm4,%xmm0
> + addq $16,%r9
> +.byte 102,15,58,15,237,12
> + pxor %xmm1,%xmm0
> + subq $1,%rax
> +
> +.Ldec_entry:
> +
> + movdqa %xmm9,%xmm1
> + pandn %xmm0,%xmm1
> + movdqa %xmm11,%xmm2
> + psrld $4,%xmm1
> + pand %xmm9,%xmm0
> +.byte 102,15,56,0,208
> + movdqa %xmm10,%xmm3
> + pxor %xmm1,%xmm0
> +.byte 102,15,56,0,217
> + movdqa %xmm10,%xmm4
> + pxor %xmm2,%xmm3
> +.byte 102,15,56,0,224
> + pxor %xmm2,%xmm4
> + movdqa %xmm10,%xmm2
> +.byte 102,15,56,0,211
> + movdqa %xmm10,%xmm3
> + pxor %xmm0,%xmm2
> +.byte 102,15,56,0,220
> + movdqu (%r9),%xmm0
> + pxor %xmm1,%xmm3
> + jnz .Ldec_loop
> +
> +
> + movdqa 96(%r10),%xmm4
> +.byte 102,15,56,0,226
> + pxor %xmm0,%xmm4
> + movdqa 112(%r10),%xmm0
> + movdqa -352(%r11),%xmm2
> +.byte 102,15,56,0,195
> + pxor %xmm4,%xmm0
> +.byte 102,15,56,0,194
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _vpaes_decrypt_core,.-_vpaes_decrypt_core
> +
> +
> +
> +
> +
> +
> +.type _vpaes_schedule_core, at function
> +.align 16
> +_vpaes_schedule_core:
> +.cfi_startproc
> +
> +
> +
> +
> +
> + call _vpaes_preheat
> + movdqa .Lk_rcon(%rip),%xmm8
> + movdqu (%rdi),%xmm0
> +
> +
> + movdqa %xmm0,%xmm3
> + leaq .Lk_ipt(%rip),%r11
> + call _vpaes_schedule_transform
> + movdqa %xmm0,%xmm7
> +
> + leaq .Lk_sr(%rip),%r10
> + testq %rcx,%rcx
> + jnz .Lschedule_am_decrypting
> +
> +
> + movdqu %xmm0,(%rdx)
> + jmp .Lschedule_go
> +
> +.Lschedule_am_decrypting:
> +
> + movdqa (%r8,%r10,1),%xmm1
> +.byte 102,15,56,0,217
> + movdqu %xmm3,(%rdx)
> + xorq $0x30,%r8
> +
> +.Lschedule_go:
> + cmpl $192,%esi
> + ja .Lschedule_256
> + je .Lschedule_192
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.Lschedule_128:
> + movl $10,%esi
> +
> +.Loop_schedule_128:
> + call _vpaes_schedule_round
> + decq %rsi
> + jz .Lschedule_mangle_last
> + call _vpaes_schedule_mangle
> + jmp .Loop_schedule_128
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.align 16
> +.Lschedule_192:
> + movdqu 8(%rdi),%xmm0
> + call _vpaes_schedule_transform
> + movdqa %xmm0,%xmm6
> + pxor %xmm4,%xmm4
> + movhlps %xmm4,%xmm6
> + movl $4,%esi
> +
> +.Loop_schedule_192:
> + call _vpaes_schedule_round
> +.byte 102,15,58,15,198,8
> + call _vpaes_schedule_mangle
> + call _vpaes_schedule_192_smear
> + call _vpaes_schedule_mangle
> + call _vpaes_schedule_round
> + decq %rsi
> + jz .Lschedule_mangle_last
> + call _vpaes_schedule_mangle
> + call _vpaes_schedule_192_smear
> + jmp .Loop_schedule_192
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.align 16
> +.Lschedule_256:
> + movdqu 16(%rdi),%xmm0
> + call _vpaes_schedule_transform
> + movl $7,%esi
> +
> +.Loop_schedule_256:
> + call _vpaes_schedule_mangle
> + movdqa %xmm0,%xmm6
> +
> +
> + call _vpaes_schedule_round
> + decq %rsi
> + jz .Lschedule_mangle_last
> + call _vpaes_schedule_mangle
> +
> +
> + pshufd $0xFF,%xmm0,%xmm0
> + movdqa %xmm7,%xmm5
> + movdqa %xmm6,%xmm7
> + call _vpaes_schedule_low_round
> + movdqa %xmm5,%xmm7
> +
> + jmp .Loop_schedule_256
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.align 16
> +.Lschedule_mangle_last:
> +
> + leaq .Lk_deskew(%rip),%r11
> + testq %rcx,%rcx
> + jnz .Lschedule_mangle_last_dec
> +
> +
> + movdqa (%r8,%r10,1),%xmm1
> +.byte 102,15,56,0,193
> + leaq .Lk_opt(%rip),%r11
> + addq $32,%rdx
> +
> +.Lschedule_mangle_last_dec:
> + addq $-16,%rdx
> + pxor .Lk_s63(%rip),%xmm0
> + call _vpaes_schedule_transform
> + movdqu %xmm0,(%rdx)
> +
> +
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _vpaes_schedule_core,.-_vpaes_schedule_core
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.type _vpaes_schedule_192_smear, at function
> +.align 16
> +_vpaes_schedule_192_smear:
> +.cfi_startproc
> + pshufd $0x80,%xmm6,%xmm1
> + pshufd $0xFE,%xmm7,%xmm0
> + pxor %xmm1,%xmm6
> + pxor %xmm1,%xmm1
> + pxor %xmm0,%xmm6
> + movdqa %xmm6,%xmm0
> + movhlps %xmm1,%xmm6
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _vpaes_schedule_192_smear,.-_vpaes_schedule_192_smear
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.type _vpaes_schedule_round, at function
> +.align 16
> +_vpaes_schedule_round:
> +.cfi_startproc
> +
> + pxor %xmm1,%xmm1
> +.byte 102,65,15,58,15,200,15
> +.byte 102,69,15,58,15,192,15
> + pxor %xmm1,%xmm7
> +
> +
> + pshufd $0xFF,%xmm0,%xmm0
> +.byte 102,15,58,15,192,1
> +
> +
> +
> +
> +_vpaes_schedule_low_round:
> +
> + movdqa %xmm7,%xmm1
> + pslldq $4,%xmm7
> + pxor %xmm1,%xmm7
> + movdqa %xmm7,%xmm1
> + pslldq $8,%xmm7
> + pxor %xmm1,%xmm7
> + pxor .Lk_s63(%rip),%xmm7
> +
> +
> + movdqa %xmm9,%xmm1
> + pandn %xmm0,%xmm1
> + psrld $4,%xmm1
> + pand %xmm9,%xmm0
> + movdqa %xmm11,%xmm2
> +.byte 102,15,56,0,208
> + pxor %xmm1,%xmm0
> + movdqa %xmm10,%xmm3
> +.byte 102,15,56,0,217
> + pxor %xmm2,%xmm3
> + movdqa %xmm10,%xmm4
> +.byte 102,15,56,0,224
> + pxor %xmm2,%xmm4
> + movdqa %xmm10,%xmm2
> +.byte 102,15,56,0,211
> + pxor %xmm0,%xmm2
> + movdqa %xmm10,%xmm3
> +.byte 102,15,56,0,220
> + pxor %xmm1,%xmm3
> + movdqa %xmm13,%xmm4
> +.byte 102,15,56,0,226
> + movdqa %xmm12,%xmm0
> +.byte 102,15,56,0,195
> + pxor %xmm4,%xmm0
> +
> +
> + pxor %xmm7,%xmm0
> + movdqa %xmm0,%xmm7
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _vpaes_schedule_round,.-_vpaes_schedule_round
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.type _vpaes_schedule_transform, at function
> +.align 16
> +_vpaes_schedule_transform:
> +.cfi_startproc
> + movdqa %xmm9,%xmm1
> + pandn %xmm0,%xmm1
> + psrld $4,%xmm1
> + pand %xmm9,%xmm0
> + movdqa (%r11),%xmm2
> +.byte 102,15,56,0,208
> + movdqa 16(%r11),%xmm0
> +.byte 102,15,56,0,193
> + pxor %xmm2,%xmm0
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _vpaes_schedule_transform,.-_vpaes_schedule_transform
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +
> +.type _vpaes_schedule_mangle, at function
> +.align 16
> +_vpaes_schedule_mangle:
> +.cfi_startproc
> + movdqa %xmm0,%xmm4
> + movdqa .Lk_mc_forward(%rip),%xmm5
> + testq %rcx,%rcx
> + jnz .Lschedule_mangle_dec
> +
> +
> + addq $16,%rdx
> + pxor .Lk_s63(%rip),%xmm4
> +.byte 102,15,56,0,229
> + movdqa %xmm4,%xmm3
> +.byte 102,15,56,0,229
> + pxor %xmm4,%xmm3
> +.byte 102,15,56,0,229
> + pxor %xmm4,%xmm3
> +
> + jmp .Lschedule_mangle_both
> +.align 16
> +.Lschedule_mangle_dec:
> +
> + leaq .Lk_dksd(%rip),%r11
> + movdqa %xmm9,%xmm1
> + pandn %xmm4,%xmm1
> + psrld $4,%xmm1
> + pand %xmm9,%xmm4
> +
> + movdqa 0(%r11),%xmm2
> +.byte 102,15,56,0,212
> + movdqa 16(%r11),%xmm3
> +.byte 102,15,56,0,217
> + pxor %xmm2,%xmm3
> +.byte 102,15,56,0,221
> +
> + movdqa 32(%r11),%xmm2
> +.byte 102,15,56,0,212
> + pxor %xmm3,%xmm2
> + movdqa 48(%r11),%xmm3
> +.byte 102,15,56,0,217
> + pxor %xmm2,%xmm3
> +.byte 102,15,56,0,221
> +
> + movdqa 64(%r11),%xmm2
> +.byte 102,15,56,0,212
> + pxor %xmm3,%xmm2
> + movdqa 80(%r11),%xmm3
> +.byte 102,15,56,0,217
> + pxor %xmm2,%xmm3
> +.byte 102,15,56,0,221
> +
> + movdqa 96(%r11),%xmm2
> +.byte 102,15,56,0,212
> + pxor %xmm3,%xmm2
> + movdqa 112(%r11),%xmm3
> +.byte 102,15,56,0,217
> + pxor %xmm2,%xmm3
> +
> + addq $-16,%rdx
> +
> +.Lschedule_mangle_both:
> + movdqa (%r8,%r10,1),%xmm1
> +.byte 102,15,56,0,217
> + addq $-16,%r8
> + andq $0x30,%r8
> + movdqu %xmm3,(%rdx)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _vpaes_schedule_mangle,.-_vpaes_schedule_mangle
> +
> +
> +
> +
> +.globl vpaes_set_encrypt_key
> +.type vpaes_set_encrypt_key, at function
> +.align 16
> +vpaes_set_encrypt_key:
> +.cfi_startproc
> + movl %esi,%eax
> + shrl $5,%eax
> + addl $5,%eax
> + movl %eax,240(%rdx)
> +
> + movl $0,%ecx
> + movl $0x30,%r8d
> + call _vpaes_schedule_core
> + xorl %eax,%eax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size vpaes_set_encrypt_key,.-vpaes_set_encrypt_key
> +
> +.globl vpaes_set_decrypt_key
> +.type vpaes_set_decrypt_key, at function
> +.align 16
> +vpaes_set_decrypt_key:
> +.cfi_startproc
> + movl %esi,%eax
> + shrl $5,%eax
> + addl $5,%eax
> + movl %eax,240(%rdx)
> + shll $4,%eax
> + leaq 16(%rdx,%rax,1),%rdx
> +
> + movl $1,%ecx
> + movl %esi,%r8d
> + shrl $1,%r8d
> + andl $32,%r8d
> + xorl $32,%r8d
> + call _vpaes_schedule_core
> + xorl %eax,%eax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size vpaes_set_decrypt_key,.-vpaes_set_decrypt_key
> +
> +.globl vpaes_encrypt
> +.type vpaes_encrypt, at function
> +.align 16
> +vpaes_encrypt:
> +.cfi_startproc
> + movdqu (%rdi),%xmm0
> + call _vpaes_preheat
> + call _vpaes_encrypt_core
> + movdqu %xmm0,(%rsi)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size vpaes_encrypt,.-vpaes_encrypt
> +
> +.globl vpaes_decrypt
> +.type vpaes_decrypt, at function
> +.align 16
> +vpaes_decrypt:
> +.cfi_startproc
> + movdqu (%rdi),%xmm0
> + call _vpaes_preheat
> + call _vpaes_decrypt_core
> + movdqu %xmm0,(%rsi)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size vpaes_decrypt,.-vpaes_decrypt
> +.globl vpaes_cbc_encrypt
> +.type vpaes_cbc_encrypt, at function
> +.align 16
> +vpaes_cbc_encrypt:
> +.cfi_startproc
> + xchgq %rcx,%rdx
> + subq $16,%rcx
> + jc .Lcbc_abort
> + movdqu (%r8),%xmm6
> + subq %rdi,%rsi
> + call _vpaes_preheat
> + cmpl $0,%r9d
> + je .Lcbc_dec_loop
> + jmp .Lcbc_enc_loop
> +.align 16
> +.Lcbc_enc_loop:
> + movdqu (%rdi),%xmm0
> + pxor %xmm6,%xmm0
> + call _vpaes_encrypt_core
> + movdqa %xmm0,%xmm6
> + movdqu %xmm0,(%rsi,%rdi,1)
> + leaq 16(%rdi),%rdi
> + subq $16,%rcx
> + jnc .Lcbc_enc_loop
> + jmp .Lcbc_done
> +.align 16
> +.Lcbc_dec_loop:
> + movdqu (%rdi),%xmm0
> + movdqa %xmm0,%xmm7
> + call _vpaes_decrypt_core
> + pxor %xmm6,%xmm0
> + movdqa %xmm7,%xmm6
> + movdqu %xmm0,(%rsi,%rdi,1)
> + leaq 16(%rdi),%rdi
> + subq $16,%rcx
> + jnc .Lcbc_dec_loop
> +.Lcbc_done:
> + movdqu %xmm6,(%r8)
> +.Lcbc_abort:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size vpaes_cbc_encrypt,.-vpaes_cbc_encrypt
> +
> +
> +
> +
> +
> +
> +.type _vpaes_preheat, at function
> +.align 16
> +_vpaes_preheat:
> +.cfi_startproc
> + leaq .Lk_s0F(%rip),%r10
> + movdqa -32(%r10),%xmm10
> + movdqa -16(%r10),%xmm11
> + movdqa 0(%r10),%xmm9
> + movdqa 48(%r10),%xmm13
> + movdqa 64(%r10),%xmm12
> + movdqa 80(%r10),%xmm15
> + movdqa 96(%r10),%xmm14
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size _vpaes_preheat,.-_vpaes_preheat
> +
> +
> +
> +
> +
> +.type _vpaes_consts, at object
> +.align 64
> +_vpaes_consts:
> +.Lk_inv:
> +.quad 0x0E05060F0D080180, 0x040703090A0B0C02
> +.quad 0x01040A060F0B0780, 0x030D0E0C02050809
> +
> +.Lk_s0F:
> +.quad 0x0F0F0F0F0F0F0F0F, 0x0F0F0F0F0F0F0F0F
> +
> +.Lk_ipt:
> +.quad 0xC2B2E8985A2A7000, 0xCABAE09052227808
> +.quad 0x4C01307D317C4D00, 0xCD80B1FCB0FDCC81
> +
> +.Lk_sb1:
> +.quad 0xB19BE18FCB503E00, 0xA5DF7A6E142AF544
> +.quad 0x3618D415FAE22300, 0x3BF7CCC10D2ED9EF
> +.Lk_sb2:
> +.quad 0xE27A93C60B712400, 0x5EB7E955BC982FCD
> +.quad 0x69EB88400AE12900, 0xC2A163C8AB82234A
> +.Lk_sbo:
> +.quad 0xD0D26D176FBDC700, 0x15AABF7AC502A878
> +.quad 0xCFE474A55FBB6A00, 0x8E1E90D1412B35FA
> +
> +.Lk_mc_forward:
> +.quad 0x0407060500030201, 0x0C0F0E0D080B0A09
> +.quad 0x080B0A0904070605, 0x000302010C0F0E0D
> +.quad 0x0C0F0E0D080B0A09, 0x0407060500030201
> +.quad 0x000302010C0F0E0D, 0x080B0A0904070605
> +
> +.Lk_mc_backward:
> +.quad 0x0605040702010003, 0x0E0D0C0F0A09080B
> +.quad 0x020100030E0D0C0F, 0x0A09080B06050407
> +.quad 0x0E0D0C0F0A09080B, 0x0605040702010003
> +.quad 0x0A09080B06050407, 0x020100030E0D0C0F
> +
> +.Lk_sr:
> +.quad 0x0706050403020100, 0x0F0E0D0C0B0A0908
> +.quad 0x030E09040F0A0500, 0x0B06010C07020D08
> +.quad 0x0F060D040B020900, 0x070E050C030A0108
> +.quad 0x0B0E0104070A0D00, 0x0306090C0F020508
> +
> +.Lk_rcon:
> +.quad 0x1F8391B9AF9DEEB6, 0x702A98084D7C7D81
> +
> +.Lk_s63:
> +.quad 0x5B5B5B5B5B5B5B5B, 0x5B5B5B5B5B5B5B5B
> +
> +.Lk_opt:
> +.quad 0xFF9F4929D6B66000, 0xF7974121DEBE6808
> +.quad 0x01EDBD5150BCEC00, 0xE10D5DB1B05C0CE0
> +
> +.Lk_deskew:
> +.quad 0x07E4A34047A4E300, 0x1DFEB95A5DBEF91A
> +.quad 0x5F36B5DC83EA6900, 0x2841C2ABF49D1E77
> +
> +
> +
> +
> +
> +.Lk_dksd:
> +.quad 0xFEB91A5DA3E44700, 0x0740E3A45A1DBEF9
> +.quad 0x41C277F4B5368300, 0x5FDC69EAAB289D1E
> +.Lk_dksb:
> +.quad 0x9A4FCA1F8550D500, 0x03D653861CC94C99
> +.quad 0x115BEDA7B6FC4A00, 0xD993256F7E3482C8
> +.Lk_dkse:
> +.quad 0xD5031CCA1FC9D600, 0x53859A4C994F5086
> +.quad 0xA23196054FDC7BE8, 0xCD5EF96A20B31487
> +.Lk_dks9:
> +.quad 0xB6116FC87ED9A700, 0x4AED933482255BFC
> +.quad 0x4576516227143300, 0x8BB89FACE9DAFDCE
> +
> +
> +
> +
> +
> +.Lk_dipt:
> +.quad 0x0F505B040B545F00, 0x154A411E114E451A
> +.quad 0x86E383E660056500, 0x12771772F491F194
> +
> +.Lk_dsb9:
> +.quad 0x851C03539A86D600, 0xCAD51F504F994CC9
> +.quad 0xC03B1789ECD74900, 0x725E2C9EB2FBA565
> +.Lk_dsbd:
> +.quad 0x7D57CCDFE6B1A200, 0xF56E9B13882A4439
> +.quad 0x3CE2FAF724C6CB00, 0x2931180D15DEEFD3
> +.Lk_dsbb:
> +.quad 0xD022649296B44200, 0x602646F6B0F2D404
> +.quad 0xC19498A6CD596700, 0xF3FF0C3E3255AA6B
> +.Lk_dsbe:
> +.quad 0x46F2929626D4D000, 0x2242600464B4F6B0
> +.quad 0x0C55A6CDFFAAC100, 0x9467F36B98593E32
> +.Lk_dsbo:
> +.quad 0x1387EA537EF94000, 0xC7AA6DB9D4943E2D
> +.quad 0x12D7560F93441D00, 0xCA4B8159D8C58E9C
> +.byte
> 86,101,99,116,111,114,32,80,101,114,109,117,116,97,116,105,111,110,32,65,6
> 9,83,32,102,111,114,32,120,56,54,95,54,52,47,83,83,83,69,51,44,32,77,105,10
> 7,101,32,72,97,109,98,117,114,103,32,40,83,116,97,110,102,111,114,100,32,8
> 5,110,105,118,101,114,115,105,116,121,41,0
> +.align 64
> +.size _vpaes_consts,.-_vpaes_consts
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/modes/aesni-gcm-
> x86_64.S b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/modes/aesni-gcm-
> x86_64.S
> new file mode 100644
> index 0000000000..1201f3427a
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/modes/aesni-gcm-
> x86_64.S
> @@ -0,0 +1,29 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/modes/asm/aesni-gcm-x86_64.pl
> +#
> +# Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +.globl aesni_gcm_encrypt
> +.type aesni_gcm_encrypt, at function
> +aesni_gcm_encrypt:
> +.cfi_startproc
> + xorl %eax,%eax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_gcm_encrypt,.-aesni_gcm_encrypt
> +
> +.globl aesni_gcm_decrypt
> +.type aesni_gcm_decrypt, at function
> +aesni_gcm_decrypt:
> +.cfi_startproc
> + xorl %eax,%eax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size aesni_gcm_decrypt,.-aesni_gcm_decrypt
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/modes/ghash-
> x86_64.S b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/modes/ghash-
> x86_64.S
> new file mode 100644
> index 0000000000..3fcaa4b2ef
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/modes/ghash-x86_64.S
> @@ -0,0 +1,1386 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/modes/asm/ghash-x86_64.pl
> +#
> +# Copyright 2010-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +.globl gcm_gmult_4bit
> +.type gcm_gmult_4bit, at function
> +.align 16
> +gcm_gmult_4bit:
> +.cfi_startproc
> + pushq %rbx
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r14,-48
> + pushq %r15
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r15,-56
> + subq $280,%rsp
> +.cfi_adjust_cfa_offset 280
> +.Lgmult_prologue:
> +
> + movzbq 15(%rdi),%r8
> + leaq .Lrem_4bit(%rip),%r11
> + xorq %rax,%rax
> + xorq %rbx,%rbx
> + movb %r8b,%al
> + movb %r8b,%bl
> + shlb $4,%al
> + movq $14,%rcx
> + movq 8(%rsi,%rax,1),%r8
> + movq (%rsi,%rax,1),%r9
> + andb $0xf0,%bl
> + movq %r8,%rdx
> + jmp .Loop1
> +
> +.align 16
> +.Loop1:
> + shrq $4,%r8
> + andq $0xf,%rdx
> + movq %r9,%r10
> + movb (%rdi,%rcx,1),%al
> + shrq $4,%r9
> + xorq 8(%rsi,%rbx,1),%r8
> + shlq $60,%r10
> + xorq (%rsi,%rbx,1),%r9
> + movb %al,%bl
> + xorq (%r11,%rdx,8),%r9
> + movq %r8,%rdx
> + shlb $4,%al
> + xorq %r10,%r8
> + decq %rcx
> + js .Lbreak1
> +
> + shrq $4,%r8
> + andq $0xf,%rdx
> + movq %r9,%r10
> + shrq $4,%r9
> + xorq 8(%rsi,%rax,1),%r8
> + shlq $60,%r10
> + xorq (%rsi,%rax,1),%r9
> + andb $0xf0,%bl
> + xorq (%r11,%rdx,8),%r9
> + movq %r8,%rdx
> + xorq %r10,%r8
> + jmp .Loop1
> +
> +.align 16
> +.Lbreak1:
> + shrq $4,%r8
> + andq $0xf,%rdx
> + movq %r9,%r10
> + shrq $4,%r9
> + xorq 8(%rsi,%rax,1),%r8
> + shlq $60,%r10
> + xorq (%rsi,%rax,1),%r9
> + andb $0xf0,%bl
> + xorq (%r11,%rdx,8),%r9
> + movq %r8,%rdx
> + xorq %r10,%r8
> +
> + shrq $4,%r8
> + andq $0xf,%rdx
> + movq %r9,%r10
> + shrq $4,%r9
> + xorq 8(%rsi,%rbx,1),%r8
> + shlq $60,%r10
> + xorq (%rsi,%rbx,1),%r9
> + xorq %r10,%r8
> + xorq (%r11,%rdx,8),%r9
> +
> + bswapq %r8
> + bswapq %r9
> + movq %r8,8(%rdi)
> + movq %r9,(%rdi)
> +
> + leaq 280+48(%rsp),%rsi
> +.cfi_def_cfa %rsi,8
> + movq -8(%rsi),%rbx
> +.cfi_restore %rbx
> + leaq (%rsi),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lgmult_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size gcm_gmult_4bit,.-gcm_gmult_4bit
> +.globl gcm_ghash_4bit
> +.type gcm_ghash_4bit, at function
> +.align 16
> +gcm_ghash_4bit:
> +.cfi_startproc
> + pushq %rbx
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r14,-48
> + pushq %r15
> +.cfi_adjust_cfa_offset 8
> +.cfi_offset %r15,-56
> + subq $280,%rsp
> +.cfi_adjust_cfa_offset 280
> +.Lghash_prologue:
> + movq %rdx,%r14
> + movq %rcx,%r15
> + subq $-128,%rsi
> + leaq 16+128(%rsp),%rbp
> + xorl %edx,%edx
> + movq 0+0-128(%rsi),%r8
> + movq 0+8-128(%rsi),%rax
> + movb %al,%dl
> + shrq $4,%rax
> + movq %r8,%r10
> + shrq $4,%r8
> + movq 16+0-128(%rsi),%r9
> + shlb $4,%dl
> + movq 16+8-128(%rsi),%rbx
> + shlq $60,%r10
> + movb %dl,0(%rsp)
> + orq %r10,%rax
> + movb %bl,%dl
> + shrq $4,%rbx
> + movq %r9,%r10
> + shrq $4,%r9
> + movq %r8,0(%rbp)
> + movq 32+0-128(%rsi),%r8
> + shlb $4,%dl
> + movq %rax,0-128(%rbp)
> + movq 32+8-128(%rsi),%rax
> + shlq $60,%r10
> + movb %dl,1(%rsp)
> + orq %r10,%rbx
> + movb %al,%dl
> + shrq $4,%rax
> + movq %r8,%r10
> + shrq $4,%r8
> + movq %r9,8(%rbp)
> + movq 48+0-128(%rsi),%r9
> + shlb $4,%dl
> + movq %rbx,8-128(%rbp)
> + movq 48+8-128(%rsi),%rbx
> + shlq $60,%r10
> + movb %dl,2(%rsp)
> + orq %r10,%rax
> + movb %bl,%dl
> + shrq $4,%rbx
> + movq %r9,%r10
> + shrq $4,%r9
> + movq %r8,16(%rbp)
> + movq 64+0-128(%rsi),%r8
> + shlb $4,%dl
> + movq %rax,16-128(%rbp)
> + movq 64+8-128(%rsi),%rax
> + shlq $60,%r10
> + movb %dl,3(%rsp)
> + orq %r10,%rbx
> + movb %al,%dl
> + shrq $4,%rax
> + movq %r8,%r10
> + shrq $4,%r8
> + movq %r9,24(%rbp)
> + movq 80+0-128(%rsi),%r9
> + shlb $4,%dl
> + movq %rbx,24-128(%rbp)
> + movq 80+8-128(%rsi),%rbx
> + shlq $60,%r10
> + movb %dl,4(%rsp)
> + orq %r10,%rax
> + movb %bl,%dl
> + shrq $4,%rbx
> + movq %r9,%r10
> + shrq $4,%r9
> + movq %r8,32(%rbp)
> + movq 96+0-128(%rsi),%r8
> + shlb $4,%dl
> + movq %rax,32-128(%rbp)
> + movq 96+8-128(%rsi),%rax
> + shlq $60,%r10
> + movb %dl,5(%rsp)
> + orq %r10,%rbx
> + movb %al,%dl
> + shrq $4,%rax
> + movq %r8,%r10
> + shrq $4,%r8
> + movq %r9,40(%rbp)
> + movq 112+0-128(%rsi),%r9
> + shlb $4,%dl
> + movq %rbx,40-128(%rbp)
> + movq 112+8-128(%rsi),%rbx
> + shlq $60,%r10
> + movb %dl,6(%rsp)
> + orq %r10,%rax
> + movb %bl,%dl
> + shrq $4,%rbx
> + movq %r9,%r10
> + shrq $4,%r9
> + movq %r8,48(%rbp)
> + movq 128+0-128(%rsi),%r8
> + shlb $4,%dl
> + movq %rax,48-128(%rbp)
> + movq 128+8-128(%rsi),%rax
> + shlq $60,%r10
> + movb %dl,7(%rsp)
> + orq %r10,%rbx
> + movb %al,%dl
> + shrq $4,%rax
> + movq %r8,%r10
> + shrq $4,%r8
> + movq %r9,56(%rbp)
> + movq 144+0-128(%rsi),%r9
> + shlb $4,%dl
> + movq %rbx,56-128(%rbp)
> + movq 144+8-128(%rsi),%rbx
> + shlq $60,%r10
> + movb %dl,8(%rsp)
> + orq %r10,%rax
> + movb %bl,%dl
> + shrq $4,%rbx
> + movq %r9,%r10
> + shrq $4,%r9
> + movq %r8,64(%rbp)
> + movq 160+0-128(%rsi),%r8
> + shlb $4,%dl
> + movq %rax,64-128(%rbp)
> + movq 160+8-128(%rsi),%rax
> + shlq $60,%r10
> + movb %dl,9(%rsp)
> + orq %r10,%rbx
> + movb %al,%dl
> + shrq $4,%rax
> + movq %r8,%r10
> + shrq $4,%r8
> + movq %r9,72(%rbp)
> + movq 176+0-128(%rsi),%r9
> + shlb $4,%dl
> + movq %rbx,72-128(%rbp)
> + movq 176+8-128(%rsi),%rbx
> + shlq $60,%r10
> + movb %dl,10(%rsp)
> + orq %r10,%rax
> + movb %bl,%dl
> + shrq $4,%rbx
> + movq %r9,%r10
> + shrq $4,%r9
> + movq %r8,80(%rbp)
> + movq 192+0-128(%rsi),%r8
> + shlb $4,%dl
> + movq %rax,80-128(%rbp)
> + movq 192+8-128(%rsi),%rax
> + shlq $60,%r10
> + movb %dl,11(%rsp)
> + orq %r10,%rbx
> + movb %al,%dl
> + shrq $4,%rax
> + movq %r8,%r10
> + shrq $4,%r8
> + movq %r9,88(%rbp)
> + movq 208+0-128(%rsi),%r9
> + shlb $4,%dl
> + movq %rbx,88-128(%rbp)
> + movq 208+8-128(%rsi),%rbx
> + shlq $60,%r10
> + movb %dl,12(%rsp)
> + orq %r10,%rax
> + movb %bl,%dl
> + shrq $4,%rbx
> + movq %r9,%r10
> + shrq $4,%r9
> + movq %r8,96(%rbp)
> + movq 224+0-128(%rsi),%r8
> + shlb $4,%dl
> + movq %rax,96-128(%rbp)
> + movq 224+8-128(%rsi),%rax
> + shlq $60,%r10
> + movb %dl,13(%rsp)
> + orq %r10,%rbx
> + movb %al,%dl
> + shrq $4,%rax
> + movq %r8,%r10
> + shrq $4,%r8
> + movq %r9,104(%rbp)
> + movq 240+0-128(%rsi),%r9
> + shlb $4,%dl
> + movq %rbx,104-128(%rbp)
> + movq 240+8-128(%rsi),%rbx
> + shlq $60,%r10
> + movb %dl,14(%rsp)
> + orq %r10,%rax
> + movb %bl,%dl
> + shrq $4,%rbx
> + movq %r9,%r10
> + shrq $4,%r9
> + movq %r8,112(%rbp)
> + shlb $4,%dl
> + movq %rax,112-128(%rbp)
> + shlq $60,%r10
> + movb %dl,15(%rsp)
> + orq %r10,%rbx
> + movq %r9,120(%rbp)
> + movq %rbx,120-128(%rbp)
> + addq $-128,%rsi
> + movq 8(%rdi),%r8
> + movq 0(%rdi),%r9
> + addq %r14,%r15
> + leaq .Lrem_8bit(%rip),%r11
> + jmp .Louter_loop
> +.align 16
> +.Louter_loop:
> + xorq (%r14),%r9
> + movq 8(%r14),%rdx
> + leaq 16(%r14),%r14
> + xorq %r8,%rdx
> + movq %r9,(%rdi)
> + movq %rdx,8(%rdi)
> + shrq $32,%rdx
> + xorq %rax,%rax
> + roll $8,%edx
> + movb %dl,%al
> + movzbl %dl,%ebx
> + shlb $4,%al
> + shrl $4,%ebx
> + roll $8,%edx
> + movq 8(%rsi,%rax,1),%r8
> + movq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + movzbl %dl,%ecx
> + shlb $4,%al
> + movzbq (%rsp,%rbx,1),%r12
> + shrl $4,%ecx
> + xorq %r8,%r12
> + movq %r9,%r10
> + shrq $8,%r8
> + movzbq %r12b,%r12
> + shrq $8,%r9
> + xorq -128(%rbp,%rbx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rbx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r12,2),%r12
> + movzbl %dl,%ebx
> + shlb $4,%al
> + movzbq (%rsp,%rcx,1),%r13
> + shrl $4,%ebx
> + shlq $48,%r12
> + xorq %r8,%r13
> + movq %r9,%r10
> + xorq %r12,%r9
> + shrq $8,%r8
> + movzbq %r13b,%r13
> + shrq $8,%r9
> + xorq -128(%rbp,%rcx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rcx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r13,2),%r13
> + movzbl %dl,%ecx
> + shlb $4,%al
> + movzbq (%rsp,%rbx,1),%r12
> + shrl $4,%ecx
> + shlq $48,%r13
> + xorq %r8,%r12
> + movq %r9,%r10
> + xorq %r13,%r9
> + shrq $8,%r8
> + movzbq %r12b,%r12
> + movl 8(%rdi),%edx
> + shrq $8,%r9
> + xorq -128(%rbp,%rbx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rbx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r12,2),%r12
> + movzbl %dl,%ebx
> + shlb $4,%al
> + movzbq (%rsp,%rcx,1),%r13
> + shrl $4,%ebx
> + shlq $48,%r12
> + xorq %r8,%r13
> + movq %r9,%r10
> + xorq %r12,%r9
> + shrq $8,%r8
> + movzbq %r13b,%r13
> + shrq $8,%r9
> + xorq -128(%rbp,%rcx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rcx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r13,2),%r13
> + movzbl %dl,%ecx
> + shlb $4,%al
> + movzbq (%rsp,%rbx,1),%r12
> + shrl $4,%ecx
> + shlq $48,%r13
> + xorq %r8,%r12
> + movq %r9,%r10
> + xorq %r13,%r9
> + shrq $8,%r8
> + movzbq %r12b,%r12
> + shrq $8,%r9
> + xorq -128(%rbp,%rbx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rbx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r12,2),%r12
> + movzbl %dl,%ebx
> + shlb $4,%al
> + movzbq (%rsp,%rcx,1),%r13
> + shrl $4,%ebx
> + shlq $48,%r12
> + xorq %r8,%r13
> + movq %r9,%r10
> + xorq %r12,%r9
> + shrq $8,%r8
> + movzbq %r13b,%r13
> + shrq $8,%r9
> + xorq -128(%rbp,%rcx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rcx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r13,2),%r13
> + movzbl %dl,%ecx
> + shlb $4,%al
> + movzbq (%rsp,%rbx,1),%r12
> + shrl $4,%ecx
> + shlq $48,%r13
> + xorq %r8,%r12
> + movq %r9,%r10
> + xorq %r13,%r9
> + shrq $8,%r8
> + movzbq %r12b,%r12
> + movl 4(%rdi),%edx
> + shrq $8,%r9
> + xorq -128(%rbp,%rbx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rbx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r12,2),%r12
> + movzbl %dl,%ebx
> + shlb $4,%al
> + movzbq (%rsp,%rcx,1),%r13
> + shrl $4,%ebx
> + shlq $48,%r12
> + xorq %r8,%r13
> + movq %r9,%r10
> + xorq %r12,%r9
> + shrq $8,%r8
> + movzbq %r13b,%r13
> + shrq $8,%r9
> + xorq -128(%rbp,%rcx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rcx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r13,2),%r13
> + movzbl %dl,%ecx
> + shlb $4,%al
> + movzbq (%rsp,%rbx,1),%r12
> + shrl $4,%ecx
> + shlq $48,%r13
> + xorq %r8,%r12
> + movq %r9,%r10
> + xorq %r13,%r9
> + shrq $8,%r8
> + movzbq %r12b,%r12
> + shrq $8,%r9
> + xorq -128(%rbp,%rbx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rbx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r12,2),%r12
> + movzbl %dl,%ebx
> + shlb $4,%al
> + movzbq (%rsp,%rcx,1),%r13
> + shrl $4,%ebx
> + shlq $48,%r12
> + xorq %r8,%r13
> + movq %r9,%r10
> + xorq %r12,%r9
> + shrq $8,%r8
> + movzbq %r13b,%r13
> + shrq $8,%r9
> + xorq -128(%rbp,%rcx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rcx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r13,2),%r13
> + movzbl %dl,%ecx
> + shlb $4,%al
> + movzbq (%rsp,%rbx,1),%r12
> + shrl $4,%ecx
> + shlq $48,%r13
> + xorq %r8,%r12
> + movq %r9,%r10
> + xorq %r13,%r9
> + shrq $8,%r8
> + movzbq %r12b,%r12
> + movl 0(%rdi),%edx
> + shrq $8,%r9
> + xorq -128(%rbp,%rbx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rbx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r12,2),%r12
> + movzbl %dl,%ebx
> + shlb $4,%al
> + movzbq (%rsp,%rcx,1),%r13
> + shrl $4,%ebx
> + shlq $48,%r12
> + xorq %r8,%r13
> + movq %r9,%r10
> + xorq %r12,%r9
> + shrq $8,%r8
> + movzbq %r13b,%r13
> + shrq $8,%r9
> + xorq -128(%rbp,%rcx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rcx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r13,2),%r13
> + movzbl %dl,%ecx
> + shlb $4,%al
> + movzbq (%rsp,%rbx,1),%r12
> + shrl $4,%ecx
> + shlq $48,%r13
> + xorq %r8,%r12
> + movq %r9,%r10
> + xorq %r13,%r9
> + shrq $8,%r8
> + movzbq %r12b,%r12
> + shrq $8,%r9
> + xorq -128(%rbp,%rbx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rbx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r12,2),%r12
> + movzbl %dl,%ebx
> + shlb $4,%al
> + movzbq (%rsp,%rcx,1),%r13
> + shrl $4,%ebx
> + shlq $48,%r12
> + xorq %r8,%r13
> + movq %r9,%r10
> + xorq %r12,%r9
> + shrq $8,%r8
> + movzbq %r13b,%r13
> + shrq $8,%r9
> + xorq -128(%rbp,%rcx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rcx,8),%r9
> + roll $8,%edx
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + movb %dl,%al
> + xorq %r10,%r8
> + movzwq (%r11,%r13,2),%r13
> + movzbl %dl,%ecx
> + shlb $4,%al
> + movzbq (%rsp,%rbx,1),%r12
> + andl $240,%ecx
> + shlq $48,%r13
> + xorq %r8,%r12
> + movq %r9,%r10
> + xorq %r13,%r9
> + shrq $8,%r8
> + movzbq %r12b,%r12
> + movl -4(%rdi),%edx
> + shrq $8,%r9
> + xorq -128(%rbp,%rbx,8),%r8
> + shlq $56,%r10
> + xorq (%rbp,%rbx,8),%r9
> + movzwq (%r11,%r12,2),%r12
> + xorq 8(%rsi,%rax,1),%r8
> + xorq (%rsi,%rax,1),%r9
> + shlq $48,%r12
> + xorq %r10,%r8
> + xorq %r12,%r9
> + movzbq %r8b,%r13
> + shrq $4,%r8
> + movq %r9,%r10
> + shlb $4,%r13b
> + shrq $4,%r9
> + xorq 8(%rsi,%rcx,1),%r8
> + movzwq (%r11,%r13,2),%r13
> + shlq $60,%r10
> + xorq (%rsi,%rcx,1),%r9
> + xorq %r10,%r8
> + shlq $48,%r13
> + bswapq %r8
> + xorq %r13,%r9
> + bswapq %r9
> + cmpq %r15,%r14
> + jb .Louter_loop
> + movq %r8,8(%rdi)
> + movq %r9,(%rdi)
> +
> + leaq 280+48(%rsp),%rsi
> +.cfi_def_cfa %rsi,8
> + movq -48(%rsi),%r15
> +.cfi_restore %r15
> + movq -40(%rsi),%r14
> +.cfi_restore %r14
> + movq -32(%rsi),%r13
> +.cfi_restore %r13
> + movq -24(%rsi),%r12
> +.cfi_restore %r12
> + movq -16(%rsi),%rbp
> +.cfi_restore %rbp
> + movq -8(%rsi),%rbx
> +.cfi_restore %rbx
> + leaq 0(%rsi),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lghash_epilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size gcm_ghash_4bit,.-gcm_ghash_4bit
> +.globl gcm_init_clmul
> +.type gcm_init_clmul, at function
> +.align 16
> +gcm_init_clmul:
> +.cfi_startproc
> +.L_init_clmul:
> + movdqu (%rsi),%xmm2
> + pshufd $78,%xmm2,%xmm2
> +
> +
> + pshufd $255,%xmm2,%xmm4
> + movdqa %xmm2,%xmm3
> + psllq $1,%xmm2
> + pxor %xmm5,%xmm5
> + psrlq $63,%xmm3
> + pcmpgtd %xmm4,%xmm5
> + pslldq $8,%xmm3
> + por %xmm3,%xmm2
> +
> +
> + pand .L0x1c2_polynomial(%rip),%xmm5
> + pxor %xmm5,%xmm2
> +
> +
> + pshufd $78,%xmm2,%xmm6
> + movdqa %xmm2,%xmm0
> + pxor %xmm2,%xmm6
> + movdqa %xmm0,%xmm1
> + pshufd $78,%xmm0,%xmm3
> + pxor %xmm0,%xmm3
> +.byte 102,15,58,68,194,0
> +.byte 102,15,58,68,202,17
> +.byte 102,15,58,68,222,0
> + pxor %xmm0,%xmm3
> + pxor %xmm1,%xmm3
> +
> + movdqa %xmm3,%xmm4
> + psrldq $8,%xmm3
> + pslldq $8,%xmm4
> + pxor %xmm3,%xmm1
> + pxor %xmm4,%xmm0
> +
> + movdqa %xmm0,%xmm4
> + movdqa %xmm0,%xmm3
> + psllq $5,%xmm0
> + pxor %xmm0,%xmm3
> + psllq $1,%xmm0
> + pxor %xmm3,%xmm0
> + psllq $57,%xmm0
> + movdqa %xmm0,%xmm3
> + pslldq $8,%xmm0
> + psrldq $8,%xmm3
> + pxor %xmm4,%xmm0
> + pxor %xmm3,%xmm1
> +
> +
> + movdqa %xmm0,%xmm4
> + psrlq $1,%xmm0
> + pxor %xmm4,%xmm1
> + pxor %xmm0,%xmm4
> + psrlq $5,%xmm0
> + pxor %xmm4,%xmm0
> + psrlq $1,%xmm0
> + pxor %xmm1,%xmm0
> + pshufd $78,%xmm2,%xmm3
> + pshufd $78,%xmm0,%xmm4
> + pxor %xmm2,%xmm3
> + movdqu %xmm2,0(%rdi)
> + pxor %xmm0,%xmm4
> + movdqu %xmm0,16(%rdi)
> +.byte 102,15,58,15,227,8
> + movdqu %xmm4,32(%rdi)
> + movdqa %xmm0,%xmm1
> + pshufd $78,%xmm0,%xmm3
> + pxor %xmm0,%xmm3
> +.byte 102,15,58,68,194,0
> +.byte 102,15,58,68,202,17
> +.byte 102,15,58,68,222,0
> + pxor %xmm0,%xmm3
> + pxor %xmm1,%xmm3
> +
> + movdqa %xmm3,%xmm4
> + psrldq $8,%xmm3
> + pslldq $8,%xmm4
> + pxor %xmm3,%xmm1
> + pxor %xmm4,%xmm0
> +
> + movdqa %xmm0,%xmm4
> + movdqa %xmm0,%xmm3
> + psllq $5,%xmm0
> + pxor %xmm0,%xmm3
> + psllq $1,%xmm0
> + pxor %xmm3,%xmm0
> + psllq $57,%xmm0
> + movdqa %xmm0,%xmm3
> + pslldq $8,%xmm0
> + psrldq $8,%xmm3
> + pxor %xmm4,%xmm0
> + pxor %xmm3,%xmm1
> +
> +
> + movdqa %xmm0,%xmm4
> + psrlq $1,%xmm0
> + pxor %xmm4,%xmm1
> + pxor %xmm0,%xmm4
> + psrlq $5,%xmm0
> + pxor %xmm4,%xmm0
> + psrlq $1,%xmm0
> + pxor %xmm1,%xmm0
> + movdqa %xmm0,%xmm5
> + movdqa %xmm0,%xmm1
> + pshufd $78,%xmm0,%xmm3
> + pxor %xmm0,%xmm3
> +.byte 102,15,58,68,194,0
> +.byte 102,15,58,68,202,17
> +.byte 102,15,58,68,222,0
> + pxor %xmm0,%xmm3
> + pxor %xmm1,%xmm3
> +
> + movdqa %xmm3,%xmm4
> + psrldq $8,%xmm3
> + pslldq $8,%xmm4
> + pxor %xmm3,%xmm1
> + pxor %xmm4,%xmm0
> +
> + movdqa %xmm0,%xmm4
> + movdqa %xmm0,%xmm3
> + psllq $5,%xmm0
> + pxor %xmm0,%xmm3
> + psllq $1,%xmm0
> + pxor %xmm3,%xmm0
> + psllq $57,%xmm0
> + movdqa %xmm0,%xmm3
> + pslldq $8,%xmm0
> + psrldq $8,%xmm3
> + pxor %xmm4,%xmm0
> + pxor %xmm3,%xmm1
> +
> +
> + movdqa %xmm0,%xmm4
> + psrlq $1,%xmm0
> + pxor %xmm4,%xmm1
> + pxor %xmm0,%xmm4
> + psrlq $5,%xmm0
> + pxor %xmm4,%xmm0
> + psrlq $1,%xmm0
> + pxor %xmm1,%xmm0
> + pshufd $78,%xmm5,%xmm3
> + pshufd $78,%xmm0,%xmm4
> + pxor %xmm5,%xmm3
> + movdqu %xmm5,48(%rdi)
> + pxor %xmm0,%xmm4
> + movdqu %xmm0,64(%rdi)
> +.byte 102,15,58,15,227,8
> + movdqu %xmm4,80(%rdi)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size gcm_init_clmul,.-gcm_init_clmul
> +.globl gcm_gmult_clmul
> +.type gcm_gmult_clmul, at function
> +.align 16
> +gcm_gmult_clmul:
> +.cfi_startproc
> +.L_gmult_clmul:
> + movdqu (%rdi),%xmm0
> + movdqa .Lbswap_mask(%rip),%xmm5
> + movdqu (%rsi),%xmm2
> + movdqu 32(%rsi),%xmm4
> +.byte 102,15,56,0,197
> + movdqa %xmm0,%xmm1
> + pshufd $78,%xmm0,%xmm3
> + pxor %xmm0,%xmm3
> +.byte 102,15,58,68,194,0
> +.byte 102,15,58,68,202,17
> +.byte 102,15,58,68,220,0
> + pxor %xmm0,%xmm3
> + pxor %xmm1,%xmm3
> +
> + movdqa %xmm3,%xmm4
> + psrldq $8,%xmm3
> + pslldq $8,%xmm4
> + pxor %xmm3,%xmm1
> + pxor %xmm4,%xmm0
> +
> + movdqa %xmm0,%xmm4
> + movdqa %xmm0,%xmm3
> + psllq $5,%xmm0
> + pxor %xmm0,%xmm3
> + psllq $1,%xmm0
> + pxor %xmm3,%xmm0
> + psllq $57,%xmm0
> + movdqa %xmm0,%xmm3
> + pslldq $8,%xmm0
> + psrldq $8,%xmm3
> + pxor %xmm4,%xmm0
> + pxor %xmm3,%xmm1
> +
> +
> + movdqa %xmm0,%xmm4
> + psrlq $1,%xmm0
> + pxor %xmm4,%xmm1
> + pxor %xmm0,%xmm4
> + psrlq $5,%xmm0
> + pxor %xmm4,%xmm0
> + psrlq $1,%xmm0
> + pxor %xmm1,%xmm0
> +.byte 102,15,56,0,197
> + movdqu %xmm0,(%rdi)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size gcm_gmult_clmul,.-gcm_gmult_clmul
> +.globl gcm_ghash_clmul
> +.type gcm_ghash_clmul, at function
> +.align 32
> +gcm_ghash_clmul:
> +.cfi_startproc
> +.L_ghash_clmul:
> + movdqa .Lbswap_mask(%rip),%xmm10
> +
> + movdqu (%rdi),%xmm0
> + movdqu (%rsi),%xmm2
> + movdqu 32(%rsi),%xmm7
> +.byte 102,65,15,56,0,194
> +
> + subq $0x10,%rcx
> + jz .Lodd_tail
> +
> + movdqu 16(%rsi),%xmm6
> + movl OPENSSL_ia32cap_P+4(%rip),%eax
> + cmpq $0x30,%rcx
> + jb .Lskip4x
> +
> + andl $71303168,%eax
> + cmpl $4194304,%eax
> + je .Lskip4x
> +
> + subq $0x30,%rcx
> + movq $0xA040608020C0E000,%rax
> + movdqu 48(%rsi),%xmm14
> + movdqu 64(%rsi),%xmm15
> +
> +
> +
> +
> + movdqu 48(%rdx),%xmm3
> + movdqu 32(%rdx),%xmm11
> +.byte 102,65,15,56,0,218
> +.byte 102,69,15,56,0,218
> + movdqa %xmm3,%xmm5
> + pshufd $78,%xmm3,%xmm4
> + pxor %xmm3,%xmm4
> +.byte 102,15,58,68,218,0
> +.byte 102,15,58,68,234,17
> +.byte 102,15,58,68,231,0
> +
> + movdqa %xmm11,%xmm13
> + pshufd $78,%xmm11,%xmm12
> + pxor %xmm11,%xmm12
> +.byte 102,68,15,58,68,222,0
> +.byte 102,68,15,58,68,238,17
> +.byte 102,68,15,58,68,231,16
> + xorps %xmm11,%xmm3
> + xorps %xmm13,%xmm5
> + movups 80(%rsi),%xmm7
> + xorps %xmm12,%xmm4
> +
> + movdqu 16(%rdx),%xmm11
> + movdqu 0(%rdx),%xmm8
> +.byte 102,69,15,56,0,218
> +.byte 102,69,15,56,0,194
> + movdqa %xmm11,%xmm13
> + pshufd $78,%xmm11,%xmm12
> + pxor %xmm8,%xmm0
> + pxor %xmm11,%xmm12
> +.byte 102,69,15,58,68,222,0
> + movdqa %xmm0,%xmm1
> + pshufd $78,%xmm0,%xmm8
> + pxor %xmm0,%xmm8
> +.byte 102,69,15,58,68,238,17
> +.byte 102,68,15,58,68,231,0
> + xorps %xmm11,%xmm3
> + xorps %xmm13,%xmm5
> +
> + leaq 64(%rdx),%rdx
> + subq $0x40,%rcx
> + jc .Ltail4x
> +
> + jmp .Lmod4_loop
> +.align 32
> +.Lmod4_loop:
> +.byte 102,65,15,58,68,199,0
> + xorps %xmm12,%xmm4
> + movdqu 48(%rdx),%xmm11
> +.byte 102,69,15,56,0,218
> +.byte 102,65,15,58,68,207,17
> + xorps %xmm3,%xmm0
> + movdqu 32(%rdx),%xmm3
> + movdqa %xmm11,%xmm13
> +.byte 102,68,15,58,68,199,16
> + pshufd $78,%xmm11,%xmm12
> + xorps %xmm5,%xmm1
> + pxor %xmm11,%xmm12
> +.byte 102,65,15,56,0,218
> + movups 32(%rsi),%xmm7
> + xorps %xmm4,%xmm8
> +.byte 102,68,15,58,68,218,0
> + pshufd $78,%xmm3,%xmm4
> +
> + pxor %xmm0,%xmm8
> + movdqa %xmm3,%xmm5
> + pxor %xmm1,%xmm8
> + pxor %xmm3,%xmm4
> + movdqa %xmm8,%xmm9
> +.byte 102,68,15,58,68,234,17
> + pslldq $8,%xmm8
> + psrldq $8,%xmm9
> + pxor %xmm8,%xmm0
> + movdqa .L7_mask(%rip),%xmm8
> + pxor %xmm9,%xmm1
> +.byte 102,76,15,110,200
> +
> + pand %xmm0,%xmm8
> +.byte 102,69,15,56,0,200
> + pxor %xmm0,%xmm9
> +.byte 102,68,15,58,68,231,0
> + psllq $57,%xmm9
> + movdqa %xmm9,%xmm8
> + pslldq $8,%xmm9
> +.byte 102,15,58,68,222,0
> + psrldq $8,%xmm8
> + pxor %xmm9,%xmm0
> + pxor %xmm8,%xmm1
> + movdqu 0(%rdx),%xmm8
> +
> + movdqa %xmm0,%xmm9
> + psrlq $1,%xmm0
> +.byte 102,15,58,68,238,17
> + xorps %xmm11,%xmm3
> + movdqu 16(%rdx),%xmm11
> +.byte 102,69,15,56,0,218
> +.byte 102,15,58,68,231,16
> + xorps %xmm13,%xmm5
> + movups 80(%rsi),%xmm7
> +.byte 102,69,15,56,0,194
> + pxor %xmm9,%xmm1
> + pxor %xmm0,%xmm9
> + psrlq $5,%xmm0
> +
> + movdqa %xmm11,%xmm13
> + pxor %xmm12,%xmm4
> + pshufd $78,%xmm11,%xmm12
> + pxor %xmm9,%xmm0
> + pxor %xmm8,%xmm1
> + pxor %xmm11,%xmm12
> +.byte 102,69,15,58,68,222,0
> + psrlq $1,%xmm0
> + pxor %xmm1,%xmm0
> + movdqa %xmm0,%xmm1
> +.byte 102,69,15,58,68,238,17
> + xorps %xmm11,%xmm3
> + pshufd $78,%xmm0,%xmm8
> + pxor %xmm0,%xmm8
> +
> +.byte 102,68,15,58,68,231,0
> + xorps %xmm13,%xmm5
> +
> + leaq 64(%rdx),%rdx
> + subq $0x40,%rcx
> + jnc .Lmod4_loop
> +
> +.Ltail4x:
> +.byte 102,65,15,58,68,199,0
> +.byte 102,65,15,58,68,207,17
> +.byte 102,68,15,58,68,199,16
> + xorps %xmm12,%xmm4
> + xorps %xmm3,%xmm0
> + xorps %xmm5,%xmm1
> + pxor %xmm0,%xmm1
> + pxor %xmm4,%xmm8
> +
> + pxor %xmm1,%xmm8
> + pxor %xmm0,%xmm1
> +
> + movdqa %xmm8,%xmm9
> + psrldq $8,%xmm8
> + pslldq $8,%xmm9
> + pxor %xmm8,%xmm1
> + pxor %xmm9,%xmm0
> +
> + movdqa %xmm0,%xmm4
> + movdqa %xmm0,%xmm3
> + psllq $5,%xmm0
> + pxor %xmm0,%xmm3
> + psllq $1,%xmm0
> + pxor %xmm3,%xmm0
> + psllq $57,%xmm0
> + movdqa %xmm0,%xmm3
> + pslldq $8,%xmm0
> + psrldq $8,%xmm3
> + pxor %xmm4,%xmm0
> + pxor %xmm3,%xmm1
> +
> +
> + movdqa %xmm0,%xmm4
> + psrlq $1,%xmm0
> + pxor %xmm4,%xmm1
> + pxor %xmm0,%xmm4
> + psrlq $5,%xmm0
> + pxor %xmm4,%xmm0
> + psrlq $1,%xmm0
> + pxor %xmm1,%xmm0
> + addq $0x40,%rcx
> + jz .Ldone
> + movdqu 32(%rsi),%xmm7
> + subq $0x10,%rcx
> + jz .Lodd_tail
> +.Lskip4x:
> +
> +
> +
> +
> +
> + movdqu (%rdx),%xmm8
> + movdqu 16(%rdx),%xmm3
> +.byte 102,69,15,56,0,194
> +.byte 102,65,15,56,0,218
> + pxor %xmm8,%xmm0
> +
> + movdqa %xmm3,%xmm5
> + pshufd $78,%xmm3,%xmm4
> + pxor %xmm3,%xmm4
> +.byte 102,15,58,68,218,0
> +.byte 102,15,58,68,234,17
> +.byte 102,15,58,68,231,0
> +
> + leaq 32(%rdx),%rdx
> + nop
> + subq $0x20,%rcx
> + jbe .Leven_tail
> + nop
> + jmp .Lmod_loop
> +
> +.align 32
> +.Lmod_loop:
> + movdqa %xmm0,%xmm1
> + movdqa %xmm4,%xmm8
> + pshufd $78,%xmm0,%xmm4
> + pxor %xmm0,%xmm4
> +
> +.byte 102,15,58,68,198,0
> +.byte 102,15,58,68,206,17
> +.byte 102,15,58,68,231,16
> +
> + pxor %xmm3,%xmm0
> + pxor %xmm5,%xmm1
> + movdqu (%rdx),%xmm9
> + pxor %xmm0,%xmm8
> +.byte 102,69,15,56,0,202
> + movdqu 16(%rdx),%xmm3
> +
> + pxor %xmm1,%xmm8
> + pxor %xmm9,%xmm1
> + pxor %xmm8,%xmm4
> +.byte 102,65,15,56,0,218
> + movdqa %xmm4,%xmm8
> + psrldq $8,%xmm8
> + pslldq $8,%xmm4
> + pxor %xmm8,%xmm1
> + pxor %xmm4,%xmm0
> +
> + movdqa %xmm3,%xmm5
> +
> + movdqa %xmm0,%xmm9
> + movdqa %xmm0,%xmm8
> + psllq $5,%xmm0
> + pxor %xmm0,%xmm8
> +.byte 102,15,58,68,218,0
> + psllq $1,%xmm0
> + pxor %xmm8,%xmm0
> + psllq $57,%xmm0
> + movdqa %xmm0,%xmm8
> + pslldq $8,%xmm0
> + psrldq $8,%xmm8
> + pxor %xmm9,%xmm0
> + pshufd $78,%xmm5,%xmm4
> + pxor %xmm8,%xmm1
> + pxor %xmm5,%xmm4
> +
> + movdqa %xmm0,%xmm9
> + psrlq $1,%xmm0
> +.byte 102,15,58,68,234,17
> + pxor %xmm9,%xmm1
> + pxor %xmm0,%xmm9
> + psrlq $5,%xmm0
> + pxor %xmm9,%xmm0
> + leaq 32(%rdx),%rdx
> + psrlq $1,%xmm0
> +.byte 102,15,58,68,231,0
> + pxor %xmm1,%xmm0
> +
> + subq $0x20,%rcx
> + ja .Lmod_loop
> +
> +.Leven_tail:
> + movdqa %xmm0,%xmm1
> + movdqa %xmm4,%xmm8
> + pshufd $78,%xmm0,%xmm4
> + pxor %xmm0,%xmm4
> +
> +.byte 102,15,58,68,198,0
> +.byte 102,15,58,68,206,17
> +.byte 102,15,58,68,231,16
> +
> + pxor %xmm3,%xmm0
> + pxor %xmm5,%xmm1
> + pxor %xmm0,%xmm8
> + pxor %xmm1,%xmm8
> + pxor %xmm8,%xmm4
> + movdqa %xmm4,%xmm8
> + psrldq $8,%xmm8
> + pslldq $8,%xmm4
> + pxor %xmm8,%xmm1
> + pxor %xmm4,%xmm0
> +
> + movdqa %xmm0,%xmm4
> + movdqa %xmm0,%xmm3
> + psllq $5,%xmm0
> + pxor %xmm0,%xmm3
> + psllq $1,%xmm0
> + pxor %xmm3,%xmm0
> + psllq $57,%xmm0
> + movdqa %xmm0,%xmm3
> + pslldq $8,%xmm0
> + psrldq $8,%xmm3
> + pxor %xmm4,%xmm0
> + pxor %xmm3,%xmm1
> +
> +
> + movdqa %xmm0,%xmm4
> + psrlq $1,%xmm0
> + pxor %xmm4,%xmm1
> + pxor %xmm0,%xmm4
> + psrlq $5,%xmm0
> + pxor %xmm4,%xmm0
> + psrlq $1,%xmm0
> + pxor %xmm1,%xmm0
> + testq %rcx,%rcx
> + jnz .Ldone
> +
> +.Lodd_tail:
> + movdqu (%rdx),%xmm8
> +.byte 102,69,15,56,0,194
> + pxor %xmm8,%xmm0
> + movdqa %xmm0,%xmm1
> + pshufd $78,%xmm0,%xmm3
> + pxor %xmm0,%xmm3
> +.byte 102,15,58,68,194,0
> +.byte 102,15,58,68,202,17
> +.byte 102,15,58,68,223,0
> + pxor %xmm0,%xmm3
> + pxor %xmm1,%xmm3
> +
> + movdqa %xmm3,%xmm4
> + psrldq $8,%xmm3
> + pslldq $8,%xmm4
> + pxor %xmm3,%xmm1
> + pxor %xmm4,%xmm0
> +
> + movdqa %xmm0,%xmm4
> + movdqa %xmm0,%xmm3
> + psllq $5,%xmm0
> + pxor %xmm0,%xmm3
> + psllq $1,%xmm0
> + pxor %xmm3,%xmm0
> + psllq $57,%xmm0
> + movdqa %xmm0,%xmm3
> + pslldq $8,%xmm0
> + psrldq $8,%xmm3
> + pxor %xmm4,%xmm0
> + pxor %xmm3,%xmm1
> +
> +
> + movdqa %xmm0,%xmm4
> + psrlq $1,%xmm0
> + pxor %xmm4,%xmm1
> + pxor %xmm0,%xmm4
> + psrlq $5,%xmm0
> + pxor %xmm4,%xmm0
> + psrlq $1,%xmm0
> + pxor %xmm1,%xmm0
> +.Ldone:
> +.byte 102,65,15,56,0,194
> + movdqu %xmm0,(%rdi)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size gcm_ghash_clmul,.-gcm_ghash_clmul
> +.globl gcm_init_avx
> +.type gcm_init_avx, at function
> +.align 32
> +gcm_init_avx:
> +.cfi_startproc
> + jmp .L_init_clmul
> +.cfi_endproc
> +.size gcm_init_avx,.-gcm_init_avx
> +.globl gcm_gmult_avx
> +.type gcm_gmult_avx, at function
> +.align 32
> +gcm_gmult_avx:
> +.cfi_startproc
> + jmp .L_gmult_clmul
> +.cfi_endproc
> +.size gcm_gmult_avx,.-gcm_gmult_avx
> +.globl gcm_ghash_avx
> +.type gcm_ghash_avx, at function
> +.align 32
> +gcm_ghash_avx:
> +.cfi_startproc
> + jmp .L_ghash_clmul
> +.cfi_endproc
> +.size gcm_ghash_avx,.-gcm_ghash_avx
> +.align 64
> +.Lbswap_mask:
> +.byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
> +.L0x1c2_polynomial:
> +.byte 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xc2
> +.L7_mask:
> +.long 7,0,7,0
> +.L7_mask_poly:
> +.long 7,0,450,0
> +.align 64
> +.type .Lrem_4bit, at object
> +.Lrem_4bit:
> +.long 0,0,0,471859200,0,943718400,0,610271232
> +.long 0,1887436800,0,1822425088,0,1220542464,0,1423966208
> +.long 0,3774873600,0,4246732800,0,3644850176,0,3311403008
> +.long 0,2441084928,0,2376073216,0,2847932416,0,3051356160
> +.type .Lrem_8bit, at object
> +.Lrem_8bit:
> +.value 0x0000,0x01C2,0x0384,0x0246,0x0708,0x06CA,0x048C,0x054E
> +.value 0x0E10,0x0FD2,0x0D94,0x0C56,0x0918,0x08DA,0x0A9C,0x0B5E
> +.value 0x1C20,0x1DE2,0x1FA4,0x1E66,0x1B28,0x1AEA,0x18AC,0x196E
> +.value 0x1230,0x13F2,0x11B4,0x1076,0x1538,0x14FA,0x16BC,0x177E
> +.value 0x3840,0x3982,0x3BC4,0x3A06,0x3F48,0x3E8A,0x3CCC,0x3D0E
> +.value 0x3650,0x3792,0x35D4,0x3416,0x3158,0x309A,0x32DC,0x331E
> +.value 0x2460,0x25A2,0x27E4,0x2626,0x2368,0x22AA,0x20EC,0x212E
> +.value 0x2A70,0x2BB2,0x29F4,0x2836,0x2D78,0x2CBA,0x2EFC,0x2F3E
> +.value 0x7080,0x7142,0x7304,0x72C6,0x7788,0x764A,0x740C,0x75CE
> +.value 0x7E90,0x7F52,0x7D14,0x7CD6,0x7998,0x785A,0x7A1C,0x7BDE
> +.value 0x6CA0,0x6D62,0x6F24,0x6EE6,0x6BA8,0x6A6A,0x682C,0x69EE
> +.value 0x62B0,0x6372,0x6134,0x60F6,0x65B8,0x647A,0x663C,0x67FE
> +.value 0x48C0,0x4902,0x4B44,0x4A86,0x4FC8,0x4E0A,0x4C4C,0x4D8E
> +.value 0x46D0,0x4712,0x4554,0x4496,0x41D8,0x401A,0x425C,0x439E
> +.value 0x54E0,0x5522,0x5764,0x56A6,0x53E8,0x522A,0x506C,0x51AE
> +.value 0x5AF0,0x5B32,0x5974,0x58B6,0x5DF8,0x5C3A,0x5E7C,0x5FBE
> +.value 0xE100,0xE0C2,0xE284,0xE346,0xE608,0xE7CA,0xE58C,0xE44E
> +.value 0xEF10,0xEED2,0xEC94,0xED56,0xE818,0xE9DA,0xEB9C,0xEA5E
> +.value 0xFD20,0xFCE2,0xFEA4,0xFF66,0xFA28,0xFBEA,0xF9AC,0xF86E
> +.value 0xF330,0xF2F2,0xF0B4,0xF176,0xF438,0xF5FA,0xF7BC,0xF67E
> +.value 0xD940,0xD882,0xDAC4,0xDB06,0xDE48,0xDF8A,0xDDCC,0xDC0E
> +.value 0xD750,0xD692,0xD4D4,0xD516,0xD058,0xD19A,0xD3DC,0xD21E
> +.value 0xC560,0xC4A2,0xC6E4,0xC726,0xC268,0xC3AA,0xC1EC,0xC02E
> +.value 0xCB70,0xCAB2,0xC8F4,0xC936,0xCC78,0xCDBA,0xCFFC,0xCE3E
> +.value 0x9180,0x9042,0x9204,0x93C6,0x9688,0x974A,0x950C,0x94CE
> +.value 0x9F90,0x9E52,0x9C14,0x9DD6,0x9898,0x995A,0x9B1C,0x9ADE
> +.value 0x8DA0,0x8C62,0x8E24,0x8FE6,0x8AA8,0x8B6A,0x892C,0x88EE
> +.value 0x83B0,0x8272,0x8034,0x81F6,0x84B8,0x857A,0x873C,0x86FE
> +.value 0xA9C0,0xA802,0xAA44,0xAB86,0xAEC8,0xAF0A,0xAD4C,0xAC8E
> +.value 0xA7D0,0xA612,0xA454,0xA596,0xA0D8,0xA11A,0xA35C,0xA29E
> +.value 0xB5E0,0xB422,0xB664,0xB7A6,0xB2E8,0xB32A,0xB16C,0xB0AE
> +.value 0xBBF0,0xBA32,0xB874,0xB9B6,0xBCF8,0xBD3A,0xBF7C,0xBEBE
> +
> +.byte
> 71,72,65,83,72,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,
> 71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,
> 108,46,111,114,103,62,0
> +.align 64
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha1-mb-x86_64.S
> b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha1-mb-x86_64.S
> new file mode 100644
> index 0000000000..4572bc7227
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha1-mb-x86_64.S
> @@ -0,0 +1,2962 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/sha/asm/sha1-mb-x86_64.pl
> +#
> +# Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +
> +.globl sha1_multi_block
> +.type sha1_multi_block, at function
> +.align 32
> +sha1_multi_block:
> +.cfi_startproc
> + movq OPENSSL_ia32cap_P+4(%rip),%rcx
> + btq $61,%rcx
> + jc _shaext_shortcut
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbx,-24
> + subq $288,%rsp
> + andq $-256,%rsp
> + movq %rax,272(%rsp)
> +.cfi_escape 0x0f,0x06,0x77,0x90,0x02,0x06,0x23,0x08
> +.Lbody:
> + leaq K_XX_XX(%rip),%rbp
> + leaq 256(%rsp),%rbx
> +
> +.Loop_grande:
> + movl %edx,280(%rsp)
> + xorl %edx,%edx
> + movq 0(%rsi),%r8
> + movl 8(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,0(%rbx)
> + cmovleq %rbp,%r8
> + movq 16(%rsi),%r9
> + movl 24(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,4(%rbx)
> + cmovleq %rbp,%r9
> + movq 32(%rsi),%r10
> + movl 40(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,8(%rbx)
> + cmovleq %rbp,%r10
> + movq 48(%rsi),%r11
> + movl 56(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,12(%rbx)
> + cmovleq %rbp,%r11
> + testl %edx,%edx
> + jz .Ldone
> +
> + movdqu 0(%rdi),%xmm10
> + leaq 128(%rsp),%rax
> + movdqu 32(%rdi),%xmm11
> + movdqu 64(%rdi),%xmm12
> + movdqu 96(%rdi),%xmm13
> + movdqu 128(%rdi),%xmm14
> + movdqa 96(%rbp),%xmm5
> + movdqa -32(%rbp),%xmm15
> + jmp .Loop
> +
> +.align 32
> +.Loop:
> + movd (%r8),%xmm0
> + leaq 64(%r8),%r8
> + movd (%r9),%xmm2
> + leaq 64(%r9),%r9
> + movd (%r10),%xmm3
> + leaq 64(%r10),%r10
> + movd (%r11),%xmm4
> + leaq 64(%r11),%r11
> + punpckldq %xmm3,%xmm0
> + movd -60(%r8),%xmm1
> + punpckldq %xmm4,%xmm2
> + movd -60(%r9),%xmm9
> + punpckldq %xmm2,%xmm0
> + movd -60(%r10),%xmm8
> +.byte 102,15,56,0,197
> + movd -60(%r11),%xmm7
> + punpckldq %xmm8,%xmm1
> + movdqa %xmm10,%xmm8
> + paddd %xmm15,%xmm14
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm11,%xmm7
> + movdqa %xmm11,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm13,%xmm7
> + pand %xmm12,%xmm6
> + punpckldq %xmm9,%xmm1
> + movdqa %xmm10,%xmm9
> +
> + movdqa %xmm0,0-128(%rax)
> + paddd %xmm0,%xmm14
> + movd -56(%r8),%xmm2
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -56(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm14
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> +.byte 102,15,56,0,205
> + movd -56(%r10),%xmm8
> + por %xmm7,%xmm11
> + movd -56(%r11),%xmm7
> + punpckldq %xmm8,%xmm2
> + movdqa %xmm14,%xmm8
> + paddd %xmm15,%xmm13
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm10,%xmm7
> + movdqa %xmm10,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm12,%xmm7
> + pand %xmm11,%xmm6
> + punpckldq %xmm9,%xmm2
> + movdqa %xmm14,%xmm9
> +
> + movdqa %xmm1,16-128(%rax)
> + paddd %xmm1,%xmm13
> + movd -52(%r8),%xmm3
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -52(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm13
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> +.byte 102,15,56,0,213
> + movd -52(%r10),%xmm8
> + por %xmm7,%xmm10
> + movd -52(%r11),%xmm7
> + punpckldq %xmm8,%xmm3
> + movdqa %xmm13,%xmm8
> + paddd %xmm15,%xmm12
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm14,%xmm7
> + movdqa %xmm14,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm11,%xmm7
> + pand %xmm10,%xmm6
> + punpckldq %xmm9,%xmm3
> + movdqa %xmm13,%xmm9
> +
> + movdqa %xmm2,32-128(%rax)
> + paddd %xmm2,%xmm12
> + movd -48(%r8),%xmm4
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -48(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm12
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> +.byte 102,15,56,0,221
> + movd -48(%r10),%xmm8
> + por %xmm7,%xmm14
> + movd -48(%r11),%xmm7
> + punpckldq %xmm8,%xmm4
> + movdqa %xmm12,%xmm8
> + paddd %xmm15,%xmm11
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm13,%xmm7
> + movdqa %xmm13,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm10,%xmm7
> + pand %xmm14,%xmm6
> + punpckldq %xmm9,%xmm4
> + movdqa %xmm12,%xmm9
> +
> + movdqa %xmm3,48-128(%rax)
> + paddd %xmm3,%xmm11
> + movd -44(%r8),%xmm0
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -44(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm11
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> +.byte 102,15,56,0,229
> + movd -44(%r10),%xmm8
> + por %xmm7,%xmm13
> + movd -44(%r11),%xmm7
> + punpckldq %xmm8,%xmm0
> + movdqa %xmm11,%xmm8
> + paddd %xmm15,%xmm10
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm12,%xmm7
> + movdqa %xmm12,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm14,%xmm7
> + pand %xmm13,%xmm6
> + punpckldq %xmm9,%xmm0
> + movdqa %xmm11,%xmm9
> +
> + movdqa %xmm4,64-128(%rax)
> + paddd %xmm4,%xmm10
> + movd -40(%r8),%xmm1
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -40(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm10
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> +.byte 102,15,56,0,197
> + movd -40(%r10),%xmm8
> + por %xmm7,%xmm12
> + movd -40(%r11),%xmm7
> + punpckldq %xmm8,%xmm1
> + movdqa %xmm10,%xmm8
> + paddd %xmm15,%xmm14
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm11,%xmm7
> + movdqa %xmm11,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm13,%xmm7
> + pand %xmm12,%xmm6
> + punpckldq %xmm9,%xmm1
> + movdqa %xmm10,%xmm9
> +
> + movdqa %xmm0,80-128(%rax)
> + paddd %xmm0,%xmm14
> + movd -36(%r8),%xmm2
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -36(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm14
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> +.byte 102,15,56,0,205
> + movd -36(%r10),%xmm8
> + por %xmm7,%xmm11
> + movd -36(%r11),%xmm7
> + punpckldq %xmm8,%xmm2
> + movdqa %xmm14,%xmm8
> + paddd %xmm15,%xmm13
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm10,%xmm7
> + movdqa %xmm10,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm12,%xmm7
> + pand %xmm11,%xmm6
> + punpckldq %xmm9,%xmm2
> + movdqa %xmm14,%xmm9
> +
> + movdqa %xmm1,96-128(%rax)
> + paddd %xmm1,%xmm13
> + movd -32(%r8),%xmm3
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -32(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm13
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> +.byte 102,15,56,0,213
> + movd -32(%r10),%xmm8
> + por %xmm7,%xmm10
> + movd -32(%r11),%xmm7
> + punpckldq %xmm8,%xmm3
> + movdqa %xmm13,%xmm8
> + paddd %xmm15,%xmm12
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm14,%xmm7
> + movdqa %xmm14,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm11,%xmm7
> + pand %xmm10,%xmm6
> + punpckldq %xmm9,%xmm3
> + movdqa %xmm13,%xmm9
> +
> + movdqa %xmm2,112-128(%rax)
> + paddd %xmm2,%xmm12
> + movd -28(%r8),%xmm4
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -28(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm12
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> +.byte 102,15,56,0,221
> + movd -28(%r10),%xmm8
> + por %xmm7,%xmm14
> + movd -28(%r11),%xmm7
> + punpckldq %xmm8,%xmm4
> + movdqa %xmm12,%xmm8
> + paddd %xmm15,%xmm11
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm13,%xmm7
> + movdqa %xmm13,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm10,%xmm7
> + pand %xmm14,%xmm6
> + punpckldq %xmm9,%xmm4
> + movdqa %xmm12,%xmm9
> +
> + movdqa %xmm3,128-128(%rax)
> + paddd %xmm3,%xmm11
> + movd -24(%r8),%xmm0
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -24(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm11
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> +.byte 102,15,56,0,229
> + movd -24(%r10),%xmm8
> + por %xmm7,%xmm13
> + movd -24(%r11),%xmm7
> + punpckldq %xmm8,%xmm0
> + movdqa %xmm11,%xmm8
> + paddd %xmm15,%xmm10
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm12,%xmm7
> + movdqa %xmm12,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm14,%xmm7
> + pand %xmm13,%xmm6
> + punpckldq %xmm9,%xmm0
> + movdqa %xmm11,%xmm9
> +
> + movdqa %xmm4,144-128(%rax)
> + paddd %xmm4,%xmm10
> + movd -20(%r8),%xmm1
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -20(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm10
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> +.byte 102,15,56,0,197
> + movd -20(%r10),%xmm8
> + por %xmm7,%xmm12
> + movd -20(%r11),%xmm7
> + punpckldq %xmm8,%xmm1
> + movdqa %xmm10,%xmm8
> + paddd %xmm15,%xmm14
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm11,%xmm7
> + movdqa %xmm11,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm13,%xmm7
> + pand %xmm12,%xmm6
> + punpckldq %xmm9,%xmm1
> + movdqa %xmm10,%xmm9
> +
> + movdqa %xmm0,160-128(%rax)
> + paddd %xmm0,%xmm14
> + movd -16(%r8),%xmm2
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -16(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm14
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> +.byte 102,15,56,0,205
> + movd -16(%r10),%xmm8
> + por %xmm7,%xmm11
> + movd -16(%r11),%xmm7
> + punpckldq %xmm8,%xmm2
> + movdqa %xmm14,%xmm8
> + paddd %xmm15,%xmm13
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm10,%xmm7
> + movdqa %xmm10,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm12,%xmm7
> + pand %xmm11,%xmm6
> + punpckldq %xmm9,%xmm2
> + movdqa %xmm14,%xmm9
> +
> + movdqa %xmm1,176-128(%rax)
> + paddd %xmm1,%xmm13
> + movd -12(%r8),%xmm3
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -12(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm13
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> +.byte 102,15,56,0,213
> + movd -12(%r10),%xmm8
> + por %xmm7,%xmm10
> + movd -12(%r11),%xmm7
> + punpckldq %xmm8,%xmm3
> + movdqa %xmm13,%xmm8
> + paddd %xmm15,%xmm12
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm14,%xmm7
> + movdqa %xmm14,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm11,%xmm7
> + pand %xmm10,%xmm6
> + punpckldq %xmm9,%xmm3
> + movdqa %xmm13,%xmm9
> +
> + movdqa %xmm2,192-128(%rax)
> + paddd %xmm2,%xmm12
> + movd -8(%r8),%xmm4
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -8(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm12
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> +.byte 102,15,56,0,221
> + movd -8(%r10),%xmm8
> + por %xmm7,%xmm14
> + movd -8(%r11),%xmm7
> + punpckldq %xmm8,%xmm4
> + movdqa %xmm12,%xmm8
> + paddd %xmm15,%xmm11
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm13,%xmm7
> + movdqa %xmm13,%xmm6
> + pslld $5,%xmm8
> + pandn %xmm10,%xmm7
> + pand %xmm14,%xmm6
> + punpckldq %xmm9,%xmm4
> + movdqa %xmm12,%xmm9
> +
> + movdqa %xmm3,208-128(%rax)
> + paddd %xmm3,%xmm11
> + movd -4(%r8),%xmm0
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + por %xmm9,%xmm8
> + movd -4(%r9),%xmm9
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm11
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> +.byte 102,15,56,0,229
> + movd -4(%r10),%xmm8
> + por %xmm7,%xmm13
> + movdqa 0-128(%rax),%xmm1
> + movd -4(%r11),%xmm7
> + punpckldq %xmm8,%xmm0
> + movdqa %xmm11,%xmm8
> + paddd %xmm15,%xmm10
> + punpckldq %xmm7,%xmm9
> + movdqa %xmm12,%xmm7
> + movdqa %xmm12,%xmm6
> + pslld $5,%xmm8
> + prefetcht0 63(%r8)
> + pandn %xmm14,%xmm7
> + pand %xmm13,%xmm6
> + punpckldq %xmm9,%xmm0
> + movdqa %xmm11,%xmm9
> +
> + movdqa %xmm4,224-128(%rax)
> + paddd %xmm4,%xmm10
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> + movdqa %xmm12,%xmm7
> + prefetcht0 63(%r9)
> +
> + por %xmm9,%xmm8
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm10
> + prefetcht0 63(%r10)
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> +.byte 102,15,56,0,197
> + prefetcht0 63(%r11)
> + por %xmm7,%xmm12
> + movdqa 16-128(%rax),%xmm2
> + pxor %xmm3,%xmm1
> + movdqa 32-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + pxor 128-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + movdqa %xmm11,%xmm7
> + pslld $5,%xmm8
> + pxor %xmm3,%xmm1
> + movdqa %xmm11,%xmm6
> + pandn %xmm13,%xmm7
> + movdqa %xmm1,%xmm5
> + pand %xmm12,%xmm6
> + movdqa %xmm10,%xmm9
> + psrld $31,%xmm5
> + paddd %xmm1,%xmm1
> +
> + movdqa %xmm0,240-128(%rax)
> + paddd %xmm0,%xmm14
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> +
> + movdqa %xmm11,%xmm7
> + por %xmm9,%xmm8
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm14
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 48-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + pxor 144-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + movdqa %xmm10,%xmm7
> + pslld $5,%xmm8
> + pxor %xmm4,%xmm2
> + movdqa %xmm10,%xmm6
> + pandn %xmm12,%xmm7
> + movdqa %xmm2,%xmm5
> + pand %xmm11,%xmm6
> + movdqa %xmm14,%xmm9
> + psrld $31,%xmm5
> + paddd %xmm2,%xmm2
> +
> + movdqa %xmm1,0-128(%rax)
> + paddd %xmm1,%xmm13
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> +
> + movdqa %xmm10,%xmm7
> + por %xmm9,%xmm8
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm13
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 64-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + pxor 160-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + movdqa %xmm14,%xmm7
> + pslld $5,%xmm8
> + pxor %xmm0,%xmm3
> + movdqa %xmm14,%xmm6
> + pandn %xmm11,%xmm7
> + movdqa %xmm3,%xmm5
> + pand %xmm10,%xmm6
> + movdqa %xmm13,%xmm9
> + psrld $31,%xmm5
> + paddd %xmm3,%xmm3
> +
> + movdqa %xmm2,16-128(%rax)
> + paddd %xmm2,%xmm12
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> +
> + movdqa %xmm14,%xmm7
> + por %xmm9,%xmm8
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm12
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 80-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + pxor 176-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + movdqa %xmm13,%xmm7
> + pslld $5,%xmm8
> + pxor %xmm1,%xmm4
> + movdqa %xmm13,%xmm6
> + pandn %xmm10,%xmm7
> + movdqa %xmm4,%xmm5
> + pand %xmm14,%xmm6
> + movdqa %xmm12,%xmm9
> + psrld $31,%xmm5
> + paddd %xmm4,%xmm4
> +
> + movdqa %xmm3,32-128(%rax)
> + paddd %xmm3,%xmm11
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> +
> + movdqa %xmm13,%xmm7
> + por %xmm9,%xmm8
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm11
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 96-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + pxor 192-128(%rax),%xmm0
> + paddd %xmm15,%xmm10
> + movdqa %xmm12,%xmm7
> + pslld $5,%xmm8
> + pxor %xmm2,%xmm0
> + movdqa %xmm12,%xmm6
> + pandn %xmm14,%xmm7
> + movdqa %xmm0,%xmm5
> + pand %xmm13,%xmm6
> + movdqa %xmm11,%xmm9
> + psrld $31,%xmm5
> + paddd %xmm0,%xmm0
> +
> + movdqa %xmm4,48-128(%rax)
> + paddd %xmm4,%xmm10
> + psrld $27,%xmm9
> + pxor %xmm7,%xmm6
> +
> + movdqa %xmm12,%xmm7
> + por %xmm9,%xmm8
> + pslld $30,%xmm7
> + paddd %xmm6,%xmm10
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + movdqa 0(%rbp),%xmm15
> + pxor %xmm3,%xmm1
> + movdqa 112-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm6
> + pxor 208-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm10,%xmm9
> + movdqa %xmm0,64-128(%rax)
> + paddd %xmm0,%xmm14
> + pxor %xmm3,%xmm1
> + psrld $27,%xmm9
> + pxor %xmm12,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm1,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm14
> + paddd %xmm1,%xmm1
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 128-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm6
> + pxor 224-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm14,%xmm9
> + movdqa %xmm1,80-128(%rax)
> + paddd %xmm1,%xmm13
> + pxor %xmm4,%xmm2
> + psrld $27,%xmm9
> + pxor %xmm11,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm2,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm13
> + paddd %xmm2,%xmm2
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 144-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm6
> + pxor 240-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm13,%xmm9
> + movdqa %xmm2,96-128(%rax)
> + paddd %xmm2,%xmm12
> + pxor %xmm0,%xmm3
> + psrld $27,%xmm9
> + pxor %xmm10,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm3,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm12
> + paddd %xmm3,%xmm3
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 160-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm6
> + pxor 0-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm12,%xmm9
> + movdqa %xmm3,112-128(%rax)
> + paddd %xmm3,%xmm11
> + pxor %xmm1,%xmm4
> + psrld $27,%xmm9
> + pxor %xmm14,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm4,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm11
> + paddd %xmm4,%xmm4
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 176-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm6
> + pxor 16-128(%rax),%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm11,%xmm9
> + movdqa %xmm4,128-128(%rax)
> + paddd %xmm4,%xmm10
> + pxor %xmm2,%xmm0
> + psrld $27,%xmm9
> + pxor %xmm13,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm0,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm10
> + paddd %xmm0,%xmm0
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 192-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm6
> + pxor 32-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm10,%xmm9
> + movdqa %xmm0,144-128(%rax)
> + paddd %xmm0,%xmm14
> + pxor %xmm3,%xmm1
> + psrld $27,%xmm9
> + pxor %xmm12,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm1,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm14
> + paddd %xmm1,%xmm1
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 208-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm6
> + pxor 48-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm14,%xmm9
> + movdqa %xmm1,160-128(%rax)
> + paddd %xmm1,%xmm13
> + pxor %xmm4,%xmm2
> + psrld $27,%xmm9
> + pxor %xmm11,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm2,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm13
> + paddd %xmm2,%xmm2
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 224-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm6
> + pxor 64-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm13,%xmm9
> + movdqa %xmm2,176-128(%rax)
> + paddd %xmm2,%xmm12
> + pxor %xmm0,%xmm3
> + psrld $27,%xmm9
> + pxor %xmm10,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm3,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm12
> + paddd %xmm3,%xmm3
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 240-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm6
> + pxor 80-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm12,%xmm9
> + movdqa %xmm3,192-128(%rax)
> + paddd %xmm3,%xmm11
> + pxor %xmm1,%xmm4
> + psrld $27,%xmm9
> + pxor %xmm14,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm4,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm11
> + paddd %xmm4,%xmm4
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 0-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm6
> + pxor 96-128(%rax),%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm11,%xmm9
> + movdqa %xmm4,208-128(%rax)
> + paddd %xmm4,%xmm10
> + pxor %xmm2,%xmm0
> + psrld $27,%xmm9
> + pxor %xmm13,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm0,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm10
> + paddd %xmm0,%xmm0
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 16-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm6
> + pxor 112-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm10,%xmm9
> + movdqa %xmm0,224-128(%rax)
> + paddd %xmm0,%xmm14
> + pxor %xmm3,%xmm1
> + psrld $27,%xmm9
> + pxor %xmm12,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm1,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm14
> + paddd %xmm1,%xmm1
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 32-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm6
> + pxor 128-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm14,%xmm9
> + movdqa %xmm1,240-128(%rax)
> + paddd %xmm1,%xmm13
> + pxor %xmm4,%xmm2
> + psrld $27,%xmm9
> + pxor %xmm11,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm2,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm13
> + paddd %xmm2,%xmm2
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 48-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm6
> + pxor 144-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm13,%xmm9
> + movdqa %xmm2,0-128(%rax)
> + paddd %xmm2,%xmm12
> + pxor %xmm0,%xmm3
> + psrld $27,%xmm9
> + pxor %xmm10,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm3,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm12
> + paddd %xmm3,%xmm3
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 64-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm6
> + pxor 160-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm12,%xmm9
> + movdqa %xmm3,16-128(%rax)
> + paddd %xmm3,%xmm11
> + pxor %xmm1,%xmm4
> + psrld $27,%xmm9
> + pxor %xmm14,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm4,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm11
> + paddd %xmm4,%xmm4
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 80-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm6
> + pxor 176-128(%rax),%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm11,%xmm9
> + movdqa %xmm4,32-128(%rax)
> + paddd %xmm4,%xmm10
> + pxor %xmm2,%xmm0
> + psrld $27,%xmm9
> + pxor %xmm13,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm0,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm10
> + paddd %xmm0,%xmm0
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 96-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm6
> + pxor 192-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm10,%xmm9
> + movdqa %xmm0,48-128(%rax)
> + paddd %xmm0,%xmm14
> + pxor %xmm3,%xmm1
> + psrld $27,%xmm9
> + pxor %xmm12,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm1,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm14
> + paddd %xmm1,%xmm1
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 112-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm6
> + pxor 208-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm14,%xmm9
> + movdqa %xmm1,64-128(%rax)
> + paddd %xmm1,%xmm13
> + pxor %xmm4,%xmm2
> + psrld $27,%xmm9
> + pxor %xmm11,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm2,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm13
> + paddd %xmm2,%xmm2
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 128-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm6
> + pxor 224-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm13,%xmm9
> + movdqa %xmm2,80-128(%rax)
> + paddd %xmm2,%xmm12
> + pxor %xmm0,%xmm3
> + psrld $27,%xmm9
> + pxor %xmm10,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm3,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm12
> + paddd %xmm3,%xmm3
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 144-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm6
> + pxor 240-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm12,%xmm9
> + movdqa %xmm3,96-128(%rax)
> + paddd %xmm3,%xmm11
> + pxor %xmm1,%xmm4
> + psrld $27,%xmm9
> + pxor %xmm14,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm4,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm11
> + paddd %xmm4,%xmm4
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 160-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm6
> + pxor 0-128(%rax),%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm11,%xmm9
> + movdqa %xmm4,112-128(%rax)
> + paddd %xmm4,%xmm10
> + pxor %xmm2,%xmm0
> + psrld $27,%xmm9
> + pxor %xmm13,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm0,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm10
> + paddd %xmm0,%xmm0
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + movdqa 32(%rbp),%xmm15
> + pxor %xmm3,%xmm1
> + movdqa 176-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm7
> + pxor 16-128(%rax),%xmm1
> + pxor %xmm3,%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + movdqa %xmm10,%xmm9
> + pand %xmm12,%xmm7
> +
> + movdqa %xmm13,%xmm6
> + movdqa %xmm1,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm14
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm0,128-128(%rax)
> + paddd %xmm0,%xmm14
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm11,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm1,%xmm1
> + paddd %xmm6,%xmm14
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 192-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm7
> + pxor 32-128(%rax),%xmm2
> + pxor %xmm4,%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + movdqa %xmm14,%xmm9
> + pand %xmm11,%xmm7
> +
> + movdqa %xmm12,%xmm6
> + movdqa %xmm2,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm13
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm1,144-128(%rax)
> + paddd %xmm1,%xmm13
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm10,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm2,%xmm2
> + paddd %xmm6,%xmm13
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 208-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm7
> + pxor 48-128(%rax),%xmm3
> + pxor %xmm0,%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + movdqa %xmm13,%xmm9
> + pand %xmm10,%xmm7
> +
> + movdqa %xmm11,%xmm6
> + movdqa %xmm3,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm12
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm2,160-128(%rax)
> + paddd %xmm2,%xmm12
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm14,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm3,%xmm3
> + paddd %xmm6,%xmm12
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 224-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm7
> + pxor 64-128(%rax),%xmm4
> + pxor %xmm1,%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + movdqa %xmm12,%xmm9
> + pand %xmm14,%xmm7
> +
> + movdqa %xmm10,%xmm6
> + movdqa %xmm4,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm11
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm3,176-128(%rax)
> + paddd %xmm3,%xmm11
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm13,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm4,%xmm4
> + paddd %xmm6,%xmm11
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 240-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm7
> + pxor 80-128(%rax),%xmm0
> + pxor %xmm2,%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + movdqa %xmm11,%xmm9
> + pand %xmm13,%xmm7
> +
> + movdqa %xmm14,%xmm6
> + movdqa %xmm0,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm10
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm4,192-128(%rax)
> + paddd %xmm4,%xmm10
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm12,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm0,%xmm0
> + paddd %xmm6,%xmm10
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 0-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm7
> + pxor 96-128(%rax),%xmm1
> + pxor %xmm3,%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + movdqa %xmm10,%xmm9
> + pand %xmm12,%xmm7
> +
> + movdqa %xmm13,%xmm6
> + movdqa %xmm1,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm14
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm0,208-128(%rax)
> + paddd %xmm0,%xmm14
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm11,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm1,%xmm1
> + paddd %xmm6,%xmm14
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 16-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm7
> + pxor 112-128(%rax),%xmm2
> + pxor %xmm4,%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + movdqa %xmm14,%xmm9
> + pand %xmm11,%xmm7
> +
> + movdqa %xmm12,%xmm6
> + movdqa %xmm2,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm13
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm1,224-128(%rax)
> + paddd %xmm1,%xmm13
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm10,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm2,%xmm2
> + paddd %xmm6,%xmm13
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 32-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm7
> + pxor 128-128(%rax),%xmm3
> + pxor %xmm0,%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + movdqa %xmm13,%xmm9
> + pand %xmm10,%xmm7
> +
> + movdqa %xmm11,%xmm6
> + movdqa %xmm3,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm12
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm2,240-128(%rax)
> + paddd %xmm2,%xmm12
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm14,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm3,%xmm3
> + paddd %xmm6,%xmm12
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 48-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm7
> + pxor 144-128(%rax),%xmm4
> + pxor %xmm1,%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + movdqa %xmm12,%xmm9
> + pand %xmm14,%xmm7
> +
> + movdqa %xmm10,%xmm6
> + movdqa %xmm4,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm11
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm3,0-128(%rax)
> + paddd %xmm3,%xmm11
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm13,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm4,%xmm4
> + paddd %xmm6,%xmm11
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 64-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm7
> + pxor 160-128(%rax),%xmm0
> + pxor %xmm2,%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + movdqa %xmm11,%xmm9
> + pand %xmm13,%xmm7
> +
> + movdqa %xmm14,%xmm6
> + movdqa %xmm0,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm10
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm4,16-128(%rax)
> + paddd %xmm4,%xmm10
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm12,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm0,%xmm0
> + paddd %xmm6,%xmm10
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 80-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm7
> + pxor 176-128(%rax),%xmm1
> + pxor %xmm3,%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + movdqa %xmm10,%xmm9
> + pand %xmm12,%xmm7
> +
> + movdqa %xmm13,%xmm6
> + movdqa %xmm1,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm14
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm0,32-128(%rax)
> + paddd %xmm0,%xmm14
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm11,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm1,%xmm1
> + paddd %xmm6,%xmm14
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 96-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm7
> + pxor 192-128(%rax),%xmm2
> + pxor %xmm4,%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + movdqa %xmm14,%xmm9
> + pand %xmm11,%xmm7
> +
> + movdqa %xmm12,%xmm6
> + movdqa %xmm2,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm13
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm1,48-128(%rax)
> + paddd %xmm1,%xmm13
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm10,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm2,%xmm2
> + paddd %xmm6,%xmm13
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 112-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm7
> + pxor 208-128(%rax),%xmm3
> + pxor %xmm0,%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + movdqa %xmm13,%xmm9
> + pand %xmm10,%xmm7
> +
> + movdqa %xmm11,%xmm6
> + movdqa %xmm3,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm12
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm2,64-128(%rax)
> + paddd %xmm2,%xmm12
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm14,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm3,%xmm3
> + paddd %xmm6,%xmm12
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 128-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm7
> + pxor 224-128(%rax),%xmm4
> + pxor %xmm1,%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + movdqa %xmm12,%xmm9
> + pand %xmm14,%xmm7
> +
> + movdqa %xmm10,%xmm6
> + movdqa %xmm4,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm11
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm3,80-128(%rax)
> + paddd %xmm3,%xmm11
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm13,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm4,%xmm4
> + paddd %xmm6,%xmm11
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 144-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm7
> + pxor 240-128(%rax),%xmm0
> + pxor %xmm2,%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + movdqa %xmm11,%xmm9
> + pand %xmm13,%xmm7
> +
> + movdqa %xmm14,%xmm6
> + movdqa %xmm0,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm10
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm4,96-128(%rax)
> + paddd %xmm4,%xmm10
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm12,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm0,%xmm0
> + paddd %xmm6,%xmm10
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 160-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm7
> + pxor 0-128(%rax),%xmm1
> + pxor %xmm3,%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + movdqa %xmm10,%xmm9
> + pand %xmm12,%xmm7
> +
> + movdqa %xmm13,%xmm6
> + movdqa %xmm1,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm14
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm0,112-128(%rax)
> + paddd %xmm0,%xmm14
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm11,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm1,%xmm1
> + paddd %xmm6,%xmm14
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 176-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm7
> + pxor 16-128(%rax),%xmm2
> + pxor %xmm4,%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + movdqa %xmm14,%xmm9
> + pand %xmm11,%xmm7
> +
> + movdqa %xmm12,%xmm6
> + movdqa %xmm2,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm13
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm1,128-128(%rax)
> + paddd %xmm1,%xmm13
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm10,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm2,%xmm2
> + paddd %xmm6,%xmm13
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 192-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm7
> + pxor 32-128(%rax),%xmm3
> + pxor %xmm0,%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + movdqa %xmm13,%xmm9
> + pand %xmm10,%xmm7
> +
> + movdqa %xmm11,%xmm6
> + movdqa %xmm3,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm12
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm2,144-128(%rax)
> + paddd %xmm2,%xmm12
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm14,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm3,%xmm3
> + paddd %xmm6,%xmm12
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 208-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm7
> + pxor 48-128(%rax),%xmm4
> + pxor %xmm1,%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + movdqa %xmm12,%xmm9
> + pand %xmm14,%xmm7
> +
> + movdqa %xmm10,%xmm6
> + movdqa %xmm4,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm11
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm3,160-128(%rax)
> + paddd %xmm3,%xmm11
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm13,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm4,%xmm4
> + paddd %xmm6,%xmm11
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 224-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm7
> + pxor 64-128(%rax),%xmm0
> + pxor %xmm2,%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + movdqa %xmm11,%xmm9
> + pand %xmm13,%xmm7
> +
> + movdqa %xmm14,%xmm6
> + movdqa %xmm0,%xmm5
> + psrld $27,%xmm9
> + paddd %xmm7,%xmm10
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm4,176-128(%rax)
> + paddd %xmm4,%xmm10
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + pand %xmm12,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + paddd %xmm0,%xmm0
> + paddd %xmm6,%xmm10
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + movdqa 64(%rbp),%xmm15
> + pxor %xmm3,%xmm1
> + movdqa 240-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm6
> + pxor 80-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm10,%xmm9
> + movdqa %xmm0,192-128(%rax)
> + paddd %xmm0,%xmm14
> + pxor %xmm3,%xmm1
> + psrld $27,%xmm9
> + pxor %xmm12,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm1,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm14
> + paddd %xmm1,%xmm1
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 0-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm6
> + pxor 96-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm14,%xmm9
> + movdqa %xmm1,208-128(%rax)
> + paddd %xmm1,%xmm13
> + pxor %xmm4,%xmm2
> + psrld $27,%xmm9
> + pxor %xmm11,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm2,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm13
> + paddd %xmm2,%xmm2
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 16-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm6
> + pxor 112-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm13,%xmm9
> + movdqa %xmm2,224-128(%rax)
> + paddd %xmm2,%xmm12
> + pxor %xmm0,%xmm3
> + psrld $27,%xmm9
> + pxor %xmm10,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm3,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm12
> + paddd %xmm3,%xmm3
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 32-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm6
> + pxor 128-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm12,%xmm9
> + movdqa %xmm3,240-128(%rax)
> + paddd %xmm3,%xmm11
> + pxor %xmm1,%xmm4
> + psrld $27,%xmm9
> + pxor %xmm14,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm4,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm11
> + paddd %xmm4,%xmm4
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 48-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm6
> + pxor 144-128(%rax),%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm11,%xmm9
> + movdqa %xmm4,0-128(%rax)
> + paddd %xmm4,%xmm10
> + pxor %xmm2,%xmm0
> + psrld $27,%xmm9
> + pxor %xmm13,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm0,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm10
> + paddd %xmm0,%xmm0
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 64-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm6
> + pxor 160-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm10,%xmm9
> + movdqa %xmm0,16-128(%rax)
> + paddd %xmm0,%xmm14
> + pxor %xmm3,%xmm1
> + psrld $27,%xmm9
> + pxor %xmm12,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm1,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm14
> + paddd %xmm1,%xmm1
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 80-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm6
> + pxor 176-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm14,%xmm9
> + movdqa %xmm1,32-128(%rax)
> + paddd %xmm1,%xmm13
> + pxor %xmm4,%xmm2
> + psrld $27,%xmm9
> + pxor %xmm11,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm2,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm13
> + paddd %xmm2,%xmm2
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 96-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm6
> + pxor 192-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm13,%xmm9
> + movdqa %xmm2,48-128(%rax)
> + paddd %xmm2,%xmm12
> + pxor %xmm0,%xmm3
> + psrld $27,%xmm9
> + pxor %xmm10,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm3,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm12
> + paddd %xmm3,%xmm3
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 112-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm6
> + pxor 208-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm12,%xmm9
> + movdqa %xmm3,64-128(%rax)
> + paddd %xmm3,%xmm11
> + pxor %xmm1,%xmm4
> + psrld $27,%xmm9
> + pxor %xmm14,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm4,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm11
> + paddd %xmm4,%xmm4
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 128-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm6
> + pxor 224-128(%rax),%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm11,%xmm9
> + movdqa %xmm4,80-128(%rax)
> + paddd %xmm4,%xmm10
> + pxor %xmm2,%xmm0
> + psrld $27,%xmm9
> + pxor %xmm13,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm0,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm10
> + paddd %xmm0,%xmm0
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 144-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm6
> + pxor 240-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm10,%xmm9
> + movdqa %xmm0,96-128(%rax)
> + paddd %xmm0,%xmm14
> + pxor %xmm3,%xmm1
> + psrld $27,%xmm9
> + pxor %xmm12,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm1,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm14
> + paddd %xmm1,%xmm1
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 160-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm6
> + pxor 0-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm14,%xmm9
> + movdqa %xmm1,112-128(%rax)
> + paddd %xmm1,%xmm13
> + pxor %xmm4,%xmm2
> + psrld $27,%xmm9
> + pxor %xmm11,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm2,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm13
> + paddd %xmm2,%xmm2
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 176-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm6
> + pxor 16-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm13,%xmm9
> + paddd %xmm2,%xmm12
> + pxor %xmm0,%xmm3
> + psrld $27,%xmm9
> + pxor %xmm10,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm3,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm12
> + paddd %xmm3,%xmm3
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 192-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm6
> + pxor 32-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm12,%xmm9
> + paddd %xmm3,%xmm11
> + pxor %xmm1,%xmm4
> + psrld $27,%xmm9
> + pxor %xmm14,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm4,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm11
> + paddd %xmm4,%xmm4
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + pxor %xmm2,%xmm0
> + movdqa 208-128(%rax),%xmm2
> +
> + movdqa %xmm11,%xmm8
> + movdqa %xmm14,%xmm6
> + pxor 48-128(%rax),%xmm0
> + paddd %xmm15,%xmm10
> + pslld $5,%xmm8
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm11,%xmm9
> + paddd %xmm4,%xmm10
> + pxor %xmm2,%xmm0
> + psrld $27,%xmm9
> + pxor %xmm13,%xmm6
> + movdqa %xmm12,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm0,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm10
> + paddd %xmm0,%xmm0
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm5,%xmm0
> + por %xmm7,%xmm12
> + pxor %xmm3,%xmm1
> + movdqa 224-128(%rax),%xmm3
> +
> + movdqa %xmm10,%xmm8
> + movdqa %xmm13,%xmm6
> + pxor 64-128(%rax),%xmm1
> + paddd %xmm15,%xmm14
> + pslld $5,%xmm8
> + pxor %xmm11,%xmm6
> +
> + movdqa %xmm10,%xmm9
> + paddd %xmm0,%xmm14
> + pxor %xmm3,%xmm1
> + psrld $27,%xmm9
> + pxor %xmm12,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm1,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm14
> + paddd %xmm1,%xmm1
> +
> + psrld $2,%xmm11
> + paddd %xmm8,%xmm14
> + por %xmm5,%xmm1
> + por %xmm7,%xmm11
> + pxor %xmm4,%xmm2
> + movdqa 240-128(%rax),%xmm4
> +
> + movdqa %xmm14,%xmm8
> + movdqa %xmm12,%xmm6
> + pxor 80-128(%rax),%xmm2
> + paddd %xmm15,%xmm13
> + pslld $5,%xmm8
> + pxor %xmm10,%xmm6
> +
> + movdqa %xmm14,%xmm9
> + paddd %xmm1,%xmm13
> + pxor %xmm4,%xmm2
> + psrld $27,%xmm9
> + pxor %xmm11,%xmm6
> + movdqa %xmm10,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm2,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm13
> + paddd %xmm2,%xmm2
> +
> + psrld $2,%xmm10
> + paddd %xmm8,%xmm13
> + por %xmm5,%xmm2
> + por %xmm7,%xmm10
> + pxor %xmm0,%xmm3
> + movdqa 0-128(%rax),%xmm0
> +
> + movdqa %xmm13,%xmm8
> + movdqa %xmm11,%xmm6
> + pxor 96-128(%rax),%xmm3
> + paddd %xmm15,%xmm12
> + pslld $5,%xmm8
> + pxor %xmm14,%xmm6
> +
> + movdqa %xmm13,%xmm9
> + paddd %xmm2,%xmm12
> + pxor %xmm0,%xmm3
> + psrld $27,%xmm9
> + pxor %xmm10,%xmm6
> + movdqa %xmm14,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm3,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm12
> + paddd %xmm3,%xmm3
> +
> + psrld $2,%xmm14
> + paddd %xmm8,%xmm12
> + por %xmm5,%xmm3
> + por %xmm7,%xmm14
> + pxor %xmm1,%xmm4
> + movdqa 16-128(%rax),%xmm1
> +
> + movdqa %xmm12,%xmm8
> + movdqa %xmm10,%xmm6
> + pxor 112-128(%rax),%xmm4
> + paddd %xmm15,%xmm11
> + pslld $5,%xmm8
> + pxor %xmm13,%xmm6
> +
> + movdqa %xmm12,%xmm9
> + paddd %xmm3,%xmm11
> + pxor %xmm1,%xmm4
> + psrld $27,%xmm9
> + pxor %xmm14,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + pslld $30,%xmm7
> + movdqa %xmm4,%xmm5
> + por %xmm9,%xmm8
> + psrld $31,%xmm5
> + paddd %xmm6,%xmm11
> + paddd %xmm4,%xmm4
> +
> + psrld $2,%xmm13
> + paddd %xmm8,%xmm11
> + por %xmm5,%xmm4
> + por %xmm7,%xmm13
> + movdqa %xmm11,%xmm8
> + paddd %xmm15,%xmm10
> + movdqa %xmm14,%xmm6
> + pslld $5,%xmm8
> + pxor %xmm12,%xmm6
> +
> + movdqa %xmm11,%xmm9
> + paddd %xmm4,%xmm10
> + psrld $27,%xmm9
> + movdqa %xmm12,%xmm7
> + pxor %xmm13,%xmm6
> +
> + pslld $30,%xmm7
> + por %xmm9,%xmm8
> + paddd %xmm6,%xmm10
> +
> + psrld $2,%xmm12
> + paddd %xmm8,%xmm10
> + por %xmm7,%xmm12
> + movdqa (%rbx),%xmm0
> + movl $1,%ecx
> + cmpl 0(%rbx),%ecx
> + pxor %xmm8,%xmm8
> + cmovgeq %rbp,%r8
> + cmpl 4(%rbx),%ecx
> + movdqa %xmm0,%xmm1
> + cmovgeq %rbp,%r9
> + cmpl 8(%rbx),%ecx
> + pcmpgtd %xmm8,%xmm1
> + cmovgeq %rbp,%r10
> + cmpl 12(%rbx),%ecx
> + paddd %xmm1,%xmm0
> + cmovgeq %rbp,%r11
> +
> + movdqu 0(%rdi),%xmm6
> + pand %xmm1,%xmm10
> + movdqu 32(%rdi),%xmm7
> + pand %xmm1,%xmm11
> + paddd %xmm6,%xmm10
> + movdqu 64(%rdi),%xmm8
> + pand %xmm1,%xmm12
> + paddd %xmm7,%xmm11
> + movdqu 96(%rdi),%xmm9
> + pand %xmm1,%xmm13
> + paddd %xmm8,%xmm12
> + movdqu 128(%rdi),%xmm5
> + pand %xmm1,%xmm14
> + movdqu %xmm10,0(%rdi)
> + paddd %xmm9,%xmm13
> + movdqu %xmm11,32(%rdi)
> + paddd %xmm5,%xmm14
> + movdqu %xmm12,64(%rdi)
> + movdqu %xmm13,96(%rdi)
> + movdqu %xmm14,128(%rdi)
> +
> + movdqa %xmm0,(%rbx)
> + movdqa 96(%rbp),%xmm5
> + movdqa -32(%rbp),%xmm15
> + decl %edx
> + jnz .Loop
> +
> + movl 280(%rsp),%edx
> + leaq 16(%rdi),%rdi
> + leaq 64(%rsi),%rsi
> + decl %edx
> + jnz .Loop_grande
> +
> +.Ldone:
> + movq 272(%rsp),%rax
> +.cfi_def_cfa %rax,8
> + movq -16(%rax),%rbp
> +.cfi_restore %rbp
> + movq -8(%rax),%rbx
> +.cfi_restore %rbx
> + leaq (%rax),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha1_multi_block,.-sha1_multi_block
> +.type sha1_multi_block_shaext, at function
> +.align 32
> +sha1_multi_block_shaext:
> +.cfi_startproc
> +_shaext_shortcut:
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + subq $288,%rsp
> + shll $1,%edx
> + andq $-256,%rsp
> + leaq 64(%rdi),%rdi
> + movq %rax,272(%rsp)
> +.Lbody_shaext:
> + leaq 256(%rsp),%rbx
> + movdqa K_XX_XX+128(%rip),%xmm3
> +
> +.Loop_grande_shaext:
> + movl %edx,280(%rsp)
> + xorl %edx,%edx
> + movq 0(%rsi),%r8
> + movl 8(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,0(%rbx)
> + cmovleq %rsp,%r8
> + movq 16(%rsi),%r9
> + movl 24(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,4(%rbx)
> + cmovleq %rsp,%r9
> + testl %edx,%edx
> + jz .Ldone_shaext
> +
> + movq 0-64(%rdi),%xmm0
> + movq 32-64(%rdi),%xmm4
> + movq 64-64(%rdi),%xmm5
> + movq 96-64(%rdi),%xmm6
> + movq 128-64(%rdi),%xmm7
> +
> + punpckldq %xmm4,%xmm0
> + punpckldq %xmm6,%xmm5
> +
> + movdqa %xmm0,%xmm8
> + punpcklqdq %xmm5,%xmm0
> + punpckhqdq %xmm5,%xmm8
> +
> + pshufd $63,%xmm7,%xmm1
> + pshufd $127,%xmm7,%xmm9
> + pshufd $27,%xmm0,%xmm0
> + pshufd $27,%xmm8,%xmm8
> + jmp .Loop_shaext
> +
> +.align 32
> +.Loop_shaext:
> + movdqu 0(%r8),%xmm4
> + movdqu 0(%r9),%xmm11
> + movdqu 16(%r8),%xmm5
> + movdqu 16(%r9),%xmm12
> + movdqu 32(%r8),%xmm6
> +.byte 102,15,56,0,227
> + movdqu 32(%r9),%xmm13
> +.byte 102,68,15,56,0,219
> + movdqu 48(%r8),%xmm7
> + leaq 64(%r8),%r8
> +.byte 102,15,56,0,235
> + movdqu 48(%r9),%xmm14
> + leaq 64(%r9),%r9
> +.byte 102,68,15,56,0,227
> +
> + movdqa %xmm1,80(%rsp)
> + paddd %xmm4,%xmm1
> + movdqa %xmm9,112(%rsp)
> + paddd %xmm11,%xmm9
> + movdqa %xmm0,64(%rsp)
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,96(%rsp)
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,0
> +.byte 15,56,200,213
> +.byte 69,15,58,204,193,0
> +.byte 69,15,56,200,212
> +.byte 102,15,56,0,243
> + prefetcht0 127(%r8)
> +.byte 15,56,201,229
> +.byte 102,68,15,56,0,235
> + prefetcht0 127(%r9)
> +.byte 69,15,56,201,220
> +
> +.byte 102,15,56,0,251
> + movdqa %xmm0,%xmm1
> +.byte 102,68,15,56,0,243
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,0
> +.byte 15,56,200,206
> +.byte 69,15,58,204,194,0
> +.byte 69,15,56,200,205
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> + pxor %xmm13,%xmm11
> +.byte 69,15,56,201,229
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,0
> +.byte 15,56,200,215
> +.byte 69,15,58,204,193,0
> +.byte 69,15,56,200,214
> +.byte 15,56,202,231
> +.byte 69,15,56,202,222
> + pxor %xmm7,%xmm5
> +.byte 15,56,201,247
> + pxor %xmm14,%xmm12
> +.byte 69,15,56,201,238
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,0
> +.byte 15,56,200,204
> +.byte 69,15,58,204,194,0
> +.byte 69,15,56,200,203
> +.byte 15,56,202,236
> +.byte 69,15,56,202,227
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,252
> + pxor %xmm11,%xmm13
> +.byte 69,15,56,201,243
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,0
> +.byte 15,56,200,213
> +.byte 69,15,58,204,193,0
> +.byte 69,15,56,200,212
> +.byte 15,56,202,245
> +.byte 69,15,56,202,236
> + pxor %xmm5,%xmm7
> +.byte 15,56,201,229
> + pxor %xmm12,%xmm14
> +.byte 69,15,56,201,220
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,1
> +.byte 15,56,200,206
> +.byte 69,15,58,204,194,1
> +.byte 69,15,56,200,205
> +.byte 15,56,202,254
> +.byte 69,15,56,202,245
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> + pxor %xmm13,%xmm11
> +.byte 69,15,56,201,229
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,1
> +.byte 15,56,200,215
> +.byte 69,15,58,204,193,1
> +.byte 69,15,56,200,214
> +.byte 15,56,202,231
> +.byte 69,15,56,202,222
> + pxor %xmm7,%xmm5
> +.byte 15,56,201,247
> + pxor %xmm14,%xmm12
> +.byte 69,15,56,201,238
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,1
> +.byte 15,56,200,204
> +.byte 69,15,58,204,194,1
> +.byte 69,15,56,200,203
> +.byte 15,56,202,236
> +.byte 69,15,56,202,227
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,252
> + pxor %xmm11,%xmm13
> +.byte 69,15,56,201,243
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,1
> +.byte 15,56,200,213
> +.byte 69,15,58,204,193,1
> +.byte 69,15,56,200,212
> +.byte 15,56,202,245
> +.byte 69,15,56,202,236
> + pxor %xmm5,%xmm7
> +.byte 15,56,201,229
> + pxor %xmm12,%xmm14
> +.byte 69,15,56,201,220
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,1
> +.byte 15,56,200,206
> +.byte 69,15,58,204,194,1
> +.byte 69,15,56,200,205
> +.byte 15,56,202,254
> +.byte 69,15,56,202,245
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> + pxor %xmm13,%xmm11
> +.byte 69,15,56,201,229
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,2
> +.byte 15,56,200,215
> +.byte 69,15,58,204,193,2
> +.byte 69,15,56,200,214
> +.byte 15,56,202,231
> +.byte 69,15,56,202,222
> + pxor %xmm7,%xmm5
> +.byte 15,56,201,247
> + pxor %xmm14,%xmm12
> +.byte 69,15,56,201,238
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,2
> +.byte 15,56,200,204
> +.byte 69,15,58,204,194,2
> +.byte 69,15,56,200,203
> +.byte 15,56,202,236
> +.byte 69,15,56,202,227
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,252
> + pxor %xmm11,%xmm13
> +.byte 69,15,56,201,243
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,2
> +.byte 15,56,200,213
> +.byte 69,15,58,204,193,2
> +.byte 69,15,56,200,212
> +.byte 15,56,202,245
> +.byte 69,15,56,202,236
> + pxor %xmm5,%xmm7
> +.byte 15,56,201,229
> + pxor %xmm12,%xmm14
> +.byte 69,15,56,201,220
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,2
> +.byte 15,56,200,206
> +.byte 69,15,58,204,194,2
> +.byte 69,15,56,200,205
> +.byte 15,56,202,254
> +.byte 69,15,56,202,245
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> + pxor %xmm13,%xmm11
> +.byte 69,15,56,201,229
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,2
> +.byte 15,56,200,215
> +.byte 69,15,58,204,193,2
> +.byte 69,15,56,200,214
> +.byte 15,56,202,231
> +.byte 69,15,56,202,222
> + pxor %xmm7,%xmm5
> +.byte 15,56,201,247
> + pxor %xmm14,%xmm12
> +.byte 69,15,56,201,238
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,3
> +.byte 15,56,200,204
> +.byte 69,15,58,204,194,3
> +.byte 69,15,56,200,203
> +.byte 15,56,202,236
> +.byte 69,15,56,202,227
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,252
> + pxor %xmm11,%xmm13
> +.byte 69,15,56,201,243
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,3
> +.byte 15,56,200,213
> +.byte 69,15,58,204,193,3
> +.byte 69,15,56,200,212
> +.byte 15,56,202,245
> +.byte 69,15,56,202,236
> + pxor %xmm5,%xmm7
> + pxor %xmm12,%xmm14
> +
> + movl $1,%ecx
> + pxor %xmm4,%xmm4
> + cmpl 0(%rbx),%ecx
> + cmovgeq %rsp,%r8
> +
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,3
> +.byte 15,56,200,206
> +.byte 69,15,58,204,194,3
> +.byte 69,15,56,200,205
> +.byte 15,56,202,254
> +.byte 69,15,56,202,245
> +
> + cmpl 4(%rbx),%ecx
> + cmovgeq %rsp,%r9
> + movq (%rbx),%xmm6
> +
> + movdqa %xmm0,%xmm2
> + movdqa %xmm8,%xmm10
> +.byte 15,58,204,193,3
> +.byte 15,56,200,215
> +.byte 69,15,58,204,193,3
> +.byte 69,15,56,200,214
> +
> + pshufd $0x00,%xmm6,%xmm11
> + pshufd $0x55,%xmm6,%xmm12
> + movdqa %xmm6,%xmm7
> + pcmpgtd %xmm4,%xmm11
> + pcmpgtd %xmm4,%xmm12
> +
> + movdqa %xmm0,%xmm1
> + movdqa %xmm8,%xmm9
> +.byte 15,58,204,194,3
> +.byte 15,56,200,204
> +.byte 69,15,58,204,194,3
> +.byte 68,15,56,200,204
> +
> + pcmpgtd %xmm4,%xmm7
> + pand %xmm11,%xmm0
> + pand %xmm11,%xmm1
> + pand %xmm12,%xmm8
> + pand %xmm12,%xmm9
> + paddd %xmm7,%xmm6
> +
> + paddd 64(%rsp),%xmm0
> + paddd 80(%rsp),%xmm1
> + paddd 96(%rsp),%xmm8
> + paddd 112(%rsp),%xmm9
> +
> + movq %xmm6,(%rbx)
> + decl %edx
> + jnz .Loop_shaext
> +
> + movl 280(%rsp),%edx
> +
> + pshufd $27,%xmm0,%xmm0
> + pshufd $27,%xmm8,%xmm8
> +
> + movdqa %xmm0,%xmm6
> + punpckldq %xmm8,%xmm0
> + punpckhdq %xmm8,%xmm6
> + punpckhdq %xmm9,%xmm1
> + movq %xmm0,0-64(%rdi)
> + psrldq $8,%xmm0
> + movq %xmm6,64-64(%rdi)
> + psrldq $8,%xmm6
> + movq %xmm0,32-64(%rdi)
> + psrldq $8,%xmm1
> + movq %xmm6,96-64(%rdi)
> + movq %xmm1,128-64(%rdi)
> +
> + leaq 8(%rdi),%rdi
> + leaq 32(%rsi),%rsi
> + decl %edx
> + jnz .Loop_grande_shaext
> +
> +.Ldone_shaext:
> +
> + movq -16(%rax),%rbp
> +.cfi_restore %rbp
> + movq -8(%rax),%rbx
> +.cfi_restore %rbx
> + leaq (%rax),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue_shaext:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha1_multi_block_shaext,.-sha1_multi_block_shaext
> +
> +.align 256
> +.long 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> +.long 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> +K_XX_XX:
> +.long 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> +.long 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> +.long 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> +.long 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> +.long 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> +.long 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.byte 0xf,0xe,0xd,0xc,0xb,0xa,0x9,0x8,0x7,0x6,0x5,0x4,0x3,0x2,0x1,0x0
> +.byte
> 83,72,65,49,32,109,117,108,116,105,45,98,108,111,99,107,32,116,114,97,110,
> 115,102,111,114,109,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,
> 84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,11
> 5,115,108,46,111,114,103,62,0
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha1-x86_64.S
> b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha1-x86_64.S
> new file mode 100644
> index 0000000000..0b59726ae4
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha1-x86_64.S
> @@ -0,0 +1,2631 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/sha/asm/sha1-x86_64.pl
> +#
> +# Copyright 2006-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +.globl sha1_block_data_order
> +.type sha1_block_data_order, at function
> +.align 16
> +sha1_block_data_order:
> +.cfi_startproc
> + movl OPENSSL_ia32cap_P+0(%rip),%r9d
> + movl OPENSSL_ia32cap_P+4(%rip),%r8d
> + movl OPENSSL_ia32cap_P+8(%rip),%r10d
> + testl $512,%r8d
> + jz .Lialu
> + testl $536870912,%r10d
> + jnz _shaext_shortcut
> + jmp _ssse3_shortcut
> +
> +.align 16
> +.Lialu:
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_offset %r14,-48
> + movq %rdi,%r8
> + subq $72,%rsp
> + movq %rsi,%r9
> + andq $-64,%rsp
> + movq %rdx,%r10
> + movq %rax,64(%rsp)
> +.cfi_escape 0x0f,0x06,0x77,0xc0,0x00,0x06,0x23,0x08
> +.Lprologue:
> +
> + movl 0(%r8),%esi
> + movl 4(%r8),%edi
> + movl 8(%r8),%r11d
> + movl 12(%r8),%r12d
> + movl 16(%r8),%r13d
> + jmp .Lloop
> +
> +.align 16
> +.Lloop:
> + movl 0(%r9),%edx
> + bswapl %edx
> + movl 4(%r9),%ebp
> + movl %r12d,%eax
> + movl %edx,0(%rsp)
> + movl %esi,%ecx
> + bswapl %ebp
> + xorl %r11d,%eax
> + roll $5,%ecx
> + andl %edi,%eax
> + leal 1518500249(%rdx,%r13,1),%r13d
> + addl %ecx,%r13d
> + xorl %r12d,%eax
> + roll $30,%edi
> + addl %eax,%r13d
> + movl 8(%r9),%r14d
> + movl %r11d,%eax
> + movl %ebp,4(%rsp)
> + movl %r13d,%ecx
> + bswapl %r14d
> + xorl %edi,%eax
> + roll $5,%ecx
> + andl %esi,%eax
> + leal 1518500249(%rbp,%r12,1),%r12d
> + addl %ecx,%r12d
> + xorl %r11d,%eax
> + roll $30,%esi
> + addl %eax,%r12d
> + movl 12(%r9),%edx
> + movl %edi,%eax
> + movl %r14d,8(%rsp)
> + movl %r12d,%ecx
> + bswapl %edx
> + xorl %esi,%eax
> + roll $5,%ecx
> + andl %r13d,%eax
> + leal 1518500249(%r14,%r11,1),%r11d
> + addl %ecx,%r11d
> + xorl %edi,%eax
> + roll $30,%r13d
> + addl %eax,%r11d
> + movl 16(%r9),%ebp
> + movl %esi,%eax
> + movl %edx,12(%rsp)
> + movl %r11d,%ecx
> + bswapl %ebp
> + xorl %r13d,%eax
> + roll $5,%ecx
> + andl %r12d,%eax
> + leal 1518500249(%rdx,%rdi,1),%edi
> + addl %ecx,%edi
> + xorl %esi,%eax
> + roll $30,%r12d
> + addl %eax,%edi
> + movl 20(%r9),%r14d
> + movl %r13d,%eax
> + movl %ebp,16(%rsp)
> + movl %edi,%ecx
> + bswapl %r14d
> + xorl %r12d,%eax
> + roll $5,%ecx
> + andl %r11d,%eax
> + leal 1518500249(%rbp,%rsi,1),%esi
> + addl %ecx,%esi
> + xorl %r13d,%eax
> + roll $30,%r11d
> + addl %eax,%esi
> + movl 24(%r9),%edx
> + movl %r12d,%eax
> + movl %r14d,20(%rsp)
> + movl %esi,%ecx
> + bswapl %edx
> + xorl %r11d,%eax
> + roll $5,%ecx
> + andl %edi,%eax
> + leal 1518500249(%r14,%r13,1),%r13d
> + addl %ecx,%r13d
> + xorl %r12d,%eax
> + roll $30,%edi
> + addl %eax,%r13d
> + movl 28(%r9),%ebp
> + movl %r11d,%eax
> + movl %edx,24(%rsp)
> + movl %r13d,%ecx
> + bswapl %ebp
> + xorl %edi,%eax
> + roll $5,%ecx
> + andl %esi,%eax
> + leal 1518500249(%rdx,%r12,1),%r12d
> + addl %ecx,%r12d
> + xorl %r11d,%eax
> + roll $30,%esi
> + addl %eax,%r12d
> + movl 32(%r9),%r14d
> + movl %edi,%eax
> + movl %ebp,28(%rsp)
> + movl %r12d,%ecx
> + bswapl %r14d
> + xorl %esi,%eax
> + roll $5,%ecx
> + andl %r13d,%eax
> + leal 1518500249(%rbp,%r11,1),%r11d
> + addl %ecx,%r11d
> + xorl %edi,%eax
> + roll $30,%r13d
> + addl %eax,%r11d
> + movl 36(%r9),%edx
> + movl %esi,%eax
> + movl %r14d,32(%rsp)
> + movl %r11d,%ecx
> + bswapl %edx
> + xorl %r13d,%eax
> + roll $5,%ecx
> + andl %r12d,%eax
> + leal 1518500249(%r14,%rdi,1),%edi
> + addl %ecx,%edi
> + xorl %esi,%eax
> + roll $30,%r12d
> + addl %eax,%edi
> + movl 40(%r9),%ebp
> + movl %r13d,%eax
> + movl %edx,36(%rsp)
> + movl %edi,%ecx
> + bswapl %ebp
> + xorl %r12d,%eax
> + roll $5,%ecx
> + andl %r11d,%eax
> + leal 1518500249(%rdx,%rsi,1),%esi
> + addl %ecx,%esi
> + xorl %r13d,%eax
> + roll $30,%r11d
> + addl %eax,%esi
> + movl 44(%r9),%r14d
> + movl %r12d,%eax
> + movl %ebp,40(%rsp)
> + movl %esi,%ecx
> + bswapl %r14d
> + xorl %r11d,%eax
> + roll $5,%ecx
> + andl %edi,%eax
> + leal 1518500249(%rbp,%r13,1),%r13d
> + addl %ecx,%r13d
> + xorl %r12d,%eax
> + roll $30,%edi
> + addl %eax,%r13d
> + movl 48(%r9),%edx
> + movl %r11d,%eax
> + movl %r14d,44(%rsp)
> + movl %r13d,%ecx
> + bswapl %edx
> + xorl %edi,%eax
> + roll $5,%ecx
> + andl %esi,%eax
> + leal 1518500249(%r14,%r12,1),%r12d
> + addl %ecx,%r12d
> + xorl %r11d,%eax
> + roll $30,%esi
> + addl %eax,%r12d
> + movl 52(%r9),%ebp
> + movl %edi,%eax
> + movl %edx,48(%rsp)
> + movl %r12d,%ecx
> + bswapl %ebp
> + xorl %esi,%eax
> + roll $5,%ecx
> + andl %r13d,%eax
> + leal 1518500249(%rdx,%r11,1),%r11d
> + addl %ecx,%r11d
> + xorl %edi,%eax
> + roll $30,%r13d
> + addl %eax,%r11d
> + movl 56(%r9),%r14d
> + movl %esi,%eax
> + movl %ebp,52(%rsp)
> + movl %r11d,%ecx
> + bswapl %r14d
> + xorl %r13d,%eax
> + roll $5,%ecx
> + andl %r12d,%eax
> + leal 1518500249(%rbp,%rdi,1),%edi
> + addl %ecx,%edi
> + xorl %esi,%eax
> + roll $30,%r12d
> + addl %eax,%edi
> + movl 60(%r9),%edx
> + movl %r13d,%eax
> + movl %r14d,56(%rsp)
> + movl %edi,%ecx
> + bswapl %edx
> + xorl %r12d,%eax
> + roll $5,%ecx
> + andl %r11d,%eax
> + leal 1518500249(%r14,%rsi,1),%esi
> + addl %ecx,%esi
> + xorl %r13d,%eax
> + roll $30,%r11d
> + addl %eax,%esi
> + xorl 0(%rsp),%ebp
> + movl %r12d,%eax
> + movl %edx,60(%rsp)
> + movl %esi,%ecx
> + xorl 8(%rsp),%ebp
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 32(%rsp),%ebp
> + andl %edi,%eax
> + leal 1518500249(%rdx,%r13,1),%r13d
> + roll $30,%edi
> + xorl %r12d,%eax
> + addl %ecx,%r13d
> + roll $1,%ebp
> + addl %eax,%r13d
> + xorl 4(%rsp),%r14d
> + movl %r11d,%eax
> + movl %ebp,0(%rsp)
> + movl %r13d,%ecx
> + xorl 12(%rsp),%r14d
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 36(%rsp),%r14d
> + andl %esi,%eax
> + leal 1518500249(%rbp,%r12,1),%r12d
> + roll $30,%esi
> + xorl %r11d,%eax
> + addl %ecx,%r12d
> + roll $1,%r14d
> + addl %eax,%r12d
> + xorl 8(%rsp),%edx
> + movl %edi,%eax
> + movl %r14d,4(%rsp)
> + movl %r12d,%ecx
> + xorl 16(%rsp),%edx
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 40(%rsp),%edx
> + andl %r13d,%eax
> + leal 1518500249(%r14,%r11,1),%r11d
> + roll $30,%r13d
> + xorl %edi,%eax
> + addl %ecx,%r11d
> + roll $1,%edx
> + addl %eax,%r11d
> + xorl 12(%rsp),%ebp
> + movl %esi,%eax
> + movl %edx,8(%rsp)
> + movl %r11d,%ecx
> + xorl 20(%rsp),%ebp
> + xorl %r13d,%eax
> + roll $5,%ecx
> + xorl 44(%rsp),%ebp
> + andl %r12d,%eax
> + leal 1518500249(%rdx,%rdi,1),%edi
> + roll $30,%r12d
> + xorl %esi,%eax
> + addl %ecx,%edi
> + roll $1,%ebp
> + addl %eax,%edi
> + xorl 16(%rsp),%r14d
> + movl %r13d,%eax
> + movl %ebp,12(%rsp)
> + movl %edi,%ecx
> + xorl 24(%rsp),%r14d
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 48(%rsp),%r14d
> + andl %r11d,%eax
> + leal 1518500249(%rbp,%rsi,1),%esi
> + roll $30,%r11d
> + xorl %r13d,%eax
> + addl %ecx,%esi
> + roll $1,%r14d
> + addl %eax,%esi
> + xorl 20(%rsp),%edx
> + movl %edi,%eax
> + movl %r14d,16(%rsp)
> + movl %esi,%ecx
> + xorl 28(%rsp),%edx
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 52(%rsp),%edx
> + leal 1859775393(%r14,%r13,1),%r13d
> + xorl %r11d,%eax
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %eax,%r13d
> + roll $1,%edx
> + xorl 24(%rsp),%ebp
> + movl %esi,%eax
> + movl %edx,20(%rsp)
> + movl %r13d,%ecx
> + xorl 32(%rsp),%ebp
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 56(%rsp),%ebp
> + leal 1859775393(%rdx,%r12,1),%r12d
> + xorl %edi,%eax
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %eax,%r12d
> + roll $1,%ebp
> + xorl 28(%rsp),%r14d
> + movl %r13d,%eax
> + movl %ebp,24(%rsp)
> + movl %r12d,%ecx
> + xorl 36(%rsp),%r14d
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 60(%rsp),%r14d
> + leal 1859775393(%rbp,%r11,1),%r11d
> + xorl %esi,%eax
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %eax,%r11d
> + roll $1,%r14d
> + xorl 32(%rsp),%edx
> + movl %r12d,%eax
> + movl %r14d,28(%rsp)
> + movl %r11d,%ecx
> + xorl 40(%rsp),%edx
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 0(%rsp),%edx
> + leal 1859775393(%r14,%rdi,1),%edi
> + xorl %r13d,%eax
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %eax,%edi
> + roll $1,%edx
> + xorl 36(%rsp),%ebp
> + movl %r11d,%eax
> + movl %edx,32(%rsp)
> + movl %edi,%ecx
> + xorl 44(%rsp),%ebp
> + xorl %r13d,%eax
> + roll $5,%ecx
> + xorl 4(%rsp),%ebp
> + leal 1859775393(%rdx,%rsi,1),%esi
> + xorl %r12d,%eax
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %eax,%esi
> + roll $1,%ebp
> + xorl 40(%rsp),%r14d
> + movl %edi,%eax
> + movl %ebp,36(%rsp)
> + movl %esi,%ecx
> + xorl 48(%rsp),%r14d
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 8(%rsp),%r14d
> + leal 1859775393(%rbp,%r13,1),%r13d
> + xorl %r11d,%eax
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %eax,%r13d
> + roll $1,%r14d
> + xorl 44(%rsp),%edx
> + movl %esi,%eax
> + movl %r14d,40(%rsp)
> + movl %r13d,%ecx
> + xorl 52(%rsp),%edx
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 12(%rsp),%edx
> + leal 1859775393(%r14,%r12,1),%r12d
> + xorl %edi,%eax
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %eax,%r12d
> + roll $1,%edx
> + xorl 48(%rsp),%ebp
> + movl %r13d,%eax
> + movl %edx,44(%rsp)
> + movl %r12d,%ecx
> + xorl 56(%rsp),%ebp
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 16(%rsp),%ebp
> + leal 1859775393(%rdx,%r11,1),%r11d
> + xorl %esi,%eax
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %eax,%r11d
> + roll $1,%ebp
> + xorl 52(%rsp),%r14d
> + movl %r12d,%eax
> + movl %ebp,48(%rsp)
> + movl %r11d,%ecx
> + xorl 60(%rsp),%r14d
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 20(%rsp),%r14d
> + leal 1859775393(%rbp,%rdi,1),%edi
> + xorl %r13d,%eax
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %eax,%edi
> + roll $1,%r14d
> + xorl 56(%rsp),%edx
> + movl %r11d,%eax
> + movl %r14d,52(%rsp)
> + movl %edi,%ecx
> + xorl 0(%rsp),%edx
> + xorl %r13d,%eax
> + roll $5,%ecx
> + xorl 24(%rsp),%edx
> + leal 1859775393(%r14,%rsi,1),%esi
> + xorl %r12d,%eax
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %eax,%esi
> + roll $1,%edx
> + xorl 60(%rsp),%ebp
> + movl %edi,%eax
> + movl %edx,56(%rsp)
> + movl %esi,%ecx
> + xorl 4(%rsp),%ebp
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 28(%rsp),%ebp
> + leal 1859775393(%rdx,%r13,1),%r13d
> + xorl %r11d,%eax
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %eax,%r13d
> + roll $1,%ebp
> + xorl 0(%rsp),%r14d
> + movl %esi,%eax
> + movl %ebp,60(%rsp)
> + movl %r13d,%ecx
> + xorl 8(%rsp),%r14d
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 32(%rsp),%r14d
> + leal 1859775393(%rbp,%r12,1),%r12d
> + xorl %edi,%eax
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %eax,%r12d
> + roll $1,%r14d
> + xorl 4(%rsp),%edx
> + movl %r13d,%eax
> + movl %r14d,0(%rsp)
> + movl %r12d,%ecx
> + xorl 12(%rsp),%edx
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 36(%rsp),%edx
> + leal 1859775393(%r14,%r11,1),%r11d
> + xorl %esi,%eax
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %eax,%r11d
> + roll $1,%edx
> + xorl 8(%rsp),%ebp
> + movl %r12d,%eax
> + movl %edx,4(%rsp)
> + movl %r11d,%ecx
> + xorl 16(%rsp),%ebp
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 40(%rsp),%ebp
> + leal 1859775393(%rdx,%rdi,1),%edi
> + xorl %r13d,%eax
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %eax,%edi
> + roll $1,%ebp
> + xorl 12(%rsp),%r14d
> + movl %r11d,%eax
> + movl %ebp,8(%rsp)
> + movl %edi,%ecx
> + xorl 20(%rsp),%r14d
> + xorl %r13d,%eax
> + roll $5,%ecx
> + xorl 44(%rsp),%r14d
> + leal 1859775393(%rbp,%rsi,1),%esi
> + xorl %r12d,%eax
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %eax,%esi
> + roll $1,%r14d
> + xorl 16(%rsp),%edx
> + movl %edi,%eax
> + movl %r14d,12(%rsp)
> + movl %esi,%ecx
> + xorl 24(%rsp),%edx
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 48(%rsp),%edx
> + leal 1859775393(%r14,%r13,1),%r13d
> + xorl %r11d,%eax
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %eax,%r13d
> + roll $1,%edx
> + xorl 20(%rsp),%ebp
> + movl %esi,%eax
> + movl %edx,16(%rsp)
> + movl %r13d,%ecx
> + xorl 28(%rsp),%ebp
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 52(%rsp),%ebp
> + leal 1859775393(%rdx,%r12,1),%r12d
> + xorl %edi,%eax
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %eax,%r12d
> + roll $1,%ebp
> + xorl 24(%rsp),%r14d
> + movl %r13d,%eax
> + movl %ebp,20(%rsp)
> + movl %r12d,%ecx
> + xorl 32(%rsp),%r14d
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 56(%rsp),%r14d
> + leal 1859775393(%rbp,%r11,1),%r11d
> + xorl %esi,%eax
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %eax,%r11d
> + roll $1,%r14d
> + xorl 28(%rsp),%edx
> + movl %r12d,%eax
> + movl %r14d,24(%rsp)
> + movl %r11d,%ecx
> + xorl 36(%rsp),%edx
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 60(%rsp),%edx
> + leal 1859775393(%r14,%rdi,1),%edi
> + xorl %r13d,%eax
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %eax,%edi
> + roll $1,%edx
> + xorl 32(%rsp),%ebp
> + movl %r11d,%eax
> + movl %edx,28(%rsp)
> + movl %edi,%ecx
> + xorl 40(%rsp),%ebp
> + xorl %r13d,%eax
> + roll $5,%ecx
> + xorl 0(%rsp),%ebp
> + leal 1859775393(%rdx,%rsi,1),%esi
> + xorl %r12d,%eax
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %eax,%esi
> + roll $1,%ebp
> + xorl 36(%rsp),%r14d
> + movl %r12d,%eax
> + movl %ebp,32(%rsp)
> + movl %r12d,%ebx
> + xorl 44(%rsp),%r14d
> + andl %r11d,%eax
> + movl %esi,%ecx
> + xorl 4(%rsp),%r14d
> + leal -1894007588(%rbp,%r13,1),%r13d
> + xorl %r11d,%ebx
> + roll $5,%ecx
> + addl %eax,%r13d
> + roll $1,%r14d
> + andl %edi,%ebx
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %ebx,%r13d
> + xorl 40(%rsp),%edx
> + movl %r11d,%eax
> + movl %r14d,36(%rsp)
> + movl %r11d,%ebx
> + xorl 48(%rsp),%edx
> + andl %edi,%eax
> + movl %r13d,%ecx
> + xorl 8(%rsp),%edx
> + leal -1894007588(%r14,%r12,1),%r12d
> + xorl %edi,%ebx
> + roll $5,%ecx
> + addl %eax,%r12d
> + roll $1,%edx
> + andl %esi,%ebx
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %ebx,%r12d
> + xorl 44(%rsp),%ebp
> + movl %edi,%eax
> + movl %edx,40(%rsp)
> + movl %edi,%ebx
> + xorl 52(%rsp),%ebp
> + andl %esi,%eax
> + movl %r12d,%ecx
> + xorl 12(%rsp),%ebp
> + leal -1894007588(%rdx,%r11,1),%r11d
> + xorl %esi,%ebx
> + roll $5,%ecx
> + addl %eax,%r11d
> + roll $1,%ebp
> + andl %r13d,%ebx
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %ebx,%r11d
> + xorl 48(%rsp),%r14d
> + movl %esi,%eax
> + movl %ebp,44(%rsp)
> + movl %esi,%ebx
> + xorl 56(%rsp),%r14d
> + andl %r13d,%eax
> + movl %r11d,%ecx
> + xorl 16(%rsp),%r14d
> + leal -1894007588(%rbp,%rdi,1),%edi
> + xorl %r13d,%ebx
> + roll $5,%ecx
> + addl %eax,%edi
> + roll $1,%r14d
> + andl %r12d,%ebx
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %ebx,%edi
> + xorl 52(%rsp),%edx
> + movl %r13d,%eax
> + movl %r14d,48(%rsp)
> + movl %r13d,%ebx
> + xorl 60(%rsp),%edx
> + andl %r12d,%eax
> + movl %edi,%ecx
> + xorl 20(%rsp),%edx
> + leal -1894007588(%r14,%rsi,1),%esi
> + xorl %r12d,%ebx
> + roll $5,%ecx
> + addl %eax,%esi
> + roll $1,%edx
> + andl %r11d,%ebx
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %ebx,%esi
> + xorl 56(%rsp),%ebp
> + movl %r12d,%eax
> + movl %edx,52(%rsp)
> + movl %r12d,%ebx
> + xorl 0(%rsp),%ebp
> + andl %r11d,%eax
> + movl %esi,%ecx
> + xorl 24(%rsp),%ebp
> + leal -1894007588(%rdx,%r13,1),%r13d
> + xorl %r11d,%ebx
> + roll $5,%ecx
> + addl %eax,%r13d
> + roll $1,%ebp
> + andl %edi,%ebx
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %ebx,%r13d
> + xorl 60(%rsp),%r14d
> + movl %r11d,%eax
> + movl %ebp,56(%rsp)
> + movl %r11d,%ebx
> + xorl 4(%rsp),%r14d
> + andl %edi,%eax
> + movl %r13d,%ecx
> + xorl 28(%rsp),%r14d
> + leal -1894007588(%rbp,%r12,1),%r12d
> + xorl %edi,%ebx
> + roll $5,%ecx
> + addl %eax,%r12d
> + roll $1,%r14d
> + andl %esi,%ebx
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %ebx,%r12d
> + xorl 0(%rsp),%edx
> + movl %edi,%eax
> + movl %r14d,60(%rsp)
> + movl %edi,%ebx
> + xorl 8(%rsp),%edx
> + andl %esi,%eax
> + movl %r12d,%ecx
> + xorl 32(%rsp),%edx
> + leal -1894007588(%r14,%r11,1),%r11d
> + xorl %esi,%ebx
> + roll $5,%ecx
> + addl %eax,%r11d
> + roll $1,%edx
> + andl %r13d,%ebx
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %ebx,%r11d
> + xorl 4(%rsp),%ebp
> + movl %esi,%eax
> + movl %edx,0(%rsp)
> + movl %esi,%ebx
> + xorl 12(%rsp),%ebp
> + andl %r13d,%eax
> + movl %r11d,%ecx
> + xorl 36(%rsp),%ebp
> + leal -1894007588(%rdx,%rdi,1),%edi
> + xorl %r13d,%ebx
> + roll $5,%ecx
> + addl %eax,%edi
> + roll $1,%ebp
> + andl %r12d,%ebx
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %ebx,%edi
> + xorl 8(%rsp),%r14d
> + movl %r13d,%eax
> + movl %ebp,4(%rsp)
> + movl %r13d,%ebx
> + xorl 16(%rsp),%r14d
> + andl %r12d,%eax
> + movl %edi,%ecx
> + xorl 40(%rsp),%r14d
> + leal -1894007588(%rbp,%rsi,1),%esi
> + xorl %r12d,%ebx
> + roll $5,%ecx
> + addl %eax,%esi
> + roll $1,%r14d
> + andl %r11d,%ebx
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %ebx,%esi
> + xorl 12(%rsp),%edx
> + movl %r12d,%eax
> + movl %r14d,8(%rsp)
> + movl %r12d,%ebx
> + xorl 20(%rsp),%edx
> + andl %r11d,%eax
> + movl %esi,%ecx
> + xorl 44(%rsp),%edx
> + leal -1894007588(%r14,%r13,1),%r13d
> + xorl %r11d,%ebx
> + roll $5,%ecx
> + addl %eax,%r13d
> + roll $1,%edx
> + andl %edi,%ebx
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %ebx,%r13d
> + xorl 16(%rsp),%ebp
> + movl %r11d,%eax
> + movl %edx,12(%rsp)
> + movl %r11d,%ebx
> + xorl 24(%rsp),%ebp
> + andl %edi,%eax
> + movl %r13d,%ecx
> + xorl 48(%rsp),%ebp
> + leal -1894007588(%rdx,%r12,1),%r12d
> + xorl %edi,%ebx
> + roll $5,%ecx
> + addl %eax,%r12d
> + roll $1,%ebp
> + andl %esi,%ebx
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %ebx,%r12d
> + xorl 20(%rsp),%r14d
> + movl %edi,%eax
> + movl %ebp,16(%rsp)
> + movl %edi,%ebx
> + xorl 28(%rsp),%r14d
> + andl %esi,%eax
> + movl %r12d,%ecx
> + xorl 52(%rsp),%r14d
> + leal -1894007588(%rbp,%r11,1),%r11d
> + xorl %esi,%ebx
> + roll $5,%ecx
> + addl %eax,%r11d
> + roll $1,%r14d
> + andl %r13d,%ebx
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %ebx,%r11d
> + xorl 24(%rsp),%edx
> + movl %esi,%eax
> + movl %r14d,20(%rsp)
> + movl %esi,%ebx
> + xorl 32(%rsp),%edx
> + andl %r13d,%eax
> + movl %r11d,%ecx
> + xorl 56(%rsp),%edx
> + leal -1894007588(%r14,%rdi,1),%edi
> + xorl %r13d,%ebx
> + roll $5,%ecx
> + addl %eax,%edi
> + roll $1,%edx
> + andl %r12d,%ebx
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %ebx,%edi
> + xorl 28(%rsp),%ebp
> + movl %r13d,%eax
> + movl %edx,24(%rsp)
> + movl %r13d,%ebx
> + xorl 36(%rsp),%ebp
> + andl %r12d,%eax
> + movl %edi,%ecx
> + xorl 60(%rsp),%ebp
> + leal -1894007588(%rdx,%rsi,1),%esi
> + xorl %r12d,%ebx
> + roll $5,%ecx
> + addl %eax,%esi
> + roll $1,%ebp
> + andl %r11d,%ebx
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %ebx,%esi
> + xorl 32(%rsp),%r14d
> + movl %r12d,%eax
> + movl %ebp,28(%rsp)
> + movl %r12d,%ebx
> + xorl 40(%rsp),%r14d
> + andl %r11d,%eax
> + movl %esi,%ecx
> + xorl 0(%rsp),%r14d
> + leal -1894007588(%rbp,%r13,1),%r13d
> + xorl %r11d,%ebx
> + roll $5,%ecx
> + addl %eax,%r13d
> + roll $1,%r14d
> + andl %edi,%ebx
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %ebx,%r13d
> + xorl 36(%rsp),%edx
> + movl %r11d,%eax
> + movl %r14d,32(%rsp)
> + movl %r11d,%ebx
> + xorl 44(%rsp),%edx
> + andl %edi,%eax
> + movl %r13d,%ecx
> + xorl 4(%rsp),%edx
> + leal -1894007588(%r14,%r12,1),%r12d
> + xorl %edi,%ebx
> + roll $5,%ecx
> + addl %eax,%r12d
> + roll $1,%edx
> + andl %esi,%ebx
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %ebx,%r12d
> + xorl 40(%rsp),%ebp
> + movl %edi,%eax
> + movl %edx,36(%rsp)
> + movl %edi,%ebx
> + xorl 48(%rsp),%ebp
> + andl %esi,%eax
> + movl %r12d,%ecx
> + xorl 8(%rsp),%ebp
> + leal -1894007588(%rdx,%r11,1),%r11d
> + xorl %esi,%ebx
> + roll $5,%ecx
> + addl %eax,%r11d
> + roll $1,%ebp
> + andl %r13d,%ebx
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %ebx,%r11d
> + xorl 44(%rsp),%r14d
> + movl %esi,%eax
> + movl %ebp,40(%rsp)
> + movl %esi,%ebx
> + xorl 52(%rsp),%r14d
> + andl %r13d,%eax
> + movl %r11d,%ecx
> + xorl 12(%rsp),%r14d
> + leal -1894007588(%rbp,%rdi,1),%edi
> + xorl %r13d,%ebx
> + roll $5,%ecx
> + addl %eax,%edi
> + roll $1,%r14d
> + andl %r12d,%ebx
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %ebx,%edi
> + xorl 48(%rsp),%edx
> + movl %r13d,%eax
> + movl %r14d,44(%rsp)
> + movl %r13d,%ebx
> + xorl 56(%rsp),%edx
> + andl %r12d,%eax
> + movl %edi,%ecx
> + xorl 16(%rsp),%edx
> + leal -1894007588(%r14,%rsi,1),%esi
> + xorl %r12d,%ebx
> + roll $5,%ecx
> + addl %eax,%esi
> + roll $1,%edx
> + andl %r11d,%ebx
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %ebx,%esi
> + xorl 52(%rsp),%ebp
> + movl %edi,%eax
> + movl %edx,48(%rsp)
> + movl %esi,%ecx
> + xorl 60(%rsp),%ebp
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 20(%rsp),%ebp
> + leal -899497514(%rdx,%r13,1),%r13d
> + xorl %r11d,%eax
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %eax,%r13d
> + roll $1,%ebp
> + xorl 56(%rsp),%r14d
> + movl %esi,%eax
> + movl %ebp,52(%rsp)
> + movl %r13d,%ecx
> + xorl 0(%rsp),%r14d
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 24(%rsp),%r14d
> + leal -899497514(%rbp,%r12,1),%r12d
> + xorl %edi,%eax
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %eax,%r12d
> + roll $1,%r14d
> + xorl 60(%rsp),%edx
> + movl %r13d,%eax
> + movl %r14d,56(%rsp)
> + movl %r12d,%ecx
> + xorl 4(%rsp),%edx
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 28(%rsp),%edx
> + leal -899497514(%r14,%r11,1),%r11d
> + xorl %esi,%eax
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %eax,%r11d
> + roll $1,%edx
> + xorl 0(%rsp),%ebp
> + movl %r12d,%eax
> + movl %edx,60(%rsp)
> + movl %r11d,%ecx
> + xorl 8(%rsp),%ebp
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 32(%rsp),%ebp
> + leal -899497514(%rdx,%rdi,1),%edi
> + xorl %r13d,%eax
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %eax,%edi
> + roll $1,%ebp
> + xorl 4(%rsp),%r14d
> + movl %r11d,%eax
> + movl %ebp,0(%rsp)
> + movl %edi,%ecx
> + xorl 12(%rsp),%r14d
> + xorl %r13d,%eax
> + roll $5,%ecx
> + xorl 36(%rsp),%r14d
> + leal -899497514(%rbp,%rsi,1),%esi
> + xorl %r12d,%eax
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %eax,%esi
> + roll $1,%r14d
> + xorl 8(%rsp),%edx
> + movl %edi,%eax
> + movl %r14d,4(%rsp)
> + movl %esi,%ecx
> + xorl 16(%rsp),%edx
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 40(%rsp),%edx
> + leal -899497514(%r14,%r13,1),%r13d
> + xorl %r11d,%eax
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %eax,%r13d
> + roll $1,%edx
> + xorl 12(%rsp),%ebp
> + movl %esi,%eax
> + movl %edx,8(%rsp)
> + movl %r13d,%ecx
> + xorl 20(%rsp),%ebp
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 44(%rsp),%ebp
> + leal -899497514(%rdx,%r12,1),%r12d
> + xorl %edi,%eax
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %eax,%r12d
> + roll $1,%ebp
> + xorl 16(%rsp),%r14d
> + movl %r13d,%eax
> + movl %ebp,12(%rsp)
> + movl %r12d,%ecx
> + xorl 24(%rsp),%r14d
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 48(%rsp),%r14d
> + leal -899497514(%rbp,%r11,1),%r11d
> + xorl %esi,%eax
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %eax,%r11d
> + roll $1,%r14d
> + xorl 20(%rsp),%edx
> + movl %r12d,%eax
> + movl %r14d,16(%rsp)
> + movl %r11d,%ecx
> + xorl 28(%rsp),%edx
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 52(%rsp),%edx
> + leal -899497514(%r14,%rdi,1),%edi
> + xorl %r13d,%eax
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %eax,%edi
> + roll $1,%edx
> + xorl 24(%rsp),%ebp
> + movl %r11d,%eax
> + movl %edx,20(%rsp)
> + movl %edi,%ecx
> + xorl 32(%rsp),%ebp
> + xorl %r13d,%eax
> + roll $5,%ecx
> + xorl 56(%rsp),%ebp
> + leal -899497514(%rdx,%rsi,1),%esi
> + xorl %r12d,%eax
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %eax,%esi
> + roll $1,%ebp
> + xorl 28(%rsp),%r14d
> + movl %edi,%eax
> + movl %ebp,24(%rsp)
> + movl %esi,%ecx
> + xorl 36(%rsp),%r14d
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 60(%rsp),%r14d
> + leal -899497514(%rbp,%r13,1),%r13d
> + xorl %r11d,%eax
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %eax,%r13d
> + roll $1,%r14d
> + xorl 32(%rsp),%edx
> + movl %esi,%eax
> + movl %r14d,28(%rsp)
> + movl %r13d,%ecx
> + xorl 40(%rsp),%edx
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 0(%rsp),%edx
> + leal -899497514(%r14,%r12,1),%r12d
> + xorl %edi,%eax
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %eax,%r12d
> + roll $1,%edx
> + xorl 36(%rsp),%ebp
> + movl %r13d,%eax
> +
> + movl %r12d,%ecx
> + xorl 44(%rsp),%ebp
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 4(%rsp),%ebp
> + leal -899497514(%rdx,%r11,1),%r11d
> + xorl %esi,%eax
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %eax,%r11d
> + roll $1,%ebp
> + xorl 40(%rsp),%r14d
> + movl %r12d,%eax
> +
> + movl %r11d,%ecx
> + xorl 48(%rsp),%r14d
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 8(%rsp),%r14d
> + leal -899497514(%rbp,%rdi,1),%edi
> + xorl %r13d,%eax
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %eax,%edi
> + roll $1,%r14d
> + xorl 44(%rsp),%edx
> + movl %r11d,%eax
> +
> + movl %edi,%ecx
> + xorl 52(%rsp),%edx
> + xorl %r13d,%eax
> + roll $5,%ecx
> + xorl 12(%rsp),%edx
> + leal -899497514(%r14,%rsi,1),%esi
> + xorl %r12d,%eax
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %eax,%esi
> + roll $1,%edx
> + xorl 48(%rsp),%ebp
> + movl %edi,%eax
> +
> + movl %esi,%ecx
> + xorl 56(%rsp),%ebp
> + xorl %r12d,%eax
> + roll $5,%ecx
> + xorl 16(%rsp),%ebp
> + leal -899497514(%rdx,%r13,1),%r13d
> + xorl %r11d,%eax
> + addl %ecx,%r13d
> + roll $30,%edi
> + addl %eax,%r13d
> + roll $1,%ebp
> + xorl 52(%rsp),%r14d
> + movl %esi,%eax
> +
> + movl %r13d,%ecx
> + xorl 60(%rsp),%r14d
> + xorl %r11d,%eax
> + roll $5,%ecx
> + xorl 20(%rsp),%r14d
> + leal -899497514(%rbp,%r12,1),%r12d
> + xorl %edi,%eax
> + addl %ecx,%r12d
> + roll $30,%esi
> + addl %eax,%r12d
> + roll $1,%r14d
> + xorl 56(%rsp),%edx
> + movl %r13d,%eax
> +
> + movl %r12d,%ecx
> + xorl 0(%rsp),%edx
> + xorl %edi,%eax
> + roll $5,%ecx
> + xorl 24(%rsp),%edx
> + leal -899497514(%r14,%r11,1),%r11d
> + xorl %esi,%eax
> + addl %ecx,%r11d
> + roll $30,%r13d
> + addl %eax,%r11d
> + roll $1,%edx
> + xorl 60(%rsp),%ebp
> + movl %r12d,%eax
> +
> + movl %r11d,%ecx
> + xorl 4(%rsp),%ebp
> + xorl %esi,%eax
> + roll $5,%ecx
> + xorl 28(%rsp),%ebp
> + leal -899497514(%rdx,%rdi,1),%edi
> + xorl %r13d,%eax
> + addl %ecx,%edi
> + roll $30,%r12d
> + addl %eax,%edi
> + roll $1,%ebp
> + movl %r11d,%eax
> + movl %edi,%ecx
> + xorl %r13d,%eax
> + leal -899497514(%rbp,%rsi,1),%esi
> + roll $5,%ecx
> + xorl %r12d,%eax
> + addl %ecx,%esi
> + roll $30,%r11d
> + addl %eax,%esi
> + addl 0(%r8),%esi
> + addl 4(%r8),%edi
> + addl 8(%r8),%r11d
> + addl 12(%r8),%r12d
> + addl 16(%r8),%r13d
> + movl %esi,0(%r8)
> + movl %edi,4(%r8)
> + movl %r11d,8(%r8)
> + movl %r12d,12(%r8)
> + movl %r13d,16(%r8)
> +
> + subq $1,%r10
> + leaq 64(%r9),%r9
> + jnz .Lloop
> +
> + movq 64(%rsp),%rsi
> +.cfi_def_cfa %rsi,8
> + movq -40(%rsi),%r14
> +.cfi_restore %r14
> + movq -32(%rsi),%r13
> +.cfi_restore %r13
> + movq -24(%rsi),%r12
> +.cfi_restore %r12
> + movq -16(%rsi),%rbp
> +.cfi_restore %rbp
> + movq -8(%rsi),%rbx
> +.cfi_restore %rbx
> + leaq (%rsi),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha1_block_data_order,.-sha1_block_data_order
> +.type sha1_block_data_order_shaext, at function
> +.align 32
> +sha1_block_data_order_shaext:
> +_shaext_shortcut:
> +.cfi_startproc
> + movdqu (%rdi),%xmm0
> + movd 16(%rdi),%xmm1
> + movdqa K_XX_XX+160(%rip),%xmm3
> +
> + movdqu (%rsi),%xmm4
> + pshufd $27,%xmm0,%xmm0
> + movdqu 16(%rsi),%xmm5
> + pshufd $27,%xmm1,%xmm1
> + movdqu 32(%rsi),%xmm6
> +.byte 102,15,56,0,227
> + movdqu 48(%rsi),%xmm7
> +.byte 102,15,56,0,235
> +.byte 102,15,56,0,243
> + movdqa %xmm1,%xmm9
> +.byte 102,15,56,0,251
> + jmp .Loop_shaext
> +
> +.align 16
> +.Loop_shaext:
> + decq %rdx
> + leaq 64(%rsi),%r8
> + paddd %xmm4,%xmm1
> + cmovneq %r8,%rsi
> + movdqa %xmm0,%xmm8
> +.byte 15,56,201,229
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,0
> +.byte 15,56,200,213
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> +.byte 15,56,202,231
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,0
> +.byte 15,56,200,206
> + pxor %xmm7,%xmm5
> +.byte 15,56,202,236
> +.byte 15,56,201,247
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,0
> +.byte 15,56,200,215
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,252
> +.byte 15,56,202,245
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,0
> +.byte 15,56,200,204
> + pxor %xmm5,%xmm7
> +.byte 15,56,202,254
> +.byte 15,56,201,229
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,0
> +.byte 15,56,200,213
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> +.byte 15,56,202,231
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,1
> +.byte 15,56,200,206
> + pxor %xmm7,%xmm5
> +.byte 15,56,202,236
> +.byte 15,56,201,247
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,1
> +.byte 15,56,200,215
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,252
> +.byte 15,56,202,245
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,1
> +.byte 15,56,200,204
> + pxor %xmm5,%xmm7
> +.byte 15,56,202,254
> +.byte 15,56,201,229
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,1
> +.byte 15,56,200,213
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> +.byte 15,56,202,231
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,1
> +.byte 15,56,200,206
> + pxor %xmm7,%xmm5
> +.byte 15,56,202,236
> +.byte 15,56,201,247
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,2
> +.byte 15,56,200,215
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,252
> +.byte 15,56,202,245
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,2
> +.byte 15,56,200,204
> + pxor %xmm5,%xmm7
> +.byte 15,56,202,254
> +.byte 15,56,201,229
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,2
> +.byte 15,56,200,213
> + pxor %xmm6,%xmm4
> +.byte 15,56,201,238
> +.byte 15,56,202,231
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,2
> +.byte 15,56,200,206
> + pxor %xmm7,%xmm5
> +.byte 15,56,202,236
> +.byte 15,56,201,247
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,2
> +.byte 15,56,200,215
> + pxor %xmm4,%xmm6
> +.byte 15,56,201,252
> +.byte 15,56,202,245
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,3
> +.byte 15,56,200,204
> + pxor %xmm5,%xmm7
> +.byte 15,56,202,254
> + movdqu (%rsi),%xmm4
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,3
> +.byte 15,56,200,213
> + movdqu 16(%rsi),%xmm5
> +.byte 102,15,56,0,227
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,3
> +.byte 15,56,200,206
> + movdqu 32(%rsi),%xmm6
> +.byte 102,15,56,0,235
> +
> + movdqa %xmm0,%xmm2
> +.byte 15,58,204,193,3
> +.byte 15,56,200,215
> + movdqu 48(%rsi),%xmm7
> +.byte 102,15,56,0,243
> +
> + movdqa %xmm0,%xmm1
> +.byte 15,58,204,194,3
> +.byte 65,15,56,200,201
> +.byte 102,15,56,0,251
> +
> + paddd %xmm8,%xmm0
> + movdqa %xmm1,%xmm9
> +
> + jnz .Loop_shaext
> +
> + pshufd $27,%xmm0,%xmm0
> + pshufd $27,%xmm1,%xmm1
> + movdqu %xmm0,(%rdi)
> + movd %xmm1,16(%rdi)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha1_block_data_order_shaext,.-sha1_block_data_order_shaext
> +.type sha1_block_data_order_ssse3, at function
> +.align 16
> +sha1_block_data_order_ssse3:
> +_ssse3_shortcut:
> +.cfi_startproc
> + movq %rsp,%r11
> +.cfi_def_cfa_register %r11
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_offset %r14,-48
> + leaq -64(%rsp),%rsp
> + andq $-64,%rsp
> + movq %rdi,%r8
> + movq %rsi,%r9
> + movq %rdx,%r10
> +
> + shlq $6,%r10
> + addq %r9,%r10
> + leaq K_XX_XX+64(%rip),%r14
> +
> + movl 0(%r8),%eax
> + movl 4(%r8),%ebx
> + movl 8(%r8),%ecx
> + movl 12(%r8),%edx
> + movl %ebx,%esi
> + movl 16(%r8),%ebp
> + movl %ecx,%edi
> + xorl %edx,%edi
> + andl %edi,%esi
> +
> + movdqa 64(%r14),%xmm6
> + movdqa -64(%r14),%xmm9
> + movdqu 0(%r9),%xmm0
> + movdqu 16(%r9),%xmm1
> + movdqu 32(%r9),%xmm2
> + movdqu 48(%r9),%xmm3
> +.byte 102,15,56,0,198
> +.byte 102,15,56,0,206
> +.byte 102,15,56,0,214
> + addq $64,%r9
> + paddd %xmm9,%xmm0
> +.byte 102,15,56,0,222
> + paddd %xmm9,%xmm1
> + paddd %xmm9,%xmm2
> + movdqa %xmm0,0(%rsp)
> + psubd %xmm9,%xmm0
> + movdqa %xmm1,16(%rsp)
> + psubd %xmm9,%xmm1
> + movdqa %xmm2,32(%rsp)
> + psubd %xmm9,%xmm2
> + jmp .Loop_ssse3
> +.align 16
> +.Loop_ssse3:
> + rorl $2,%ebx
> + pshufd $238,%xmm0,%xmm4
> + xorl %edx,%esi
> + movdqa %xmm3,%xmm8
> + paddd %xmm3,%xmm9
> + movl %eax,%edi
> + addl 0(%rsp),%ebp
> + punpcklqdq %xmm1,%xmm4
> + xorl %ecx,%ebx
> + roll $5,%eax
> + addl %esi,%ebp
> + psrldq $4,%xmm8
> + andl %ebx,%edi
> + xorl %ecx,%ebx
> + pxor %xmm0,%xmm4
> + addl %eax,%ebp
> + rorl $7,%eax
> + pxor %xmm2,%xmm8
> + xorl %ecx,%edi
> + movl %ebp,%esi
> + addl 4(%rsp),%edx
> + pxor %xmm8,%xmm4
> + xorl %ebx,%eax
> + roll $5,%ebp
> + movdqa %xmm9,48(%rsp)
> + addl %edi,%edx
> + andl %eax,%esi
> + movdqa %xmm4,%xmm10
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + rorl $7,%ebp
> + movdqa %xmm4,%xmm8
> + xorl %ebx,%esi
> + pslldq $12,%xmm10
> + paddd %xmm4,%xmm4
> + movl %edx,%edi
> + addl 8(%rsp),%ecx
> + psrld $31,%xmm8
> + xorl %eax,%ebp
> + roll $5,%edx
> + addl %esi,%ecx
> + movdqa %xmm10,%xmm9
> + andl %ebp,%edi
> + xorl %eax,%ebp
> + psrld $30,%xmm10
> + addl %edx,%ecx
> + rorl $7,%edx
> + por %xmm8,%xmm4
> + xorl %eax,%edi
> + movl %ecx,%esi
> + addl 12(%rsp),%ebx
> + pslld $2,%xmm9
> + pxor %xmm10,%xmm4
> + xorl %ebp,%edx
> + movdqa -64(%r14),%xmm10
> + roll $5,%ecx
> + addl %edi,%ebx
> + andl %edx,%esi
> + pxor %xmm9,%xmm4
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + rorl $7,%ecx
> + pshufd $238,%xmm1,%xmm5
> + xorl %ebp,%esi
> + movdqa %xmm4,%xmm9
> + paddd %xmm4,%xmm10
> + movl %ebx,%edi
> + addl 16(%rsp),%eax
> + punpcklqdq %xmm2,%xmm5
> + xorl %edx,%ecx
> + roll $5,%ebx
> + addl %esi,%eax
> + psrldq $4,%xmm9
> + andl %ecx,%edi
> + xorl %edx,%ecx
> + pxor %xmm1,%xmm5
> + addl %ebx,%eax
> + rorl $7,%ebx
> + pxor %xmm3,%xmm9
> + xorl %edx,%edi
> + movl %eax,%esi
> + addl 20(%rsp),%ebp
> + pxor %xmm9,%xmm5
> + xorl %ecx,%ebx
> + roll $5,%eax
> + movdqa %xmm10,0(%rsp)
> + addl %edi,%ebp
> + andl %ebx,%esi
> + movdqa %xmm5,%xmm8
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + rorl $7,%eax
> + movdqa %xmm5,%xmm9
> + xorl %ecx,%esi
> + pslldq $12,%xmm8
> + paddd %xmm5,%xmm5
> + movl %ebp,%edi
> + addl 24(%rsp),%edx
> + psrld $31,%xmm9
> + xorl %ebx,%eax
> + roll $5,%ebp
> + addl %esi,%edx
> + movdqa %xmm8,%xmm10
> + andl %eax,%edi
> + xorl %ebx,%eax
> + psrld $30,%xmm8
> + addl %ebp,%edx
> + rorl $7,%ebp
> + por %xmm9,%xmm5
> + xorl %ebx,%edi
> + movl %edx,%esi
> + addl 28(%rsp),%ecx
> + pslld $2,%xmm10
> + pxor %xmm8,%xmm5
> + xorl %eax,%ebp
> + movdqa -32(%r14),%xmm8
> + roll $5,%edx
> + addl %edi,%ecx
> + andl %ebp,%esi
> + pxor %xmm10,%xmm5
> + xorl %eax,%ebp
> + addl %edx,%ecx
> + rorl $7,%edx
> + pshufd $238,%xmm2,%xmm6
> + xorl %eax,%esi
> + movdqa %xmm5,%xmm10
> + paddd %xmm5,%xmm8
> + movl %ecx,%edi
> + addl 32(%rsp),%ebx
> + punpcklqdq %xmm3,%xmm6
> + xorl %ebp,%edx
> + roll $5,%ecx
> + addl %esi,%ebx
> + psrldq $4,%xmm10
> + andl %edx,%edi
> + xorl %ebp,%edx
> + pxor %xmm2,%xmm6
> + addl %ecx,%ebx
> + rorl $7,%ecx
> + pxor %xmm4,%xmm10
> + xorl %ebp,%edi
> + movl %ebx,%esi
> + addl 36(%rsp),%eax
> + pxor %xmm10,%xmm6
> + xorl %edx,%ecx
> + roll $5,%ebx
> + movdqa %xmm8,16(%rsp)
> + addl %edi,%eax
> + andl %ecx,%esi
> + movdqa %xmm6,%xmm9
> + xorl %edx,%ecx
> + addl %ebx,%eax
> + rorl $7,%ebx
> + movdqa %xmm6,%xmm10
> + xorl %edx,%esi
> + pslldq $12,%xmm9
> + paddd %xmm6,%xmm6
> + movl %eax,%edi
> + addl 40(%rsp),%ebp
> + psrld $31,%xmm10
> + xorl %ecx,%ebx
> + roll $5,%eax
> + addl %esi,%ebp
> + movdqa %xmm9,%xmm8
> + andl %ebx,%edi
> + xorl %ecx,%ebx
> + psrld $30,%xmm9
> + addl %eax,%ebp
> + rorl $7,%eax
> + por %xmm10,%xmm6
> + xorl %ecx,%edi
> + movl %ebp,%esi
> + addl 44(%rsp),%edx
> + pslld $2,%xmm8
> + pxor %xmm9,%xmm6
> + xorl %ebx,%eax
> + movdqa -32(%r14),%xmm9
> + roll $5,%ebp
> + addl %edi,%edx
> + andl %eax,%esi
> + pxor %xmm8,%xmm6
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + rorl $7,%ebp
> + pshufd $238,%xmm3,%xmm7
> + xorl %ebx,%esi
> + movdqa %xmm6,%xmm8
> + paddd %xmm6,%xmm9
> + movl %edx,%edi
> + addl 48(%rsp),%ecx
> + punpcklqdq %xmm4,%xmm7
> + xorl %eax,%ebp
> + roll $5,%edx
> + addl %esi,%ecx
> + psrldq $4,%xmm8
> + andl %ebp,%edi
> + xorl %eax,%ebp
> + pxor %xmm3,%xmm7
> + addl %edx,%ecx
> + rorl $7,%edx
> + pxor %xmm5,%xmm8
> + xorl %eax,%edi
> + movl %ecx,%esi
> + addl 52(%rsp),%ebx
> + pxor %xmm8,%xmm7
> + xorl %ebp,%edx
> + roll $5,%ecx
> + movdqa %xmm9,32(%rsp)
> + addl %edi,%ebx
> + andl %edx,%esi
> + movdqa %xmm7,%xmm10
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + rorl $7,%ecx
> + movdqa %xmm7,%xmm8
> + xorl %ebp,%esi
> + pslldq $12,%xmm10
> + paddd %xmm7,%xmm7
> + movl %ebx,%edi
> + addl 56(%rsp),%eax
> + psrld $31,%xmm8
> + xorl %edx,%ecx
> + roll $5,%ebx
> + addl %esi,%eax
> + movdqa %xmm10,%xmm9
> + andl %ecx,%edi
> + xorl %edx,%ecx
> + psrld $30,%xmm10
> + addl %ebx,%eax
> + rorl $7,%ebx
> + por %xmm8,%xmm7
> + xorl %edx,%edi
> + movl %eax,%esi
> + addl 60(%rsp),%ebp
> + pslld $2,%xmm9
> + pxor %xmm10,%xmm7
> + xorl %ecx,%ebx
> + movdqa -32(%r14),%xmm10
> + roll $5,%eax
> + addl %edi,%ebp
> + andl %ebx,%esi
> + pxor %xmm9,%xmm7
> + pshufd $238,%xmm6,%xmm9
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + rorl $7,%eax
> + pxor %xmm4,%xmm0
> + xorl %ecx,%esi
> + movl %ebp,%edi
> + addl 0(%rsp),%edx
> + punpcklqdq %xmm7,%xmm9
> + xorl %ebx,%eax
> + roll $5,%ebp
> + pxor %xmm1,%xmm0
> + addl %esi,%edx
> + andl %eax,%edi
> + movdqa %xmm10,%xmm8
> + xorl %ebx,%eax
> + paddd %xmm7,%xmm10
> + addl %ebp,%edx
> + pxor %xmm9,%xmm0
> + rorl $7,%ebp
> + xorl %ebx,%edi
> + movl %edx,%esi
> + addl 4(%rsp),%ecx
> + movdqa %xmm0,%xmm9
> + xorl %eax,%ebp
> + roll $5,%edx
> + movdqa %xmm10,48(%rsp)
> + addl %edi,%ecx
> + andl %ebp,%esi
> + xorl %eax,%ebp
> + pslld $2,%xmm0
> + addl %edx,%ecx
> + rorl $7,%edx
> + psrld $30,%xmm9
> + xorl %eax,%esi
> + movl %ecx,%edi
> + addl 8(%rsp),%ebx
> + por %xmm9,%xmm0
> + xorl %ebp,%edx
> + roll $5,%ecx
> + pshufd $238,%xmm7,%xmm10
> + addl %esi,%ebx
> + andl %edx,%edi
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + addl 12(%rsp),%eax
> + xorl %ebp,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %edx,%esi
> + rorl $7,%ecx
> + addl %ebx,%eax
> + pxor %xmm5,%xmm1
> + addl 16(%rsp),%ebp
> + xorl %ecx,%esi
> + punpcklqdq %xmm0,%xmm10
> + movl %eax,%edi
> + roll $5,%eax
> + pxor %xmm2,%xmm1
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + movdqa %xmm8,%xmm9
> + rorl $7,%ebx
> + paddd %xmm0,%xmm8
> + addl %eax,%ebp
> + pxor %xmm10,%xmm1
> + addl 20(%rsp),%edx
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + movdqa %xmm1,%xmm10
> + addl %edi,%edx
> + xorl %ebx,%esi
> + movdqa %xmm8,0(%rsp)
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 24(%rsp),%ecx
> + pslld $2,%xmm1
> + xorl %eax,%esi
> + movl %edx,%edi
> + psrld $30,%xmm10
> + roll $5,%edx
> + addl %esi,%ecx
> + xorl %eax,%edi
> + rorl $7,%ebp
> + por %xmm10,%xmm1
> + addl %edx,%ecx
> + addl 28(%rsp),%ebx
> + pshufd $238,%xmm0,%xmm8
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + pxor %xmm6,%xmm2
> + addl 32(%rsp),%eax
> + xorl %edx,%esi
> + punpcklqdq %xmm1,%xmm8
> + movl %ebx,%edi
> + roll $5,%ebx
> + pxor %xmm3,%xmm2
> + addl %esi,%eax
> + xorl %edx,%edi
> + movdqa 0(%r14),%xmm10
> + rorl $7,%ecx
> + paddd %xmm1,%xmm9
> + addl %ebx,%eax
> + pxor %xmm8,%xmm2
> + addl 36(%rsp),%ebp
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + movdqa %xmm2,%xmm8
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + movdqa %xmm9,16(%rsp)
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 40(%rsp),%edx
> + pslld $2,%xmm2
> + xorl %ebx,%esi
> + movl %ebp,%edi
> + psrld $30,%xmm8
> + roll $5,%ebp
> + addl %esi,%edx
> + xorl %ebx,%edi
> + rorl $7,%eax
> + por %xmm8,%xmm2
> + addl %ebp,%edx
> + addl 44(%rsp),%ecx
> + pshufd $238,%xmm1,%xmm9
> + xorl %eax,%edi
> + movl %edx,%esi
> + roll $5,%edx
> + addl %edi,%ecx
> + xorl %eax,%esi
> + rorl $7,%ebp
> + addl %edx,%ecx
> + pxor %xmm7,%xmm3
> + addl 48(%rsp),%ebx
> + xorl %ebp,%esi
> + punpcklqdq %xmm2,%xmm9
> + movl %ecx,%edi
> + roll $5,%ecx
> + pxor %xmm4,%xmm3
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + movdqa %xmm10,%xmm8
> + rorl $7,%edx
> + paddd %xmm2,%xmm10
> + addl %ecx,%ebx
> + pxor %xmm9,%xmm3
> + addl 52(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + movdqa %xmm3,%xmm9
> + addl %edi,%eax
> + xorl %edx,%esi
> + movdqa %xmm10,32(%rsp)
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 56(%rsp),%ebp
> + pslld $2,%xmm3
> + xorl %ecx,%esi
> + movl %eax,%edi
> + psrld $30,%xmm9
> + roll $5,%eax
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + rorl $7,%ebx
> + por %xmm9,%xmm3
> + addl %eax,%ebp
> + addl 60(%rsp),%edx
> + pshufd $238,%xmm2,%xmm10
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + addl %edi,%edx
> + xorl %ebx,%esi
> + rorl $7,%eax
> + addl %ebp,%edx
> + pxor %xmm0,%xmm4
> + addl 0(%rsp),%ecx
> + xorl %eax,%esi
> + punpcklqdq %xmm3,%xmm10
> + movl %edx,%edi
> + roll $5,%edx
> + pxor %xmm5,%xmm4
> + addl %esi,%ecx
> + xorl %eax,%edi
> + movdqa %xmm8,%xmm9
> + rorl $7,%ebp
> + paddd %xmm3,%xmm8
> + addl %edx,%ecx
> + pxor %xmm10,%xmm4
> + addl 4(%rsp),%ebx
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + roll $5,%ecx
> + movdqa %xmm4,%xmm10
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + movdqa %xmm8,48(%rsp)
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 8(%rsp),%eax
> + pslld $2,%xmm4
> + xorl %edx,%esi
> + movl %ebx,%edi
> + psrld $30,%xmm10
> + roll $5,%ebx
> + addl %esi,%eax
> + xorl %edx,%edi
> + rorl $7,%ecx
> + por %xmm10,%xmm4
> + addl %ebx,%eax
> + addl 12(%rsp),%ebp
> + pshufd $238,%xmm3,%xmm8
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + pxor %xmm1,%xmm5
> + addl 16(%rsp),%edx
> + xorl %ebx,%esi
> + punpcklqdq %xmm4,%xmm8
> + movl %ebp,%edi
> + roll $5,%ebp
> + pxor %xmm6,%xmm5
> + addl %esi,%edx
> + xorl %ebx,%edi
> + movdqa %xmm9,%xmm10
> + rorl $7,%eax
> + paddd %xmm4,%xmm9
> + addl %ebp,%edx
> + pxor %xmm8,%xmm5
> + addl 20(%rsp),%ecx
> + xorl %eax,%edi
> + movl %edx,%esi
> + roll $5,%edx
> + movdqa %xmm5,%xmm8
> + addl %edi,%ecx
> + xorl %eax,%esi
> + movdqa %xmm9,0(%rsp)
> + rorl $7,%ebp
> + addl %edx,%ecx
> + addl 24(%rsp),%ebx
> + pslld $2,%xmm5
> + xorl %ebp,%esi
> + movl %ecx,%edi
> + psrld $30,%xmm8
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + por %xmm8,%xmm5
> + addl %ecx,%ebx
> + addl 28(%rsp),%eax
> + pshufd $238,%xmm4,%xmm9
> + rorl $7,%ecx
> + movl %ebx,%esi
> + xorl %edx,%edi
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %ecx,%esi
> + xorl %edx,%ecx
> + addl %ebx,%eax
> + pxor %xmm2,%xmm6
> + addl 32(%rsp),%ebp
> + andl %ecx,%esi
> + xorl %edx,%ecx
> + rorl $7,%ebx
> + punpcklqdq %xmm5,%xmm9
> + movl %eax,%edi
> + xorl %ecx,%esi
> + pxor %xmm7,%xmm6
> + roll $5,%eax
> + addl %esi,%ebp
> + movdqa %xmm10,%xmm8
> + xorl %ebx,%edi
> + paddd %xmm5,%xmm10
> + xorl %ecx,%ebx
> + pxor %xmm9,%xmm6
> + addl %eax,%ebp
> + addl 36(%rsp),%edx
> + andl %ebx,%edi
> + xorl %ecx,%ebx
> + rorl $7,%eax
> + movdqa %xmm6,%xmm9
> + movl %ebp,%esi
> + xorl %ebx,%edi
> + movdqa %xmm10,16(%rsp)
> + roll $5,%ebp
> + addl %edi,%edx
> + xorl %eax,%esi
> + pslld $2,%xmm6
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + psrld $30,%xmm9
> + addl 40(%rsp),%ecx
> + andl %eax,%esi
> + xorl %ebx,%eax
> + por %xmm9,%xmm6
> + rorl $7,%ebp
> + movl %edx,%edi
> + xorl %eax,%esi
> + roll $5,%edx
> + pshufd $238,%xmm5,%xmm10
> + addl %esi,%ecx
> + xorl %ebp,%edi
> + xorl %eax,%ebp
> + addl %edx,%ecx
> + addl 44(%rsp),%ebx
> + andl %ebp,%edi
> + xorl %eax,%ebp
> + rorl $7,%edx
> + movl %ecx,%esi
> + xorl %ebp,%edi
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %edx,%esi
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + pxor %xmm3,%xmm7
> + addl 48(%rsp),%eax
> + andl %edx,%esi
> + xorl %ebp,%edx
> + rorl $7,%ecx
> + punpcklqdq %xmm6,%xmm10
> + movl %ebx,%edi
> + xorl %edx,%esi
> + pxor %xmm0,%xmm7
> + roll $5,%ebx
> + addl %esi,%eax
> + movdqa 32(%r14),%xmm9
> + xorl %ecx,%edi
> + paddd %xmm6,%xmm8
> + xorl %edx,%ecx
> + pxor %xmm10,%xmm7
> + addl %ebx,%eax
> + addl 52(%rsp),%ebp
> + andl %ecx,%edi
> + xorl %edx,%ecx
> + rorl $7,%ebx
> + movdqa %xmm7,%xmm10
> + movl %eax,%esi
> + xorl %ecx,%edi
> + movdqa %xmm8,32(%rsp)
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ebx,%esi
> + pslld $2,%xmm7
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + psrld $30,%xmm10
> + addl 56(%rsp),%edx
> + andl %ebx,%esi
> + xorl %ecx,%ebx
> + por %xmm10,%xmm7
> + rorl $7,%eax
> + movl %ebp,%edi
> + xorl %ebx,%esi
> + roll $5,%ebp
> + pshufd $238,%xmm6,%xmm8
> + addl %esi,%edx
> + xorl %eax,%edi
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + addl 60(%rsp),%ecx
> + andl %eax,%edi
> + xorl %ebx,%eax
> + rorl $7,%ebp
> + movl %edx,%esi
> + xorl %eax,%edi
> + roll $5,%edx
> + addl %edi,%ecx
> + xorl %ebp,%esi
> + xorl %eax,%ebp
> + addl %edx,%ecx
> + pxor %xmm4,%xmm0
> + addl 0(%rsp),%ebx
> + andl %ebp,%esi
> + xorl %eax,%ebp
> + rorl $7,%edx
> + punpcklqdq %xmm7,%xmm8
> + movl %ecx,%edi
> + xorl %ebp,%esi
> + pxor %xmm1,%xmm0
> + roll $5,%ecx
> + addl %esi,%ebx
> + movdqa %xmm9,%xmm10
> + xorl %edx,%edi
> + paddd %xmm7,%xmm9
> + xorl %ebp,%edx
> + pxor %xmm8,%xmm0
> + addl %ecx,%ebx
> + addl 4(%rsp),%eax
> + andl %edx,%edi
> + xorl %ebp,%edx
> + rorl $7,%ecx
> + movdqa %xmm0,%xmm8
> + movl %ebx,%esi
> + xorl %edx,%edi
> + movdqa %xmm9,48(%rsp)
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %ecx,%esi
> + pslld $2,%xmm0
> + xorl %edx,%ecx
> + addl %ebx,%eax
> + psrld $30,%xmm8
> + addl 8(%rsp),%ebp
> + andl %ecx,%esi
> + xorl %edx,%ecx
> + por %xmm8,%xmm0
> + rorl $7,%ebx
> + movl %eax,%edi
> + xorl %ecx,%esi
> + roll $5,%eax
> + pshufd $238,%xmm7,%xmm9
> + addl %esi,%ebp
> + xorl %ebx,%edi
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + addl 12(%rsp),%edx
> + andl %ebx,%edi
> + xorl %ecx,%ebx
> + rorl $7,%eax
> + movl %ebp,%esi
> + xorl %ebx,%edi
> + roll $5,%ebp
> + addl %edi,%edx
> + xorl %eax,%esi
> + xorl %ebx,%eax
> + addl %ebp,%edx
> + pxor %xmm5,%xmm1
> + addl 16(%rsp),%ecx
> + andl %eax,%esi
> + xorl %ebx,%eax
> + rorl $7,%ebp
> + punpcklqdq %xmm0,%xmm9
> + movl %edx,%edi
> + xorl %eax,%esi
> + pxor %xmm2,%xmm1
> + roll $5,%edx
> + addl %esi,%ecx
> + movdqa %xmm10,%xmm8
> + xorl %ebp,%edi
> + paddd %xmm0,%xmm10
> + xorl %eax,%ebp
> + pxor %xmm9,%xmm1
> + addl %edx,%ecx
> + addl 20(%rsp),%ebx
> + andl %ebp,%edi
> + xorl %eax,%ebp
> + rorl $7,%edx
> + movdqa %xmm1,%xmm9
> + movl %ecx,%esi
> + xorl %ebp,%edi
> + movdqa %xmm10,0(%rsp)
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %edx,%esi
> + pslld $2,%xmm1
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + psrld $30,%xmm9
> + addl 24(%rsp),%eax
> + andl %edx,%esi
> + xorl %ebp,%edx
> + por %xmm9,%xmm1
> + rorl $7,%ecx
> + movl %ebx,%edi
> + xorl %edx,%esi
> + roll $5,%ebx
> + pshufd $238,%xmm0,%xmm10
> + addl %esi,%eax
> + xorl %ecx,%edi
> + xorl %edx,%ecx
> + addl %ebx,%eax
> + addl 28(%rsp),%ebp
> + andl %ecx,%edi
> + xorl %edx,%ecx
> + rorl $7,%ebx
> + movl %eax,%esi
> + xorl %ecx,%edi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ebx,%esi
> + xorl %ecx,%ebx
> + addl %eax,%ebp
> + pxor %xmm6,%xmm2
> + addl 32(%rsp),%edx
> + andl %ebx,%esi
> + xorl %ecx,%ebx
> + rorl $7,%eax
> + punpcklqdq %xmm1,%xmm10
> + movl %ebp,%edi
> + xorl %ebx,%esi
> + pxor %xmm3,%xmm2
> + roll $5,%ebp
> + addl %esi,%edx
> + movdqa %xmm8,%xmm9
> + xorl %eax,%edi
> + paddd %xmm1,%xmm8
> + xorl %ebx,%eax
> + pxor %xmm10,%xmm2
> + addl %ebp,%edx
> + addl 36(%rsp),%ecx
> + andl %eax,%edi
> + xorl %ebx,%eax
> + rorl $7,%ebp
> + movdqa %xmm2,%xmm10
> + movl %edx,%esi
> + xorl %eax,%edi
> + movdqa %xmm8,16(%rsp)
> + roll $5,%edx
> + addl %edi,%ecx
> + xorl %ebp,%esi
> + pslld $2,%xmm2
> + xorl %eax,%ebp
> + addl %edx,%ecx
> + psrld $30,%xmm10
> + addl 40(%rsp),%ebx
> + andl %ebp,%esi
> + xorl %eax,%ebp
> + por %xmm10,%xmm2
> + rorl $7,%edx
> + movl %ecx,%edi
> + xorl %ebp,%esi
> + roll $5,%ecx
> + pshufd $238,%xmm1,%xmm8
> + addl %esi,%ebx
> + xorl %edx,%edi
> + xorl %ebp,%edx
> + addl %ecx,%ebx
> + addl 44(%rsp),%eax
> + andl %edx,%edi
> + xorl %ebp,%edx
> + rorl $7,%ecx
> + movl %ebx,%esi
> + xorl %edx,%edi
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %edx,%esi
> + addl %ebx,%eax
> + pxor %xmm7,%xmm3
> + addl 48(%rsp),%ebp
> + xorl %ecx,%esi
> + punpcklqdq %xmm2,%xmm8
> + movl %eax,%edi
> + roll $5,%eax
> + pxor %xmm4,%xmm3
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + movdqa %xmm9,%xmm10
> + rorl $7,%ebx
> + paddd %xmm2,%xmm9
> + addl %eax,%ebp
> + pxor %xmm8,%xmm3
> + addl 52(%rsp),%edx
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + movdqa %xmm3,%xmm8
> + addl %edi,%edx
> + xorl %ebx,%esi
> + movdqa %xmm9,32(%rsp)
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 56(%rsp),%ecx
> + pslld $2,%xmm3
> + xorl %eax,%esi
> + movl %edx,%edi
> + psrld $30,%xmm8
> + roll $5,%edx
> + addl %esi,%ecx
> + xorl %eax,%edi
> + rorl $7,%ebp
> + por %xmm8,%xmm3
> + addl %edx,%ecx
> + addl 60(%rsp),%ebx
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 0(%rsp),%eax
> + xorl %edx,%esi
> + movl %ebx,%edi
> + roll $5,%ebx
> + paddd %xmm3,%xmm10
> + addl %esi,%eax
> + xorl %edx,%edi
> + movdqa %xmm10,48(%rsp)
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 4(%rsp),%ebp
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 8(%rsp),%edx
> + xorl %ebx,%esi
> + movl %ebp,%edi
> + roll $5,%ebp
> + addl %esi,%edx
> + xorl %ebx,%edi
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 12(%rsp),%ecx
> + xorl %eax,%edi
> + movl %edx,%esi
> + roll $5,%edx
> + addl %edi,%ecx
> + xorl %eax,%esi
> + rorl $7,%ebp
> + addl %edx,%ecx
> + cmpq %r10,%r9
> + je .Ldone_ssse3
> + movdqa 64(%r14),%xmm6
> + movdqa -64(%r14),%xmm9
> + movdqu 0(%r9),%xmm0
> + movdqu 16(%r9),%xmm1
> + movdqu 32(%r9),%xmm2
> + movdqu 48(%r9),%xmm3
> +.byte 102,15,56,0,198
> + addq $64,%r9
> + addl 16(%rsp),%ebx
> + xorl %ebp,%esi
> + movl %ecx,%edi
> +.byte 102,15,56,0,206
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + paddd %xmm9,%xmm0
> + addl %ecx,%ebx
> + addl 20(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + movdqa %xmm0,0(%rsp)
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %edx,%esi
> + rorl $7,%ecx
> + psubd %xmm9,%xmm0
> + addl %ebx,%eax
> + addl 24(%rsp),%ebp
> + xorl %ecx,%esi
> + movl %eax,%edi
> + roll $5,%eax
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 28(%rsp),%edx
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + addl %edi,%edx
> + xorl %ebx,%esi
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 32(%rsp),%ecx
> + xorl %eax,%esi
> + movl %edx,%edi
> +.byte 102,15,56,0,214
> + roll $5,%edx
> + addl %esi,%ecx
> + xorl %eax,%edi
> + rorl $7,%ebp
> + paddd %xmm9,%xmm1
> + addl %edx,%ecx
> + addl 36(%rsp),%ebx
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + movdqa %xmm1,16(%rsp)
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + rorl $7,%edx
> + psubd %xmm9,%xmm1
> + addl %ecx,%ebx
> + addl 40(%rsp),%eax
> + xorl %edx,%esi
> + movl %ebx,%edi
> + roll $5,%ebx
> + addl %esi,%eax
> + xorl %edx,%edi
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 44(%rsp),%ebp
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 48(%rsp),%edx
> + xorl %ebx,%esi
> + movl %ebp,%edi
> +.byte 102,15,56,0,222
> + roll $5,%ebp
> + addl %esi,%edx
> + xorl %ebx,%edi
> + rorl $7,%eax
> + paddd %xmm9,%xmm2
> + addl %ebp,%edx
> + addl 52(%rsp),%ecx
> + xorl %eax,%edi
> + movl %edx,%esi
> + movdqa %xmm2,32(%rsp)
> + roll $5,%edx
> + addl %edi,%ecx
> + xorl %eax,%esi
> + rorl $7,%ebp
> + psubd %xmm9,%xmm2
> + addl %edx,%ecx
> + addl 56(%rsp),%ebx
> + xorl %ebp,%esi
> + movl %ecx,%edi
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 60(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + addl %edi,%eax
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 0(%r8),%eax
> + addl 4(%r8),%esi
> + addl 8(%r8),%ecx
> + addl 12(%r8),%edx
> + movl %eax,0(%r8)
> + addl 16(%r8),%ebp
> + movl %esi,4(%r8)
> + movl %esi,%ebx
> + movl %ecx,8(%r8)
> + movl %ecx,%edi
> + movl %edx,12(%r8)
> + xorl %edx,%edi
> + movl %ebp,16(%r8)
> + andl %edi,%esi
> + jmp .Loop_ssse3
> +
> +.align 16
> +.Ldone_ssse3:
> + addl 16(%rsp),%ebx
> + xorl %ebp,%esi
> + movl %ecx,%edi
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 20(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + addl %edi,%eax
> + xorl %edx,%esi
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 24(%rsp),%ebp
> + xorl %ecx,%esi
> + movl %eax,%edi
> + roll $5,%eax
> + addl %esi,%ebp
> + xorl %ecx,%edi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 28(%rsp),%edx
> + xorl %ebx,%edi
> + movl %ebp,%esi
> + roll $5,%ebp
> + addl %edi,%edx
> + xorl %ebx,%esi
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 32(%rsp),%ecx
> + xorl %eax,%esi
> + movl %edx,%edi
> + roll $5,%edx
> + addl %esi,%ecx
> + xorl %eax,%edi
> + rorl $7,%ebp
> + addl %edx,%ecx
> + addl 36(%rsp),%ebx
> + xorl %ebp,%edi
> + movl %ecx,%esi
> + roll $5,%ecx
> + addl %edi,%ebx
> + xorl %ebp,%esi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 40(%rsp),%eax
> + xorl %edx,%esi
> + movl %ebx,%edi
> + roll $5,%ebx
> + addl %esi,%eax
> + xorl %edx,%edi
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 44(%rsp),%ebp
> + xorl %ecx,%edi
> + movl %eax,%esi
> + roll $5,%eax
> + addl %edi,%ebp
> + xorl %ecx,%esi
> + rorl $7,%ebx
> + addl %eax,%ebp
> + addl 48(%rsp),%edx
> + xorl %ebx,%esi
> + movl %ebp,%edi
> + roll $5,%ebp
> + addl %esi,%edx
> + xorl %ebx,%edi
> + rorl $7,%eax
> + addl %ebp,%edx
> + addl 52(%rsp),%ecx
> + xorl %eax,%edi
> + movl %edx,%esi
> + roll $5,%edx
> + addl %edi,%ecx
> + xorl %eax,%esi
> + rorl $7,%ebp
> + addl %edx,%ecx
> + addl 56(%rsp),%ebx
> + xorl %ebp,%esi
> + movl %ecx,%edi
> + roll $5,%ecx
> + addl %esi,%ebx
> + xorl %ebp,%edi
> + rorl $7,%edx
> + addl %ecx,%ebx
> + addl 60(%rsp),%eax
> + xorl %edx,%edi
> + movl %ebx,%esi
> + roll $5,%ebx
> + addl %edi,%eax
> + rorl $7,%ecx
> + addl %ebx,%eax
> + addl 0(%r8),%eax
> + addl 4(%r8),%esi
> + addl 8(%r8),%ecx
> + movl %eax,0(%r8)
> + addl 12(%r8),%edx
> + movl %esi,4(%r8)
> + addl 16(%r8),%ebp
> + movl %ecx,8(%r8)
> + movl %edx,12(%r8)
> + movl %ebp,16(%r8)
> + movq -40(%r11),%r14
> +.cfi_restore %r14
> + movq -32(%r11),%r13
> +.cfi_restore %r13
> + movq -24(%r11),%r12
> +.cfi_restore %r12
> + movq -16(%r11),%rbp
> +.cfi_restore %rbp
> + movq -8(%r11),%rbx
> +.cfi_restore %rbx
> + leaq (%r11),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue_ssse3:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha1_block_data_order_ssse3,.-sha1_block_data_order_ssse3
> +.align 64
> +K_XX_XX:
> +.long 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> +.long 0x5a827999,0x5a827999,0x5a827999,0x5a827999
> +.long 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> +.long 0x6ed9eba1,0x6ed9eba1,0x6ed9eba1,0x6ed9eba1
> +.long 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> +.long 0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc,0x8f1bbcdc
> +.long 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> +.long 0xca62c1d6,0xca62c1d6,0xca62c1d6,0xca62c1d6
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.byte 0xf,0xe,0xd,0xc,0xb,0xa,0x9,0x8,0x7,0x6,0x5,0x4,0x3,0x2,0x1,0x0
> +.byte
> 83,72,65,49,32,98,108,111,99,107,32,116,114,97,110,115,102,111,114,109,32,
> 102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,
> 121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,10
> 3,62,0
> +.align 64
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha256-mb-
> x86_64.S b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha256-mb-
> x86_64.S
> new file mode 100644
> index 0000000000..25dee488b8
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha256-mb-x86_64.S
> @@ -0,0 +1,3286 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/sha/asm/sha256-mb-x86_64.pl
> +#
> +# Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +
> +.globl sha256_multi_block
> +.type sha256_multi_block, at function
> +.align 32
> +sha256_multi_block:
> +.cfi_startproc
> + movq OPENSSL_ia32cap_P+4(%rip),%rcx
> + btq $61,%rcx
> + jc _shaext_shortcut
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + subq $288,%rsp
> + andq $-256,%rsp
> + movq %rax,272(%rsp)
> +.cfi_escape 0x0f,0x06,0x77,0x90,0x02,0x06,0x23,0x08
> +.Lbody:
> + leaq K256+128(%rip),%rbp
> + leaq 256(%rsp),%rbx
> + leaq 128(%rdi),%rdi
> +
> +.Loop_grande:
> + movl %edx,280(%rsp)
> + xorl %edx,%edx
> + movq 0(%rsi),%r8
> + movl 8(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,0(%rbx)
> + cmovleq %rbp,%r8
> + movq 16(%rsi),%r9
> + movl 24(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,4(%rbx)
> + cmovleq %rbp,%r9
> + movq 32(%rsi),%r10
> + movl 40(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,8(%rbx)
> + cmovleq %rbp,%r10
> + movq 48(%rsi),%r11
> + movl 56(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,12(%rbx)
> + cmovleq %rbp,%r11
> + testl %edx,%edx
> + jz .Ldone
> +
> + movdqu 0-128(%rdi),%xmm8
> + leaq 128(%rsp),%rax
> + movdqu 32-128(%rdi),%xmm9
> + movdqu 64-128(%rdi),%xmm10
> + movdqu 96-128(%rdi),%xmm11
> + movdqu 128-128(%rdi),%xmm12
> + movdqu 160-128(%rdi),%xmm13
> + movdqu 192-128(%rdi),%xmm14
> + movdqu 224-128(%rdi),%xmm15
> + movdqu .Lpbswap(%rip),%xmm6
> + jmp .Loop
> +
> +.align 32
> +.Loop:
> + movdqa %xmm10,%xmm4
> + pxor %xmm9,%xmm4
> + movd 0(%r8),%xmm5
> + movd 0(%r9),%xmm0
> + movd 0(%r10),%xmm1
> + movd 0(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm12,%xmm7
> +.byte 102,15,56,0,238
> + movdqa %xmm12,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm12,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,0-128(%rax)
> + paddd %xmm15,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -128(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm12,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm12,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm14,%xmm0
> + pand %xmm13,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm8,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm8,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm9,%xmm3
> + movdqa %xmm8,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm8,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm9,%xmm15
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm15
> + paddd %xmm5,%xmm11
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm15
> + paddd %xmm7,%xmm15
> + movd 4(%r8),%xmm5
> + movd 4(%r9),%xmm0
> + movd 4(%r10),%xmm1
> + movd 4(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm11,%xmm7
> +
> + movdqa %xmm11,%xmm2
> +.byte 102,15,56,0,238
> + psrld $6,%xmm7
> + movdqa %xmm11,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,16-128(%rax)
> + paddd %xmm14,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -96(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm11,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm11,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm13,%xmm0
> + pand %xmm12,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm15,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm15,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa %xmm8,%xmm4
> + movdqa %xmm15,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm15,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm8,%xmm14
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm14
> + paddd %xmm5,%xmm10
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm14
> + paddd %xmm7,%xmm14
> + movd 8(%r8),%xmm5
> + movd 8(%r9),%xmm0
> + movd 8(%r10),%xmm1
> + movd 8(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm10,%xmm7
> +.byte 102,15,56,0,238
> + movdqa %xmm10,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm10,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,32-128(%rax)
> + paddd %xmm13,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -64(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm10,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm10,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm12,%xmm0
> + pand %xmm11,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm14,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm14,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm15,%xmm3
> + movdqa %xmm14,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm14,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm15,%xmm13
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm13
> + paddd %xmm5,%xmm9
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm13
> + paddd %xmm7,%xmm13
> + movd 12(%r8),%xmm5
> + movd 12(%r9),%xmm0
> + movd 12(%r10),%xmm1
> + movd 12(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm9,%xmm7
> +
> + movdqa %xmm9,%xmm2
> +.byte 102,15,56,0,238
> + psrld $6,%xmm7
> + movdqa %xmm9,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,48-128(%rax)
> + paddd %xmm12,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -32(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm9,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm9,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm11,%xmm0
> + pand %xmm10,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm13,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm13,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa %xmm14,%xmm4
> + movdqa %xmm13,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm13,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm14,%xmm12
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm12
> + paddd %xmm5,%xmm8
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm12
> + paddd %xmm7,%xmm12
> + movd 16(%r8),%xmm5
> + movd 16(%r9),%xmm0
> + movd 16(%r10),%xmm1
> + movd 16(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm8,%xmm7
> +.byte 102,15,56,0,238
> + movdqa %xmm8,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm8,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,64-128(%rax)
> + paddd %xmm11,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 0(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm8,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm8,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm10,%xmm0
> + pand %xmm9,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm12,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm12,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm13,%xmm3
> + movdqa %xmm12,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm12,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm13,%xmm11
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm11
> + paddd %xmm5,%xmm15
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm11
> + paddd %xmm7,%xmm11
> + movd 20(%r8),%xmm5
> + movd 20(%r9),%xmm0
> + movd 20(%r10),%xmm1
> + movd 20(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm15,%xmm7
> +
> + movdqa %xmm15,%xmm2
> +.byte 102,15,56,0,238
> + psrld $6,%xmm7
> + movdqa %xmm15,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,80-128(%rax)
> + paddd %xmm10,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 32(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm15,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm15,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm9,%xmm0
> + pand %xmm8,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm11,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm11,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa %xmm12,%xmm4
> + movdqa %xmm11,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm11,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm12,%xmm10
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm10
> + paddd %xmm5,%xmm14
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm10
> + paddd %xmm7,%xmm10
> + movd 24(%r8),%xmm5
> + movd 24(%r9),%xmm0
> + movd 24(%r10),%xmm1
> + movd 24(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm14,%xmm7
> +.byte 102,15,56,0,238
> + movdqa %xmm14,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm14,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,96-128(%rax)
> + paddd %xmm9,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 64(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm14,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm14,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm8,%xmm0
> + pand %xmm15,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm10,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm10,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm11,%xmm3
> + movdqa %xmm10,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm10,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm11,%xmm9
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm9
> + paddd %xmm5,%xmm13
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm9
> + paddd %xmm7,%xmm9
> + movd 28(%r8),%xmm5
> + movd 28(%r9),%xmm0
> + movd 28(%r10),%xmm1
> + movd 28(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm13,%xmm7
> +
> + movdqa %xmm13,%xmm2
> +.byte 102,15,56,0,238
> + psrld $6,%xmm7
> + movdqa %xmm13,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,112-128(%rax)
> + paddd %xmm8,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 96(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm13,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm13,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm15,%xmm0
> + pand %xmm14,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm9,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm9,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa %xmm10,%xmm4
> + movdqa %xmm9,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm9,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm10,%xmm8
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm8
> + paddd %xmm5,%xmm12
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm8
> + paddd %xmm7,%xmm8
> + leaq 256(%rbp),%rbp
> + movd 32(%r8),%xmm5
> + movd 32(%r9),%xmm0
> + movd 32(%r10),%xmm1
> + movd 32(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm12,%xmm7
> +.byte 102,15,56,0,238
> + movdqa %xmm12,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm12,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,128-128(%rax)
> + paddd %xmm15,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -128(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm12,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm12,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm14,%xmm0
> + pand %xmm13,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm8,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm8,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm9,%xmm3
> + movdqa %xmm8,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm8,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm9,%xmm15
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm15
> + paddd %xmm5,%xmm11
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm15
> + paddd %xmm7,%xmm15
> + movd 36(%r8),%xmm5
> + movd 36(%r9),%xmm0
> + movd 36(%r10),%xmm1
> + movd 36(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm11,%xmm7
> +
> + movdqa %xmm11,%xmm2
> +.byte 102,15,56,0,238
> + psrld $6,%xmm7
> + movdqa %xmm11,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,144-128(%rax)
> + paddd %xmm14,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -96(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm11,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm11,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm13,%xmm0
> + pand %xmm12,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm15,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm15,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa %xmm8,%xmm4
> + movdqa %xmm15,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm15,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm8,%xmm14
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm14
> + paddd %xmm5,%xmm10
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm14
> + paddd %xmm7,%xmm14
> + movd 40(%r8),%xmm5
> + movd 40(%r9),%xmm0
> + movd 40(%r10),%xmm1
> + movd 40(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm10,%xmm7
> +.byte 102,15,56,0,238
> + movdqa %xmm10,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm10,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,160-128(%rax)
> + paddd %xmm13,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -64(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm10,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm10,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm12,%xmm0
> + pand %xmm11,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm14,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm14,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm15,%xmm3
> + movdqa %xmm14,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm14,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm15,%xmm13
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm13
> + paddd %xmm5,%xmm9
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm13
> + paddd %xmm7,%xmm13
> + movd 44(%r8),%xmm5
> + movd 44(%r9),%xmm0
> + movd 44(%r10),%xmm1
> + movd 44(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm9,%xmm7
> +
> + movdqa %xmm9,%xmm2
> +.byte 102,15,56,0,238
> + psrld $6,%xmm7
> + movdqa %xmm9,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,176-128(%rax)
> + paddd %xmm12,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -32(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm9,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm9,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm11,%xmm0
> + pand %xmm10,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm13,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm13,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa %xmm14,%xmm4
> + movdqa %xmm13,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm13,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm14,%xmm12
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm12
> + paddd %xmm5,%xmm8
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm12
> + paddd %xmm7,%xmm12
> + movd 48(%r8),%xmm5
> + movd 48(%r9),%xmm0
> + movd 48(%r10),%xmm1
> + movd 48(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm8,%xmm7
> +.byte 102,15,56,0,238
> + movdqa %xmm8,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm8,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,192-128(%rax)
> + paddd %xmm11,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 0(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm8,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm8,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm10,%xmm0
> + pand %xmm9,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm12,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm12,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm13,%xmm3
> + movdqa %xmm12,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm12,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm13,%xmm11
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm11
> + paddd %xmm5,%xmm15
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm11
> + paddd %xmm7,%xmm11
> + movd 52(%r8),%xmm5
> + movd 52(%r9),%xmm0
> + movd 52(%r10),%xmm1
> + movd 52(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm15,%xmm7
> +
> + movdqa %xmm15,%xmm2
> +.byte 102,15,56,0,238
> + psrld $6,%xmm7
> + movdqa %xmm15,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,208-128(%rax)
> + paddd %xmm10,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 32(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm15,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm15,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm9,%xmm0
> + pand %xmm8,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm11,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm11,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa %xmm12,%xmm4
> + movdqa %xmm11,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm11,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm12,%xmm10
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm10
> + paddd %xmm5,%xmm14
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm10
> + paddd %xmm7,%xmm10
> + movd 56(%r8),%xmm5
> + movd 56(%r9),%xmm0
> + movd 56(%r10),%xmm1
> + movd 56(%r11),%xmm2
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm14,%xmm7
> +.byte 102,15,56,0,238
> + movdqa %xmm14,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm14,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,224-128(%rax)
> + paddd %xmm9,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 64(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm14,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm14,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm8,%xmm0
> + pand %xmm15,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm10,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm10,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm11,%xmm3
> + movdqa %xmm10,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm10,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm11,%xmm9
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm9
> + paddd %xmm5,%xmm13
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm9
> + paddd %xmm7,%xmm9
> + movd 60(%r8),%xmm5
> + leaq 64(%r8),%r8
> + movd 60(%r9),%xmm0
> + leaq 64(%r9),%r9
> + movd 60(%r10),%xmm1
> + leaq 64(%r10),%r10
> + movd 60(%r11),%xmm2
> + leaq 64(%r11),%r11
> + punpckldq %xmm1,%xmm5
> + punpckldq %xmm2,%xmm0
> + punpckldq %xmm0,%xmm5
> + movdqa %xmm13,%xmm7
> +
> + movdqa %xmm13,%xmm2
> +.byte 102,15,56,0,238
> + psrld $6,%xmm7
> + movdqa %xmm13,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,240-128(%rax)
> + paddd %xmm8,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 96(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm13,%xmm0
> + prefetcht0 63(%r8)
> + pxor %xmm2,%xmm7
> + movdqa %xmm13,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm15,%xmm0
> + pand %xmm14,%xmm4
> + pxor %xmm1,%xmm7
> +
> + prefetcht0 63(%r9)
> + movdqa %xmm9,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm9,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm4,%xmm0
> + movdqa %xmm10,%xmm4
> + movdqa %xmm9,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm9,%xmm4
> +
> + prefetcht0 63(%r10)
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> + prefetcht0 63(%r11)
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm10,%xmm8
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm8
> + paddd %xmm5,%xmm12
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm8
> + paddd %xmm7,%xmm8
> + leaq 256(%rbp),%rbp
> + movdqu 0-128(%rax),%xmm5
> + movl $3,%ecx
> + jmp .Loop_16_xx
> +.align 32
> +.Loop_16_xx:
> + movdqa 16-128(%rax),%xmm6
> + paddd 144-128(%rax),%xmm5
> +
> + movdqa %xmm6,%xmm7
> + movdqa %xmm6,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm6,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 224-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm3,%xmm1
> +
> + psrld $17,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + psrld $19-17,%xmm3
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm3,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm5
> + movdqa %xmm12,%xmm7
> +
> + movdqa %xmm12,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm12,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,0-128(%rax)
> + paddd %xmm15,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -128(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm12,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm12,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm14,%xmm0
> + pand %xmm13,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm8,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm8,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm9,%xmm3
> + movdqa %xmm8,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm8,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm9,%xmm15
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm15
> + paddd %xmm5,%xmm11
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm15
> + paddd %xmm7,%xmm15
> + movdqa 32-128(%rax),%xmm5
> + paddd 160-128(%rax),%xmm6
> +
> + movdqa %xmm5,%xmm7
> + movdqa %xmm5,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm5,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 240-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm4,%xmm1
> +
> + psrld $17,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + psrld $19-17,%xmm4
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm4,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + movdqa %xmm11,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm11,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm6,16-128(%rax)
> + paddd %xmm14,%xmm6
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -96(%rbp),%xmm6
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm11,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm11,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm13,%xmm0
> + pand %xmm12,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm15,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm15,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + movdqa %xmm8,%xmm4
> + movdqa %xmm15,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm15,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm6
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm8,%xmm14
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm14
> + paddd %xmm6,%xmm10
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm6,%xmm14
> + paddd %xmm7,%xmm14
> + movdqa 48-128(%rax),%xmm6
> + paddd 176-128(%rax),%xmm5
> +
> + movdqa %xmm6,%xmm7
> + movdqa %xmm6,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm6,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 0-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm3,%xmm1
> +
> + psrld $17,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + psrld $19-17,%xmm3
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm3,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm5
> + movdqa %xmm10,%xmm7
> +
> + movdqa %xmm10,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm10,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,32-128(%rax)
> + paddd %xmm13,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -64(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm10,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm10,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm12,%xmm0
> + pand %xmm11,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm14,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm14,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm15,%xmm3
> + movdqa %xmm14,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm14,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm15,%xmm13
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm13
> + paddd %xmm5,%xmm9
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm13
> + paddd %xmm7,%xmm13
> + movdqa 64-128(%rax),%xmm5
> + paddd 192-128(%rax),%xmm6
> +
> + movdqa %xmm5,%xmm7
> + movdqa %xmm5,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm5,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 16-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm4,%xmm1
> +
> + psrld $17,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + psrld $19-17,%xmm4
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm4,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm6
> + movdqa %xmm9,%xmm7
> +
> + movdqa %xmm9,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm9,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm6,48-128(%rax)
> + paddd %xmm12,%xmm6
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -32(%rbp),%xmm6
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm9,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm9,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm11,%xmm0
> + pand %xmm10,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm13,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm13,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + movdqa %xmm14,%xmm4
> + movdqa %xmm13,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm13,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm6
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm14,%xmm12
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm12
> + paddd %xmm6,%xmm8
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm6,%xmm12
> + paddd %xmm7,%xmm12
> + movdqa 80-128(%rax),%xmm6
> + paddd 208-128(%rax),%xmm5
> +
> + movdqa %xmm6,%xmm7
> + movdqa %xmm6,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm6,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 32-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm3,%xmm1
> +
> + psrld $17,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + psrld $19-17,%xmm3
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm3,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm5
> + movdqa %xmm8,%xmm7
> +
> + movdqa %xmm8,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm8,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,64-128(%rax)
> + paddd %xmm11,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 0(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm8,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm8,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm10,%xmm0
> + pand %xmm9,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm12,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm12,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm13,%xmm3
> + movdqa %xmm12,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm12,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm13,%xmm11
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm11
> + paddd %xmm5,%xmm15
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm11
> + paddd %xmm7,%xmm11
> + movdqa 96-128(%rax),%xmm5
> + paddd 224-128(%rax),%xmm6
> +
> + movdqa %xmm5,%xmm7
> + movdqa %xmm5,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm5,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 48-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm4,%xmm1
> +
> + psrld $17,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + psrld $19-17,%xmm4
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm4,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm6
> + movdqa %xmm15,%xmm7
> +
> + movdqa %xmm15,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm15,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm6,80-128(%rax)
> + paddd %xmm10,%xmm6
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 32(%rbp),%xmm6
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm15,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm15,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm9,%xmm0
> + pand %xmm8,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm11,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm11,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + movdqa %xmm12,%xmm4
> + movdqa %xmm11,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm11,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm6
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm12,%xmm10
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm10
> + paddd %xmm6,%xmm14
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm6,%xmm10
> + paddd %xmm7,%xmm10
> + movdqa 112-128(%rax),%xmm6
> + paddd 240-128(%rax),%xmm5
> +
> + movdqa %xmm6,%xmm7
> + movdqa %xmm6,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm6,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 64-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm3,%xmm1
> +
> + psrld $17,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + psrld $19-17,%xmm3
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm3,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm5
> + movdqa %xmm14,%xmm7
> +
> + movdqa %xmm14,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm14,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,96-128(%rax)
> + paddd %xmm9,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 64(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm14,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm14,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm8,%xmm0
> + pand %xmm15,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm10,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm10,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm11,%xmm3
> + movdqa %xmm10,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm10,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm11,%xmm9
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm9
> + paddd %xmm5,%xmm13
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm9
> + paddd %xmm7,%xmm9
> + movdqa 128-128(%rax),%xmm5
> + paddd 0-128(%rax),%xmm6
> +
> + movdqa %xmm5,%xmm7
> + movdqa %xmm5,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm5,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 80-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm4,%xmm1
> +
> + psrld $17,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + psrld $19-17,%xmm4
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm4,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + movdqa %xmm13,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm13,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm6,112-128(%rax)
> + paddd %xmm8,%xmm6
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 96(%rbp),%xmm6
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm13,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm13,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm15,%xmm0
> + pand %xmm14,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm9,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm9,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + movdqa %xmm10,%xmm4
> + movdqa %xmm9,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm9,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm6
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm10,%xmm8
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm8
> + paddd %xmm6,%xmm12
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm6,%xmm8
> + paddd %xmm7,%xmm8
> + leaq 256(%rbp),%rbp
> + movdqa 144-128(%rax),%xmm6
> + paddd 16-128(%rax),%xmm5
> +
> + movdqa %xmm6,%xmm7
> + movdqa %xmm6,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm6,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 96-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm3,%xmm1
> +
> + psrld $17,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + psrld $19-17,%xmm3
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm3,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm5
> + movdqa %xmm12,%xmm7
> +
> + movdqa %xmm12,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm12,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,128-128(%rax)
> + paddd %xmm15,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -128(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm12,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm12,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm14,%xmm0
> + pand %xmm13,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm8,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm8,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm9,%xmm3
> + movdqa %xmm8,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm8,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm9,%xmm15
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm15
> + paddd %xmm5,%xmm11
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm15
> + paddd %xmm7,%xmm15
> + movdqa 160-128(%rax),%xmm5
> + paddd 32-128(%rax),%xmm6
> +
> + movdqa %xmm5,%xmm7
> + movdqa %xmm5,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm5,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 112-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm4,%xmm1
> +
> + psrld $17,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + psrld $19-17,%xmm4
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm4,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm6
> + movdqa %xmm11,%xmm7
> +
> + movdqa %xmm11,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm11,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm6,144-128(%rax)
> + paddd %xmm14,%xmm6
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -96(%rbp),%xmm6
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm11,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm11,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm13,%xmm0
> + pand %xmm12,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm15,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm15,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + movdqa %xmm8,%xmm4
> + movdqa %xmm15,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm15,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm6
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm8,%xmm14
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm14
> + paddd %xmm6,%xmm10
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm6,%xmm14
> + paddd %xmm7,%xmm14
> + movdqa 176-128(%rax),%xmm6
> + paddd 48-128(%rax),%xmm5
> +
> + movdqa %xmm6,%xmm7
> + movdqa %xmm6,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm6,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 128-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm3,%xmm1
> +
> + psrld $17,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + psrld $19-17,%xmm3
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm3,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm5
> + movdqa %xmm10,%xmm7
> +
> + movdqa %xmm10,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm10,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,160-128(%rax)
> + paddd %xmm13,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -64(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm10,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm10,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm12,%xmm0
> + pand %xmm11,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm14,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm14,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm15,%xmm3
> + movdqa %xmm14,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm14,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm15,%xmm13
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm13
> + paddd %xmm5,%xmm9
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm13
> + paddd %xmm7,%xmm13
> + movdqa 192-128(%rax),%xmm5
> + paddd 64-128(%rax),%xmm6
> +
> + movdqa %xmm5,%xmm7
> + movdqa %xmm5,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm5,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 144-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm4,%xmm1
> +
> + psrld $17,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + psrld $19-17,%xmm4
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm4,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm6
> + movdqa %xmm9,%xmm7
> +
> + movdqa %xmm9,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm9,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm6,176-128(%rax)
> + paddd %xmm12,%xmm6
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd -32(%rbp),%xmm6
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm9,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm9,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm11,%xmm0
> + pand %xmm10,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm13,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm13,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + movdqa %xmm14,%xmm4
> + movdqa %xmm13,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm13,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm6
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm14,%xmm12
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm12
> + paddd %xmm6,%xmm8
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm6,%xmm12
> + paddd %xmm7,%xmm12
> + movdqa 208-128(%rax),%xmm6
> + paddd 80-128(%rax),%xmm5
> +
> + movdqa %xmm6,%xmm7
> + movdqa %xmm6,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm6,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 160-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm3,%xmm1
> +
> + psrld $17,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + psrld $19-17,%xmm3
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm3,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm5
> + movdqa %xmm8,%xmm7
> +
> + movdqa %xmm8,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm8,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,192-128(%rax)
> + paddd %xmm11,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 0(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm8,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm8,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm10,%xmm0
> + pand %xmm9,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm12,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm12,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm13,%xmm3
> + movdqa %xmm12,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm12,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm13,%xmm11
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm11
> + paddd %xmm5,%xmm15
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm11
> + paddd %xmm7,%xmm11
> + movdqa 224-128(%rax),%xmm5
> + paddd 96-128(%rax),%xmm6
> +
> + movdqa %xmm5,%xmm7
> + movdqa %xmm5,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm5,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 176-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm4,%xmm1
> +
> + psrld $17,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + psrld $19-17,%xmm4
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm4,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm6
> + movdqa %xmm15,%xmm7
> +
> + movdqa %xmm15,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm15,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm6,208-128(%rax)
> + paddd %xmm10,%xmm6
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 32(%rbp),%xmm6
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm15,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm15,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm9,%xmm0
> + pand %xmm8,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm11,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm11,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + movdqa %xmm12,%xmm4
> + movdqa %xmm11,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm11,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm6
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm12,%xmm10
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm10
> + paddd %xmm6,%xmm14
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm6,%xmm10
> + paddd %xmm7,%xmm10
> + movdqa 240-128(%rax),%xmm6
> + paddd 112-128(%rax),%xmm5
> +
> + movdqa %xmm6,%xmm7
> + movdqa %xmm6,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm6,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 192-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm3,%xmm1
> +
> + psrld $17,%xmm3
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + psrld $19-17,%xmm3
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm3,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm5
> + movdqa %xmm14,%xmm7
> +
> + movdqa %xmm14,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm14,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm5,224-128(%rax)
> + paddd %xmm9,%xmm5
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 64(%rbp),%xmm5
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm14,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm14,%xmm3
> + pslld $26-21,%xmm2
> + pandn %xmm8,%xmm0
> + pand %xmm15,%xmm3
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm10,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm10,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm5
> + pxor %xmm3,%xmm0
> + movdqa %xmm11,%xmm3
> + movdqa %xmm10,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm10,%xmm3
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm5
> + pslld $19-10,%xmm2
> + pand %xmm3,%xmm4
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm11,%xmm9
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm4,%xmm9
> + paddd %xmm5,%xmm13
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm5,%xmm9
> + paddd %xmm7,%xmm9
> + movdqa 0-128(%rax),%xmm5
> + paddd 128-128(%rax),%xmm6
> +
> + movdqa %xmm5,%xmm7
> + movdqa %xmm5,%xmm1
> + psrld $3,%xmm7
> + movdqa %xmm5,%xmm2
> +
> + psrld $7,%xmm1
> + movdqa 208-128(%rax),%xmm0
> + pslld $14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $18-7,%xmm1
> + movdqa %xmm0,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $25-14,%xmm2
> + pxor %xmm1,%xmm7
> + psrld $10,%xmm0
> + movdqa %xmm4,%xmm1
> +
> + psrld $17,%xmm4
> + pxor %xmm2,%xmm7
> + pslld $13,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + psrld $19-17,%xmm4
> + pxor %xmm1,%xmm0
> + pslld $15-13,%xmm1
> + pxor %xmm4,%xmm0
> + pxor %xmm1,%xmm0
> + paddd %xmm0,%xmm6
> + movdqa %xmm13,%xmm7
> +
> + movdqa %xmm13,%xmm2
> +
> + psrld $6,%xmm7
> + movdqa %xmm13,%xmm1
> + pslld $7,%xmm2
> + movdqa %xmm6,240-128(%rax)
> + paddd %xmm8,%xmm6
> +
> + psrld $11,%xmm1
> + pxor %xmm2,%xmm7
> + pslld $21-7,%xmm2
> + paddd 96(%rbp),%xmm6
> + pxor %xmm1,%xmm7
> +
> + psrld $25-11,%xmm1
> + movdqa %xmm13,%xmm0
> +
> + pxor %xmm2,%xmm7
> + movdqa %xmm13,%xmm4
> + pslld $26-21,%xmm2
> + pandn %xmm15,%xmm0
> + pand %xmm14,%xmm4
> + pxor %xmm1,%xmm7
> +
> +
> + movdqa %xmm9,%xmm1
> + pxor %xmm2,%xmm7
> + movdqa %xmm9,%xmm2
> + psrld $2,%xmm1
> + paddd %xmm7,%xmm6
> + pxor %xmm4,%xmm0
> + movdqa %xmm10,%xmm4
> + movdqa %xmm9,%xmm7
> + pslld $10,%xmm2
> + pxor %xmm9,%xmm4
> +
> +
> + psrld $13,%xmm7
> + pxor %xmm2,%xmm1
> + paddd %xmm0,%xmm6
> + pslld $19-10,%xmm2
> + pand %xmm4,%xmm3
> + pxor %xmm7,%xmm1
> +
> +
> + psrld $22-13,%xmm7
> + pxor %xmm2,%xmm1
> + movdqa %xmm10,%xmm8
> + pslld $30-19,%xmm2
> + pxor %xmm1,%xmm7
> + pxor %xmm3,%xmm8
> + paddd %xmm6,%xmm12
> + pxor %xmm2,%xmm7
> +
> + paddd %xmm6,%xmm8
> + paddd %xmm7,%xmm8
> + leaq 256(%rbp),%rbp
> + decl %ecx
> + jnz .Loop_16_xx
> +
> + movl $1,%ecx
> + leaq K256+128(%rip),%rbp
> +
> + movdqa (%rbx),%xmm7
> + cmpl 0(%rbx),%ecx
> + pxor %xmm0,%xmm0
> + cmovgeq %rbp,%r8
> + cmpl 4(%rbx),%ecx
> + movdqa %xmm7,%xmm6
> + cmovgeq %rbp,%r9
> + cmpl 8(%rbx),%ecx
> + pcmpgtd %xmm0,%xmm6
> + cmovgeq %rbp,%r10
> + cmpl 12(%rbx),%ecx
> + paddd %xmm6,%xmm7
> + cmovgeq %rbp,%r11
> +
> + movdqu 0-128(%rdi),%xmm0
> + pand %xmm6,%xmm8
> + movdqu 32-128(%rdi),%xmm1
> + pand %xmm6,%xmm9
> + movdqu 64-128(%rdi),%xmm2
> + pand %xmm6,%xmm10
> + movdqu 96-128(%rdi),%xmm5
> + pand %xmm6,%xmm11
> + paddd %xmm0,%xmm8
> + movdqu 128-128(%rdi),%xmm0
> + pand %xmm6,%xmm12
> + paddd %xmm1,%xmm9
> + movdqu 160-128(%rdi),%xmm1
> + pand %xmm6,%xmm13
> + paddd %xmm2,%xmm10
> + movdqu 192-128(%rdi),%xmm2
> + pand %xmm6,%xmm14
> + paddd %xmm5,%xmm11
> + movdqu 224-128(%rdi),%xmm5
> + pand %xmm6,%xmm15
> + paddd %xmm0,%xmm12
> + paddd %xmm1,%xmm13
> + movdqu %xmm8,0-128(%rdi)
> + paddd %xmm2,%xmm14
> + movdqu %xmm9,32-128(%rdi)
> + paddd %xmm5,%xmm15
> + movdqu %xmm10,64-128(%rdi)
> + movdqu %xmm11,96-128(%rdi)
> + movdqu %xmm12,128-128(%rdi)
> + movdqu %xmm13,160-128(%rdi)
> + movdqu %xmm14,192-128(%rdi)
> + movdqu %xmm15,224-128(%rdi)
> +
> + movdqa %xmm7,(%rbx)
> + movdqa .Lpbswap(%rip),%xmm6
> + decl %edx
> + jnz .Loop
> +
> + movl 280(%rsp),%edx
> + leaq 16(%rdi),%rdi
> + leaq 64(%rsi),%rsi
> + decl %edx
> + jnz .Loop_grande
> +
> +.Ldone:
> + movq 272(%rsp),%rax
> +.cfi_def_cfa %rax,8
> + movq -16(%rax),%rbp
> +.cfi_restore %rbp
> + movq -8(%rax),%rbx
> +.cfi_restore %rbx
> + leaq (%rax),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha256_multi_block,.-sha256_multi_block
> +.type sha256_multi_block_shaext, at function
> +.align 32
> +sha256_multi_block_shaext:
> +.cfi_startproc
> +_shaext_shortcut:
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + subq $288,%rsp
> + shll $1,%edx
> + andq $-256,%rsp
> + leaq 128(%rdi),%rdi
> + movq %rax,272(%rsp)
> +.Lbody_shaext:
> + leaq 256(%rsp),%rbx
> + leaq K256_shaext+128(%rip),%rbp
> +
> +.Loop_grande_shaext:
> + movl %edx,280(%rsp)
> + xorl %edx,%edx
> + movq 0(%rsi),%r8
> + movl 8(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,0(%rbx)
> + cmovleq %rsp,%r8
> + movq 16(%rsi),%r9
> + movl 24(%rsi),%ecx
> + cmpl %edx,%ecx
> + cmovgl %ecx,%edx
> + testl %ecx,%ecx
> + movl %ecx,4(%rbx)
> + cmovleq %rsp,%r9
> + testl %edx,%edx
> + jz .Ldone_shaext
> +
> + movq 0-128(%rdi),%xmm12
> + movq 32-128(%rdi),%xmm4
> + movq 64-128(%rdi),%xmm13
> + movq 96-128(%rdi),%xmm5
> + movq 128-128(%rdi),%xmm8
> + movq 160-128(%rdi),%xmm9
> + movq 192-128(%rdi),%xmm10
> + movq 224-128(%rdi),%xmm11
> +
> + punpckldq %xmm4,%xmm12
> + punpckldq %xmm5,%xmm13
> + punpckldq %xmm9,%xmm8
> + punpckldq %xmm11,%xmm10
> + movdqa K256_shaext-16(%rip),%xmm3
> +
> + movdqa %xmm12,%xmm14
> + movdqa %xmm13,%xmm15
> + punpcklqdq %xmm8,%xmm12
> + punpcklqdq %xmm10,%xmm13
> + punpckhqdq %xmm8,%xmm14
> + punpckhqdq %xmm10,%xmm15
> +
> + pshufd $27,%xmm12,%xmm12
> + pshufd $27,%xmm13,%xmm13
> + pshufd $27,%xmm14,%xmm14
> + pshufd $27,%xmm15,%xmm15
> + jmp .Loop_shaext
> +
> +.align 32
> +.Loop_shaext:
> + movdqu 0(%r8),%xmm4
> + movdqu 0(%r9),%xmm8
> + movdqu 16(%r8),%xmm5
> + movdqu 16(%r9),%xmm9
> + movdqu 32(%r8),%xmm6
> +.byte 102,15,56,0,227
> + movdqu 32(%r9),%xmm10
> +.byte 102,68,15,56,0,195
> + movdqu 48(%r8),%xmm7
> + leaq 64(%r8),%r8
> + movdqu 48(%r9),%xmm11
> + leaq 64(%r9),%r9
> +
> + movdqa 0-128(%rbp),%xmm0
> +.byte 102,15,56,0,235
> + paddd %xmm4,%xmm0
> + pxor %xmm12,%xmm4
> + movdqa %xmm0,%xmm1
> + movdqa 0-128(%rbp),%xmm2
> +.byte 102,68,15,56,0,203
> + paddd %xmm8,%xmm2
> + movdqa %xmm13,80(%rsp)
> +.byte 69,15,56,203,236
> + pxor %xmm14,%xmm8
> + movdqa %xmm2,%xmm0
> + movdqa %xmm15,112(%rsp)
> +.byte 69,15,56,203,254
> + pshufd $0x0e,%xmm1,%xmm0
> + pxor %xmm12,%xmm4
> + movdqa %xmm12,64(%rsp)
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + pxor %xmm14,%xmm8
> + movdqa %xmm14,96(%rsp)
> + movdqa 16-128(%rbp),%xmm1
> + paddd %xmm5,%xmm1
> +.byte 102,15,56,0,243
> +.byte 69,15,56,203,247
> +
> + movdqa %xmm1,%xmm0
> + movdqa 16-128(%rbp),%xmm2
> + paddd %xmm9,%xmm2
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + prefetcht0 127(%r8)
> +.byte 102,15,56,0,251
> +.byte 102,68,15,56,0,211
> + prefetcht0 127(%r9)
> +.byte 69,15,56,203,254
> + pshufd $0x0e,%xmm1,%xmm0
> +.byte 102,68,15,56,0,219
> +.byte 15,56,204,229
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 32-128(%rbp),%xmm1
> + paddd %xmm6,%xmm1
> +.byte 69,15,56,203,247
> +
> + movdqa %xmm1,%xmm0
> + movdqa 32-128(%rbp),%xmm2
> + paddd %xmm10,%xmm2
> +.byte 69,15,56,203,236
> +.byte 69,15,56,204,193
> + movdqa %xmm2,%xmm0
> + movdqa %xmm7,%xmm3
> +.byte 69,15,56,203,254
> + pshufd $0x0e,%xmm1,%xmm0
> +.byte 102,15,58,15,222,4
> + paddd %xmm3,%xmm4
> + movdqa %xmm11,%xmm3
> +.byte 102,65,15,58,15,218,4
> +.byte 15,56,204,238
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 48-128(%rbp),%xmm1
> + paddd %xmm7,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,202
> +
> + movdqa %xmm1,%xmm0
> + movdqa 48-128(%rbp),%xmm2
> + paddd %xmm3,%xmm8
> + paddd %xmm11,%xmm2
> +.byte 15,56,205,231
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm4,%xmm3
> +.byte 102,15,58,15,223,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,195
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm5
> + movdqa %xmm8,%xmm3
> +.byte 102,65,15,58,15,219,4
> +.byte 15,56,204,247
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 64-128(%rbp),%xmm1
> + paddd %xmm4,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,211
> + movdqa %xmm1,%xmm0
> + movdqa 64-128(%rbp),%xmm2
> + paddd %xmm3,%xmm9
> + paddd %xmm8,%xmm2
> +.byte 15,56,205,236
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm5,%xmm3
> +.byte 102,15,58,15,220,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,200
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm6
> + movdqa %xmm9,%xmm3
> +.byte 102,65,15,58,15,216,4
> +.byte 15,56,204,252
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 80-128(%rbp),%xmm1
> + paddd %xmm5,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,216
> + movdqa %xmm1,%xmm0
> + movdqa 80-128(%rbp),%xmm2
> + paddd %xmm3,%xmm10
> + paddd %xmm9,%xmm2
> +.byte 15,56,205,245
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm6,%xmm3
> +.byte 102,15,58,15,221,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,209
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm7
> + movdqa %xmm10,%xmm3
> +.byte 102,65,15,58,15,217,4
> +.byte 15,56,204,229
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 96-128(%rbp),%xmm1
> + paddd %xmm6,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,193
> + movdqa %xmm1,%xmm0
> + movdqa 96-128(%rbp),%xmm2
> + paddd %xmm3,%xmm11
> + paddd %xmm10,%xmm2
> +.byte 15,56,205,254
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm7,%xmm3
> +.byte 102,15,58,15,222,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,218
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm4
> + movdqa %xmm11,%xmm3
> +.byte 102,65,15,58,15,218,4
> +.byte 15,56,204,238
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 112-128(%rbp),%xmm1
> + paddd %xmm7,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,202
> + movdqa %xmm1,%xmm0
> + movdqa 112-128(%rbp),%xmm2
> + paddd %xmm3,%xmm8
> + paddd %xmm11,%xmm2
> +.byte 15,56,205,231
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm4,%xmm3
> +.byte 102,15,58,15,223,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,195
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm5
> + movdqa %xmm8,%xmm3
> +.byte 102,65,15,58,15,219,4
> +.byte 15,56,204,247
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 128-128(%rbp),%xmm1
> + paddd %xmm4,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,211
> + movdqa %xmm1,%xmm0
> + movdqa 128-128(%rbp),%xmm2
> + paddd %xmm3,%xmm9
> + paddd %xmm8,%xmm2
> +.byte 15,56,205,236
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm5,%xmm3
> +.byte 102,15,58,15,220,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,200
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm6
> + movdqa %xmm9,%xmm3
> +.byte 102,65,15,58,15,216,4
> +.byte 15,56,204,252
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 144-128(%rbp),%xmm1
> + paddd %xmm5,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,216
> + movdqa %xmm1,%xmm0
> + movdqa 144-128(%rbp),%xmm2
> + paddd %xmm3,%xmm10
> + paddd %xmm9,%xmm2
> +.byte 15,56,205,245
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm6,%xmm3
> +.byte 102,15,58,15,221,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,209
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm7
> + movdqa %xmm10,%xmm3
> +.byte 102,65,15,58,15,217,4
> +.byte 15,56,204,229
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 160-128(%rbp),%xmm1
> + paddd %xmm6,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,193
> + movdqa %xmm1,%xmm0
> + movdqa 160-128(%rbp),%xmm2
> + paddd %xmm3,%xmm11
> + paddd %xmm10,%xmm2
> +.byte 15,56,205,254
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm7,%xmm3
> +.byte 102,15,58,15,222,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,218
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm4
> + movdqa %xmm11,%xmm3
> +.byte 102,65,15,58,15,218,4
> +.byte 15,56,204,238
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 176-128(%rbp),%xmm1
> + paddd %xmm7,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,202
> + movdqa %xmm1,%xmm0
> + movdqa 176-128(%rbp),%xmm2
> + paddd %xmm3,%xmm8
> + paddd %xmm11,%xmm2
> +.byte 15,56,205,231
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm4,%xmm3
> +.byte 102,15,58,15,223,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,195
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm5
> + movdqa %xmm8,%xmm3
> +.byte 102,65,15,58,15,219,4
> +.byte 15,56,204,247
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 192-128(%rbp),%xmm1
> + paddd %xmm4,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,211
> + movdqa %xmm1,%xmm0
> + movdqa 192-128(%rbp),%xmm2
> + paddd %xmm3,%xmm9
> + paddd %xmm8,%xmm2
> +.byte 15,56,205,236
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm5,%xmm3
> +.byte 102,15,58,15,220,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,200
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm6
> + movdqa %xmm9,%xmm3
> +.byte 102,65,15,58,15,216,4
> +.byte 15,56,204,252
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 208-128(%rbp),%xmm1
> + paddd %xmm5,%xmm1
> +.byte 69,15,56,203,247
> +.byte 69,15,56,204,216
> + movdqa %xmm1,%xmm0
> + movdqa 208-128(%rbp),%xmm2
> + paddd %xmm3,%xmm10
> + paddd %xmm9,%xmm2
> +.byte 15,56,205,245
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movdqa %xmm6,%xmm3
> +.byte 102,15,58,15,221,4
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,209
> + pshufd $0x0e,%xmm1,%xmm0
> + paddd %xmm3,%xmm7
> + movdqa %xmm10,%xmm3
> +.byte 102,65,15,58,15,217,4
> + nop
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 224-128(%rbp),%xmm1
> + paddd %xmm6,%xmm1
> +.byte 69,15,56,203,247
> +
> + movdqa %xmm1,%xmm0
> + movdqa 224-128(%rbp),%xmm2
> + paddd %xmm3,%xmm11
> + paddd %xmm10,%xmm2
> +.byte 15,56,205,254
> + nop
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + movl $1,%ecx
> + pxor %xmm6,%xmm6
> +.byte 69,15,56,203,254
> +.byte 69,15,56,205,218
> + pshufd $0x0e,%xmm1,%xmm0
> + movdqa 240-128(%rbp),%xmm1
> + paddd %xmm7,%xmm1
> + movq (%rbx),%xmm7
> + nop
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + movdqa 240-128(%rbp),%xmm2
> + paddd %xmm11,%xmm2
> +.byte 69,15,56,203,247
> +
> + movdqa %xmm1,%xmm0
> + cmpl 0(%rbx),%ecx
> + cmovgeq %rsp,%r8
> + cmpl 4(%rbx),%ecx
> + cmovgeq %rsp,%r9
> + pshufd $0x00,%xmm7,%xmm9
> +.byte 69,15,56,203,236
> + movdqa %xmm2,%xmm0
> + pshufd $0x55,%xmm7,%xmm10
> + movdqa %xmm7,%xmm11
> +.byte 69,15,56,203,254
> + pshufd $0x0e,%xmm1,%xmm0
> + pcmpgtd %xmm6,%xmm9
> + pcmpgtd %xmm6,%xmm10
> +.byte 69,15,56,203,229
> + pshufd $0x0e,%xmm2,%xmm0
> + pcmpgtd %xmm6,%xmm11
> + movdqa K256_shaext-16(%rip),%xmm3
> +.byte 69,15,56,203,247
> +
> + pand %xmm9,%xmm13
> + pand %xmm10,%xmm15
> + pand %xmm9,%xmm12
> + pand %xmm10,%xmm14
> + paddd %xmm7,%xmm11
> +
> + paddd 80(%rsp),%xmm13
> + paddd 112(%rsp),%xmm15
> + paddd 64(%rsp),%xmm12
> + paddd 96(%rsp),%xmm14
> +
> + movq %xmm11,(%rbx)
> + decl %edx
> + jnz .Loop_shaext
> +
> + movl 280(%rsp),%edx
> +
> + pshufd $27,%xmm12,%xmm12
> + pshufd $27,%xmm13,%xmm13
> + pshufd $27,%xmm14,%xmm14
> + pshufd $27,%xmm15,%xmm15
> +
> + movdqa %xmm12,%xmm5
> + movdqa %xmm13,%xmm6
> + punpckldq %xmm14,%xmm12
> + punpckhdq %xmm14,%xmm5
> + punpckldq %xmm15,%xmm13
> + punpckhdq %xmm15,%xmm6
> +
> + movq %xmm12,0-128(%rdi)
> + psrldq $8,%xmm12
> + movq %xmm5,128-128(%rdi)
> + psrldq $8,%xmm5
> + movq %xmm12,32-128(%rdi)
> + movq %xmm5,160-128(%rdi)
> +
> + movq %xmm13,64-128(%rdi)
> + psrldq $8,%xmm13
> + movq %xmm6,192-128(%rdi)
> + psrldq $8,%xmm6
> + movq %xmm13,96-128(%rdi)
> + movq %xmm6,224-128(%rdi)
> +
> + leaq 8(%rdi),%rdi
> + leaq 32(%rsi),%rsi
> + decl %edx
> + jnz .Loop_grande_shaext
> +
> +.Ldone_shaext:
> +
> + movq -16(%rax),%rbp
> +.cfi_restore %rbp
> + movq -8(%rax),%rbx
> +.cfi_restore %rbx
> + leaq (%rax),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue_shaext:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha256_multi_block_shaext,.-sha256_multi_block_shaext
> +.align 256
> +K256:
> +.long 1116352408,1116352408,1116352408,1116352408
> +.long 1116352408,1116352408,1116352408,1116352408
> +.long 1899447441,1899447441,1899447441,1899447441
> +.long 1899447441,1899447441,1899447441,1899447441
> +.long 3049323471,3049323471,3049323471,3049323471
> +.long 3049323471,3049323471,3049323471,3049323471
> +.long 3921009573,3921009573,3921009573,3921009573
> +.long 3921009573,3921009573,3921009573,3921009573
> +.long 961987163,961987163,961987163,961987163
> +.long 961987163,961987163,961987163,961987163
> +.long 1508970993,1508970993,1508970993,1508970993
> +.long 1508970993,1508970993,1508970993,1508970993
> +.long 2453635748,2453635748,2453635748,2453635748
> +.long 2453635748,2453635748,2453635748,2453635748
> +.long 2870763221,2870763221,2870763221,2870763221
> +.long 2870763221,2870763221,2870763221,2870763221
> +.long 3624381080,3624381080,3624381080,3624381080
> +.long 3624381080,3624381080,3624381080,3624381080
> +.long 310598401,310598401,310598401,310598401
> +.long 310598401,310598401,310598401,310598401
> +.long 607225278,607225278,607225278,607225278
> +.long 607225278,607225278,607225278,607225278
> +.long 1426881987,1426881987,1426881987,1426881987
> +.long 1426881987,1426881987,1426881987,1426881987
> +.long 1925078388,1925078388,1925078388,1925078388
> +.long 1925078388,1925078388,1925078388,1925078388
> +.long 2162078206,2162078206,2162078206,2162078206
> +.long 2162078206,2162078206,2162078206,2162078206
> +.long 2614888103,2614888103,2614888103,2614888103
> +.long 2614888103,2614888103,2614888103,2614888103
> +.long 3248222580,3248222580,3248222580,3248222580
> +.long 3248222580,3248222580,3248222580,3248222580
> +.long 3835390401,3835390401,3835390401,3835390401
> +.long 3835390401,3835390401,3835390401,3835390401
> +.long 4022224774,4022224774,4022224774,4022224774
> +.long 4022224774,4022224774,4022224774,4022224774
> +.long 264347078,264347078,264347078,264347078
> +.long 264347078,264347078,264347078,264347078
> +.long 604807628,604807628,604807628,604807628
> +.long 604807628,604807628,604807628,604807628
> +.long 770255983,770255983,770255983,770255983
> +.long 770255983,770255983,770255983,770255983
> +.long 1249150122,1249150122,1249150122,1249150122
> +.long 1249150122,1249150122,1249150122,1249150122
> +.long 1555081692,1555081692,1555081692,1555081692
> +.long 1555081692,1555081692,1555081692,1555081692
> +.long 1996064986,1996064986,1996064986,1996064986
> +.long 1996064986,1996064986,1996064986,1996064986
> +.long 2554220882,2554220882,2554220882,2554220882
> +.long 2554220882,2554220882,2554220882,2554220882
> +.long 2821834349,2821834349,2821834349,2821834349
> +.long 2821834349,2821834349,2821834349,2821834349
> +.long 2952996808,2952996808,2952996808,2952996808
> +.long 2952996808,2952996808,2952996808,2952996808
> +.long 3210313671,3210313671,3210313671,3210313671
> +.long 3210313671,3210313671,3210313671,3210313671
> +.long 3336571891,3336571891,3336571891,3336571891
> +.long 3336571891,3336571891,3336571891,3336571891
> +.long 3584528711,3584528711,3584528711,3584528711
> +.long 3584528711,3584528711,3584528711,3584528711
> +.long 113926993,113926993,113926993,113926993
> +.long 113926993,113926993,113926993,113926993
> +.long 338241895,338241895,338241895,338241895
> +.long 338241895,338241895,338241895,338241895
> +.long 666307205,666307205,666307205,666307205
> +.long 666307205,666307205,666307205,666307205
> +.long 773529912,773529912,773529912,773529912
> +.long 773529912,773529912,773529912,773529912
> +.long 1294757372,1294757372,1294757372,1294757372
> +.long 1294757372,1294757372,1294757372,1294757372
> +.long 1396182291,1396182291,1396182291,1396182291
> +.long 1396182291,1396182291,1396182291,1396182291
> +.long 1695183700,1695183700,1695183700,1695183700
> +.long 1695183700,1695183700,1695183700,1695183700
> +.long 1986661051,1986661051,1986661051,1986661051
> +.long 1986661051,1986661051,1986661051,1986661051
> +.long 2177026350,2177026350,2177026350,2177026350
> +.long 2177026350,2177026350,2177026350,2177026350
> +.long 2456956037,2456956037,2456956037,2456956037
> +.long 2456956037,2456956037,2456956037,2456956037
> +.long 2730485921,2730485921,2730485921,2730485921
> +.long 2730485921,2730485921,2730485921,2730485921
> +.long 2820302411,2820302411,2820302411,2820302411
> +.long 2820302411,2820302411,2820302411,2820302411
> +.long 3259730800,3259730800,3259730800,3259730800
> +.long 3259730800,3259730800,3259730800,3259730800
> +.long 3345764771,3345764771,3345764771,3345764771
> +.long 3345764771,3345764771,3345764771,3345764771
> +.long 3516065817,3516065817,3516065817,3516065817
> +.long 3516065817,3516065817,3516065817,3516065817
> +.long 3600352804,3600352804,3600352804,3600352804
> +.long 3600352804,3600352804,3600352804,3600352804
> +.long 4094571909,4094571909,4094571909,4094571909
> +.long 4094571909,4094571909,4094571909,4094571909
> +.long 275423344,275423344,275423344,275423344
> +.long 275423344,275423344,275423344,275423344
> +.long 430227734,430227734,430227734,430227734
> +.long 430227734,430227734,430227734,430227734
> +.long 506948616,506948616,506948616,506948616
> +.long 506948616,506948616,506948616,506948616
> +.long 659060556,659060556,659060556,659060556
> +.long 659060556,659060556,659060556,659060556
> +.long 883997877,883997877,883997877,883997877
> +.long 883997877,883997877,883997877,883997877
> +.long 958139571,958139571,958139571,958139571
> +.long 958139571,958139571,958139571,958139571
> +.long 1322822218,1322822218,1322822218,1322822218
> +.long 1322822218,1322822218,1322822218,1322822218
> +.long 1537002063,1537002063,1537002063,1537002063
> +.long 1537002063,1537002063,1537002063,1537002063
> +.long 1747873779,1747873779,1747873779,1747873779
> +.long 1747873779,1747873779,1747873779,1747873779
> +.long 1955562222,1955562222,1955562222,1955562222
> +.long 1955562222,1955562222,1955562222,1955562222
> +.long 2024104815,2024104815,2024104815,2024104815
> +.long 2024104815,2024104815,2024104815,2024104815
> +.long 2227730452,2227730452,2227730452,2227730452
> +.long 2227730452,2227730452,2227730452,2227730452
> +.long 2361852424,2361852424,2361852424,2361852424
> +.long 2361852424,2361852424,2361852424,2361852424
> +.long 2428436474,2428436474,2428436474,2428436474
> +.long 2428436474,2428436474,2428436474,2428436474
> +.long 2756734187,2756734187,2756734187,2756734187
> +.long 2756734187,2756734187,2756734187,2756734187
> +.long 3204031479,3204031479,3204031479,3204031479
> +.long 3204031479,3204031479,3204031479,3204031479
> +.long 3329325298,3329325298,3329325298,3329325298
> +.long 3329325298,3329325298,3329325298,3329325298
> +.Lpbswap:
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +K256_shaext:
> +.long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> +.long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> +.long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> +.long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> +.long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> +.long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> +.long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> +.long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> +.long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> +.long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> +.long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> +.long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> +.long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> +.long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> +.long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> +.long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> +.byte
> 83,72,65,50,53,54,32,109,117,108,116,105,45,98,108,111,99,107,32,116,114,9
> 7,110,115,102,111,114,109,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,
> 89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,1
> 10,115,115,108,46,111,114,103,62,0
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha256-x86_64.S
> b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha256-x86_64.S
> new file mode 100644
> index 0000000000..a5d3cf5068
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha256-x86_64.S
> @@ -0,0 +1,3097 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/sha/asm/sha512-x86_64.pl
> +#
> +# Copyright 2005-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +.globl sha256_block_data_order
> +.type sha256_block_data_order, at function
> +.align 16
> +sha256_block_data_order:
> +.cfi_startproc
> + leaq OPENSSL_ia32cap_P(%rip),%r11
> + movl 0(%r11),%r9d
> + movl 4(%r11),%r10d
> + movl 8(%r11),%r11d
> + testl $536870912,%r11d
> + jnz _shaext_shortcut
> + testl $512,%r10d
> + jnz .Lssse3_shortcut
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_offset %r14,-48
> + pushq %r15
> +.cfi_offset %r15,-56
> + shlq $4,%rdx
> + subq $64+32,%rsp
> + leaq (%rsi,%rdx,4),%rdx
> + andq $-64,%rsp
> + movq %rdi,64+0(%rsp)
> + movq %rsi,64+8(%rsp)
> + movq %rdx,64+16(%rsp)
> + movq %rax,88(%rsp)
> +.cfi_escape 0x0f,0x06,0x77,0xd8,0x00,0x06,0x23,0x08
> +.Lprologue:
> +
> + movl 0(%rdi),%eax
> + movl 4(%rdi),%ebx
> + movl 8(%rdi),%ecx
> + movl 12(%rdi),%edx
> + movl 16(%rdi),%r8d
> + movl 20(%rdi),%r9d
> + movl 24(%rdi),%r10d
> + movl 28(%rdi),%r11d
> + jmp .Lloop
> +
> +.align 16
> +.Lloop:
> + movl %ebx,%edi
> + leaq K256(%rip),%rbp
> + xorl %ecx,%edi
> + movl 0(%rsi),%r12d
> + movl %r8d,%r13d
> + movl %eax,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %r9d,%r15d
> +
> + xorl %r8d,%r13d
> + rorl $9,%r14d
> + xorl %r10d,%r15d
> +
> + movl %r12d,0(%rsp)
> + xorl %eax,%r14d
> + andl %r8d,%r15d
> +
> + rorl $5,%r13d
> + addl %r11d,%r12d
> + xorl %r10d,%r15d
> +
> + rorl $11,%r14d
> + xorl %r8d,%r13d
> + addl %r15d,%r12d
> +
> + movl %eax,%r15d
> + addl (%rbp),%r12d
> + xorl %eax,%r14d
> +
> + xorl %ebx,%r15d
> + rorl $6,%r13d
> + movl %ebx,%r11d
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%r11d
> + addl %r12d,%edx
> + addl %r12d,%r11d
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%r11d
> + movl 4(%rsi),%r12d
> + movl %edx,%r13d
> + movl %r11d,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %r8d,%edi
> +
> + xorl %edx,%r13d
> + rorl $9,%r14d
> + xorl %r9d,%edi
> +
> + movl %r12d,4(%rsp)
> + xorl %r11d,%r14d
> + andl %edx,%edi
> +
> + rorl $5,%r13d
> + addl %r10d,%r12d
> + xorl %r9d,%edi
> +
> + rorl $11,%r14d
> + xorl %edx,%r13d
> + addl %edi,%r12d
> +
> + movl %r11d,%edi
> + addl (%rbp),%r12d
> + xorl %r11d,%r14d
> +
> + xorl %eax,%edi
> + rorl $6,%r13d
> + movl %eax,%r10d
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%r10d
> + addl %r12d,%ecx
> + addl %r12d,%r10d
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%r10d
> + movl 8(%rsi),%r12d
> + movl %ecx,%r13d
> + movl %r10d,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %edx,%r15d
> +
> + xorl %ecx,%r13d
> + rorl $9,%r14d
> + xorl %r8d,%r15d
> +
> + movl %r12d,8(%rsp)
> + xorl %r10d,%r14d
> + andl %ecx,%r15d
> +
> + rorl $5,%r13d
> + addl %r9d,%r12d
> + xorl %r8d,%r15d
> +
> + rorl $11,%r14d
> + xorl %ecx,%r13d
> + addl %r15d,%r12d
> +
> + movl %r10d,%r15d
> + addl (%rbp),%r12d
> + xorl %r10d,%r14d
> +
> + xorl %r11d,%r15d
> + rorl $6,%r13d
> + movl %r11d,%r9d
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%r9d
> + addl %r12d,%ebx
> + addl %r12d,%r9d
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%r9d
> + movl 12(%rsi),%r12d
> + movl %ebx,%r13d
> + movl %r9d,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %ecx,%edi
> +
> + xorl %ebx,%r13d
> + rorl $9,%r14d
> + xorl %edx,%edi
> +
> + movl %r12d,12(%rsp)
> + xorl %r9d,%r14d
> + andl %ebx,%edi
> +
> + rorl $5,%r13d
> + addl %r8d,%r12d
> + xorl %edx,%edi
> +
> + rorl $11,%r14d
> + xorl %ebx,%r13d
> + addl %edi,%r12d
> +
> + movl %r9d,%edi
> + addl (%rbp),%r12d
> + xorl %r9d,%r14d
> +
> + xorl %r10d,%edi
> + rorl $6,%r13d
> + movl %r10d,%r8d
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%r8d
> + addl %r12d,%eax
> + addl %r12d,%r8d
> +
> + leaq 20(%rbp),%rbp
> + addl %r14d,%r8d
> + movl 16(%rsi),%r12d
> + movl %eax,%r13d
> + movl %r8d,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %ebx,%r15d
> +
> + xorl %eax,%r13d
> + rorl $9,%r14d
> + xorl %ecx,%r15d
> +
> + movl %r12d,16(%rsp)
> + xorl %r8d,%r14d
> + andl %eax,%r15d
> +
> + rorl $5,%r13d
> + addl %edx,%r12d
> + xorl %ecx,%r15d
> +
> + rorl $11,%r14d
> + xorl %eax,%r13d
> + addl %r15d,%r12d
> +
> + movl %r8d,%r15d
> + addl (%rbp),%r12d
> + xorl %r8d,%r14d
> +
> + xorl %r9d,%r15d
> + rorl $6,%r13d
> + movl %r9d,%edx
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%edx
> + addl %r12d,%r11d
> + addl %r12d,%edx
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%edx
> + movl 20(%rsi),%r12d
> + movl %r11d,%r13d
> + movl %edx,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %eax,%edi
> +
> + xorl %r11d,%r13d
> + rorl $9,%r14d
> + xorl %ebx,%edi
> +
> + movl %r12d,20(%rsp)
> + xorl %edx,%r14d
> + andl %r11d,%edi
> +
> + rorl $5,%r13d
> + addl %ecx,%r12d
> + xorl %ebx,%edi
> +
> + rorl $11,%r14d
> + xorl %r11d,%r13d
> + addl %edi,%r12d
> +
> + movl %edx,%edi
> + addl (%rbp),%r12d
> + xorl %edx,%r14d
> +
> + xorl %r8d,%edi
> + rorl $6,%r13d
> + movl %r8d,%ecx
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%ecx
> + addl %r12d,%r10d
> + addl %r12d,%ecx
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%ecx
> + movl 24(%rsi),%r12d
> + movl %r10d,%r13d
> + movl %ecx,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %r11d,%r15d
> +
> + xorl %r10d,%r13d
> + rorl $9,%r14d
> + xorl %eax,%r15d
> +
> + movl %r12d,24(%rsp)
> + xorl %ecx,%r14d
> + andl %r10d,%r15d
> +
> + rorl $5,%r13d
> + addl %ebx,%r12d
> + xorl %eax,%r15d
> +
> + rorl $11,%r14d
> + xorl %r10d,%r13d
> + addl %r15d,%r12d
> +
> + movl %ecx,%r15d
> + addl (%rbp),%r12d
> + xorl %ecx,%r14d
> +
> + xorl %edx,%r15d
> + rorl $6,%r13d
> + movl %edx,%ebx
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%ebx
> + addl %r12d,%r9d
> + addl %r12d,%ebx
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%ebx
> + movl 28(%rsi),%r12d
> + movl %r9d,%r13d
> + movl %ebx,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %r10d,%edi
> +
> + xorl %r9d,%r13d
> + rorl $9,%r14d
> + xorl %r11d,%edi
> +
> + movl %r12d,28(%rsp)
> + xorl %ebx,%r14d
> + andl %r9d,%edi
> +
> + rorl $5,%r13d
> + addl %eax,%r12d
> + xorl %r11d,%edi
> +
> + rorl $11,%r14d
> + xorl %r9d,%r13d
> + addl %edi,%r12d
> +
> + movl %ebx,%edi
> + addl (%rbp),%r12d
> + xorl %ebx,%r14d
> +
> + xorl %ecx,%edi
> + rorl $6,%r13d
> + movl %ecx,%eax
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%eax
> + addl %r12d,%r8d
> + addl %r12d,%eax
> +
> + leaq 20(%rbp),%rbp
> + addl %r14d,%eax
> + movl 32(%rsi),%r12d
> + movl %r8d,%r13d
> + movl %eax,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %r9d,%r15d
> +
> + xorl %r8d,%r13d
> + rorl $9,%r14d
> + xorl %r10d,%r15d
> +
> + movl %r12d,32(%rsp)
> + xorl %eax,%r14d
> + andl %r8d,%r15d
> +
> + rorl $5,%r13d
> + addl %r11d,%r12d
> + xorl %r10d,%r15d
> +
> + rorl $11,%r14d
> + xorl %r8d,%r13d
> + addl %r15d,%r12d
> +
> + movl %eax,%r15d
> + addl (%rbp),%r12d
> + xorl %eax,%r14d
> +
> + xorl %ebx,%r15d
> + rorl $6,%r13d
> + movl %ebx,%r11d
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%r11d
> + addl %r12d,%edx
> + addl %r12d,%r11d
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%r11d
> + movl 36(%rsi),%r12d
> + movl %edx,%r13d
> + movl %r11d,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %r8d,%edi
> +
> + xorl %edx,%r13d
> + rorl $9,%r14d
> + xorl %r9d,%edi
> +
> + movl %r12d,36(%rsp)
> + xorl %r11d,%r14d
> + andl %edx,%edi
> +
> + rorl $5,%r13d
> + addl %r10d,%r12d
> + xorl %r9d,%edi
> +
> + rorl $11,%r14d
> + xorl %edx,%r13d
> + addl %edi,%r12d
> +
> + movl %r11d,%edi
> + addl (%rbp),%r12d
> + xorl %r11d,%r14d
> +
> + xorl %eax,%edi
> + rorl $6,%r13d
> + movl %eax,%r10d
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%r10d
> + addl %r12d,%ecx
> + addl %r12d,%r10d
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%r10d
> + movl 40(%rsi),%r12d
> + movl %ecx,%r13d
> + movl %r10d,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %edx,%r15d
> +
> + xorl %ecx,%r13d
> + rorl $9,%r14d
> + xorl %r8d,%r15d
> +
> + movl %r12d,40(%rsp)
> + xorl %r10d,%r14d
> + andl %ecx,%r15d
> +
> + rorl $5,%r13d
> + addl %r9d,%r12d
> + xorl %r8d,%r15d
> +
> + rorl $11,%r14d
> + xorl %ecx,%r13d
> + addl %r15d,%r12d
> +
> + movl %r10d,%r15d
> + addl (%rbp),%r12d
> + xorl %r10d,%r14d
> +
> + xorl %r11d,%r15d
> + rorl $6,%r13d
> + movl %r11d,%r9d
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%r9d
> + addl %r12d,%ebx
> + addl %r12d,%r9d
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%r9d
> + movl 44(%rsi),%r12d
> + movl %ebx,%r13d
> + movl %r9d,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %ecx,%edi
> +
> + xorl %ebx,%r13d
> + rorl $9,%r14d
> + xorl %edx,%edi
> +
> + movl %r12d,44(%rsp)
> + xorl %r9d,%r14d
> + andl %ebx,%edi
> +
> + rorl $5,%r13d
> + addl %r8d,%r12d
> + xorl %edx,%edi
> +
> + rorl $11,%r14d
> + xorl %ebx,%r13d
> + addl %edi,%r12d
> +
> + movl %r9d,%edi
> + addl (%rbp),%r12d
> + xorl %r9d,%r14d
> +
> + xorl %r10d,%edi
> + rorl $6,%r13d
> + movl %r10d,%r8d
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%r8d
> + addl %r12d,%eax
> + addl %r12d,%r8d
> +
> + leaq 20(%rbp),%rbp
> + addl %r14d,%r8d
> + movl 48(%rsi),%r12d
> + movl %eax,%r13d
> + movl %r8d,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %ebx,%r15d
> +
> + xorl %eax,%r13d
> + rorl $9,%r14d
> + xorl %ecx,%r15d
> +
> + movl %r12d,48(%rsp)
> + xorl %r8d,%r14d
> + andl %eax,%r15d
> +
> + rorl $5,%r13d
> + addl %edx,%r12d
> + xorl %ecx,%r15d
> +
> + rorl $11,%r14d
> + xorl %eax,%r13d
> + addl %r15d,%r12d
> +
> + movl %r8d,%r15d
> + addl (%rbp),%r12d
> + xorl %r8d,%r14d
> +
> + xorl %r9d,%r15d
> + rorl $6,%r13d
> + movl %r9d,%edx
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%edx
> + addl %r12d,%r11d
> + addl %r12d,%edx
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%edx
> + movl 52(%rsi),%r12d
> + movl %r11d,%r13d
> + movl %edx,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %eax,%edi
> +
> + xorl %r11d,%r13d
> + rorl $9,%r14d
> + xorl %ebx,%edi
> +
> + movl %r12d,52(%rsp)
> + xorl %edx,%r14d
> + andl %r11d,%edi
> +
> + rorl $5,%r13d
> + addl %ecx,%r12d
> + xorl %ebx,%edi
> +
> + rorl $11,%r14d
> + xorl %r11d,%r13d
> + addl %edi,%r12d
> +
> + movl %edx,%edi
> + addl (%rbp),%r12d
> + xorl %edx,%r14d
> +
> + xorl %r8d,%edi
> + rorl $6,%r13d
> + movl %r8d,%ecx
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%ecx
> + addl %r12d,%r10d
> + addl %r12d,%ecx
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%ecx
> + movl 56(%rsi),%r12d
> + movl %r10d,%r13d
> + movl %ecx,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %r11d,%r15d
> +
> + xorl %r10d,%r13d
> + rorl $9,%r14d
> + xorl %eax,%r15d
> +
> + movl %r12d,56(%rsp)
> + xorl %ecx,%r14d
> + andl %r10d,%r15d
> +
> + rorl $5,%r13d
> + addl %ebx,%r12d
> + xorl %eax,%r15d
> +
> + rorl $11,%r14d
> + xorl %r10d,%r13d
> + addl %r15d,%r12d
> +
> + movl %ecx,%r15d
> + addl (%rbp),%r12d
> + xorl %ecx,%r14d
> +
> + xorl %edx,%r15d
> + rorl $6,%r13d
> + movl %edx,%ebx
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%ebx
> + addl %r12d,%r9d
> + addl %r12d,%ebx
> +
> + leaq 4(%rbp),%rbp
> + addl %r14d,%ebx
> + movl 60(%rsi),%r12d
> + movl %r9d,%r13d
> + movl %ebx,%r14d
> + bswapl %r12d
> + rorl $14,%r13d
> + movl %r10d,%edi
> +
> + xorl %r9d,%r13d
> + rorl $9,%r14d
> + xorl %r11d,%edi
> +
> + movl %r12d,60(%rsp)
> + xorl %ebx,%r14d
> + andl %r9d,%edi
> +
> + rorl $5,%r13d
> + addl %eax,%r12d
> + xorl %r11d,%edi
> +
> + rorl $11,%r14d
> + xorl %r9d,%r13d
> + addl %edi,%r12d
> +
> + movl %ebx,%edi
> + addl (%rbp),%r12d
> + xorl %ebx,%r14d
> +
> + xorl %ecx,%edi
> + rorl $6,%r13d
> + movl %ecx,%eax
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%eax
> + addl %r12d,%r8d
> + addl %r12d,%eax
> +
> + leaq 20(%rbp),%rbp
> + jmp .Lrounds_16_xx
> +.align 16
> +.Lrounds_16_xx:
> + movl 4(%rsp),%r13d
> + movl 56(%rsp),%r15d
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%eax
> + movl %r15d,%r14d
> + rorl $2,%r15d
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%r15d
> + shrl $10,%r14d
> +
> + rorl $17,%r15d
> + xorl %r13d,%r12d
> + xorl %r14d,%r15d
> + addl 36(%rsp),%r12d
> +
> + addl 0(%rsp),%r12d
> + movl %r8d,%r13d
> + addl %r15d,%r12d
> + movl %eax,%r14d
> + rorl $14,%r13d
> + movl %r9d,%r15d
> +
> + xorl %r8d,%r13d
> + rorl $9,%r14d
> + xorl %r10d,%r15d
> +
> + movl %r12d,0(%rsp)
> + xorl %eax,%r14d
> + andl %r8d,%r15d
> +
> + rorl $5,%r13d
> + addl %r11d,%r12d
> + xorl %r10d,%r15d
> +
> + rorl $11,%r14d
> + xorl %r8d,%r13d
> + addl %r15d,%r12d
> +
> + movl %eax,%r15d
> + addl (%rbp),%r12d
> + xorl %eax,%r14d
> +
> + xorl %ebx,%r15d
> + rorl $6,%r13d
> + movl %ebx,%r11d
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%r11d
> + addl %r12d,%edx
> + addl %r12d,%r11d
> +
> + leaq 4(%rbp),%rbp
> + movl 8(%rsp),%r13d
> + movl 60(%rsp),%edi
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%r11d
> + movl %edi,%r14d
> + rorl $2,%edi
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%edi
> + shrl $10,%r14d
> +
> + rorl $17,%edi
> + xorl %r13d,%r12d
> + xorl %r14d,%edi
> + addl 40(%rsp),%r12d
> +
> + addl 4(%rsp),%r12d
> + movl %edx,%r13d
> + addl %edi,%r12d
> + movl %r11d,%r14d
> + rorl $14,%r13d
> + movl %r8d,%edi
> +
> + xorl %edx,%r13d
> + rorl $9,%r14d
> + xorl %r9d,%edi
> +
> + movl %r12d,4(%rsp)
> + xorl %r11d,%r14d
> + andl %edx,%edi
> +
> + rorl $5,%r13d
> + addl %r10d,%r12d
> + xorl %r9d,%edi
> +
> + rorl $11,%r14d
> + xorl %edx,%r13d
> + addl %edi,%r12d
> +
> + movl %r11d,%edi
> + addl (%rbp),%r12d
> + xorl %r11d,%r14d
> +
> + xorl %eax,%edi
> + rorl $6,%r13d
> + movl %eax,%r10d
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%r10d
> + addl %r12d,%ecx
> + addl %r12d,%r10d
> +
> + leaq 4(%rbp),%rbp
> + movl 12(%rsp),%r13d
> + movl 0(%rsp),%r15d
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%r10d
> + movl %r15d,%r14d
> + rorl $2,%r15d
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%r15d
> + shrl $10,%r14d
> +
> + rorl $17,%r15d
> + xorl %r13d,%r12d
> + xorl %r14d,%r15d
> + addl 44(%rsp),%r12d
> +
> + addl 8(%rsp),%r12d
> + movl %ecx,%r13d
> + addl %r15d,%r12d
> + movl %r10d,%r14d
> + rorl $14,%r13d
> + movl %edx,%r15d
> +
> + xorl %ecx,%r13d
> + rorl $9,%r14d
> + xorl %r8d,%r15d
> +
> + movl %r12d,8(%rsp)
> + xorl %r10d,%r14d
> + andl %ecx,%r15d
> +
> + rorl $5,%r13d
> + addl %r9d,%r12d
> + xorl %r8d,%r15d
> +
> + rorl $11,%r14d
> + xorl %ecx,%r13d
> + addl %r15d,%r12d
> +
> + movl %r10d,%r15d
> + addl (%rbp),%r12d
> + xorl %r10d,%r14d
> +
> + xorl %r11d,%r15d
> + rorl $6,%r13d
> + movl %r11d,%r9d
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%r9d
> + addl %r12d,%ebx
> + addl %r12d,%r9d
> +
> + leaq 4(%rbp),%rbp
> + movl 16(%rsp),%r13d
> + movl 4(%rsp),%edi
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%r9d
> + movl %edi,%r14d
> + rorl $2,%edi
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%edi
> + shrl $10,%r14d
> +
> + rorl $17,%edi
> + xorl %r13d,%r12d
> + xorl %r14d,%edi
> + addl 48(%rsp),%r12d
> +
> + addl 12(%rsp),%r12d
> + movl %ebx,%r13d
> + addl %edi,%r12d
> + movl %r9d,%r14d
> + rorl $14,%r13d
> + movl %ecx,%edi
> +
> + xorl %ebx,%r13d
> + rorl $9,%r14d
> + xorl %edx,%edi
> +
> + movl %r12d,12(%rsp)
> + xorl %r9d,%r14d
> + andl %ebx,%edi
> +
> + rorl $5,%r13d
> + addl %r8d,%r12d
> + xorl %edx,%edi
> +
> + rorl $11,%r14d
> + xorl %ebx,%r13d
> + addl %edi,%r12d
> +
> + movl %r9d,%edi
> + addl (%rbp),%r12d
> + xorl %r9d,%r14d
> +
> + xorl %r10d,%edi
> + rorl $6,%r13d
> + movl %r10d,%r8d
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%r8d
> + addl %r12d,%eax
> + addl %r12d,%r8d
> +
> + leaq 20(%rbp),%rbp
> + movl 20(%rsp),%r13d
> + movl 8(%rsp),%r15d
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%r8d
> + movl %r15d,%r14d
> + rorl $2,%r15d
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%r15d
> + shrl $10,%r14d
> +
> + rorl $17,%r15d
> + xorl %r13d,%r12d
> + xorl %r14d,%r15d
> + addl 52(%rsp),%r12d
> +
> + addl 16(%rsp),%r12d
> + movl %eax,%r13d
> + addl %r15d,%r12d
> + movl %r8d,%r14d
> + rorl $14,%r13d
> + movl %ebx,%r15d
> +
> + xorl %eax,%r13d
> + rorl $9,%r14d
> + xorl %ecx,%r15d
> +
> + movl %r12d,16(%rsp)
> + xorl %r8d,%r14d
> + andl %eax,%r15d
> +
> + rorl $5,%r13d
> + addl %edx,%r12d
> + xorl %ecx,%r15d
> +
> + rorl $11,%r14d
> + xorl %eax,%r13d
> + addl %r15d,%r12d
> +
> + movl %r8d,%r15d
> + addl (%rbp),%r12d
> + xorl %r8d,%r14d
> +
> + xorl %r9d,%r15d
> + rorl $6,%r13d
> + movl %r9d,%edx
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%edx
> + addl %r12d,%r11d
> + addl %r12d,%edx
> +
> + leaq 4(%rbp),%rbp
> + movl 24(%rsp),%r13d
> + movl 12(%rsp),%edi
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%edx
> + movl %edi,%r14d
> + rorl $2,%edi
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%edi
> + shrl $10,%r14d
> +
> + rorl $17,%edi
> + xorl %r13d,%r12d
> + xorl %r14d,%edi
> + addl 56(%rsp),%r12d
> +
> + addl 20(%rsp),%r12d
> + movl %r11d,%r13d
> + addl %edi,%r12d
> + movl %edx,%r14d
> + rorl $14,%r13d
> + movl %eax,%edi
> +
> + xorl %r11d,%r13d
> + rorl $9,%r14d
> + xorl %ebx,%edi
> +
> + movl %r12d,20(%rsp)
> + xorl %edx,%r14d
> + andl %r11d,%edi
> +
> + rorl $5,%r13d
> + addl %ecx,%r12d
> + xorl %ebx,%edi
> +
> + rorl $11,%r14d
> + xorl %r11d,%r13d
> + addl %edi,%r12d
> +
> + movl %edx,%edi
> + addl (%rbp),%r12d
> + xorl %edx,%r14d
> +
> + xorl %r8d,%edi
> + rorl $6,%r13d
> + movl %r8d,%ecx
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%ecx
> + addl %r12d,%r10d
> + addl %r12d,%ecx
> +
> + leaq 4(%rbp),%rbp
> + movl 28(%rsp),%r13d
> + movl 16(%rsp),%r15d
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%ecx
> + movl %r15d,%r14d
> + rorl $2,%r15d
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%r15d
> + shrl $10,%r14d
> +
> + rorl $17,%r15d
> + xorl %r13d,%r12d
> + xorl %r14d,%r15d
> + addl 60(%rsp),%r12d
> +
> + addl 24(%rsp),%r12d
> + movl %r10d,%r13d
> + addl %r15d,%r12d
> + movl %ecx,%r14d
> + rorl $14,%r13d
> + movl %r11d,%r15d
> +
> + xorl %r10d,%r13d
> + rorl $9,%r14d
> + xorl %eax,%r15d
> +
> + movl %r12d,24(%rsp)
> + xorl %ecx,%r14d
> + andl %r10d,%r15d
> +
> + rorl $5,%r13d
> + addl %ebx,%r12d
> + xorl %eax,%r15d
> +
> + rorl $11,%r14d
> + xorl %r10d,%r13d
> + addl %r15d,%r12d
> +
> + movl %ecx,%r15d
> + addl (%rbp),%r12d
> + xorl %ecx,%r14d
> +
> + xorl %edx,%r15d
> + rorl $6,%r13d
> + movl %edx,%ebx
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%ebx
> + addl %r12d,%r9d
> + addl %r12d,%ebx
> +
> + leaq 4(%rbp),%rbp
> + movl 32(%rsp),%r13d
> + movl 20(%rsp),%edi
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%ebx
> + movl %edi,%r14d
> + rorl $2,%edi
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%edi
> + shrl $10,%r14d
> +
> + rorl $17,%edi
> + xorl %r13d,%r12d
> + xorl %r14d,%edi
> + addl 0(%rsp),%r12d
> +
> + addl 28(%rsp),%r12d
> + movl %r9d,%r13d
> + addl %edi,%r12d
> + movl %ebx,%r14d
> + rorl $14,%r13d
> + movl %r10d,%edi
> +
> + xorl %r9d,%r13d
> + rorl $9,%r14d
> + xorl %r11d,%edi
> +
> + movl %r12d,28(%rsp)
> + xorl %ebx,%r14d
> + andl %r9d,%edi
> +
> + rorl $5,%r13d
> + addl %eax,%r12d
> + xorl %r11d,%edi
> +
> + rorl $11,%r14d
> + xorl %r9d,%r13d
> + addl %edi,%r12d
> +
> + movl %ebx,%edi
> + addl (%rbp),%r12d
> + xorl %ebx,%r14d
> +
> + xorl %ecx,%edi
> + rorl $6,%r13d
> + movl %ecx,%eax
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%eax
> + addl %r12d,%r8d
> + addl %r12d,%eax
> +
> + leaq 20(%rbp),%rbp
> + movl 36(%rsp),%r13d
> + movl 24(%rsp),%r15d
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%eax
> + movl %r15d,%r14d
> + rorl $2,%r15d
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%r15d
> + shrl $10,%r14d
> +
> + rorl $17,%r15d
> + xorl %r13d,%r12d
> + xorl %r14d,%r15d
> + addl 4(%rsp),%r12d
> +
> + addl 32(%rsp),%r12d
> + movl %r8d,%r13d
> + addl %r15d,%r12d
> + movl %eax,%r14d
> + rorl $14,%r13d
> + movl %r9d,%r15d
> +
> + xorl %r8d,%r13d
> + rorl $9,%r14d
> + xorl %r10d,%r15d
> +
> + movl %r12d,32(%rsp)
> + xorl %eax,%r14d
> + andl %r8d,%r15d
> +
> + rorl $5,%r13d
> + addl %r11d,%r12d
> + xorl %r10d,%r15d
> +
> + rorl $11,%r14d
> + xorl %r8d,%r13d
> + addl %r15d,%r12d
> +
> + movl %eax,%r15d
> + addl (%rbp),%r12d
> + xorl %eax,%r14d
> +
> + xorl %ebx,%r15d
> + rorl $6,%r13d
> + movl %ebx,%r11d
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%r11d
> + addl %r12d,%edx
> + addl %r12d,%r11d
> +
> + leaq 4(%rbp),%rbp
> + movl 40(%rsp),%r13d
> + movl 28(%rsp),%edi
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%r11d
> + movl %edi,%r14d
> + rorl $2,%edi
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%edi
> + shrl $10,%r14d
> +
> + rorl $17,%edi
> + xorl %r13d,%r12d
> + xorl %r14d,%edi
> + addl 8(%rsp),%r12d
> +
> + addl 36(%rsp),%r12d
> + movl %edx,%r13d
> + addl %edi,%r12d
> + movl %r11d,%r14d
> + rorl $14,%r13d
> + movl %r8d,%edi
> +
> + xorl %edx,%r13d
> + rorl $9,%r14d
> + xorl %r9d,%edi
> +
> + movl %r12d,36(%rsp)
> + xorl %r11d,%r14d
> + andl %edx,%edi
> +
> + rorl $5,%r13d
> + addl %r10d,%r12d
> + xorl %r9d,%edi
> +
> + rorl $11,%r14d
> + xorl %edx,%r13d
> + addl %edi,%r12d
> +
> + movl %r11d,%edi
> + addl (%rbp),%r12d
> + xorl %r11d,%r14d
> +
> + xorl %eax,%edi
> + rorl $6,%r13d
> + movl %eax,%r10d
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%r10d
> + addl %r12d,%ecx
> + addl %r12d,%r10d
> +
> + leaq 4(%rbp),%rbp
> + movl 44(%rsp),%r13d
> + movl 32(%rsp),%r15d
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%r10d
> + movl %r15d,%r14d
> + rorl $2,%r15d
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%r15d
> + shrl $10,%r14d
> +
> + rorl $17,%r15d
> + xorl %r13d,%r12d
> + xorl %r14d,%r15d
> + addl 12(%rsp),%r12d
> +
> + addl 40(%rsp),%r12d
> + movl %ecx,%r13d
> + addl %r15d,%r12d
> + movl %r10d,%r14d
> + rorl $14,%r13d
> + movl %edx,%r15d
> +
> + xorl %ecx,%r13d
> + rorl $9,%r14d
> + xorl %r8d,%r15d
> +
> + movl %r12d,40(%rsp)
> + xorl %r10d,%r14d
> + andl %ecx,%r15d
> +
> + rorl $5,%r13d
> + addl %r9d,%r12d
> + xorl %r8d,%r15d
> +
> + rorl $11,%r14d
> + xorl %ecx,%r13d
> + addl %r15d,%r12d
> +
> + movl %r10d,%r15d
> + addl (%rbp),%r12d
> + xorl %r10d,%r14d
> +
> + xorl %r11d,%r15d
> + rorl $6,%r13d
> + movl %r11d,%r9d
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%r9d
> + addl %r12d,%ebx
> + addl %r12d,%r9d
> +
> + leaq 4(%rbp),%rbp
> + movl 48(%rsp),%r13d
> + movl 36(%rsp),%edi
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%r9d
> + movl %edi,%r14d
> + rorl $2,%edi
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%edi
> + shrl $10,%r14d
> +
> + rorl $17,%edi
> + xorl %r13d,%r12d
> + xorl %r14d,%edi
> + addl 16(%rsp),%r12d
> +
> + addl 44(%rsp),%r12d
> + movl %ebx,%r13d
> + addl %edi,%r12d
> + movl %r9d,%r14d
> + rorl $14,%r13d
> + movl %ecx,%edi
> +
> + xorl %ebx,%r13d
> + rorl $9,%r14d
> + xorl %edx,%edi
> +
> + movl %r12d,44(%rsp)
> + xorl %r9d,%r14d
> + andl %ebx,%edi
> +
> + rorl $5,%r13d
> + addl %r8d,%r12d
> + xorl %edx,%edi
> +
> + rorl $11,%r14d
> + xorl %ebx,%r13d
> + addl %edi,%r12d
> +
> + movl %r9d,%edi
> + addl (%rbp),%r12d
> + xorl %r9d,%r14d
> +
> + xorl %r10d,%edi
> + rorl $6,%r13d
> + movl %r10d,%r8d
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%r8d
> + addl %r12d,%eax
> + addl %r12d,%r8d
> +
> + leaq 20(%rbp),%rbp
> + movl 52(%rsp),%r13d
> + movl 40(%rsp),%r15d
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%r8d
> + movl %r15d,%r14d
> + rorl $2,%r15d
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%r15d
> + shrl $10,%r14d
> +
> + rorl $17,%r15d
> + xorl %r13d,%r12d
> + xorl %r14d,%r15d
> + addl 20(%rsp),%r12d
> +
> + addl 48(%rsp),%r12d
> + movl %eax,%r13d
> + addl %r15d,%r12d
> + movl %r8d,%r14d
> + rorl $14,%r13d
> + movl %ebx,%r15d
> +
> + xorl %eax,%r13d
> + rorl $9,%r14d
> + xorl %ecx,%r15d
> +
> + movl %r12d,48(%rsp)
> + xorl %r8d,%r14d
> + andl %eax,%r15d
> +
> + rorl $5,%r13d
> + addl %edx,%r12d
> + xorl %ecx,%r15d
> +
> + rorl $11,%r14d
> + xorl %eax,%r13d
> + addl %r15d,%r12d
> +
> + movl %r8d,%r15d
> + addl (%rbp),%r12d
> + xorl %r8d,%r14d
> +
> + xorl %r9d,%r15d
> + rorl $6,%r13d
> + movl %r9d,%edx
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%edx
> + addl %r12d,%r11d
> + addl %r12d,%edx
> +
> + leaq 4(%rbp),%rbp
> + movl 56(%rsp),%r13d
> + movl 44(%rsp),%edi
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%edx
> + movl %edi,%r14d
> + rorl $2,%edi
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%edi
> + shrl $10,%r14d
> +
> + rorl $17,%edi
> + xorl %r13d,%r12d
> + xorl %r14d,%edi
> + addl 24(%rsp),%r12d
> +
> + addl 52(%rsp),%r12d
> + movl %r11d,%r13d
> + addl %edi,%r12d
> + movl %edx,%r14d
> + rorl $14,%r13d
> + movl %eax,%edi
> +
> + xorl %r11d,%r13d
> + rorl $9,%r14d
> + xorl %ebx,%edi
> +
> + movl %r12d,52(%rsp)
> + xorl %edx,%r14d
> + andl %r11d,%edi
> +
> + rorl $5,%r13d
> + addl %ecx,%r12d
> + xorl %ebx,%edi
> +
> + rorl $11,%r14d
> + xorl %r11d,%r13d
> + addl %edi,%r12d
> +
> + movl %edx,%edi
> + addl (%rbp),%r12d
> + xorl %edx,%r14d
> +
> + xorl %r8d,%edi
> + rorl $6,%r13d
> + movl %r8d,%ecx
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%ecx
> + addl %r12d,%r10d
> + addl %r12d,%ecx
> +
> + leaq 4(%rbp),%rbp
> + movl 60(%rsp),%r13d
> + movl 48(%rsp),%r15d
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%ecx
> + movl %r15d,%r14d
> + rorl $2,%r15d
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%r15d
> + shrl $10,%r14d
> +
> + rorl $17,%r15d
> + xorl %r13d,%r12d
> + xorl %r14d,%r15d
> + addl 28(%rsp),%r12d
> +
> + addl 56(%rsp),%r12d
> + movl %r10d,%r13d
> + addl %r15d,%r12d
> + movl %ecx,%r14d
> + rorl $14,%r13d
> + movl %r11d,%r15d
> +
> + xorl %r10d,%r13d
> + rorl $9,%r14d
> + xorl %eax,%r15d
> +
> + movl %r12d,56(%rsp)
> + xorl %ecx,%r14d
> + andl %r10d,%r15d
> +
> + rorl $5,%r13d
> + addl %ebx,%r12d
> + xorl %eax,%r15d
> +
> + rorl $11,%r14d
> + xorl %r10d,%r13d
> + addl %r15d,%r12d
> +
> + movl %ecx,%r15d
> + addl (%rbp),%r12d
> + xorl %ecx,%r14d
> +
> + xorl %edx,%r15d
> + rorl $6,%r13d
> + movl %edx,%ebx
> +
> + andl %r15d,%edi
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %edi,%ebx
> + addl %r12d,%r9d
> + addl %r12d,%ebx
> +
> + leaq 4(%rbp),%rbp
> + movl 0(%rsp),%r13d
> + movl 52(%rsp),%edi
> +
> + movl %r13d,%r12d
> + rorl $11,%r13d
> + addl %r14d,%ebx
> + movl %edi,%r14d
> + rorl $2,%edi
> +
> + xorl %r12d,%r13d
> + shrl $3,%r12d
> + rorl $7,%r13d
> + xorl %r14d,%edi
> + shrl $10,%r14d
> +
> + rorl $17,%edi
> + xorl %r13d,%r12d
> + xorl %r14d,%edi
> + addl 32(%rsp),%r12d
> +
> + addl 60(%rsp),%r12d
> + movl %r9d,%r13d
> + addl %edi,%r12d
> + movl %ebx,%r14d
> + rorl $14,%r13d
> + movl %r10d,%edi
> +
> + xorl %r9d,%r13d
> + rorl $9,%r14d
> + xorl %r11d,%edi
> +
> + movl %r12d,60(%rsp)
> + xorl %ebx,%r14d
> + andl %r9d,%edi
> +
> + rorl $5,%r13d
> + addl %eax,%r12d
> + xorl %r11d,%edi
> +
> + rorl $11,%r14d
> + xorl %r9d,%r13d
> + addl %edi,%r12d
> +
> + movl %ebx,%edi
> + addl (%rbp),%r12d
> + xorl %ebx,%r14d
> +
> + xorl %ecx,%edi
> + rorl $6,%r13d
> + movl %ecx,%eax
> +
> + andl %edi,%r15d
> + rorl $2,%r14d
> + addl %r13d,%r12d
> +
> + xorl %r15d,%eax
> + addl %r12d,%r8d
> + addl %r12d,%eax
> +
> + leaq 20(%rbp),%rbp
> + cmpb $0,3(%rbp)
> + jnz .Lrounds_16_xx
> +
> + movq 64+0(%rsp),%rdi
> + addl %r14d,%eax
> + leaq 64(%rsi),%rsi
> +
> + addl 0(%rdi),%eax
> + addl 4(%rdi),%ebx
> + addl 8(%rdi),%ecx
> + addl 12(%rdi),%edx
> + addl 16(%rdi),%r8d
> + addl 20(%rdi),%r9d
> + addl 24(%rdi),%r10d
> + addl 28(%rdi),%r11d
> +
> + cmpq 64+16(%rsp),%rsi
> +
> + movl %eax,0(%rdi)
> + movl %ebx,4(%rdi)
> + movl %ecx,8(%rdi)
> + movl %edx,12(%rdi)
> + movl %r8d,16(%rdi)
> + movl %r9d,20(%rdi)
> + movl %r10d,24(%rdi)
> + movl %r11d,28(%rdi)
> + jb .Lloop
> +
> + movq 88(%rsp),%rsi
> +.cfi_def_cfa %rsi,8
> + movq -48(%rsi),%r15
> +.cfi_restore %r15
> + movq -40(%rsi),%r14
> +.cfi_restore %r14
> + movq -32(%rsi),%r13
> +.cfi_restore %r13
> + movq -24(%rsi),%r12
> +.cfi_restore %r12
> + movq -16(%rsi),%rbp
> +.cfi_restore %rbp
> + movq -8(%rsi),%rbx
> +.cfi_restore %rbx
> + leaq (%rsi),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha256_block_data_order,.-sha256_block_data_order
> +.align 64
> +.type K256, at object
> +K256:
> +.long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> +.long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
> +.long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> +.long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
> +.long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> +.long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
> +.long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> +.long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
> +.long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> +.long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
> +.long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> +.long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
> +.long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> +.long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
> +.long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> +.long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
> +.long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> +.long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
> +.long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> +.long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
> +.long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> +.long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
> +.long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> +.long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
> +.long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> +.long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
> +.long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> +.long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
> +.long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> +.long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
> +.long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> +.long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
> +
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f
> +.long 0x03020100,0x0b0a0908,0xffffffff,0xffffffff
> +.long 0x03020100,0x0b0a0908,0xffffffff,0xffffffff
> +.long 0xffffffff,0xffffffff,0x03020100,0x0b0a0908
> +.long 0xffffffff,0xffffffff,0x03020100,0x0b0a0908
> +.byte
> 83,72,65,50,53,54,32,98,108,111,99,107,32,116,114,97,110,115,102,111,114,1
> 09,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,
> 32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,1
> 14,103,62,0
> +.type sha256_block_data_order_shaext, at function
> +.align 64
> +sha256_block_data_order_shaext:
> +_shaext_shortcut:
> +.cfi_startproc
> + leaq K256+128(%rip),%rcx
> + movdqu (%rdi),%xmm1
> + movdqu 16(%rdi),%xmm2
> + movdqa 512-128(%rcx),%xmm7
> +
> + pshufd $0x1b,%xmm1,%xmm0
> + pshufd $0xb1,%xmm1,%xmm1
> + pshufd $0x1b,%xmm2,%xmm2
> + movdqa %xmm7,%xmm8
> +.byte 102,15,58,15,202,8
> + punpcklqdq %xmm0,%xmm2
> + jmp .Loop_shaext
> +
> +.align 16
> +.Loop_shaext:
> + movdqu (%rsi),%xmm3
> + movdqu 16(%rsi),%xmm4
> + movdqu 32(%rsi),%xmm5
> +.byte 102,15,56,0,223
> + movdqu 48(%rsi),%xmm6
> +
> + movdqa 0-128(%rcx),%xmm0
> + paddd %xmm3,%xmm0
> +.byte 102,15,56,0,231
> + movdqa %xmm2,%xmm10
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + nop
> + movdqa %xmm1,%xmm9
> +.byte 15,56,203,202
> +
> + movdqa 32-128(%rcx),%xmm0
> + paddd %xmm4,%xmm0
> +.byte 102,15,56,0,239
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + leaq 64(%rsi),%rsi
> +.byte 15,56,204,220
> +.byte 15,56,203,202
> +
> + movdqa 64-128(%rcx),%xmm0
> + paddd %xmm5,%xmm0
> +.byte 102,15,56,0,247
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm6,%xmm7
> +.byte 102,15,58,15,253,4
> + nop
> + paddd %xmm7,%xmm3
> +.byte 15,56,204,229
> +.byte 15,56,203,202
> +
> + movdqa 96-128(%rcx),%xmm0
> + paddd %xmm6,%xmm0
> +.byte 15,56,205,222
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm3,%xmm7
> +.byte 102,15,58,15,254,4
> + nop
> + paddd %xmm7,%xmm4
> +.byte 15,56,204,238
> +.byte 15,56,203,202
> + movdqa 128-128(%rcx),%xmm0
> + paddd %xmm3,%xmm0
> +.byte 15,56,205,227
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm4,%xmm7
> +.byte 102,15,58,15,251,4
> + nop
> + paddd %xmm7,%xmm5
> +.byte 15,56,204,243
> +.byte 15,56,203,202
> + movdqa 160-128(%rcx),%xmm0
> + paddd %xmm4,%xmm0
> +.byte 15,56,205,236
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm5,%xmm7
> +.byte 102,15,58,15,252,4
> + nop
> + paddd %xmm7,%xmm6
> +.byte 15,56,204,220
> +.byte 15,56,203,202
> + movdqa 192-128(%rcx),%xmm0
> + paddd %xmm5,%xmm0
> +.byte 15,56,205,245
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm6,%xmm7
> +.byte 102,15,58,15,253,4
> + nop
> + paddd %xmm7,%xmm3
> +.byte 15,56,204,229
> +.byte 15,56,203,202
> + movdqa 224-128(%rcx),%xmm0
> + paddd %xmm6,%xmm0
> +.byte 15,56,205,222
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm3,%xmm7
> +.byte 102,15,58,15,254,4
> + nop
> + paddd %xmm7,%xmm4
> +.byte 15,56,204,238
> +.byte 15,56,203,202
> + movdqa 256-128(%rcx),%xmm0
> + paddd %xmm3,%xmm0
> +.byte 15,56,205,227
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm4,%xmm7
> +.byte 102,15,58,15,251,4
> + nop
> + paddd %xmm7,%xmm5
> +.byte 15,56,204,243
> +.byte 15,56,203,202
> + movdqa 288-128(%rcx),%xmm0
> + paddd %xmm4,%xmm0
> +.byte 15,56,205,236
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm5,%xmm7
> +.byte 102,15,58,15,252,4
> + nop
> + paddd %xmm7,%xmm6
> +.byte 15,56,204,220
> +.byte 15,56,203,202
> + movdqa 320-128(%rcx),%xmm0
> + paddd %xmm5,%xmm0
> +.byte 15,56,205,245
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm6,%xmm7
> +.byte 102,15,58,15,253,4
> + nop
> + paddd %xmm7,%xmm3
> +.byte 15,56,204,229
> +.byte 15,56,203,202
> + movdqa 352-128(%rcx),%xmm0
> + paddd %xmm6,%xmm0
> +.byte 15,56,205,222
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm3,%xmm7
> +.byte 102,15,58,15,254,4
> + nop
> + paddd %xmm7,%xmm4
> +.byte 15,56,204,238
> +.byte 15,56,203,202
> + movdqa 384-128(%rcx),%xmm0
> + paddd %xmm3,%xmm0
> +.byte 15,56,205,227
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm4,%xmm7
> +.byte 102,15,58,15,251,4
> + nop
> + paddd %xmm7,%xmm5
> +.byte 15,56,204,243
> +.byte 15,56,203,202
> + movdqa 416-128(%rcx),%xmm0
> + paddd %xmm4,%xmm0
> +.byte 15,56,205,236
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + movdqa %xmm5,%xmm7
> +.byte 102,15,58,15,252,4
> +.byte 15,56,203,202
> + paddd %xmm7,%xmm6
> +
> + movdqa 448-128(%rcx),%xmm0
> + paddd %xmm5,%xmm0
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> +.byte 15,56,205,245
> + movdqa %xmm8,%xmm7
> +.byte 15,56,203,202
> +
> + movdqa 480-128(%rcx),%xmm0
> + paddd %xmm6,%xmm0
> + nop
> +.byte 15,56,203,209
> + pshufd $0x0e,%xmm0,%xmm0
> + decq %rdx
> + nop
> +.byte 15,56,203,202
> +
> + paddd %xmm10,%xmm2
> + paddd %xmm9,%xmm1
> + jnz .Loop_shaext
> +
> + pshufd $0xb1,%xmm2,%xmm2
> + pshufd $0x1b,%xmm1,%xmm7
> + pshufd $0xb1,%xmm1,%xmm1
> + punpckhqdq %xmm2,%xmm1
> +.byte 102,15,58,15,215,8
> +
> + movdqu %xmm1,(%rdi)
> + movdqu %xmm2,16(%rdi)
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha256_block_data_order_shaext,.-sha256_block_data_order_shaext
> +.type sha256_block_data_order_ssse3, at function
> +.align 64
> +sha256_block_data_order_ssse3:
> +.cfi_startproc
> +.Lssse3_shortcut:
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_offset %r14,-48
> + pushq %r15
> +.cfi_offset %r15,-56
> + shlq $4,%rdx
> + subq $96,%rsp
> + leaq (%rsi,%rdx,4),%rdx
> + andq $-64,%rsp
> + movq %rdi,64+0(%rsp)
> + movq %rsi,64+8(%rsp)
> + movq %rdx,64+16(%rsp)
> + movq %rax,88(%rsp)
> +.cfi_escape 0x0f,0x06,0x77,0xd8,0x00,0x06,0x23,0x08
> +.Lprologue_ssse3:
> +
> + movl 0(%rdi),%eax
> + movl 4(%rdi),%ebx
> + movl 8(%rdi),%ecx
> + movl 12(%rdi),%edx
> + movl 16(%rdi),%r8d
> + movl 20(%rdi),%r9d
> + movl 24(%rdi),%r10d
> + movl 28(%rdi),%r11d
> +
> +
> + jmp .Lloop_ssse3
> +.align 16
> +.Lloop_ssse3:
> + movdqa K256+512(%rip),%xmm7
> + movdqu 0(%rsi),%xmm0
> + movdqu 16(%rsi),%xmm1
> + movdqu 32(%rsi),%xmm2
> +.byte 102,15,56,0,199
> + movdqu 48(%rsi),%xmm3
> + leaq K256(%rip),%rbp
> +.byte 102,15,56,0,207
> + movdqa 0(%rbp),%xmm4
> + movdqa 32(%rbp),%xmm5
> +.byte 102,15,56,0,215
> + paddd %xmm0,%xmm4
> + movdqa 64(%rbp),%xmm6
> +.byte 102,15,56,0,223
> + movdqa 96(%rbp),%xmm7
> + paddd %xmm1,%xmm5
> + paddd %xmm2,%xmm6
> + paddd %xmm3,%xmm7
> + movdqa %xmm4,0(%rsp)
> + movl %eax,%r14d
> + movdqa %xmm5,16(%rsp)
> + movl %ebx,%edi
> + movdqa %xmm6,32(%rsp)
> + xorl %ecx,%edi
> + movdqa %xmm7,48(%rsp)
> + movl %r8d,%r13d
> + jmp .Lssse3_00_47
> +
> +.align 16
> +.Lssse3_00_47:
> + subq $-128,%rbp
> + rorl $14,%r13d
> + movdqa %xmm1,%xmm4
> + movl %r14d,%eax
> + movl %r9d,%r12d
> + movdqa %xmm3,%xmm7
> + rorl $9,%r14d
> + xorl %r8d,%r13d
> + xorl %r10d,%r12d
> + rorl $5,%r13d
> + xorl %eax,%r14d
> +.byte 102,15,58,15,224,4
> + andl %r8d,%r12d
> + xorl %r8d,%r13d
> +.byte 102,15,58,15,250,4
> + addl 0(%rsp),%r11d
> + movl %eax,%r15d
> + xorl %r10d,%r12d
> + rorl $11,%r14d
> + movdqa %xmm4,%xmm5
> + xorl %ebx,%r15d
> + addl %r12d,%r11d
> + movdqa %xmm4,%xmm6
> + rorl $6,%r13d
> + andl %r15d,%edi
> + psrld $3,%xmm4
> + xorl %eax,%r14d
> + addl %r13d,%r11d
> + xorl %ebx,%edi
> + paddd %xmm7,%xmm0
> + rorl $2,%r14d
> + addl %r11d,%edx
> + psrld $7,%xmm6
> + addl %edi,%r11d
> + movl %edx,%r13d
> + pshufd $250,%xmm3,%xmm7
> + addl %r11d,%r14d
> + rorl $14,%r13d
> + pslld $14,%xmm5
> + movl %r14d,%r11d
> + movl %r8d,%r12d
> + pxor %xmm6,%xmm4
> + rorl $9,%r14d
> + xorl %edx,%r13d
> + xorl %r9d,%r12d
> + rorl $5,%r13d
> + psrld $11,%xmm6
> + xorl %r11d,%r14d
> + pxor %xmm5,%xmm4
> + andl %edx,%r12d
> + xorl %edx,%r13d
> + pslld $11,%xmm5
> + addl 4(%rsp),%r10d
> + movl %r11d,%edi
> + pxor %xmm6,%xmm4
> + xorl %r9d,%r12d
> + rorl $11,%r14d
> + movdqa %xmm7,%xmm6
> + xorl %eax,%edi
> + addl %r12d,%r10d
> + pxor %xmm5,%xmm4
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %r11d,%r14d
> + psrld $10,%xmm7
> + addl %r13d,%r10d
> + xorl %eax,%r15d
> + paddd %xmm4,%xmm0
> + rorl $2,%r14d
> + addl %r10d,%ecx
> + psrlq $17,%xmm6
> + addl %r15d,%r10d
> + movl %ecx,%r13d
> + addl %r10d,%r14d
> + pxor %xmm6,%xmm7
> + rorl $14,%r13d
> + movl %r14d,%r10d
> + movl %edx,%r12d
> + rorl $9,%r14d
> + psrlq $2,%xmm6
> + xorl %ecx,%r13d
> + xorl %r8d,%r12d
> + pxor %xmm6,%xmm7
> + rorl $5,%r13d
> + xorl %r10d,%r14d
> + andl %ecx,%r12d
> + pshufd $128,%xmm7,%xmm7
> + xorl %ecx,%r13d
> + addl 8(%rsp),%r9d
> + movl %r10d,%r15d
> + psrldq $8,%xmm7
> + xorl %r8d,%r12d
> + rorl $11,%r14d
> + xorl %r11d,%r15d
> + addl %r12d,%r9d
> + rorl $6,%r13d
> + paddd %xmm7,%xmm0
> + andl %r15d,%edi
> + xorl %r10d,%r14d
> + addl %r13d,%r9d
> + pshufd $80,%xmm0,%xmm7
> + xorl %r11d,%edi
> + rorl $2,%r14d
> + addl %r9d,%ebx
> + movdqa %xmm7,%xmm6
> + addl %edi,%r9d
> + movl %ebx,%r13d
> + psrld $10,%xmm7
> + addl %r9d,%r14d
> + rorl $14,%r13d
> + psrlq $17,%xmm6
> + movl %r14d,%r9d
> + movl %ecx,%r12d
> + pxor %xmm6,%xmm7
> + rorl $9,%r14d
> + xorl %ebx,%r13d
> + xorl %edx,%r12d
> + rorl $5,%r13d
> + xorl %r9d,%r14d
> + psrlq $2,%xmm6
> + andl %ebx,%r12d
> + xorl %ebx,%r13d
> + addl 12(%rsp),%r8d
> + pxor %xmm6,%xmm7
> + movl %r9d,%edi
> + xorl %edx,%r12d
> + rorl $11,%r14d
> + pshufd $8,%xmm7,%xmm7
> + xorl %r10d,%edi
> + addl %r12d,%r8d
> + movdqa 0(%rbp),%xmm6
> + rorl $6,%r13d
> + andl %edi,%r15d
> + pslldq $8,%xmm7
> + xorl %r9d,%r14d
> + addl %r13d,%r8d
> + xorl %r10d,%r15d
> + paddd %xmm7,%xmm0
> + rorl $2,%r14d
> + addl %r8d,%eax
> + addl %r15d,%r8d
> + paddd %xmm0,%xmm6
> + movl %eax,%r13d
> + addl %r8d,%r14d
> + movdqa %xmm6,0(%rsp)
> + rorl $14,%r13d
> + movdqa %xmm2,%xmm4
> + movl %r14d,%r8d
> + movl %ebx,%r12d
> + movdqa %xmm0,%xmm7
> + rorl $9,%r14d
> + xorl %eax,%r13d
> + xorl %ecx,%r12d
> + rorl $5,%r13d
> + xorl %r8d,%r14d
> +.byte 102,15,58,15,225,4
> + andl %eax,%r12d
> + xorl %eax,%r13d
> +.byte 102,15,58,15,251,4
> + addl 16(%rsp),%edx
> + movl %r8d,%r15d
> + xorl %ecx,%r12d
> + rorl $11,%r14d
> + movdqa %xmm4,%xmm5
> + xorl %r9d,%r15d
> + addl %r12d,%edx
> + movdqa %xmm4,%xmm6
> + rorl $6,%r13d
> + andl %r15d,%edi
> + psrld $3,%xmm4
> + xorl %r8d,%r14d
> + addl %r13d,%edx
> + xorl %r9d,%edi
> + paddd %xmm7,%xmm1
> + rorl $2,%r14d
> + addl %edx,%r11d
> + psrld $7,%xmm6
> + addl %edi,%edx
> + movl %r11d,%r13d
> + pshufd $250,%xmm0,%xmm7
> + addl %edx,%r14d
> + rorl $14,%r13d
> + pslld $14,%xmm5
> + movl %r14d,%edx
> + movl %eax,%r12d
> + pxor %xmm6,%xmm4
> + rorl $9,%r14d
> + xorl %r11d,%r13d
> + xorl %ebx,%r12d
> + rorl $5,%r13d
> + psrld $11,%xmm6
> + xorl %edx,%r14d
> + pxor %xmm5,%xmm4
> + andl %r11d,%r12d
> + xorl %r11d,%r13d
> + pslld $11,%xmm5
> + addl 20(%rsp),%ecx
> + movl %edx,%edi
> + pxor %xmm6,%xmm4
> + xorl %ebx,%r12d
> + rorl $11,%r14d
> + movdqa %xmm7,%xmm6
> + xorl %r8d,%edi
> + addl %r12d,%ecx
> + pxor %xmm5,%xmm4
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %edx,%r14d
> + psrld $10,%xmm7
> + addl %r13d,%ecx
> + xorl %r8d,%r15d
> + paddd %xmm4,%xmm1
> + rorl $2,%r14d
> + addl %ecx,%r10d
> + psrlq $17,%xmm6
> + addl %r15d,%ecx
> + movl %r10d,%r13d
> + addl %ecx,%r14d
> + pxor %xmm6,%xmm7
> + rorl $14,%r13d
> + movl %r14d,%ecx
> + movl %r11d,%r12d
> + rorl $9,%r14d
> + psrlq $2,%xmm6
> + xorl %r10d,%r13d
> + xorl %eax,%r12d
> + pxor %xmm6,%xmm7
> + rorl $5,%r13d
> + xorl %ecx,%r14d
> + andl %r10d,%r12d
> + pshufd $128,%xmm7,%xmm7
> + xorl %r10d,%r13d
> + addl 24(%rsp),%ebx
> + movl %ecx,%r15d
> + psrldq $8,%xmm7
> + xorl %eax,%r12d
> + rorl $11,%r14d
> + xorl %edx,%r15d
> + addl %r12d,%ebx
> + rorl $6,%r13d
> + paddd %xmm7,%xmm1
> + andl %r15d,%edi
> + xorl %ecx,%r14d
> + addl %r13d,%ebx
> + pshufd $80,%xmm1,%xmm7
> + xorl %edx,%edi
> + rorl $2,%r14d
> + addl %ebx,%r9d
> + movdqa %xmm7,%xmm6
> + addl %edi,%ebx
> + movl %r9d,%r13d
> + psrld $10,%xmm7
> + addl %ebx,%r14d
> + rorl $14,%r13d
> + psrlq $17,%xmm6
> + movl %r14d,%ebx
> + movl %r10d,%r12d
> + pxor %xmm6,%xmm7
> + rorl $9,%r14d
> + xorl %r9d,%r13d
> + xorl %r11d,%r12d
> + rorl $5,%r13d
> + xorl %ebx,%r14d
> + psrlq $2,%xmm6
> + andl %r9d,%r12d
> + xorl %r9d,%r13d
> + addl 28(%rsp),%eax
> + pxor %xmm6,%xmm7
> + movl %ebx,%edi
> + xorl %r11d,%r12d
> + rorl $11,%r14d
> + pshufd $8,%xmm7,%xmm7
> + xorl %ecx,%edi
> + addl %r12d,%eax
> + movdqa 32(%rbp),%xmm6
> + rorl $6,%r13d
> + andl %edi,%r15d
> + pslldq $8,%xmm7
> + xorl %ebx,%r14d
> + addl %r13d,%eax
> + xorl %ecx,%r15d
> + paddd %xmm7,%xmm1
> + rorl $2,%r14d
> + addl %eax,%r8d
> + addl %r15d,%eax
> + paddd %xmm1,%xmm6
> + movl %r8d,%r13d
> + addl %eax,%r14d
> + movdqa %xmm6,16(%rsp)
> + rorl $14,%r13d
> + movdqa %xmm3,%xmm4
> + movl %r14d,%eax
> + movl %r9d,%r12d
> + movdqa %xmm1,%xmm7
> + rorl $9,%r14d
> + xorl %r8d,%r13d
> + xorl %r10d,%r12d
> + rorl $5,%r13d
> + xorl %eax,%r14d
> +.byte 102,15,58,15,226,4
> + andl %r8d,%r12d
> + xorl %r8d,%r13d
> +.byte 102,15,58,15,248,4
> + addl 32(%rsp),%r11d
> + movl %eax,%r15d
> + xorl %r10d,%r12d
> + rorl $11,%r14d
> + movdqa %xmm4,%xmm5
> + xorl %ebx,%r15d
> + addl %r12d,%r11d
> + movdqa %xmm4,%xmm6
> + rorl $6,%r13d
> + andl %r15d,%edi
> + psrld $3,%xmm4
> + xorl %eax,%r14d
> + addl %r13d,%r11d
> + xorl %ebx,%edi
> + paddd %xmm7,%xmm2
> + rorl $2,%r14d
> + addl %r11d,%edx
> + psrld $7,%xmm6
> + addl %edi,%r11d
> + movl %edx,%r13d
> + pshufd $250,%xmm1,%xmm7
> + addl %r11d,%r14d
> + rorl $14,%r13d
> + pslld $14,%xmm5
> + movl %r14d,%r11d
> + movl %r8d,%r12d
> + pxor %xmm6,%xmm4
> + rorl $9,%r14d
> + xorl %edx,%r13d
> + xorl %r9d,%r12d
> + rorl $5,%r13d
> + psrld $11,%xmm6
> + xorl %r11d,%r14d
> + pxor %xmm5,%xmm4
> + andl %edx,%r12d
> + xorl %edx,%r13d
> + pslld $11,%xmm5
> + addl 36(%rsp),%r10d
> + movl %r11d,%edi
> + pxor %xmm6,%xmm4
> + xorl %r9d,%r12d
> + rorl $11,%r14d
> + movdqa %xmm7,%xmm6
> + xorl %eax,%edi
> + addl %r12d,%r10d
> + pxor %xmm5,%xmm4
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %r11d,%r14d
> + psrld $10,%xmm7
> + addl %r13d,%r10d
> + xorl %eax,%r15d
> + paddd %xmm4,%xmm2
> + rorl $2,%r14d
> + addl %r10d,%ecx
> + psrlq $17,%xmm6
> + addl %r15d,%r10d
> + movl %ecx,%r13d
> + addl %r10d,%r14d
> + pxor %xmm6,%xmm7
> + rorl $14,%r13d
> + movl %r14d,%r10d
> + movl %edx,%r12d
> + rorl $9,%r14d
> + psrlq $2,%xmm6
> + xorl %ecx,%r13d
> + xorl %r8d,%r12d
> + pxor %xmm6,%xmm7
> + rorl $5,%r13d
> + xorl %r10d,%r14d
> + andl %ecx,%r12d
> + pshufd $128,%xmm7,%xmm7
> + xorl %ecx,%r13d
> + addl 40(%rsp),%r9d
> + movl %r10d,%r15d
> + psrldq $8,%xmm7
> + xorl %r8d,%r12d
> + rorl $11,%r14d
> + xorl %r11d,%r15d
> + addl %r12d,%r9d
> + rorl $6,%r13d
> + paddd %xmm7,%xmm2
> + andl %r15d,%edi
> + xorl %r10d,%r14d
> + addl %r13d,%r9d
> + pshufd $80,%xmm2,%xmm7
> + xorl %r11d,%edi
> + rorl $2,%r14d
> + addl %r9d,%ebx
> + movdqa %xmm7,%xmm6
> + addl %edi,%r9d
> + movl %ebx,%r13d
> + psrld $10,%xmm7
> + addl %r9d,%r14d
> + rorl $14,%r13d
> + psrlq $17,%xmm6
> + movl %r14d,%r9d
> + movl %ecx,%r12d
> + pxor %xmm6,%xmm7
> + rorl $9,%r14d
> + xorl %ebx,%r13d
> + xorl %edx,%r12d
> + rorl $5,%r13d
> + xorl %r9d,%r14d
> + psrlq $2,%xmm6
> + andl %ebx,%r12d
> + xorl %ebx,%r13d
> + addl 44(%rsp),%r8d
> + pxor %xmm6,%xmm7
> + movl %r9d,%edi
> + xorl %edx,%r12d
> + rorl $11,%r14d
> + pshufd $8,%xmm7,%xmm7
> + xorl %r10d,%edi
> + addl %r12d,%r8d
> + movdqa 64(%rbp),%xmm6
> + rorl $6,%r13d
> + andl %edi,%r15d
> + pslldq $8,%xmm7
> + xorl %r9d,%r14d
> + addl %r13d,%r8d
> + xorl %r10d,%r15d
> + paddd %xmm7,%xmm2
> + rorl $2,%r14d
> + addl %r8d,%eax
> + addl %r15d,%r8d
> + paddd %xmm2,%xmm6
> + movl %eax,%r13d
> + addl %r8d,%r14d
> + movdqa %xmm6,32(%rsp)
> + rorl $14,%r13d
> + movdqa %xmm0,%xmm4
> + movl %r14d,%r8d
> + movl %ebx,%r12d
> + movdqa %xmm2,%xmm7
> + rorl $9,%r14d
> + xorl %eax,%r13d
> + xorl %ecx,%r12d
> + rorl $5,%r13d
> + xorl %r8d,%r14d
> +.byte 102,15,58,15,227,4
> + andl %eax,%r12d
> + xorl %eax,%r13d
> +.byte 102,15,58,15,249,4
> + addl 48(%rsp),%edx
> + movl %r8d,%r15d
> + xorl %ecx,%r12d
> + rorl $11,%r14d
> + movdqa %xmm4,%xmm5
> + xorl %r9d,%r15d
> + addl %r12d,%edx
> + movdqa %xmm4,%xmm6
> + rorl $6,%r13d
> + andl %r15d,%edi
> + psrld $3,%xmm4
> + xorl %r8d,%r14d
> + addl %r13d,%edx
> + xorl %r9d,%edi
> + paddd %xmm7,%xmm3
> + rorl $2,%r14d
> + addl %edx,%r11d
> + psrld $7,%xmm6
> + addl %edi,%edx
> + movl %r11d,%r13d
> + pshufd $250,%xmm2,%xmm7
> + addl %edx,%r14d
> + rorl $14,%r13d
> + pslld $14,%xmm5
> + movl %r14d,%edx
> + movl %eax,%r12d
> + pxor %xmm6,%xmm4
> + rorl $9,%r14d
> + xorl %r11d,%r13d
> + xorl %ebx,%r12d
> + rorl $5,%r13d
> + psrld $11,%xmm6
> + xorl %edx,%r14d
> + pxor %xmm5,%xmm4
> + andl %r11d,%r12d
> + xorl %r11d,%r13d
> + pslld $11,%xmm5
> + addl 52(%rsp),%ecx
> + movl %edx,%edi
> + pxor %xmm6,%xmm4
> + xorl %ebx,%r12d
> + rorl $11,%r14d
> + movdqa %xmm7,%xmm6
> + xorl %r8d,%edi
> + addl %r12d,%ecx
> + pxor %xmm5,%xmm4
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %edx,%r14d
> + psrld $10,%xmm7
> + addl %r13d,%ecx
> + xorl %r8d,%r15d
> + paddd %xmm4,%xmm3
> + rorl $2,%r14d
> + addl %ecx,%r10d
> + psrlq $17,%xmm6
> + addl %r15d,%ecx
> + movl %r10d,%r13d
> + addl %ecx,%r14d
> + pxor %xmm6,%xmm7
> + rorl $14,%r13d
> + movl %r14d,%ecx
> + movl %r11d,%r12d
> + rorl $9,%r14d
> + psrlq $2,%xmm6
> + xorl %r10d,%r13d
> + xorl %eax,%r12d
> + pxor %xmm6,%xmm7
> + rorl $5,%r13d
> + xorl %ecx,%r14d
> + andl %r10d,%r12d
> + pshufd $128,%xmm7,%xmm7
> + xorl %r10d,%r13d
> + addl 56(%rsp),%ebx
> + movl %ecx,%r15d
> + psrldq $8,%xmm7
> + xorl %eax,%r12d
> + rorl $11,%r14d
> + xorl %edx,%r15d
> + addl %r12d,%ebx
> + rorl $6,%r13d
> + paddd %xmm7,%xmm3
> + andl %r15d,%edi
> + xorl %ecx,%r14d
> + addl %r13d,%ebx
> + pshufd $80,%xmm3,%xmm7
> + xorl %edx,%edi
> + rorl $2,%r14d
> + addl %ebx,%r9d
> + movdqa %xmm7,%xmm6
> + addl %edi,%ebx
> + movl %r9d,%r13d
> + psrld $10,%xmm7
> + addl %ebx,%r14d
> + rorl $14,%r13d
> + psrlq $17,%xmm6
> + movl %r14d,%ebx
> + movl %r10d,%r12d
> + pxor %xmm6,%xmm7
> + rorl $9,%r14d
> + xorl %r9d,%r13d
> + xorl %r11d,%r12d
> + rorl $5,%r13d
> + xorl %ebx,%r14d
> + psrlq $2,%xmm6
> + andl %r9d,%r12d
> + xorl %r9d,%r13d
> + addl 60(%rsp),%eax
> + pxor %xmm6,%xmm7
> + movl %ebx,%edi
> + xorl %r11d,%r12d
> + rorl $11,%r14d
> + pshufd $8,%xmm7,%xmm7
> + xorl %ecx,%edi
> + addl %r12d,%eax
> + movdqa 96(%rbp),%xmm6
> + rorl $6,%r13d
> + andl %edi,%r15d
> + pslldq $8,%xmm7
> + xorl %ebx,%r14d
> + addl %r13d,%eax
> + xorl %ecx,%r15d
> + paddd %xmm7,%xmm3
> + rorl $2,%r14d
> + addl %eax,%r8d
> + addl %r15d,%eax
> + paddd %xmm3,%xmm6
> + movl %r8d,%r13d
> + addl %eax,%r14d
> + movdqa %xmm6,48(%rsp)
> + cmpb $0,131(%rbp)
> + jne .Lssse3_00_47
> + rorl $14,%r13d
> + movl %r14d,%eax
> + movl %r9d,%r12d
> + rorl $9,%r14d
> + xorl %r8d,%r13d
> + xorl %r10d,%r12d
> + rorl $5,%r13d
> + xorl %eax,%r14d
> + andl %r8d,%r12d
> + xorl %r8d,%r13d
> + addl 0(%rsp),%r11d
> + movl %eax,%r15d
> + xorl %r10d,%r12d
> + rorl $11,%r14d
> + xorl %ebx,%r15d
> + addl %r12d,%r11d
> + rorl $6,%r13d
> + andl %r15d,%edi
> + xorl %eax,%r14d
> + addl %r13d,%r11d
> + xorl %ebx,%edi
> + rorl $2,%r14d
> + addl %r11d,%edx
> + addl %edi,%r11d
> + movl %edx,%r13d
> + addl %r11d,%r14d
> + rorl $14,%r13d
> + movl %r14d,%r11d
> + movl %r8d,%r12d
> + rorl $9,%r14d
> + xorl %edx,%r13d
> + xorl %r9d,%r12d
> + rorl $5,%r13d
> + xorl %r11d,%r14d
> + andl %edx,%r12d
> + xorl %edx,%r13d
> + addl 4(%rsp),%r10d
> + movl %r11d,%edi
> + xorl %r9d,%r12d
> + rorl $11,%r14d
> + xorl %eax,%edi
> + addl %r12d,%r10d
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %r11d,%r14d
> + addl %r13d,%r10d
> + xorl %eax,%r15d
> + rorl $2,%r14d
> + addl %r10d,%ecx
> + addl %r15d,%r10d
> + movl %ecx,%r13d
> + addl %r10d,%r14d
> + rorl $14,%r13d
> + movl %r14d,%r10d
> + movl %edx,%r12d
> + rorl $9,%r14d
> + xorl %ecx,%r13d
> + xorl %r8d,%r12d
> + rorl $5,%r13d
> + xorl %r10d,%r14d
> + andl %ecx,%r12d
> + xorl %ecx,%r13d
> + addl 8(%rsp),%r9d
> + movl %r10d,%r15d
> + xorl %r8d,%r12d
> + rorl $11,%r14d
> + xorl %r11d,%r15d
> + addl %r12d,%r9d
> + rorl $6,%r13d
> + andl %r15d,%edi
> + xorl %r10d,%r14d
> + addl %r13d,%r9d
> + xorl %r11d,%edi
> + rorl $2,%r14d
> + addl %r9d,%ebx
> + addl %edi,%r9d
> + movl %ebx,%r13d
> + addl %r9d,%r14d
> + rorl $14,%r13d
> + movl %r14d,%r9d
> + movl %ecx,%r12d
> + rorl $9,%r14d
> + xorl %ebx,%r13d
> + xorl %edx,%r12d
> + rorl $5,%r13d
> + xorl %r9d,%r14d
> + andl %ebx,%r12d
> + xorl %ebx,%r13d
> + addl 12(%rsp),%r8d
> + movl %r9d,%edi
> + xorl %edx,%r12d
> + rorl $11,%r14d
> + xorl %r10d,%edi
> + addl %r12d,%r8d
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %r9d,%r14d
> + addl %r13d,%r8d
> + xorl %r10d,%r15d
> + rorl $2,%r14d
> + addl %r8d,%eax
> + addl %r15d,%r8d
> + movl %eax,%r13d
> + addl %r8d,%r14d
> + rorl $14,%r13d
> + movl %r14d,%r8d
> + movl %ebx,%r12d
> + rorl $9,%r14d
> + xorl %eax,%r13d
> + xorl %ecx,%r12d
> + rorl $5,%r13d
> + xorl %r8d,%r14d
> + andl %eax,%r12d
> + xorl %eax,%r13d
> + addl 16(%rsp),%edx
> + movl %r8d,%r15d
> + xorl %ecx,%r12d
> + rorl $11,%r14d
> + xorl %r9d,%r15d
> + addl %r12d,%edx
> + rorl $6,%r13d
> + andl %r15d,%edi
> + xorl %r8d,%r14d
> + addl %r13d,%edx
> + xorl %r9d,%edi
> + rorl $2,%r14d
> + addl %edx,%r11d
> + addl %edi,%edx
> + movl %r11d,%r13d
> + addl %edx,%r14d
> + rorl $14,%r13d
> + movl %r14d,%edx
> + movl %eax,%r12d
> + rorl $9,%r14d
> + xorl %r11d,%r13d
> + xorl %ebx,%r12d
> + rorl $5,%r13d
> + xorl %edx,%r14d
> + andl %r11d,%r12d
> + xorl %r11d,%r13d
> + addl 20(%rsp),%ecx
> + movl %edx,%edi
> + xorl %ebx,%r12d
> + rorl $11,%r14d
> + xorl %r8d,%edi
> + addl %r12d,%ecx
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %edx,%r14d
> + addl %r13d,%ecx
> + xorl %r8d,%r15d
> + rorl $2,%r14d
> + addl %ecx,%r10d
> + addl %r15d,%ecx
> + movl %r10d,%r13d
> + addl %ecx,%r14d
> + rorl $14,%r13d
> + movl %r14d,%ecx
> + movl %r11d,%r12d
> + rorl $9,%r14d
> + xorl %r10d,%r13d
> + xorl %eax,%r12d
> + rorl $5,%r13d
> + xorl %ecx,%r14d
> + andl %r10d,%r12d
> + xorl %r10d,%r13d
> + addl 24(%rsp),%ebx
> + movl %ecx,%r15d
> + xorl %eax,%r12d
> + rorl $11,%r14d
> + xorl %edx,%r15d
> + addl %r12d,%ebx
> + rorl $6,%r13d
> + andl %r15d,%edi
> + xorl %ecx,%r14d
> + addl %r13d,%ebx
> + xorl %edx,%edi
> + rorl $2,%r14d
> + addl %ebx,%r9d
> + addl %edi,%ebx
> + movl %r9d,%r13d
> + addl %ebx,%r14d
> + rorl $14,%r13d
> + movl %r14d,%ebx
> + movl %r10d,%r12d
> + rorl $9,%r14d
> + xorl %r9d,%r13d
> + xorl %r11d,%r12d
> + rorl $5,%r13d
> + xorl %ebx,%r14d
> + andl %r9d,%r12d
> + xorl %r9d,%r13d
> + addl 28(%rsp),%eax
> + movl %ebx,%edi
> + xorl %r11d,%r12d
> + rorl $11,%r14d
> + xorl %ecx,%edi
> + addl %r12d,%eax
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %ebx,%r14d
> + addl %r13d,%eax
> + xorl %ecx,%r15d
> + rorl $2,%r14d
> + addl %eax,%r8d
> + addl %r15d,%eax
> + movl %r8d,%r13d
> + addl %eax,%r14d
> + rorl $14,%r13d
> + movl %r14d,%eax
> + movl %r9d,%r12d
> + rorl $9,%r14d
> + xorl %r8d,%r13d
> + xorl %r10d,%r12d
> + rorl $5,%r13d
> + xorl %eax,%r14d
> + andl %r8d,%r12d
> + xorl %r8d,%r13d
> + addl 32(%rsp),%r11d
> + movl %eax,%r15d
> + xorl %r10d,%r12d
> + rorl $11,%r14d
> + xorl %ebx,%r15d
> + addl %r12d,%r11d
> + rorl $6,%r13d
> + andl %r15d,%edi
> + xorl %eax,%r14d
> + addl %r13d,%r11d
> + xorl %ebx,%edi
> + rorl $2,%r14d
> + addl %r11d,%edx
> + addl %edi,%r11d
> + movl %edx,%r13d
> + addl %r11d,%r14d
> + rorl $14,%r13d
> + movl %r14d,%r11d
> + movl %r8d,%r12d
> + rorl $9,%r14d
> + xorl %edx,%r13d
> + xorl %r9d,%r12d
> + rorl $5,%r13d
> + xorl %r11d,%r14d
> + andl %edx,%r12d
> + xorl %edx,%r13d
> + addl 36(%rsp),%r10d
> + movl %r11d,%edi
> + xorl %r9d,%r12d
> + rorl $11,%r14d
> + xorl %eax,%edi
> + addl %r12d,%r10d
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %r11d,%r14d
> + addl %r13d,%r10d
> + xorl %eax,%r15d
> + rorl $2,%r14d
> + addl %r10d,%ecx
> + addl %r15d,%r10d
> + movl %ecx,%r13d
> + addl %r10d,%r14d
> + rorl $14,%r13d
> + movl %r14d,%r10d
> + movl %edx,%r12d
> + rorl $9,%r14d
> + xorl %ecx,%r13d
> + xorl %r8d,%r12d
> + rorl $5,%r13d
> + xorl %r10d,%r14d
> + andl %ecx,%r12d
> + xorl %ecx,%r13d
> + addl 40(%rsp),%r9d
> + movl %r10d,%r15d
> + xorl %r8d,%r12d
> + rorl $11,%r14d
> + xorl %r11d,%r15d
> + addl %r12d,%r9d
> + rorl $6,%r13d
> + andl %r15d,%edi
> + xorl %r10d,%r14d
> + addl %r13d,%r9d
> + xorl %r11d,%edi
> + rorl $2,%r14d
> + addl %r9d,%ebx
> + addl %edi,%r9d
> + movl %ebx,%r13d
> + addl %r9d,%r14d
> + rorl $14,%r13d
> + movl %r14d,%r9d
> + movl %ecx,%r12d
> + rorl $9,%r14d
> + xorl %ebx,%r13d
> + xorl %edx,%r12d
> + rorl $5,%r13d
> + xorl %r9d,%r14d
> + andl %ebx,%r12d
> + xorl %ebx,%r13d
> + addl 44(%rsp),%r8d
> + movl %r9d,%edi
> + xorl %edx,%r12d
> + rorl $11,%r14d
> + xorl %r10d,%edi
> + addl %r12d,%r8d
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %r9d,%r14d
> + addl %r13d,%r8d
> + xorl %r10d,%r15d
> + rorl $2,%r14d
> + addl %r8d,%eax
> + addl %r15d,%r8d
> + movl %eax,%r13d
> + addl %r8d,%r14d
> + rorl $14,%r13d
> + movl %r14d,%r8d
> + movl %ebx,%r12d
> + rorl $9,%r14d
> + xorl %eax,%r13d
> + xorl %ecx,%r12d
> + rorl $5,%r13d
> + xorl %r8d,%r14d
> + andl %eax,%r12d
> + xorl %eax,%r13d
> + addl 48(%rsp),%edx
> + movl %r8d,%r15d
> + xorl %ecx,%r12d
> + rorl $11,%r14d
> + xorl %r9d,%r15d
> + addl %r12d,%edx
> + rorl $6,%r13d
> + andl %r15d,%edi
> + xorl %r8d,%r14d
> + addl %r13d,%edx
> + xorl %r9d,%edi
> + rorl $2,%r14d
> + addl %edx,%r11d
> + addl %edi,%edx
> + movl %r11d,%r13d
> + addl %edx,%r14d
> + rorl $14,%r13d
> + movl %r14d,%edx
> + movl %eax,%r12d
> + rorl $9,%r14d
> + xorl %r11d,%r13d
> + xorl %ebx,%r12d
> + rorl $5,%r13d
> + xorl %edx,%r14d
> + andl %r11d,%r12d
> + xorl %r11d,%r13d
> + addl 52(%rsp),%ecx
> + movl %edx,%edi
> + xorl %ebx,%r12d
> + rorl $11,%r14d
> + xorl %r8d,%edi
> + addl %r12d,%ecx
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %edx,%r14d
> + addl %r13d,%ecx
> + xorl %r8d,%r15d
> + rorl $2,%r14d
> + addl %ecx,%r10d
> + addl %r15d,%ecx
> + movl %r10d,%r13d
> + addl %ecx,%r14d
> + rorl $14,%r13d
> + movl %r14d,%ecx
> + movl %r11d,%r12d
> + rorl $9,%r14d
> + xorl %r10d,%r13d
> + xorl %eax,%r12d
> + rorl $5,%r13d
> + xorl %ecx,%r14d
> + andl %r10d,%r12d
> + xorl %r10d,%r13d
> + addl 56(%rsp),%ebx
> + movl %ecx,%r15d
> + xorl %eax,%r12d
> + rorl $11,%r14d
> + xorl %edx,%r15d
> + addl %r12d,%ebx
> + rorl $6,%r13d
> + andl %r15d,%edi
> + xorl %ecx,%r14d
> + addl %r13d,%ebx
> + xorl %edx,%edi
> + rorl $2,%r14d
> + addl %ebx,%r9d
> + addl %edi,%ebx
> + movl %r9d,%r13d
> + addl %ebx,%r14d
> + rorl $14,%r13d
> + movl %r14d,%ebx
> + movl %r10d,%r12d
> + rorl $9,%r14d
> + xorl %r9d,%r13d
> + xorl %r11d,%r12d
> + rorl $5,%r13d
> + xorl %ebx,%r14d
> + andl %r9d,%r12d
> + xorl %r9d,%r13d
> + addl 60(%rsp),%eax
> + movl %ebx,%edi
> + xorl %r11d,%r12d
> + rorl $11,%r14d
> + xorl %ecx,%edi
> + addl %r12d,%eax
> + rorl $6,%r13d
> + andl %edi,%r15d
> + xorl %ebx,%r14d
> + addl %r13d,%eax
> + xorl %ecx,%r15d
> + rorl $2,%r14d
> + addl %eax,%r8d
> + addl %r15d,%eax
> + movl %r8d,%r13d
> + addl %eax,%r14d
> + movq 64+0(%rsp),%rdi
> + movl %r14d,%eax
> +
> + addl 0(%rdi),%eax
> + leaq 64(%rsi),%rsi
> + addl 4(%rdi),%ebx
> + addl 8(%rdi),%ecx
> + addl 12(%rdi),%edx
> + addl 16(%rdi),%r8d
> + addl 20(%rdi),%r9d
> + addl 24(%rdi),%r10d
> + addl 28(%rdi),%r11d
> +
> + cmpq 64+16(%rsp),%rsi
> +
> + movl %eax,0(%rdi)
> + movl %ebx,4(%rdi)
> + movl %ecx,8(%rdi)
> + movl %edx,12(%rdi)
> + movl %r8d,16(%rdi)
> + movl %r9d,20(%rdi)
> + movl %r10d,24(%rdi)
> + movl %r11d,28(%rdi)
> + jb .Lloop_ssse3
> +
> + movq 88(%rsp),%rsi
> +.cfi_def_cfa %rsi,8
> + movq -48(%rsi),%r15
> +.cfi_restore %r15
> + movq -40(%rsi),%r14
> +.cfi_restore %r14
> + movq -32(%rsi),%r13
> +.cfi_restore %r13
> + movq -24(%rsi),%r12
> +.cfi_restore %r12
> + movq -16(%rsi),%rbp
> +.cfi_restore %rbp
> + movq -8(%rsi),%rbx
> +.cfi_restore %rbx
> + leaq (%rsi),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue_ssse3:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha256_block_data_order_ssse3,.-sha256_block_data_order_ssse3
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha512-x86_64.S
> b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha512-x86_64.S
> new file mode 100644
> index 0000000000..11e67e5ba1
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/sha/sha512-x86_64.S
> @@ -0,0 +1,1811 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/sha/asm/sha512-x86_64.pl
> +#
> +# Copyright 2005-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +.text
> +
> +
> +.globl sha512_block_data_order
> +.type sha512_block_data_order, at function
> +.align 16
> +sha512_block_data_order:
> +.cfi_startproc
> + movq %rsp,%rax
> +.cfi_def_cfa_register %rax
> + pushq %rbx
> +.cfi_offset %rbx,-16
> + pushq %rbp
> +.cfi_offset %rbp,-24
> + pushq %r12
> +.cfi_offset %r12,-32
> + pushq %r13
> +.cfi_offset %r13,-40
> + pushq %r14
> +.cfi_offset %r14,-48
> + pushq %r15
> +.cfi_offset %r15,-56
> + shlq $4,%rdx
> + subq $128+32,%rsp
> + leaq (%rsi,%rdx,8),%rdx
> + andq $-64,%rsp
> + movq %rdi,128+0(%rsp)
> + movq %rsi,128+8(%rsp)
> + movq %rdx,128+16(%rsp)
> + movq %rax,152(%rsp)
> +.cfi_escape 0x0f,0x06,0x77,0x98,0x01,0x06,0x23,0x08
> +.Lprologue:
> +
> + movq 0(%rdi),%rax
> + movq 8(%rdi),%rbx
> + movq 16(%rdi),%rcx
> + movq 24(%rdi),%rdx
> + movq 32(%rdi),%r8
> + movq 40(%rdi),%r9
> + movq 48(%rdi),%r10
> + movq 56(%rdi),%r11
> + jmp .Lloop
> +
> +.align 16
> +.Lloop:
> + movq %rbx,%rdi
> + leaq K512(%rip),%rbp
> + xorq %rcx,%rdi
> + movq 0(%rsi),%r12
> + movq %r8,%r13
> + movq %rax,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %r9,%r15
> +
> + xorq %r8,%r13
> + rorq $5,%r14
> + xorq %r10,%r15
> +
> + movq %r12,0(%rsp)
> + xorq %rax,%r14
> + andq %r8,%r15
> +
> + rorq $4,%r13
> + addq %r11,%r12
> + xorq %r10,%r15
> +
> + rorq $6,%r14
> + xorq %r8,%r13
> + addq %r15,%r12
> +
> + movq %rax,%r15
> + addq (%rbp),%r12
> + xorq %rax,%r14
> +
> + xorq %rbx,%r15
> + rorq $14,%r13
> + movq %rbx,%r11
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%r11
> + addq %r12,%rdx
> + addq %r12,%r11
> +
> + leaq 8(%rbp),%rbp
> + addq %r14,%r11
> + movq 8(%rsi),%r12
> + movq %rdx,%r13
> + movq %r11,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %r8,%rdi
> +
> + xorq %rdx,%r13
> + rorq $5,%r14
> + xorq %r9,%rdi
> +
> + movq %r12,8(%rsp)
> + xorq %r11,%r14
> + andq %rdx,%rdi
> +
> + rorq $4,%r13
> + addq %r10,%r12
> + xorq %r9,%rdi
> +
> + rorq $6,%r14
> + xorq %rdx,%r13
> + addq %rdi,%r12
> +
> + movq %r11,%rdi
> + addq (%rbp),%r12
> + xorq %r11,%r14
> +
> + xorq %rax,%rdi
> + rorq $14,%r13
> + movq %rax,%r10
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%r10
> + addq %r12,%rcx
> + addq %r12,%r10
> +
> + leaq 24(%rbp),%rbp
> + addq %r14,%r10
> + movq 16(%rsi),%r12
> + movq %rcx,%r13
> + movq %r10,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %rdx,%r15
> +
> + xorq %rcx,%r13
> + rorq $5,%r14
> + xorq %r8,%r15
> +
> + movq %r12,16(%rsp)
> + xorq %r10,%r14
> + andq %rcx,%r15
> +
> + rorq $4,%r13
> + addq %r9,%r12
> + xorq %r8,%r15
> +
> + rorq $6,%r14
> + xorq %rcx,%r13
> + addq %r15,%r12
> +
> + movq %r10,%r15
> + addq (%rbp),%r12
> + xorq %r10,%r14
> +
> + xorq %r11,%r15
> + rorq $14,%r13
> + movq %r11,%r9
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%r9
> + addq %r12,%rbx
> + addq %r12,%r9
> +
> + leaq 8(%rbp),%rbp
> + addq %r14,%r9
> + movq 24(%rsi),%r12
> + movq %rbx,%r13
> + movq %r9,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %rcx,%rdi
> +
> + xorq %rbx,%r13
> + rorq $5,%r14
> + xorq %rdx,%rdi
> +
> + movq %r12,24(%rsp)
> + xorq %r9,%r14
> + andq %rbx,%rdi
> +
> + rorq $4,%r13
> + addq %r8,%r12
> + xorq %rdx,%rdi
> +
> + rorq $6,%r14
> + xorq %rbx,%r13
> + addq %rdi,%r12
> +
> + movq %r9,%rdi
> + addq (%rbp),%r12
> + xorq %r9,%r14
> +
> + xorq %r10,%rdi
> + rorq $14,%r13
> + movq %r10,%r8
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%r8
> + addq %r12,%rax
> + addq %r12,%r8
> +
> + leaq 24(%rbp),%rbp
> + addq %r14,%r8
> + movq 32(%rsi),%r12
> + movq %rax,%r13
> + movq %r8,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %rbx,%r15
> +
> + xorq %rax,%r13
> + rorq $5,%r14
> + xorq %rcx,%r15
> +
> + movq %r12,32(%rsp)
> + xorq %r8,%r14
> + andq %rax,%r15
> +
> + rorq $4,%r13
> + addq %rdx,%r12
> + xorq %rcx,%r15
> +
> + rorq $6,%r14
> + xorq %rax,%r13
> + addq %r15,%r12
> +
> + movq %r8,%r15
> + addq (%rbp),%r12
> + xorq %r8,%r14
> +
> + xorq %r9,%r15
> + rorq $14,%r13
> + movq %r9,%rdx
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%rdx
> + addq %r12,%r11
> + addq %r12,%rdx
> +
> + leaq 8(%rbp),%rbp
> + addq %r14,%rdx
> + movq 40(%rsi),%r12
> + movq %r11,%r13
> + movq %rdx,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %rax,%rdi
> +
> + xorq %r11,%r13
> + rorq $5,%r14
> + xorq %rbx,%rdi
> +
> + movq %r12,40(%rsp)
> + xorq %rdx,%r14
> + andq %r11,%rdi
> +
> + rorq $4,%r13
> + addq %rcx,%r12
> + xorq %rbx,%rdi
> +
> + rorq $6,%r14
> + xorq %r11,%r13
> + addq %rdi,%r12
> +
> + movq %rdx,%rdi
> + addq (%rbp),%r12
> + xorq %rdx,%r14
> +
> + xorq %r8,%rdi
> + rorq $14,%r13
> + movq %r8,%rcx
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%rcx
> + addq %r12,%r10
> + addq %r12,%rcx
> +
> + leaq 24(%rbp),%rbp
> + addq %r14,%rcx
> + movq 48(%rsi),%r12
> + movq %r10,%r13
> + movq %rcx,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %r11,%r15
> +
> + xorq %r10,%r13
> + rorq $5,%r14
> + xorq %rax,%r15
> +
> + movq %r12,48(%rsp)
> + xorq %rcx,%r14
> + andq %r10,%r15
> +
> + rorq $4,%r13
> + addq %rbx,%r12
> + xorq %rax,%r15
> +
> + rorq $6,%r14
> + xorq %r10,%r13
> + addq %r15,%r12
> +
> + movq %rcx,%r15
> + addq (%rbp),%r12
> + xorq %rcx,%r14
> +
> + xorq %rdx,%r15
> + rorq $14,%r13
> + movq %rdx,%rbx
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%rbx
> + addq %r12,%r9
> + addq %r12,%rbx
> +
> + leaq 8(%rbp),%rbp
> + addq %r14,%rbx
> + movq 56(%rsi),%r12
> + movq %r9,%r13
> + movq %rbx,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %r10,%rdi
> +
> + xorq %r9,%r13
> + rorq $5,%r14
> + xorq %r11,%rdi
> +
> + movq %r12,56(%rsp)
> + xorq %rbx,%r14
> + andq %r9,%rdi
> +
> + rorq $4,%r13
> + addq %rax,%r12
> + xorq %r11,%rdi
> +
> + rorq $6,%r14
> + xorq %r9,%r13
> + addq %rdi,%r12
> +
> + movq %rbx,%rdi
> + addq (%rbp),%r12
> + xorq %rbx,%r14
> +
> + xorq %rcx,%rdi
> + rorq $14,%r13
> + movq %rcx,%rax
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%rax
> + addq %r12,%r8
> + addq %r12,%rax
> +
> + leaq 24(%rbp),%rbp
> + addq %r14,%rax
> + movq 64(%rsi),%r12
> + movq %r8,%r13
> + movq %rax,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %r9,%r15
> +
> + xorq %r8,%r13
> + rorq $5,%r14
> + xorq %r10,%r15
> +
> + movq %r12,64(%rsp)
> + xorq %rax,%r14
> + andq %r8,%r15
> +
> + rorq $4,%r13
> + addq %r11,%r12
> + xorq %r10,%r15
> +
> + rorq $6,%r14
> + xorq %r8,%r13
> + addq %r15,%r12
> +
> + movq %rax,%r15
> + addq (%rbp),%r12
> + xorq %rax,%r14
> +
> + xorq %rbx,%r15
> + rorq $14,%r13
> + movq %rbx,%r11
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%r11
> + addq %r12,%rdx
> + addq %r12,%r11
> +
> + leaq 8(%rbp),%rbp
> + addq %r14,%r11
> + movq 72(%rsi),%r12
> + movq %rdx,%r13
> + movq %r11,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %r8,%rdi
> +
> + xorq %rdx,%r13
> + rorq $5,%r14
> + xorq %r9,%rdi
> +
> + movq %r12,72(%rsp)
> + xorq %r11,%r14
> + andq %rdx,%rdi
> +
> + rorq $4,%r13
> + addq %r10,%r12
> + xorq %r9,%rdi
> +
> + rorq $6,%r14
> + xorq %rdx,%r13
> + addq %rdi,%r12
> +
> + movq %r11,%rdi
> + addq (%rbp),%r12
> + xorq %r11,%r14
> +
> + xorq %rax,%rdi
> + rorq $14,%r13
> + movq %rax,%r10
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%r10
> + addq %r12,%rcx
> + addq %r12,%r10
> +
> + leaq 24(%rbp),%rbp
> + addq %r14,%r10
> + movq 80(%rsi),%r12
> + movq %rcx,%r13
> + movq %r10,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %rdx,%r15
> +
> + xorq %rcx,%r13
> + rorq $5,%r14
> + xorq %r8,%r15
> +
> + movq %r12,80(%rsp)
> + xorq %r10,%r14
> + andq %rcx,%r15
> +
> + rorq $4,%r13
> + addq %r9,%r12
> + xorq %r8,%r15
> +
> + rorq $6,%r14
> + xorq %rcx,%r13
> + addq %r15,%r12
> +
> + movq %r10,%r15
> + addq (%rbp),%r12
> + xorq %r10,%r14
> +
> + xorq %r11,%r15
> + rorq $14,%r13
> + movq %r11,%r9
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%r9
> + addq %r12,%rbx
> + addq %r12,%r9
> +
> + leaq 8(%rbp),%rbp
> + addq %r14,%r9
> + movq 88(%rsi),%r12
> + movq %rbx,%r13
> + movq %r9,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %rcx,%rdi
> +
> + xorq %rbx,%r13
> + rorq $5,%r14
> + xorq %rdx,%rdi
> +
> + movq %r12,88(%rsp)
> + xorq %r9,%r14
> + andq %rbx,%rdi
> +
> + rorq $4,%r13
> + addq %r8,%r12
> + xorq %rdx,%rdi
> +
> + rorq $6,%r14
> + xorq %rbx,%r13
> + addq %rdi,%r12
> +
> + movq %r9,%rdi
> + addq (%rbp),%r12
> + xorq %r9,%r14
> +
> + xorq %r10,%rdi
> + rorq $14,%r13
> + movq %r10,%r8
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%r8
> + addq %r12,%rax
> + addq %r12,%r8
> +
> + leaq 24(%rbp),%rbp
> + addq %r14,%r8
> + movq 96(%rsi),%r12
> + movq %rax,%r13
> + movq %r8,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %rbx,%r15
> +
> + xorq %rax,%r13
> + rorq $5,%r14
> + xorq %rcx,%r15
> +
> + movq %r12,96(%rsp)
> + xorq %r8,%r14
> + andq %rax,%r15
> +
> + rorq $4,%r13
> + addq %rdx,%r12
> + xorq %rcx,%r15
> +
> + rorq $6,%r14
> + xorq %rax,%r13
> + addq %r15,%r12
> +
> + movq %r8,%r15
> + addq (%rbp),%r12
> + xorq %r8,%r14
> +
> + xorq %r9,%r15
> + rorq $14,%r13
> + movq %r9,%rdx
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%rdx
> + addq %r12,%r11
> + addq %r12,%rdx
> +
> + leaq 8(%rbp),%rbp
> + addq %r14,%rdx
> + movq 104(%rsi),%r12
> + movq %r11,%r13
> + movq %rdx,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %rax,%rdi
> +
> + xorq %r11,%r13
> + rorq $5,%r14
> + xorq %rbx,%rdi
> +
> + movq %r12,104(%rsp)
> + xorq %rdx,%r14
> + andq %r11,%rdi
> +
> + rorq $4,%r13
> + addq %rcx,%r12
> + xorq %rbx,%rdi
> +
> + rorq $6,%r14
> + xorq %r11,%r13
> + addq %rdi,%r12
> +
> + movq %rdx,%rdi
> + addq (%rbp),%r12
> + xorq %rdx,%r14
> +
> + xorq %r8,%rdi
> + rorq $14,%r13
> + movq %r8,%rcx
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%rcx
> + addq %r12,%r10
> + addq %r12,%rcx
> +
> + leaq 24(%rbp),%rbp
> + addq %r14,%rcx
> + movq 112(%rsi),%r12
> + movq %r10,%r13
> + movq %rcx,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %r11,%r15
> +
> + xorq %r10,%r13
> + rorq $5,%r14
> + xorq %rax,%r15
> +
> + movq %r12,112(%rsp)
> + xorq %rcx,%r14
> + andq %r10,%r15
> +
> + rorq $4,%r13
> + addq %rbx,%r12
> + xorq %rax,%r15
> +
> + rorq $6,%r14
> + xorq %r10,%r13
> + addq %r15,%r12
> +
> + movq %rcx,%r15
> + addq (%rbp),%r12
> + xorq %rcx,%r14
> +
> + xorq %rdx,%r15
> + rorq $14,%r13
> + movq %rdx,%rbx
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%rbx
> + addq %r12,%r9
> + addq %r12,%rbx
> +
> + leaq 8(%rbp),%rbp
> + addq %r14,%rbx
> + movq 120(%rsi),%r12
> + movq %r9,%r13
> + movq %rbx,%r14
> + bswapq %r12
> + rorq $23,%r13
> + movq %r10,%rdi
> +
> + xorq %r9,%r13
> + rorq $5,%r14
> + xorq %r11,%rdi
> +
> + movq %r12,120(%rsp)
> + xorq %rbx,%r14
> + andq %r9,%rdi
> +
> + rorq $4,%r13
> + addq %rax,%r12
> + xorq %r11,%rdi
> +
> + rorq $6,%r14
> + xorq %r9,%r13
> + addq %rdi,%r12
> +
> + movq %rbx,%rdi
> + addq (%rbp),%r12
> + xorq %rbx,%r14
> +
> + xorq %rcx,%rdi
> + rorq $14,%r13
> + movq %rcx,%rax
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%rax
> + addq %r12,%r8
> + addq %r12,%rax
> +
> + leaq 24(%rbp),%rbp
> + jmp .Lrounds_16_xx
> +.align 16
> +.Lrounds_16_xx:
> + movq 8(%rsp),%r13
> + movq 112(%rsp),%r15
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%rax
> + movq %r15,%r14
> + rorq $42,%r15
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%r15
> + shrq $6,%r14
> +
> + rorq $19,%r15
> + xorq %r13,%r12
> + xorq %r14,%r15
> + addq 72(%rsp),%r12
> +
> + addq 0(%rsp),%r12
> + movq %r8,%r13
> + addq %r15,%r12
> + movq %rax,%r14
> + rorq $23,%r13
> + movq %r9,%r15
> +
> + xorq %r8,%r13
> + rorq $5,%r14
> + xorq %r10,%r15
> +
> + movq %r12,0(%rsp)
> + xorq %rax,%r14
> + andq %r8,%r15
> +
> + rorq $4,%r13
> + addq %r11,%r12
> + xorq %r10,%r15
> +
> + rorq $6,%r14
> + xorq %r8,%r13
> + addq %r15,%r12
> +
> + movq %rax,%r15
> + addq (%rbp),%r12
> + xorq %rax,%r14
> +
> + xorq %rbx,%r15
> + rorq $14,%r13
> + movq %rbx,%r11
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%r11
> + addq %r12,%rdx
> + addq %r12,%r11
> +
> + leaq 8(%rbp),%rbp
> + movq 16(%rsp),%r13
> + movq 120(%rsp),%rdi
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%r11
> + movq %rdi,%r14
> + rorq $42,%rdi
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%rdi
> + shrq $6,%r14
> +
> + rorq $19,%rdi
> + xorq %r13,%r12
> + xorq %r14,%rdi
> + addq 80(%rsp),%r12
> +
> + addq 8(%rsp),%r12
> + movq %rdx,%r13
> + addq %rdi,%r12
> + movq %r11,%r14
> + rorq $23,%r13
> + movq %r8,%rdi
> +
> + xorq %rdx,%r13
> + rorq $5,%r14
> + xorq %r9,%rdi
> +
> + movq %r12,8(%rsp)
> + xorq %r11,%r14
> + andq %rdx,%rdi
> +
> + rorq $4,%r13
> + addq %r10,%r12
> + xorq %r9,%rdi
> +
> + rorq $6,%r14
> + xorq %rdx,%r13
> + addq %rdi,%r12
> +
> + movq %r11,%rdi
> + addq (%rbp),%r12
> + xorq %r11,%r14
> +
> + xorq %rax,%rdi
> + rorq $14,%r13
> + movq %rax,%r10
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%r10
> + addq %r12,%rcx
> + addq %r12,%r10
> +
> + leaq 24(%rbp),%rbp
> + movq 24(%rsp),%r13
> + movq 0(%rsp),%r15
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%r10
> + movq %r15,%r14
> + rorq $42,%r15
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%r15
> + shrq $6,%r14
> +
> + rorq $19,%r15
> + xorq %r13,%r12
> + xorq %r14,%r15
> + addq 88(%rsp),%r12
> +
> + addq 16(%rsp),%r12
> + movq %rcx,%r13
> + addq %r15,%r12
> + movq %r10,%r14
> + rorq $23,%r13
> + movq %rdx,%r15
> +
> + xorq %rcx,%r13
> + rorq $5,%r14
> + xorq %r8,%r15
> +
> + movq %r12,16(%rsp)
> + xorq %r10,%r14
> + andq %rcx,%r15
> +
> + rorq $4,%r13
> + addq %r9,%r12
> + xorq %r8,%r15
> +
> + rorq $6,%r14
> + xorq %rcx,%r13
> + addq %r15,%r12
> +
> + movq %r10,%r15
> + addq (%rbp),%r12
> + xorq %r10,%r14
> +
> + xorq %r11,%r15
> + rorq $14,%r13
> + movq %r11,%r9
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%r9
> + addq %r12,%rbx
> + addq %r12,%r9
> +
> + leaq 8(%rbp),%rbp
> + movq 32(%rsp),%r13
> + movq 8(%rsp),%rdi
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%r9
> + movq %rdi,%r14
> + rorq $42,%rdi
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%rdi
> + shrq $6,%r14
> +
> + rorq $19,%rdi
> + xorq %r13,%r12
> + xorq %r14,%rdi
> + addq 96(%rsp),%r12
> +
> + addq 24(%rsp),%r12
> + movq %rbx,%r13
> + addq %rdi,%r12
> + movq %r9,%r14
> + rorq $23,%r13
> + movq %rcx,%rdi
> +
> + xorq %rbx,%r13
> + rorq $5,%r14
> + xorq %rdx,%rdi
> +
> + movq %r12,24(%rsp)
> + xorq %r9,%r14
> + andq %rbx,%rdi
> +
> + rorq $4,%r13
> + addq %r8,%r12
> + xorq %rdx,%rdi
> +
> + rorq $6,%r14
> + xorq %rbx,%r13
> + addq %rdi,%r12
> +
> + movq %r9,%rdi
> + addq (%rbp),%r12
> + xorq %r9,%r14
> +
> + xorq %r10,%rdi
> + rorq $14,%r13
> + movq %r10,%r8
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%r8
> + addq %r12,%rax
> + addq %r12,%r8
> +
> + leaq 24(%rbp),%rbp
> + movq 40(%rsp),%r13
> + movq 16(%rsp),%r15
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%r8
> + movq %r15,%r14
> + rorq $42,%r15
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%r15
> + shrq $6,%r14
> +
> + rorq $19,%r15
> + xorq %r13,%r12
> + xorq %r14,%r15
> + addq 104(%rsp),%r12
> +
> + addq 32(%rsp),%r12
> + movq %rax,%r13
> + addq %r15,%r12
> + movq %r8,%r14
> + rorq $23,%r13
> + movq %rbx,%r15
> +
> + xorq %rax,%r13
> + rorq $5,%r14
> + xorq %rcx,%r15
> +
> + movq %r12,32(%rsp)
> + xorq %r8,%r14
> + andq %rax,%r15
> +
> + rorq $4,%r13
> + addq %rdx,%r12
> + xorq %rcx,%r15
> +
> + rorq $6,%r14
> + xorq %rax,%r13
> + addq %r15,%r12
> +
> + movq %r8,%r15
> + addq (%rbp),%r12
> + xorq %r8,%r14
> +
> + xorq %r9,%r15
> + rorq $14,%r13
> + movq %r9,%rdx
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%rdx
> + addq %r12,%r11
> + addq %r12,%rdx
> +
> + leaq 8(%rbp),%rbp
> + movq 48(%rsp),%r13
> + movq 24(%rsp),%rdi
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%rdx
> + movq %rdi,%r14
> + rorq $42,%rdi
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%rdi
> + shrq $6,%r14
> +
> + rorq $19,%rdi
> + xorq %r13,%r12
> + xorq %r14,%rdi
> + addq 112(%rsp),%r12
> +
> + addq 40(%rsp),%r12
> + movq %r11,%r13
> + addq %rdi,%r12
> + movq %rdx,%r14
> + rorq $23,%r13
> + movq %rax,%rdi
> +
> + xorq %r11,%r13
> + rorq $5,%r14
> + xorq %rbx,%rdi
> +
> + movq %r12,40(%rsp)
> + xorq %rdx,%r14
> + andq %r11,%rdi
> +
> + rorq $4,%r13
> + addq %rcx,%r12
> + xorq %rbx,%rdi
> +
> + rorq $6,%r14
> + xorq %r11,%r13
> + addq %rdi,%r12
> +
> + movq %rdx,%rdi
> + addq (%rbp),%r12
> + xorq %rdx,%r14
> +
> + xorq %r8,%rdi
> + rorq $14,%r13
> + movq %r8,%rcx
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%rcx
> + addq %r12,%r10
> + addq %r12,%rcx
> +
> + leaq 24(%rbp),%rbp
> + movq 56(%rsp),%r13
> + movq 32(%rsp),%r15
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%rcx
> + movq %r15,%r14
> + rorq $42,%r15
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%r15
> + shrq $6,%r14
> +
> + rorq $19,%r15
> + xorq %r13,%r12
> + xorq %r14,%r15
> + addq 120(%rsp),%r12
> +
> + addq 48(%rsp),%r12
> + movq %r10,%r13
> + addq %r15,%r12
> + movq %rcx,%r14
> + rorq $23,%r13
> + movq %r11,%r15
> +
> + xorq %r10,%r13
> + rorq $5,%r14
> + xorq %rax,%r15
> +
> + movq %r12,48(%rsp)
> + xorq %rcx,%r14
> + andq %r10,%r15
> +
> + rorq $4,%r13
> + addq %rbx,%r12
> + xorq %rax,%r15
> +
> + rorq $6,%r14
> + xorq %r10,%r13
> + addq %r15,%r12
> +
> + movq %rcx,%r15
> + addq (%rbp),%r12
> + xorq %rcx,%r14
> +
> + xorq %rdx,%r15
> + rorq $14,%r13
> + movq %rdx,%rbx
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%rbx
> + addq %r12,%r9
> + addq %r12,%rbx
> +
> + leaq 8(%rbp),%rbp
> + movq 64(%rsp),%r13
> + movq 40(%rsp),%rdi
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%rbx
> + movq %rdi,%r14
> + rorq $42,%rdi
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%rdi
> + shrq $6,%r14
> +
> + rorq $19,%rdi
> + xorq %r13,%r12
> + xorq %r14,%rdi
> + addq 0(%rsp),%r12
> +
> + addq 56(%rsp),%r12
> + movq %r9,%r13
> + addq %rdi,%r12
> + movq %rbx,%r14
> + rorq $23,%r13
> + movq %r10,%rdi
> +
> + xorq %r9,%r13
> + rorq $5,%r14
> + xorq %r11,%rdi
> +
> + movq %r12,56(%rsp)
> + xorq %rbx,%r14
> + andq %r9,%rdi
> +
> + rorq $4,%r13
> + addq %rax,%r12
> + xorq %r11,%rdi
> +
> + rorq $6,%r14
> + xorq %r9,%r13
> + addq %rdi,%r12
> +
> + movq %rbx,%rdi
> + addq (%rbp),%r12
> + xorq %rbx,%r14
> +
> + xorq %rcx,%rdi
> + rorq $14,%r13
> + movq %rcx,%rax
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%rax
> + addq %r12,%r8
> + addq %r12,%rax
> +
> + leaq 24(%rbp),%rbp
> + movq 72(%rsp),%r13
> + movq 48(%rsp),%r15
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%rax
> + movq %r15,%r14
> + rorq $42,%r15
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%r15
> + shrq $6,%r14
> +
> + rorq $19,%r15
> + xorq %r13,%r12
> + xorq %r14,%r15
> + addq 8(%rsp),%r12
> +
> + addq 64(%rsp),%r12
> + movq %r8,%r13
> + addq %r15,%r12
> + movq %rax,%r14
> + rorq $23,%r13
> + movq %r9,%r15
> +
> + xorq %r8,%r13
> + rorq $5,%r14
> + xorq %r10,%r15
> +
> + movq %r12,64(%rsp)
> + xorq %rax,%r14
> + andq %r8,%r15
> +
> + rorq $4,%r13
> + addq %r11,%r12
> + xorq %r10,%r15
> +
> + rorq $6,%r14
> + xorq %r8,%r13
> + addq %r15,%r12
> +
> + movq %rax,%r15
> + addq (%rbp),%r12
> + xorq %rax,%r14
> +
> + xorq %rbx,%r15
> + rorq $14,%r13
> + movq %rbx,%r11
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%r11
> + addq %r12,%rdx
> + addq %r12,%r11
> +
> + leaq 8(%rbp),%rbp
> + movq 80(%rsp),%r13
> + movq 56(%rsp),%rdi
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%r11
> + movq %rdi,%r14
> + rorq $42,%rdi
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%rdi
> + shrq $6,%r14
> +
> + rorq $19,%rdi
> + xorq %r13,%r12
> + xorq %r14,%rdi
> + addq 16(%rsp),%r12
> +
> + addq 72(%rsp),%r12
> + movq %rdx,%r13
> + addq %rdi,%r12
> + movq %r11,%r14
> + rorq $23,%r13
> + movq %r8,%rdi
> +
> + xorq %rdx,%r13
> + rorq $5,%r14
> + xorq %r9,%rdi
> +
> + movq %r12,72(%rsp)
> + xorq %r11,%r14
> + andq %rdx,%rdi
> +
> + rorq $4,%r13
> + addq %r10,%r12
> + xorq %r9,%rdi
> +
> + rorq $6,%r14
> + xorq %rdx,%r13
> + addq %rdi,%r12
> +
> + movq %r11,%rdi
> + addq (%rbp),%r12
> + xorq %r11,%r14
> +
> + xorq %rax,%rdi
> + rorq $14,%r13
> + movq %rax,%r10
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%r10
> + addq %r12,%rcx
> + addq %r12,%r10
> +
> + leaq 24(%rbp),%rbp
> + movq 88(%rsp),%r13
> + movq 64(%rsp),%r15
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%r10
> + movq %r15,%r14
> + rorq $42,%r15
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%r15
> + shrq $6,%r14
> +
> + rorq $19,%r15
> + xorq %r13,%r12
> + xorq %r14,%r15
> + addq 24(%rsp),%r12
> +
> + addq 80(%rsp),%r12
> + movq %rcx,%r13
> + addq %r15,%r12
> + movq %r10,%r14
> + rorq $23,%r13
> + movq %rdx,%r15
> +
> + xorq %rcx,%r13
> + rorq $5,%r14
> + xorq %r8,%r15
> +
> + movq %r12,80(%rsp)
> + xorq %r10,%r14
> + andq %rcx,%r15
> +
> + rorq $4,%r13
> + addq %r9,%r12
> + xorq %r8,%r15
> +
> + rorq $6,%r14
> + xorq %rcx,%r13
> + addq %r15,%r12
> +
> + movq %r10,%r15
> + addq (%rbp),%r12
> + xorq %r10,%r14
> +
> + xorq %r11,%r15
> + rorq $14,%r13
> + movq %r11,%r9
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%r9
> + addq %r12,%rbx
> + addq %r12,%r9
> +
> + leaq 8(%rbp),%rbp
> + movq 96(%rsp),%r13
> + movq 72(%rsp),%rdi
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%r9
> + movq %rdi,%r14
> + rorq $42,%rdi
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%rdi
> + shrq $6,%r14
> +
> + rorq $19,%rdi
> + xorq %r13,%r12
> + xorq %r14,%rdi
> + addq 32(%rsp),%r12
> +
> + addq 88(%rsp),%r12
> + movq %rbx,%r13
> + addq %rdi,%r12
> + movq %r9,%r14
> + rorq $23,%r13
> + movq %rcx,%rdi
> +
> + xorq %rbx,%r13
> + rorq $5,%r14
> + xorq %rdx,%rdi
> +
> + movq %r12,88(%rsp)
> + xorq %r9,%r14
> + andq %rbx,%rdi
> +
> + rorq $4,%r13
> + addq %r8,%r12
> + xorq %rdx,%rdi
> +
> + rorq $6,%r14
> + xorq %rbx,%r13
> + addq %rdi,%r12
> +
> + movq %r9,%rdi
> + addq (%rbp),%r12
> + xorq %r9,%r14
> +
> + xorq %r10,%rdi
> + rorq $14,%r13
> + movq %r10,%r8
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%r8
> + addq %r12,%rax
> + addq %r12,%r8
> +
> + leaq 24(%rbp),%rbp
> + movq 104(%rsp),%r13
> + movq 80(%rsp),%r15
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%r8
> + movq %r15,%r14
> + rorq $42,%r15
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%r15
> + shrq $6,%r14
> +
> + rorq $19,%r15
> + xorq %r13,%r12
> + xorq %r14,%r15
> + addq 40(%rsp),%r12
> +
> + addq 96(%rsp),%r12
> + movq %rax,%r13
> + addq %r15,%r12
> + movq %r8,%r14
> + rorq $23,%r13
> + movq %rbx,%r15
> +
> + xorq %rax,%r13
> + rorq $5,%r14
> + xorq %rcx,%r15
> +
> + movq %r12,96(%rsp)
> + xorq %r8,%r14
> + andq %rax,%r15
> +
> + rorq $4,%r13
> + addq %rdx,%r12
> + xorq %rcx,%r15
> +
> + rorq $6,%r14
> + xorq %rax,%r13
> + addq %r15,%r12
> +
> + movq %r8,%r15
> + addq (%rbp),%r12
> + xorq %r8,%r14
> +
> + xorq %r9,%r15
> + rorq $14,%r13
> + movq %r9,%rdx
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%rdx
> + addq %r12,%r11
> + addq %r12,%rdx
> +
> + leaq 8(%rbp),%rbp
> + movq 112(%rsp),%r13
> + movq 88(%rsp),%rdi
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%rdx
> + movq %rdi,%r14
> + rorq $42,%rdi
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%rdi
> + shrq $6,%r14
> +
> + rorq $19,%rdi
> + xorq %r13,%r12
> + xorq %r14,%rdi
> + addq 48(%rsp),%r12
> +
> + addq 104(%rsp),%r12
> + movq %r11,%r13
> + addq %rdi,%r12
> + movq %rdx,%r14
> + rorq $23,%r13
> + movq %rax,%rdi
> +
> + xorq %r11,%r13
> + rorq $5,%r14
> + xorq %rbx,%rdi
> +
> + movq %r12,104(%rsp)
> + xorq %rdx,%r14
> + andq %r11,%rdi
> +
> + rorq $4,%r13
> + addq %rcx,%r12
> + xorq %rbx,%rdi
> +
> + rorq $6,%r14
> + xorq %r11,%r13
> + addq %rdi,%r12
> +
> + movq %rdx,%rdi
> + addq (%rbp),%r12
> + xorq %rdx,%r14
> +
> + xorq %r8,%rdi
> + rorq $14,%r13
> + movq %r8,%rcx
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%rcx
> + addq %r12,%r10
> + addq %r12,%rcx
> +
> + leaq 24(%rbp),%rbp
> + movq 120(%rsp),%r13
> + movq 96(%rsp),%r15
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%rcx
> + movq %r15,%r14
> + rorq $42,%r15
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%r15
> + shrq $6,%r14
> +
> + rorq $19,%r15
> + xorq %r13,%r12
> + xorq %r14,%r15
> + addq 56(%rsp),%r12
> +
> + addq 112(%rsp),%r12
> + movq %r10,%r13
> + addq %r15,%r12
> + movq %rcx,%r14
> + rorq $23,%r13
> + movq %r11,%r15
> +
> + xorq %r10,%r13
> + rorq $5,%r14
> + xorq %rax,%r15
> +
> + movq %r12,112(%rsp)
> + xorq %rcx,%r14
> + andq %r10,%r15
> +
> + rorq $4,%r13
> + addq %rbx,%r12
> + xorq %rax,%r15
> +
> + rorq $6,%r14
> + xorq %r10,%r13
> + addq %r15,%r12
> +
> + movq %rcx,%r15
> + addq (%rbp),%r12
> + xorq %rcx,%r14
> +
> + xorq %rdx,%r15
> + rorq $14,%r13
> + movq %rdx,%rbx
> +
> + andq %r15,%rdi
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %rdi,%rbx
> + addq %r12,%r9
> + addq %r12,%rbx
> +
> + leaq 8(%rbp),%rbp
> + movq 0(%rsp),%r13
> + movq 104(%rsp),%rdi
> +
> + movq %r13,%r12
> + rorq $7,%r13
> + addq %r14,%rbx
> + movq %rdi,%r14
> + rorq $42,%rdi
> +
> + xorq %r12,%r13
> + shrq $7,%r12
> + rorq $1,%r13
> + xorq %r14,%rdi
> + shrq $6,%r14
> +
> + rorq $19,%rdi
> + xorq %r13,%r12
> + xorq %r14,%rdi
> + addq 64(%rsp),%r12
> +
> + addq 120(%rsp),%r12
> + movq %r9,%r13
> + addq %rdi,%r12
> + movq %rbx,%r14
> + rorq $23,%r13
> + movq %r10,%rdi
> +
> + xorq %r9,%r13
> + rorq $5,%r14
> + xorq %r11,%rdi
> +
> + movq %r12,120(%rsp)
> + xorq %rbx,%r14
> + andq %r9,%rdi
> +
> + rorq $4,%r13
> + addq %rax,%r12
> + xorq %r11,%rdi
> +
> + rorq $6,%r14
> + xorq %r9,%r13
> + addq %rdi,%r12
> +
> + movq %rbx,%rdi
> + addq (%rbp),%r12
> + xorq %rbx,%r14
> +
> + xorq %rcx,%rdi
> + rorq $14,%r13
> + movq %rcx,%rax
> +
> + andq %rdi,%r15
> + rorq $28,%r14
> + addq %r13,%r12
> +
> + xorq %r15,%rax
> + addq %r12,%r8
> + addq %r12,%rax
> +
> + leaq 24(%rbp),%rbp
> + cmpb $0,7(%rbp)
> + jnz .Lrounds_16_xx
> +
> + movq 128+0(%rsp),%rdi
> + addq %r14,%rax
> + leaq 128(%rsi),%rsi
> +
> + addq 0(%rdi),%rax
> + addq 8(%rdi),%rbx
> + addq 16(%rdi),%rcx
> + addq 24(%rdi),%rdx
> + addq 32(%rdi),%r8
> + addq 40(%rdi),%r9
> + addq 48(%rdi),%r10
> + addq 56(%rdi),%r11
> +
> + cmpq 128+16(%rsp),%rsi
> +
> + movq %rax,0(%rdi)
> + movq %rbx,8(%rdi)
> + movq %rcx,16(%rdi)
> + movq %rdx,24(%rdi)
> + movq %r8,32(%rdi)
> + movq %r9,40(%rdi)
> + movq %r10,48(%rdi)
> + movq %r11,56(%rdi)
> + jb .Lloop
> +
> + movq 152(%rsp),%rsi
> +.cfi_def_cfa %rsi,8
> + movq -48(%rsi),%r15
> +.cfi_restore %r15
> + movq -40(%rsi),%r14
> +.cfi_restore %r14
> + movq -32(%rsi),%r13
> +.cfi_restore %r13
> + movq -24(%rsi),%r12
> +.cfi_restore %r12
> + movq -16(%rsi),%rbp
> +.cfi_restore %rbp
> + movq -8(%rsi),%rbx
> +.cfi_restore %rbx
> + leaq (%rsi),%rsp
> +.cfi_def_cfa_register %rsp
> +.Lepilogue:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size sha512_block_data_order,.-sha512_block_data_order
> +.align 64
> +.type K512, at object
> +K512:
> +.quad 0x428a2f98d728ae22,0x7137449123ef65cd
> +.quad 0x428a2f98d728ae22,0x7137449123ef65cd
> +.quad 0xb5c0fbcfec4d3b2f,0xe9b5dba58189dbbc
> +.quad 0xb5c0fbcfec4d3b2f,0xe9b5dba58189dbbc
> +.quad 0x3956c25bf348b538,0x59f111f1b605d019
> +.quad 0x3956c25bf348b538,0x59f111f1b605d019
> +.quad 0x923f82a4af194f9b,0xab1c5ed5da6d8118
> +.quad 0x923f82a4af194f9b,0xab1c5ed5da6d8118
> +.quad 0xd807aa98a3030242,0x12835b0145706fbe
> +.quad 0xd807aa98a3030242,0x12835b0145706fbe
> +.quad 0x243185be4ee4b28c,0x550c7dc3d5ffb4e2
> +.quad 0x243185be4ee4b28c,0x550c7dc3d5ffb4e2
> +.quad 0x72be5d74f27b896f,0x80deb1fe3b1696b1
> +.quad 0x72be5d74f27b896f,0x80deb1fe3b1696b1
> +.quad 0x9bdc06a725c71235,0xc19bf174cf692694
> +.quad 0x9bdc06a725c71235,0xc19bf174cf692694
> +.quad 0xe49b69c19ef14ad2,0xefbe4786384f25e3
> +.quad 0xe49b69c19ef14ad2,0xefbe4786384f25e3
> +.quad 0x0fc19dc68b8cd5b5,0x240ca1cc77ac9c65
> +.quad 0x0fc19dc68b8cd5b5,0x240ca1cc77ac9c65
> +.quad 0x2de92c6f592b0275,0x4a7484aa6ea6e483
> +.quad 0x2de92c6f592b0275,0x4a7484aa6ea6e483
> +.quad 0x5cb0a9dcbd41fbd4,0x76f988da831153b5
> +.quad 0x5cb0a9dcbd41fbd4,0x76f988da831153b5
> +.quad 0x983e5152ee66dfab,0xa831c66d2db43210
> +.quad 0x983e5152ee66dfab,0xa831c66d2db43210
> +.quad 0xb00327c898fb213f,0xbf597fc7beef0ee4
> +.quad 0xb00327c898fb213f,0xbf597fc7beef0ee4
> +.quad 0xc6e00bf33da88fc2,0xd5a79147930aa725
> +.quad 0xc6e00bf33da88fc2,0xd5a79147930aa725
> +.quad 0x06ca6351e003826f,0x142929670a0e6e70
> +.quad 0x06ca6351e003826f,0x142929670a0e6e70
> +.quad 0x27b70a8546d22ffc,0x2e1b21385c26c926
> +.quad 0x27b70a8546d22ffc,0x2e1b21385c26c926
> +.quad 0x4d2c6dfc5ac42aed,0x53380d139d95b3df
> +.quad 0x4d2c6dfc5ac42aed,0x53380d139d95b3df
> +.quad 0x650a73548baf63de,0x766a0abb3c77b2a8
> +.quad 0x650a73548baf63de,0x766a0abb3c77b2a8
> +.quad 0x81c2c92e47edaee6,0x92722c851482353b
> +.quad 0x81c2c92e47edaee6,0x92722c851482353b
> +.quad 0xa2bfe8a14cf10364,0xa81a664bbc423001
> +.quad 0xa2bfe8a14cf10364,0xa81a664bbc423001
> +.quad 0xc24b8b70d0f89791,0xc76c51a30654be30
> +.quad 0xc24b8b70d0f89791,0xc76c51a30654be30
> +.quad 0xd192e819d6ef5218,0xd69906245565a910
> +.quad 0xd192e819d6ef5218,0xd69906245565a910
> +.quad 0xf40e35855771202a,0x106aa07032bbd1b8
> +.quad 0xf40e35855771202a,0x106aa07032bbd1b8
> +.quad 0x19a4c116b8d2d0c8,0x1e376c085141ab53
> +.quad 0x19a4c116b8d2d0c8,0x1e376c085141ab53
> +.quad 0x2748774cdf8eeb99,0x34b0bcb5e19b48a8
> +.quad 0x2748774cdf8eeb99,0x34b0bcb5e19b48a8
> +.quad 0x391c0cb3c5c95a63,0x4ed8aa4ae3418acb
> +.quad 0x391c0cb3c5c95a63,0x4ed8aa4ae3418acb
> +.quad 0x5b9cca4f7763e373,0x682e6ff3d6b2b8a3
> +.quad 0x5b9cca4f7763e373,0x682e6ff3d6b2b8a3
> +.quad 0x748f82ee5defb2fc,0x78a5636f43172f60
> +.quad 0x748f82ee5defb2fc,0x78a5636f43172f60
> +.quad 0x84c87814a1f0ab72,0x8cc702081a6439ec
> +.quad 0x84c87814a1f0ab72,0x8cc702081a6439ec
> +.quad 0x90befffa23631e28,0xa4506cebde82bde9
> +.quad 0x90befffa23631e28,0xa4506cebde82bde9
> +.quad 0xbef9a3f7b2c67915,0xc67178f2e372532b
> +.quad 0xbef9a3f7b2c67915,0xc67178f2e372532b
> +.quad 0xca273eceea26619c,0xd186b8c721c0c207
> +.quad 0xca273eceea26619c,0xd186b8c721c0c207
> +.quad 0xeada7dd6cde0eb1e,0xf57d4f7fee6ed178
> +.quad 0xeada7dd6cde0eb1e,0xf57d4f7fee6ed178
> +.quad 0x06f067aa72176fba,0x0a637dc5a2c898a6
> +.quad 0x06f067aa72176fba,0x0a637dc5a2c898a6
> +.quad 0x113f9804bef90dae,0x1b710b35131c471b
> +.quad 0x113f9804bef90dae,0x1b710b35131c471b
> +.quad 0x28db77f523047d84,0x32caab7b40c72493
> +.quad 0x28db77f523047d84,0x32caab7b40c72493
> +.quad 0x3c9ebe0a15c9bebc,0x431d67c49c100d4c
> +.quad 0x3c9ebe0a15c9bebc,0x431d67c49c100d4c
> +.quad 0x4cc5d4becb3e42b6,0x597f299cfc657e2a
> +.quad 0x4cc5d4becb3e42b6,0x597f299cfc657e2a
> +.quad 0x5fcb6fab3ad6faec,0x6c44198c4a475817
> +.quad 0x5fcb6fab3ad6faec,0x6c44198c4a475817
> +
> +.quad 0x0001020304050607,0x08090a0b0c0d0e0f
> +.quad 0x0001020304050607,0x08090a0b0c0d0e0f
> +.byte
> 83,72,65,53,49,50,32,98,108,111,99,107,32,116,114,97,110,115,102,111,114,1
> 09,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,
> 32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,1
> 14,103,62,0
> diff --git a/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/x86_64cpuid.S
> b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/x86_64cpuid.S
> new file mode 100644
> index 0000000000..cac5f8f32c
> --- /dev/null
> +++ b/CryptoPkg/Library/OpensslLib/X64Gcc/crypto/x86_64cpuid.S
> @@ -0,0 +1,491 @@
> +# WARNING: do not edit!
> +# Generated from openssl/crypto/x86_64cpuid.pl
> +#
> +# Copyright 2005-2020 The OpenSSL Project Authors. All Rights Reserved.
> +#
> +# Licensed under the OpenSSL license (the "License"). You may not use
> +# this file except in compliance with the License. You can obtain a copy
> +# in the file LICENSE in the source distribution or at
> +# https://www.openssl.org/source/license.html
> +
> +
> +.hidden OPENSSL_cpuid_setup
> +.section .init
> + call OPENSSL_cpuid_setup
> +
> +.hidden OPENSSL_ia32cap_P
> +.comm OPENSSL_ia32cap_P,16,4
> +
> +.text
> +
> +.globl OPENSSL_atomic_add
> +.type OPENSSL_atomic_add, at function
> +.align 16
> +OPENSSL_atomic_add:
> +.cfi_startproc
> + movl (%rdi),%eax
> +.Lspin: leaq (%rsi,%rax,1),%r8
> +.byte 0xf0
> + cmpxchgl %r8d,(%rdi)
> + jne .Lspin
> + movl %r8d,%eax
> +.byte 0x48,0x98
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_atomic_add,.-OPENSSL_atomic_add
> +
> +.globl OPENSSL_rdtsc
> +.type OPENSSL_rdtsc, at function
> +.align 16
> +OPENSSL_rdtsc:
> +.cfi_startproc
> + rdtsc
> + shlq $32,%rdx
> + orq %rdx,%rax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_rdtsc,.-OPENSSL_rdtsc
> +
> +.globl OPENSSL_ia32_cpuid
> +.type OPENSSL_ia32_cpuid, at function
> +.align 16
> +OPENSSL_ia32_cpuid:
> +.cfi_startproc
> + movq %rbx,%r8
> +.cfi_register %rbx,%r8
> +
> + xorl %eax,%eax
> + movq %rax,8(%rdi)
> + cpuid
> + movl %eax,%r11d
> +
> + xorl %eax,%eax
> + cmpl $0x756e6547,%ebx
> + setne %al
> + movl %eax,%r9d
> + cmpl $0x49656e69,%edx
> + setne %al
> + orl %eax,%r9d
> + cmpl $0x6c65746e,%ecx
> + setne %al
> + orl %eax,%r9d
> + jz .Lintel
> +
> + cmpl $0x68747541,%ebx
> + setne %al
> + movl %eax,%r10d
> + cmpl $0x69746E65,%edx
> + setne %al
> + orl %eax,%r10d
> + cmpl $0x444D4163,%ecx
> + setne %al
> + orl %eax,%r10d
> + jnz .Lintel
> +
> +
> + movl $0x80000000,%eax
> + cpuid
> + cmpl $0x80000001,%eax
> + jb .Lintel
> + movl %eax,%r10d
> + movl $0x80000001,%eax
> + cpuid
> + orl %ecx,%r9d
> + andl $0x00000801,%r9d
> +
> + cmpl $0x80000008,%r10d
> + jb .Lintel
> +
> + movl $0x80000008,%eax
> + cpuid
> + movzbq %cl,%r10
> + incq %r10
> +
> + movl $1,%eax
> + cpuid
> + btl $28,%edx
> + jnc .Lgeneric
> + shrl $16,%ebx
> + cmpb %r10b,%bl
> + ja .Lgeneric
> + andl $0xefffffff,%edx
> + jmp .Lgeneric
> +
> +.Lintel:
> + cmpl $4,%r11d
> + movl $-1,%r10d
> + jb .Lnocacheinfo
> +
> + movl $4,%eax
> + movl $0,%ecx
> + cpuid
> + movl %eax,%r10d
> + shrl $14,%r10d
> + andl $0xfff,%r10d
> +
> +.Lnocacheinfo:
> + movl $1,%eax
> + cpuid
> + movd %eax,%xmm0
> + andl $0xbfefffff,%edx
> + cmpl $0,%r9d
> + jne .Lnotintel
> + orl $0x40000000,%edx
> + andb $15,%ah
> + cmpb $15,%ah
> + jne .LnotP4
> + orl $0x00100000,%edx
> +.LnotP4:
> + cmpb $6,%ah
> + jne .Lnotintel
> + andl $0x0fff0ff0,%eax
> + cmpl $0x00050670,%eax
> + je .Lknights
> + cmpl $0x00080650,%eax
> + jne .Lnotintel
> +.Lknights:
> + andl $0xfbffffff,%ecx
> +
> +.Lnotintel:
> + btl $28,%edx
> + jnc .Lgeneric
> + andl $0xefffffff,%edx
> + cmpl $0,%r10d
> + je .Lgeneric
> +
> + orl $0x10000000,%edx
> + shrl $16,%ebx
> + cmpb $1,%bl
> + ja .Lgeneric
> + andl $0xefffffff,%edx
> +.Lgeneric:
> + andl $0x00000800,%r9d
> + andl $0xfffff7ff,%ecx
> + orl %ecx,%r9d
> +
> + movl %edx,%r10d
> +
> + cmpl $7,%r11d
> + jb .Lno_extended_info
> + movl $7,%eax
> + xorl %ecx,%ecx
> + cpuid
> + btl $26,%r9d
> + jc .Lnotknights
> + andl $0xfff7ffff,%ebx
> +.Lnotknights:
> + movd %xmm0,%eax
> + andl $0x0fff0ff0,%eax
> + cmpl $0x00050650,%eax
> + jne .Lnotskylakex
> + andl $0xfffeffff,%ebx
> +
> +.Lnotskylakex:
> + movl %ebx,8(%rdi)
> + movl %ecx,12(%rdi)
> +.Lno_extended_info:
> +
> + btl $27,%r9d
> + jnc .Lclear_avx
> + xorl %ecx,%ecx
> +.byte 0x0f,0x01,0xd0
> + andl $0xe6,%eax
> + cmpl $0xe6,%eax
> + je .Ldone
> + andl $0x3fdeffff,8(%rdi)
> +
> +
> +
> +
> + andl $6,%eax
> + cmpl $6,%eax
> + je .Ldone
> +.Lclear_avx:
> + movl $0xefffe7ff,%eax
> + andl %eax,%r9d
> + movl $0x3fdeffdf,%eax
> + andl %eax,8(%rdi)
> +.Ldone:
> + shlq $32,%r9
> + movl %r10d,%eax
> + movq %r8,%rbx
> +.cfi_restore %rbx
> + orq %r9,%rax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_ia32_cpuid,.-OPENSSL_ia32_cpuid
> +
> +.globl OPENSSL_cleanse
> +.type OPENSSL_cleanse, at function
> +.align 16
> +OPENSSL_cleanse:
> +.cfi_startproc
> + xorq %rax,%rax
> + cmpq $15,%rsi
> + jae .Lot
> + cmpq $0,%rsi
> + je .Lret
> +.Little:
> + movb %al,(%rdi)
> + subq $1,%rsi
> + leaq 1(%rdi),%rdi
> + jnz .Little
> +.Lret:
> + .byte 0xf3,0xc3
> +.align 16
> +.Lot:
> + testq $7,%rdi
> + jz .Laligned
> + movb %al,(%rdi)
> + leaq -1(%rsi),%rsi
> + leaq 1(%rdi),%rdi
> + jmp .Lot
> +.Laligned:
> + movq %rax,(%rdi)
> + leaq -8(%rsi),%rsi
> + testq $-8,%rsi
> + leaq 8(%rdi),%rdi
> + jnz .Laligned
> + cmpq $0,%rsi
> + jne .Little
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_cleanse,.-OPENSSL_cleanse
> +
> +.globl CRYPTO_memcmp
> +.type CRYPTO_memcmp, at function
> +.align 16
> +CRYPTO_memcmp:
> +.cfi_startproc
> + xorq %rax,%rax
> + xorq %r10,%r10
> + cmpq $0,%rdx
> + je .Lno_data
> + cmpq $16,%rdx
> + jne .Loop_cmp
> + movq (%rdi),%r10
> + movq 8(%rdi),%r11
> + movq $1,%rdx
> + xorq (%rsi),%r10
> + xorq 8(%rsi),%r11
> + orq %r11,%r10
> + cmovnzq %rdx,%rax
> + .byte 0xf3,0xc3
> +
> +.align 16
> +.Loop_cmp:
> + movb (%rdi),%r10b
> + leaq 1(%rdi),%rdi
> + xorb (%rsi),%r10b
> + leaq 1(%rsi),%rsi
> + orb %r10b,%al
> + decq %rdx
> + jnz .Loop_cmp
> + negq %rax
> + shrq $63,%rax
> +.Lno_data:
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size CRYPTO_memcmp,.-CRYPTO_memcmp
> +.globl OPENSSL_wipe_cpu
> +.type OPENSSL_wipe_cpu, at function
> +.align 16
> +OPENSSL_wipe_cpu:
> +.cfi_startproc
> + pxor %xmm0,%xmm0
> + pxor %xmm1,%xmm1
> + pxor %xmm2,%xmm2
> + pxor %xmm3,%xmm3
> + pxor %xmm4,%xmm4
> + pxor %xmm5,%xmm5
> + pxor %xmm6,%xmm6
> + pxor %xmm7,%xmm7
> + pxor %xmm8,%xmm8
> + pxor %xmm9,%xmm9
> + pxor %xmm10,%xmm10
> + pxor %xmm11,%xmm11
> + pxor %xmm12,%xmm12
> + pxor %xmm13,%xmm13
> + pxor %xmm14,%xmm14
> + pxor %xmm15,%xmm15
> + xorq %rcx,%rcx
> + xorq %rdx,%rdx
> + xorq %rsi,%rsi
> + xorq %rdi,%rdi
> + xorq %r8,%r8
> + xorq %r9,%r9
> + xorq %r10,%r10
> + xorq %r11,%r11
> + leaq 8(%rsp),%rax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_wipe_cpu,.-OPENSSL_wipe_cpu
> +.globl OPENSSL_instrument_bus
> +.type OPENSSL_instrument_bus, at function
> +.align 16
> +OPENSSL_instrument_bus:
> +.cfi_startproc
> + movq %rdi,%r10
> + movq %rsi,%rcx
> + movq %rsi,%r11
> +
> + rdtsc
> + movl %eax,%r8d
> + movl $0,%r9d
> + clflush (%r10)
> +.byte 0xf0
> + addl %r9d,(%r10)
> + jmp .Loop
> +.align 16
> +.Loop: rdtsc
> + movl %eax,%edx
> + subl %r8d,%eax
> + movl %edx,%r8d
> + movl %eax,%r9d
> + clflush (%r10)
> +.byte 0xf0
> + addl %eax,(%r10)
> + leaq 4(%r10),%r10
> + subq $1,%rcx
> + jnz .Loop
> +
> + movq %r11,%rax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_instrument_bus,.-OPENSSL_instrument_bus
> +
> +.globl OPENSSL_instrument_bus2
> +.type OPENSSL_instrument_bus2, at function
> +.align 16
> +OPENSSL_instrument_bus2:
> +.cfi_startproc
> + movq %rdi,%r10
> + movq %rsi,%rcx
> + movq %rdx,%r11
> + movq %rcx,8(%rsp)
> +
> + rdtsc
> + movl %eax,%r8d
> + movl $0,%r9d
> +
> + clflush (%r10)
> +.byte 0xf0
> + addl %r9d,(%r10)
> +
> + rdtsc
> + movl %eax,%edx
> + subl %r8d,%eax
> + movl %edx,%r8d
> + movl %eax,%r9d
> +.Loop2:
> + clflush (%r10)
> +.byte 0xf0
> + addl %eax,(%r10)
> +
> + subq $1,%r11
> + jz .Ldone2
> +
> + rdtsc
> + movl %eax,%edx
> + subl %r8d,%eax
> + movl %edx,%r8d
> + cmpl %r9d,%eax
> + movl %eax,%r9d
> + movl $0,%edx
> + setne %dl
> + subq %rdx,%rcx
> + leaq (%r10,%rdx,4),%r10
> + jnz .Loop2
> +
> +.Ldone2:
> + movq 8(%rsp),%rax
> + subq %rcx,%rax
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_instrument_bus2,.-OPENSSL_instrument_bus2
> +.globl OPENSSL_ia32_rdrand_bytes
> +.type OPENSSL_ia32_rdrand_bytes, at function
> +.align 16
> +OPENSSL_ia32_rdrand_bytes:
> +.cfi_startproc
> + xorq %rax,%rax
> + cmpq $0,%rsi
> + je .Ldone_rdrand_bytes
> +
> + movq $8,%r11
> +.Loop_rdrand_bytes:
> +.byte 73,15,199,242
> + jc .Lbreak_rdrand_bytes
> + decq %r11
> + jnz .Loop_rdrand_bytes
> + jmp .Ldone_rdrand_bytes
> +
> +.align 16
> +.Lbreak_rdrand_bytes:
> + cmpq $8,%rsi
> + jb .Ltail_rdrand_bytes
> + movq %r10,(%rdi)
> + leaq 8(%rdi),%rdi
> + addq $8,%rax
> + subq $8,%rsi
> + jz .Ldone_rdrand_bytes
> + movq $8,%r11
> + jmp .Loop_rdrand_bytes
> +
> +.align 16
> +.Ltail_rdrand_bytes:
> + movb %r10b,(%rdi)
> + leaq 1(%rdi),%rdi
> + incq %rax
> + shrq $8,%r10
> + decq %rsi
> + jnz .Ltail_rdrand_bytes
> +
> +.Ldone_rdrand_bytes:
> + xorq %r10,%r10
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_ia32_rdrand_bytes,.-OPENSSL_ia32_rdrand_bytes
> +.globl OPENSSL_ia32_rdseed_bytes
> +.type OPENSSL_ia32_rdseed_bytes, at function
> +.align 16
> +OPENSSL_ia32_rdseed_bytes:
> +.cfi_startproc
> + xorq %rax,%rax
> + cmpq $0,%rsi
> + je .Ldone_rdseed_bytes
> +
> + movq $8,%r11
> +.Loop_rdseed_bytes:
> +.byte 73,15,199,250
> + jc .Lbreak_rdseed_bytes
> + decq %r11
> + jnz .Loop_rdseed_bytes
> + jmp .Ldone_rdseed_bytes
> +
> +.align 16
> +.Lbreak_rdseed_bytes:
> + cmpq $8,%rsi
> + jb .Ltail_rdseed_bytes
> + movq %r10,(%rdi)
> + leaq 8(%rdi),%rdi
> + addq $8,%rax
> + subq $8,%rsi
> + jz .Ldone_rdseed_bytes
> + movq $8,%r11
> + jmp .Loop_rdseed_bytes
> +
> +.align 16
> +.Ltail_rdseed_bytes:
> + movb %r10b,(%rdi)
> + leaq 1(%rdi),%rdi
> + incq %rax
> + shrq $8,%r10
> + decq %rsi
> + jnz .Ltail_rdseed_bytes
> +
> +.Ldone_rdseed_bytes:
> + xorq %r10,%r10
> + .byte 0xf3,0xc3
> +.cfi_endproc
> +.size OPENSSL_ia32_rdseed_bytes,.-OPENSSL_ia32_rdseed_bytes
> --
> 2.32.0.windows.1
-=-=-=-=-=-=-=-=-=-=-=-
Groups.io Links: You receive all messages sent to this group.
View/Reply Online (#78021): https://edk2.groups.io/g/devel/message/78021
Mute This Topic: https://groups.io/mt/84344212/1813853
Group Owner: devel+owner at edk2.groups.io
Unsubscribe: https://edk2.groups.io/g/devel/unsub [edk2-devel-archive at redhat.com]
-=-=-=-=-=-=-=-=-=-=-=-
More information about the edk2-devel-archive
mailing list