Source-Changes-HG archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

[src/trunk]: src/crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64 regen



details:   https://anonhg.NetBSD.org/src/rev/5d215b117849
branches:  trunk
changeset: 350233:5d215b117849
user:      christos <christos%NetBSD.org@localhost>
date:      Sun Jan 08 01:59:43 2017 +0000

description:
regen

diffstat:

 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aes-x86_64.S          |     2 +-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aesni-gcm-x86_64.S    |   754 +-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aesni-mb-x86_64.S     |   929 ++
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aesni-sha1-x86_64.S   |  1327 ++-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aesni-sha256-x86_64.S |  4297 +++++++++
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/ecp_nistz256-x86_64.S |  1509 +++-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/ghash-x86_64.S        |   475 +-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/rc4-x86_64.S          |     6 +-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/rsaz-avx2.S           |  1705 +++-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/rsaz-x86_64.S         |   654 +
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/sha1-mb-x86_64.S      |  4287 +++++++++
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/sha1-x86_64.S         |  2805 ++++++
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/sha256-mb-x86_64.S    |  4644 ++++++++++
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/sha512-x86_64.S       |  2311 ++++-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/uplink-x86_64.S       |    88 +-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/x86_64-gf2m.S         |     5 +-
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/x86_64-mont.S         |   363 +
 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/x86_64-mont5.S        |  1317 ++
 18 files changed, 27387 insertions(+), 91 deletions(-)

diffs (truncated from 28305 to 300 lines):

diff -r 1ed056e4e9de -r 5d215b117849 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aes-x86_64.S
--- a/crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aes-x86_64.S        Sun Jan 08 01:56:31 2017 +0000
+++ b/crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aes-x86_64.S        Sun Jan 08 01:59:43 2017 +0000
@@ -1318,7 +1318,7 @@
        leaq    .LAES_Td(%rip),%r14
 .Lcbc_picked_te:
 
-       movl    OPENSSL_ia32cap_P@GOTPCREL(%rip),%r10d
+       movl    OPENSSL_ia32cap_P(%rip),%r10d
        cmpq    $512,%rdx
        jb      .Lcbc_slow_prologue
        testq   $15,%rdx
diff -r 1ed056e4e9de -r 5d215b117849 crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aesni-gcm-x86_64.S
--- a/crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aesni-gcm-x86_64.S  Sun Jan 08 01:56:31 2017 +0000
+++ b/crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/aesni-gcm-x86_64.S  Sun Jan 08 01:59:43 2017 +0000
@@ -1,16 +1,754 @@
 #include <machine/asm.h>
 .text  
 
+.type  _aesni_ctr32_ghash_6x,@function
+.align 32
+_aesni_ctr32_ghash_6x:
+       vmovdqu 32(%r11),%xmm2
+       subq    $6,%rdx
+       vpxor   %xmm4,%xmm4,%xmm4
+       vmovdqu 0-128(%rcx),%xmm15
+       vpaddb  %xmm2,%xmm1,%xmm10
+       vpaddb  %xmm2,%xmm10,%xmm11
+       vpaddb  %xmm2,%xmm11,%xmm12
+       vpaddb  %xmm2,%xmm12,%xmm13
+       vpaddb  %xmm2,%xmm13,%xmm14
+       vpxor   %xmm15,%xmm1,%xmm9
+       vmovdqu %xmm4,16+8(%rsp)
+       jmp     .Loop6x
+
+.align 32
+.Loop6x:
+       addl    $100663296,%ebx
+       jc      .Lhandle_ctr32
+       vmovdqu 0-32(%r9),%xmm3
+       vpaddb  %xmm2,%xmm14,%xmm1
+       vpxor   %xmm15,%xmm10,%xmm10
+       vpxor   %xmm15,%xmm11,%xmm11
+
+.Lresume_ctr32:
+       vmovdqu %xmm1,(%r8)
+       vpclmulqdq      $0x10,%xmm3,%xmm7,%xmm5
+       vpxor   %xmm15,%xmm12,%xmm12
+       vmovups 16-128(%rcx),%xmm2
+       vpclmulqdq      $0x01,%xmm3,%xmm7,%xmm6
+       xorq    %r12,%r12
+       cmpq    %r14,%r15
+
+       vaesenc %xmm2,%xmm9,%xmm9
+       vmovdqu 48+8(%rsp),%xmm0
+       vpxor   %xmm15,%xmm13,%xmm13
+       vpclmulqdq      $0x00,%xmm3,%xmm7,%xmm1
+       vaesenc %xmm2,%xmm10,%xmm10
+       vpxor   %xmm15,%xmm14,%xmm14
+       setnc   %r12b
+       vpclmulqdq      $0x11,%xmm3,%xmm7,%xmm7
+       vaesenc %xmm2,%xmm11,%xmm11
+       vmovdqu 16-32(%r9),%xmm3
+       negq    %r12
+       vaesenc %xmm2,%xmm12,%xmm12
+       vpxor   %xmm5,%xmm6,%xmm6
+       vpclmulqdq      $0x00,%xmm3,%xmm0,%xmm5
+       vpxor   %xmm4,%xmm8,%xmm8
+       vaesenc %xmm2,%xmm13,%xmm13
+       vpxor   %xmm5,%xmm1,%xmm4
+       andq    $0x60,%r12
+       vmovups 32-128(%rcx),%xmm15
+       vpclmulqdq      $0x10,%xmm3,%xmm0,%xmm1
+       vaesenc %xmm2,%xmm14,%xmm14
+
+       vpclmulqdq      $0x01,%xmm3,%xmm0,%xmm2
+       leaq    (%r14,%r12,1),%r14
+       vaesenc %xmm15,%xmm9,%xmm9
+       vpxor   16+8(%rsp),%xmm8,%xmm8
+       vpclmulqdq      $0x11,%xmm3,%xmm0,%xmm3
+       vmovdqu 64+8(%rsp),%xmm0
+       vaesenc %xmm15,%xmm10,%xmm10
+       movbeq  88(%r14),%r13
+       vaesenc %xmm15,%xmm11,%xmm11
+       movbeq  80(%r14),%r12
+       vaesenc %xmm15,%xmm12,%xmm12
+       movq    %r13,32+8(%rsp)
+       vaesenc %xmm15,%xmm13,%xmm13
+       movq    %r12,40+8(%rsp)
+       vmovdqu 48-32(%r9),%xmm5
+       vaesenc %xmm15,%xmm14,%xmm14
+
+       vmovups 48-128(%rcx),%xmm15
+       vpxor   %xmm1,%xmm6,%xmm6
+       vpclmulqdq      $0x00,%xmm5,%xmm0,%xmm1
+       vaesenc %xmm15,%xmm9,%xmm9
+       vpxor   %xmm2,%xmm6,%xmm6
+       vpclmulqdq      $0x10,%xmm5,%xmm0,%xmm2
+       vaesenc %xmm15,%xmm10,%xmm10
+       vpxor   %xmm3,%xmm7,%xmm7
+       vpclmulqdq      $0x01,%xmm5,%xmm0,%xmm3
+       vaesenc %xmm15,%xmm11,%xmm11
+       vpclmulqdq      $0x11,%xmm5,%xmm0,%xmm5
+       vmovdqu 80+8(%rsp),%xmm0
+       vaesenc %xmm15,%xmm12,%xmm12
+       vaesenc %xmm15,%xmm13,%xmm13
+       vpxor   %xmm1,%xmm4,%xmm4
+       vmovdqu 64-32(%r9),%xmm1
+       vaesenc %xmm15,%xmm14,%xmm14
+
+       vmovups 64-128(%rcx),%xmm15
+       vpxor   %xmm2,%xmm6,%xmm6
+       vpclmulqdq      $0x00,%xmm1,%xmm0,%xmm2
+       vaesenc %xmm15,%xmm9,%xmm9
+       vpxor   %xmm3,%xmm6,%xmm6
+       vpclmulqdq      $0x10,%xmm1,%xmm0,%xmm3
+       vaesenc %xmm15,%xmm10,%xmm10
+       movbeq  72(%r14),%r13
+       vpxor   %xmm5,%xmm7,%xmm7
+       vpclmulqdq      $0x01,%xmm1,%xmm0,%xmm5
+       vaesenc %xmm15,%xmm11,%xmm11
+       movbeq  64(%r14),%r12
+       vpclmulqdq      $0x11,%xmm1,%xmm0,%xmm1
+       vmovdqu 96+8(%rsp),%xmm0
+       vaesenc %xmm15,%xmm12,%xmm12
+       movq    %r13,48+8(%rsp)
+       vaesenc %xmm15,%xmm13,%xmm13
+       movq    %r12,56+8(%rsp)
+       vpxor   %xmm2,%xmm4,%xmm4
+       vmovdqu 96-32(%r9),%xmm2
+       vaesenc %xmm15,%xmm14,%xmm14
+
+       vmovups 80-128(%rcx),%xmm15
+       vpxor   %xmm3,%xmm6,%xmm6
+       vpclmulqdq      $0x00,%xmm2,%xmm0,%xmm3
+       vaesenc %xmm15,%xmm9,%xmm9
+       vpxor   %xmm5,%xmm6,%xmm6
+       vpclmulqdq      $0x10,%xmm2,%xmm0,%xmm5
+       vaesenc %xmm15,%xmm10,%xmm10
+       movbeq  56(%r14),%r13
+       vpxor   %xmm1,%xmm7,%xmm7
+       vpclmulqdq      $0x01,%xmm2,%xmm0,%xmm1
+       vpxor   112+8(%rsp),%xmm8,%xmm8
+       vaesenc %xmm15,%xmm11,%xmm11
+       movbeq  48(%r14),%r12
+       vpclmulqdq      $0x11,%xmm2,%xmm0,%xmm2
+       vaesenc %xmm15,%xmm12,%xmm12
+       movq    %r13,64+8(%rsp)
+       vaesenc %xmm15,%xmm13,%xmm13
+       movq    %r12,72+8(%rsp)
+       vpxor   %xmm3,%xmm4,%xmm4
+       vmovdqu 112-32(%r9),%xmm3
+       vaesenc %xmm15,%xmm14,%xmm14
+
+       vmovups 96-128(%rcx),%xmm15
+       vpxor   %xmm5,%xmm6,%xmm6
+       vpclmulqdq      $0x10,%xmm3,%xmm8,%xmm5
+       vaesenc %xmm15,%xmm9,%xmm9
+       vpxor   %xmm1,%xmm6,%xmm6
+       vpclmulqdq      $0x01,%xmm3,%xmm8,%xmm1
+       vaesenc %xmm15,%xmm10,%xmm10
+       movbeq  40(%r14),%r13
+       vpxor   %xmm2,%xmm7,%xmm7
+       vpclmulqdq      $0x00,%xmm3,%xmm8,%xmm2
+       vaesenc %xmm15,%xmm11,%xmm11
+       movbeq  32(%r14),%r12
+       vpclmulqdq      $0x11,%xmm3,%xmm8,%xmm8
+       vaesenc %xmm15,%xmm12,%xmm12
+       movq    %r13,80+8(%rsp)
+       vaesenc %xmm15,%xmm13,%xmm13
+       movq    %r12,88+8(%rsp)
+       vpxor   %xmm5,%xmm6,%xmm6
+       vaesenc %xmm15,%xmm14,%xmm14
+       vpxor   %xmm1,%xmm6,%xmm6
+
+       vmovups 112-128(%rcx),%xmm15
+       vpslldq $8,%xmm6,%xmm5
+       vpxor   %xmm2,%xmm4,%xmm4
+       vmovdqu 16(%r11),%xmm3
+
+       vaesenc %xmm15,%xmm9,%xmm9
+       vpxor   %xmm8,%xmm7,%xmm7
+       vaesenc %xmm15,%xmm10,%xmm10
+       vpxor   %xmm5,%xmm4,%xmm4
+       movbeq  24(%r14),%r13
+       vaesenc %xmm15,%xmm11,%xmm11
+       movbeq  16(%r14),%r12
+       vpalignr        $8,%xmm4,%xmm4,%xmm0
+       vpclmulqdq      $0x10,%xmm3,%xmm4,%xmm4
+       movq    %r13,96+8(%rsp)
+       vaesenc %xmm15,%xmm12,%xmm12
+       movq    %r12,104+8(%rsp)
+       vaesenc %xmm15,%xmm13,%xmm13
+       vmovups 128-128(%rcx),%xmm1
+       vaesenc %xmm15,%xmm14,%xmm14
+
+       vaesenc %xmm1,%xmm9,%xmm9
+       vmovups 144-128(%rcx),%xmm15
+       vaesenc %xmm1,%xmm10,%xmm10
+       vpsrldq $8,%xmm6,%xmm6
+       vaesenc %xmm1,%xmm11,%xmm11
+       vpxor   %xmm6,%xmm7,%xmm7
+       vaesenc %xmm1,%xmm12,%xmm12
+       vpxor   %xmm0,%xmm4,%xmm4
+       movbeq  8(%r14),%r13
+       vaesenc %xmm1,%xmm13,%xmm13
+       movbeq  0(%r14),%r12
+       vaesenc %xmm1,%xmm14,%xmm14
+       vmovups 160-128(%rcx),%xmm1
+       cmpl    $11,%ebp
+       jb      .Lenc_tail
+
+       vaesenc %xmm15,%xmm9,%xmm9
+       vaesenc %xmm15,%xmm10,%xmm10
+       vaesenc %xmm15,%xmm11,%xmm11
+       vaesenc %xmm15,%xmm12,%xmm12
+       vaesenc %xmm15,%xmm13,%xmm13
+       vaesenc %xmm15,%xmm14,%xmm14
+
+       vaesenc %xmm1,%xmm9,%xmm9
+       vaesenc %xmm1,%xmm10,%xmm10
+       vaesenc %xmm1,%xmm11,%xmm11
+       vaesenc %xmm1,%xmm12,%xmm12
+       vaesenc %xmm1,%xmm13,%xmm13
+       vmovups 176-128(%rcx),%xmm15
+       vaesenc %xmm1,%xmm14,%xmm14
+       vmovups 192-128(%rcx),%xmm1
+       je      .Lenc_tail
+
+       vaesenc %xmm15,%xmm9,%xmm9
+       vaesenc %xmm15,%xmm10,%xmm10
+       vaesenc %xmm15,%xmm11,%xmm11
+       vaesenc %xmm15,%xmm12,%xmm12
+       vaesenc %xmm15,%xmm13,%xmm13
+       vaesenc %xmm15,%xmm14,%xmm14
+
+       vaesenc %xmm1,%xmm9,%xmm9
+       vaesenc %xmm1,%xmm10,%xmm10
+       vaesenc %xmm1,%xmm11,%xmm11
+       vaesenc %xmm1,%xmm12,%xmm12
+       vaesenc %xmm1,%xmm13,%xmm13
+       vmovups 208-128(%rcx),%xmm15
+       vaesenc %xmm1,%xmm14,%xmm14
+       vmovups 224-128(%rcx),%xmm1
+       jmp     .Lenc_tail
+
+.align 32
+.Lhandle_ctr32:
+       vmovdqu (%r11),%xmm0
+       vpshufb %xmm0,%xmm1,%xmm6
+       vmovdqu 48(%r11),%xmm5
+       vpaddd  64(%r11),%xmm6,%xmm10
+       vpaddd  %xmm5,%xmm6,%xmm11
+       vmovdqu 0-32(%r9),%xmm3
+       vpaddd  %xmm5,%xmm10,%xmm12
+       vpshufb %xmm0,%xmm10,%xmm10
+       vpaddd  %xmm5,%xmm11,%xmm13
+       vpshufb %xmm0,%xmm11,%xmm11
+       vpxor   %xmm15,%xmm10,%xmm10
+       vpaddd  %xmm5,%xmm12,%xmm14
+       vpshufb %xmm0,%xmm12,%xmm12
+       vpxor   %xmm15,%xmm11,%xmm11
+       vpaddd  %xmm5,%xmm13,%xmm1
+       vpshufb %xmm0,%xmm13,%xmm13
+       vpshufb %xmm0,%xmm14,%xmm14
+       vpshufb %xmm0,%xmm1,%xmm1
+       jmp     .Lresume_ctr32
+
+.align 32
+.Lenc_tail:
+       vaesenc %xmm15,%xmm9,%xmm9
+       vmovdqu %xmm7,16+8(%rsp)
+       vpalignr        $8,%xmm4,%xmm4,%xmm8
+       vaesenc %xmm15,%xmm10,%xmm10
+       vpclmulqdq      $0x10,%xmm3,%xmm4,%xmm4
+       vpxor   0(%rdi),%xmm1,%xmm2
+       vaesenc %xmm15,%xmm11,%xmm11
+       vpxor   16(%rdi),%xmm1,%xmm0
+       vaesenc %xmm15,%xmm12,%xmm12
+       vpxor   32(%rdi),%xmm1,%xmm5
+       vaesenc %xmm15,%xmm13,%xmm13
+       vpxor   48(%rdi),%xmm1,%xmm6
+       vaesenc %xmm15,%xmm14,%xmm14
+       vpxor   64(%rdi),%xmm1,%xmm7
+       vpxor   80(%rdi),%xmm1,%xmm3
+       vmovdqu (%r8),%xmm1
+
+       vaesenclast     %xmm2,%xmm9,%xmm9
+       vmovdqu 32(%r11),%xmm2
+       vaesenclast     %xmm0,%xmm10,%xmm10
+       vpaddb  %xmm2,%xmm1,%xmm0
+       movq    %r13,112+8(%rsp)
+       leaq    96(%rdi),%rdi
+       vaesenclast     %xmm5,%xmm11,%xmm11
+       vpaddb  %xmm2,%xmm0,%xmm5
+       movq    %r12,120+8(%rsp)
+       leaq    96(%rsi),%rsi
+       vmovdqu 0-128(%rcx),%xmm15
+       vaesenclast     %xmm6,%xmm12,%xmm12
+       vpaddb  %xmm2,%xmm5,%xmm6



Home | Main Index | Thread Index | Old Index