]> Cypherpunks repositories - gostls13.git/commitdiff
crypto/sha256: Use AVX2 if possible
authorIlya Tocar <ilya.tocar@intel.com>
Fri, 29 Apr 2016 13:19:17 +0000 (16:19 +0300)
committerRuss Cox <rsc@golang.org>
Fri, 6 May 2016 15:20:37 +0000 (15:20 +0000)
name          old time/op    new time/op    delta
Hash8Bytes-4     376ns ± 0%     246ns ± 0%  -34.57%  (p=0.000 n=20+20)
Hash1K-4        5.21µs ± 0%    2.82µs ± 0%  -45.83%  (p=0.000 n=20+20)
Hash8K-4        38.6µs ± 0%    20.8µs ± 0%  -46.05%  (p=0.000 n=20+20)

name          old speed      new speed      delta
Hash8Bytes-4  21.2MB/s ± 0%  32.4MB/s ± 0%  +52.70%  (p=0.000 n=15+19)
Hash1K-4       197MB/s ± 0%   363MB/s ± 0%  +84.60%  (p=0.000 n=20+20)
Hash8K-4       212MB/s ± 0%   393MB/s ± 0%  +85.36%  (p=0.000 n=20+20)

Change-Id: Ib50119c591074ff486d11d3566e24b691bcc6598
Reviewed-on: https://go-review.googlesource.com/22608
Run-TryBot: Russ Cox <rsc@golang.org>
TryBot-Result: Gobot Gobot <gobot@golang.org>
Reviewed-by: Russ Cox <rsc@golang.org>
src/crypto/sha256/sha256block_amd64.s

index c9f134c1069082977e79a88f036e9ae30d2ce44c..6ab3b52d65e16336b5036f26207d9b9296160d33 100644 (file)
@@ -1,4 +1,4 @@
-// Copyright 2013 The Go Authors. All rights reserved.
+// Copyright 2013 The Go Authors.  All rights reserved.
 // Use of this source code is governed by a BSD-style
 // license that can be found in the LICENSE file.
 
@@ -9,7 +9,18 @@
 // The algorithm is detailed in FIPS 180-4:
 //
 //  http://csrc.nist.gov/publications/fips/fips180-4/fips-180-4.pdf
-//
+
+// The avx2-version is described in an Intel White-Paper:
+// "Fast SHA-256 Implementations on Intel Architecture Processors"
+// To find it, surf to http://www.intel.com/p/en_US/embedded
+// and search for that title.
+// AVX2 version by Intel, same algorithm as code in Linux kernel:
+// https://github.com/torvalds/linux/blob/master/arch/x86/crypto/sha256-avx2-asm.S
+// by
+//     James Guilford <james.guilford@intel.com>
+//     Kirk Yap <kirk.s.yap@intel.com>
+//     Tim Chen <tim.c.chen@linux.intel.com>
+
 // Wt = Mt; for 0 <= t <= 15
 // Wt = SIGMA1(Wt-2) + SIGMA0(Wt-15) + Wt-16; for 16 <= t <= 63
 //
        MSGSCHEDULE1(index); \
        SHA256ROUND(index, const, a, b, c, d, e, f, g, h)
 
-TEXT ·block(SB),0,$264-32
-       MOVQ    p_base+8(FP), SI
-       MOVQ    p_len+16(FP), DX
-       SHRQ    $6, DX
-       SHLQ    $6, DX
-
-       LEAQ    (SI)(DX*1), DI
-       MOVQ    DI, 256(SP)
-       CMPQ    SI, DI
-       JEQ     end
-
-       MOVQ    dig+0(FP), BP
-       MOVL    (0*4)(BP), R8           // a = H0
-       MOVL    (1*4)(BP), R9           // b = H1
-       MOVL    (2*4)(BP), R10          // c = H2
-       MOVL    (3*4)(BP), R11          // d = H3
-       MOVL    (4*4)(BP), R12          // e = H4
-       MOVL    (5*4)(BP), R13          // f = H5
-       MOVL    (6*4)(BP), R14          // g = H6
-       MOVL    (7*4)(BP), R15          // h = H7
+
+// Definitions for AVX2 version
+
+// addm (mem), reg
+// Add reg to mem using reg-mem add and store
+#define addm(P1, P2) \
+       ADDL P2, P1; \
+       MOVL P1, P2
+
+#define XDWORD0 Y4
+#define XDWORD1 Y5
+#define XDWORD2 Y6
+#define XDWORD3 Y7
+
+#define XWORD0 X4
+#define XWORD1 X5
+#define XWORD2 X6
+#define XWORD3 X7
+
+#define XTMP0 Y0
+#define XTMP1 Y1
+#define XTMP2 Y2
+#define XTMP3 Y3
+#define XTMP4 Y8
+#define XTMP5 Y11
+
+#define XFER  Y9
+
+#define BYTE_FLIP_MASK         Y13 // mask to convert LE -> BE
+#define X_BYTE_FLIP_MASK X13
+
+#define NUM_BYTES DX
+#define INP    DI
+
+#define CTX SI // Beginning of digest in memory (a, b, c, ... , h)
+
+#define a AX
+#define b BX
+#define c CX
+#define d R8
+#define e DX
+#define f R9
+#define g R10
+#define h R11
+
+#define old_h R11
+
+#define TBL BP
+
+#define SRND SI // SRND is same register as CTX
+
+#define T1 R12
+
+#define y0 R13
+#define y1 R14
+#define y2 R15
+#define y3 DI
+
+// Offsets
+#define XFER_SIZE 2*64*4
+#define INP_END_SIZE 8
+#define INP_SIZE 8
+#define TMP_SIZE 4
+
+#define _XFER 0
+#define _INP_END _XFER + XFER_SIZE
+#define _INP _INP_END + INP_END_SIZE
+#define _TMP _INP + INP_SIZE
+#define STACK_SIZE _TMP + TMP_SIZE
+
+#define ROUND_AND_SCHED_N_0(disp, a, b, c, d, e, f, g, h, XDWORD0, XDWORD1, XDWORD2, XDWORD3) \
+       ;                                     \ // #############################  RND N + 0 ############################//
+       MOVL     a, y3;                       \ // y3 = a                                       // MAJA
+       RORXL    $25, e, y0;                  \ // y0 = e >> 25                         // S1A
+       RORXL    $11, e, y1;                  \ // y1 = e >> 11                         // S1B
+       ;                                     \
+       ADDL     (disp + 0*4)(SP)(SRND*1), h; \ // h = k + w + h        // disp = k + w
+       ORL      c, y3;                       \ // y3 = a|c                             // MAJA
+       VPALIGNR $4, XDWORD2, XDWORD3, XTMP0; \ // XTMP0 = W[-7]
+       MOVL     f, y2;                       \ // y2 = f                               // CH
+       RORXL    $13, a, T1;                  \ // T1 = a >> 13                 // S0B
+       ;                                     \
+       XORL     y1, y0;                      \ // y0 = (e>>25) ^ (e>>11)                                       // S1
+       XORL     g, y2;                       \ // y2 = f^g                                     // CH
+       VPADDD   XDWORD0, XTMP0, XTMP0;       \ // XTMP0 = W[-7] + W[-16]       // y1 = (e >> 6)        // S1
+       RORXL    $6, e, y1;                   \ // y1 = (e >> 6)                                                // S1
+       ;                                     \
+       ANDL     e, y2;                       \ // y2 = (f^g)&e                         // CH
+       XORL     y1, y0;                      \ // y0 = (e>>25) ^ (e>>11) ^ (e>>6)              // S1
+       RORXL    $22, a, y1;                  \ // y1 = a >> 22                                                 // S0A
+       ADDL     h, d;                        \ // d = k + w + h + d                            // --
+       ;                                     \
+       ANDL     b, y3;                       \ // y3 = (a|c)&b                                                 // MAJA
+       VPALIGNR $4, XDWORD0, XDWORD1, XTMP1; \ // XTMP1 = W[-15]
+       XORL     T1, y1;                      \ // y1 = (a>>22) ^ (a>>13)                               // S0
+       RORXL    $2, a, T1;                   \ // T1 = (a >> 2)                                                // S0
+       ;                                     \
+       XORL     g, y2;                       \ // y2 = CH = ((f^g)&e)^g                                // CH
+       VPSRLD   $7, XTMP1, XTMP2;            \
+       XORL     T1, y1;                      \ // y1 = (a>>22) ^ (a>>13) ^ (a>>2)              // S0
+       MOVL     a, T1;                       \ // T1 = a                                                               // MAJB
+       ANDL     c, T1;                       \ // T1 = a&c                                                             // MAJB
+       ;                                     \
+       ADDL     y0, y2;                      \ // y2 = S1 + CH                                                 // --
+       VPSLLD   $(32-7), XTMP1, XTMP3;       \
+       ORL      T1, y3;                      \ // y3 = MAJ = (a|c)&b)|(a&c)                    // MAJ
+       ADDL     y1, h;                       \ // h = k + w + h + S0                                   // --
+       ;                                     \
+       ADDL     y2, d;                       \ // d = k + w + h + d + S1 + CH = d + t1  // --
+       VPOR     XTMP2, XTMP3, XTMP3;         \ // XTMP3 = W[-15] ror 7
+       ;                                     \
+       VPSRLD   $18, XTMP1, XTMP2;           \
+       ADDL     y2, h;                       \ // h = k + w + h + S0 + S1 + CH = t1 + S0// --
+       ADDL     y3, h                        // h = t1 + S0 + MAJ                     // --
+
+#define ROUND_AND_SCHED_N_1(disp, a, b, c, d, e, f, g, h, XDWORD0, XDWORD1, XDWORD2, XDWORD3) \
+       ;                                    \ // ################################### RND N + 1 ############################
+       ;                                    \
+       MOVL    a, y3;                       \ // y3 = a                       // MAJA
+       RORXL   $25, e, y0;                  \ // y0 = e >> 25                                  // S1A
+       RORXL   $11, e, y1;                  \ // y1 = e >> 11                                  // S1B
+       ADDL    (disp + 1*4)(SP)(SRND*1), h; \ // h = k + w + h                         // --
+       ORL     c, y3;                       \ // y3 = a|c                                              // MAJA
+       ;                                    \
+       VPSRLD  $3, XTMP1, XTMP4;            \ // XTMP4 = W[-15] >> 3
+       MOVL    f, y2;                       \ // y2 = f                                                // CH
+       RORXL   $13, a, T1;                  \ // T1 = a >> 13                                  // S0B
+       XORL    y1, y0;                      \ // y0 = (e>>25) ^ (e>>11)                // S1
+       XORL    g, y2;                       \ // y2 = f^g                                              // CH
+       ;                                    \
+       RORXL   $6, e, y1;                   \ // y1 = (e >> 6)                         // S1
+       XORL    y1, y0;                      \ // y0 = (e>>25) ^ (e>>11) ^ (e>>6)       // S1
+       RORXL   $22, a, y1;                  \ // y1 = a >> 22                                          // S0A
+       ANDL    e, y2;                       \ // y2 = (f^g)&e                                          // CH
+       ADDL    h, d;                        \ // d = k + w + h + d                             // --
+       ;                                    \
+       VPSLLD  $(32-18), XTMP1, XTMP1;      \
+       ANDL    b, y3;                       \ // y3 = (a|c)&b                                  // MAJA
+       XORL    T1, y1;                      \ // y1 = (a>>22) ^ (a>>13)                // S0
+       ;                                    \
+       VPXOR   XTMP1, XTMP3, XTMP3;         \
+       RORXL   $2, a, T1;                   \ // T1 = (a >> 2)                         // S0
+       XORL    g, y2;                       \ // y2 = CH = ((f^g)&e)^g         // CH
+       ;                                    \
+       VPXOR   XTMP2, XTMP3, XTMP3;         \ // XTMP3 = W[-15] ror 7 ^ W[-15] ror 18
+       XORL    T1, y1;                      \ // y1 = (a>>22) ^ (a>>13) ^ (a>>2)               // S0
+       MOVL    a, T1;                       \ // T1 = a                                                // MAJB
+       ANDL    c, T1;                       \ // T1 = a&c                                              // MAJB
+       ADDL    y0, y2;                      \ // y2 = S1 + CH                                  // --
+       ;                                    \
+       VPXOR   XTMP4, XTMP3, XTMP1;         \ // XTMP1 = s0
+       VPSHUFD $-6, XDWORD3, XTMP2;         \ // XTMP2 = W[-2] {BBAA}
+       ORL     T1, y3;                      \ // y3 = MAJ = (a|c)&b)|(a&c)             // MAJ
+       ADDL    y1, h;                       \ // h = k + w + h + S0                    // --
+       ;                                    \
+       VPADDD  XTMP1, XTMP0, XTMP0;         \ // XTMP0 = W[-16] + W[-7] + s0
+       ADDL    y2, d;                       \ // d = k + w + h + d + S1 + CH = d + t1  // --
+       ADDL    y2, h;                       \ // h = k + w + h + S0 + S1 + CH = t1 + S0// --
+       ADDL    y3, h;                       \ // h = t1 + S0 + MAJ                     // --
+       ;                                    \
+       VPSRLD  $10, XTMP2, XTMP4            // XTMP4 = W[-2] >> 10 {BBAA}
+
+#define ROUND_AND_SCHED_N_2(disp, a, b, c, d, e, f, g, h, XDWORD0, XDWORD1, XDWORD2, XDWORD3) \
+       ;                                    \ // ################################### RND N + 2 ############################
+       ;                                    \
+       MOVL    a, y3;                       \ // y3 = a                                                        // MAJA
+       RORXL   $25, e, y0;                  \ // y0 = e >> 25                                          // S1A
+       ADDL    (disp + 2*4)(SP)(SRND*1), h; \ // h = k + w + h                         // --
+       ;                                    \
+       VPSRLQ  $19, XTMP2, XTMP3;           \ // XTMP3 = W[-2] ror 19 {xBxA}
+       RORXL   $11, e, y1;                  \ // y1 = e >> 11                                          // S1B
+       ORL     c, y3;                       \ // y3 = a|c                         // MAJA
+       MOVL    f, y2;                       \ // y2 = f                           // CH
+       XORL    g, y2;                       \ // y2 = f^g                         // CH
+       ;                                    \
+       RORXL   $13, a, T1;                  \ // T1 = a >> 13                                          // S0B
+       XORL    y1, y0;                      \ // y0 = (e>>25) ^ (e>>11)                        // S1
+       VPSRLQ  $17, XTMP2, XTMP2;           \ // XTMP2 = W[-2] ror 17 {xBxA}
+       ANDL    e, y2;                       \ // y2 = (f^g)&e                                          // CH
+       ;                                    \
+       RORXL   $6, e, y1;                   \ // y1 = (e >> 6)                                 // S1
+       VPXOR   XTMP3, XTMP2, XTMP2;         \
+       ADDL    h, d;                        \ // d = k + w + h + d                             // --
+       ANDL    b, y3;                       \ // y3 = (a|c)&b                                          // MAJA
+       ;                                    \
+       XORL    y1, y0;                      \ // y0 = (e>>25) ^ (e>>11) ^ (e>>6)       // S1
+       RORXL   $22, a, y1;                  \ // y1 = a >> 22                                          // S0A
+       VPXOR   XTMP2, XTMP4, XTMP4;         \ // XTMP4 = s1 {xBxA}
+       XORL    g, y2;                       \ // y2 = CH = ((f^g)&e)^g                 // CH
+       ;                                    \
+       MOVL    f, _TMP(SP);                 \
+       MOVQ    $shuff_00BA<>(SB), f;        \ // f is used to keep SHUF_00BA
+       VPSHUFB (f), XTMP4, XTMP4;           \ // XTMP4 = s1 {00BA}
+       MOVL    _TMP(SP), f;                 \ // f is restored
+       ;                                    \
+       XORL    T1, y1;                      \ // y1 = (a>>22) ^ (a>>13)                // S0
+       RORXL   $2, a, T1;                   \ // T1 = (a >> 2)                         // S0
+       VPADDD  XTMP4, XTMP0, XTMP0;         \ // XTMP0 = {..., ..., W[1], W[0]}
+       ;                                    \
+       XORL    T1, y1;                      \ // y1 = (a>>22) ^ (a>>13) ^ (a>>2)       // S0
+       MOVL    a, T1;                       \ // T1 = a                                // MAJB
+       ANDL    c, T1;                       \ // T1 = a&c                              // MAJB
+       ADDL    y0, y2;                      \ // y2 = S1 + CH                          // --
+       VPSHUFD $80, XTMP0, XTMP2;           \ // XTMP2 = W[-2] {DDCC}
+       ;                                    \
+       ORL     T1, y3;                      \ // y3 = MAJ = (a|c)&b)|(a&c)             // MAJ
+       ADDL    y1, h;                       \ // h = k + w + h + S0                    // --
+       ADDL    y2, d;                       \ // d = k + w + h + d + S1 + CH = d + t1  // --
+       ADDL    y2, h;                       \ // h = k + w + h + S0 + S1 + CH = t1 + S0// --
+       ;                                    \
+       ADDL    y3, h                        // h = t1 + S0 + MAJ                     // --
+
+#define ROUND_AND_SCHED_N_3(disp, a, b, c, d, e, f, g, h, XDWORD0, XDWORD1, XDWORD2, XDWORD3) \
+       ;                                    \ // ################################### RND N + 3 ############################
+       ;                                    \
+       MOVL    a, y3;                       \ // y3 = a                                                // MAJA
+       RORXL   $25, e, y0;                  \ // y0 = e >> 25                                  // S1A
+       RORXL   $11, e, y1;                  \ // y1 = e >> 11                                  // S1B
+       ADDL    (disp + 3*4)(SP)(SRND*1), h; \ // h = k + w + h                         // --
+       ORL     c, y3;                       \ // y3 = a|c                     // MAJA
+       ;                                    \
+       VPSRLD  $10, XTMP2, XTMP5;           \ // XTMP5 = W[-2] >> 10 {DDCC}
+       MOVL    f, y2;                       \ // y2 = f                                                // CH
+       RORXL   $13, a, T1;                  \ // T1 = a >> 13                                  // S0B
+       XORL    y1, y0;                      \ // y0 = (e>>25) ^ (e>>11)                // S1
+       XORL    g, y2;                       \ // y2 = f^g                                              // CH
+       ;                                    \
+       VPSRLQ  $19, XTMP2, XTMP3;           \ // XTMP3 = W[-2] ror 19 {xDxC}
+       RORXL   $6, e, y1;                   \ // y1 = (e >> 6)                         // S1
+       ANDL    e, y2;                       \ // y2 = (f^g)&e                                  // CH
+       ADDL    h, d;                        \ // d = k + w + h + d                     // --
+       ANDL    b, y3;                       \ // y3 = (a|c)&b                                  // MAJA
+       ;                                    \
+       VPSRLQ  $17, XTMP2, XTMP2;           \ // XTMP2 = W[-2] ror 17 {xDxC}
+       XORL    y1, y0;                      \ // y0 = (e>>25) ^ (e>>11) ^ (e>>6)       // S1
+       XORL    g, y2;                       \ // y2 = CH = ((f^g)&e)^g                 // CH
+       ;                                    \
+       VPXOR   XTMP3, XTMP2, XTMP2;         \
+       RORXL   $22, a, y1;                  \ // y1 = a >> 22                                  // S0A
+       ADDL    y0, y2;                      \ // y2 = S1 + CH                                  // --
+       ;                                    \
+       VPXOR   XTMP2, XTMP5, XTMP5;         \ // XTMP5 = s1 {xDxC}
+       XORL    T1, y1;                      \ // y1 = (a>>22) ^ (a>>13)                // S0
+       ADDL    y2, d;                       \ // d = k + w + h + d + S1 + CH = d + t1  // --
+       ;                                    \
+       RORXL   $2, a, T1;                   \ // T1 = (a >> 2)                         // S0
+       ;                                    \
+       MOVL    f, _TMP(SP);                 \ // Save f
+       MOVQ    $shuff_DC00<>(SB), f;        \ // SHUF_00DC
+       VPSHUFB (f), XTMP5, XTMP5;           \ // XTMP5 = s1 {DC00}
+       MOVL    _TMP(SP), f;                 \ // Restore f
+       ;                                    \
+       VPADDD  XTMP0, XTMP5, XDWORD0;       \ // XDWORD0 = {W[3], W[2], W[1], W[0]}
+       XORL    T1, y1;                      \ // y1 = (a>>22) ^ (a>>13) ^ (a>>2)       // S0
+       MOVL    a, T1;                       \ // T1 = a                                                        // MAJB
+       ANDL    c, T1;                       \ // T1 = a&c                                                      // MAJB
+       ORL     T1, y3;                      \ // y3 = MAJ = (a|c)&b)|(a&c)             // MAJ
+       ;                                    \
+       ADDL    y1, h;                       \ // h = k + w + h + S0                            // --
+       ADDL    y2, h;                       \ // h = k + w + h + S0 + S1 + CH = t1 + S0// --
+       ADDL    y3, h                        // h = t1 + S0 + MAJ                               // --
+
+#define DO_ROUND_N_0(disp, a, b, c, d, e, f, g, h, old_h) \
+       ;                                  \ // ################################### RND N + 0 ###########################
+       MOVL  f, y2;                       \ // y2 = f                                  // CH
+       RORXL $25, e, y0;                  \ // y0 = e >> 25                            // S1A
+       RORXL $11, e, y1;                  \ // y1 = e >> 11                            // S1B
+       XORL  g, y2;                       \ // y2 = f^g                                        // CH
+       ;                                  \
+       XORL  y1, y0;                      \ // y0 = (e>>25) ^ (e>>11)  // S1
+       RORXL $6, e, y1;                   \ // y1 = (e >> 6)                   // S1
+       ANDL  e, y2;                       \ // y2 = (f^g)&e                            // CH
+       ;                                  \
+       XORL  y1, y0;                      \ // y0 = (e>>25) ^ (e>>11) ^ (e>>6) // S1
+       RORXL $13, a, T1;                  \ // T1 = a >> 13                                            // S0B
+       XORL  g, y2;                       \ // y2 = CH = ((f^g)&e)^g                   // CH
+       RORXL $22, a, y1;                  \ // y1 = a >> 22                                            // S0A
+       MOVL  a, y3;                       \ // y3 = a                                                  // MAJA
+       ;                                  \
+       XORL  T1, y1;                      \ // y1 = (a>>22) ^ (a>>13)                  // S0
+       RORXL $2, a, T1;                   \ // T1 = (a >> 2)                                   // S0
+       ADDL  (disp + 0*4)(SP)(SRND*1), h; \ // h = k + w + h // --
+       ORL   c, y3;                       \ // y3 = a|c                                                        // MAJA
+       ;                                  \
+       XORL  T1, y1;                      \ // y1 = (a>>22) ^ (a>>13) ^ (a>>2) // S0
+       MOVL  a, T1;                       \ // T1 = a                                                  // MAJB
+       ANDL  b, y3;                       \ // y3 = (a|c)&b                                            // MAJA
+       ANDL  c, T1;                       \ // T1 = a&c                                                        // MAJB
+       ADDL  y0, y2;                      \ // y2 = S1 + CH                                            // --
+       ;                                  \
+       ADDL  h, d;                        \ // d = k + w + h + d                                       // --
+       ORL   T1, y3;                      \ // y3 = MAJ = (a|c)&b)|(a&c)                       // MAJ
+       ADDL  y1, h;                       \ // h = k + w + h + S0                                      // --
+       ADDL  y2, d                        // d = k + w + h + d + S1 + CH = d + t1      // --
+
+#define DO_ROUND_N_1(disp, a, b, c, d, e, f, g, h, old_h) \
+       ;                                  \ // ################################### RND N + 1 ###########################
+       ADDL  y2, old_h;                   \ // h = k + w + h + S0 + S1 + CH = t1 + S0 // --
+       MOVL  f, y2;                       \ // y2 = f                                // CH
+       RORXL $25, e, y0;                  \ // y0 = e >> 25                            // S1A
+       RORXL $11, e, y1;                  \ // y1 = e >> 11                            // S1B
+       XORL  g, y2;                       \ // y2 = f^g                             // CH
+       ;                                  \
+       XORL  y1, y0;                      \ // y0 = (e>>25) ^ (e>>11)                          // S1
+       RORXL $6, e, y1;                   \ // y1 = (e >> 6)                                           // S1
+       ANDL  e, y2;                       \ // y2 = (f^g)&e                         // CH
+       ADDL  y3, old_h;                   \ // h = t1 + S0 + MAJ                    // --
+       ;                                  \
+       XORL  y1, y0;                      \ // y0 = (e>>25) ^ (e>>11) ^ (e>>6)         // S1
+       RORXL $13, a, T1;                  \ // T1 = a >> 13                                                    // S0B
+       XORL  g, y2;                       \ // y2 = CH = ((f^g)&e)^g                // CH
+       RORXL $22, a, y1;                  \ // y1 = a >> 22                                                    // S0A
+       MOVL  a, y3;                       \ // y3 = a                               // MAJA
+       ;                                  \
+       XORL  T1, y1;                      \ // y1 = (a>>22) ^ (a>>13)                          // S0
+       RORXL $2, a, T1;                   \ // T1 = (a >> 2)                                           // S0
+       ADDL  (disp + 1*4)(SP)(SRND*1), h; \ // h = k + w + h // --
+       ORL   c, y3;                       \ // y3 = a|c                             // MAJA
+       ;                                  \
+       XORL  T1, y1;                      \ // y1 = (a>>22) ^ (a>>13) ^ (a>>2)         // S0
+       MOVL  a, T1;                       \ // T1 = a                               // MAJB
+       ANDL  b, y3;                       \ // y3 = (a|c)&b                         // MAJA
+       ANDL  c, T1;                       \ // T1 = a&c                             // MAJB
+       ADDL  y0, y2;                      \ // y2 = S1 + CH                         // --
+       ;                                  \
+       ADDL  h, d;                        \ // d = k + w + h + d                    // --
+       ORL   T1, y3;                      \ // y3 = MAJ = (a|c)&b)|(a&c)            // MAJ
+       ADDL  y1, h;                       \ // h = k + w + h + S0                   // --
+       ;                                  \
+       ADDL  y2, d                        // d = k + w + h + d + S1 + CH = d + t1 // --
+
+#define DO_ROUND_N_2(disp, a, b, c, d, e, f, g, h, old_h) \
+       ;                                  \ // ################################### RND N + 2 ##############################
+       ADDL  y2, old_h;                   \ // h = k + w + h + S0 + S1 + CH = t1 + S0// --
+       MOVL  f, y2;                       \ // y2 = f                                                          // CH
+       RORXL $25, e, y0;                  \ // y0 = e >> 25                                                    // S1A
+       RORXL $11, e, y1;                  \ // y1 = e >> 11                                                    // S1B
+       XORL  g, y2;                       \ // y2 = f^g                                                                // CH
+       ;                                  \
+       XORL  y1, y0;                      \ // y0 = (e>>25) ^ (e>>11)                          // S1
+       RORXL $6, e, y1;                   \ // y1 = (e >> 6)                                           // S1
+       ANDL  e, y2;                       \ // y2 = (f^g)&e                                                    // CH
+       ADDL  y3, old_h;                   \ // h = t1 + S0 + MAJ                                       // --
+       ;                                  \
+       XORL  y1, y0;                      \ // y0 = (e>>25) ^ (e>>11) ^ (e>>6)         // S1
+       RORXL $13, a, T1;                  \ // T1 = a >> 13                                                    // S0B
+       XORL  g, y2;                       \ // y2 = CH = ((f^g)&e)^g                // CH
+       RORXL $22, a, y1;                  \ // y1 = a >> 22                                                    // S0A
+       MOVL  a, y3;                       \ // y3 = a                                                          // MAJA
+       ;                                  \
+       XORL  T1, y1;                      \ // y1 = (a>>22) ^ (a>>13)                          // S0
+       RORXL $2, a, T1;                   \ // T1 = (a >> 2)                                           // S0
+       ADDL  (disp + 2*4)(SP)(SRND*1), h; \ // h = k + w + h   // --
+       ORL   c, y3;                       \ // y3 = a|c                                                                // MAJA
+       ;                                  \
+       XORL  T1, y1;                      \ // y1 = (a>>22) ^ (a>>13) ^ (a>>2)         // S0
+       MOVL  a, T1;                       \ // T1 = a                                                          // MAJB
+       ANDL  b, y3;                       \ // y3 = (a|c)&b                                                    // MAJA
+       ANDL  c, T1;                       \ // T1 = a&c                                                                // MAJB
+       ADDL  y0, y2;                      \ // y2 = S1 + CH                                                    // --
+       ;                                  \
+       ADDL  h, d;                        \ // d = k + w + h + d                                       // --
+       ORL   T1, y3;                      \ // y3 = MAJ = (a|c)&b)|(a&c)                       // MAJ
+       ADDL  y1, h;                       \ // h = k + w + h + S0                                      // --
+       ;                                  \
+       ADDL  y2, d                        // d = k + w + h + d + S1 + CH = d + t1 // --
+
+#define DO_ROUND_N_3(disp, a, b, c, d, e, f, g, h, old_h) \
+       ;                                  \ // ################################### RND N + 3 ###########################
+       ADDL  y2, old_h;                   \ // h = k + w + h + S0 + S1 + CH = t1 + S0// --
+       MOVL  f, y2;                       \ // y2 = f                                                          // CH
+       RORXL $25, e, y0;                  \ // y0 = e >> 25                                                    // S1A
+       RORXL $11, e, y1;                  \ // y1 = e >> 11                                                    // S1B
+       XORL  g, y2;                       \ // y2 = f^g                                                                // CH
+       ;                                  \
+       XORL  y1, y0;                      \ // y0 = (e>>25) ^ (e>>11)                          // S1
+       RORXL $6, e, y1;                   \ // y1 = (e >> 6)                                           // S1
+       ANDL  e, y2;                       \ // y2 = (f^g)&e                                                    // CH
+       ADDL  y3, old_h;                   \ // h = t1 + S0 + MAJ                                       // --
+       ;                                  \
+       XORL  y1, y0;                      \ // y0 = (e>>25) ^ (e>>11) ^ (e>>6)         // S1
+       RORXL $13, a, T1;                  \ // T1 = a >> 13                                                    // S0B
+       XORL  g, y2;                       \ // y2 = CH = ((f^g)&e)^g                           // CH
+       RORXL $22, a, y1;                  \ // y1 = a >> 22                                                    // S0A
+       MOVL  a, y3;                       \ // y3 = a                                                          // MAJA
+       ;                                  \
+       XORL  T1, y1;                      \ // y1 = (a>>22) ^ (a>>13)                          // S0
+       RORXL $2, a, T1;                   \ // T1 = (a >> 2)                                           // S0
+       ADDL  (disp + 3*4)(SP)(SRND*1), h; \ // h = k + w + h   // --
+       ORL   c, y3;                       \ // y3 = a|c                                                                // MAJA
+       ;                                  \
+       XORL  T1, y1;                      \ // y1 = (a>>22) ^ (a>>13) ^ (a>>2)         // S0
+       MOVL  a, T1;                       \ // T1 = a                                                          // MAJB
+       ANDL  b, y3;                       \ // y3 = (a|c)&b                                                    // MAJA
+       ANDL  c, T1;                       \ // T1 = a&c                                                                // MAJB
+       ADDL  y0, y2;                      \ // y2 = S1 + CH                                                    // --
+       ;                                  \
+       ADDL  h, d;                        \ // d = k + w + h + d                                       // --
+       ORL   T1, y3;                      \ // y3 = MAJ = (a|c)&b)|(a&c)                       // MAJ
+       ADDL  y1, h;                       \ // h = k + w + h + S0                                      // --
+       ;                                  \
+       ADDL  y2, d;                       \ // d = k + w + h + d + S1 + CH = d + t1    // --
+       ;                                  \
+       ADDL  y2, h;                       \ // h = k + w + h + S0 + S1 + CH = t1 + S0// --
+       ;                                  \
+       ADDL  y3, h                        // h = t1 + S0 + MAJ                                 // --
+
+TEXT ·block(SB), 0, $536-32
+       CMPB runtime·support_avx2(SB), $1
+       JE   avx2
+
+       MOVQ p_base+8(FP), SI
+       MOVQ p_len+16(FP), DX
+       SHRQ $6, DX
+       SHLQ $6, DX
+
+       LEAQ (SI)(DX*1), DI
+       MOVQ DI, 256(SP)
+       CMPQ SI, DI
+       JEQ  end
+
+       MOVQ dig+0(FP), BP
+       MOVL (0*4)(BP), R8  // a = H0
+       MOVL (1*4)(BP), R9  // b = H1
+       MOVL (2*4)(BP), R10 // c = H2
+       MOVL (3*4)(BP), R11 // d = H3
+       MOVL (4*4)(BP), R12 // e = H4
+       MOVL (5*4)(BP), R13 // f = H5
+       MOVL (6*4)(BP), R14 // g = H6
+       MOVL (7*4)(BP), R15 // h = H7
 
 loop:
-       MOVQ    SP, BP                  // message schedule
+       MOVQ SP, BP
 
        SHA256ROUND0(0, 0x428a2f98, R8, R9, R10, R11, R12, R13, R14, R15)
        SHA256ROUND0(1, 0x71374491, R15, R8, R9, R10, R11, R12, R13, R14)
@@ -230,27 +651,391 @@ loop:
        SHA256ROUND1(62, 0xbef9a3f7, R10, R11, R12, R13, R14, R15, R8, R9)
        SHA256ROUND1(63, 0xc67178f2, R9, R10, R11, R12, R13, R14, R15, R8)
 
-       MOVQ    dig+0(FP), BP
-       ADDL    (0*4)(BP), R8   // H0 = a + H0
-       MOVL    R8, (0*4)(BP)
-       ADDL    (1*4)(BP), R9   // H1 = b + H1
-       MOVL    R9, (1*4)(BP)
-       ADDL    (2*4)(BP), R10  // H2 = c + H2
-       MOVL    R10, (2*4)(BP)
-       ADDL    (3*4)(BP), R11  // H3 = d + H3
-       MOVL    R11, (3*4)(BP)
-       ADDL    (4*4)(BP), R12  // H4 = e + H4
-       MOVL    R12, (4*4)(BP)
-       ADDL    (5*4)(BP), R13  // H5 = f + H5
-       MOVL    R13, (5*4)(BP)
-       ADDL    (6*4)(BP), R14  // H6 = g + H6
-       MOVL    R14, (6*4)(BP)
-       ADDL    (7*4)(BP), R15  // H7 = h + H7
-       MOVL    R15, (7*4)(BP)
-
-       ADDQ    $64, SI
-       CMPQ    SI, 256(SP)
-       JB      loop
+       MOVQ dig+0(FP), BP
+       ADDL (0*4)(BP), R8  // H0 = a + H0
+       MOVL R8, (0*4)(BP)
+       ADDL (1*4)(BP), R9  // H1 = b + H1
+       MOVL R9, (1*4)(BP)
+       ADDL (2*4)(BP), R10 // H2 = c + H2
+       MOVL R10, (2*4)(BP)
+       ADDL (3*4)(BP), R11 // H3 = d + H3
+       MOVL R11, (3*4)(BP)
+       ADDL (4*4)(BP), R12 // H4 = e + H4
+       MOVL R12, (4*4)(BP)
+       ADDL (5*4)(BP), R13 // H5 = f + H5
+       MOVL R13, (5*4)(BP)
+       ADDL (6*4)(BP), R14 // H6 = g + H6
+       MOVL R14, (6*4)(BP)
+       ADDL (7*4)(BP), R15 // H7 = h + H7
+       MOVL R15, (7*4)(BP)
+
+       ADDQ $64, SI
+       CMPQ SI, 256(SP)
+       JB   loop
 
 end:
        RET
+
+avx2:
+       MOVQ dig+0(FP), CTX          // d.h[8]
+       MOVQ p_base+8(FP), INP
+       MOVQ p_len+16(FP), NUM_BYTES
+
+       LEAQ -64(INP)(NUM_BYTES*1), NUM_BYTES // Pointer to the last block
+       MOVQ NUM_BYTES, _INP_END(SP)
+
+       CMPQ NUM_BYTES, INP
+       JE   avx2_only_one_block
+
+       // Load initial digest
+       MOVL 0(CTX), a  // a = H0
+       MOVL 4(CTX), b  // b = H1
+       MOVL 8(CTX), c  // c = H2
+       MOVL 12(CTX), d // d = H3
+       MOVL 16(CTX), e // e = H4
+       MOVL 20(CTX), f // f = H5
+       MOVL 24(CTX), g // g = H6
+       MOVL 28(CTX), h // h = H7
+
+avx2_loop0: // at each iteration works with one block (512 bit)
+
+       VMOVDQU (0*32)(INP), XTMP0
+       VMOVDQU (1*32)(INP), XTMP1
+       VMOVDQU (2*32)(INP), XTMP2
+       VMOVDQU (3*32)(INP), XTMP3
+
+       MOVQ    $flip_mask<>(SB), BP // BYTE_FLIP_MASK
+       VMOVDQU (BP), BYTE_FLIP_MASK
+
+       // Apply Byte Flip Mask: LE -> BE
+       VPSHUFB BYTE_FLIP_MASK, XTMP0, XTMP0
+       VPSHUFB BYTE_FLIP_MASK, XTMP1, XTMP1
+       VPSHUFB BYTE_FLIP_MASK, XTMP2, XTMP2
+       VPSHUFB BYTE_FLIP_MASK, XTMP3, XTMP3
+
+       // Transpose data into high/low parts
+       VPERM2I128 $0x20, XTMP2, XTMP0, XDWORD0 // w3, w2, w1, w0
+       VPERM2I128 $0x31, XTMP2, XTMP0, XDWORD1 // w7, w6, w5, w4
+       VPERM2I128 $0x20, XTMP3, XTMP1, XDWORD2 // w11, w10, w9, w8
+       VPERM2I128 $0x31, XTMP3, XTMP1, XDWORD3 // w15, w14, w13, w12
+
+       MOVQ $K256<>(SB), TBL // Loading address of table with round-specific constants
+
+avx2_last_block_enter:
+       ADDQ $64, INP
+       MOVQ INP, _INP(SP)
+       XORQ SRND, SRND
+
+avx2_loop1: // for w0 - w47
+       // Do 4 rounds and scheduling
+       VPADDD  0*32(TBL)(SRND*1), XDWORD0, XFER
+       VMOVDQU XFER, (_XFER + 0*32)(SP)(SRND*1)
+       ROUND_AND_SCHED_N_0(_XFER + 0*32, a, b, c, d, e, f, g, h, XDWORD0, XDWORD1, XDWORD2, XDWORD3)
+       ROUND_AND_SCHED_N_1(_XFER + 0*32, h, a, b, c, d, e, f, g, XDWORD0, XDWORD1, XDWORD2, XDWORD3)
+       ROUND_AND_SCHED_N_2(_XFER + 0*32, g, h, a, b, c, d, e, f, XDWORD0, XDWORD1, XDWORD2, XDWORD3)
+       ROUND_AND_SCHED_N_3(_XFER + 0*32, f, g, h, a, b, c, d, e, XDWORD0, XDWORD1, XDWORD2, XDWORD3)
+
+       // Do 4 rounds and scheduling
+       VPADDD  1*32(TBL)(SRND*1), XDWORD1, XFER
+       VMOVDQU XFER, (_XFER + 1*32)(SP)(SRND*1)
+       ROUND_AND_SCHED_N_0(_XFER + 1*32, e, f, g, h, a, b, c, d, XDWORD1, XDWORD2, XDWORD3, XDWORD0)
+       ROUND_AND_SCHED_N_1(_XFER + 1*32, d, e, f, g, h, a, b, c, XDWORD1, XDWORD2, XDWORD3, XDWORD0)
+       ROUND_AND_SCHED_N_2(_XFER + 1*32, c, d, e, f, g, h, a, b, XDWORD1, XDWORD2, XDWORD3, XDWORD0)
+       ROUND_AND_SCHED_N_3(_XFER + 1*32, b, c, d, e, f, g, h, a, XDWORD1, XDWORD2, XDWORD3, XDWORD0)
+
+       // Do 4 rounds and scheduling
+       VPADDD  2*32(TBL)(SRND*1), XDWORD2, XFER
+       VMOVDQU XFER, (_XFER + 2*32)(SP)(SRND*1)
+       ROUND_AND_SCHED_N_0(_XFER + 2*32, a, b, c, d, e, f, g, h, XDWORD2, XDWORD3, XDWORD0, XDWORD1)
+       ROUND_AND_SCHED_N_1(_XFER + 2*32, h, a, b, c, d, e, f, g, XDWORD2, XDWORD3, XDWORD0, XDWORD1)
+       ROUND_AND_SCHED_N_2(_XFER + 2*32, g, h, a, b, c, d, e, f, XDWORD2, XDWORD3, XDWORD0, XDWORD1)
+       ROUND_AND_SCHED_N_3(_XFER + 2*32, f, g, h, a, b, c, d, e, XDWORD2, XDWORD3, XDWORD0, XDWORD1)
+
+       // Do 4 rounds and scheduling
+       VPADDD  3*32(TBL)(SRND*1), XDWORD3, XFER
+       VMOVDQU XFER, (_XFER + 3*32)(SP)(SRND*1)
+       ROUND_AND_SCHED_N_0(_XFER + 3*32, e, f, g, h, a, b, c, d, XDWORD3, XDWORD0, XDWORD1, XDWORD2)
+       ROUND_AND_SCHED_N_1(_XFER + 3*32, d, e, f, g, h, a, b, c, XDWORD3, XDWORD0, XDWORD1, XDWORD2)
+       ROUND_AND_SCHED_N_2(_XFER + 3*32, c, d, e, f, g, h, a, b, XDWORD3, XDWORD0, XDWORD1, XDWORD2)
+       ROUND_AND_SCHED_N_3(_XFER + 3*32, b, c, d, e, f, g, h, a, XDWORD3, XDWORD0, XDWORD1, XDWORD2)
+
+       ADDQ $4*32, SRND
+       CMPQ SRND, $3*4*32
+       JB   avx2_loop1
+
+avx2_loop2:
+       // w48 - w63 processed with no scheduliung (last 16 rounds)
+       VPADDD  0*32(TBL)(SRND*1), XDWORD0, XFER
+       VMOVDQU XFER, (_XFER + 0*32)(SP)(SRND*1)
+       DO_ROUND_N_0(_XFER + 0*32, a, b, c, d, e, f, g, h, h)
+       DO_ROUND_N_1(_XFER + 0*32, h, a, b, c, d, e, f, g, h)
+       DO_ROUND_N_2(_XFER + 0*32, g, h, a, b, c, d, e, f, g)
+       DO_ROUND_N_3(_XFER + 0*32, f, g, h, a, b, c, d, e, f)
+
+       VPADDD  1*32(TBL)(SRND*1), XDWORD1, XFER
+       VMOVDQU XFER, (_XFER + 1*32)(SP)(SRND*1)
+       DO_ROUND_N_0(_XFER + 1*32, e, f, g, h, a, b, c, d, e)
+       DO_ROUND_N_1(_XFER + 1*32, d, e, f, g, h, a, b, c, d)
+       DO_ROUND_N_2(_XFER + 1*32, c, d, e, f, g, h, a, b, c)
+       DO_ROUND_N_3(_XFER + 1*32, b, c, d, e, f, g, h, a, b)
+
+       ADDQ $2*32, SRND
+
+       VMOVDQU XDWORD2, XDWORD0
+       VMOVDQU XDWORD3, XDWORD1
+
+       CMPQ SRND, $4*4*32
+       JB   avx2_loop2
+
+       MOVQ dig+0(FP), CTX // d.h[8]
+       MOVQ _INP(SP), INP
+
+       addm(  0(CTX), a)
+       addm(  4(CTX), b)
+       addm(  8(CTX), c)
+       addm( 12(CTX), d)
+       addm( 16(CTX), e)
+       addm( 20(CTX), f)
+       addm( 24(CTX), g)
+       addm( 28(CTX), h)
+
+       CMPQ _INP_END(SP), INP
+       JB   done_hash
+
+       XORQ SRND, SRND
+
+avx2_loop3: // Do second block using previously scheduled results
+       DO_ROUND_N_0(_XFER + 0*32 + 16, a, b, c, d, e, f, g, h, a)
+       DO_ROUND_N_1(_XFER + 0*32 + 16, h, a, b, c, d, e, f, g, h)
+       DO_ROUND_N_2(_XFER + 0*32 + 16, g, h, a, b, c, d, e, f, g)
+       DO_ROUND_N_3(_XFER + 0*32 + 16, f, g, h, a, b, c, d, e, f)
+
+       DO_ROUND_N_0(_XFER + 1*32 + 16, e, f, g, h, a, b, c, d, e)
+       DO_ROUND_N_1(_XFER + 1*32 + 16, d, e, f, g, h, a, b, c, d)
+       DO_ROUND_N_2(_XFER + 1*32 + 16, c, d, e, f, g, h, a, b, c)
+       DO_ROUND_N_3(_XFER + 1*32 + 16, b, c, d, e, f, g, h, a, b)
+
+       ADDQ $2*32, SRND
+       CMPQ SRND, $4*4*32
+       JB   avx2_loop3
+
+       MOVQ dig+0(FP), CTX // d.h[8]
+       MOVQ _INP(SP), INP
+       ADDQ $64, INP
+
+       addm(  0(CTX), a)
+       addm(  4(CTX), b)
+       addm(  8(CTX), c)
+       addm( 12(CTX), d)
+       addm( 16(CTX), e)
+       addm( 20(CTX), f)
+       addm( 24(CTX), g)
+       addm( 28(CTX), h)
+
+       CMPQ _INP_END(SP), INP
+       JA   avx2_loop0
+       JB   done_hash
+
+avx2_do_last_block:
+
+       VMOVDQU 0(INP), XWORD0
+       VMOVDQU 16(INP), XWORD1
+       VMOVDQU 32(INP), XWORD2
+       VMOVDQU 48(INP), XWORD3
+
+       MOVQ    $flip_mask<>(SB), BP
+       VMOVDQU (BP), X_BYTE_FLIP_MASK
+
+       VPSHUFB X_BYTE_FLIP_MASK, XWORD0, XWORD0
+       VPSHUFB X_BYTE_FLIP_MASK, XWORD1, XWORD1
+       VPSHUFB X_BYTE_FLIP_MASK, XWORD2, XWORD2
+       VPSHUFB X_BYTE_FLIP_MASK, XWORD3, XWORD3
+
+       MOVQ $K256<>(SB), TBL
+
+       JMP avx2_last_block_enter
+
+avx2_only_one_block:
+       // Load initial digest
+       MOVL 0(CTX), a  // a = H0
+       MOVL 4(CTX), b  // b = H1
+       MOVL 8(CTX), c  // c = H2
+       MOVL 12(CTX), d // d = H3
+       MOVL 16(CTX), e // e = H4
+       MOVL 20(CTX), f // f = H5
+       MOVL 24(CTX), g // g = H6
+       MOVL 28(CTX), h // h = H7
+
+       JMP avx2_do_last_block
+
+done_hash:
+       VZEROUPPER
+       RET
+
+// shuffle byte order from LE to BE
+DATA flip_mask<>+0x00(SB)/8, $0x0405060700010203
+DATA flip_mask<>+0x08(SB)/8, $0x0c0d0e0f08090a0b
+DATA flip_mask<>+0x10(SB)/8, $0x0405060700010203
+DATA flip_mask<>+0x18(SB)/8, $0x0c0d0e0f08090a0b
+GLOBL flip_mask<>(SB), 8, $32
+
+// shuffle xBxA -> 00BA
+DATA shuff_00BA<>+0x00(SB)/8, $0x0b0a090803020100
+DATA shuff_00BA<>+0x08(SB)/8, $0xFFFFFFFFFFFFFFFF
+DATA shuff_00BA<>+0x10(SB)/8, $0x0b0a090803020100
+DATA shuff_00BA<>+0x18(SB)/8, $0xFFFFFFFFFFFFFFFF
+GLOBL shuff_00BA<>(SB), 8, $32
+
+// shuffle xDxC -> DC00
+DATA shuff_DC00<>+0x00(SB)/8, $0xFFFFFFFFFFFFFFFF
+DATA shuff_DC00<>+0x08(SB)/8, $0x0b0a090803020100
+DATA shuff_DC00<>+0x10(SB)/8, $0xFFFFFFFFFFFFFFFF
+DATA shuff_DC00<>+0x18(SB)/8, $0x0b0a090803020100
+GLOBL shuff_DC00<>(SB), 8, $32
+
+// Round specific constants
+DATA K256<>+0x00(SB)/4, $0x428a2f98 // k1
+DATA K256<>+0x04(SB)/4, $0x71374491 // k2
+DATA K256<>+0x08(SB)/4, $0xb5c0fbcf // k3
+DATA K256<>+0x0c(SB)/4, $0xe9b5dba5 // k4
+DATA K256<>+0x10(SB)/4, $0x428a2f98 // k1
+DATA K256<>+0x14(SB)/4, $0x71374491 // k2
+DATA K256<>+0x18(SB)/4, $0xb5c0fbcf // k3
+DATA K256<>+0x1c(SB)/4, $0xe9b5dba5 // k4
+
+DATA K256<>+0x20(SB)/4, $0x3956c25b // k5 - k8
+DATA K256<>+0x24(SB)/4, $0x59f111f1
+DATA K256<>+0x28(SB)/4, $0x923f82a4
+DATA K256<>+0x2c(SB)/4, $0xab1c5ed5
+DATA K256<>+0x30(SB)/4, $0x3956c25b
+DATA K256<>+0x34(SB)/4, $0x59f111f1
+DATA K256<>+0x38(SB)/4, $0x923f82a4
+DATA K256<>+0x3c(SB)/4, $0xab1c5ed5
+
+DATA K256<>+0x40(SB)/4, $0xd807aa98 // k9 - k12
+DATA K256<>+0x44(SB)/4, $0x12835b01
+DATA K256<>+0x48(SB)/4, $0x243185be
+DATA K256<>+0x4c(SB)/4, $0x550c7dc3
+DATA K256<>+0x50(SB)/4, $0xd807aa98
+DATA K256<>+0x54(SB)/4, $0x12835b01
+DATA K256<>+0x58(SB)/4, $0x243185be
+DATA K256<>+0x5c(SB)/4, $0x550c7dc3
+
+DATA K256<>+0x60(SB)/4, $0x72be5d74 // k13 - k16
+DATA K256<>+0x64(SB)/4, $0x80deb1fe
+DATA K256<>+0x68(SB)/4, $0x9bdc06a7
+DATA K256<>+0x6c(SB)/4, $0xc19bf174
+DATA K256<>+0x70(SB)/4, $0x72be5d74
+DATA K256<>+0x74(SB)/4, $0x80deb1fe
+DATA K256<>+0x78(SB)/4, $0x9bdc06a7
+DATA K256<>+0x7c(SB)/4, $0xc19bf174
+
+DATA K256<>+0x80(SB)/4, $0xe49b69c1 // k17 - k20
+DATA K256<>+0x84(SB)/4, $0xefbe4786
+DATA K256<>+0x88(SB)/4, $0x0fc19dc6
+DATA K256<>+0x8c(SB)/4, $0x240ca1cc
+DATA K256<>+0x90(SB)/4, $0xe49b69c1
+DATA K256<>+0x94(SB)/4, $0xefbe4786
+DATA K256<>+0x98(SB)/4, $0x0fc19dc6
+DATA K256<>+0x9c(SB)/4, $0x240ca1cc
+
+DATA K256<>+0xa0(SB)/4, $0x2de92c6f // k21 - k24
+DATA K256<>+0xa4(SB)/4, $0x4a7484aa
+DATA K256<>+0xa8(SB)/4, $0x5cb0a9dc
+DATA K256<>+0xac(SB)/4, $0x76f988da
+DATA K256<>+0xb0(SB)/4, $0x2de92c6f
+DATA K256<>+0xb4(SB)/4, $0x4a7484aa
+DATA K256<>+0xb8(SB)/4, $0x5cb0a9dc
+DATA K256<>+0xbc(SB)/4, $0x76f988da
+
+DATA K256<>+0xc0(SB)/4, $0x983e5152 // k25 - k28
+DATA K256<>+0xc4(SB)/4, $0xa831c66d
+DATA K256<>+0xc8(SB)/4, $0xb00327c8
+DATA K256<>+0xcc(SB)/4, $0xbf597fc7
+DATA K256<>+0xd0(SB)/4, $0x983e5152
+DATA K256<>+0xd4(SB)/4, $0xa831c66d
+DATA K256<>+0xd8(SB)/4, $0xb00327c8
+DATA K256<>+0xdc(SB)/4, $0xbf597fc7
+
+DATA K256<>+0xe0(SB)/4, $0xc6e00bf3 // k29 - k32
+DATA K256<>+0xe4(SB)/4, $0xd5a79147
+DATA K256<>+0xe8(SB)/4, $0x06ca6351
+DATA K256<>+0xec(SB)/4, $0x14292967
+DATA K256<>+0xf0(SB)/4, $0xc6e00bf3
+DATA K256<>+0xf4(SB)/4, $0xd5a79147
+DATA K256<>+0xf8(SB)/4, $0x06ca6351
+DATA K256<>+0xfc(SB)/4, $0x14292967
+
+DATA K256<>+0x100(SB)/4, $0x27b70a85
+DATA K256<>+0x104(SB)/4, $0x2e1b2138
+DATA K256<>+0x108(SB)/4, $0x4d2c6dfc
+DATA K256<>+0x10c(SB)/4, $0x53380d13
+DATA K256<>+0x110(SB)/4, $0x27b70a85
+DATA K256<>+0x114(SB)/4, $0x2e1b2138
+DATA K256<>+0x118(SB)/4, $0x4d2c6dfc
+DATA K256<>+0x11c(SB)/4, $0x53380d13
+
+DATA K256<>+0x120(SB)/4, $0x650a7354
+DATA K256<>+0x124(SB)/4, $0x766a0abb
+DATA K256<>+0x128(SB)/4, $0x81c2c92e
+DATA K256<>+0x12c(SB)/4, $0x92722c85
+DATA K256<>+0x130(SB)/4, $0x650a7354
+DATA K256<>+0x134(SB)/4, $0x766a0abb
+DATA K256<>+0x138(SB)/4, $0x81c2c92e
+DATA K256<>+0x13c(SB)/4, $0x92722c85
+
+DATA K256<>+0x140(SB)/4, $0xa2bfe8a1
+DATA K256<>+0x144(SB)/4, $0xa81a664b
+DATA K256<>+0x148(SB)/4, $0xc24b8b70
+DATA K256<>+0x14c(SB)/4, $0xc76c51a3
+DATA K256<>+0x150(SB)/4, $0xa2bfe8a1
+DATA K256<>+0x154(SB)/4, $0xa81a664b
+DATA K256<>+0x158(SB)/4, $0xc24b8b70
+DATA K256<>+0x15c(SB)/4, $0xc76c51a3
+
+DATA K256<>+0x160(SB)/4, $0xd192e819
+DATA K256<>+0x164(SB)/4, $0xd6990624
+DATA K256<>+0x168(SB)/4, $0xf40e3585
+DATA K256<>+0x16c(SB)/4, $0x106aa070
+DATA K256<>+0x170(SB)/4, $0xd192e819
+DATA K256<>+0x174(SB)/4, $0xd6990624
+DATA K256<>+0x178(SB)/4, $0xf40e3585
+DATA K256<>+0x17c(SB)/4, $0x106aa070
+
+DATA K256<>+0x180(SB)/4, $0x19a4c116
+DATA K256<>+0x184(SB)/4, $0x1e376c08
+DATA K256<>+0x188(SB)/4, $0x2748774c
+DATA K256<>+0x18c(SB)/4, $0x34b0bcb5
+DATA K256<>+0x190(SB)/4, $0x19a4c116
+DATA K256<>+0x194(SB)/4, $0x1e376c08
+DATA K256<>+0x198(SB)/4, $0x2748774c
+DATA K256<>+0x19c(SB)/4, $0x34b0bcb5
+
+DATA K256<>+0x1a0(SB)/4, $0x391c0cb3
+DATA K256<>+0x1a4(SB)/4, $0x4ed8aa4a
+DATA K256<>+0x1a8(SB)/4, $0x5b9cca4f
+DATA K256<>+0x1ac(SB)/4, $0x682e6ff3
+DATA K256<>+0x1b0(SB)/4, $0x391c0cb3
+DATA K256<>+0x1b4(SB)/4, $0x4ed8aa4a
+DATA K256<>+0x1b8(SB)/4, $0x5b9cca4f
+DATA K256<>+0x1bc(SB)/4, $0x682e6ff3
+
+DATA K256<>+0x1c0(SB)/4, $0x748f82ee
+DATA K256<>+0x1c4(SB)/4, $0x78a5636f
+DATA K256<>+0x1c8(SB)/4, $0x84c87814
+DATA K256<>+0x1cc(SB)/4, $0x8cc70208
+DATA K256<>+0x1d0(SB)/4, $0x748f82ee
+DATA K256<>+0x1d4(SB)/4, $0x78a5636f
+DATA K256<>+0x1d8(SB)/4, $0x84c87814
+DATA K256<>+0x1dc(SB)/4, $0x8cc70208
+
+DATA K256<>+0x1e0(SB)/4, $0x90befffa
+DATA K256<>+0x1e4(SB)/4, $0xa4506ceb
+DATA K256<>+0x1e8(SB)/4, $0xbef9a3f7
+DATA K256<>+0x1ec(SB)/4, $0xc67178f2
+DATA K256<>+0x1f0(SB)/4, $0x90befffa
+DATA K256<>+0x1f4(SB)/4, $0xa4506ceb
+DATA K256<>+0x1f8(SB)/4, $0xbef9a3f7
+DATA K256<>+0x1fc(SB)/4, $0xc67178f2
+
+GLOBL K256<>(SB), (NOPTR + RODATA), $512