/* SPDX-License-Identifier: GPL-2.0-or-later */1/*2* SM3 AVX accelerated transform.3* specified in: https://datatracker.ietf.org/doc/html/draft-sca-cfrg-sm3-024*5* Copyright (C) 2021 Jussi Kivilinna <[email protected]>6* Copyright (C) 2021 Tianjia Zhang <[email protected]>7*/89/* Based on SM3 AES/BMI2 accelerated work by libgcrypt at:10* https://gnupg.org/software/libgcrypt/index.html11*/1213#include <linux/linkage.h>14#include <linux/cfi_types.h>15#include <asm/frame.h>1617/* Context structure */1819#define state_h0 020#define state_h1 421#define state_h2 822#define state_h3 1223#define state_h4 1624#define state_h5 2025#define state_h6 2426#define state_h7 282728/* Constants */2930/* Round constant macros */3132#define K0 2043430169 /* 0x79cc4519 */33#define K1 -208106958 /* 0xf3988a32 */34#define K2 -416213915 /* 0xe7311465 */35#define K3 -832427829 /* 0xce6228cb */36#define K4 -1664855657 /* 0x9cc45197 */37#define K5 965255983 /* 0x3988a32f */38#define K6 1930511966 /* 0x7311465e */39#define K7 -433943364 /* 0xe6228cbc */40#define K8 -867886727 /* 0xcc451979 */41#define K9 -1735773453 /* 0x988a32f3 */42#define K10 823420391 /* 0x311465e7 */43#define K11 1646840782 /* 0x6228cbce */44#define K12 -1001285732 /* 0xc451979c */45#define K13 -2002571463 /* 0x88a32f39 */46#define K14 289824371 /* 0x11465e73 */47#define K15 579648742 /* 0x228cbce6 */48#define K16 -1651869049 /* 0x9d8a7a87 */49#define K17 991229199 /* 0x3b14f50f */50#define K18 1982458398 /* 0x7629ea1e */51#define K19 -330050500 /* 0xec53d43c */52#define K20 -660100999 /* 0xd8a7a879 */53#define K21 -1320201997 /* 0xb14f50f3 */54#define K22 1654563303 /* 0x629ea1e7 */55#define K23 -985840690 /* 0xc53d43ce */56#define K24 -1971681379 /* 0x8a7a879d */57#define K25 351604539 /* 0x14f50f3b */58#define K26 703209078 /* 0x29ea1e76 */59#define K27 1406418156 /* 0x53d43cec */60#define K28 -1482130984 /* 0xa7a879d8 */61#define K29 1330705329 /* 0x4f50f3b1 */62#define K30 -1633556638 /* 0x9ea1e762 */63#define K31 1027854021 /* 0x3d43cec5 */64#define K32 2055708042 /* 0x7a879d8a */65#define K33 -183551212 /* 0xf50f3b14 */66#define K34 -367102423 /* 0xea1e7629 */67#define K35 -734204845 /* 0xd43cec53 */68#define K36 -1468409689 /* 0xa879d8a7 */69#define K37 1358147919 /* 0x50f3b14f */70#define K38 -1578671458 /* 0xa1e7629e */71#define K39 1137624381 /* 0x43cec53d */72#define K40 -2019718534 /* 0x879d8a7a */73#define K41 255530229 /* 0x0f3b14f5 */74#define K42 511060458 /* 0x1e7629ea */75#define K43 1022120916 /* 0x3cec53d4 */76#define K44 2044241832 /* 0x79d8a7a8 */77#define K45 -206483632 /* 0xf3b14f50 */78#define K46 -412967263 /* 0xe7629ea1 */79#define K47 -825934525 /* 0xcec53d43 */80#define K48 -1651869049 /* 0x9d8a7a87 */81#define K49 991229199 /* 0x3b14f50f */82#define K50 1982458398 /* 0x7629ea1e */83#define K51 -330050500 /* 0xec53d43c */84#define K52 -660100999 /* 0xd8a7a879 */85#define K53 -1320201997 /* 0xb14f50f3 */86#define K54 1654563303 /* 0x629ea1e7 */87#define K55 -985840690 /* 0xc53d43ce */88#define K56 -1971681379 /* 0x8a7a879d */89#define K57 351604539 /* 0x14f50f3b */90#define K58 703209078 /* 0x29ea1e76 */91#define K59 1406418156 /* 0x53d43cec */92#define K60 -1482130984 /* 0xa7a879d8 */93#define K61 1330705329 /* 0x4f50f3b1 */94#define K62 -1633556638 /* 0x9ea1e762 */95#define K63 1027854021 /* 0x3d43cec5 */9697/* Register macros */9899#define RSTATE %rdi100#define RDATA %rsi101#define RNBLKS %rdx102103#define t0 %eax104#define t1 %ebx105#define t2 %ecx106107#define a %r8d108#define b %r9d109#define c %r10d110#define d %r11d111#define e %r12d112#define f %r13d113#define g %r14d114#define h %r15d115116#define W0 %xmm0117#define W1 %xmm1118#define W2 %xmm2119#define W3 %xmm3120#define W4 %xmm4121#define W5 %xmm5122123#define XTMP0 %xmm6124#define XTMP1 %xmm7125#define XTMP2 %xmm8126#define XTMP3 %xmm9127#define XTMP4 %xmm10128#define XTMP5 %xmm11129#define XTMP6 %xmm12130131#define BSWAP_REG %xmm15132133/* Stack structure */134135#define STACK_W_SIZE (32 * 2 * 3)136#define STACK_REG_SAVE_SIZE (64)137138#define STACK_W (0)139#define STACK_REG_SAVE (STACK_W + STACK_W_SIZE)140#define STACK_SIZE (STACK_REG_SAVE + STACK_REG_SAVE_SIZE)141142/* Instruction helpers. */143144#define roll2(v, reg) \145roll $(v), reg;146147#define roll3mov(v, src, dst) \148movl src, dst; \149roll $(v), dst;150151#define roll3(v, src, dst) \152rorxl $(32-(v)), src, dst;153154#define addl2(a, out) \155leal (a, out), out;156157/* Round function macros. */158159#define GG1(x, y, z, o, t) \160movl x, o; \161xorl y, o; \162xorl z, o;163164#define FF1(x, y, z, o, t) GG1(x, y, z, o, t)165166#define GG2(x, y, z, o, t) \167andnl z, x, o; \168movl y, t; \169andl x, t; \170addl2(t, o);171172#define FF2(x, y, z, o, t) \173movl y, o; \174xorl x, o; \175movl y, t; \176andl x, t; \177andl z, o; \178xorl t, o;179180#define R(i, a, b, c, d, e, f, g, h, round, widx, wtype) \181/* rol(a, 12) => t0 */ \182roll3mov(12, a, t0); /* rorxl here would reduce perf by 6% on zen3 */ \183/* rol (t0 + e + t), 7) => t1 */ \184leal K##round(t0, e, 1), t1; \185roll2(7, t1); \186/* h + w1 => h */ \187addl wtype##_W1_ADDR(round, widx), h; \188/* h + t1 => h */ \189addl2(t1, h); \190/* t1 ^ t0 => t0 */ \191xorl t1, t0; \192/* w1w2 + d => d */ \193addl wtype##_W1W2_ADDR(round, widx), d; \194/* FF##i(a,b,c) => t1 */ \195FF##i(a, b, c, t1, t2); \196/* d + t1 => d */ \197addl2(t1, d); \198/* GG#i(e,f,g) => t2 */ \199GG##i(e, f, g, t2, t1); \200/* h + t2 => h */ \201addl2(t2, h); \202/* rol (f, 19) => f */ \203roll2(19, f); \204/* d + t0 => d */ \205addl2(t0, d); \206/* rol (b, 9) => b */ \207roll2(9, b); \208/* P0(h) => h */ \209roll3(9, h, t2); \210roll3(17, h, t1); \211xorl t2, h; \212xorl t1, h;213214#define R1(a, b, c, d, e, f, g, h, round, widx, wtype) \215R(1, a, b, c, d, e, f, g, h, round, widx, wtype)216217#define R2(a, b, c, d, e, f, g, h, round, widx, wtype) \218R(2, a, b, c, d, e, f, g, h, round, widx, wtype)219220/* Input expansion macros. */221222/* Byte-swapped input address. */223#define IW_W_ADDR(round, widx, offs) \224(STACK_W + ((round) / 4) * 64 + (offs) + ((widx) * 4))(%rsp)225226/* Expanded input address. */227#define XW_W_ADDR(round, widx, offs) \228(STACK_W + ((((round) / 3) - 4) % 2) * 64 + (offs) + ((widx) * 4))(%rsp)229230/* Rounds 1-12, byte-swapped input block addresses. */231#define IW_W1_ADDR(round, widx) IW_W_ADDR(round, widx, 0)232#define IW_W1W2_ADDR(round, widx) IW_W_ADDR(round, widx, 32)233234/* Rounds 1-12, expanded input block addresses. */235#define XW_W1_ADDR(round, widx) XW_W_ADDR(round, widx, 0)236#define XW_W1W2_ADDR(round, widx) XW_W_ADDR(round, widx, 32)237238/* Input block loading. */239#define LOAD_W_XMM_1() \240vmovdqu 0*16(RDATA), XTMP0; /* XTMP0: w3, w2, w1, w0 */ \241vmovdqu 1*16(RDATA), XTMP1; /* XTMP1: w7, w6, w5, w4 */ \242vmovdqu 2*16(RDATA), XTMP2; /* XTMP2: w11, w10, w9, w8 */ \243vmovdqu 3*16(RDATA), XTMP3; /* XTMP3: w15, w14, w13, w12 */ \244vpshufb BSWAP_REG, XTMP0, XTMP0; \245vpshufb BSWAP_REG, XTMP1, XTMP1; \246vpshufb BSWAP_REG, XTMP2, XTMP2; \247vpshufb BSWAP_REG, XTMP3, XTMP3; \248vpxor XTMP0, XTMP1, XTMP4; \249vpxor XTMP1, XTMP2, XTMP5; \250vpxor XTMP2, XTMP3, XTMP6; \251leaq 64(RDATA), RDATA; \252vmovdqa XTMP0, IW_W1_ADDR(0, 0); \253vmovdqa XTMP4, IW_W1W2_ADDR(0, 0); \254vmovdqa XTMP1, IW_W1_ADDR(4, 0); \255vmovdqa XTMP5, IW_W1W2_ADDR(4, 0);256257#define LOAD_W_XMM_2() \258vmovdqa XTMP2, IW_W1_ADDR(8, 0); \259vmovdqa XTMP6, IW_W1W2_ADDR(8, 0);260261#define LOAD_W_XMM_3() \262vpshufd $0b00000000, XTMP0, W0; /* W0: xx, w0, xx, xx */ \263vpshufd $0b11111001, XTMP0, W1; /* W1: xx, w3, w2, w1 */ \264vmovdqa XTMP1, W2; /* W2: xx, w6, w5, w4 */ \265vpalignr $12, XTMP1, XTMP2, W3; /* W3: xx, w9, w8, w7 */ \266vpalignr $8, XTMP2, XTMP3, W4; /* W4: xx, w12, w11, w10 */ \267vpshufd $0b11111001, XTMP3, W5; /* W5: xx, w15, w14, w13 */268269/* Message scheduling. Note: 3 words per XMM register. */270#define SCHED_W_0(round, w0, w1, w2, w3, w4, w5) \271/* Load (w[i - 16]) => XTMP0 */ \272vpshufd $0b10111111, w0, XTMP0; \273vpalignr $12, XTMP0, w1, XTMP0; /* XTMP0: xx, w2, w1, w0 */ \274/* Load (w[i - 13]) => XTMP1 */ \275vpshufd $0b10111111, w1, XTMP1; \276vpalignr $12, XTMP1, w2, XTMP1; \277/* w[i - 9] == w3 */ \278/* XMM3 ^ XTMP0 => XTMP0 */ \279vpxor w3, XTMP0, XTMP0;280281#define SCHED_W_1(round, w0, w1, w2, w3, w4, w5) \282/* w[i - 3] == w5 */ \283/* rol(XMM5, 15) ^ XTMP0 => XTMP0 */ \284vpslld $15, w5, XTMP2; \285vpsrld $(32-15), w5, XTMP3; \286vpxor XTMP2, XTMP3, XTMP3; \287vpxor XTMP3, XTMP0, XTMP0; \288/* rol(XTMP1, 7) => XTMP1 */ \289vpslld $7, XTMP1, XTMP5; \290vpsrld $(32-7), XTMP1, XTMP1; \291vpxor XTMP5, XTMP1, XTMP1; \292/* XMM4 ^ XTMP1 => XTMP1 */ \293vpxor w4, XTMP1, XTMP1; \294/* w[i - 6] == XMM4 */ \295/* P1(XTMP0) ^ XTMP1 => XMM0 */ \296vpslld $15, XTMP0, XTMP5; \297vpsrld $(32-15), XTMP0, XTMP6; \298vpslld $23, XTMP0, XTMP2; \299vpsrld $(32-23), XTMP0, XTMP3; \300vpxor XTMP0, XTMP1, XTMP1; \301vpxor XTMP6, XTMP5, XTMP5; \302vpxor XTMP3, XTMP2, XTMP2; \303vpxor XTMP2, XTMP5, XTMP5; \304vpxor XTMP5, XTMP1, w0;305306#define SCHED_W_2(round, w0, w1, w2, w3, w4, w5) \307/* W1 in XMM12 */ \308vpshufd $0b10111111, w4, XTMP4; \309vpalignr $12, XTMP4, w5, XTMP4; \310vmovdqa XTMP4, XW_W1_ADDR((round), 0); \311/* W1 ^ W2 => XTMP1 */ \312vpxor w0, XTMP4, XTMP1; \313vmovdqa XTMP1, XW_W1W2_ADDR((round), 0);314315316.section .rodata.cst16, "aM", @progbits, 16317.align 16318319.Lbe32mask:320.long 0x00010203, 0x04050607, 0x08090a0b, 0x0c0d0e0f321322.text323324/*325* Transform nblocks*64 bytes (nblocks*16 32-bit words) at DATA.326*327* void sm3_transform_avx(struct sm3_state *state,328* const u8 *data, int nblocks);329*/330SYM_TYPED_FUNC_START(sm3_transform_avx)331/* input:332* %rdi: ctx, CTX333* %rsi: data (64*nblks bytes)334* %rdx: nblocks335*/336vzeroupper;337338pushq %rbp;339movq %rsp, %rbp;340341movq %rdx, RNBLKS;342343subq $STACK_SIZE, %rsp;344andq $(~63), %rsp;345346movq %rbx, (STACK_REG_SAVE + 0 * 8)(%rsp);347movq %r15, (STACK_REG_SAVE + 1 * 8)(%rsp);348movq %r14, (STACK_REG_SAVE + 2 * 8)(%rsp);349movq %r13, (STACK_REG_SAVE + 3 * 8)(%rsp);350movq %r12, (STACK_REG_SAVE + 4 * 8)(%rsp);351352vmovdqa .Lbe32mask (%rip), BSWAP_REG;353354/* Get the values of the chaining variables. */355movl state_h0(RSTATE), a;356movl state_h1(RSTATE), b;357movl state_h2(RSTATE), c;358movl state_h3(RSTATE), d;359movl state_h4(RSTATE), e;360movl state_h5(RSTATE), f;361movl state_h6(RSTATE), g;362movl state_h7(RSTATE), h;363364.align 16365.Loop:366/* Load data part1. */367LOAD_W_XMM_1();368369leaq -1(RNBLKS), RNBLKS;370371/* Transform 0-3 + Load data part2. */372R1(a, b, c, d, e, f, g, h, 0, 0, IW); LOAD_W_XMM_2();373R1(d, a, b, c, h, e, f, g, 1, 1, IW);374R1(c, d, a, b, g, h, e, f, 2, 2, IW);375R1(b, c, d, a, f, g, h, e, 3, 3, IW); LOAD_W_XMM_3();376377/* Transform 4-7 + Precalc 12-14. */378R1(a, b, c, d, e, f, g, h, 4, 0, IW);379R1(d, a, b, c, h, e, f, g, 5, 1, IW);380R1(c, d, a, b, g, h, e, f, 6, 2, IW); SCHED_W_0(12, W0, W1, W2, W3, W4, W5);381R1(b, c, d, a, f, g, h, e, 7, 3, IW); SCHED_W_1(12, W0, W1, W2, W3, W4, W5);382383/* Transform 8-11 + Precalc 12-17. */384R1(a, b, c, d, e, f, g, h, 8, 0, IW); SCHED_W_2(12, W0, W1, W2, W3, W4, W5);385R1(d, a, b, c, h, e, f, g, 9, 1, IW); SCHED_W_0(15, W1, W2, W3, W4, W5, W0);386R1(c, d, a, b, g, h, e, f, 10, 2, IW); SCHED_W_1(15, W1, W2, W3, W4, W5, W0);387R1(b, c, d, a, f, g, h, e, 11, 3, IW); SCHED_W_2(15, W1, W2, W3, W4, W5, W0);388389/* Transform 12-14 + Precalc 18-20 */390R1(a, b, c, d, e, f, g, h, 12, 0, XW); SCHED_W_0(18, W2, W3, W4, W5, W0, W1);391R1(d, a, b, c, h, e, f, g, 13, 1, XW); SCHED_W_1(18, W2, W3, W4, W5, W0, W1);392R1(c, d, a, b, g, h, e, f, 14, 2, XW); SCHED_W_2(18, W2, W3, W4, W5, W0, W1);393394/* Transform 15-17 + Precalc 21-23 */395R1(b, c, d, a, f, g, h, e, 15, 0, XW); SCHED_W_0(21, W3, W4, W5, W0, W1, W2);396R2(a, b, c, d, e, f, g, h, 16, 1, XW); SCHED_W_1(21, W3, W4, W5, W0, W1, W2);397R2(d, a, b, c, h, e, f, g, 17, 2, XW); SCHED_W_2(21, W3, W4, W5, W0, W1, W2);398399/* Transform 18-20 + Precalc 24-26 */400R2(c, d, a, b, g, h, e, f, 18, 0, XW); SCHED_W_0(24, W4, W5, W0, W1, W2, W3);401R2(b, c, d, a, f, g, h, e, 19, 1, XW); SCHED_W_1(24, W4, W5, W0, W1, W2, W3);402R2(a, b, c, d, e, f, g, h, 20, 2, XW); SCHED_W_2(24, W4, W5, W0, W1, W2, W3);403404/* Transform 21-23 + Precalc 27-29 */405R2(d, a, b, c, h, e, f, g, 21, 0, XW); SCHED_W_0(27, W5, W0, W1, W2, W3, W4);406R2(c, d, a, b, g, h, e, f, 22, 1, XW); SCHED_W_1(27, W5, W0, W1, W2, W3, W4);407R2(b, c, d, a, f, g, h, e, 23, 2, XW); SCHED_W_2(27, W5, W0, W1, W2, W3, W4);408409/* Transform 24-26 + Precalc 30-32 */410R2(a, b, c, d, e, f, g, h, 24, 0, XW); SCHED_W_0(30, W0, W1, W2, W3, W4, W5);411R2(d, a, b, c, h, e, f, g, 25, 1, XW); SCHED_W_1(30, W0, W1, W2, W3, W4, W5);412R2(c, d, a, b, g, h, e, f, 26, 2, XW); SCHED_W_2(30, W0, W1, W2, W3, W4, W5);413414/* Transform 27-29 + Precalc 33-35 */415R2(b, c, d, a, f, g, h, e, 27, 0, XW); SCHED_W_0(33, W1, W2, W3, W4, W5, W0);416R2(a, b, c, d, e, f, g, h, 28, 1, XW); SCHED_W_1(33, W1, W2, W3, W4, W5, W0);417R2(d, a, b, c, h, e, f, g, 29, 2, XW); SCHED_W_2(33, W1, W2, W3, W4, W5, W0);418419/* Transform 30-32 + Precalc 36-38 */420R2(c, d, a, b, g, h, e, f, 30, 0, XW); SCHED_W_0(36, W2, W3, W4, W5, W0, W1);421R2(b, c, d, a, f, g, h, e, 31, 1, XW); SCHED_W_1(36, W2, W3, W4, W5, W0, W1);422R2(a, b, c, d, e, f, g, h, 32, 2, XW); SCHED_W_2(36, W2, W3, W4, W5, W0, W1);423424/* Transform 33-35 + Precalc 39-41 */425R2(d, a, b, c, h, e, f, g, 33, 0, XW); SCHED_W_0(39, W3, W4, W5, W0, W1, W2);426R2(c, d, a, b, g, h, e, f, 34, 1, XW); SCHED_W_1(39, W3, W4, W5, W0, W1, W2);427R2(b, c, d, a, f, g, h, e, 35, 2, XW); SCHED_W_2(39, W3, W4, W5, W0, W1, W2);428429/* Transform 36-38 + Precalc 42-44 */430R2(a, b, c, d, e, f, g, h, 36, 0, XW); SCHED_W_0(42, W4, W5, W0, W1, W2, W3);431R2(d, a, b, c, h, e, f, g, 37, 1, XW); SCHED_W_1(42, W4, W5, W0, W1, W2, W3);432R2(c, d, a, b, g, h, e, f, 38, 2, XW); SCHED_W_2(42, W4, W5, W0, W1, W2, W3);433434/* Transform 39-41 + Precalc 45-47 */435R2(b, c, d, a, f, g, h, e, 39, 0, XW); SCHED_W_0(45, W5, W0, W1, W2, W3, W4);436R2(a, b, c, d, e, f, g, h, 40, 1, XW); SCHED_W_1(45, W5, W0, W1, W2, W3, W4);437R2(d, a, b, c, h, e, f, g, 41, 2, XW); SCHED_W_2(45, W5, W0, W1, W2, W3, W4);438439/* Transform 42-44 + Precalc 48-50 */440R2(c, d, a, b, g, h, e, f, 42, 0, XW); SCHED_W_0(48, W0, W1, W2, W3, W4, W5);441R2(b, c, d, a, f, g, h, e, 43, 1, XW); SCHED_W_1(48, W0, W1, W2, W3, W4, W5);442R2(a, b, c, d, e, f, g, h, 44, 2, XW); SCHED_W_2(48, W0, W1, W2, W3, W4, W5);443444/* Transform 45-47 + Precalc 51-53 */445R2(d, a, b, c, h, e, f, g, 45, 0, XW); SCHED_W_0(51, W1, W2, W3, W4, W5, W0);446R2(c, d, a, b, g, h, e, f, 46, 1, XW); SCHED_W_1(51, W1, W2, W3, W4, W5, W0);447R2(b, c, d, a, f, g, h, e, 47, 2, XW); SCHED_W_2(51, W1, W2, W3, W4, W5, W0);448449/* Transform 48-50 + Precalc 54-56 */450R2(a, b, c, d, e, f, g, h, 48, 0, XW); SCHED_W_0(54, W2, W3, W4, W5, W0, W1);451R2(d, a, b, c, h, e, f, g, 49, 1, XW); SCHED_W_1(54, W2, W3, W4, W5, W0, W1);452R2(c, d, a, b, g, h, e, f, 50, 2, XW); SCHED_W_2(54, W2, W3, W4, W5, W0, W1);453454/* Transform 51-53 + Precalc 57-59 */455R2(b, c, d, a, f, g, h, e, 51, 0, XW); SCHED_W_0(57, W3, W4, W5, W0, W1, W2);456R2(a, b, c, d, e, f, g, h, 52, 1, XW); SCHED_W_1(57, W3, W4, W5, W0, W1, W2);457R2(d, a, b, c, h, e, f, g, 53, 2, XW); SCHED_W_2(57, W3, W4, W5, W0, W1, W2);458459/* Transform 54-56 + Precalc 60-62 */460R2(c, d, a, b, g, h, e, f, 54, 0, XW); SCHED_W_0(60, W4, W5, W0, W1, W2, W3);461R2(b, c, d, a, f, g, h, e, 55, 1, XW); SCHED_W_1(60, W4, W5, W0, W1, W2, W3);462R2(a, b, c, d, e, f, g, h, 56, 2, XW); SCHED_W_2(60, W4, W5, W0, W1, W2, W3);463464/* Transform 57-59 + Precalc 63 */465R2(d, a, b, c, h, e, f, g, 57, 0, XW); SCHED_W_0(63, W5, W0, W1, W2, W3, W4);466R2(c, d, a, b, g, h, e, f, 58, 1, XW);467R2(b, c, d, a, f, g, h, e, 59, 2, XW); SCHED_W_1(63, W5, W0, W1, W2, W3, W4);468469/* Transform 60-62 + Precalc 63 */470R2(a, b, c, d, e, f, g, h, 60, 0, XW);471R2(d, a, b, c, h, e, f, g, 61, 1, XW); SCHED_W_2(63, W5, W0, W1, W2, W3, W4);472R2(c, d, a, b, g, h, e, f, 62, 2, XW);473474/* Transform 63 */475R2(b, c, d, a, f, g, h, e, 63, 0, XW);476477/* Update the chaining variables. */478xorl state_h0(RSTATE), a;479xorl state_h1(RSTATE), b;480xorl state_h2(RSTATE), c;481xorl state_h3(RSTATE), d;482movl a, state_h0(RSTATE);483movl b, state_h1(RSTATE);484movl c, state_h2(RSTATE);485movl d, state_h3(RSTATE);486xorl state_h4(RSTATE), e;487xorl state_h5(RSTATE), f;488xorl state_h6(RSTATE), g;489xorl state_h7(RSTATE), h;490movl e, state_h4(RSTATE);491movl f, state_h5(RSTATE);492movl g, state_h6(RSTATE);493movl h, state_h7(RSTATE);494495cmpq $0, RNBLKS;496jne .Loop;497498vzeroall;499500movq (STACK_REG_SAVE + 0 * 8)(%rsp), %rbx;501movq (STACK_REG_SAVE + 1 * 8)(%rsp), %r15;502movq (STACK_REG_SAVE + 2 * 8)(%rsp), %r14;503movq (STACK_REG_SAVE + 3 * 8)(%rsp), %r13;504movq (STACK_REG_SAVE + 4 * 8)(%rsp), %r12;505506vmovdqa %xmm0, IW_W1_ADDR(0, 0);507vmovdqa %xmm0, IW_W1W2_ADDR(0, 0);508vmovdqa %xmm0, IW_W1_ADDR(4, 0);509vmovdqa %xmm0, IW_W1W2_ADDR(4, 0);510vmovdqa %xmm0, IW_W1_ADDR(8, 0);511vmovdqa %xmm0, IW_W1W2_ADDR(8, 0);512513movq %rbp, %rsp;514popq %rbp;515RET;516SYM_FUNC_END(sm3_transform_avx)517518519