Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/lib/crypto/riscv/sm3-riscv64-zvksh-zvkb.S
170891 views
1
/* SPDX-License-Identifier: Apache-2.0 OR BSD-2-Clause */
2
//
3
// This file is dual-licensed, meaning that you can use it under your
4
// choice of either of the following two licenses:
5
//
6
// Copyright 2023 The OpenSSL Project Authors. All Rights Reserved.
7
//
8
// Licensed under the Apache License 2.0 (the "License"). You can obtain
9
// a copy in the file LICENSE in the source distribution or at
10
// https://www.openssl.org/source/license.html
11
//
12
// or
13
//
14
// Copyright (c) 2023, Christoph Müllner <christoph.muellner@vrull.eu>
15
// Copyright (c) 2023, Jerry Shih <jerry.shih@sifive.com>
16
// Copyright 2024 Google LLC
17
// All rights reserved.
18
//
19
// Redistribution and use in source and binary forms, with or without
20
// modification, are permitted provided that the following conditions
21
// are met:
22
// 1. Redistributions of source code must retain the above copyright
23
// notice, this list of conditions and the following disclaimer.
24
// 2. Redistributions in binary form must reproduce the above copyright
25
// notice, this list of conditions and the following disclaimer in the
26
// documentation and/or other materials provided with the distribution.
27
//
28
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
29
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
30
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
31
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
32
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
33
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
34
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
35
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
36
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
37
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
38
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
39
40
// The generated code of this file depends on the following RISC-V extensions:
41
// - RV64I
42
// - RISC-V Vector ('V') with VLEN >= 128
43
// - RISC-V Vector SM3 Secure Hash extension ('Zvksh')
44
// - RISC-V Vector Cryptography Bit-manipulation extension ('Zvkb')
45
46
#include <linux/linkage.h>
47
48
.text
49
.option arch, +zvksh, +zvkb
50
51
#define STATEP a0
52
#define DATA a1
53
#define NUM_BLOCKS a2
54
55
#define STATE v0 // LMUL=2
56
#define PREV_STATE v2 // LMUL=2
57
#define W0 v4 // LMUL=2
58
#define W1 v6 // LMUL=2
59
#define VTMP v8 // LMUL=2
60
61
.macro sm3_8rounds i, w0, w1
62
// Do 4 rounds using W_{0+i}..W_{7+i}.
63
vsm3c.vi STATE, \w0, \i + 0
64
vslidedown.vi VTMP, \w0, 2
65
vsm3c.vi STATE, VTMP, \i + 1
66
67
// Compute W_{4+i}..W_{11+i}.
68
vslidedown.vi VTMP, \w0, 4
69
vslideup.vi VTMP, \w1, 4
70
71
// Do 4 rounds using W_{4+i}..W_{11+i}.
72
vsm3c.vi STATE, VTMP, \i + 2
73
vslidedown.vi VTMP, VTMP, 2
74
vsm3c.vi STATE, VTMP, \i + 3
75
76
.if \i < 28
77
// Compute W_{16+i}..W_{23+i}.
78
vsm3me.vv \w0, \w1, \w0
79
.endif
80
// For the next 8 rounds, w0 and w1 are swapped.
81
.endm
82
83
// void sm3_transform_zvksh_zvkb(struct sm3_block_state *state,
84
// const u8 *data, size_t nblocks);
85
SYM_FUNC_START(sm3_transform_zvksh_zvkb)
86
87
// Load the state and endian-swap each 32-bit word.
88
vsetivli zero, 8, e32, m2, ta, ma
89
vle32.v STATE, (STATEP)
90
vrev8.v STATE, STATE
91
92
.Lnext_block:
93
addi NUM_BLOCKS, NUM_BLOCKS, -1
94
95
// Save the previous state, as it's needed later.
96
vmv.v.v PREV_STATE, STATE
97
98
// Load the next 512-bit message block into W0-W1.
99
vle32.v W0, (DATA)
100
addi DATA, DATA, 32
101
vle32.v W1, (DATA)
102
addi DATA, DATA, 32
103
104
// Do the 64 rounds of SM3.
105
sm3_8rounds 0, W0, W1
106
sm3_8rounds 4, W1, W0
107
sm3_8rounds 8, W0, W1
108
sm3_8rounds 12, W1, W0
109
sm3_8rounds 16, W0, W1
110
sm3_8rounds 20, W1, W0
111
sm3_8rounds 24, W0, W1
112
sm3_8rounds 28, W1, W0
113
114
// XOR in the previous state.
115
vxor.vv STATE, STATE, PREV_STATE
116
117
// Repeat if more blocks remain.
118
bnez NUM_BLOCKS, .Lnext_block
119
120
// Store the new state and return.
121
vrev8.v STATE, STATE
122
vse32.v STATE, (STATEP)
123
ret
124
SYM_FUNC_END(sm3_transform_zvksh_zvkb)
125
126