Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

crypto: sm3 - create SM3 stand-alone library

Stand-alone implementation of the SM3 algorithm. It is designed
to have as little dependencies as possible. In other cases you
should generally use the hash APIs from include/crypto/hash.h.
Especially when hashing large amounts of data as those APIs may
be hw-accelerated. In the new SM3 stand-alone library,
sm3_transform() has also been optimized, instead of simply using
the code in sm3_generic.

Signed-off-by: Tianjia Zhang <tianjia.zhang@linux.alibaba.com>
Reviewed-by: Gilad Ben-Yossef <gilad@benyossef.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>

authored by

Tianjia Zhang and committed by
Herbert Xu
eb90686d 68ce6126

+284
+32
include/crypto/sm3.h
··· 1 + /* SPDX-License-Identifier: GPL-2.0-only */ 1 2 /* 2 3 * Common values for SM3 algorithm 4 + * 5 + * Copyright (C) 2017 ARM Limited or its affiliates. 6 + * Copyright (C) 2017 Gilad Ben-Yossef <gilad@benyossef.com> 7 + * Copyright (C) 2021 Tianjia Zhang <tianjia.zhang@linux.alibaba.com> 3 8 */ 4 9 5 10 #ifndef _CRYPTO_SM3_H ··· 44 39 45 40 extern int crypto_sm3_finup(struct shash_desc *desc, const u8 *data, 46 41 unsigned int len, u8 *hash); 42 + 43 + /* 44 + * Stand-alone implementation of the SM3 algorithm. It is designed to 45 + * have as little dependencies as possible so it can be used in the 46 + * kexec_file purgatory. In other cases you should generally use the 47 + * hash APIs from include/crypto/hash.h. Especially when hashing large 48 + * amounts of data as those APIs may be hw-accelerated. 49 + * 50 + * For details see lib/crypto/sm3.c 51 + */ 52 + 53 + static inline void sm3_init(struct sm3_state *sctx) 54 + { 55 + sctx->state[0] = SM3_IVA; 56 + sctx->state[1] = SM3_IVB; 57 + sctx->state[2] = SM3_IVC; 58 + sctx->state[3] = SM3_IVD; 59 + sctx->state[4] = SM3_IVE; 60 + sctx->state[5] = SM3_IVF; 61 + sctx->state[6] = SM3_IVG; 62 + sctx->state[7] = SM3_IVH; 63 + sctx->count = 0; 64 + } 65 + 66 + void sm3_update(struct sm3_state *sctx, const u8 *data, unsigned int len); 67 + void sm3_final(struct sm3_state *sctx, u8 *out); 68 + 47 69 #endif
+3
lib/crypto/Kconfig
··· 123 123 config CRYPTO_LIB_SHA256 124 124 tristate 125 125 126 + config CRYPTO_LIB_SM3 127 + tristate 128 + 126 129 config CRYPTO_LIB_SM4 127 130 tristate 128 131
+3
lib/crypto/Makefile
··· 37 37 obj-$(CONFIG_CRYPTO_LIB_SHA256) += libsha256.o 38 38 libsha256-y := sha256.o 39 39 40 + obj-$(CONFIG_CRYPTO_LIB_SM3) += libsm3.o 41 + libsm3-y := sm3.o 42 + 40 43 obj-$(CONFIG_CRYPTO_LIB_SM4) += libsm4.o 41 44 libsm4-y := sm4.o 42 45
+246
lib/crypto/sm3.c
··· 1 + /* SPDX-License-Identifier: GPL-2.0-only */ 2 + /* 3 + * SM3 secure hash, as specified by OSCCA GM/T 0004-2012 SM3 and described 4 + * at https://datatracker.ietf.org/doc/html/draft-sca-cfrg-sm3-02 5 + * 6 + * Copyright (C) 2017 ARM Limited or its affiliates. 7 + * Copyright (C) 2017 Gilad Ben-Yossef <gilad@benyossef.com> 8 + * Copyright (C) 2021 Tianjia Zhang <tianjia.zhang@linux.alibaba.com> 9 + */ 10 + 11 + #include <linux/module.h> 12 + #include <asm/unaligned.h> 13 + #include <crypto/sm3.h> 14 + 15 + static const u32 ____cacheline_aligned K[64] = { 16 + 0x79cc4519, 0xf3988a32, 0xe7311465, 0xce6228cb, 17 + 0x9cc45197, 0x3988a32f, 0x7311465e, 0xe6228cbc, 18 + 0xcc451979, 0x988a32f3, 0x311465e7, 0x6228cbce, 19 + 0xc451979c, 0x88a32f39, 0x11465e73, 0x228cbce6, 20 + 0x9d8a7a87, 0x3b14f50f, 0x7629ea1e, 0xec53d43c, 21 + 0xd8a7a879, 0xb14f50f3, 0x629ea1e7, 0xc53d43ce, 22 + 0x8a7a879d, 0x14f50f3b, 0x29ea1e76, 0x53d43cec, 23 + 0xa7a879d8, 0x4f50f3b1, 0x9ea1e762, 0x3d43cec5, 24 + 0x7a879d8a, 0xf50f3b14, 0xea1e7629, 0xd43cec53, 25 + 0xa879d8a7, 0x50f3b14f, 0xa1e7629e, 0x43cec53d, 26 + 0x879d8a7a, 0x0f3b14f5, 0x1e7629ea, 0x3cec53d4, 27 + 0x79d8a7a8, 0xf3b14f50, 0xe7629ea1, 0xcec53d43, 28 + 0x9d8a7a87, 0x3b14f50f, 0x7629ea1e, 0xec53d43c, 29 + 0xd8a7a879, 0xb14f50f3, 0x629ea1e7, 0xc53d43ce, 30 + 0x8a7a879d, 0x14f50f3b, 0x29ea1e76, 0x53d43cec, 31 + 0xa7a879d8, 0x4f50f3b1, 0x9ea1e762, 0x3d43cec5 32 + }; 33 + 34 + /* 35 + * Transform the message X which consists of 16 32-bit-words. See 36 + * GM/T 004-2012 for details. 37 + */ 38 + #define R(i, a, b, c, d, e, f, g, h, t, w1, w2) \ 39 + do { \ 40 + ss1 = rol32((rol32((a), 12) + (e) + (t)), 7); \ 41 + ss2 = ss1 ^ rol32((a), 12); \ 42 + d += FF ## i(a, b, c) + ss2 + ((w1) ^ (w2)); \ 43 + h += GG ## i(e, f, g) + ss1 + (w1); \ 44 + b = rol32((b), 9); \ 45 + f = rol32((f), 19); \ 46 + h = P0((h)); \ 47 + } while (0) 48 + 49 + #define R1(a, b, c, d, e, f, g, h, t, w1, w2) \ 50 + R(1, a, b, c, d, e, f, g, h, t, w1, w2) 51 + #define R2(a, b, c, d, e, f, g, h, t, w1, w2) \ 52 + R(2, a, b, c, d, e, f, g, h, t, w1, w2) 53 + 54 + #define FF1(x, y, z) (x ^ y ^ z) 55 + #define FF2(x, y, z) ((x & y) | (x & z) | (y & z)) 56 + 57 + #define GG1(x, y, z) FF1(x, y, z) 58 + #define GG2(x, y, z) ((x & y) | (~x & z)) 59 + 60 + /* Message expansion */ 61 + #define P0(x) ((x) ^ rol32((x), 9) ^ rol32((x), 17)) 62 + #define P1(x) ((x) ^ rol32((x), 15) ^ rol32((x), 23)) 63 + #define I(i) (W[i] = get_unaligned_be32(data + i * 4)) 64 + #define W1(i) (W[i & 0x0f]) 65 + #define W2(i) (W[i & 0x0f] = \ 66 + P1(W[i & 0x0f] \ 67 + ^ W[(i-9) & 0x0f] \ 68 + ^ rol32(W[(i-3) & 0x0f], 15)) \ 69 + ^ rol32(W[(i-13) & 0x0f], 7) \ 70 + ^ W[(i-6) & 0x0f]) 71 + 72 + static void sm3_transform(struct sm3_state *sctx, u8 const *data, u32 W[16]) 73 + { 74 + u32 a, b, c, d, e, f, g, h, ss1, ss2; 75 + 76 + a = sctx->state[0]; 77 + b = sctx->state[1]; 78 + c = sctx->state[2]; 79 + d = sctx->state[3]; 80 + e = sctx->state[4]; 81 + f = sctx->state[5]; 82 + g = sctx->state[6]; 83 + h = sctx->state[7]; 84 + 85 + R1(a, b, c, d, e, f, g, h, K[0], I(0), I(4)); 86 + R1(d, a, b, c, h, e, f, g, K[1], I(1), I(5)); 87 + R1(c, d, a, b, g, h, e, f, K[2], I(2), I(6)); 88 + R1(b, c, d, a, f, g, h, e, K[3], I(3), I(7)); 89 + R1(a, b, c, d, e, f, g, h, K[4], W1(4), I(8)); 90 + R1(d, a, b, c, h, e, f, g, K[5], W1(5), I(9)); 91 + R1(c, d, a, b, g, h, e, f, K[6], W1(6), I(10)); 92 + R1(b, c, d, a, f, g, h, e, K[7], W1(7), I(11)); 93 + R1(a, b, c, d, e, f, g, h, K[8], W1(8), I(12)); 94 + R1(d, a, b, c, h, e, f, g, K[9], W1(9), I(13)); 95 + R1(c, d, a, b, g, h, e, f, K[10], W1(10), I(14)); 96 + R1(b, c, d, a, f, g, h, e, K[11], W1(11), I(15)); 97 + R1(a, b, c, d, e, f, g, h, K[12], W1(12), W2(16)); 98 + R1(d, a, b, c, h, e, f, g, K[13], W1(13), W2(17)); 99 + R1(c, d, a, b, g, h, e, f, K[14], W1(14), W2(18)); 100 + R1(b, c, d, a, f, g, h, e, K[15], W1(15), W2(19)); 101 + 102 + R2(a, b, c, d, e, f, g, h, K[16], W1(16), W2(20)); 103 + R2(d, a, b, c, h, e, f, g, K[17], W1(17), W2(21)); 104 + R2(c, d, a, b, g, h, e, f, K[18], W1(18), W2(22)); 105 + R2(b, c, d, a, f, g, h, e, K[19], W1(19), W2(23)); 106 + R2(a, b, c, d, e, f, g, h, K[20], W1(20), W2(24)); 107 + R2(d, a, b, c, h, e, f, g, K[21], W1(21), W2(25)); 108 + R2(c, d, a, b, g, h, e, f, K[22], W1(22), W2(26)); 109 + R2(b, c, d, a, f, g, h, e, K[23], W1(23), W2(27)); 110 + R2(a, b, c, d, e, f, g, h, K[24], W1(24), W2(28)); 111 + R2(d, a, b, c, h, e, f, g, K[25], W1(25), W2(29)); 112 + R2(c, d, a, b, g, h, e, f, K[26], W1(26), W2(30)); 113 + R2(b, c, d, a, f, g, h, e, K[27], W1(27), W2(31)); 114 + R2(a, b, c, d, e, f, g, h, K[28], W1(28), W2(32)); 115 + R2(d, a, b, c, h, e, f, g, K[29], W1(29), W2(33)); 116 + R2(c, d, a, b, g, h, e, f, K[30], W1(30), W2(34)); 117 + R2(b, c, d, a, f, g, h, e, K[31], W1(31), W2(35)); 118 + 119 + R2(a, b, c, d, e, f, g, h, K[32], W1(32), W2(36)); 120 + R2(d, a, b, c, h, e, f, g, K[33], W1(33), W2(37)); 121 + R2(c, d, a, b, g, h, e, f, K[34], W1(34), W2(38)); 122 + R2(b, c, d, a, f, g, h, e, K[35], W1(35), W2(39)); 123 + R2(a, b, c, d, e, f, g, h, K[36], W1(36), W2(40)); 124 + R2(d, a, b, c, h, e, f, g, K[37], W1(37), W2(41)); 125 + R2(c, d, a, b, g, h, e, f, K[38], W1(38), W2(42)); 126 + R2(b, c, d, a, f, g, h, e, K[39], W1(39), W2(43)); 127 + R2(a, b, c, d, e, f, g, h, K[40], W1(40), W2(44)); 128 + R2(d, a, b, c, h, e, f, g, K[41], W1(41), W2(45)); 129 + R2(c, d, a, b, g, h, e, f, K[42], W1(42), W2(46)); 130 + R2(b, c, d, a, f, g, h, e, K[43], W1(43), W2(47)); 131 + R2(a, b, c, d, e, f, g, h, K[44], W1(44), W2(48)); 132 + R2(d, a, b, c, h, e, f, g, K[45], W1(45), W2(49)); 133 + R2(c, d, a, b, g, h, e, f, K[46], W1(46), W2(50)); 134 + R2(b, c, d, a, f, g, h, e, K[47], W1(47), W2(51)); 135 + 136 + R2(a, b, c, d, e, f, g, h, K[48], W1(48), W2(52)); 137 + R2(d, a, b, c, h, e, f, g, K[49], W1(49), W2(53)); 138 + R2(c, d, a, b, g, h, e, f, K[50], W1(50), W2(54)); 139 + R2(b, c, d, a, f, g, h, e, K[51], W1(51), W2(55)); 140 + R2(a, b, c, d, e, f, g, h, K[52], W1(52), W2(56)); 141 + R2(d, a, b, c, h, e, f, g, K[53], W1(53), W2(57)); 142 + R2(c, d, a, b, g, h, e, f, K[54], W1(54), W2(58)); 143 + R2(b, c, d, a, f, g, h, e, K[55], W1(55), W2(59)); 144 + R2(a, b, c, d, e, f, g, h, K[56], W1(56), W2(60)); 145 + R2(d, a, b, c, h, e, f, g, K[57], W1(57), W2(61)); 146 + R2(c, d, a, b, g, h, e, f, K[58], W1(58), W2(62)); 147 + R2(b, c, d, a, f, g, h, e, K[59], W1(59), W2(63)); 148 + R2(a, b, c, d, e, f, g, h, K[60], W1(60), W2(64)); 149 + R2(d, a, b, c, h, e, f, g, K[61], W1(61), W2(65)); 150 + R2(c, d, a, b, g, h, e, f, K[62], W1(62), W2(66)); 151 + R2(b, c, d, a, f, g, h, e, K[63], W1(63), W2(67)); 152 + 153 + sctx->state[0] ^= a; 154 + sctx->state[1] ^= b; 155 + sctx->state[2] ^= c; 156 + sctx->state[3] ^= d; 157 + sctx->state[4] ^= e; 158 + sctx->state[5] ^= f; 159 + sctx->state[6] ^= g; 160 + sctx->state[7] ^= h; 161 + } 162 + #undef R 163 + #undef R1 164 + #undef R2 165 + #undef I 166 + #undef W1 167 + #undef W2 168 + 169 + static inline void sm3_block(struct sm3_state *sctx, 170 + u8 const *data, int blocks, u32 W[16]) 171 + { 172 + while (blocks--) { 173 + sm3_transform(sctx, data, W); 174 + data += SM3_BLOCK_SIZE; 175 + } 176 + } 177 + 178 + void sm3_update(struct sm3_state *sctx, const u8 *data, unsigned int len) 179 + { 180 + unsigned int partial = sctx->count % SM3_BLOCK_SIZE; 181 + u32 W[16]; 182 + 183 + sctx->count += len; 184 + 185 + if ((partial + len) >= SM3_BLOCK_SIZE) { 186 + int blocks; 187 + 188 + if (partial) { 189 + int p = SM3_BLOCK_SIZE - partial; 190 + 191 + memcpy(sctx->buffer + partial, data, p); 192 + data += p; 193 + len -= p; 194 + 195 + sm3_block(sctx, sctx->buffer, 1, W); 196 + } 197 + 198 + blocks = len / SM3_BLOCK_SIZE; 199 + len %= SM3_BLOCK_SIZE; 200 + 201 + if (blocks) { 202 + sm3_block(sctx, data, blocks, W); 203 + data += blocks * SM3_BLOCK_SIZE; 204 + } 205 + 206 + memzero_explicit(W, sizeof(W)); 207 + 208 + partial = 0; 209 + } 210 + if (len) 211 + memcpy(sctx->buffer + partial, data, len); 212 + } 213 + EXPORT_SYMBOL_GPL(sm3_update); 214 + 215 + void sm3_final(struct sm3_state *sctx, u8 *out) 216 + { 217 + const int bit_offset = SM3_BLOCK_SIZE - sizeof(u64); 218 + __be64 *bits = (__be64 *)(sctx->buffer + bit_offset); 219 + __be32 *digest = (__be32 *)out; 220 + unsigned int partial = sctx->count % SM3_BLOCK_SIZE; 221 + u32 W[16]; 222 + int i; 223 + 224 + sctx->buffer[partial++] = 0x80; 225 + if (partial > bit_offset) { 226 + memset(sctx->buffer + partial, 0, SM3_BLOCK_SIZE - partial); 227 + partial = 0; 228 + 229 + sm3_block(sctx, sctx->buffer, 1, W); 230 + } 231 + 232 + memset(sctx->buffer + partial, 0, bit_offset - partial); 233 + *bits = cpu_to_be64(sctx->count << 3); 234 + sm3_block(sctx, sctx->buffer, 1, W); 235 + 236 + for (i = 0; i < 8; i++) 237 + put_unaligned_be32(sctx->state[i], digest++); 238 + 239 + /* Zeroize sensitive information. */ 240 + memzero_explicit(W, sizeof(W)); 241 + memzero_explicit(sctx, sizeof(*sctx)); 242 + } 243 + EXPORT_SYMBOL_GPL(sm3_final); 244 + 245 + MODULE_DESCRIPTION("Generic SM3 library"); 246 + MODULE_LICENSE("GPL v2");