1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* 3 * SM3 secure hash, as specified by OSCCA GM/T 0004-2012 SM3 and described 4 * at https://datatracker.ietf.org/doc/html/draft-sca-cfrg-sm3-02 5 * 6 * Copyright (C) 2017 ARM Limited or its affiliates. 7 * Copyright (C) 2017 Gilad Ben-Yossef <gilad@benyossef.com> 8 * Copyright (C) 2021 Tianjia Zhang <tianjia.zhang@linux.alibaba.com> 9 */ 10 11 #include <crypto/sm3.h> 12 #include <linux/export.h> 13 #include <linux/kernel.h> 14 #include <linux/module.h> 15 #include <linux/string.h> 16 #include <linux/unaligned.h> 17 18 static const struct sm3_block_state sm3_iv = { 19 .h = { 20 SM3_IVA, SM3_IVB, SM3_IVC, SM3_IVD, 21 SM3_IVE, SM3_IVF, SM3_IVG, SM3_IVH, 22 }, 23 }; 24 25 static const u32 ____cacheline_aligned K[64] = { 26 0x79cc4519, 0xf3988a32, 0xe7311465, 0xce6228cb, 27 0x9cc45197, 0x3988a32f, 0x7311465e, 0xe6228cbc, 28 0xcc451979, 0x988a32f3, 0x311465e7, 0x6228cbce, 29 0xc451979c, 0x88a32f39, 0x11465e73, 0x228cbce6, 30 0x9d8a7a87, 0x3b14f50f, 0x7629ea1e, 0xec53d43c, 31 0xd8a7a879, 0xb14f50f3, 0x629ea1e7, 0xc53d43ce, 32 0x8a7a879d, 0x14f50f3b, 0x29ea1e76, 0x53d43cec, 33 0xa7a879d8, 0x4f50f3b1, 0x9ea1e762, 0x3d43cec5, 34 0x7a879d8a, 0xf50f3b14, 0xea1e7629, 0xd43cec53, 35 0xa879d8a7, 0x50f3b14f, 0xa1e7629e, 0x43cec53d, 36 0x879d8a7a, 0x0f3b14f5, 0x1e7629ea, 0x3cec53d4, 37 0x79d8a7a8, 0xf3b14f50, 0xe7629ea1, 0xcec53d43, 38 0x9d8a7a87, 0x3b14f50f, 0x7629ea1e, 0xec53d43c, 39 0xd8a7a879, 0xb14f50f3, 0x629ea1e7, 0xc53d43ce, 40 0x8a7a879d, 0x14f50f3b, 0x29ea1e76, 0x53d43cec, 41 0xa7a879d8, 0x4f50f3b1, 0x9ea1e762, 0x3d43cec5 42 }; 43 44 /* 45 * Transform the message X which consists of 16 32-bit-words. See 46 * GM/T 004-2012 for details. 47 */ 48 #define R(i, a, b, c, d, e, f, g, h, t, w1, w2) \ 49 do { \ 50 ss1 = rol32((rol32((a), 12) + (e) + (t)), 7); \ 51 ss2 = ss1 ^ rol32((a), 12); \ 52 d += FF ## i(a, b, c) + ss2 + ((w1) ^ (w2)); \ 53 h += GG ## i(e, f, g) + ss1 + (w1); \ 54 b = rol32((b), 9); \ 55 f = rol32((f), 19); \ 56 h = P0((h)); \ 57 } while (0) 58 59 #define R1(a, b, c, d, e, f, g, h, t, w1, w2) \ 60 R(1, a, b, c, d, e, f, g, h, t, w1, w2) 61 #define R2(a, b, c, d, e, f, g, h, t, w1, w2) \ 62 R(2, a, b, c, d, e, f, g, h, t, w1, w2) 63 64 #define FF1(x, y, z) (x ^ y ^ z) 65 #define FF2(x, y, z) ((x & y) | (x & z) | (y & z)) 66 67 #define GG1(x, y, z) FF1(x, y, z) 68 #define GG2(x, y, z) ((x & y) | (~x & z)) 69 70 /* Message expansion */ 71 #define P0(x) ((x) ^ rol32((x), 9) ^ rol32((x), 17)) 72 #define P1(x) ((x) ^ rol32((x), 15) ^ rol32((x), 23)) 73 #define I(i) (W[i] = get_unaligned_be32(data + i * 4)) 74 #define W1(i) (W[i & 0x0f]) 75 #define W2(i) (W[i & 0x0f] = \ 76 P1(W[i & 0x0f] \ 77 ^ W[(i-9) & 0x0f] \ 78 ^ rol32(W[(i-3) & 0x0f], 15)) \ 79 ^ rol32(W[(i-13) & 0x0f], 7) \ 80 ^ W[(i-6) & 0x0f]) 81 82 static void sm3_block_generic(struct sm3_block_state *state, 83 const u8 data[SM3_BLOCK_SIZE], u32 W[16]) 84 { 85 u32 a, b, c, d, e, f, g, h, ss1, ss2; 86 87 a = state->h[0]; 88 b = state->h[1]; 89 c = state->h[2]; 90 d = state->h[3]; 91 e = state->h[4]; 92 f = state->h[5]; 93 g = state->h[6]; 94 h = state->h[7]; 95 96 R1(a, b, c, d, e, f, g, h, K[0], I(0), I(4)); 97 R1(d, a, b, c, h, e, f, g, K[1], I(1), I(5)); 98 R1(c, d, a, b, g, h, e, f, K[2], I(2), I(6)); 99 R1(b, c, d, a, f, g, h, e, K[3], I(3), I(7)); 100 R1(a, b, c, d, e, f, g, h, K[4], W1(4), I(8)); 101 R1(d, a, b, c, h, e, f, g, K[5], W1(5), I(9)); 102 R1(c, d, a, b, g, h, e, f, K[6], W1(6), I(10)); 103 R1(b, c, d, a, f, g, h, e, K[7], W1(7), I(11)); 104 R1(a, b, c, d, e, f, g, h, K[8], W1(8), I(12)); 105 R1(d, a, b, c, h, e, f, g, K[9], W1(9), I(13)); 106 R1(c, d, a, b, g, h, e, f, K[10], W1(10), I(14)); 107 R1(b, c, d, a, f, g, h, e, K[11], W1(11), I(15)); 108 R1(a, b, c, d, e, f, g, h, K[12], W1(12), W2(16)); 109 R1(d, a, b, c, h, e, f, g, K[13], W1(13), W2(17)); 110 R1(c, d, a, b, g, h, e, f, K[14], W1(14), W2(18)); 111 R1(b, c, d, a, f, g, h, e, K[15], W1(15), W2(19)); 112 113 R2(a, b, c, d, e, f, g, h, K[16], W1(16), W2(20)); 114 R2(d, a, b, c, h, e, f, g, K[17], W1(17), W2(21)); 115 R2(c, d, a, b, g, h, e, f, K[18], W1(18), W2(22)); 116 R2(b, c, d, a, f, g, h, e, K[19], W1(19), W2(23)); 117 R2(a, b, c, d, e, f, g, h, K[20], W1(20), W2(24)); 118 R2(d, a, b, c, h, e, f, g, K[21], W1(21), W2(25)); 119 R2(c, d, a, b, g, h, e, f, K[22], W1(22), W2(26)); 120 R2(b, c, d, a, f, g, h, e, K[23], W1(23), W2(27)); 121 R2(a, b, c, d, e, f, g, h, K[24], W1(24), W2(28)); 122 R2(d, a, b, c, h, e, f, g, K[25], W1(25), W2(29)); 123 R2(c, d, a, b, g, h, e, f, K[26], W1(26), W2(30)); 124 R2(b, c, d, a, f, g, h, e, K[27], W1(27), W2(31)); 125 R2(a, b, c, d, e, f, g, h, K[28], W1(28), W2(32)); 126 R2(d, a, b, c, h, e, f, g, K[29], W1(29), W2(33)); 127 R2(c, d, a, b, g, h, e, f, K[30], W1(30), W2(34)); 128 R2(b, c, d, a, f, g, h, e, K[31], W1(31), W2(35)); 129 130 R2(a, b, c, d, e, f, g, h, K[32], W1(32), W2(36)); 131 R2(d, a, b, c, h, e, f, g, K[33], W1(33), W2(37)); 132 R2(c, d, a, b, g, h, e, f, K[34], W1(34), W2(38)); 133 R2(b, c, d, a, f, g, h, e, K[35], W1(35), W2(39)); 134 R2(a, b, c, d, e, f, g, h, K[36], W1(36), W2(40)); 135 R2(d, a, b, c, h, e, f, g, K[37], W1(37), W2(41)); 136 R2(c, d, a, b, g, h, e, f, K[38], W1(38), W2(42)); 137 R2(b, c, d, a, f, g, h, e, K[39], W1(39), W2(43)); 138 R2(a, b, c, d, e, f, g, h, K[40], W1(40), W2(44)); 139 R2(d, a, b, c, h, e, f, g, K[41], W1(41), W2(45)); 140 R2(c, d, a, b, g, h, e, f, K[42], W1(42), W2(46)); 141 R2(b, c, d, a, f, g, h, e, K[43], W1(43), W2(47)); 142 R2(a, b, c, d, e, f, g, h, K[44], W1(44), W2(48)); 143 R2(d, a, b, c, h, e, f, g, K[45], W1(45), W2(49)); 144 R2(c, d, a, b, g, h, e, f, K[46], W1(46), W2(50)); 145 R2(b, c, d, a, f, g, h, e, K[47], W1(47), W2(51)); 146 147 R2(a, b, c, d, e, f, g, h, K[48], W1(48), W2(52)); 148 R2(d, a, b, c, h, e, f, g, K[49], W1(49), W2(53)); 149 R2(c, d, a, b, g, h, e, f, K[50], W1(50), W2(54)); 150 R2(b, c, d, a, f, g, h, e, K[51], W1(51), W2(55)); 151 R2(a, b, c, d, e, f, g, h, K[52], W1(52), W2(56)); 152 R2(d, a, b, c, h, e, f, g, K[53], W1(53), W2(57)); 153 R2(c, d, a, b, g, h, e, f, K[54], W1(54), W2(58)); 154 R2(b, c, d, a, f, g, h, e, K[55], W1(55), W2(59)); 155 R2(a, b, c, d, e, f, g, h, K[56], W1(56), W2(60)); 156 R2(d, a, b, c, h, e, f, g, K[57], W1(57), W2(61)); 157 R2(c, d, a, b, g, h, e, f, K[58], W1(58), W2(62)); 158 R2(b, c, d, a, f, g, h, e, K[59], W1(59), W2(63)); 159 R2(a, b, c, d, e, f, g, h, K[60], W1(60), W2(64)); 160 R2(d, a, b, c, h, e, f, g, K[61], W1(61), W2(65)); 161 R2(c, d, a, b, g, h, e, f, K[62], W1(62), W2(66)); 162 R2(b, c, d, a, f, g, h, e, K[63], W1(63), W2(67)); 163 164 state->h[0] ^= a; 165 state->h[1] ^= b; 166 state->h[2] ^= c; 167 state->h[3] ^= d; 168 state->h[4] ^= e; 169 state->h[5] ^= f; 170 state->h[6] ^= g; 171 state->h[7] ^= h; 172 } 173 #undef R 174 #undef R1 175 #undef R2 176 #undef I 177 #undef W1 178 #undef W2 179 180 static void __maybe_unused sm3_blocks_generic(struct sm3_block_state *state, 181 const u8 *data, size_t nblocks) 182 { 183 u32 W[16]; 184 185 do { 186 sm3_block_generic(state, data, W); 187 data += SM3_BLOCK_SIZE; 188 } while (--nblocks); 189 190 memzero_explicit(W, sizeof(W)); 191 } 192 193 #ifdef CONFIG_CRYPTO_LIB_SM3_ARCH 194 #include "sm3.h" /* $(SRCARCH)/sm3.h */ 195 #else 196 #define sm3_blocks sm3_blocks_generic 197 #endif 198 199 void sm3_init(struct sm3_ctx *ctx) 200 { 201 ctx->state = sm3_iv; 202 ctx->bytecount = 0; 203 } 204 EXPORT_SYMBOL_GPL(sm3_init); 205 206 void sm3_update(struct sm3_ctx *ctx, const u8 *data, size_t len) 207 { 208 size_t partial = ctx->bytecount % SM3_BLOCK_SIZE; 209 210 ctx->bytecount += len; 211 212 if (partial + len >= SM3_BLOCK_SIZE) { 213 size_t nblocks; 214 215 if (partial) { 216 size_t l = SM3_BLOCK_SIZE - partial; 217 218 memcpy(&ctx->buf[partial], data, l); 219 data += l; 220 len -= l; 221 222 sm3_blocks(&ctx->state, ctx->buf, 1); 223 } 224 225 nblocks = len / SM3_BLOCK_SIZE; 226 len %= SM3_BLOCK_SIZE; 227 228 if (nblocks) { 229 sm3_blocks(&ctx->state, data, nblocks); 230 data += nblocks * SM3_BLOCK_SIZE; 231 } 232 partial = 0; 233 } 234 if (len) 235 memcpy(&ctx->buf[partial], data, len); 236 } 237 EXPORT_SYMBOL_GPL(sm3_update); 238 239 static void __sm3_final(struct sm3_ctx *ctx, u8 out[SM3_DIGEST_SIZE]) 240 { 241 u64 bitcount = ctx->bytecount << 3; 242 size_t partial = ctx->bytecount % SM3_BLOCK_SIZE; 243 244 ctx->buf[partial++] = 0x80; 245 if (partial > SM3_BLOCK_SIZE - 8) { 246 memset(&ctx->buf[partial], 0, SM3_BLOCK_SIZE - partial); 247 sm3_blocks(&ctx->state, ctx->buf, 1); 248 partial = 0; 249 } 250 memset(&ctx->buf[partial], 0, SM3_BLOCK_SIZE - 8 - partial); 251 *(__be64 *)&ctx->buf[SM3_BLOCK_SIZE - 8] = cpu_to_be64(bitcount); 252 sm3_blocks(&ctx->state, ctx->buf, 1); 253 254 for (size_t i = 0; i < SM3_DIGEST_SIZE; i += 4) 255 put_unaligned_be32(ctx->state.h[i / 4], out + i); 256 } 257 258 void sm3_final(struct sm3_ctx *ctx, u8 out[SM3_DIGEST_SIZE]) 259 { 260 __sm3_final(ctx, out); 261 memzero_explicit(ctx, sizeof(*ctx)); 262 } 263 EXPORT_SYMBOL_GPL(sm3_final); 264 265 void sm3(const u8 *data, size_t len, u8 out[SM3_DIGEST_SIZE]) 266 { 267 struct sm3_ctx ctx; 268 269 sm3_init(&ctx); 270 sm3_update(&ctx, data, len); 271 sm3_final(&ctx, out); 272 } 273 EXPORT_SYMBOL_GPL(sm3); 274 275 #ifdef sm3_mod_init_arch 276 static int __init sm3_mod_init(void) 277 { 278 sm3_mod_init_arch(); 279 return 0; 280 } 281 subsys_initcall(sm3_mod_init); 282 283 static void __exit sm3_mod_exit(void) 284 { 285 } 286 module_exit(sm3_mod_exit); 287 #endif 288 289 MODULE_DESCRIPTION("SM3 library functions"); 290 MODULE_LICENSE("GPL v2"); 291