[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [xen staging-4.18] xen/lib: Introduce SHA2-256
commit ea0110cb3c49f351bdb39f8800fe421a22fc04ce Author: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> AuthorDate: Fri Dec 13 14:34:00 2024 +0000 Commit: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> CommitDate: Fri Apr 11 17:27:59 2025 +0100 xen/lib: Introduce SHA2-256 A future change will need to calculate SHA2-256 digests. Introduce an implementation in lib/, derived from Trenchboot which itself is derived from Linux. In order to be useful to other architectures, it is careful with endianness and misaligned accesses as well as being more MISRA friendly, but is only wired up for x86 in the short term. Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> Acked-by: Roger Pau Monné <roger.pau@xxxxxxxxxx> (cherry picked from commit 372af524411f5a013bcb0b117073d8d07c026563) Xen: CI fix from XSN-2 * Add U suffix to the K[] table to fix MISRA Rule 7.2 violations. Fixes: 372af524411f ("xen/lib: Introduce SHA2-256") Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> Reviewed-by: Stefano Stabellini <sstabellini@xxxxxxxxxx> (cherry picked from commit 15fe2eb5f1bac8a212c0ba3d6dfe60d1fdf851cf) --- xen/include/xen/sha2.h | 15 ++++ xen/lib/Makefile | 1 + xen/lib/sha2-256.c | 216 +++++++++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 232 insertions(+) diff --git a/xen/include/xen/sha2.h b/xen/include/xen/sha2.h new file mode 100644 index 0000000000..47d97fbf01 --- /dev/null +++ b/xen/include/xen/sha2.h @@ -0,0 +1,15 @@ +/* SPDX-License-Identifier: GPL-2.0-or-later */ +/* + * SHA2-256: https://csrc.nist.gov/pubs/fips/180-2/upd1/final + */ +#ifndef XEN_SHA2_H +#define XEN_SHA2_H + +#include <xen/types.h> + +#define SHA2_256_DIGEST_SIZE 32 + +void sha2_256_digest(uint8_t digest[SHA2_256_DIGEST_SIZE], + const void *msg, size_t len); + +#endif /* XEN_SHA2_H */ diff --git a/xen/lib/Makefile b/xen/lib/Makefile index 2d9ebb945f..90766e37ce 100644 --- a/xen/lib/Makefile +++ b/xen/lib/Makefile @@ -13,6 +13,7 @@ lib-y += memset.o lib-y += muldiv64.o lib-y += parse-size.o lib-y += rbtree.o +lib-$(CONFIG_X86) += sha2-256.o lib-y += sort.o lib-y += strcasecmp.o lib-y += strchr.o diff --git a/xen/lib/sha2-256.c b/xen/lib/sha2-256.c new file mode 100644 index 0000000000..cb02ec1857 --- /dev/null +++ b/xen/lib/sha2-256.c @@ -0,0 +1,216 @@ +/* SPDX-License-Identifier: GPL-2.0-or-later */ +/* + * SHA2-256: https://csrc.nist.gov/pubs/fips/180-2/upd1/final + * + * Originally derived from Linux. Modified substantially to optimise for size + * and Xen's expected usecases. + */ +#include <xen/bitops.h> +#include <xen/sha2.h> +#include <xen/string.h> + +#include <asm/unaligned.h> + +struct sha2_256_state { + uint32_t state[SHA2_256_DIGEST_SIZE / sizeof(uint32_t)]; + uint8_t buf[64]; + size_t count; /* Byte count. */ +}; + +static uint32_t choose(uint32_t x, uint32_t y, uint32_t z) +{ + return z ^ (x & (y ^ z)); +} + +static uint32_t majority(uint32_t x, uint32_t y, uint32_t z) +{ + return (x & y) | (z & (x | y)); +} + +static uint32_t e0(uint32_t x) +{ + return ror32(x, 2) ^ ror32(x, 13) ^ ror32(x, 22); +} + +static uint32_t e1(uint32_t x) +{ + return ror32(x, 6) ^ ror32(x, 11) ^ ror32(x, 25); +} + +static uint32_t s0(uint32_t x) +{ + return ror32(x, 7) ^ ror32(x, 18) ^ (x >> 3); +} + +static uint32_t s1(uint32_t x) +{ + return ror32(x, 17) ^ ror32(x, 19) ^ (x >> 10); +} + +static uint32_t blend(uint32_t W[16], unsigned int i) +{ +#define W(i) W[(i) & 15] + + return W(i) += s1(W(i - 2)) + W(i - 7) + s0(W(i - 15)); + +#undef W +} + +static const uint32_t K[] = { + 0x428a2f98U, 0x71374491U, 0xb5c0fbcfU, 0xe9b5dba5U, + 0x3956c25bU, 0x59f111f1U, 0x923f82a4U, 0xab1c5ed5U, + 0xd807aa98U, 0x12835b01U, 0x243185beU, 0x550c7dc3U, + 0x72be5d74U, 0x80deb1feU, 0x9bdc06a7U, 0xc19bf174U, + 0xe49b69c1U, 0xefbe4786U, 0x0fc19dc6U, 0x240ca1ccU, + 0x2de92c6fU, 0x4a7484aaU, 0x5cb0a9dcU, 0x76f988daU, + 0x983e5152U, 0xa831c66dU, 0xb00327c8U, 0xbf597fc7U, + 0xc6e00bf3U, 0xd5a79147U, 0x06ca6351U, 0x14292967U, + 0x27b70a85U, 0x2e1b2138U, 0x4d2c6dfcU, 0x53380d13U, + 0x650a7354U, 0x766a0abbU, 0x81c2c92eU, 0x92722c85U, + 0xa2bfe8a1U, 0xa81a664bU, 0xc24b8b70U, 0xc76c51a3U, + 0xd192e819U, 0xd6990624U, 0xf40e3585U, 0x106aa070U, + 0x19a4c116U, 0x1e376c08U, 0x2748774cU, 0x34b0bcb5U, + 0x391c0cb3U, 0x4ed8aa4aU, 0x5b9cca4fU, 0x682e6ff3U, + 0x748f82eeU, 0x78a5636fU, 0x84c87814U, 0x8cc70208U, + 0x90befffaU, 0xa4506cebU, 0xbef9a3f7U, 0xc67178f2U, +}; + +static void sha2_256_transform(uint32_t *state, const void *_input) +{ + const uint32_t *input = _input; + uint32_t a, b, c, d, e, f, g, h, t1, t2; + uint32_t W[16]; + unsigned int i; + + for ( i = 0; i < 16; i++ ) + W[i] = get_unaligned_be32(&input[i]); + + a = state[0]; b = state[1]; c = state[2]; d = state[3]; + e = state[4]; f = state[5]; g = state[6]; h = state[7]; + + for ( i = 0; i < 16; i += 8 ) + { + t1 = h + e1(e) + choose(e, f, g) + K[i + 0] + W[i + 0]; + t2 = e0(a) + majority(a, b, c); d += t1; h = t1 + t2; + t1 = g + e1(d) + choose(d, e, f) + K[i + 1] + W[i + 1]; + t2 = e0(h) + majority(h, a, b); c += t1; g = t1 + t2; + t1 = f + e1(c) + choose(c, d, e) + K[i + 2] + W[i + 2]; + t2 = e0(g) + majority(g, h, a); b += t1; f = t1 + t2; + t1 = e + e1(b) + choose(b, c, d) + K[i + 3] + W[i + 3]; + t2 = e0(f) + majority(f, g, h); a += t1; e = t1 + t2; + t1 = d + e1(a) + choose(a, b, c) + K[i + 4] + W[i + 4]; + t2 = e0(e) + majority(e, f, g); h += t1; d = t1 + t2; + t1 = c + e1(h) + choose(h, a, b) + K[i + 5] + W[i + 5]; + t2 = e0(d) + majority(d, e, f); g += t1; c = t1 + t2; + t1 = b + e1(g) + choose(g, h, a) + K[i + 6] + W[i + 6]; + t2 = e0(c) + majority(c, d, e); f += t1; b = t1 + t2; + t1 = a + e1(f) + choose(f, g, h) + K[i + 7] + W[i + 7]; + t2 = e0(b) + majority(b, c, d); e += t1; a = t1 + t2; + } + + for ( ; i < 64; i += 8 ) + { + t1 = h + e1(e) + choose(e, f, g) + K[i + 0] + blend(W, i + 0); + t2 = e0(a) + majority(a, b, c); d += t1; h = t1 + t2; + t1 = g + e1(d) + choose(d, e, f) + K[i + 1] + blend(W, i + 1); + t2 = e0(h) + majority(h, a, b); c += t1; g = t1 + t2; + t1 = f + e1(c) + choose(c, d, e) + K[i + 2] + blend(W, i + 2); + t2 = e0(g) + majority(g, h, a); b += t1; f = t1 + t2; + t1 = e + e1(b) + choose(b, c, d) + K[i + 3] + blend(W, i + 3); + t2 = e0(f) + majority(f, g, h); a += t1; e = t1 + t2; + t1 = d + e1(a) + choose(a, b, c) + K[i + 4] + blend(W, i + 4); + t2 = e0(e) + majority(e, f, g); h += t1; d = t1 + t2; + t1 = c + e1(h) + choose(h, a, b) + K[i + 5] + blend(W, i + 5); + t2 = e0(d) + majority(d, e, f); g += t1; c = t1 + t2; + t1 = b + e1(g) + choose(g, h, a) + K[i + 6] + blend(W, i + 6); + t2 = e0(c) + majority(c, d, e); f += t1; b = t1 + t2; + t1 = a + e1(f) + choose(f, g, h) + K[i + 7] + blend(W, i + 7); + t2 = e0(b) + majority(b, c, d); e += t1; a = t1 + t2; + } + + state[0] += a; state[1] += b; state[2] += c; state[3] += d; + state[4] += e; state[5] += f; state[6] += g; state[7] += h; +} + +static void sha2_256_init(struct sha2_256_state *s) +{ + *s = (struct sha2_256_state){ + .state = { + 0x6a09e667UL, + 0xbb67ae85UL, + 0x3c6ef372UL, + 0xa54ff53aUL, + 0x510e527fUL, + 0x9b05688cUL, + 0x1f83d9abUL, + 0x5be0cd19UL, + }, + }; +} + +static void sha2_256_update(struct sha2_256_state *s, const void *msg, + size_t len) +{ + unsigned int partial = s->count & 63; + + s->count += len; + + if ( (partial + len) >= 64 ) + { + if ( partial ) + { + unsigned int rem = 64 - partial; + + /* Fill the partial block. */ + memcpy(s->buf + partial, msg, rem); + msg += rem; + len -= rem; + + sha2_256_transform(s->state, s->buf); + partial = 0; + } + + for ( ; len >= 64; msg += 64, len -= 64 ) + sha2_256_transform(s->state, msg); + } + + /* Remaining data becomes partial. */ + memcpy(s->buf + partial, msg, len); +} + +static void sha2_256_final(struct sha2_256_state *s, void *_dst) +{ + uint32_t *dst = _dst; + unsigned int i, partial = s->count & 63; + + /* Start padding */ + s->buf[partial++] = 0x80; + + if ( partial > 56 ) + { + /* Need one extra block - pad to 64 */ + memset(s->buf + partial, 0, 64 - partial); + sha2_256_transform(s->state, s->buf); + partial = 0; + } + /* Pad to 56 */ + memset(s->buf + partial, 0, 56 - partial); + + /* Append the bit count */ + put_unaligned_be64((uint64_t)s->count << 3, &s->buf[56]); + sha2_256_transform(s->state, s->buf); + + /* Store state in digest */ + for ( i = 0; i < 8; i++ ) + put_unaligned_be32(s->state[i], &dst[i]); +} + +void sha2_256_digest(uint8_t digest[SHA2_256_DIGEST_SIZE], + const void *msg, size_t len) +{ + struct sha2_256_state s; + + sha2_256_init(&s); + sha2_256_update(&s, msg, len); + sha2_256_final(&s, digest); +} -- generated by git-patchbot for /home/xen/git/xen.git#staging-4.18
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |