Date: Tue, 6 Jan 2026 19:13:46 -0800 From: Adrian Chadd <adrian@freebsd.org> To: =?UTF-8?B?SmVhbi1Tw6liYXN0aWVuIFDDqWRyb24=?= <dumbbell@freebsd.org> Cc: src-committers@freebsd.org, dev-commits-src-all@freebsd.org, dev-commits-src-main@freebsd.org Subject: Re: git: d448578b445d - main - linuxkpi: Add <linux/siphash.h> Message-ID: <CAJ-VmomHFgOcC1nWUXWS0zLbcjKPaPLCtbBosYOy=7xb4X0wTA@mail.gmail.com> In-Reply-To: <695c1a7c.859c.12adca29@gitrepo.freebsd.org>
index | next in thread | previous in thread | raw e-mail
Hi! This looks like it's at least failing on armv7: https://ci.freebsd.org/job/FreeBSD-main-armv7-build/26773/ 02:56:27 --- all_subdir_linuxkpi --- 02:56:27 --- linux_siphash.o --- 02:56:27 /usr/src/sys/compat/linuxkpi/common/src/linux_siphash.c:425:3: error: call = 02:56:27 to undeclared function 'rol32'; ISO C99 and later do not support implicit f= 02:56:27 unction declarations [-Werror,-Wimplicit-function-declaration] 02:56:27 425 | HSIPROUND; 02:56:27 | ^ -adrian On Mon, 5 Jan 2026 at 12:11, Jean-Sébastien Pédron <dumbbell@freebsd.org> wrote: > > The branch main has been updated by dumbbell: > > URL: https://cgit.FreeBSD.org/src/commit/?id=d448578b445da95806ef9af996a0db9754daadeb > > commit d448578b445da95806ef9af996a0db9754daadeb > Author: Jean-Sébastien Pédron <dumbbell@FreeBSD.org> > AuthorDate: 2025-09-07 13:43:11 +0000 > Commit: Jean-Sébastien Pédron <dumbbell@FreeBSD.org> > CommitDate: 2026-01-05 19:32:50 +0000 > > linuxkpi: Add <linux/siphash.h> > > The file is copied as is from Linux 6.10 as it dual-licensend under the > GPLv2 and BSD 3-clause. > > The amdgpu DRM driver started to use it in Linux 6.10. > > Reviewed by: bz, emaste > Sponsored by: The FreeBSD Foundation > Differential Revision: https://reviews.freebsd.org/D54501 > --- > sys/compat/linuxkpi/common/include/linux/siphash.h | 168 +++++++ > sys/compat/linuxkpi/common/src/linux_siphash.c | 546 +++++++++++++++++++++ > sys/conf/files | 2 + > sys/modules/linuxkpi/Makefile | 1 + > 4 files changed, 717 insertions(+) > > diff --git a/sys/compat/linuxkpi/common/include/linux/siphash.h b/sys/compat/linuxkpi/common/include/linux/siphash.h > new file mode 100644 > index 000000000000..9153e77382e1 > --- /dev/null > +++ b/sys/compat/linuxkpi/common/include/linux/siphash.h > @@ -0,0 +1,168 @@ > +/* SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) */ > +/* Copyright (C) 2016-2022 Jason A. Donenfeld <Jason@zx2c4.com>. All Rights Reserved. > + * > + * SipHash: a fast short-input PRF > + * https://131002.net/siphash/ > + * > + * This implementation is specifically for SipHash2-4 for a secure PRF > + * and HalfSipHash1-3/SipHash1-3 for an insecure PRF only suitable for > + * hashtables. > + */ > + > +#ifndef _LINUX_SIPHASH_H > +#define _LINUX_SIPHASH_H > + > +#include <linux/types.h> > +#include <linux/kernel.h> > + > +#define SIPHASH_ALIGNMENT __alignof__(u64) > +typedef struct { > + u64 key[2]; > +} siphash_key_t; > + > +#define siphash_aligned_key_t siphash_key_t __aligned(16) > + > +static inline bool siphash_key_is_zero(const siphash_key_t *key) > +{ > + return !(key->key[0] | key->key[1]); > +} > + > +u64 __siphash_aligned(const void *data, size_t len, const siphash_key_t *key); > +u64 __siphash_unaligned(const void *data, size_t len, const siphash_key_t *key); > + > +u64 siphash_1u64(const u64 a, const siphash_key_t *key); > +u64 siphash_2u64(const u64 a, const u64 b, const siphash_key_t *key); > +u64 siphash_3u64(const u64 a, const u64 b, const u64 c, > + const siphash_key_t *key); > +u64 siphash_4u64(const u64 a, const u64 b, const u64 c, const u64 d, > + const siphash_key_t *key); > +u64 siphash_1u32(const u32 a, const siphash_key_t *key); > +u64 siphash_3u32(const u32 a, const u32 b, const u32 c, > + const siphash_key_t *key); > + > +static inline u64 siphash_2u32(const u32 a, const u32 b, > + const siphash_key_t *key) > +{ > + return siphash_1u64((u64)b << 32 | a, key); > +} > +static inline u64 siphash_4u32(const u32 a, const u32 b, const u32 c, > + const u32 d, const siphash_key_t *key) > +{ > + return siphash_2u64((u64)b << 32 | a, (u64)d << 32 | c, key); > +} > + > + > +static inline u64 ___siphash_aligned(const __le64 *data, size_t len, > + const siphash_key_t *key) > +{ > + if (__builtin_constant_p(len) && len == 4) > + return siphash_1u32(le32_to_cpup((const __le32 *)data), key); > + if (__builtin_constant_p(len) && len == 8) > + return siphash_1u64(le64_to_cpu(data[0]), key); > + if (__builtin_constant_p(len) && len == 16) > + return siphash_2u64(le64_to_cpu(data[0]), le64_to_cpu(data[1]), > + key); > + if (__builtin_constant_p(len) && len == 24) > + return siphash_3u64(le64_to_cpu(data[0]), le64_to_cpu(data[1]), > + le64_to_cpu(data[2]), key); > + if (__builtin_constant_p(len) && len == 32) > + return siphash_4u64(le64_to_cpu(data[0]), le64_to_cpu(data[1]), > + le64_to_cpu(data[2]), le64_to_cpu(data[3]), > + key); > + return __siphash_aligned(data, len, key); > +} > + > +/** > + * siphash - compute 64-bit siphash PRF value > + * @data: buffer to hash > + * @size: size of @data > + * @key: the siphash key > + */ > +static inline u64 siphash(const void *data, size_t len, > + const siphash_key_t *key) > +{ > + if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) || > + !IS_ALIGNED((unsigned long)data, SIPHASH_ALIGNMENT)) > + return __siphash_unaligned(data, len, key); > + return ___siphash_aligned(data, len, key); > +} > + > +#define HSIPHASH_ALIGNMENT __alignof__(unsigned long) > +typedef struct { > + unsigned long key[2]; > +} hsiphash_key_t; > + > +u32 __hsiphash_aligned(const void *data, size_t len, > + const hsiphash_key_t *key); > +u32 __hsiphash_unaligned(const void *data, size_t len, > + const hsiphash_key_t *key); > + > +u32 hsiphash_1u32(const u32 a, const hsiphash_key_t *key); > +u32 hsiphash_2u32(const u32 a, const u32 b, const hsiphash_key_t *key); > +u32 hsiphash_3u32(const u32 a, const u32 b, const u32 c, > + const hsiphash_key_t *key); > +u32 hsiphash_4u32(const u32 a, const u32 b, const u32 c, const u32 d, > + const hsiphash_key_t *key); > + > +static inline u32 ___hsiphash_aligned(const __le32 *data, size_t len, > + const hsiphash_key_t *key) > +{ > + if (__builtin_constant_p(len) && len == 4) > + return hsiphash_1u32(le32_to_cpu(data[0]), key); > + if (__builtin_constant_p(len) && len == 8) > + return hsiphash_2u32(le32_to_cpu(data[0]), le32_to_cpu(data[1]), > + key); > + if (__builtin_constant_p(len) && len == 12) > + return hsiphash_3u32(le32_to_cpu(data[0]), le32_to_cpu(data[1]), > + le32_to_cpu(data[2]), key); > + if (__builtin_constant_p(len) && len == 16) > + return hsiphash_4u32(le32_to_cpu(data[0]), le32_to_cpu(data[1]), > + le32_to_cpu(data[2]), le32_to_cpu(data[3]), > + key); > + return __hsiphash_aligned(data, len, key); > +} > + > +/** > + * hsiphash - compute 32-bit hsiphash PRF value > + * @data: buffer to hash > + * @size: size of @data > + * @key: the hsiphash key > + */ > +static inline u32 hsiphash(const void *data, size_t len, > + const hsiphash_key_t *key) > +{ > + if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) || > + !IS_ALIGNED((unsigned long)data, HSIPHASH_ALIGNMENT)) > + return __hsiphash_unaligned(data, len, key); > + return ___hsiphash_aligned(data, len, key); > +} > + > +/* > + * These macros expose the raw SipHash and HalfSipHash permutations. > + * Do not use them directly! If you think you have a use for them, > + * be sure to CC the maintainer of this file explaining why. > + */ > + > +#define SIPHASH_PERMUTATION(a, b, c, d) ( \ > + (a) += (b), (b) = rol64((b), 13), (b) ^= (a), (a) = rol64((a), 32), \ > + (c) += (d), (d) = rol64((d), 16), (d) ^= (c), \ > + (a) += (d), (d) = rol64((d), 21), (d) ^= (a), \ > + (c) += (b), (b) = rol64((b), 17), (b) ^= (c), (c) = rol64((c), 32)) > + > +#define SIPHASH_CONST_0 0x736f6d6570736575ULL > +#define SIPHASH_CONST_1 0x646f72616e646f6dULL > +#define SIPHASH_CONST_2 0x6c7967656e657261ULL > +#define SIPHASH_CONST_3 0x7465646279746573ULL > + > +#define HSIPHASH_PERMUTATION(a, b, c, d) ( \ > + (a) += (b), (b) = rol32((b), 5), (b) ^= (a), (a) = rol32((a), 16), \ > + (c) += (d), (d) = rol32((d), 8), (d) ^= (c), \ > + (a) += (d), (d) = rol32((d), 7), (d) ^= (a), \ > + (c) += (b), (b) = rol32((b), 13), (b) ^= (c), (c) = rol32((c), 16)) > + > +#define HSIPHASH_CONST_0 0U > +#define HSIPHASH_CONST_1 0U > +#define HSIPHASH_CONST_2 0x6c796765U > +#define HSIPHASH_CONST_3 0x74656462U > + > +#endif /* _LINUX_SIPHASH_H */ > diff --git a/sys/compat/linuxkpi/common/src/linux_siphash.c b/sys/compat/linuxkpi/common/src/linux_siphash.c > new file mode 100644 > index 000000000000..b4842a8250e1 > --- /dev/null > +++ b/sys/compat/linuxkpi/common/src/linux_siphash.c > @@ -0,0 +1,546 @@ > +// SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) > +/* Copyright (C) 2016-2022 Jason A. Donenfeld <Jason@zx2c4.com>. All Rights Reserved. > + * > + * SipHash: a fast short-input PRF > + * https://131002.net/siphash/ > + * > + * This implementation is specifically for SipHash2-4 for a secure PRF > + * and HalfSipHash1-3/SipHash1-3 for an insecure PRF only suitable for > + * hashtables. > + */ > + > +#include <linux/siphash.h> > +#include <asm/unaligned.h> > + > +#if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64 > +#include <linux/dcache.h> > +#include <asm/word-at-a-time.h> > +#endif > + > +#define EXPORT_SYMBOL(name) > + > +#define SIPROUND SIPHASH_PERMUTATION(v0, v1, v2, v3) > + > +#define PREAMBLE(len) \ > + u64 v0 = SIPHASH_CONST_0; \ > + u64 v1 = SIPHASH_CONST_1; \ > + u64 v2 = SIPHASH_CONST_2; \ > + u64 v3 = SIPHASH_CONST_3; \ > + u64 b = ((u64)(len)) << 56; \ > + v3 ^= key->key[1]; \ > + v2 ^= key->key[0]; \ > + v1 ^= key->key[1]; \ > + v0 ^= key->key[0]; > + > +#define POSTAMBLE \ > + v3 ^= b; \ > + SIPROUND; \ > + SIPROUND; \ > + v0 ^= b; \ > + v2 ^= 0xff; \ > + SIPROUND; \ > + SIPROUND; \ > + SIPROUND; \ > + SIPROUND; \ > + return (v0 ^ v1) ^ (v2 ^ v3); > + > +#ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS > +u64 __siphash_aligned(const void *_data, size_t len, const siphash_key_t *key) > +{ > + const u8 *data = _data; > + const u8 *end = data + len - (len % sizeof(u64)); > + const u8 left = len & (sizeof(u64) - 1); > + u64 m; > + PREAMBLE(len) > + for (; data != end; data += sizeof(u64)) { > + m = le64_to_cpup(data); > + v3 ^= m; > + SIPROUND; > + SIPROUND; > + v0 ^= m; > + } > +#if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64 > + if (left) > + b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & > + bytemask_from_count(left))); > +#else > + switch (left) { > + case 7: b |= ((u64)end[6]) << 48; fallthrough; > + case 6: b |= ((u64)end[5]) << 40; fallthrough; > + case 5: b |= ((u64)end[4]) << 32; fallthrough; > + case 4: b |= le32_to_cpup(data); break; > + case 3: b |= ((u64)end[2]) << 16; fallthrough; > + case 2: b |= le16_to_cpup(data); break; > + case 1: b |= end[0]; > + } > +#endif > + POSTAMBLE > +} > +EXPORT_SYMBOL(__siphash_aligned); > +#endif > + > +u64 __siphash_unaligned(const void *_data, size_t len, const siphash_key_t *key) > +{ > + const u8 *data = _data; > + const u8 *end = data + len - (len % sizeof(u64)); > + const u8 left = len & (sizeof(u64) - 1); > + u64 m; > + PREAMBLE(len) > + for (; data != end; data += sizeof(u64)) { > + m = get_unaligned_le64(data); > + v3 ^= m; > + SIPROUND; > + SIPROUND; > + v0 ^= m; > + } > +#if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64 > + if (left) > + b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & > + bytemask_from_count(left))); > +#else > + switch (left) { > + case 7: b |= ((u64)end[6]) << 48; fallthrough; > + case 6: b |= ((u64)end[5]) << 40; fallthrough; > + case 5: b |= ((u64)end[4]) << 32; fallthrough; > + case 4: b |= get_unaligned_le32(end); break; > + case 3: b |= ((u64)end[2]) << 16; fallthrough; > + case 2: b |= get_unaligned_le16(end); break; > + case 1: b |= end[0]; > + } > +#endif > + POSTAMBLE > +} > +EXPORT_SYMBOL(__siphash_unaligned); > + > +/** > + * siphash_1u64 - compute 64-bit siphash PRF value of a u64 > + * @first: first u64 > + * @key: the siphash key > + */ > +u64 siphash_1u64(const u64 first, const siphash_key_t *key) > +{ > + PREAMBLE(8) > + v3 ^= first; > + SIPROUND; > + SIPROUND; > + v0 ^= first; > + POSTAMBLE > +} > +EXPORT_SYMBOL(siphash_1u64); > + > +/** > + * siphash_2u64 - compute 64-bit siphash PRF value of 2 u64 > + * @first: first u64 > + * @second: second u64 > + * @key: the siphash key > + */ > +u64 siphash_2u64(const u64 first, const u64 second, const siphash_key_t *key) > +{ > + PREAMBLE(16) > + v3 ^= first; > + SIPROUND; > + SIPROUND; > + v0 ^= first; > + v3 ^= second; > + SIPROUND; > + SIPROUND; > + v0 ^= second; > + POSTAMBLE > +} > +EXPORT_SYMBOL(siphash_2u64); > + > +/** > + * siphash_3u64 - compute 64-bit siphash PRF value of 3 u64 > + * @first: first u64 > + * @second: second u64 > + * @third: third u64 > + * @key: the siphash key > + */ > +u64 siphash_3u64(const u64 first, const u64 second, const u64 third, > + const siphash_key_t *key) > +{ > + PREAMBLE(24) > + v3 ^= first; > + SIPROUND; > + SIPROUND; > + v0 ^= first; > + v3 ^= second; > + SIPROUND; > + SIPROUND; > + v0 ^= second; > + v3 ^= third; > + SIPROUND; > + SIPROUND; > + v0 ^= third; > + POSTAMBLE > +} > +EXPORT_SYMBOL(siphash_3u64); > + > +/** > + * siphash_4u64 - compute 64-bit siphash PRF value of 4 u64 > + * @first: first u64 > + * @second: second u64 > + * @third: third u64 > + * @forth: forth u64 > + * @key: the siphash key > + */ > +u64 siphash_4u64(const u64 first, const u64 second, const u64 third, > + const u64 forth, const siphash_key_t *key) > +{ > + PREAMBLE(32) > + v3 ^= first; > + SIPROUND; > + SIPROUND; > + v0 ^= first; > + v3 ^= second; > + SIPROUND; > + SIPROUND; > + v0 ^= second; > + v3 ^= third; > + SIPROUND; > + SIPROUND; > + v0 ^= third; > + v3 ^= forth; > + SIPROUND; > + SIPROUND; > + v0 ^= forth; > + POSTAMBLE > +} > +EXPORT_SYMBOL(siphash_4u64); > + > +u64 siphash_1u32(const u32 first, const siphash_key_t *key) > +{ > + PREAMBLE(4) > + b |= first; > + POSTAMBLE > +} > +EXPORT_SYMBOL(siphash_1u32); > + > +u64 siphash_3u32(const u32 first, const u32 second, const u32 third, > + const siphash_key_t *key) > +{ > + u64 combined = (u64)second << 32 | first; > + PREAMBLE(12) > + v3 ^= combined; > + SIPROUND; > + SIPROUND; > + v0 ^= combined; > + b |= third; > + POSTAMBLE > +} > +EXPORT_SYMBOL(siphash_3u32); > + > +#if BITS_PER_LONG == 64 > +/* Note that on 64-bit, we make HalfSipHash1-3 actually be SipHash1-3, for > + * performance reasons. On 32-bit, below, we actually implement HalfSipHash1-3. > + */ > + > +#define HSIPROUND SIPROUND > +#define HPREAMBLE(len) PREAMBLE(len) > +#define HPOSTAMBLE \ > + v3 ^= b; \ > + HSIPROUND; \ > + v0 ^= b; \ > + v2 ^= 0xff; \ > + HSIPROUND; \ > + HSIPROUND; \ > + HSIPROUND; \ > + return (v0 ^ v1) ^ (v2 ^ v3); > + > +#ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS > +u32 __hsiphash_aligned(const void *_data, size_t len, const hsiphash_key_t *key) > +{ > + const u8 *data = _data; > + const u8 *end = data + len - (len % sizeof(u64)); > + const u8 left = len & (sizeof(u64) - 1); > + u64 m; > + HPREAMBLE(len) > + for (; data != end; data += sizeof(u64)) { > + m = le64_to_cpup(data); > + v3 ^= m; > + HSIPROUND; > + v0 ^= m; > + } > +#if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64 > + if (left) > + b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & > + bytemask_from_count(left))); > +#else > + switch (left) { > + case 7: b |= ((u64)end[6]) << 48; fallthrough; > + case 6: b |= ((u64)end[5]) << 40; fallthrough; > + case 5: b |= ((u64)end[4]) << 32; fallthrough; > + case 4: b |= le32_to_cpup(data); break; > + case 3: b |= ((u64)end[2]) << 16; fallthrough; > + case 2: b |= le16_to_cpup(data); break; > + case 1: b |= end[0]; > + } > +#endif > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(__hsiphash_aligned); > +#endif > + > +u32 __hsiphash_unaligned(const void *_data, size_t len, > + const hsiphash_key_t *key) > +{ > + const u8 *data = _data; > + const u8 *end = data + len - (len % sizeof(u64)); > + const u8 left = len & (sizeof(u64) - 1); > + u64 m; > + HPREAMBLE(len) > + for (; data != end; data += sizeof(u64)) { > + m = get_unaligned_le64(data); > + v3 ^= m; > + HSIPROUND; > + v0 ^= m; > + } > +#if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64 > + if (left) > + b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & > + bytemask_from_count(left))); > +#else > + switch (left) { > + case 7: b |= ((u64)end[6]) << 48; fallthrough; > + case 6: b |= ((u64)end[5]) << 40; fallthrough; > + case 5: b |= ((u64)end[4]) << 32; fallthrough; > + case 4: b |= get_unaligned_le32(end); break; > + case 3: b |= ((u64)end[2]) << 16; fallthrough; > + case 2: b |= get_unaligned_le16(end); break; > + case 1: b |= end[0]; > + } > +#endif > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(__hsiphash_unaligned); > + > +/** > + * hsiphash_1u32 - compute 64-bit hsiphash PRF value of a u32 > + * @first: first u32 > + * @key: the hsiphash key > + */ > +u32 hsiphash_1u32(const u32 first, const hsiphash_key_t *key) > +{ > + HPREAMBLE(4) > + b |= first; > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(hsiphash_1u32); > + > +/** > + * hsiphash_2u32 - compute 32-bit hsiphash PRF value of 2 u32 > + * @first: first u32 > + * @second: second u32 > + * @key: the hsiphash key > + */ > +u32 hsiphash_2u32(const u32 first, const u32 second, const hsiphash_key_t *key) > +{ > + u64 combined = (u64)second << 32 | first; > + HPREAMBLE(8) > + v3 ^= combined; > + HSIPROUND; > + v0 ^= combined; > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(hsiphash_2u32); > + > +/** > + * hsiphash_3u32 - compute 32-bit hsiphash PRF value of 3 u32 > + * @first: first u32 > + * @second: second u32 > + * @third: third u32 > + * @key: the hsiphash key > + */ > +u32 hsiphash_3u32(const u32 first, const u32 second, const u32 third, > + const hsiphash_key_t *key) > +{ > + u64 combined = (u64)second << 32 | first; > + HPREAMBLE(12) > + v3 ^= combined; > + HSIPROUND; > + v0 ^= combined; > + b |= third; > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(hsiphash_3u32); > + > +/** > + * hsiphash_4u32 - compute 32-bit hsiphash PRF value of 4 u32 > + * @first: first u32 > + * @second: second u32 > + * @third: third u32 > + * @forth: forth u32 > + * @key: the hsiphash key > + */ > +u32 hsiphash_4u32(const u32 first, const u32 second, const u32 third, > + const u32 forth, const hsiphash_key_t *key) > +{ > + u64 combined = (u64)second << 32 | first; > + HPREAMBLE(16) > + v3 ^= combined; > + HSIPROUND; > + v0 ^= combined; > + combined = (u64)forth << 32 | third; > + v3 ^= combined; > + HSIPROUND; > + v0 ^= combined; > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(hsiphash_4u32); > +#else > +#define HSIPROUND HSIPHASH_PERMUTATION(v0, v1, v2, v3) > + > +#define HPREAMBLE(len) \ > + u32 v0 = HSIPHASH_CONST_0; \ > + u32 v1 = HSIPHASH_CONST_1; \ > + u32 v2 = HSIPHASH_CONST_2; \ > + u32 v3 = HSIPHASH_CONST_3; \ > + u32 b = ((u32)(len)) << 24; \ > + v3 ^= key->key[1]; \ > + v2 ^= key->key[0]; \ > + v1 ^= key->key[1]; \ > + v0 ^= key->key[0]; > + > +#define HPOSTAMBLE \ > + v3 ^= b; \ > + HSIPROUND; \ > + v0 ^= b; \ > + v2 ^= 0xff; \ > + HSIPROUND; \ > + HSIPROUND; \ > + HSIPROUND; \ > + return v1 ^ v3; > + > +#ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS > +u32 __hsiphash_aligned(const void *_data, size_t len, const hsiphash_key_t *key) > +{ > + const u8 *data = _data; > + const u8 *end = data + len - (len % sizeof(u32)); > + const u8 left = len & (sizeof(u32) - 1); > + u32 m; > + HPREAMBLE(len) > + for (; data != end; data += sizeof(u32)) { > + m = le32_to_cpup(data); > + v3 ^= m; > + HSIPROUND; > + v0 ^= m; > + } > + switch (left) { > + case 3: b |= ((u32)end[2]) << 16; fallthrough; > + case 2: b |= le16_to_cpup(data); break; > + case 1: b |= end[0]; > + } > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(__hsiphash_aligned); > +#endif > + > +u32 __hsiphash_unaligned(const void *_data, size_t len, > + const hsiphash_key_t *key) > +{ > + const u8 *data = _data; > + const u8 *end = data + len - (len % sizeof(u32)); > + const u8 left = len & (sizeof(u32) - 1); > + u32 m; > + HPREAMBLE(len) > + for (; data != end; data += sizeof(u32)) { > + m = get_unaligned_le32(data); > + v3 ^= m; > + HSIPROUND; > + v0 ^= m; > + } > + switch (left) { > + case 3: b |= ((u32)end[2]) << 16; fallthrough; > + case 2: b |= get_unaligned_le16(end); break; > + case 1: b |= end[0]; > + } > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(__hsiphash_unaligned); > + > +/** > + * hsiphash_1u32 - compute 32-bit hsiphash PRF value of a u32 > + * @first: first u32 > + * @key: the hsiphash key > + */ > +u32 hsiphash_1u32(const u32 first, const hsiphash_key_t *key) > +{ > + HPREAMBLE(4) > + v3 ^= first; > + HSIPROUND; > + v0 ^= first; > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(hsiphash_1u32); > + > +/** > + * hsiphash_2u32 - compute 32-bit hsiphash PRF value of 2 u32 > + * @first: first u32 > + * @second: second u32 > + * @key: the hsiphash key > + */ > +u32 hsiphash_2u32(const u32 first, const u32 second, const hsiphash_key_t *key) > +{ > + HPREAMBLE(8) > + v3 ^= first; > + HSIPROUND; > + v0 ^= first; > + v3 ^= second; > + HSIPROUND; > + v0 ^= second; > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(hsiphash_2u32); > + > +/** > + * hsiphash_3u32 - compute 32-bit hsiphash PRF value of 3 u32 > + * @first: first u32 > + * @second: second u32 > + * @third: third u32 > + * @key: the hsiphash key > + */ > +u32 hsiphash_3u32(const u32 first, const u32 second, const u32 third, > + const hsiphash_key_t *key) > +{ > + HPREAMBLE(12) > + v3 ^= first; > + HSIPROUND; > + v0 ^= first; > + v3 ^= second; > + HSIPROUND; > + v0 ^= second; > + v3 ^= third; > + HSIPROUND; > + v0 ^= third; > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(hsiphash_3u32); > + > +/** > + * hsiphash_4u32 - compute 32-bit hsiphash PRF value of 4 u32 > + * @first: first u32 > + * @second: second u32 > + * @third: third u32 > + * @forth: forth u32 > + * @key: the hsiphash key > + */ > +u32 hsiphash_4u32(const u32 first, const u32 second, const u32 third, > + const u32 forth, const hsiphash_key_t *key) > +{ > + HPREAMBLE(16) > + v3 ^= first; > + HSIPROUND; > + v0 ^= first; > + v3 ^= second; > + HSIPROUND; > + v0 ^= second; > + v3 ^= third; > + HSIPROUND; > + v0 ^= third; > + v3 ^= forth; > + HSIPROUND; > + v0 ^= forth; > + HPOSTAMBLE > +} > +EXPORT_SYMBOL(hsiphash_4u32); > +#endif > diff --git a/sys/conf/files b/sys/conf/files > index 8deb2bd400c0..d0c4ea5f544d 100644 > --- a/sys/conf/files > +++ b/sys/conf/files > @@ -4704,6 +4704,8 @@ compat/linuxkpi/common/src/linux_shmemfs.c optional compat_linuxkpi \ > compile-with "${LINUXKPI_C}" > compat/linuxkpi/common/src/linux_shrinker.c optional compat_linuxkpi \ > compile-with "${LINUXKPI_C}" > +compat/linuxkpi/common/src/linux_siphash.c optional compat_linuxkpi \ > + compile-with "${LINUXKPI_C}" > compat/linuxkpi/common/src/linux_skbuff.c optional compat_linuxkpi \ > compile-with "${LINUXKPI_C}" > compat/linuxkpi/common/src/linux_slab.c optional compat_linuxkpi \ > diff --git a/sys/modules/linuxkpi/Makefile b/sys/modules/linuxkpi/Makefile > index a662f5dffbb6..c465c76a7626 100644 > --- a/sys/modules/linuxkpi/Makefile > +++ b/sys/modules/linuxkpi/Makefile > @@ -28,6 +28,7 @@ SRCS= linux_compat.c \ > linux_shmemfs.c \ > linux_shrinker.c \ > linux_simple_attr.c \ > + linux_siphash.c \ > linux_skbuff.c \ > linux_slab.c \ > linux_tasklet.c \ >home | help
Want to link to this message? Use this
URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?CAJ-VmomHFgOcC1nWUXWS0zLbcjKPaPLCtbBosYOy=7xb4X0wTA>
