123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552 |
- /* Copyright (C) 2016 Jason A. Donenfeld <Jason@zx2c4.com>. All Rights Reserved.
- *
- * This file is provided under a dual BSD/GPLv2 license.
- *
- * SipHash: a fast short-input PRF
- * https://131002.net/siphash/
- *
- * This implementation is specifically for SipHash2-4 for a secure PRF
- * and HalfSipHash1-3/SipHash1-3 for an insecure PRF only suitable for
- * hashtables.
- */
- #include <linux/siphash.h>
- #include <asm/unaligned.h>
- #if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64
- #include <linux/dcache.h>
- #include <asm/word-at-a-time.h>
- #endif
- #define SIPROUND \
- do { \
- v0 += v1; v1 = rol64(v1, 13); v1 ^= v0; v0 = rol64(v0, 32); \
- v2 += v3; v3 = rol64(v3, 16); v3 ^= v2; \
- v0 += v3; v3 = rol64(v3, 21); v3 ^= v0; \
- v2 += v1; v1 = rol64(v1, 17); v1 ^= v2; v2 = rol64(v2, 32); \
- } while (0)
- #define PREAMBLE(len) \
- u64 v0 = 0x736f6d6570736575ULL; \
- u64 v1 = 0x646f72616e646f6dULL; \
- u64 v2 = 0x6c7967656e657261ULL; \
- u64 v3 = 0x7465646279746573ULL; \
- u64 b = ((u64)(len)) << 56; \
- v3 ^= key->key[1]; \
- v2 ^= key->key[0]; \
- v1 ^= key->key[1]; \
- v0 ^= key->key[0];
- #define POSTAMBLE \
- v3 ^= b; \
- SIPROUND; \
- SIPROUND; \
- v0 ^= b; \
- v2 ^= 0xff; \
- SIPROUND; \
- SIPROUND; \
- SIPROUND; \
- SIPROUND; \
- return (v0 ^ v1) ^ (v2 ^ v3);
- u64 __siphash_aligned(const void *data, size_t len, const siphash_key_t *key)
- {
- const u8 *end = data + len - (len % sizeof(u64));
- const u8 left = len & (sizeof(u64) - 1);
- u64 m;
- PREAMBLE(len)
- for (; data != end; data += sizeof(u64)) {
- m = le64_to_cpup(data);
- v3 ^= m;
- SIPROUND;
- SIPROUND;
- v0 ^= m;
- }
- #if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64
- if (left)
- b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) &
- bytemask_from_count(left)));
- #else
- switch (left) {
- case 7: b |= ((u64)end[6]) << 48;
- case 6: b |= ((u64)end[5]) << 40;
- case 5: b |= ((u64)end[4]) << 32;
- case 4: b |= le32_to_cpup(data); break;
- case 3: b |= ((u64)end[2]) << 16;
- case 2: b |= le16_to_cpup(data); break;
- case 1: b |= end[0];
- }
- #endif
- POSTAMBLE
- }
- EXPORT_SYMBOL(__siphash_aligned);
- #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
- u64 __siphash_unaligned(const void *data, size_t len, const siphash_key_t *key)
- {
- const u8 *end = data + len - (len % sizeof(u64));
- const u8 left = len & (sizeof(u64) - 1);
- u64 m;
- PREAMBLE(len)
- for (; data != end; data += sizeof(u64)) {
- m = get_unaligned_le64(data);
- v3 ^= m;
- SIPROUND;
- SIPROUND;
- v0 ^= m;
- }
- #if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64
- if (left)
- b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) &
- bytemask_from_count(left)));
- #else
- switch (left) {
- case 7: b |= ((u64)end[6]) << 48;
- case 6: b |= ((u64)end[5]) << 40;
- case 5: b |= ((u64)end[4]) << 32;
- case 4: b |= get_unaligned_le32(end); break;
- case 3: b |= ((u64)end[2]) << 16;
- case 2: b |= get_unaligned_le16(end); break;
- case 1: b |= end[0];
- }
- #endif
- POSTAMBLE
- }
- EXPORT_SYMBOL(__siphash_unaligned);
- #endif
- /**
- * siphash_1u64 - compute 64-bit siphash PRF value of a u64
- * @first: first u64
- * @key: the siphash key
- */
- u64 siphash_1u64(const u64 first, const siphash_key_t *key)
- {
- PREAMBLE(8)
- v3 ^= first;
- SIPROUND;
- SIPROUND;
- v0 ^= first;
- POSTAMBLE
- }
- EXPORT_SYMBOL(siphash_1u64);
- /**
- * siphash_2u64 - compute 64-bit siphash PRF value of 2 u64
- * @first: first u64
- * @second: second u64
- * @key: the siphash key
- */
- u64 siphash_2u64(const u64 first, const u64 second, const siphash_key_t *key)
- {
- PREAMBLE(16)
- v3 ^= first;
- SIPROUND;
- SIPROUND;
- v0 ^= first;
- v3 ^= second;
- SIPROUND;
- SIPROUND;
- v0 ^= second;
- POSTAMBLE
- }
- EXPORT_SYMBOL(siphash_2u64);
- /**
- * siphash_3u64 - compute 64-bit siphash PRF value of 3 u64
- * @first: first u64
- * @second: second u64
- * @third: third u64
- * @key: the siphash key
- */
- u64 siphash_3u64(const u64 first, const u64 second, const u64 third,
- const siphash_key_t *key)
- {
- PREAMBLE(24)
- v3 ^= first;
- SIPROUND;
- SIPROUND;
- v0 ^= first;
- v3 ^= second;
- SIPROUND;
- SIPROUND;
- v0 ^= second;
- v3 ^= third;
- SIPROUND;
- SIPROUND;
- v0 ^= third;
- POSTAMBLE
- }
- EXPORT_SYMBOL(siphash_3u64);
- /**
- * siphash_4u64 - compute 64-bit siphash PRF value of 4 u64
- * @first: first u64
- * @second: second u64
- * @third: third u64
- * @forth: forth u64
- * @key: the siphash key
- */
- u64 siphash_4u64(const u64 first, const u64 second, const u64 third,
- const u64 forth, const siphash_key_t *key)
- {
- PREAMBLE(32)
- v3 ^= first;
- SIPROUND;
- SIPROUND;
- v0 ^= first;
- v3 ^= second;
- SIPROUND;
- SIPROUND;
- v0 ^= second;
- v3 ^= third;
- SIPROUND;
- SIPROUND;
- v0 ^= third;
- v3 ^= forth;
- SIPROUND;
- SIPROUND;
- v0 ^= forth;
- POSTAMBLE
- }
- EXPORT_SYMBOL(siphash_4u64);
- u64 siphash_1u32(const u32 first, const siphash_key_t *key)
- {
- PREAMBLE(4)
- b |= first;
- POSTAMBLE
- }
- EXPORT_SYMBOL(siphash_1u32);
- u64 siphash_3u32(const u32 first, const u32 second, const u32 third,
- const siphash_key_t *key)
- {
- u64 combined = (u64)second << 32 | first;
- PREAMBLE(12)
- v3 ^= combined;
- SIPROUND;
- SIPROUND;
- v0 ^= combined;
- b |= third;
- POSTAMBLE
- }
- EXPORT_SYMBOL(siphash_3u32);
- #if BITS_PER_LONG == 64
- /* Note that on 64-bit, we make HalfSipHash1-3 actually be SipHash1-3, for
- * performance reasons. On 32-bit, below, we actually implement HalfSipHash1-3.
- */
- #define HSIPROUND SIPROUND
- #define HPREAMBLE(len) PREAMBLE(len)
- #define HPOSTAMBLE \
- v3 ^= b; \
- HSIPROUND; \
- v0 ^= b; \
- v2 ^= 0xff; \
- HSIPROUND; \
- HSIPROUND; \
- HSIPROUND; \
- return (v0 ^ v1) ^ (v2 ^ v3);
- u32 __hsiphash_aligned(const void *data, size_t len, const hsiphash_key_t *key)
- {
- const u8 *end = data + len - (len % sizeof(u64));
- const u8 left = len & (sizeof(u64) - 1);
- u64 m;
- HPREAMBLE(len)
- for (; data != end; data += sizeof(u64)) {
- m = le64_to_cpup(data);
- v3 ^= m;
- HSIPROUND;
- v0 ^= m;
- }
- #if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64
- if (left)
- b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) &
- bytemask_from_count(left)));
- #else
- switch (left) {
- case 7: b |= ((u64)end[6]) << 48;
- case 6: b |= ((u64)end[5]) << 40;
- case 5: b |= ((u64)end[4]) << 32;
- case 4: b |= le32_to_cpup(data); break;
- case 3: b |= ((u64)end[2]) << 16;
- case 2: b |= le16_to_cpup(data); break;
- case 1: b |= end[0];
- }
- #endif
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(__hsiphash_aligned);
- #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
- u32 __hsiphash_unaligned(const void *data, size_t len,
- const hsiphash_key_t *key)
- {
- const u8 *end = data + len - (len % sizeof(u64));
- const u8 left = len & (sizeof(u64) - 1);
- u64 m;
- HPREAMBLE(len)
- for (; data != end; data += sizeof(u64)) {
- m = get_unaligned_le64(data);
- v3 ^= m;
- HSIPROUND;
- v0 ^= m;
- }
- #if defined(CONFIG_DCACHE_WORD_ACCESS) && BITS_PER_LONG == 64
- if (left)
- b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) &
- bytemask_from_count(left)));
- #else
- switch (left) {
- case 7: b |= ((u64)end[6]) << 48;
- case 6: b |= ((u64)end[5]) << 40;
- case 5: b |= ((u64)end[4]) << 32;
- case 4: b |= get_unaligned_le32(end); break;
- case 3: b |= ((u64)end[2]) << 16;
- case 2: b |= get_unaligned_le16(end); break;
- case 1: b |= end[0];
- }
- #endif
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(__hsiphash_unaligned);
- #endif
- /**
- * hsiphash_1u32 - compute 64-bit hsiphash PRF value of a u32
- * @first: first u32
- * @key: the hsiphash key
- */
- u32 hsiphash_1u32(const u32 first, const hsiphash_key_t *key)
- {
- HPREAMBLE(4)
- b |= first;
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(hsiphash_1u32);
- /**
- * hsiphash_2u32 - compute 32-bit hsiphash PRF value of 2 u32
- * @first: first u32
- * @second: second u32
- * @key: the hsiphash key
- */
- u32 hsiphash_2u32(const u32 first, const u32 second, const hsiphash_key_t *key)
- {
- u64 combined = (u64)second << 32 | first;
- HPREAMBLE(8)
- v3 ^= combined;
- HSIPROUND;
- v0 ^= combined;
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(hsiphash_2u32);
- /**
- * hsiphash_3u32 - compute 32-bit hsiphash PRF value of 3 u32
- * @first: first u32
- * @second: second u32
- * @third: third u32
- * @key: the hsiphash key
- */
- u32 hsiphash_3u32(const u32 first, const u32 second, const u32 third,
- const hsiphash_key_t *key)
- {
- u64 combined = (u64)second << 32 | first;
- HPREAMBLE(12)
- v3 ^= combined;
- HSIPROUND;
- v0 ^= combined;
- b |= third;
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(hsiphash_3u32);
- /**
- * hsiphash_4u32 - compute 32-bit hsiphash PRF value of 4 u32
- * @first: first u32
- * @second: second u32
- * @third: third u32
- * @forth: forth u32
- * @key: the hsiphash key
- */
- u32 hsiphash_4u32(const u32 first, const u32 second, const u32 third,
- const u32 forth, const hsiphash_key_t *key)
- {
- u64 combined = (u64)second << 32 | first;
- HPREAMBLE(16)
- v3 ^= combined;
- HSIPROUND;
- v0 ^= combined;
- combined = (u64)forth << 32 | third;
- v3 ^= combined;
- HSIPROUND;
- v0 ^= combined;
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(hsiphash_4u32);
- #else
- #define HSIPROUND \
- do { \
- v0 += v1; v1 = rol32(v1, 5); v1 ^= v0; v0 = rol32(v0, 16); \
- v2 += v3; v3 = rol32(v3, 8); v3 ^= v2; \
- v0 += v3; v3 = rol32(v3, 7); v3 ^= v0; \
- v2 += v1; v1 = rol32(v1, 13); v1 ^= v2; v2 = rol32(v2, 16); \
- } while (0)
- #define HPREAMBLE(len) \
- u32 v0 = 0; \
- u32 v1 = 0; \
- u32 v2 = 0x6c796765U; \
- u32 v3 = 0x74656462U; \
- u32 b = ((u32)(len)) << 24; \
- v3 ^= key->key[1]; \
- v2 ^= key->key[0]; \
- v1 ^= key->key[1]; \
- v0 ^= key->key[0];
- #define HPOSTAMBLE \
- v3 ^= b; \
- HSIPROUND; \
- v0 ^= b; \
- v2 ^= 0xff; \
- HSIPROUND; \
- HSIPROUND; \
- HSIPROUND; \
- return v1 ^ v3;
- u32 __hsiphash_aligned(const void *data, size_t len, const hsiphash_key_t *key)
- {
- const u8 *end = data + len - (len % sizeof(u32));
- const u8 left = len & (sizeof(u32) - 1);
- u32 m;
- HPREAMBLE(len)
- for (; data != end; data += sizeof(u32)) {
- m = le32_to_cpup(data);
- v3 ^= m;
- HSIPROUND;
- v0 ^= m;
- }
- switch (left) {
- case 3: b |= ((u32)end[2]) << 16;
- case 2: b |= le16_to_cpup(data); break;
- case 1: b |= end[0];
- }
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(__hsiphash_aligned);
- #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
- u32 __hsiphash_unaligned(const void *data, size_t len,
- const hsiphash_key_t *key)
- {
- const u8 *end = data + len - (len % sizeof(u32));
- const u8 left = len & (sizeof(u32) - 1);
- u32 m;
- HPREAMBLE(len)
- for (; data != end; data += sizeof(u32)) {
- m = get_unaligned_le32(data);
- v3 ^= m;
- HSIPROUND;
- v0 ^= m;
- }
- switch (left) {
- case 3: b |= ((u32)end[2]) << 16;
- case 2: b |= get_unaligned_le16(end); break;
- case 1: b |= end[0];
- }
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(__hsiphash_unaligned);
- #endif
- /**
- * hsiphash_1u32 - compute 32-bit hsiphash PRF value of a u32
- * @first: first u32
- * @key: the hsiphash key
- */
- u32 hsiphash_1u32(const u32 first, const hsiphash_key_t *key)
- {
- HPREAMBLE(4)
- v3 ^= first;
- HSIPROUND;
- v0 ^= first;
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(hsiphash_1u32);
- /**
- * hsiphash_2u32 - compute 32-bit hsiphash PRF value of 2 u32
- * @first: first u32
- * @second: second u32
- * @key: the hsiphash key
- */
- u32 hsiphash_2u32(const u32 first, const u32 second, const hsiphash_key_t *key)
- {
- HPREAMBLE(8)
- v3 ^= first;
- HSIPROUND;
- v0 ^= first;
- v3 ^= second;
- HSIPROUND;
- v0 ^= second;
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(hsiphash_2u32);
- /**
- * hsiphash_3u32 - compute 32-bit hsiphash PRF value of 3 u32
- * @first: first u32
- * @second: second u32
- * @third: third u32
- * @key: the hsiphash key
- */
- u32 hsiphash_3u32(const u32 first, const u32 second, const u32 third,
- const hsiphash_key_t *key)
- {
- HPREAMBLE(12)
- v3 ^= first;
- HSIPROUND;
- v0 ^= first;
- v3 ^= second;
- HSIPROUND;
- v0 ^= second;
- v3 ^= third;
- HSIPROUND;
- v0 ^= third;
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(hsiphash_3u32);
- /**
- * hsiphash_4u32 - compute 32-bit hsiphash PRF value of 4 u32
- * @first: first u32
- * @second: second u32
- * @third: third u32
- * @forth: forth u32
- * @key: the hsiphash key
- */
- u32 hsiphash_4u32(const u32 first, const u32 second, const u32 third,
- const u32 forth, const hsiphash_key_t *key)
- {
- HPREAMBLE(16)
- v3 ^= first;
- HSIPROUND;
- v0 ^= first;
- v3 ^= second;
- HSIPROUND;
- v0 ^= second;
- v3 ^= third;
- HSIPROUND;
- v0 ^= third;
- v3 ^= forth;
- HSIPROUND;
- v0 ^= forth;
- HPOSTAMBLE
- }
- EXPORT_SYMBOL(hsiphash_4u32);
- #endif
|