aboutsummaryrefslogtreecommitdiff
path: root/core/hash/xxhash/xxhash_64.odin
diff options
context:
space:
mode:
authorJeroen van Rijn <Kelimion@users.noreply.github.com>2021-09-09 05:34:13 +0200
committerJeroen van Rijn <Kelimion@users.noreply.github.com>2021-09-09 15:26:57 +0200
commitf5d5417af75dba8fa2c631f9c101ba2a43bb38f3 (patch)
treee3daa06faa23e483d75878ea242c01dcf3914d2a /core/hash/xxhash/xxhash_64.odin
parente65e0b5db2ab9880d278555798194b875025899a (diff)
xxhash: Initial implementations.
Diffstat (limited to 'core/hash/xxhash/xxhash_64.odin')
-rw-r--r--core/hash/xxhash/xxhash_64.odin294
1 files changed, 294 insertions, 0 deletions
diff --git a/core/hash/xxhash/xxhash_64.odin b/core/hash/xxhash/xxhash_64.odin
new file mode 100644
index 000000000..d535a134c
--- /dev/null
+++ b/core/hash/xxhash/xxhash_64.odin
@@ -0,0 +1,294 @@
+/*
+ An implementation of Yann Collet's [xxhash Fast Hash Algorithm](https://cyan4973.github.io/xxHash/).
+ Copyright 2021 Jeroen van Rijn <nom@duclavier.com>.
+
+ Made available under Odin's BSD-3 license, based on the original C code.
+
+ List of contributors:
+ Jeroen van Rijn: Initial implementation.
+*/
+package xxhash
+
+import "core:intrinsics"
+
+/*
+ 64-bit hash functions
+*/
+XXH64_hash :: u64
+xxh_u64 :: u64
+XXH64_DEFAULT_SEED :: XXH64_hash(0)
+
+XXH64_state :: struct {
+ total_len: XXH64_hash, /*!< Total length hashed. This is always 64-bit. */
+ v1: XXH64_hash, /*!< First accumulator lane */
+ v2: XXH64_hash, /*!< Second accumulator lane */
+ v3: XXH64_hash, /*!< Third accumulator lane */
+ v4: XXH64_hash, /*!< Fourth accumulator lane */
+ mem64: [4]XXH64_hash, /*!< Internal buffer for partial reads. Treated as unsigned char[32]. */
+ memsize: XXH32_hash, /*!< Amount of data in @ref mem64 */
+ reserved32: XXH32_hash, /*!< Reserved field, needed for padding anyways*/
+ reserved64: XXH64_hash, /*!< Reserved field. Do not read or write to it, it may be removed. */
+}
+
+XXH64_canonical :: struct {
+ digest: [8]u8,
+}
+
+XXH_PRIME64_1 :: 0x9E3779B185EBCA87 /*!< 0b1001111000110111011110011011000110000101111010111100101010000111 */
+XXH_PRIME64_2 :: 0xC2B2AE3D27D4EB4F /*!< 0b1100001010110010101011100011110100100111110101001110101101001111 */
+XXH_PRIME64_3 :: 0x165667B19E3779F9 /*!< 0b0001011001010110011001111011000110011110001101110111100111111001 */
+XXH_PRIME64_4 :: 0x85EBCA77C2B2AE63 /*!< 0b1000010111101011110010100111011111000010101100101010111001100011 */
+XXH_PRIME64_5 :: 0x27D4EB2F165667C5 /*!< 0b0010011111010100111010110010111100010110010101100110011111000101 */
+
+@(optimization_mode="speed")
+XXH64_round :: proc(acc, input: xxh_u64) -> (res: xxh_u64) {
+ acc := acc
+
+ acc += input * XXH_PRIME64_2
+ acc = XXH_rotl64(acc, 31)
+ acc *= XXH_PRIME64_1
+ return acc
+}
+
+@(optimization_mode="speed")
+XXH64_mergeRound :: proc(acc, val: xxh_u64) -> (res: xxh_u64) {
+ res = acc ~ XXH64_round(0, val)
+ res = res * XXH_PRIME64_1 + XXH_PRIME64_4
+ return res
+}
+
+@(optimization_mode="speed")
+XXH64_avalanche :: proc(h64: xxh_u64) -> (res: xxh_u64) {
+ res = h64
+ res ~= res >> 33
+ res *= XXH_PRIME64_2
+ res ~= res >> 29
+ res *= XXH_PRIME64_3
+ res ~= res >> 32
+ return res
+}
+
+@(optimization_mode="speed")
+XXH64_finalize :: proc(h64: xxh_u64, buf: []u8, alignment: Alignment) -> (res: xxh_u64) {
+ buf := buf
+ length := len(buf) & 31
+ res = h64
+
+ for length >= 8 {
+ b := XXH64_read64(buf, alignment)
+ k1 := XXH64_round(0, b)
+ #no_bounds_check buf = buf[8:]
+ res ~= k1
+ res = XXH_rotl64(res, 27) * XXH_PRIME64_1 + XXH_PRIME64_4
+ length -= 8
+ }
+
+ if length >= 4 {
+ res ~= xxh_u64(XXH32_read32(buf, alignment)) * XXH_PRIME64_1
+ #no_bounds_check buf = buf[4:]
+ res = XXH_rotl64(res, 23) * XXH_PRIME64_2 + XXH_PRIME64_3
+ length -= 4
+ }
+
+ for length > 0 {
+ #no_bounds_check b := xxh_u64(buf[0])
+ buf = buf[1:]
+ res ~= b * XXH_PRIME64_5
+ res = XXH_rotl64(res, 11) * XXH_PRIME64_1
+ length -= 1
+ }
+ return XXH64_avalanche(res)
+}
+
+@(optimization_mode="speed")
+XXH64_endian_align :: proc(input: []u8, seed := XXH64_DEFAULT_SEED, alignment := Alignment.Unaligned) -> (res: xxh_u64) {
+ buf := input
+ length := len(buf)
+
+ if length >= 32 {
+ v1 := seed + XXH_PRIME64_1 + XXH_PRIME64_2
+ v2 := seed + XXH_PRIME64_2
+ v3 := seed + 0
+ v4 := seed - XXH_PRIME64_1
+
+ for len(buf) >= 32 {
+ v1 = XXH64_round(v1, XXH64_read64(buf, alignment)); buf = buf[8:]
+ v2 = XXH64_round(v2, XXH64_read64(buf, alignment)); buf = buf[8:]
+ v3 = XXH64_round(v3, XXH64_read64(buf, alignment)); buf = buf[8:]
+ v4 = XXH64_round(v4, XXH64_read64(buf, alignment)); buf = buf[8:]
+ }
+
+ res = XXH_rotl64(v1, 1) + XXH_rotl64(v2, 7) + XXH_rotl64(v3, 12) + XXH_rotl64(v4, 18)
+ res = XXH64_mergeRound(res, v1)
+ res = XXH64_mergeRound(res, v2)
+ res = XXH64_mergeRound(res, v3)
+ res = XXH64_mergeRound(res, v4)
+ } else {
+ res = seed + XXH_PRIME64_5
+ }
+ res += xxh_u64(length)
+
+ return XXH64_finalize(res, buf, alignment)
+}
+
+XXH64 :: proc(input: []u8, seed := XXH64_DEFAULT_SEED) -> (digest: XXH64_hash) {
+ when false {
+ /*
+ Simple version, good for code maintenance, but unfortunately slow for small inputs.
+ */
+ state: XXH64_state
+ XXH64_reset_state(&state, seed)
+ buf := input
+ for len(buf) > 0 {
+ l := min(65536, len(buf))
+ XXH64_update(&state, buf[:l])
+ buf = buf[l:]
+ }
+ return XXH64_digest(&state)
+ } else {
+ when XXH_FORCE_ALIGN_CHECK {
+ if uintptr(raw_data(input)) & uintptr(7) == 0 {
+ /*
+ Input is 8-bytes aligned, leverage the speed benefit.
+ */
+ return XXH64_endian_align(input, seed, .Aligned)
+ }
+ }
+ return XXH64_endian_align(input, seed, .Unaligned)
+ }
+}
+
+/*
+ ****** Hash Streaming ******
+*/
+XXH64_create_state :: proc(allocator := context.allocator) -> (res: ^XXH64_state, err: Error) {
+ state := new(XXH64_state, allocator)
+ return state, nil if state != nil else .Error
+}
+
+XXH64_destroy_state :: proc(state: ^XXH64_state, allocator := context.allocator) -> (err: Error) {
+ free(state, allocator)
+ return nil
+}
+
+XXH64_copy_state :: proc(dest, src: ^XXH64_state) {
+ assert(dest != nil && src != nil)
+ mem_copy(dest, src, size_of(XXH64_state))
+}
+
+XXH64_reset_state :: proc(state_ptr: ^XXH64_state, seed := XXH64_DEFAULT_SEED) -> (err: Error) {
+ state := XXH64_state{}
+
+ state.v1 = seed + XXH_PRIME64_1 + XXH_PRIME64_2
+ state.v2 = seed + XXH_PRIME64_2
+ state.v3 = seed + 0
+ state.v4 = seed - XXH_PRIME64_1
+ /*
+ Fo not write into reserved64, might be removed in a future version.
+ */
+ mem_copy(state_ptr, &state, size_of(state) - size_of(state.reserved64))
+ return nil
+}
+
+@(optimization_mode="speed")
+XXH64_update :: proc(state: ^XXH64_state, input: []u8) -> (err: Error) {
+ buf := input
+ length := len(buf)
+
+ state.total_len += u64(length)
+
+ if state.memsize + u32(length) < 32 { /* fill in tmp buffer */
+ ptr := uintptr(raw_data(state.mem64[:])) + uintptr(state.memsize)
+ mem_copy(rawptr(ptr), raw_data(input), int(length))
+ state.memsize += u32(length)
+ return nil
+ }
+
+ if state.memsize > 0 { /* tmp buffer is full */
+ ptr := uintptr(raw_data(state.mem64[:])) + uintptr(state.memsize)
+ mem_copy(rawptr(ptr), raw_data(input), int(32 - state.memsize))
+ {
+ #no_bounds_check state.v1 = XXH64_round(state.v1, state.mem64[0])
+ #no_bounds_check state.v2 = XXH64_round(state.v2, state.mem64[1])
+ #no_bounds_check state.v3 = XXH64_round(state.v3, state.mem64[2])
+ #no_bounds_check state.v4 = XXH64_round(state.v4, state.mem64[3])
+ }
+ buf = buf[32 - state.memsize:]
+ state.memsize = 0
+ }
+
+ if len(buf) >= 32 {
+ v1 := state.v1
+ v2 := state.v2
+ v3 := state.v3
+ v4 := state.v4
+
+ for len(buf) >= 32 {
+ #no_bounds_check v1 = XXH64_round(v1, XXH64_read64(buf, .Unaligned)); buf = buf[8:]
+ #no_bounds_check v2 = XXH64_round(v2, XXH64_read64(buf, .Unaligned)); buf = buf[8:]
+ #no_bounds_check v3 = XXH64_round(v3, XXH64_read64(buf, .Unaligned)); buf = buf[8:]
+ #no_bounds_check v4 = XXH64_round(v4, XXH64_read64(buf, .Unaligned)); buf = buf[8:]
+ }
+
+ state.v1 = v1
+ state.v2 = v2
+ state.v3 = v3
+ state.v4 = v4
+ }
+
+ length = len(buf)
+ if length > 0 {
+ mem_copy(raw_data(state.mem64[:]), raw_data(buf[:]), int(length))
+ state.memsize = u32(length)
+ }
+ return nil
+}
+
+@(optimization_mode="speed")
+XXH64_digest :: proc(state: ^XXH64_state) -> (res: XXH64_hash) {
+ if state.total_len >= 32 {
+ v1 := state.v1
+ v2 := state.v2
+ v3 := state.v3
+ v4 := state.v4
+
+ res = XXH_rotl64(v1, 1) + XXH_rotl64(v2, 7) + XXH_rotl64(v3, 12) + XXH_rotl64(v4, 18)
+ res = XXH64_mergeRound(res, v1)
+ res = XXH64_mergeRound(res, v2)
+ res = XXH64_mergeRound(res, v3)
+ res = XXH64_mergeRound(res, v4)
+ } else {
+ res = state.v3 /*seed*/ + XXH_PRIME64_5
+ }
+ res += XXH64_hash(state.total_len)
+
+ buf := (^[32]u8)(&state.mem64)^
+ alignment: Alignment = .Aligned if uintptr(&state.mem64) & 15 == 0 else .Unaligned
+ return XXH64_finalize(res, buf[:state.memsize], alignment)
+}
+
+/*
+ ****** Canonical representation ******
+
+ The default return values from XXH functions are unsigned 32 and 64 bit integers.
+
+ The canonical representation uses big endian convention,
+ the same convention as human-readable numbers (large digits first).
+
+ This way, hash values can be written into a file or buffer, remaining
+ comparable across different systems.
+
+ The following functions allow transformation of hash values to and from their
+ canonical format.
+*/
+XXH64_canonical_from_hash :: proc(hash: XXH64_hash) -> (canonical: XXH64_canonical) {
+ #assert(size_of(XXH64_canonical) == size_of(XXH64_hash))
+ h := u64be(hash)
+ mem_copy(&canonical, &h, size_of(canonical))
+ return
+}
+
+XXH64_hash_from_canonical :: proc(canonical: ^XXH64_canonical) -> (hash: XXH64_hash) {
+ h := (^u64be)(&canonical.digest)^
+ return XXH64_hash(h)
+} \ No newline at end of file