lighttpd 1.4.x
https://www.lighttpd.net/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
5672 lines
207 KiB
5672 lines
207 KiB
#ifndef INCLUDED_ALGO_XXHASH_H |
|
#define INCLUDED_ALGO_XXHASH_H |
|
#include "first.h" |
|
|
|
/*(lighttpd customization)*/ |
|
/*#define XXH_NO_INLINE_HINTS 1*/ |
|
/*#define XXH_REROLL 1*/ |
|
#define XXH_NO_LONG_LONG |
|
|
|
|
|
#ifdef HAVE_XXHASH_H |
|
|
|
#include <xxhash.h> |
|
|
|
#else /* ! HAVE_XXHASH_H */ |
|
|
|
/* |
|
* xxHash - Extremely Fast Hash algorithm |
|
* Header File |
|
* Copyright (C) 2012-2020 Yann Collet |
|
* |
|
* BSD 2-Clause License (https://www.opensource.org/licenses/bsd-license.php) |
|
* |
|
* Redistribution and use in source and binary forms, with or without |
|
* modification, are permitted provided that the following conditions are |
|
* met: |
|
* |
|
* * Redistributions of source code must retain the above copyright |
|
* notice, this list of conditions and the following disclaimer. |
|
* * Redistributions in binary form must reproduce the above |
|
* copyright notice, this list of conditions and the following disclaimer |
|
* in the documentation and/or other materials provided with the |
|
* distribution. |
|
* |
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
|
* |
|
* You can contact the author at: |
|
* - xxHash homepage: https://www.xxhash.com |
|
* - xxHash source repository: https://github.com/Cyan4973/xxHash |
|
*/ |
|
/*! |
|
* @mainpage xxHash |
|
* |
|
* @file xxhash.h |
|
* xxHash prototypes and implementation |
|
*/ |
|
/* TODO: update */ |
|
/* Notice extracted from xxHash homepage: |
|
|
|
xxHash is an extremely fast hash algorithm, running at RAM speed limits. |
|
It also successfully passes all tests from the SMHasher suite. |
|
|
|
Comparison (single thread, Windows Seven 32 bits, using SMHasher on a Core 2 Duo @3GHz) |
|
|
|
Name Speed Q.Score Author |
|
xxHash 5.4 GB/s 10 |
|
CrapWow 3.2 GB/s 2 Andrew |
|
MurmurHash 3a 2.7 GB/s 10 Austin Appleby |
|
SpookyHash 2.0 GB/s 10 Bob Jenkins |
|
SBox 1.4 GB/s 9 Bret Mulvey |
|
Lookup3 1.2 GB/s 9 Bob Jenkins |
|
SuperFastHash 1.2 GB/s 1 Paul Hsieh |
|
CityHash64 1.05 GB/s 10 Pike & Alakuijala |
|
FNV 0.55 GB/s 5 Fowler, Noll, Vo |
|
CRC32 0.43 GB/s 9 |
|
MD5-32 0.33 GB/s 10 Ronald L. Rivest |
|
SHA1-32 0.28 GB/s 10 |
|
|
|
Q.Score is a measure of quality of the hash function. |
|
It depends on successfully passing SMHasher test set. |
|
10 is a perfect score. |
|
|
|
Note: SMHasher's CRC32 implementation is not the fastest one. |
|
Other speed-oriented implementations can be faster, |
|
especially in combination with PCLMUL instruction: |
|
https://fastcompression.blogspot.com/2019/03/presenting-xxh3.html?showComment=1552696407071#c3490092340461170735 |
|
|
|
A 64-bit version, named XXH64, is available since r35. |
|
It offers much better speed, but for 64-bit applications only. |
|
Name Speed on 64 bits Speed on 32 bits |
|
XXH64 13.8 GB/s 1.9 GB/s |
|
XXH32 6.8 GB/s 6.0 GB/s |
|
*/ |
|
|
|
#if defined (__cplusplus) |
|
extern "C" { |
|
#endif |
|
|
|
/* **************************** |
|
* INLINE mode |
|
******************************/ |
|
/*! |
|
* XXH_INLINE_ALL (and XXH_PRIVATE_API) |
|
* Use these build macros to inline xxhash into the target unit. |
|
* Inlining improves performance on small inputs, especially when the length is |
|
* expressed as a compile-time constant: |
|
* |
|
* https://fastcompression.blogspot.com/2018/03/xxhash-for-small-keys-impressive-power.html |
|
* |
|
* It also keeps xxHash symbols private to the unit, so they are not exported. |
|
* |
|
* Usage: |
|
* #define XXH_INLINE_ALL |
|
* #include "xxhash.h" |
|
* |
|
* Do not compile and link xxhash.o as a separate object, as it is not useful. |
|
*/ |
|
#if (defined(XXH_INLINE_ALL) || defined(XXH_PRIVATE_API)) \ |
|
&& !defined(XXH_INLINE_ALL_31684351384) |
|
/* this section should be traversed only once */ |
|
# define XXH_INLINE_ALL_31684351384 |
|
/* give access to the advanced API, required to compile implementations */ |
|
# undef XXH_STATIC_LINKING_ONLY /* avoid macro redef */ |
|
# define XXH_STATIC_LINKING_ONLY |
|
/* make all functions private */ |
|
# undef XXH_PUBLIC_API |
|
# if defined(__GNUC__) |
|
# define XXH_PUBLIC_API static __inline __attribute__((unused)) |
|
# elif defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) |
|
# define XXH_PUBLIC_API static inline |
|
# elif defined(_MSC_VER) |
|
# define XXH_PUBLIC_API static __inline |
|
# else |
|
/* note: this version may generate warnings for unused static functions */ |
|
# define XXH_PUBLIC_API static |
|
# endif |
|
|
|
/* |
|
* This part deals with the special case where a unit wants to inline xxHash, |
|
* but "xxhash.h" has previously been included without XXH_INLINE_ALL, |
|
* such as part of some previously included *.h header file. |
|
* Without further action, the new include would just be ignored, |
|
* and functions would effectively _not_ be inlined (silent failure). |
|
* The following macros solve this situation by prefixing all inlined names, |
|
* avoiding naming collision with previous inclusions. |
|
*/ |
|
/* Before that, we unconditionally #undef all symbols, |
|
* in case they were already defined with XXH_NAMESPACE. |
|
* They will then be redefined for XXH_INLINE_ALL |
|
*/ |
|
# undef XXH_versionNumber |
|
/* XXH32 */ |
|
# undef XXH32 |
|
# undef XXH32_createState |
|
# undef XXH32_freeState |
|
# undef XXH32_reset |
|
# undef XXH32_update |
|
# undef XXH32_digest |
|
# undef XXH32_copyState |
|
# undef XXH32_canonicalFromHash |
|
# undef XXH32_hashFromCanonical |
|
/* XXH64 */ |
|
# undef XXH64 |
|
# undef XXH64_createState |
|
# undef XXH64_freeState |
|
# undef XXH64_reset |
|
# undef XXH64_update |
|
# undef XXH64_digest |
|
# undef XXH64_copyState |
|
# undef XXH64_canonicalFromHash |
|
# undef XXH64_hashFromCanonical |
|
/* XXH3_64bits */ |
|
# undef XXH3_64bits |
|
# undef XXH3_64bits_withSecret |
|
# undef XXH3_64bits_withSeed |
|
# undef XXH3_64bits_withSecretandSeed |
|
# undef XXH3_createState |
|
# undef XXH3_freeState |
|
# undef XXH3_copyState |
|
# undef XXH3_64bits_reset |
|
# undef XXH3_64bits_reset_withSeed |
|
# undef XXH3_64bits_reset_withSecret |
|
# undef XXH3_64bits_update |
|
# undef XXH3_64bits_digest |
|
# undef XXH3_generateSecret |
|
/* XXH3_128bits */ |
|
# undef XXH128 |
|
# undef XXH3_128bits |
|
# undef XXH3_128bits_withSeed |
|
# undef XXH3_128bits_withSecret |
|
# undef XXH3_128bits_reset |
|
# undef XXH3_128bits_reset_withSeed |
|
# undef XXH3_128bits_reset_withSecret |
|
# undef XXH3_128bits_reset_withSecretandSeed |
|
# undef XXH3_128bits_update |
|
# undef XXH3_128bits_digest |
|
# undef XXH128_isEqual |
|
# undef XXH128_cmp |
|
# undef XXH128_canonicalFromHash |
|
# undef XXH128_hashFromCanonical |
|
/* Finally, free the namespace itself */ |
|
# undef XXH_NAMESPACE |
|
|
|
/* employ the namespace for XXH_INLINE_ALL */ |
|
# define XXH_NAMESPACE XXH_INLINE_ |
|
/* |
|
* Some identifiers (enums, type names) are not symbols, |
|
* but they must nonetheless be renamed to avoid redeclaration. |
|
* Alternative solution: do not redeclare them. |
|
* However, this requires some #ifdefs, and has a more dispersed impact. |
|
* Meanwhile, renaming can be achieved in a single place. |
|
*/ |
|
# define XXH_IPREF(Id) XXH_NAMESPACE ## Id |
|
# define XXH_OK XXH_IPREF(XXH_OK) |
|
# define XXH_ERROR XXH_IPREF(XXH_ERROR) |
|
# define XXH_errorcode XXH_IPREF(XXH_errorcode) |
|
# define XXH32_canonical_t XXH_IPREF(XXH32_canonical_t) |
|
# define XXH64_canonical_t XXH_IPREF(XXH64_canonical_t) |
|
# define XXH128_canonical_t XXH_IPREF(XXH128_canonical_t) |
|
# define XXH32_state_s XXH_IPREF(XXH32_state_s) |
|
# define XXH32_state_t XXH_IPREF(XXH32_state_t) |
|
# define XXH64_state_s XXH_IPREF(XXH64_state_s) |
|
# define XXH64_state_t XXH_IPREF(XXH64_state_t) |
|
# define XXH3_state_s XXH_IPREF(XXH3_state_s) |
|
# define XXH3_state_t XXH_IPREF(XXH3_state_t) |
|
# define XXH128_hash_t XXH_IPREF(XXH128_hash_t) |
|
/* Ensure the header is parsed again, even if it was previously included */ |
|
# undef XXHASH_H_5627135585666179 |
|
# undef XXHASH_H_STATIC_13879238742 |
|
#endif /* XXH_INLINE_ALL || XXH_PRIVATE_API */ |
|
|
|
|
|
|
|
/* **************************************************************** |
|
* Stable API |
|
*****************************************************************/ |
|
#ifndef XXHASH_H_5627135585666179 |
|
#define XXHASH_H_5627135585666179 1 |
|
|
|
|
|
/*! |
|
* @defgroup public Public API |
|
* Contains details on the public xxHash functions. |
|
* @{ |
|
*/ |
|
/* specific declaration modes for Windows */ |
|
#if !defined(XXH_INLINE_ALL) && !defined(XXH_PRIVATE_API) |
|
# if defined(WIN32) && defined(_MSC_VER) && (defined(XXH_IMPORT) || defined(XXH_EXPORT)) |
|
# ifdef XXH_EXPORT |
|
# define XXH_PUBLIC_API __declspec(dllexport) |
|
# elif XXH_IMPORT |
|
# define XXH_PUBLIC_API __declspec(dllimport) |
|
# endif |
|
# else |
|
# define XXH_PUBLIC_API /* do nothing */ |
|
# endif |
|
#endif |
|
|
|
#ifdef XXH_DOXYGEN |
|
/*! |
|
* @brief Emulate a namespace by transparently prefixing all symbols. |
|
* |
|
* If you want to include _and expose_ xxHash functions from within your own |
|
* library, but also want to avoid symbol collisions with other libraries which |
|
* may also include xxHash, you can use XXH_NAMESPACE to automatically prefix |
|
* any public symbol from xxhash library with the value of XXH_NAMESPACE |
|
* (therefore, avoid empty or numeric values). |
|
* |
|
* Note that no change is required within the calling program as long as it |
|
* includes `xxhash.h`: Regular symbol names will be automatically translated |
|
* by this header. |
|
*/ |
|
# define XXH_NAMESPACE /* YOUR NAME HERE */ |
|
# undef XXH_NAMESPACE |
|
#endif |
|
|
|
#ifdef XXH_NAMESPACE |
|
# define XXH_CAT(A,B) A##B |
|
# define XXH_NAME2(A,B) XXH_CAT(A,B) |
|
# define XXH_versionNumber XXH_NAME2(XXH_NAMESPACE, XXH_versionNumber) |
|
/* XXH32 */ |
|
# define XXH32 XXH_NAME2(XXH_NAMESPACE, XXH32) |
|
# define XXH32_createState XXH_NAME2(XXH_NAMESPACE, XXH32_createState) |
|
# define XXH32_freeState XXH_NAME2(XXH_NAMESPACE, XXH32_freeState) |
|
# define XXH32_reset XXH_NAME2(XXH_NAMESPACE, XXH32_reset) |
|
# define XXH32_update XXH_NAME2(XXH_NAMESPACE, XXH32_update) |
|
# define XXH32_digest XXH_NAME2(XXH_NAMESPACE, XXH32_digest) |
|
# define XXH32_copyState XXH_NAME2(XXH_NAMESPACE, XXH32_copyState) |
|
# define XXH32_canonicalFromHash XXH_NAME2(XXH_NAMESPACE, XXH32_canonicalFromHash) |
|
# define XXH32_hashFromCanonical XXH_NAME2(XXH_NAMESPACE, XXH32_hashFromCanonical) |
|
/* XXH64 */ |
|
# define XXH64 XXH_NAME2(XXH_NAMESPACE, XXH64) |
|
# define XXH64_createState XXH_NAME2(XXH_NAMESPACE, XXH64_createState) |
|
# define XXH64_freeState XXH_NAME2(XXH_NAMESPACE, XXH64_freeState) |
|
# define XXH64_reset XXH_NAME2(XXH_NAMESPACE, XXH64_reset) |
|
# define XXH64_update XXH_NAME2(XXH_NAMESPACE, XXH64_update) |
|
# define XXH64_digest XXH_NAME2(XXH_NAMESPACE, XXH64_digest) |
|
# define XXH64_copyState XXH_NAME2(XXH_NAMESPACE, XXH64_copyState) |
|
# define XXH64_canonicalFromHash XXH_NAME2(XXH_NAMESPACE, XXH64_canonicalFromHash) |
|
# define XXH64_hashFromCanonical XXH_NAME2(XXH_NAMESPACE, XXH64_hashFromCanonical) |
|
/* XXH3_64bits */ |
|
# define XXH3_64bits XXH_NAME2(XXH_NAMESPACE, XXH3_64bits) |
|
# define XXH3_64bits_withSecret XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_withSecret) |
|
# define XXH3_64bits_withSeed XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_withSeed) |
|
# define XXH3_64bits_withSecretandSeed XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_withSecretandSeed) |
|
# define XXH3_createState XXH_NAME2(XXH_NAMESPACE, XXH3_createState) |
|
# define XXH3_freeState XXH_NAME2(XXH_NAMESPACE, XXH3_freeState) |
|
# define XXH3_copyState XXH_NAME2(XXH_NAMESPACE, XXH3_copyState) |
|
# define XXH3_64bits_reset XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_reset) |
|
# define XXH3_64bits_reset_withSeed XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_reset_withSeed) |
|
# define XXH3_64bits_reset_withSecret XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_reset_withSecret) |
|
# define XXH3_64bits_reset_withSecretandSeed XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_reset_withSecretandSeed) |
|
# define XXH3_64bits_update XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_update) |
|
# define XXH3_64bits_digest XXH_NAME2(XXH_NAMESPACE, XXH3_64bits_digest) |
|
# define XXH3_generateSecret XXH_NAME2(XXH_NAMESPACE, XXH3_generateSecret) |
|
# define XXH3_generateSecret_fromSeed XXH_NAME2(XXH_NAMESPACE, XXH3_generateSecret_fromSeed) |
|
/* XXH3_128bits */ |
|
# define XXH128 XXH_NAME2(XXH_NAMESPACE, XXH128) |
|
# define XXH3_128bits XXH_NAME2(XXH_NAMESPACE, XXH3_128bits) |
|
# define XXH3_128bits_withSeed XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_withSeed) |
|
# define XXH3_128bits_withSecret XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_withSecret) |
|
# define XXH3_128bits_withSecretandSeed XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_withSecretandSeed) |
|
# define XXH3_128bits_reset XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_reset) |
|
# define XXH3_128bits_reset_withSeed XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_reset_withSeed) |
|
# define XXH3_128bits_reset_withSecret XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_reset_withSecret) |
|
# define XXH3_128bits_reset_withSecretandSeed XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_reset_withSecretandSeed) |
|
# define XXH3_128bits_update XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_update) |
|
# define XXH3_128bits_digest XXH_NAME2(XXH_NAMESPACE, XXH3_128bits_digest) |
|
# define XXH128_isEqual XXH_NAME2(XXH_NAMESPACE, XXH128_isEqual) |
|
# define XXH128_cmp XXH_NAME2(XXH_NAMESPACE, XXH128_cmp) |
|
# define XXH128_canonicalFromHash XXH_NAME2(XXH_NAMESPACE, XXH128_canonicalFromHash) |
|
# define XXH128_hashFromCanonical XXH_NAME2(XXH_NAMESPACE, XXH128_hashFromCanonical) |
|
#endif |
|
|
|
|
|
/* ************************************* |
|
* Version |
|
***************************************/ |
|
#define XXH_VERSION_MAJOR 0 |
|
#define XXH_VERSION_MINOR 8 |
|
#define XXH_VERSION_RELEASE 1 |
|
#define XXH_VERSION_NUMBER (XXH_VERSION_MAJOR *100*100 + XXH_VERSION_MINOR *100 + XXH_VERSION_RELEASE) |
|
|
|
/*! |
|
* @brief Obtains the xxHash version. |
|
* |
|
* This is mostly useful when xxHash is compiled as a shared library, |
|
* since the returned value comes from the library, as opposed to header file. |
|
* |
|
* @return `XXH_VERSION_NUMBER` of the invoked library. |
|
*/ |
|
XXH_PUBLIC_API unsigned XXH_versionNumber (void); |
|
|
|
|
|
/* **************************** |
|
* Common basic types |
|
******************************/ |
|
#include <stddef.h> /* size_t */ |
|
typedef enum { XXH_OK=0, XXH_ERROR } XXH_errorcode; |
|
|
|
|
|
/*-********************************************************************** |
|
* 32-bit hash |
|
************************************************************************/ |
|
#if defined(XXH_DOXYGEN) /* Don't show <stdint.h> include */ |
|
/*! |
|
* @brief An unsigned 32-bit integer. |
|
* |
|
* Not necessarily defined to `uint32_t` but functionally equivalent. |
|
*/ |
|
typedef uint32_t XXH32_hash_t; |
|
|
|
#elif !defined (__VMS) \ |
|
&& (defined (__cplusplus) \ |
|
|| (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) ) |
|
# include <stdint.h> |
|
typedef uint32_t XXH32_hash_t; |
|
|
|
#else |
|
# include <limits.h> |
|
# if UINT_MAX == 0xFFFFFFFFUL |
|
typedef unsigned int XXH32_hash_t; |
|
# else |
|
# if ULONG_MAX == 0xFFFFFFFFUL |
|
typedef unsigned long XXH32_hash_t; |
|
# else |
|
# error "unsupported platform: need a 32-bit type" |
|
# endif |
|
# endif |
|
#endif |
|
|
|
/*! |
|
* @} |
|
* |
|
* @defgroup xxh32_family XXH32 family |
|
* @ingroup public |
|
* Contains functions used in the classic 32-bit xxHash algorithm. |
|
* |
|
* @note |
|
* XXH32 is useful for older platforms, with no or poor 64-bit performance. |
|
* Note that @ref xxh3_family provides competitive speed |
|
* for both 32-bit and 64-bit systems, and offers true 64/128 bit hash results. |
|
* |
|
* @see @ref xxh64_family, @ref xxh3_family : Other xxHash families |
|
* @see @ref xxh32_impl for implementation details |
|
* @{ |
|
*/ |
|
|
|
/*! |
|
* @brief Calculates the 32-bit hash of @p input using xxHash32. |
|
* |
|
* Speed on Core 2 Duo @ 3 GHz (single thread, SMHasher benchmark): 5.4 GB/s |
|
* |
|
* @param input The block of data to be hashed, at least @p length bytes in size. |
|
* @param length The length of @p input, in bytes. |
|
* @param seed The 32-bit seed to alter the hash's output predictably. |
|
* |
|
* @pre |
|
* The memory between @p input and @p input + @p length must be valid, |
|
* readable, contiguous memory. However, if @p length is `0`, @p input may be |
|
* `NULL`. In C++, this also must be *TriviallyCopyable*. |
|
* |
|
* @return The calculated 32-bit hash value. |
|
* |
|
* @see |
|
* XXH64(), XXH3_64bits_withSeed(), XXH3_128bits_withSeed(), XXH128(): |
|
* Direct equivalents for the other variants of xxHash. |
|
* @see |
|
* XXH32_createState(), XXH32_update(), XXH32_digest(): Streaming version. |
|
*/ |
|
XXH_PUBLIC_API XXH32_hash_t XXH32 (const void* input, size_t length, XXH32_hash_t seed); |
|
|
|
/*! |
|
* Streaming functions generate the xxHash value from an incremental input. |
|
* This method is slower than single-call functions, due to state management. |
|
* For small inputs, prefer `XXH32()` and `XXH64()`, which are better optimized. |
|
* |
|
* An XXH state must first be allocated using `XXH*_createState()`. |
|
* |
|
* Start a new hash by initializing the state with a seed using `XXH*_reset()`. |
|
* |
|
* Then, feed the hash state by calling `XXH*_update()` as many times as necessary. |
|
* |
|
* The function returns an error code, with 0 meaning OK, and any other value |
|
* meaning there is an error. |
|
* |
|
* Finally, a hash value can be produced anytime, by using `XXH*_digest()`. |
|
* This function returns the nn-bits hash as an int or long long. |
|
* |
|
* It's still possible to continue inserting input into the hash state after a |
|
* digest, and generate new hash values later on by invoking `XXH*_digest()`. |
|
* |
|
* When done, release the state using `XXH*_freeState()`. |
|
* |
|
* Example code for incrementally hashing a file: |
|
* @code{.c} |
|
* #include <stdio.h> |
|
* #include <xxhash.h> |
|
* #define BUFFER_SIZE 256 |
|
* |
|
* // Note: XXH64 and XXH3 use the same interface. |
|
* XXH32_hash_t |
|
* hashFile(FILE* stream) |
|
* { |
|
* XXH32_state_t* state; |
|
* unsigned char buf[BUFFER_SIZE]; |
|
* size_t amt; |
|
* XXH32_hash_t hash; |
|
* |
|
* state = XXH32_createState(); // Create a state |
|
* assert(state != NULL); // Error check here |
|
* XXH32_reset(state, 0xbaad5eed); // Reset state with our seed |
|
* while ((amt = fread(buf, 1, sizeof(buf), stream)) != 0) { |
|
* XXH32_update(state, buf, amt); // Hash the file in chunks |
|
* } |
|
* hash = XXH32_digest(state); // Finalize the hash |
|
* XXH32_freeState(state); // Clean up |
|
* return hash; |
|
* } |
|
* @endcode |
|
*/ |
|
|
|
/*! |
|
* @typedef struct XXH32_state_s XXH32_state_t |
|
* @brief The opaque state struct for the XXH32 streaming API. |
|
* |
|
* @see XXH32_state_s for details. |
|
*/ |
|
typedef struct XXH32_state_s XXH32_state_t; |
|
|
|
/*! |
|
* @brief Allocates an @ref XXH32_state_t. |
|
* |
|
* Must be freed with XXH32_freeState(). |
|
* @return An allocated XXH32_state_t on success, `NULL` on failure. |
|
*/ |
|
XXH_PUBLIC_API XXH32_state_t* XXH32_createState(void); |
|
/*! |
|
* @brief Frees an @ref XXH32_state_t. |
|
* |
|
* Must be allocated with XXH32_createState(). |
|
* @param statePtr A pointer to an @ref XXH32_state_t allocated with @ref XXH32_createState(). |
|
* @return XXH_OK. |
|
*/ |
|
XXH_PUBLIC_API XXH_errorcode XXH32_freeState(XXH32_state_t* statePtr); |
|
/*! |
|
* @brief Copies one @ref XXH32_state_t to another. |
|
* |
|
* @param dst_state The state to copy to. |
|
* @param src_state The state to copy from. |
|
* @pre |
|
* @p dst_state and @p src_state must not be `NULL` and must not overlap. |
|
*/ |
|
XXH_PUBLIC_API void XXH32_copyState(XXH32_state_t* dst_state, const XXH32_state_t* src_state); |
|
|
|
/*! |
|
* @brief Resets an @ref XXH32_state_t to begin a new hash. |
|
* |
|
* This function resets and seeds a state. Call it before @ref XXH32_update(). |
|
* |
|
* @param statePtr The state struct to reset. |
|
* @param seed The 32-bit seed to alter the hash result predictably. |
|
* |
|
* @pre |
|
* @p statePtr must not be `NULL`. |
|
* |
|
* @return @ref XXH_OK on success, @ref XXH_ERROR on failure. |
|
*/ |
|
XXH_PUBLIC_API XXH_errorcode XXH32_reset (XXH32_state_t* statePtr, XXH32_hash_t seed); |
|
|
|
/*! |
|
* @brief Consumes a block of @p input to an @ref XXH32_state_t. |
|
* |
|
* Call this to incrementally consume blocks of data. |
|
* |
|
* @param statePtr The state struct to update. |
|
* @param input The block of data to be hashed, at least @p length bytes in size. |
|
* @param length The length of @p input, in bytes. |
|
* |
|
* @pre |
|
* @p statePtr must not be `NULL`. |
|
* @pre |
|
* The memory between @p input and @p input + @p length must be valid, |
|
* readable, contiguous memory. However, if @p length is `0`, @p input may be |
|
* `NULL`. In C++, this also must be *TriviallyCopyable*. |
|
* |
|
* @return @ref XXH_OK on success, @ref XXH_ERROR on failure. |
|
*/ |
|
XXH_PUBLIC_API XXH_errorcode XXH32_update (XXH32_state_t* statePtr, const void* input, size_t length); |
|
|
|
/*! |
|
* @brief Returns the calculated hash value from an @ref XXH32_state_t. |
|
* |
|
* @note |
|
* Calling XXH32_digest() will not affect @p statePtr, so you can update, |
|
* digest, and update again. |
|
* |
|
* @param statePtr The state struct to calculate the hash from. |
|
* |
|
* @pre |
|
* @p statePtr must not be `NULL`. |
|
* |
|
* @return The calculated xxHash32 value from that state. |
|
*/ |
|
XXH_PUBLIC_API XXH32_hash_t XXH32_digest (const XXH32_state_t* statePtr); |
|
|
|
/******* Canonical representation *******/ |
|
|
|
/* |
|
* The default return values from XXH functions are unsigned 32 and 64 bit |
|
* integers. |
|
* This the simplest and fastest format for further post-processing. |
|
* |
|
* However, this leaves open the question of what is the order on the byte level, |
|
* since little and big endian conventions will store the same number differently. |
|
* |
|
* The canonical representation settles this issue by mandating big-endian |
|
* convention, the same convention as human-readable numbers (large digits first). |
|
* |
|
* When writing hash values to storage, sending them over a network, or printing |
|
* them, it's highly recommended to use the canonical representation to ensure |
|
* portability across a wider range of systems, present and future. |
|
* |
|
* The following functions allow transformation of hash values to and from |
|
* canonical format. |
|
*/ |
|
|
|
/*! |
|
* @brief Canonical (big endian) representation of @ref XXH32_hash_t. |
|
*/ |
|
typedef struct { |
|
unsigned char digest[4]; /*!< Hash bytes, big endian */ |
|
} XXH32_canonical_t; |
|
|
|
/*! |
|
* @brief Converts an @ref XXH32_hash_t to a big endian @ref XXH32_canonical_t. |
|
* |
|
* @param dst The @ref XXH32_canonical_t pointer to be stored to. |
|
* @param hash The @ref XXH32_hash_t to be converted. |
|
* |
|
* @pre |
|
* @p dst must not be `NULL`. |
|
*/ |
|
XXH_PUBLIC_API void XXH32_canonicalFromHash(XXH32_canonical_t* dst, XXH32_hash_t hash); |
|
|
|
/*! |
|
* @brief Converts an @ref XXH32_canonical_t to a native @ref XXH32_hash_t. |
|
* |
|
* @param src The @ref XXH32_canonical_t to convert. |
|
* |
|
* @pre |
|
* @p src must not be `NULL`. |
|
* |
|
* @return The converted hash. |
|
*/ |
|
XXH_PUBLIC_API XXH32_hash_t XXH32_hashFromCanonical(const XXH32_canonical_t* src); |
|
|
|
|
|
#ifdef __has_attribute |
|
# define XXH_HAS_ATTRIBUTE(x) __has_attribute(x) |
|
#else |
|
# define XXH_HAS_ATTRIBUTE(x) 0 |
|
#endif |
|
|
|
/* C-language Attributes are added in C23. */ |
|
#if defined(__STDC_VERSION__) && (__STDC_VERSION__ > 201710L) && defined(__has_c_attribute) |
|
# define XXH_HAS_C_ATTRIBUTE(x) __has_c_attribute(x) |
|
#else |
|
# define XXH_HAS_C_ATTRIBUTE(x) 0 |
|
#endif |
|
|
|
#if defined(__cplusplus) && defined(__has_cpp_attribute) |
|
# define XXH_HAS_CPP_ATTRIBUTE(x) __has_cpp_attribute(x) |
|
#else |
|
# define XXH_HAS_CPP_ATTRIBUTE(x) 0 |
|
#endif |
|
|
|
/* |
|
Define XXH_FALLTHROUGH macro for annotating switch case with the 'fallthrough' attribute |
|
introduced in CPP17 and C23. |
|
CPP17 : https://en.cppreference.com/w/cpp/language/attributes/fallthrough |
|
C23 : https://en.cppreference.com/w/c/language/attributes/fallthrough |
|
*/ |
|
#if XXH_HAS_C_ATTRIBUTE(x) |
|
# define XXH_FALLTHROUGH [[fallthrough]] |
|
#elif XXH_HAS_CPP_ATTRIBUTE(x) |
|
# define XXH_FALLTHROUGH [[fallthrough]] |
|
#elif XXH_HAS_ATTRIBUTE(__fallthrough__) |
|
# define XXH_FALLTHROUGH __attribute__ ((fallthrough)) |
|
#else |
|
# define XXH_FALLTHROUGH |
|
#endif |
|
|
|
/*! |
|
* @} |
|
* @ingroup public |
|
* @{ |
|
*/ |
|
|
|
#ifndef XXH_NO_LONG_LONG |
|
/*-********************************************************************** |
|
* 64-bit hash |
|
************************************************************************/ |
|
#if defined(XXH_DOXYGEN) /* don't include <stdint.h> */ |
|
/*! |
|
* @brief An unsigned 64-bit integer. |
|
* |
|
* Not necessarily defined to `uint64_t` but functionally equivalent. |
|
*/ |
|
typedef uint64_t XXH64_hash_t; |
|
#elif !defined (__VMS) \ |
|
&& (defined (__cplusplus) \ |
|
|| (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) ) |
|
# include <stdint.h> |
|
typedef uint64_t XXH64_hash_t; |
|
#else |
|
# include <limits.h> |
|
# if defined(__LP64__) && ULONG_MAX == 0xFFFFFFFFFFFFFFFFULL |
|
/* LP64 ABI says uint64_t is unsigned long */ |
|
typedef unsigned long XXH64_hash_t; |
|
# else |
|
/* the following type must have a width of 64-bit */ |
|
typedef unsigned long long XXH64_hash_t; |
|
# endif |
|
#endif |
|
|
|
/*! |
|
* @} |
|
* |
|
* @defgroup xxh64_family XXH64 family |
|
* @ingroup public |
|
* @{ |
|
* Contains functions used in the classic 64-bit xxHash algorithm. |
|
* |
|
* @note |
|
* XXH3 provides competitive speed for both 32-bit and 64-bit systems, |
|
* and offers true 64/128 bit hash results. |
|
* It provides better speed for systems with vector processing capabilities. |
|
*/ |
|
|
|
|
|
/*! |
|
* @brief Calculates the 64-bit hash of @p input using xxHash64. |
|
* |
|
* This function usually runs faster on 64-bit systems, but slower on 32-bit |
|
* systems (see benchmark). |
|
* |
|
* @param input The block of data to be hashed, at least @p length bytes in size. |
|
* @param length The length of @p input, in bytes. |
|
* @param seed The 64-bit seed to alter the hash's output predictably. |
|
* |
|
* @pre |
|
* The memory between @p input and @p input + @p length must be valid, |
|
* readable, contiguous memory. However, if @p length is `0`, @p input may be |
|
* `NULL`. In C++, this also must be *TriviallyCopyable*. |
|
* |
|
* @return The calculated 64-bit hash. |
|
* |
|
* @see |
|
* XXH32(), XXH3_64bits_withSeed(), XXH3_128bits_withSeed(), XXH128(): |
|
* Direct equivalents for the other variants of xxHash. |
|
* @see |
|
* XXH64_createState(), XXH64_update(), XXH64_digest(): Streaming version. |
|
*/ |
|
XXH_PUBLIC_API XXH64_hash_t XXH64(const void* input, size_t length, XXH64_hash_t seed); |
|
|
|
/******* Streaming *******/ |
|
/*! |
|
* @brief The opaque state struct for the XXH64 streaming API. |
|
* |
|
* @see XXH64_state_s for details. |
|
*/ |
|
typedef struct XXH64_state_s XXH64_state_t; /* incomplete type */ |
|
XXH_PUBLIC_API XXH64_state_t* XXH64_createState(void); |
|
XXH_PUBLIC_API XXH_errorcode XXH64_freeState(XXH64_state_t* statePtr); |
|
XXH_PUBLIC_API void XXH64_copyState(XXH64_state_t* dst_state, const XXH64_state_t* src_state); |
|
|
|
XXH_PUBLIC_API XXH_errorcode XXH64_reset (XXH64_state_t* statePtr, XXH64_hash_t seed); |
|
XXH_PUBLIC_API XXH_errorcode XXH64_update (XXH64_state_t* statePtr, const void* input, size_t length); |
|
XXH_PUBLIC_API XXH64_hash_t XXH64_digest (const XXH64_state_t* statePtr); |
|
|
|
/******* Canonical representation *******/ |
|
typedef struct { unsigned char digest[sizeof(XXH64_hash_t)]; } XXH64_canonical_t; |
|
XXH_PUBLIC_API void XXH64_canonicalFromHash(XXH64_canonical_t* dst, XXH64_hash_t hash); |
|
XXH_PUBLIC_API XXH64_hash_t XXH64_hashFromCanonical(const XXH64_canonical_t* src); |
|
|
|
/*! |
|
* @} |
|
* ************************************************************************ |
|
* @defgroup xxh3_family XXH3 family |
|
* @ingroup public |
|
* @{ |
|
* |
|
* XXH3 is a more recent hash algorithm featuring: |
|
* - Improved speed for both small and large inputs |
|
* - True 64-bit and 128-bit outputs |
|
* - SIMD acceleration |
|
* - Improved 32-bit viability |
|
* |
|
* Speed analysis methodology is explained here: |
|
* |
|
* https://fastcompression.blogspot.com/2019/03/presenting-xxh3.html |
|
* |
|
* Compared to XXH64, expect XXH3 to run approximately |
|
* ~2x faster on large inputs and >3x faster on small ones, |
|
* exact differences vary depending on platform. |
|
* |
|
* XXH3's speed benefits greatly from SIMD and 64-bit arithmetic, |
|
* but does not require it. |
|
* Any 32-bit and 64-bit targets that can run XXH32 smoothly |
|
* can run XXH3 at competitive speeds, even without vector support. |
|
* Further details are explained in the implementation. |
|
* |
|
* Optimized implementations are provided for AVX512, AVX2, SSE2, NEON, POWER8, |
|
* ZVector and scalar targets. This can be controlled via the XXH_VECTOR macro. |
|
* |
|
* XXH3 implementation is portable: |
|
* it has a generic C90 formulation that can be compiled on any platform, |
|
* all implementations generage exactly the same hash value on all platforms. |
|
* Starting from v0.8.0, it's also labelled "stable", meaning that |
|
* any future version will also generate the same hash value. |
|
* |
|
* XXH3 offers 2 variants, _64bits and _128bits. |
|
* |
|
* When only 64 bits are needed, prefer invoking the _64bits variant, as it |
|
* reduces the amount of mixing, resulting in faster speed on small inputs. |
|
* It's also generally simpler to manipulate a scalar return type than a struct. |
|
* |
|
* The API supports one-shot hashing, streaming mode, and custom secrets. |
|
*/ |
|
|
|
/*-********************************************************************** |
|
* XXH3 64-bit variant |
|
************************************************************************/ |
|
|
|
/* XXH3_64bits(): |
|
* default 64-bit variant, using default secret and default seed of 0. |
|
* It's the fastest variant. */ |
|
XXH_PUBLIC_API XXH64_hash_t XXH3_64bits(const void* data, size_t len); |
|
|
|
/* |
|
* XXH3_64bits_withSeed(): |
|
* This variant generates a custom secret on the fly |
|
* based on default secret altered using the `seed` value. |
|
* While this operation is decently fast, note that it's not completely free. |
|
* Note: seed==0 produces the same results as XXH3_64bits(). |
|
*/ |
|
XXH_PUBLIC_API XXH64_hash_t XXH3_64bits_withSeed(const void* data, size_t len, XXH64_hash_t seed); |
|
|
|
/*! |
|
* The bare minimum size for a custom secret. |
|
* |
|
* @see |
|
* XXH3_64bits_withSecret(), XXH3_64bits_reset_withSecret(), |
|
* XXH3_128bits_withSecret(), XXH3_128bits_reset_withSecret(). |
|
*/ |
|
#define XXH3_SECRET_SIZE_MIN 136 |
|
|
|
/* |
|
* XXH3_64bits_withSecret(): |
|
* It's possible to provide any blob of bytes as a "secret" to generate the hash. |
|
* This makes it more difficult for an external actor to prepare an intentional collision. |
|
* The main condition is that secretSize *must* be large enough (>= XXH3_SECRET_SIZE_MIN). |
|
* However, the quality of the secret impacts the dispersion of the hash algorithm. |
|
* Therefore, the secret _must_ look like a bunch of random bytes. |
|
* Avoid "trivial" or structured data such as repeated sequences or a text document. |
|
* Whenever in doubt about the "randomness" of the blob of bytes, |
|
* consider employing "XXH3_generateSecret()" instead (see below). |
|
* It will generate a proper high entropy secret derived from the blob of bytes. |
|
* Another advantage of using XXH3_generateSecret() is that |
|
* it guarantees that all bits within the initial blob of bytes |
|
* will impact every bit of the output. |
|
* This is not necessarily the case when using the blob of bytes directly |
|
* because, when hashing _small_ inputs, only a portion of the secret is employed. |
|
*/ |
|
XXH_PUBLIC_API XXH64_hash_t XXH3_64bits_withSecret(const void* data, size_t len, const void* secret, size_t secretSize); |
|
|
|
|
|
/******* Streaming *******/ |
|
/* |
|
* Streaming requires state maintenance. |
|
* This operation costs memory and CPU. |
|
* As a consequence, streaming is slower than one-shot hashing. |
|
* For better performance, prefer one-shot functions whenever applicable. |
|
*/ |
|
|
|
/*! |
|
* @brief The state struct for the XXH3 streaming API. |
|
* |
|
* @see XXH3_state_s for details. |
|
*/ |
|
typedef struct XXH3_state_s XXH3_state_t; |
|
XXH_PUBLIC_API XXH3_state_t* XXH3_createState(void); |
|
XXH_PUBLIC_API XXH_errorcode XXH3_freeState(XXH3_state_t* statePtr); |
|
XXH_PUBLIC_API void XXH3_copyState(XXH3_state_t* dst_state, const XXH3_state_t* src_state); |
|
|
|
/* |
|
* XXH3_64bits_reset(): |
|
* Initialize with default parameters. |
|
* digest will be equivalent to `XXH3_64bits()`. |
|
*/ |
|
XXH_PUBLIC_API XXH_errorcode XXH3_64bits_reset(XXH3_state_t* statePtr); |
|
/* |
|
* XXH3_64bits_reset_withSeed(): |
|
* Generate a custom secret from `seed`, and store it into `statePtr`. |
|
* digest will be equivalent to `XXH3_64bits_withSeed()`. |
|
*/ |
|
XXH_PUBLIC_API XXH_errorcode XXH3_64bits_reset_withSeed(XXH3_state_t* statePtr, XXH64_hash_t seed); |
|
/* |
|
* XXH3_64bits_reset_withSecret(): |
|
* `secret` is referenced, it _must outlive_ the hash streaming session. |
|
* Similar to one-shot API, `secretSize` must be >= `XXH3_SECRET_SIZE_MIN`, |
|
* and the quality of produced hash values depends on secret's entropy |
|
* (secret's content should look like a bunch of random bytes). |
|
* When in doubt about the randomness of a candidate `secret`, |
|
* consider employing `XXH3_generateSecret()` instead (see below). |
|
*/ |
|
XXH_PUBLIC_API XXH_errorcode XXH3_64bits_reset_withSecret(XXH3_state_t* statePtr, const void* secret, size_t secretSize); |
|
|
|
XXH_PUBLIC_API XXH_errorcode XXH3_64bits_update (XXH3_state_t* statePtr, const void* input, size_t length); |
|
XXH_PUBLIC_API XXH64_hash_t XXH3_64bits_digest (const XXH3_state_t* statePtr); |
|
|
|
/* note : canonical representation of XXH3 is the same as XXH64 |
|
* since they both produce XXH64_hash_t values */ |
|
|
|
|
|
/*-********************************************************************** |
|
* XXH3 128-bit variant |
|
************************************************************************/ |
|
|
|
/*! |
|
* @brief The return value from 128-bit hashes. |
|
* |
|
* Stored in little endian order, although the fields themselves are in native |
|
* endianness. |
|
*/ |
|
typedef struct { |
|
XXH64_hash_t low64; /*!< `value & 0xFFFFFFFFFFFFFFFF` */ |
|
XXH64_hash_t high64; /*!< `value >> 64` */ |
|
} XXH128_hash_t; |
|
|
|
XXH_PUBLIC_API XXH128_hash_t XXH3_128bits(const void* data, size_t len); |
|
XXH_PUBLIC_API XXH128_hash_t XXH3_128bits_withSeed(const void* data, size_t len, XXH64_hash_t seed); |
|
XXH_PUBLIC_API XXH128_hash_t XXH3_128bits_withSecret(const void* data, size_t len, const void* secret, size_t secretSize); |
|
|
|
/******* Streaming *******/ |
|
/* |
|
* Streaming requires state maintenance. |
|
* This operation costs memory and CPU. |
|
* As a consequence, streaming is slower than one-shot hashing. |
|
* For better performance, prefer one-shot functions whenever applicable. |
|
* |
|
* XXH3_128bits uses the same XXH3_state_t as XXH3_64bits(). |
|
* Use already declared XXH3_createState() and XXH3_freeState(). |
|
* |
|
* All reset and streaming functions have same meaning as their 64-bit counterpart. |
|
*/ |
|
|
|
XXH_PUBLIC_API XXH_errorcode XXH3_128bits_reset(XXH3_state_t* statePtr); |
|
XXH_PUBLIC_API XXH_errorcode XXH3_128bits_reset_withSeed(XXH3_state_t* statePtr, XXH64_hash_t seed); |
|
XXH_PUBLIC_API XXH_errorcode XXH3_128bits_reset_withSecret(XXH3_state_t* statePtr, const void* secret, size_t secretSize); |
|
|
|
XXH_PUBLIC_API XXH_errorcode XXH3_128bits_update (XXH3_state_t* statePtr, const void* input, size_t length); |
|
XXH_PUBLIC_API XXH128_hash_t XXH3_128bits_digest (const XXH3_state_t* statePtr); |
|
|
|
/* Following helper functions make it possible to compare XXH128_hast_t values. |
|
* Since XXH128_hash_t is a structure, this capability is not offered by the language. |
|
* Note: For better performance, these functions can be inlined using XXH_INLINE_ALL */ |
|
|
|
/*! |
|
* XXH128_isEqual(): |
|
* Return: 1 if `h1` and `h2` are equal, 0 if they are not. |
|
*/ |
|
XXH_PUBLIC_API int XXH128_isEqual(XXH128_hash_t h1, XXH128_hash_t h2); |
|
|
|
/*! |
|
* XXH128_cmp(): |
|
* |
|
* This comparator is compatible with stdlib's `qsort()`/`bsearch()`. |
|
* |
|
* return: >0 if *h128_1 > *h128_2 |
|
* =0 if *h128_1 == *h128_2 |
|
* <0 if *h128_1 < *h128_2 |
|
*/ |
|
XXH_PUBLIC_API int XXH128_cmp(const void* h128_1, const void* h128_2); |
|
|
|
|
|
/******* Canonical representation *******/ |
|
typedef struct { unsigned char digest[sizeof(XXH128_hash_t)]; } XXH128_canonical_t; |
|
XXH_PUBLIC_API void XXH128_canonicalFromHash(XXH128_canonical_t* dst, XXH128_hash_t hash); |
|
XXH_PUBLIC_API XXH128_hash_t XXH128_hashFromCanonical(const XXH128_canonical_t* src); |
|
|
|
|
|
#endif /* XXH_NO_LONG_LONG */ |
|
|
|
/*! |
|
* @} |
|
*/ |
|
#endif /* XXHASH_H_5627135585666179 */ |
|
|
|
|
|
|
|
#if defined(XXH_STATIC_LINKING_ONLY) && !defined(XXHASH_H_STATIC_13879238742) |
|
#define XXHASH_H_STATIC_13879238742 |
|
/* **************************************************************************** |
|
* This section contains declarations which are not guaranteed to remain stable. |
|
* They may change in future versions, becoming incompatible with a different |
|
* version of the library. |
|
* These declarations should only be used with static linking. |
|
* Never use them in association with dynamic linking! |
|
***************************************************************************** */ |
|
|
|
/* |
|
* These definitions are only present to allow static allocation |
|
* of XXH states, on stack or in a struct, for example. |
|
* Never **ever** access their members directly. |
|
*/ |
|
|
|
/*! |
|
* @internal |
|
* @brief Structure for XXH32 streaming API. |
|
* |
|
* @note This is only defined when @ref XXH_STATIC_LINKING_ONLY, |
|
* @ref XXH_INLINE_ALL, or @ref XXH_IMPLEMENTATION is defined. Otherwise it is |
|
* an opaque type. This allows fields to safely be changed. |
|
* |
|
* Typedef'd to @ref XXH32_state_t. |
|
* Do not access the members of this struct directly. |
|
* @see XXH64_state_s, XXH3_state_s |
|
*/ |
|
struct XXH32_state_s { |
|
XXH32_hash_t total_len_32; /*!< Total length hashed, modulo 2^32 */ |
|
XXH32_hash_t large_len; /*!< Whether the hash is >= 16 (handles @ref total_len_32 overflow) */ |
|
XXH32_hash_t v[4]; /*!< Accumulator lanes */ |
|
XXH32_hash_t mem32[4]; /*!< Internal buffer for partial reads. Treated as unsigned char[16]. */ |
|
XXH32_hash_t memsize; /*!< Amount of data in @ref mem32 */ |
|
XXH32_hash_t reserved; /*!< Reserved field. Do not read or write to it, it may be removed. */ |
|
}; /* typedef'd to XXH32_state_t */ |
|
|
|
|
|
#ifndef XXH_NO_LONG_LONG /* defined when there is no 64-bit support */ |
|
|
|
/*! |
|
* @internal |
|
* @brief Structure for XXH64 streaming API. |
|
* |
|
* @note This is only defined when @ref XXH_STATIC_LINKING_ONLY, |
|
* @ref XXH_INLINE_ALL, or @ref XXH_IMPLEMENTATION is defined. Otherwise it is |
|
* an opaque type. This allows fields to safely be changed. |
|
* |
|
* Typedef'd to @ref XXH64_state_t. |
|
* Do not access the members of this struct directly. |
|
* @see XXH32_state_s, XXH3_state_s |
|
*/ |
|
struct XXH64_state_s { |
|
XXH64_hash_t total_len; /*!< Total length hashed. This is always 64-bit. */ |
|
XXH64_hash_t v[4]; /*!< Accumulator lanes */ |
|
XXH64_hash_t mem64[4]; /*!< Internal buffer for partial reads. Treated as unsigned char[32]. */ |
|
XXH32_hash_t memsize; /*!< Amount of data in @ref mem64 */ |
|
XXH32_hash_t reserved32; /*!< Reserved field, needed for padding anyways*/ |
|
XXH64_hash_t reserved64; /*!< Reserved field. Do not read or write to it, it may be removed. */ |
|
}; /* typedef'd to XXH64_state_t */ |
|
|
|
#if defined(__STDC_VERSION__) && (__STDC_VERSION__ >= 201112L) /* >= C11 */ |
|
# include <stdalign.h> |
|
# define XXH_ALIGN(n) alignas(n) |
|
#elif defined(__cplusplus) && (__cplusplus >= 201103L) /* >= C++11 */ |
|
/* In C++ alignas() is a keyword */ |
|
# define XXH_ALIGN(n) alignas(n) |
|
#elif defined(__GNUC__) |
|
# define XXH_ALIGN(n) __attribute__ ((aligned(n))) |
|
#elif defined(_MSC_VER) |
|
# define XXH_ALIGN(n) __declspec(align(n)) |
|
#else |
|
# define XXH_ALIGN(n) /* disabled */ |
|
#endif |
|
|
|
/* Old GCC versions only accept the attribute after the type in structures. */ |
|
#if !(defined(__STDC_VERSION__) && (__STDC_VERSION__ >= 201112L)) /* C11+ */ \ |
|
&& ! (defined(__cplusplus) && (__cplusplus >= 201103L)) /* >= C++11 */ \ |
|
&& defined(__GNUC__) |
|
# define XXH_ALIGN_MEMBER(align, type) type XXH_ALIGN(align) |
|
#else |
|
# define XXH_ALIGN_MEMBER(align, type) XXH_ALIGN(align) type |
|
#endif |
|
|
|
/*! |
|
* @brief The size of the internal XXH3 buffer. |
|
* |
|
* This is the optimal update size for incremental hashing. |
|
* |
|
* @see XXH3_64b_update(), XXH3_128b_update(). |
|
*/ |
|
#define XXH3_INTERNALBUFFER_SIZE 256 |
|
|
|
/*! |
|
* @brief Default size of the secret buffer (and @ref XXH3_kSecret). |
|
* |
|
* This is the size used in @ref XXH3_kSecret and the seeded functions. |
|
* |
|
* Not to be confused with @ref XXH3_SECRET_SIZE_MIN. |
|
*/ |
|
#define XXH3_SECRET_DEFAULT_SIZE 192 |
|
|
|
/*! |
|
* @internal |
|
* @brief Structure for XXH3 streaming API. |
|
* |
|
* @note This is only defined when @ref XXH_STATIC_LINKING_ONLY, |
|
* @ref XXH_INLINE_ALL, or @ref XXH_IMPLEMENTATION is defined. |
|
* Otherwise it is an opaque type. |
|
* Never use this definition in combination with dynamic library. |
|
* This allows fields to safely be changed in the future. |
|
* |
|
* @note ** This structure has a strict alignment requirement of 64 bytes!! ** |
|
* Do not allocate this with `malloc()` or `new`, |
|
* it will not be sufficiently aligned. |
|
* Use @ref XXH3_createState() and @ref XXH3_freeState(), or stack allocation. |
|
* |
|
* Typedef'd to @ref XXH3_state_t. |
|
* Do never access the members of this struct directly. |
|
* |
|
* @see XXH3_INITSTATE() for stack initialization. |
|
* @see XXH3_createState(), XXH3_freeState(). |
|
* @see XXH32_state_s, XXH64_state_s |
|
*/ |
|
struct XXH3_state_s { |
|
XXH_ALIGN_MEMBER(64, XXH64_hash_t acc[8]); |
|
/*!< The 8 accumulators. Similar to `vN` in @ref XXH32_state_s::v1 and @ref XXH64_state_s */ |
|
XXH_ALIGN_MEMBER(64, unsigned char customSecret[XXH3_SECRET_DEFAULT_SIZE]); |
|
/*!< Used to store a custom secret generated from a seed. */ |
|
XXH_ALIGN_MEMBER(64, unsigned char buffer[XXH3_INTERNALBUFFER_SIZE]); |
|
/*!< The internal buffer. @see XXH32_state_s::mem32 */ |
|
XXH32_hash_t bufferedSize; |
|
/*!< The amount of memory in @ref buffer, @see XXH32_state_s::memsize */ |
|
XXH32_hash_t useSeed; |
|
/*!< Reserved field. Needed for padding on 64-bit. */ |
|
size_t nbStripesSoFar; |
|
/*!< Number or stripes processed. */ |
|
XXH64_hash_t totalLen; |
|
/*!< Total length hashed. 64-bit even on 32-bit targets. */ |
|
size_t nbStripesPerBlock; |
|
/*!< Number of stripes per block. */ |
|
size_t secretLimit; |
|
/*!< Size of @ref customSecret or @ref extSecret */ |
|
XXH64_hash_t seed; |
|
/*!< Seed for _withSeed variants. Must be zero otherwise, @see XXH3_INITSTATE() */ |
|
XXH64_hash_t reserved64; |
|
/*!< Reserved field. */ |
|
const unsigned char* extSecret; |
|
/*!< Reference to an external secret for the _withSecret variants, NULL |
|
* for other variants. */ |
|
/* note: there may be some padding at the end due to alignment on 64 bytes */ |
|
}; /* typedef'd to XXH3_state_t */ |
|
|
|
#undef XXH_ALIGN_MEMBER |
|
|
|
/*! |
|
* @brief Initializes a stack-allocated `XXH3_state_s`. |
|
* |
|
* When the @ref XXH3_state_t structure is merely emplaced on stack, |
|
* it should be initialized with XXH3_INITSTATE() or a memset() |
|
* in case its first reset uses XXH3_NNbits_reset_withSeed(). |
|
* This init can be omitted if the first reset uses default or _withSecret mode. |
|
* This operation isn't necessary when the state is created with XXH3_createState(). |
|
* Note that this doesn't prepare the state for a streaming operation, |
|
* it's still necessary to use XXH3_NNbits_reset*() afterwards. |
|
*/ |
|
#define XXH3_INITSTATE(XXH3_state_ptr) { (XXH3_state_ptr)->seed = 0; } |
|
|
|
|
|
/* XXH128() : |
|
* simple alias to pre-selected XXH3_128bits variant |
|
*/ |
|
XXH_PUBLIC_API XXH128_hash_t XXH128(const void* data, size_t len, XXH64_hash_t seed); |
|
|
|
|
|
/* === Experimental API === */ |
|
/* Symbols defined below must be considered tied to a specific library version. */ |
|
|
|
/* |
|
* XXH3_generateSecret(): |
|
* |
|
* Derive a high-entropy secret from any user-defined content, named customSeed. |
|
* The generated secret can be used in combination with `*_withSecret()` functions. |
|
* The `_withSecret()` variants are useful to provide a higher level of protection than 64-bit seed, |
|
* as it becomes much more difficult for an external actor to guess how to impact the calculation logic. |
|
* |
|
* The function accepts as input a custom seed of any length and any content, |
|
* and derives from it a high-entropy secret of length @secretSize |
|
* into an already allocated buffer @secretBuffer. |
|
* @secretSize must be >= XXH3_SECRET_SIZE_MIN |
|
* |
|
* The generated secret can then be used with any `*_withSecret()` variant. |
|
* Functions `XXH3_128bits_withSecret()`, `XXH3_64bits_withSecret()`, |
|
* `XXH3_128bits_reset_withSecret()` and `XXH3_64bits_reset_withSecret()` |
|
* are part of this list. They all accept a `secret` parameter |
|
* which must be large enough for implementation reasons (>= XXH3_SECRET_SIZE_MIN) |
|
* _and_ feature very high entropy (consist of random-looking bytes). |
|
* These conditions can be a high bar to meet, so |
|
* XXH3_generateSecret() can be employed to ensure proper quality. |
|
* |
|
* customSeed can be anything. It can have any size, even small ones, |
|
* and its content can be anything, even "poor entropy" sources such as a bunch of zeroes. |
|
* The resulting `secret` will nonetheless provide all required qualities. |
|
* |
|
* When customSeedSize > 0, supplying NULL as customSeed is undefined behavior. |
|
*/ |
|
XXH_PUBLIC_API XXH_errorcode XXH3_generateSecret(void* secretBuffer, size_t secretSize, const void* customSeed, size_t customSeedSize); |
|
|
|
|
|
/* |
|
* XXH3_generateSecret_fromSeed(): |
|
* |
|
* Generate the same secret as the _withSeed() variants. |
|
* |
|
* The resulting secret has a length of XXH3_SECRET_DEFAULT_SIZE (necessarily). |
|
* @secretBuffer must be already allocated, of size at least XXH3_SECRET_DEFAULT_SIZE bytes. |
|
* |
|
* The generated secret can be used in combination with |
|
*`*_withSecret()` and `_withSecretandSeed()` variants. |
|
* This generator is notably useful in combination with `_withSecretandSeed()`, |
|
* as a way to emulate a faster `_withSeed()` variant. |
|
*/ |
|
XXH_PUBLIC_API void XXH3_generateSecret_fromSeed(void* secretBuffer, XXH64_hash_t seed); |
|
|
|
/* |
|
* *_withSecretandSeed() : |
|
* These variants generate hash values using either |
|
* @seed for "short" keys (< XXH3_MIDSIZE_MAX = 240 bytes) |
|
* or @secret for "large" keys (>= XXH3_MIDSIZE_MAX). |
|
* |
|
* This generally benefits speed, compared to `_withSeed()` or `_withSecret()`. |
|
* `_withSeed()` has to generate the secret on the fly for "large" keys. |
|
* It's fast, but can be perceptible for "not so large" keys (< 1 KB). |
|
* `_withSecret()` has to generate the masks on the fly for "small" keys, |
|
* which requires more instructions than _withSeed() variants. |
|
* Therefore, _withSecretandSeed variant combines the best of both worlds. |
|
* |
|
* When @secret has been generated by XXH3_generateSecret_fromSeed(), |
|
* this variant produces *exactly* the same results as `_withSeed()` variant, |
|
* hence offering only a pure speed benefit on "large" input, |
|
* by skipping the need to regenerate the secret for every large input. |
|
* |
|
* Another usage scenario is to hash the secret to a 64-bit hash value, |
|
* for example with XXH3_64bits(), which then becomes the seed, |
|
* and then employ both the seed and the secret in _withSecretandSeed(). |
|
* On top of speed, an added benefit is that each bit in the secret |
|
* has a 50% chance to swap each bit in the output, |
|
* via its impact to the seed. |
|
* This is not guaranteed when using the secret directly in "small data" scenarios, |
|
* because only portions of the secret are employed for small data. |
|
*/ |
|
XXH_PUBLIC_API XXH64_hash_t |
|
XXH3_64bits_withSecretandSeed(const void* data, size_t len, |
|
const void* secret, size_t secretSize, |
|
XXH64_hash_t seed); |
|
|
|
XXH_PUBLIC_API XXH128_hash_t |
|
XXH3_128bits_withSecretandSeed(const void* data, size_t len, |
|
const void* secret, size_t secretSize, |
|
XXH64_hash_t seed64); |
|
|
|
XXH_PUBLIC_API XXH_errorcode |
|
XXH3_64bits_reset_withSecretandSeed(XXH3_state_t* statePtr, |
|
const void* secret, size_t secretSize, |
|
XXH64_hash_t seed64); |
|
|
|
XXH_PUBLIC_API XXH_errorcode |
|
XXH3_128bits_reset_withSecretandSeed(XXH3_state_t* statePtr, |
|
const void* secret, size_t secretSize, |
|
XXH64_hash_t seed64); |
|
|
|
|
|
#endif /* XXH_NO_LONG_LONG */ |
|
#if defined(XXH_INLINE_ALL) || defined(XXH_PRIVATE_API) |
|
# define XXH_IMPLEMENTATION |
|
#endif |
|
|
|
#endif /* defined(XXH_STATIC_LINKING_ONLY) && !defined(XXHASH_H_STATIC_13879238742) */ |
|
|
|
|
|
/* ======================================================================== */ |
|
/* ======================================================================== */ |
|
/* ======================================================================== */ |
|
|
|
|
|
/*-********************************************************************** |
|
* xxHash implementation |
|
*-********************************************************************** |
|
* xxHash's implementation used to be hosted inside xxhash.c. |
|
* |
|
* However, inlining requires implementation to be visible to the compiler, |
|
* hence be included alongside the header. |
|
* Previously, implementation was hosted inside xxhash.c, |
|
* which was then #included when inlining was activated. |
|
* This construction created issues with a few build and install systems, |
|
* as it required xxhash.c to be stored in /include directory. |
|
* |
|
* xxHash implementation is now directly integrated within xxhash.h. |
|
* As a consequence, xxhash.c is no longer needed in /include. |
|
* |
|
* xxhash.c is still available and is still useful. |
|
* In a "normal" setup, when xxhash is not inlined, |
|
* xxhash.h only exposes the prototypes and public symbols, |
|
* while xxhash.c can be built into an object file xxhash.o |
|
* which can then be linked into the final binary. |
|
************************************************************************/ |
|
|
|
#if ( defined(XXH_INLINE_ALL) || defined(XXH_PRIVATE_API) \ |
|
|| defined(XXH_IMPLEMENTATION) ) && !defined(XXH_IMPLEM_13a8737387) |
|
# define XXH_IMPLEM_13a8737387 |
|
|
|
/* ************************************* |
|
* Tuning parameters |
|
***************************************/ |
|
|
|
/*! |
|
* @defgroup tuning Tuning parameters |
|
* @{ |
|
* |
|
* Various macros to control xxHash's behavior. |
|
*/ |
|
#ifdef XXH_DOXYGEN |
|
/*! |
|
* @brief Define this to disable 64-bit code. |
|
* |
|
* Useful if only using the @ref xxh32_family and you have a strict C90 compiler. |
|
*/ |
|
# define XXH_NO_LONG_LONG |
|
# undef XXH_NO_LONG_LONG /* don't actually */ |
|
/*! |
|
* @brief Controls how unaligned memory is accessed. |
|
* |
|
* By default, access to unaligned memory is controlled by `memcpy()`, which is |
|
* safe and portable. |
|
* |
|
* Unfortunately, on some target/compiler combinations, the generated assembly |
|
* is sub-optimal. |
|
* |
|
* The below switch allow selection of a different access method |
|
* in the search for improved performance. |
|
* |
|
* @par Possible options: |
|
* |
|
* - `XXH_FORCE_MEMORY_ACCESS=0` (default): `memcpy` |
|
* @par |
|
* Use `memcpy()`. Safe and portable. Note that most modern compilers will |
|
* eliminate the function call and treat it as an unaligned access. |
|
* |
|
* - `XXH_FORCE_MEMORY_ACCESS=1`: `__attribute__((packed))` |
|
* @par |
|
* Depends on compiler extensions and is therefore not portable. |
|
* This method is safe _if_ your compiler supports it, |
|
* and *generally* as fast or faster than `memcpy`. |
|
* |
|
* - `XXH_FORCE_MEMORY_ACCESS=2`: Direct cast |
|
* @par |
|
* Casts directly and dereferences. This method doesn't depend on the |
|
* compiler, but it violates the C standard as it directly dereferences an |
|
* unaligned pointer. It can generate buggy code on targets which do not |
|
* support unaligned memory accesses, but in some circumstances, it's the |
|
* only known way to get the most performance. |
|
* |
|
* - `XXH_FORCE_MEMORY_ACCESS=3`: Byteshift |
|
* @par |
|
* Also portable. This can generate the best code on old compilers which don't |
|
* inline small `memcpy()` calls, and it might also be faster on big-endian |
|
* systems which lack a native byteswap instruction. However, some compilers |
|
* will emit literal byteshifts even if the target supports unaligned access. |
|
* . |
|
* |
|
* @warning |
|
* Methods 1 and 2 rely on implementation-defined behavior. Use these with |
|
* care, as what works on one compiler/platform/optimization level may cause |
|
* another to read garbage data or even crash. |
|
* |
|
* See http://fastcompression.blogspot.com/2015/08/accessing-unaligned-memory.html for details. |
|
* |
|
* Prefer these methods in priority order (0 > 3 > 1 > 2) |
|
*/ |
|
# define XXH_FORCE_MEMORY_ACCESS 0 |
|
|
|
/*! |
|
* @def XXH_FORCE_ALIGN_CHECK |
|
* @brief If defined to non-zero, adds a special path for aligned inputs (XXH32() |
|
* and XXH64() only). |
|
* |
|
* This is an important performance trick for architectures without decent |
|
* unaligned memory access performance. |
|
* |
|
* It checks for input alignment, and when conditions are met, uses a "fast |
|
* path" employing direct 32-bit/64-bit reads, resulting in _dramatically |
|
* faster_ read speed. |
|
* |
|
* The check costs one initial branch per hash, which is generally negligible, |
|
* but not zero. |
|
* |
|
* Moreover, it's not useful to generate an additional code path if memory |
|
* access uses the same instruction for both aligned and unaligned |
|
* addresses (e.g. x86 and aarch64). |
|
* |
|
* In these cases, the alignment check can be removed by setting this macro to 0. |
|
* Then the code will always use unaligned memory access. |
|
* Align check is automatically disabled on x86, x64 & arm64, |
|
* which are platforms known to offer good unaligned memory accesses performance. |
|
* |
|
* This option does not affect XXH3 (only XXH32 and XXH64). |
|
*/ |
|
# define XXH_FORCE_ALIGN_CHECK 0 |
|
|
|
/*! |
|
* @def XXH_NO_INLINE_HINTS |
|
* @brief When non-zero, sets all functions to `static`. |
|
* |
|
* By default, xxHash tries to force the compiler to inline almost all internal |
|
* functions. |
|
* |
|
* This can usually improve performance due to reduced jumping and improved |
|
* constant folding, but significantly increases the size of the binary which |
|
* might not be favorable. |
|
* |
|
* Additionally, sometimes the forced inlining can be detrimental to performance, |
|
* depending on the architecture. |
|
* |
|
* XXH_NO_INLINE_HINTS marks all internal functions as static, giving the |
|
* compiler full control on whether to inline or not. |
|
* |
|
* When not optimizing (-O0), optimizing for size (-Os, -Oz), or using |
|
* -fno-inline with GCC or Clang, this will automatically be defined. |
|
*/ |
|
# define XXH_NO_INLINE_HINTS 0 |
|
|
|
/*! |
|
* @def XXH32_ENDJMP |
|
* @brief Whether to use a jump for `XXH32_finalize`. |
|
* |
|
* For performance, `XXH32_finalize` uses multiple branches in the finalizer. |
|
* This is generally preferable for performance, |
|
* but depending on exact architecture, a jmp may be preferable. |
|
* |
|
* This setting is only possibly making a difference for very small inputs. |
|
*/ |
|
# define XXH32_ENDJMP 0 |
|
|
|
/*! |
|
* @internal |
|
* @brief Redefines old internal names. |
|
* |
|
* For compatibility with code that uses xxHash's internals before the names |
|
* were changed to improve namespacing. There is no other reason to use this. |
|
*/ |
|
# define XXH_OLD_NAMES |
|
# undef XXH_OLD_NAMES /* don't actually use, it is ugly. */ |
|
#endif /* XXH_DOXYGEN */ |
|
/*! |
|
* @} |
|
*/ |
|
|
|
#ifndef XXH_FORCE_MEMORY_ACCESS /* can be defined externally, on command line for example */ |
|
/* prefer __packed__ structures (method 1) for gcc on armv7+ and mips */ |
|
# if !defined(__clang__) && \ |
|
( \ |
|
(defined(__INTEL_COMPILER) && !defined(_WIN32)) || \ |
|
( \ |
|
defined(__GNUC__) && ( \ |
|
(defined(__ARM_ARCH) && __ARM_ARCH >= 7) || \ |
|
( \ |
|
defined(__mips__) && \ |
|
(__mips <= 5 || __mips_isa_rev < 6) && \ |
|
(!defined(__mips16) || defined(__mips_mips16e2)) \ |
|
) \ |
|
) \ |
|
) \ |
|
) |
|
# define XXH_FORCE_MEMORY_ACCESS 1 |
|
# endif |
|
#endif |
|
|
|
#ifndef XXH_FORCE_ALIGN_CHECK /* can be defined externally */ |
|
# if defined(__i386) || defined(__x86_64__) || defined(__aarch64__) \ |
|
|| defined(_M_IX86) || defined(_M_X64) || defined(_M_ARM64) /* visual */ |
|
# define XXH_FORCE_ALIGN_CHECK 0 |
|
# else |
|
# define XXH_FORCE_ALIGN_CHECK 1 |
|
# endif |
|
#endif |
|
|
|
#ifndef XXH_NO_INLINE_HINTS |
|
# if defined(__OPTIMIZE_SIZE__) /* -Os, -Oz */ \ |
|
|| defined(__NO_INLINE__) /* -O0, -fno-inline */ |
|
# define XXH_NO_INLINE_HINTS 1 |
|
# else |
|
# define XXH_NO_INLINE_HINTS 0 |
|
# endif |
|
#endif |
|
|
|
#ifndef XXH32_ENDJMP |
|
/* generally preferable for performance */ |
|
# define XXH32_ENDJMP 0 |
|
#endif |
|
|
|
/*! |
|
* @defgroup impl Implementation |
|
* @{ |
|
*/ |
|
|
|
|
|
/* ************************************* |
|
* Includes & Memory related functions |
|
***************************************/ |
|
/* |
|
* Modify the local functions below should you wish to use |
|
* different memory routines for malloc() and free() |
|
*/ |
|
#include <stdlib.h> |
|
|
|
/*! |
|
* @internal |
|
* @brief Modify this function to use a different routine than malloc(). |
|
*/ |
|
static void* XXH_malloc(size_t s) { return malloc(s); } |
|
|
|
/*! |
|
* @internal |
|
* @brief Modify this function to use a different routine than free(). |
|
*/ |
|
static void XXH_free(void* p) { free(p); } |
|
|
|
#include <string.h> |
|
|
|
/*! |
|
* @internal |
|
* @brief Modify this function to use a different routine than memcpy(). |
|
*/ |
|
static void* XXH_memcpy(void* dest, const void* src, size_t size) |
|
{ |
|
return memcpy(dest,src,size); |
|
} |
|
|
|
#include <limits.h> /* ULLONG_MAX */ |
|
|
|
|
|
/* ************************************* |
|
* Compiler Specific Options |
|
***************************************/ |
|
#ifdef _MSC_VER /* Visual Studio warning fix */ |
|
# pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */ |
|
#endif |
|
|
|
#if XXH_NO_INLINE_HINTS /* disable inlining hints */ |
|
# if defined(__GNUC__) || defined(__clang__) |
|
# define XXH_FORCE_INLINE static __attribute__((unused)) |
|
# else |
|
# define XXH_FORCE_INLINE static |
|
# endif |
|
# define XXH_NO_INLINE static |
|
/* enable inlining hints */ |
|
#elif defined(__GNUC__) || defined(__clang__) |
|
# define XXH_FORCE_INLINE static __inline__ __attribute__((always_inline, unused)) |
|
# define XXH_NO_INLINE static __attribute__((noinline)) |
|
#elif defined(_MSC_VER) /* Visual Studio */ |
|
# define XXH_FORCE_INLINE static __forceinline |
|
# define XXH_NO_INLINE static __declspec(noinline) |
|
#elif defined (__cplusplus) \ |
|
|| (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L)) /* C99 */ |
|
# define XXH_FORCE_INLINE static inline |
|
# define XXH_NO_INLINE static |
|
#else |
|
# define XXH_FORCE_INLINE static |
|
# define XXH_NO_INLINE static |
|
#endif |
|
|
|
|
|
|
|
/* ************************************* |
|
* Debug |
|
***************************************/ |
|
/*! |
|
* @ingroup tuning |
|
* @def XXH_DEBUGLEVEL |
|
* @brief Sets the debugging level. |
|
* |
|
* XXH_DEBUGLEVEL is expected to be defined externally, typically via the |
|
* compiler's command line options. The value must be a number. |
|
*/ |
|
#ifndef XXH_DEBUGLEVEL |
|
# ifdef DEBUGLEVEL /* backwards compat */ |
|
# define XXH_DEBUGLEVEL DEBUGLEVEL |
|
# else |
|
# define XXH_DEBUGLEVEL 0 |
|
# endif |
|
#endif |
|
|
|
#if (XXH_DEBUGLEVEL>=1) |
|
# include <assert.h> /* note: can still be disabled with NDEBUG */ |
|
# define XXH_ASSERT(c) assert(c) |
|
#else |
|
# define XXH_ASSERT(c) ((void)0) |
|
#endif |
|
|
|
/* note: use after variable declarations */ |
|
#ifndef XXH_STATIC_ASSERT |
|
# if defined(__STDC_VERSION__) && (__STDC_VERSION__ >= 201112L) /* C11 */ |
|
# include <assert.h> |
|
# ifndef static_assert /* uClibc does not define static_assert */ |
|
# define static_assert _Static_assert |
|
# endif |
|
# define XXH_STATIC_ASSERT_WITH_MESSAGE(c,m) do { static_assert((c),m); } while(0) |
|
# elif defined(__cplusplus) && (__cplusplus >= 201103L) /* C++11 */ |
|
# define XXH_STATIC_ASSERT_WITH_MESSAGE(c,m) do { static_assert((c),m); } while(0) |
|
# else |
|
# define XXH_STATIC_ASSERT_WITH_MESSAGE(c,m) do { struct xxh_sa { char x[(c) ? 1 : -1]; }; } while(0) |
|
# endif |
|
# define XXH_STATIC_ASSERT(c) XXH_STATIC_ASSERT_WITH_MESSAGE((c),#c) |
|
#endif |
|
|
|
/*! |
|
* @internal |
|
* @def XXH_COMPILER_GUARD(var) |
|
* @brief Used to prevent unwanted optimizations for @p var. |
|
* |
|
* It uses an empty GCC inline assembly statement with a register constraint |
|
* which forces @p var into a general purpose register (eg eax, ebx, ecx |
|
* on x86) and marks it as modified. |
|
* |
|
* This is used in a few places to avoid unwanted autovectorization (e.g. |
|
* XXH32_round()). All vectorization we want is explicit via intrinsics, |
|
* and _usually_ isn't wanted elsewhere. |
|
* |
|
* We also use it to prevent unwanted constant folding for AArch64 in |
|
* XXH3_initCustomSecret_scalar(). |
|
*/ |
|
#if defined(__GNUC__) || defined(__clang__) |
|
# define XXH_COMPILER_GUARD(var) __asm__ __volatile__("" : "+r" (var)) |
|
#else |
|
# define XXH_COMPILER_GUARD(var) ((void)0) |
|
#endif |
|
|
|
/* ************************************* |
|
* Basic Types |
|
***************************************/ |
|
#if !defined (__VMS) \ |
|
&& (defined (__cplusplus) \ |
|
|| (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) ) |
|
# include <stdint.h> |
|
typedef uint8_t xxh_u8; |
|
#else |
|
typedef unsigned char xxh_u8; |
|
#endif |
|
typedef XXH32_hash_t xxh_u32; |
|
|
|
#ifdef XXH_OLD_NAMES |
|
# define BYTE xxh_u8 |
|
# define U8 xxh_u8 |
|
# define U32 xxh_u32 |
|
#endif |
|
|
|
/* *** Memory access *** */ |
|
|
|
/*! |
|
* @internal |
|
* @fn xxh_u32 XXH_read32(const void* ptr) |
|
* @brief Reads an unaligned 32-bit integer from @p ptr in native endianness. |
|
* |
|
* Affected by @ref XXH_FORCE_MEMORY_ACCESS. |
|
* |
|
* @param ptr The pointer to read from. |
|
* @return The 32-bit native endian integer from the bytes at @p ptr. |
|
*/ |
|
|
|
/*! |
|
* @internal |
|
* @fn xxh_u32 XXH_readLE32(const void* ptr) |
|
* @brief Reads an unaligned 32-bit little endian integer from @p ptr. |
|
* |
|
* Affected by @ref XXH_FORCE_MEMORY_ACCESS. |
|
* |
|
* @param ptr The pointer to read from. |
|
* @return The 32-bit little endian integer from the bytes at @p ptr. |
|
*/ |
|
|
|
/*! |
|
* @internal |
|
* @fn xxh_u32 XXH_readBE32(const void* ptr) |
|
* @brief Reads an unaligned 32-bit big endian integer from @p ptr. |
|
* |
|
* Affected by @ref XXH_FORCE_MEMORY_ACCESS. |
|
* |
|
* @param ptr The pointer to read from. |
|
* @return The 32-bit big endian integer from the bytes at @p ptr. |
|
*/ |
|
|
|
/*! |
|
* @internal |
|
* @fn xxh_u32 XXH_readLE32_align(const void* ptr, XXH_alignment align) |
|
* @brief Like @ref XXH_readLE32(), but has an option for aligned reads. |
|
* |
|
* Affected by @ref XXH_FORCE_MEMORY_ACCESS. |
|
* Note that when @ref XXH_FORCE_ALIGN_CHECK == 0, the @p align parameter is |
|
* always @ref XXH_alignment::XXH_unaligned. |
|
* |
|
* @param ptr The pointer to read from. |
|
* @param align Whether @p ptr is aligned. |
|
* @pre |
|
* If @p align == @ref XXH_alignment::XXH_aligned, @p ptr must be 4 byte |
|
* aligned. |
|
* @return The 32-bit little endian integer from the bytes at @p ptr. |
|
*/ |
|
|
|
#if (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==3)) |
|
/* |
|
* Manual byteshift. Best for old compilers which don't inline memcpy. |
|
* We actually directly use XXH_readLE32 and XXH_readBE32. |
|
*/ |
|
#elif (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==2)) |
|
|
|
/* |
|
* Force direct memory access. Only works on CPU which support unaligned memory |
|
* access in hardware. |
|
*/ |
|
static xxh_u32 XXH_read32(const void* memPtr) { return *(const xxh_u32*) memPtr; } |
|
|
|
#elif (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==1)) |
|
|
|
/* |
|
* __pack instructions are safer but compiler specific, hence potentially |
|
* problematic for some compilers. |
|
* |
|
* Currently only defined for GCC and ICC. |
|
*/ |
|
#ifdef XXH_OLD_NAMES |
|
typedef union { xxh_u32 u32; } __attribute__((packed)) unalign; |
|
#endif |
|
static xxh_u32 XXH_read32(const void* ptr) |
|
{ |
|
typedef union { xxh_u32 u32; } __attribute__((packed)) xxh_unalign; |
|
return ((const xxh_unalign*)ptr)->u32; |
|
} |
|
|
|
#else |
|
|
|
/* |
|
* Portable and safe solution. Generally efficient. |
|
* see: http://fastcompression.blogspot.com/2015/08/accessing-unaligned-memory.html |
|
*/ |
|
static xxh_u32 XXH_read32(const void* memPtr) |
|
{ |
|
xxh_u32 val; |
|
XXH_memcpy(&val, memPtr, sizeof(val)); |
|
return val; |
|
} |
|
|
|
#endif /* XXH_FORCE_DIRECT_MEMORY_ACCESS */ |
|
|
|
|
|
/* *** Endianness *** */ |
|
|
|
/*! |
|
* @ingroup tuning |
|
* @def XXH_CPU_LITTLE_ENDIAN |
|
* @brief Whether the target is little endian. |
|
* |
|
* Defined to 1 if the target is little endian, or 0 if it is big endian. |
|
* It can be defined externally, for example on the compiler command line. |
|
* |
|
* If it is not defined, |
|
* a runtime check (which is usually constant folded) is used instead. |
|
* |
|
* @note |
|
* This is not necessarily defined to an integer constant. |
|
* |
|
* @see XXH_isLittleEndian() for the runtime check. |
|
*/ |
|
#ifndef XXH_CPU_LITTLE_ENDIAN |
|
/* |
|
* Try to detect endianness automatically, to avoid the nonstandard behavior |
|
* in `XXH_isLittleEndian()` |
|
*/ |
|
# if defined(_WIN32) /* Windows is always little endian */ \ |
|
|| defined(__LITTLE_ENDIAN__) \ |
|
|| (defined(__BYTE_ORDER__) && __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__) |
|
# define XXH_CPU_LITTLE_ENDIAN 1 |
|
# elif defined(__BIG_ENDIAN__) \ |
|
|| (defined(__BYTE_ORDER__) && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__) |
|
# define XXH_CPU_LITTLE_ENDIAN 0 |
|
# else |
|
/*! |
|
* @internal |
|
* @brief Runtime check for @ref XXH_CPU_LITTLE_ENDIAN. |
|
* |
|
* Most compilers will constant fold this. |
|
*/ |
|
static int XXH_isLittleEndian(void) |
|
{ |
|
/* |
|
* Portable and well-defined behavior. |
|
* Don't use static: it is detrimental to performance. |
|
*/ |
|
const union { xxh_u32 u; xxh_u8 c[4]; } one = { 1 }; |
|
return one.c[0]; |
|
} |
|
# define XXH_CPU_LITTLE_ENDIAN XXH_isLittleEndian() |
|
# endif |
|
#endif |
|
|
|
|
|
|
|
|
|
/* **************************************** |
|
* Compiler-specific Functions and Macros |
|
******************************************/ |
|
#define XXH_GCC_VERSION (__GNUC__ * 100 + __GNUC_MINOR__) |
|
|
|
#ifdef __has_builtin |
|
# define XXH_HAS_BUILTIN(x) __has_builtin(x) |
|
#else |
|
# define XXH_HAS_BUILTIN(x) 0 |
|
#endif |
|
|
|
/*! |
|
* @internal |
|
* @def XXH_rotl32(x,r) |
|
* @brief 32-bit rotate left. |
|
* |
|
* @param x The 32-bit integer to be rotated. |
|
* @param r The number of bits to rotate. |
|
* @pre |
|
* @p r > 0 && @p r < 32 |
|
* @note |
|
* @p x and @p r may be evaluated multiple times. |
|
* @return The rotated result. |
|
*/ |
|
#if !defined(NO_CLANG_BUILTIN) && XXH_HAS_BUILTIN(__builtin_rotateleft32) \ |
|
&& XXH_HAS_BUILTIN(__builtin_rotateleft64) |
|
# define XXH_rotl32 __builtin_rotateleft32 |
|
# define XXH_rotl64 __builtin_rotateleft64 |
|
/* Note: although _rotl exists for minGW (GCC under windows), performance seems poor */ |
|
#elif defined(_MSC_VER) |
|
# define XXH_rotl32(x,r) _rotl(x,r) |
|
# define XXH_rotl64(x,r) _rotl64(x,r) |
|
#else |
|
# define XXH_rotl32(x,r) (((x) << (r)) | ((x) >> (32 - (r)))) |
|
# define XXH_rotl64(x,r) (((x) << (r)) | ((x) >> (64 - (r)))) |
|
#endif |
|
|
|
/*! |
|
* @internal |
|
* @fn xxh_u32 XXH_swap32(xxh_u32 x) |
|
* @brief A 32-bit byteswap. |
|
* |
|
* @param x The 32-bit integer to byteswap. |
|
* @return @p x, byteswapped. |
|
*/ |
|
#if defined(_MSC_VER) /* Visual Studio */ |
|
# define XXH_swap32 _byteswap_ulong |
|
#elif XXH_GCC_VERSION >= 403 |
|
# define XXH_swap32 __builtin_bswap32 |
|
#else |
|
static xxh_u32 XXH_swap32 (xxh_u32 x) |
|
{ |
|
return ((x << 24) & 0xff000000 ) | |
|
((x << 8) & 0x00ff0000 ) | |
|
((x >> 8) & 0x0000ff00 ) | |
|
((x >> 24) & 0x000000ff ); |
|
} |
|
#endif |
|
|
|
|
|
/* *************************** |
|
* Memory reads |
|
*****************************/ |
|
|
|
/*! |
|
* @internal |
|
* @brief Enum to indicate whether a pointer is aligned. |
|
*/ |
|
typedef enum { |
|
XXH_aligned, /*!< Aligned */ |
|
XXH_unaligned /*!< Possibly unaligned */ |
|
} XXH_alignment; |
|
|
|
/* |
|
* XXH_FORCE_MEMORY_ACCESS==3 is an endian-independent byteshift load. |
|
* |
|
* This is ideal for older compilers which don't inline memcpy. |
|
*/ |
|
#if (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==3)) |
|
|
|
XXH_FORCE_INLINE xxh_u32 XXH_readLE32(const void* memPtr) |
|
{ |
|
const xxh_u8* bytePtr = (const xxh_u8 *)memPtr; |
|
return bytePtr[0] |
|
| ((xxh_u32)bytePtr[1] << 8) |
|
| ((xxh_u32)bytePtr[2] << 16) |
|
| ((xxh_u32)bytePtr[3] << 24); |
|
} |
|
|
|
XXH_FORCE_INLINE xxh_u32 XXH_readBE32(const void* memPtr) |
|
{ |
|
const xxh_u8* bytePtr = (const xxh_u8 *)memPtr; |
|
return bytePtr[3] |
|
| ((xxh_u32)bytePtr[2] << 8) |
|
| ((xxh_u32)bytePtr[1] << 16) |
|
| ((xxh_u32)bytePtr[0] << 24); |
|
} |
|
|
|
#else |
|
XXH_FORCE_INLINE xxh_u32 XXH_readLE32(const void* ptr) |
|
{ |
|
return XXH_CPU_LITTLE_ENDIAN ? XXH_read32(ptr) : XXH_swap32(XXH_read32(ptr)); |
|
} |
|
|
|
static xxh_u32 XXH_readBE32(const void* ptr) |
|
{ |
|
return XXH_CPU_LITTLE_ENDIAN ? XXH_swap32(XXH_read32(ptr)) : XXH_read32(ptr); |
|
} |
|
#endif |
|
|
|
XXH_FORCE_INLINE xxh_u32 |
|
XXH_readLE32_align(const void* ptr, XXH_alignment align) |
|
{ |
|
if (align==XXH_unaligned) { |
|
return XXH_readLE32(ptr); |
|
} else { |
|
return XXH_CPU_LITTLE_ENDIAN ? *(const xxh_u32*)ptr : XXH_swap32(*(const xxh_u32*)ptr); |
|
} |
|
} |
|
|
|
|
|
/* ************************************* |
|
* Misc |
|
***************************************/ |
|
/*! @ingroup public */ |
|
XXH_PUBLIC_API unsigned XXH_versionNumber (void) { return XXH_VERSION_NUMBER; } |
|
|
|
|
|
/* ******************************************************************* |
|
* 32-bit hash functions |
|
*********************************************************************/ |
|
/*! |
|
* @} |
|
* @defgroup xxh32_impl XXH32 implementation |
|
* @ingroup impl |
|
* @{ |
|
*/ |
|
/* #define instead of static const, to be used as initializers */ |
|
#define XXH_PRIME32_1 0x9E3779B1U /*!< 0b10011110001101110111100110110001 */ |
|
#define XXH_PRIME32_2 0x85EBCA77U /*!< 0b10000101111010111100101001110111 */ |
|
#define XXH_PRIME32_3 0xC2B2AE3DU /*!< 0b11000010101100101010111000111101 */ |
|
#define XXH_PRIME32_4 0x27D4EB2FU /*!< 0b00100111110101001110101100101111 */ |
|
#define XXH_PRIME32_5 0x165667B1U /*!< 0b00010110010101100110011110110001 */ |
|
|
|
#ifdef XXH_OLD_NAMES |
|
# define PRIME32_1 XXH_PRIME32_1 |
|
# define PRIME32_2 XXH_PRIME32_2 |
|
# define PRIME32_3 XXH_PRIME32_3 |
|
# define PRIME32_4 XXH_PRIME32_4 |
|
# define PRIME32_5 XXH_PRIME32_5 |
|
#endif |
|
|
|
/*! |
|
* @internal |
|
* @brief Normal stripe processing routine. |
|
* |
|
* This shuffles the bits so that any bit from @p input impacts several bits in |
|
* @p acc. |
|
* |
|
* @param acc The accumulator lane. |
|
* @param input The stripe of input to mix. |
|
* @return The mixed accumulator lane. |
|
*/ |
|
static xxh_u32 XXH32_round(xxh_u32 acc, xxh_u32 input) |
|
{ |
|
acc += input * XXH_PRIME32_2; |
|
acc = XXH_rotl32(acc, 13); |
|
acc *= XXH_PRIME32_1; |
|
#if (defined(__SSE4_1__) || defined(__aarch64__)) && !defined(XXH_ENABLE_AUTOVECTORIZE) |
|
/* |
|
* UGLY HACK: |
|
* A compiler fence is the only thing that prevents GCC and Clang from |
|
* autovectorizing the XXH32 loop (pragmas and attributes don't work for some |
|
* reason) without globally disabling SSE4.1. |
|
* |
|
* The reason we want to avoid vectorization is because despite working on |
|
* 4 integers at a time, there are multiple factors slowing XXH32 down on |
|
* SSE4: |
|
* - There's a ridiculous amount of lag from pmulld (10 cycles of latency on |
|
* newer chips!) making it slightly slower to multiply four integers at |
|
* once compared to four integers independently. Even when pmulld was |
|
* fastest, Sandy/Ivy Bridge, it is still not worth it to go into SSE |
|
* just to multiply unless doing a long operation. |
|
* |
|
* - Four instructions are required to rotate, |
|
* movqda tmp, v // not required with VEX encoding |
|
* pslld tmp, 13 // tmp <<= 13 |
|
* psrld v, 19 // x >>= 19 |
|
* por v, tmp // x |= tmp |
|
* compared to one for scalar: |
|
* roll v, 13 // reliably fast across the board |
|
* shldl v, v, 13 // Sandy Bridge and later prefer this for some reason |
|
* |
|
* - Instruction level parallelism is actually more beneficial here because |
|
* the SIMD actually serializes this operation: While v1 is rotating, v2 |
|
* can load data, while v3 can multiply. SSE forces them to operate |
|
* together. |
|
* |
|
* This is also enabled on AArch64, as Clang autovectorizes it incorrectly |
|
* and it is pointless writing a NEON implementation that is basically the |
|
* same speed as scalar for XXH32. |
|
*/ |
|
XXH_COMPILER_GUARD(acc); |
|
#endif |
|
return acc; |
|
} |
|
|
|
/*! |
|
* @internal |
|
* @brief Mixes all bits to finalize the hash. |
|
* |
|
* The final mix ensures that all input bits have a chance to impact any bit in |
|
* the output digest, resulting in an unbiased distribution. |
|
* |
|
* @param h32 The hash to avalanche. |
|
* @return The avalanched hash. |
|
*/ |
|
static xxh_u32 XXH32_avalanche(xxh_u32 h32) |
|
{ |
|
h32 ^= h32 >> 15; |
|
h32 *= XXH_PRIME32_2; |
|
h32 ^= h32 >> 13; |
|
h32 *= XXH_PRIME32_3; |
|
h32 ^= h32 >> 16; |
|
return(h32); |
|
} |
|
|
|
#define XXH_get32bits(p) XXH_readLE32_align(p, align) |
|
|
|
/*! |
|
* @internal |
|
* @brief Processes the last 0-15 bytes of @p ptr. |
|
* |
|
* There may be up to 15 bytes remaining to consume from the input. |
|
* This final stage will digest them to ensure that all input bytes are present |
|
* in the final mix. |
|
* |
|
* @param h32 The hash to finalize. |
|
* @param ptr The pointer to the remaining input. |
|
* @param len The remaining length, modulo 16. |
|
* @param align Whether @p ptr is aligned. |
|
* @return The finalized hash. |
|
*/ |
|
static xxh_u32 |
|
XXH32_finalize(xxh_u32 h32, const xxh_u8* ptr, size_t len, XXH_alignment align) |
|
{ |
|
#define XXH_PROCESS1 do { \ |
|
h32 += (*ptr++) * XXH_PRIME32_5; \ |
|
h32 = XXH_rotl32(h32, 11) * XXH_PRIME32_1; \ |
|
} while (0) |
|
|
|
#define XXH_PROCESS4 do { \ |
|
h32 += XXH_get32bits(ptr) * XXH_PRIME32_3; \ |
|
ptr += 4; \ |
|
h32 = XXH_rotl32(h32, 17) * XXH_PRIME32_4; \ |
|
} while (0) |
|
|
|
if (ptr==NULL) XXH_ASSERT(len == 0); |
|
|
|
/* Compact rerolled version; generally faster */ |
|
if (!XXH32_ENDJMP) { |
|
len &= 15; |
|
while (len >= 4) { |
|
XXH_PROCESS4; |
|
len -= 4; |
|
} |
|
while (len > 0) { |
|
XXH_PROCESS1; |
|
--len; |
|
} |
|
return XXH32_avalanche(h32); |
|
} else { |
|
switch(len&15) /* or switch(bEnd - p) */ { |
|
case 12: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 8: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 4: XXH_PROCESS4; |
|
return XXH32_avalanche(h32); |
|
|
|
case 13: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 9: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 5: XXH_PROCESS4; |
|
XXH_PROCESS1; |
|
return XXH32_avalanche(h32); |
|
|
|
case 14: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 10: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 6: XXH_PROCESS4; |
|
XXH_PROCESS1; |
|
XXH_PROCESS1; |
|
return XXH32_avalanche(h32); |
|
|
|
case 15: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 11: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 7: XXH_PROCESS4; |
|
XXH_FALLTHROUGH; |
|
case 3: XXH_PROCESS1; |
|
XXH_FALLTHROUGH; |
|
case 2: XXH_PROCESS1; |
|
XXH_FALLTHROUGH; |
|
case 1: XXH_PROCESS1; |
|
XXH_FALLTHROUGH; |
|
case 0: return XXH32_avalanche(h32); |
|
} |
|
XXH_ASSERT(0); |
|
return h32; /* reaching this point is deemed impossible */ |
|
} |
|
} |
|
|
|
#ifdef XXH_OLD_NAMES |
|
# define PROCESS1 XXH_PROCESS1 |
|
# define PROCESS4 XXH_PROCESS4 |
|
#else |
|
# undef XXH_PROCESS1 |
|
# undef XXH_PROCESS4 |
|
#endif |
|
|
|
/*! |
|
* @internal |
|
* @brief The implementation for @ref XXH32(). |
|
* |
|
* @param input , len , seed Directly passed from @ref XXH32(). |
|
* @param align Whether @p input is aligned. |
|
* @return The calculated hash. |
|
*/ |
|
XXH_FORCE_INLINE xxh_u32 |
|
XXH32_endian_align(const xxh_u8* input, size_t len, xxh_u32 seed, XXH_alignment align) |
|
{ |
|
xxh_u32 h32; |
|
|
|
if (input==NULL) XXH_ASSERT(len == 0); |
|
|
|
if (len>=16) { |
|
const xxh_u8* const bEnd = input + len; |
|
const xxh_u8* const limit = bEnd - 15; |
|
xxh_u32 v1 = seed + XXH_PRIME32_1 + XXH_PRIME32_2; |
|
xxh_u32 v2 = seed + XXH_PRIME32_2; |
|
xxh_u32 v3 = seed + 0; |
|
xxh_u32 v4 = seed - XXH_PRIME32_1; |
|
|
|
do { |
|
v1 = XXH32_round(v1, XXH_get32bits(input)); input += 4; |
|
v2 = XXH32_round(v2, XXH_get32bits(input)); input += 4; |
|
v3 = XXH32_round(v3, XXH_get32bits(input)); input += 4; |
|
v4 = XXH32_round(v4, XXH_get32bits(input)); input += 4; |
|
} while (input < limit); |
|
|
|
h32 = XXH_rotl32(v1, 1) + XXH_rotl32(v2, 7) |
|
+ XXH_rotl32(v3, 12) + XXH_rotl32(v4, 18); |
|
} else { |
|
h32 = seed + XXH_PRIME32_5; |
|
} |
|
|
|
h32 += (xxh_u32)len; |
|
|
|
return XXH32_finalize(h32, input, len&15, align); |
|
} |
|
|
|
/*! @ingroup xxh32_family */ |
|
XXH_PUBLIC_API XXH32_hash_t XXH32 (const void* input, size_t len, XXH32_hash_t seed) |
|
{ |
|
#if 0 |
|
/* Simple version, good for code maintenance, but unfortunately slow for small inputs */ |
|
XXH32_state_t state; |
|
XXH32_reset(&state, seed); |
|
XXH32_update(&state, (const xxh_u8*)input, len); |
|
return XXH32_digest(&state); |
|
#else |
|
if (XXH_FORCE_ALIGN_CHECK) { |
|
if ((((size_t)input) & 3) == 0) { /* Input is 4-bytes aligned, leverage the speed benefit */ |
|
return XXH32_endian_align((const xxh_u8*)input, len, seed, XXH_aligned); |
|
} } |
|
|
|
return XXH32_endian_align((const xxh_u8*)input, len, seed, XXH_unaligned); |
|
#endif |
|
} |
|
|
|
|
|
|
|
/******* Hash streaming *******/ |
|
/*! |
|
* @ingroup xxh32_family |
|
*/ |
|
XXH_PUBLIC_API XXH32_state_t* XXH32_createState(void) |
|
{ |
|
return (XXH32_state_t*)XXH_malloc(sizeof(XXH32_state_t)); |
|
} |
|
/*! @ingroup xxh32_family */ |
|
XXH_PUBLIC_API XXH_errorcode XXH32_freeState(XXH32_state_t* statePtr) |
|
{ |
|
XXH_free(statePtr); |
|
return XXH_OK; |
|
} |
|
|
|
/*! @ingroup xxh32_family */ |
|
XXH_PUBLIC_API void XXH32_copyState(XXH32_state_t* dstState, const XXH32_state_t* srcState) |
|
{ |
|
XXH_memcpy(dstState, srcState, sizeof(*dstState)); |
|
} |
|
|
|
/*! @ingroup xxh32_family */ |
|
XXH_PUBLIC_API XXH_errorcode XXH32_reset(XXH32_state_t* statePtr, XXH32_hash_t seed) |
|
{ |
|
XXH32_state_t state; /* using a local state to memcpy() in order to avoid strict-aliasing warnings */ |
|
memset(&state, 0, sizeof(state)); |
|
state.v[0] = seed + XXH_PRIME32_1 + XXH_PRIME32_2; |
|
state.v[1] = seed + XXH_PRIME32_2; |
|
state.v[2] = seed + 0; |
|
state.v[3] = seed - XXH_PRIME32_1; |
|
/* do not write into reserved, planned to be removed in a future version */ |
|
XXH_memcpy(statePtr, &state, sizeof(state) - sizeof(state.reserved)); |
|
return XXH_OK; |
|
} |
|
|
|
|
|
/*! @ingroup xxh32_family */ |
|
XXH_PUBLIC_API XXH_errorcode |
|
XXH32_update(XXH32_state_t* state, const void* input, size_t len) |
|
{ |
|
if (input==NULL) { |
|
XXH_ASSERT(len == 0); |
|
return XXH_OK; |
|
} |
|
|
|
{ const xxh_u8* p = (const xxh_u8*)input; |
|
const xxh_u8* const bEnd = p + len; |
|
|
|
state->total_len_32 += (XXH32_hash_t)len; |
|
state->large_len |= (XXH32_hash_t)((len>=16) | (state->total_len_32>=16)); |
|
|
|
if (state->memsize + len < 16) { /* fill in tmp buffer */ |
|
XXH_memcpy((xxh_u8*)(state->mem32) + state->memsize, input, len); |
|
state->memsize += (XXH32_hash_t)len; |
|
return XXH_OK; |
|
} |
|
|
|
if (state->memsize) { /* some data left from previous update */ |
|
XXH_memcpy((xxh_u8*)(state->mem32) + state->memsize, input, 16-state->memsize); |
|
{ const xxh_u32* p32 = state->mem32; |
|
state->v[0] = XXH32_round(state->v[0], XXH_readLE32(p32)); p32++; |
|
state->v[1] = XXH32_round(state->v[1], XXH_readLE32(p32)); p32++; |
|
state->v[2] = XXH32_round(state->v[2], XXH_readLE32(p32)); p32++; |
|
state->v[3] = XXH32_round(state->v[3], XXH_readLE32(p32)); |
|
} |
|
p += 16-state->memsize; |
|
state->memsize = 0; |
|
} |
|
|
|
if (p <= bEnd-16) { |
|
const xxh_u8* const limit = bEnd - 16; |
|
|
|
do { |
|
state->v[0] = XXH32_round(state->v[0], XXH_readLE32(p)); p+=4; |
|
state->v[1] = XXH32_round(state->v[1], XXH_readLE32(p)); p+=4; |
|
state->v[2] = XXH32_round(state->v[2], XXH_readLE32(p)); p+=4; |
|
state->v[3] = XXH32_round(state->v[3], XXH_readLE32(p)); p+=4; |
|
} while (p<=limit); |
|
|
|
} |
|
|
|
if (p < bEnd) { |
|
XXH_memcpy(state->mem32, p, (size_t)(bEnd-p)); |
|
state->memsize = (unsigned)(bEnd-p); |
|
} |
|
} |
|
|
|
return XXH_OK; |
|
} |
|
|
|
|
|
/*! @ingroup xxh32_family */ |
|
XXH_PUBLIC_API XXH32_hash_t XXH32_digest(const XXH32_state_t* state) |
|
{ |
|
xxh_u32 h32; |
|
|
|
if (state->large_len) { |
|
h32 = XXH_rotl32(state->v[0], 1) |
|
+ XXH_rotl32(state->v[1], 7) |
|
+ XXH_rotl32(state->v[2], 12) |
|
+ XXH_rotl32(state->v[3], 18); |
|
} else { |
|
h32 = state->v[2] /* == seed */ + XXH_PRIME32_5; |
|
} |
|
|
|
h32 += state->total_len_32; |
|
|
|
return XXH32_finalize(h32, (const xxh_u8*)state->mem32, state->memsize, XXH_aligned); |
|
} |
|
|
|
|
|
/******* Canonical representation *******/ |
|
|
|
/*! |
|
* @ingroup xxh32_family |
|
* The default return values from XXH functions are unsigned 32 and 64 bit |
|
* integers. |
|
* |
|
* The canonical representation uses big endian convention, the same convention |
|
* as human-readable numbers (large digits first). |
|
* |
|
* This way, hash values can be written into a file or buffer, remaining |
|
* comparable across different systems. |
|
* |
|
* The following functions allow transformation of hash values to and from their |
|
* canonical format. |
|
*/ |
|
XXH_PUBLIC_API void XXH32_canonicalFromHash(XXH32_canonical_t* dst, XXH32_hash_t hash) |
|
{ |
|
XXH_STATIC_ASSERT(sizeof(XXH32_canonical_t) == sizeof(XXH32_hash_t)); |
|
if (XXH_CPU_LITTLE_ENDIAN) hash = XXH_swap32(hash); |
|
XXH_memcpy(dst, &hash, sizeof(*dst)); |
|
} |
|
/*! @ingroup xxh32_family */ |
|
XXH_PUBLIC_API XXH32_hash_t XXH32_hashFromCanonical(const XXH32_canonical_t* src) |
|
{ |
|
return XXH_readBE32(src); |
|
} |
|
|
|
|
|
#ifndef XXH_NO_LONG_LONG |
|
|
|
/* ******************************************************************* |
|
* 64-bit hash functions |
|
*********************************************************************/ |
|
/*! |
|
* @} |
|
* @ingroup impl |
|
* @{ |
|
*/ |
|
/******* Memory access *******/ |
|
|
|
typedef XXH64_hash_t xxh_u64; |
|
|
|
#ifdef XXH_OLD_NAMES |
|
# define U64 xxh_u64 |
|
#endif |
|
|
|
#if (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==3)) |
|
/* |
|
* Manual byteshift. Best for old compilers which don't inline memcpy. |
|
* We actually directly use XXH_readLE64 and XXH_readBE64. |
|
*/ |
|
#elif (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==2)) |
|
|
|
/* Force direct memory access. Only works on CPU which support unaligned memory access in hardware */ |
|
static xxh_u64 XXH_read64(const void* memPtr) |
|
{ |
|
return *(const xxh_u64*) memPtr; |
|
} |
|
|
|
#elif (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==1)) |
|
|
|
/* |
|
* __pack instructions are safer, but compiler specific, hence potentially |
|
* problematic for some compilers. |
|
* |
|
* Currently only defined for GCC and ICC. |
|
*/ |
|
#ifdef XXH_OLD_NAMES |
|
typedef union { xxh_u32 u32; xxh_u64 u64; } __attribute__((packed)) unalign64; |
|
#endif |
|
static xxh_u64 XXH_read64(const void* ptr) |
|
{ |
|
typedef union { xxh_u32 u32; xxh_u64 u64; } __attribute__((packed)) xxh_unalign64; |
|
return ((const xxh_unalign64*)ptr)->u64; |
|
} |
|
|
|
#else |
|
|
|
/* |
|
* Portable and safe solution. Generally efficient. |
|
* see: http://fastcompression.blogspot.com/2015/08/accessing-unaligned-memory.html |
|
*/ |
|
static xxh_u64 XXH_read64(const void* memPtr) |
|
{ |
|
xxh_u64 val; |
|
XXH_memcpy(&val, memPtr, sizeof(val)); |
|
return val; |
|
} |
|
|
|
#endif /* XXH_FORCE_DIRECT_MEMORY_ACCESS */ |
|
|
|
#if defined(_MSC_VER) /* Visual Studio */ |
|
# define XXH_swap64 _byteswap_uint64 |
|
#elif XXH_GCC_VERSION >= 403 |
|
# define XXH_swap64 __builtin_bswap64 |
|
#else |
|
static xxh_u64 XXH_swap64(xxh_u64 x) |
|
{ |
|
return ((x << 56) & 0xff00000000000000ULL) | |
|
((x << 40) & 0x00ff000000000000ULL) | |
|
((x << 24) & 0x0000ff0000000000ULL) | |
|
((x << 8) & 0x000000ff00000000ULL) | |
|
((x >> 8) & 0x00000000ff000000ULL) | |
|
((x >> 24) & 0x0000000000ff0000ULL) | |
|
((x >> 40) & 0x000000000000ff00ULL) | |
|
((x >> 56) & 0x00000000000000ffULL); |
|
} |
|
#endif |
|
|
|
|
|
/* XXH_FORCE_MEMORY_ACCESS==3 is an endian-independent byteshift load. */ |
|
#if (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==3)) |
|
|
|
XXH_FORCE_INLINE xxh_u64 XXH_readLE64(const void* memPtr) |
|
{ |
|
const xxh_u8* bytePtr = (const xxh_u8 *)memPtr; |
|
return bytePtr[0] |
|
| ((xxh_u64)bytePtr[1] << 8) |
|
| ((xxh_u64)bytePtr[2] << 16) |
|
| ((xxh_u64)bytePtr[3] << 24) |
|
| ((xxh_u64)bytePtr[4] << 32) |
|
| ((xxh_u64)bytePtr[5] << 40) |
|
| ((xxh_u64)bytePtr[6] << 48) |
|
| ((xxh_u64)bytePtr[7] << 56); |
|
} |
|
|
|
XXH_FORCE_INLINE xxh_u64 XXH_readBE64(const void* memPtr) |
|
{ |
|
const xxh_u8* bytePtr = (const xxh_u8 *)memPtr; |
|
return bytePtr[7] |
|
| ((xxh_u64)bytePtr[6] << 8) |
|
| ((xxh_u64)bytePtr[5] << 16) |
|
| ((xxh_u64)bytePtr[4] << 24) |
|
| ((xxh_u64)bytePtr[3] << 32) |
|
| ((xxh_u64)bytePtr[2] << 40) |
|
| ((xxh_u64)bytePtr[1] << 48) |
|
| ((xxh_u64)bytePtr[0] << 56); |
|
} |
|
|
|
#else |
|
XXH_FORCE_INLINE xxh_u64 XXH_readLE64(const void* ptr) |
|
{ |
|
return XXH_CPU_LITTLE_ENDIAN ? XXH_read64(ptr) : XXH_swap64(XXH_read64(ptr)); |
|
} |
|
|
|
static xxh_u64 XXH_readBE64(const void* ptr) |
|
{ |
|
return XXH_CPU_LITTLE_ENDIAN ? XXH_swap64(XXH_read64(ptr)) : XXH_read64(ptr); |
|
} |
|
#endif |
|
|
|
XXH_FORCE_INLINE xxh_u64 |
|
XXH_readLE64_align(const void* ptr, XXH_alignment align) |
|
{ |
|
if (align==XXH_unaligned) |
|
return XXH_readLE64(ptr); |
|
else |
|
return XXH_CPU_LITTLE_ENDIAN ? *(const xxh_u64*)ptr : XXH_swap64(*(const xxh_u64*)ptr); |
|
} |
|
|
|
|
|
/******* xxh64 *******/ |
|
/*! |
|
* @} |
|
* @defgroup xxh64_impl XXH64 implementation |
|
* @ingroup impl |
|
* @{ |
|
*/ |
|
/* #define rather that static const, to be used as initializers */ |
|
#define XXH_PRIME64_1 0x9E3779B185EBCA87ULL /*!< 0b1001111000110111011110011011000110000101111010111100101010000111 */ |
|
#define XXH_PRIME64_2 0xC2B2AE3D27D4EB4FULL /*!< 0b1100001010110010101011100011110100100111110101001110101101001111 */ |
|
#define XXH_PRIME64_3 0x165667B19E3779F9ULL /*!< 0b0001011001010110011001111011000110011110001101110111100111111001 */ |
|
#define XXH_PRIME64_4 0x85EBCA77C2B2AE63ULL /*!< 0b1000010111101011110010100111011111000010101100101010111001100011 */ |
|
#define XXH_PRIME64_5 0x27D4EB2F165667C5ULL /*!< 0b0010011111010100111010110010111100010110010101100110011111000101 */ |
|
|
|
#ifdef XXH_OLD_NAMES |
|
# define PRIME64_1 XXH_PRIME64_1 |
|
# define PRIME64_2 XXH_PRIME64_2 |
|
# define PRIME64_3 XXH_PRIME64_3 |
|
# define PRIME64_4 XXH_PRIME64_4 |
|
# define PRIME64_5 XXH_PRIME64_5 |
|
#endif |
|
|
|
static xxh_u64 XXH64_round(xxh_u64 acc, xxh_u64 input) |
|
{ |
|
acc += input * XXH_PRIME64_2; |
|
acc = XXH_rotl64(acc, 31); |
|
acc *= XXH_PRIME64_1; |
|
return acc; |
|
} |
|
|
|
static xxh_u64 XXH64_mergeRound(xxh_u64 acc, xxh_u64 val) |
|
{ |
|
val = XXH64_round(0, val); |
|
acc ^= val; |
|
acc = acc * XXH_PRIME64_1 + XXH_PRIME64_4; |
|
return acc; |
|
} |
|
|
|
static xxh_u64 XXH64_avalanche(xxh_u64 h64) |
|
{ |
|
h64 ^= h64 >> 33; |
|
h64 *= XXH_PRIME64_2; |
|
h64 ^= h64 >> 29; |
|
h64 *= XXH_PRIME64_3; |
|
h64 ^= h64 >> 32; |
|
return h64; |
|
} |
|
|
|
|
|
#define XXH_get64bits(p) XXH_readLE64_align(p, align) |
|
|
|
static xxh_u64 |
|
XXH64_finalize(xxh_u64 h64, const xxh_u8* ptr, size_t len, XXH_alignment align) |
|
{ |
|
if (ptr==NULL) XXH_ASSERT(len == 0); |
|
len &= 31; |
|
while (len >= 8) { |
|
xxh_u64 const k1 = XXH64_round(0, XXH_get64bits(ptr)); |
|
ptr += 8; |
|
h64 ^= k1; |
|
h64 = XXH_rotl64(h64,27) * XXH_PRIME64_1 + XXH_PRIME64_4; |
|
len -= 8; |
|
} |
|
if (len >= 4) { |
|
h64 ^= (xxh_u64)(XXH_get32bits(ptr)) * XXH_PRIME64_1; |
|
ptr += 4; |
|
h64 = XXH_rotl64(h64, 23) * XXH_PRIME64_2 + XXH_PRIME64_3; |
|
len -= 4; |
|
} |
|
while (len > 0) { |
|
h64 ^= (*ptr++) * XXH_PRIME64_5; |
|
h64 = XXH_rotl64(h64, 11) * XXH_PRIME64_1; |
|
--len; |
|
} |
|
return XXH64_avalanche(h64); |
|
} |
|
|
|
#ifdef XXH_OLD_NAMES |
|
# define PROCESS1_64 XXH_PROCESS1_64 |
|
# define PROCESS4_64 XXH_PROCESS4_64 |
|
# define PROCESS8_64 XXH_PROCESS8_64 |
|
#else |
|
# undef XXH_PROCESS1_64 |
|
# undef XXH_PROCESS4_64 |
|
# undef XXH_PROCESS8_64 |
|
#endif |
|
|
|
XXH_FORCE_INLINE xxh_u64 |
|
XXH64_endian_align(const xxh_u8* input, size_t len, xxh_u64 seed, XXH_alignment align) |
|
{ |
|
xxh_u64 h64; |
|
if (input==NULL) XXH_ASSERT(len == 0); |
|
|
|
if (len>=32) { |
|
const xxh_u8* const bEnd = input + len; |
|
const xxh_u8* const limit = bEnd - 31; |
|
xxh_u64 v1 = seed + XXH_PRIME64_1 + XXH_PRIME64_2; |
|
xxh_u64 v2 = seed + XXH_PRIME64_2; |
|
xxh_u64 v3 = seed + 0; |
|
xxh_u64 v4 = seed - XXH_PRIME64_1; |
|
|
|
do { |
|
v1 = XXH64_round(v1, XXH_get64bits(input)); input+=8; |
|
v2 = XXH64_round(v2, XXH_get64bits(input)); input+=8; |
|
v3 = XXH64_round(v3, XXH_get64bits(input)); input+=8; |
|
v4 = XXH64_round(v4, XXH_get64bits(input)); input+=8; |
|
} while (input<limit); |
|
|
|
h64 = XXH_rotl64(v1, 1) + XXH_rotl64(v2, 7) + XXH_rotl64(v3, 12) + XXH_rotl64(v4, 18); |
|
h64 = XXH64_mergeRound(h64, v1); |
|
h64 = XXH64_mergeRound(h64, v2); |
|
h64 = XXH64_mergeRound(h64, v3); |
|
h64 = XXH64_mergeRound(h64, v4); |
|
|
|
} else { |
|
h64 = seed + XXH_PRIME64_5; |
|
} |
|
|
|
h64 += (xxh_u64) len; |
|
|
|
return XXH64_finalize(h64, input, len, align); |
|
} |
|
|
|
|
|
/*! @ingroup xxh64_family */ |
|
XXH_PUBLIC_API XXH64_hash_t XXH64 (const void* input, size_t len, XXH64_hash_t seed) |
|
{ |
|
#if 0 |
|
/* Simple version, good for code maintenance, but unfortunately slow for small inputs */ |
|
XXH64_state_t state; |
|
XXH64_reset(&state, seed); |
|
XXH64_update(&state, (const xxh_u8*)input, len); |
|
return XXH64_digest(&state); |
|
#else |
|
if (XXH_FORCE_ALIGN_CHECK) { |
|
if ((((size_t)input) & 7)==0) { /* Input is aligned, let's leverage the speed advantage */ |
|
return XXH64_endian_align((const xxh_u8*)input, len, seed, XXH_aligned); |
|
} } |
|
|
|
return XXH64_endian_align((const xxh_u8*)input, len, seed, XXH_unaligned); |
|
|
|
#endif |
|
} |
|
|
|
/******* Hash Streaming *******/ |
|
|
|
/*! @ingroup xxh64_family*/ |
|
XXH_PUBLIC_API XXH64_state_t* XXH64_createState(void) |
|
{ |
|
return (XXH64_state_t*)XXH_malloc(sizeof(XXH64_state_t)); |
|
} |
|
/*! @ingroup xxh64_family */ |
|
XXH_PUBLIC_API XXH_errorcode XXH64_freeState(XXH64_state_t* statePtr) |
|
{ |
|
XXH_free(statePtr); |
|
return XXH_OK; |
|
} |
|
|
|
/*! @ingroup xxh64_family */ |
|
XXH_PUBLIC_API void XXH64_copyState(XXH64_state_t* dstState, const XXH64_state_t* srcState) |
|
{ |
|
XXH_memcpy(dstState, srcState, sizeof(*dstState)); |
|
} |
|
|
|
/*! @ingroup xxh64_family */ |
|
XXH_PUBLIC_API XXH_errorcode XXH64_reset(XXH64_state_t* statePtr, XXH64_hash_t seed) |
|
{ |
|
XXH64_state_t state; /* use a local state to memcpy() in order to avoid strict-aliasing warnings */ |
|
memset(&state, 0, sizeof(state)); |
|
state.v[0] = seed + XXH_PRIME64_1 + XXH_PRIME64_2; |
|
state.v[1] = seed + XXH_PRIME64_2; |
|
state.v[2] = seed + 0; |
|
state.v[3] = seed - XXH_PRIME64_1; |
|
/* do not write into reserved64, might be removed in a future version */ |
|
XXH_memcpy(statePtr, &state, sizeof(state) - sizeof(state.reserved64)); |
|
return XXH_OK; |
|
} |
|
|
|
/*! @ingroup xxh64_family */ |
|
XXH_PUBLIC_API XXH_errorcode |
|
XXH64_update (XXH64_state_t* state, const void* input, size_t len) |
|
{ |
|
if (input==NULL) { |
|
XXH_ASSERT(len == 0); |
|
return XXH_OK; |
|
} |
|
|
|
{ const xxh_u8* p = (const xxh_u8*)input; |
|
const xxh_u8* const bEnd = p + len; |
|
|
|
state->total_len += len; |
|
|
|
if (state->memsize + len < 32) { /* fill in tmp buffer */ |
|
XXH_memcpy(((xxh_u8*)state->mem64) + state->memsize, input, len); |
|
state->memsize += (xxh_u32)len; |
|
return XXH_OK; |
|
} |
|
|
|
if (state->memsize) { /* tmp buffer is full */ |
|
XXH_memcpy(((xxh_u8*)state->mem64) + state->memsize, input, 32-state->memsize); |
|
state->v[0] = XXH64_round(state->v[0], XXH_readLE64(state->mem64+0)); |
|
state->v[1] = XXH64_round(state->v[1], XXH_readLE64(state->mem64+1)); |
|
state->v[2] = XXH64_round(state->v[2], XXH_readLE64(state->mem64+2)); |
|
state->v[3] = XXH64_round(state->v[3], XXH_readLE64(state->mem64+3)); |
|
p += 32 - state->memsize; |
|
state->memsize = 0; |
|
} |
|
|
|
if (p+32 <= bEnd) { |
|
const xxh_u8* const limit = bEnd - 32; |
|
|
|
do { |
|
state->v[0] = XXH64_round(state->v[0], XXH_readLE64(p)); p+=8; |
|
state->v[1] = XXH64_round(state->v[1], XXH_readLE64(p)); p+=8; |
|
state->v[2] = XXH64_round(state->v[2], XXH_readLE64(p)); p+=8; |
|
state->v[3] = XXH64_round(state->v[3], XXH_readLE64(p)); p+=8; |
|
} while (p<=limit); |
|
|
|
} |
|
|
|
if (p < bEnd) { |
|
XXH_memcpy(state->mem64, p, (size_t)(bEnd-p)); |
|
state->memsize = (unsigned)(bEnd-p); |
|
} |
|
} |
|
|
|
return XXH_OK; |
|
} |
|
|
|
|
|
/*! @ingroup xxh64_family */ |
|
XXH_PUBLIC_API XXH64_hash_t XXH64_digest(const XXH64_state_t* state) |
|
{ |
|
xxh_u64 h64; |
|
|
|
if (state->total_len >= 32) { |
|
h64 = XXH_rotl64(state->v[0], 1) + XXH_rotl64(state->v[1], 7) + XXH_rotl64(state->v[2], 12) + XXH_rotl64(state->v[3], 18); |
|
h64 = XXH64_mergeRound(h64, state->v[0]); |
|
h64 = XXH64_mergeRound(h64, state->v[1]); |
|
h64 = XXH64_mergeRound(h64, state->v[2]); |
|
h64 = XXH64_mergeRound(h64, state->v[3]); |
|
} else { |
|
h64 = state->v[2] /*seed*/ + XXH_PRIME64_5; |
|
} |
|
|
|
h64 += (xxh_u64) state->total_len; |
|
|
|
return XXH64_finalize(h64, (const xxh_u8*)state->mem64, (size_t)state->total_len, XXH_aligned); |
|
} |
|
|
|
|
|
/******* Canonical representation *******/ |
|
|
|
/*! @ingroup xxh64_family */ |
|
XXH_PUBLIC_API void XXH64_canonicalFromHash(XXH64_canonical_t* dst, XXH64_hash_t hash) |
|
{ |
|
XXH_STATIC_ASSERT(sizeof(XXH64_canonical_t) == sizeof(XXH64_hash_t)); |
|
if (XXH_CPU_LITTLE_ENDIAN) hash = XXH_swap64(hash); |
|
XXH_memcpy(dst, &hash, sizeof(*dst)); |
|
} |
|
|