[v2,1/5] hash: add new toeplitz hash implementation
Checks
Commit Message
This patch add a new Toeplitz hash implementation using
Galios Fields New Instructions (GFNI).
Signed-off-by: Vladimir Medvedkin <vladimir.medvedkin@intel.com>
---
doc/api/doxy-api-index.md | 1 +
lib/hash/meson.build | 1 +
lib/hash/rte_thash.c | 28 ++++++
lib/hash/rte_thash.h | 24 +++++
lib/hash/rte_thash_gfni.h | 232 ++++++++++++++++++++++++++++++++++++++++++++++
lib/hash/version.map | 2 +
6 files changed, 288 insertions(+)
create mode 100644 lib/hash/rte_thash_gfni.h
Comments
On Fri, 15 Oct 2021 10:30:02 +0100
Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote:
> + m[i * 8 + j] = (rss_key[i] << j)|
> + (uint8_t)((uint16_t)(rss_key[i + 1]) >>
> + (8 - j));
> + }
This ends up being harder than necessary to read. Maybe split into
multiple statements and/or use temporary variable.
> +RTE_INIT(rte_thash_gfni_init)
> +{
> + rte_thash_gfni_supported = 0;
Not necessary in C globals are initialized to zero by default.
By removing that the constructor can be totally behind #ifdef
> +__rte_internal
> +static inline __m512i
> +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple,
> + const uint8_t *secondary_tuple, int len)
> +{
> + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4,
> + 6, 5, 4, 3, 6, 5, 4, 3,
> + 5, 4, 3, 2, 5, 4, 3, 2,
> + 4, 3, 2, 1, 4, 3, 2, 1,
> + 3, 2, 1, 0, 3, 2, 1, 0,
> + 2, 1, 0, -1, 2, 1, 0, -1,
> + 1, 0, -1, -2, 1, 0, -1, -2,
> + 0, -1, -2, -3, 0, -1, -2, -3);
NAK
Please don't put the implementation in an inline. This makes it harder
to support (API/ABI) and blocks other architectures from implementing
same thing with different instructions.
> On Fri, 15 Oct 2021 10:30:02 +0100
> Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote:
>
> > + m[i * 8 + j] = (rss_key[i] << j)|
> > + (uint8_t)((uint16_t)(rss_key[i + 1]) >>
> > + (8 - j));
> > + }
>
> This ends up being harder than necessary to read. Maybe split into
> multiple statements and/or use temporary variable.
>
> > +RTE_INIT(rte_thash_gfni_init)
> > +{
> > + rte_thash_gfni_supported = 0;
>
> Not necessary in C globals are initialized to zero by default.
>
> By removing that the constructor can be totally behind #ifdef
>
> > +__rte_internal
> > +static inline __m512i
> > +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple,
> > + const uint8_t *secondary_tuple, int len)
> > +{
> > + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4,
> > + 6, 5, 4, 3, 6, 5, 4, 3,
> > + 5, 4, 3, 2, 5, 4, 3, 2,
> > + 4, 3, 2, 1, 4, 3, 2, 1,
> > + 3, 2, 1, 0, 3, 2, 1, 0,
> > + 2, 1, 0, -1, 2, 1, 0, -1,
> > + 1, 0, -1, -2, 1, 0, -1, -2,
> > + 0, -1, -2, -3, 0, -1, -2, -3);
>
> NAK
>
> Please don't put the implementation in an inline. This makes it harder
> to support (API/ABI) and blocks other architectures from implementing
> same thing with different instructions.
I don't really understand your reasoning here.
rte_thash_gfni.h is an arch-specific header, which provides
arch-specific optimizations for RSS hash calculation
(Vladimir pls correct me if I am wrong here).
We do have dozens of inline functions that do use arch-specific instructions (both x86 and arm)
for different purposes:
sync primitives, memory-ordering, cache manipulations, LPM lookup, TSX, power-saving, etc.
That's a usual trade-off taken for performance reasons, when extra function call
costs too much comparing to the operation itself.
Why it suddenly became a problem for that particular case and how exactly it blocks other architectures?
Also I don't understand how it makes things harder in terms of API/ABI stability.
As I can see this patch doesn't introduce any public structs/unions.
All functions take as arguments just raw data buffers and length.
To summarize - in general, I don't see any good reason why this patch shouldn't be allowed.
Konstantin
Hi Stephen,
Thanks for reviewing
On 15/10/2021 18:58, Stephen Hemminger wrote:
> On Fri, 15 Oct 2021 10:30:02 +0100
> Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote:
>
>> + m[i * 8 + j] = (rss_key[i] << j)|
>> + (uint8_t)((uint16_t)(rss_key[i + 1]) >>
>> + (8 - j));
>> + }
>
> This ends up being harder than necessary to read. Maybe split into
> multiple statements and/or use temporary variable.
>
>> +RTE_INIT(rte_thash_gfni_init)
>> +{
>> + rte_thash_gfni_supported = 0;
>
> Not necessary in C globals are initialized to zero by default.
>
> By removing that the constructor can be totally behind #ifdef
>
>> +__rte_internal
>> +static inline __m512i
>> +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple,
>> + const uint8_t *secondary_tuple, int len)
>> +{
>> + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4,
>> + 6, 5, 4, 3, 6, 5, 4, 3,
>> + 5, 4, 3, 2, 5, 4, 3, 2,
>> + 4, 3, 2, 1, 4, 3, 2, 1,
>> + 3, 2, 1, 0, 3, 2, 1, 0,
>> + 2, 1, 0, -1, 2, 1, 0, -1,
>> + 1, 0, -1, -2, 1, 0, -1, -2,
>> + 0, -1, -2, -3, 0, -1, -2, -3);
>
> NAK
>
> Please don't put the implementation in an inline. This makes it harder
> to support (API/ABI) and blocks other architectures from implementing
> same thing with different instructions.
>
By making this function not inline, its performance drops by about 2
times. Compiler optimization (at least with respect to the len argument)
helps a lot in the implementation.
On Mon, 18 Oct 2021 10:40:00 +0000
"Ananyev, Konstantin" <konstantin.ananyev@intel.com> wrote:
> > On Fri, 15 Oct 2021 10:30:02 +0100
> > Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote:
> >
> > > + m[i * 8 + j] = (rss_key[i] << j)|
> > > + (uint8_t)((uint16_t)(rss_key[i + 1]) >>
> > > + (8 - j));
> > > + }
> >
> > This ends up being harder than necessary to read. Maybe split into
> > multiple statements and/or use temporary variable.
> >
> > > +RTE_INIT(rte_thash_gfni_init)
> > > +{
> > > + rte_thash_gfni_supported = 0;
> >
> > Not necessary in C globals are initialized to zero by default.
> >
> > By removing that the constructor can be totally behind #ifdef
> >
> > > +__rte_internal
> > > +static inline __m512i
> > > +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple,
> > > + const uint8_t *secondary_tuple, int len)
> > > +{
> > > + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4,
> > > + 6, 5, 4, 3, 6, 5, 4, 3,
> > > + 5, 4, 3, 2, 5, 4, 3, 2,
> > > + 4, 3, 2, 1, 4, 3, 2, 1,
> > > + 3, 2, 1, 0, 3, 2, 1, 0,
> > > + 2, 1, 0, -1, 2, 1, 0, -1,
> > > + 1, 0, -1, -2, 1, 0, -1, -2,
> > > + 0, -1, -2, -3, 0, -1, -2, -3);
> >
> > NAK
> >
> > Please don't put the implementation in an inline. This makes it harder
> > to support (API/ABI) and blocks other architectures from implementing
> > same thing with different instructions.
>
> I don't really understand your reasoning here.
> rte_thash_gfni.h is an arch-specific header, which provides
> arch-specific optimizations for RSS hash calculation
> (Vladimir pls correct me if I am wrong here).
Ok, but rte_thash_gfni.h is included on all architectures.
> We do have dozens of inline functions that do use arch-specific instructions (both x86 and arm)
> for different purposes:
> sync primitives, memory-ordering, cache manipulations, LPM lookup, TSX, power-saving, etc.
> That's a usual trade-off taken for performance reasons, when extra function call
> costs too much comparing to the operation itself.
> Why it suddenly became a problem for that particular case and how exactly it blocks other architectures?
> Also I don't understand how it makes things harder in terms of API/ABI stability.
> As I can see this patch doesn't introduce any public structs/unions.
> All functions take as arguments just raw data buffers and length.
> To summarize - in general, I don't see any good reason why this patch shouldn't be allowed.
> Konstantin
The comments about rte_thash_gfni_supported initialization still apply.
Why not:
#ifdef __GFNI__
RTE_INIT(rte_thash_gfni_init)
{
if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_GFNI))
rte_thash_gfni_supported = 1;
}
#endif
Hi Stephen,
On 19/10/2021 03:15, Stephen Hemminger wrote:
> On Mon, 18 Oct 2021 10:40:00 +0000
> "Ananyev, Konstantin" <konstantin.ananyev@intel.com> wrote:
>
>>> On Fri, 15 Oct 2021 10:30:02 +0100
>>> Vladimir Medvedkin <vladimir.medvedkin@intel.com> wrote:
>>>
>>>> + m[i * 8 + j] = (rss_key[i] << j)|
>>>> + (uint8_t)((uint16_t)(rss_key[i + 1]) >>
>>>> + (8 - j));
>>>> + }
>>>
>>> This ends up being harder than necessary to read. Maybe split into
>>> multiple statements and/or use temporary variable.
>>>
>>>> +RTE_INIT(rte_thash_gfni_init)
>>>> +{
>>>> + rte_thash_gfni_supported = 0;
>>>
>>> Not necessary in C globals are initialized to zero by default.
>>>
>>> By removing that the constructor can be totally behind #ifdef
>>>
>>>> +__rte_internal
>>>> +static inline __m512i
>>>> +__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple,
>>>> + const uint8_t *secondary_tuple, int len)
>>>> +{
>>>> + __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4,
>>>> + 6, 5, 4, 3, 6, 5, 4, 3,
>>>> + 5, 4, 3, 2, 5, 4, 3, 2,
>>>> + 4, 3, 2, 1, 4, 3, 2, 1,
>>>> + 3, 2, 1, 0, 3, 2, 1, 0,
>>>> + 2, 1, 0, -1, 2, 1, 0, -1,
>>>> + 1, 0, -1, -2, 1, 0, -1, -2,
>>>> + 0, -1, -2, -3, 0, -1, -2, -3);
>>>
>>> NAK
>>>
>>> Please don't put the implementation in an inline. This makes it harder
>>> to support (API/ABI) and blocks other architectures from implementing
>>> same thing with different instructions.
>>
>> I don't really understand your reasoning here.
>> rte_thash_gfni.h is an arch-specific header, which provides
>> arch-specific optimizations for RSS hash calculation
>> (Vladimir pls correct me if I am wrong here).
>
> Ok, but rte_thash_gfni.h is included on all architectures.
>
Ok, I'll rework the patch to move x86 + avx512 related things into x86
arch specific header. Would that suit?
>> We do have dozens of inline functions that do use arch-specific instructions (both x86 and arm)
>> for different purposes:
>> sync primitives, memory-ordering, cache manipulations, LPM lookup, TSX, power-saving, etc.
>> That's a usual trade-off taken for performance reasons, when extra function call
>> costs too much comparing to the operation itself.
>> Why it suddenly became a problem for that particular case and how exactly it blocks other architectures?
>> Also I don't understand how it makes things harder in terms of API/ABI stability.
>> As I can see this patch doesn't introduce any public structs/unions.
>> All functions take as arguments just raw data buffers and length.
>> To summarize - in general, I don't see any good reason why this patch shouldn't be allowed.
>> Konstantin
>
> The comments about rte_thash_gfni_supported initialization still apply.
> Why not:
>
> #ifdef __GFNI__
> RTE_INIT(rte_thash_gfni_init)
> {
> if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_GFNI))
> rte_thash_gfni_supported = 1;
> }
> #endif
>
Agree, I'll reflect this changes in v3.
@@ -139,6 +139,7 @@ The public API headers are grouped by topics:
[hash] (@ref rte_hash.h),
[jhash] (@ref rte_jhash.h),
[thash] (@ref rte_thash.h),
+ [thash_gfni] (@ref rte_thash_gfni.h),
[FBK hash] (@ref rte_fbk_hash.h),
[CRC hash] (@ref rte_hash_crc.h)
@@ -7,6 +7,7 @@ headers = files(
'rte_hash.h',
'rte_jhash.h',
'rte_thash.h',
+ 'rte_thash_gfni.h',
)
indirect_headers += files('rte_crc_arm64.h')
@@ -13,6 +13,7 @@
#include <rte_eal_memconfig.h>
#include <rte_log.h>
#include <rte_malloc.h>
+#include <rte_thash_gfni.h>
#define THASH_NAME_LEN 64
#define TOEPLITZ_HASH_LEN 32
@@ -90,6 +91,24 @@ struct rte_thash_ctx {
uint8_t hash_key[0];
};
+/** Flag indicating GFNI support */
+uint8_t rte_thash_gfni_supported;
+
+void
+rte_thash_complete_matrix(uint64_t *matrixes, const uint8_t *rss_key, int size)
+{
+ int i, j;
+ uint8_t *m = (uint8_t *)matrixes;
+
+ for (i = 0; i < size; i++) {
+ for (j = 0; j < 8; j++) {
+ m[i * 8 + j] = (rss_key[i] << j)|
+ (uint8_t)((uint16_t)(rss_key[i + 1]) >>
+ (8 - j));
+ }
+ }
+}
+
static inline uint32_t
get_bit_lfsr(struct thash_lfsr *lfsr)
{
@@ -761,3 +780,12 @@ rte_thash_adjust_tuple(struct rte_thash_ctx *ctx,
return ret;
}
+
+RTE_INIT(rte_thash_gfni_init)
+{
+ rte_thash_gfni_supported = 0;
+#ifdef __GFNI__
+ if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_GFNI))
+ rte_thash_gfni_supported = 1;
+#endif
+}
@@ -28,6 +28,7 @@ extern "C" {
#include <rte_config.h>
#include <rte_ip.h>
#include <rte_common.h>
+#include <rte_thash_gfni.h>
#if defined(RTE_ARCH_X86) || defined(__ARM_NEON)
#include <rte_vect.h>
@@ -113,6 +114,9 @@ union rte_thash_tuple {
};
#endif
+/** Flag indicating GFNI support */
+extern uint8_t rte_thash_gfni_supported;
+
/**
* Prepare special converted key to use with rte_softrss_be()
* @param orig
@@ -223,6 +227,26 @@ rte_softrss_be(uint32_t *input_tuple, uint32_t input_len,
return ret;
}
+/**
+ * Converts Toeplitz hash key (RSS key) into matrixes required
+ * for GFNI implementation
+ *
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice.
+ *
+ * @param matrixes
+ * pointer to the memory where matrices will be written.
+ * Note: the size of this memory must be equal to size * 8
+ * @param rss_key
+ * pointer to the Toeplitz hash key
+ * @param size
+ * Size of the rss_key in bytes.
+ */
+__rte_experimental
+void
+rte_thash_complete_matrix(uint64_t *matrixes, const uint8_t *rss_key,
+ int size);
+
/** @internal Logarithm of minimum size of the RSS ReTa */
#define RTE_THASH_RETA_SZ_MIN 2U
/** @internal Logarithm of maximum size of the RSS ReTa */
new file mode 100644
@@ -0,0 +1,232 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2021 Intel Corporation
+ */
+
+#ifndef _RTE_THASH_GFNI_H_
+#define _RTE_THASH_GFNI_H_
+
+/**
+ * @file
+ *
+ * Optimized Toeplitz hash functions implementation
+ * using Galois Fields New Instructions.
+ */
+
+#include <rte_vect.h>
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
+#ifdef __GFNI__
+
+#define RTE_THASH_FIRST_ITER_MSK 0x0f0f0f0f0f0e0c08
+#define RTE_THASH_PERM_MSK 0x0f0f0f0f0f0f0f0f
+#define RTE_THASH_FIRST_ITER_MSK_2 0xf0f0f0f0f0e0c080
+#define RTE_THASH_PERM_MSK_2 0xf0f0f0f0f0f0f0f0
+#define RTE_THASH_REWIND_MSK 0x0000000000113377
+
+__rte_internal
+static inline void
+__rte_thash_xor_reduce(__m512i xor_acc, uint32_t *val_1, uint32_t *val_2)
+{
+ __m256i tmp_256_1, tmp_256_2;
+ __m128i tmp128_1, tmp128_2;
+ uint64_t tmp_1, tmp_2;
+
+ tmp_256_1 = _mm512_castsi512_si256(xor_acc);
+ tmp_256_2 = _mm512_extracti32x8_epi32(xor_acc, 1);
+ tmp_256_1 = _mm256_xor_si256(tmp_256_1, tmp_256_2);
+
+ tmp128_1 = _mm256_castsi256_si128(tmp_256_1);
+ tmp128_2 = _mm256_extracti32x4_epi32(tmp_256_1, 1);
+ tmp128_1 = _mm_xor_si128(tmp128_1, tmp128_2);
+
+ tmp_1 = _mm_extract_epi64(tmp128_1, 0);
+ tmp_2 = _mm_extract_epi64(tmp128_1, 1);
+ tmp_1 ^= tmp_2;
+
+ *val_1 = (uint32_t)tmp_1;
+ *val_2 = (uint32_t)(tmp_1 >> 32);
+}
+
+__rte_internal
+static inline __m512i
+__rte_thash_gfni(const uint64_t *mtrx, const uint8_t *tuple,
+ const uint8_t *secondary_tuple, int len)
+{
+ __m512i permute_idx = _mm512_set_epi8(7, 6, 5, 4, 7, 6, 5, 4,
+ 6, 5, 4, 3, 6, 5, 4, 3,
+ 5, 4, 3, 2, 5, 4, 3, 2,
+ 4, 3, 2, 1, 4, 3, 2, 1,
+ 3, 2, 1, 0, 3, 2, 1, 0,
+ 2, 1, 0, -1, 2, 1, 0, -1,
+ 1, 0, -1, -2, 1, 0, -1, -2,
+ 0, -1, -2, -3, 0, -1, -2, -3);
+
+ const __m512i rewind_idx = _mm512_set_epi8(0, 0, 0, 0, 0, 0, 0, 0,
+ 0, 0, 0, 0, 0, 0, 0, 0,
+ 0, 0, 0, 0, 0, 0, 0, 0,
+ 0, 0, 0, 0, 0, 0, 0, 0,
+ 0, 0, 0, 0, 0, 0, 0, 0,
+ 0, 0, 0, 59, 0, 0, 0, 59,
+ 0, 0, 59, 58, 0, 0, 59, 58,
+ 0, 59, 58, 57, 0, 59, 58, 57);
+ const __mmask64 rewind_mask = RTE_THASH_REWIND_MSK;
+ const __m512i shift_8 = _mm512_set1_epi8(8);
+ __m512i xor_acc = _mm512_setzero_si512();
+ __m512i perm_bytes = _mm512_setzero_si512();
+ __m512i vals, matrixes, tuple_bytes, tuple_bytes_2;
+ __mmask64 load_mask, permute_mask, permute_mask_2;
+ int chunk_len = 0, i = 0;
+ uint8_t mtrx_msk;
+ const int prepend = 3;
+
+ for (; len > 0; len -= 64, tuple += 64) {
+ if (i == 8)
+ perm_bytes = _mm512_maskz_permutexvar_epi8(rewind_mask,
+ rewind_idx, perm_bytes);
+
+ permute_mask = RTE_THASH_FIRST_ITER_MSK;
+ load_mask = (len >= 64) ? UINT64_MAX : ((1ULL << len) - 1);
+ tuple_bytes = _mm512_maskz_loadu_epi8(load_mask, tuple);
+ if (secondary_tuple) {
+ permute_mask_2 = RTE_THASH_FIRST_ITER_MSK_2;
+ tuple_bytes_2 = _mm512_maskz_loadu_epi8(load_mask,
+ secondary_tuple);
+ }
+
+ chunk_len = __builtin_popcountll(load_mask);
+ for (i = 0; i < ((chunk_len + prepend) / 8); i++, mtrx += 8) {
+ perm_bytes = _mm512_mask_permutexvar_epi8(perm_bytes,
+ permute_mask, permute_idx, tuple_bytes);
+
+ if (secondary_tuple)
+ perm_bytes =
+ _mm512_mask_permutexvar_epi8(perm_bytes,
+ permute_mask_2, permute_idx,
+ tuple_bytes_2);
+
+ matrixes = _mm512_maskz_loadu_epi64(UINT8_MAX, mtrx);
+ vals = _mm512_gf2p8affine_epi64_epi8(perm_bytes,
+ matrixes, 0);
+
+ xor_acc = _mm512_xor_si512(xor_acc, vals);
+ permute_idx = _mm512_add_epi8(permute_idx, shift_8);
+ permute_mask = RTE_THASH_PERM_MSK;
+ if (secondary_tuple)
+ permute_mask_2 = RTE_THASH_PERM_MSK_2;
+ }
+ }
+
+ int rest_len = (chunk_len + prepend) % 8;
+ if (rest_len != 0) {
+ mtrx_msk = (1 << (rest_len % 8)) - 1;
+ matrixes = _mm512_maskz_loadu_epi64(mtrx_msk, mtrx);
+ if (i == 8) {
+ perm_bytes = _mm512_maskz_permutexvar_epi8(rewind_mask,
+ rewind_idx, perm_bytes);
+ } else {
+ perm_bytes = _mm512_mask_permutexvar_epi8(perm_bytes,
+ permute_mask, permute_idx, tuple_bytes);
+
+ if (secondary_tuple)
+ perm_bytes =
+ _mm512_mask_permutexvar_epi8(
+ perm_bytes, permute_mask_2,
+ permute_idx, tuple_bytes_2);
+ }
+
+ vals = _mm512_gf2p8affine_epi64_epi8(perm_bytes, matrixes, 0);
+ xor_acc = _mm512_xor_si512(xor_acc, vals);
+ }
+
+ return xor_acc;
+}
+
+/**
+ * Calculate Toeplitz hash.
+ *
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice.
+ *
+ * @param m
+ * Pointer to the matrices generated from the corresponding
+ * RSS hash key using rte_thash_complete_matrix().
+ * @param tuple
+ * Pointer to the data to be hashed. Data must be in network byte order.
+ * @param len
+ * Length of the data to be hashed.
+ * @return
+ * Calculated Toeplitz hash value.
+ */
+__rte_experimental
+static inline uint32_t
+rte_thash_gfni(const uint64_t *m, const uint8_t *tuple, int len)
+{
+ uint32_t val, val_zero;
+
+ __m512i xor_acc = __rte_thash_gfni(m, tuple, NULL, len);
+ __rte_thash_xor_reduce(xor_acc, &val, &val_zero);
+
+ return val;
+}
+
+/**
+ * Calculate Toeplitz hash for two independent data buffers.
+ *
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice.
+ *
+ * @param m
+ * Pointer to the matrices generated from the corresponding
+ * RSS hash key using rte_thash_complete_matrix().
+ * @param tuple_1
+ * Pointer to the data to be hashed. Data must be in network byte order.
+ * @param tuple_2
+ * Pointer to the data to be hashed. Data must be in network byte order.
+ * @param len
+ * Length of the largest data buffer to be hashed.
+ * @param val_1
+ * Pointer to uint32_t where to put calculated Toeplitz hash value for
+ * the first tuple.
+ * @param val_2
+ * Pointer to uint32_t where to put calculated Toeplitz hash value for
+ * the second tuple.
+ */
+__rte_experimental
+static inline void
+rte_thash_gfni_x2(const uint64_t *mtrx, const uint8_t *tuple_1,
+ const uint8_t *tuple_2, int len,
+ uint32_t *val_1, uint32_t *val_2)
+{
+ __m512i xor_acc = __rte_thash_gfni(mtrx, tuple_1, tuple_2, len);
+ __rte_thash_xor_reduce(xor_acc, val_1, val_2);
+}
+
+#else /* __GFNI__ */
+
+static inline uint32_t
+rte_thash_gfni(const uint64_t *mtrx __rte_unused,
+ const uint8_t *key __rte_unused, int len __rte_unused)
+{
+ return 0;
+}
+
+static inline void
+rte_thash_gfni_x2(const uint64_t *mtrx __rte_unused,
+ const uint8_t *tuple_1 __rte_unused,
+ const uint8_t *tuple_2 __rte_unused, int len __rte_unused,
+ uint32_t *val_1 __rte_unused, uint32_t *val_2 __rte_unused)
+{
+ *val_1 = 0;
+ *val_2 = 0;
+}
+
+#endif
+
+#ifdef __cplusplus
+}
+#endif
+
+#endif /* _RTE_THASH_GFNI_H_ */
@@ -39,10 +39,12 @@ EXPERIMENTAL {
rte_hash_rcu_qsbr_add;
rte_thash_add_helper;
rte_thash_adjust_tuple;
+ rte_thash_complete_matrix;
rte_thash_find_existing;
rte_thash_free_ctx;
rte_thash_get_complement;
rte_thash_get_helper;
rte_thash_get_key;
+ rte_thash_gfni_supported;
rte_thash_init_ctx;
};