[v2,2/2] random: make rte_rand() thread safe for non-EAL threads

Message ID 20230907152456.20570-3-stephen@networkplumber.org (mailing list archive)
State Rejected, archived
Delegated to: David Marchand
Headers
Series fixes to rte_random for non-EAL threads |

Checks

Context Check Description
ci/checkpatch success coding style OK
ci/loongarch-compilation success Compilation OK
ci/loongarch-unit-testing success Unit Testing PASS
ci/Intel-compilation fail Compilation issues
ci/intel-Testing success Testing PASS
ci/intel-Functional success Functional PASS
ci/github-robot: build fail github build: failed
ci/iol-mellanox-Performance success Performance Testing PASS
ci/iol-broadcom-Performance success Performance Testing PASS
ci/iol-intel-Performance success Performance Testing PASS
ci/iol-intel-Functional success Functional Testing PASS
ci/iol-broadcom-Functional success Functional Testing PASS
ci/iol-compile-amd64-testing fail Testing issues
ci/iol-compile-arm64-testing fail Testing issues
ci/iol-unit-arm64-testing success Testing PASS
ci/iol-unit-amd64-testing fail Testing issues
ci/iol-sample-apps-testing success Testing PASS

Commit Message

Stephen Hemminger Sept. 7, 2023, 3:24 p.m. UTC
  Add missing locking so that if two non-EAL threads call rte_rand()
they will not corrupt the per-thread state.

Fixes: 3f002f069612 ("eal: replace libc-based random generation with LFSR")
Signed-off-by: Stephen Hemminger <stephen@networkplumber.org>
---
 lib/eal/common/rte_random.c | 54 ++++++++++++++++++++++++-------------
 1 file changed, 36 insertions(+), 18 deletions(-)
  

Comments

Stephen Hemminger Sept. 7, 2023, 3:47 p.m. UTC | #1
On Thu,  7 Sep 2023 08:24:56 -0700
Stephen Hemminger <stephen@networkplumber.org> wrote:

>  
> +static __rte_always_inline
> +void __rte_rand_put_state(struct rte_rand_state *state)
> +{
> +	if (state == &rand_states[RTE_MAX_LCORE])
> +		rte_spinlock_unlock(&rte_rand_lock);
> +}

Conditional locking like this make clang lock analyzer unhappy though.
  
David Marchand Sept. 7, 2023, 4:10 p.m. UTC | #2
On Thu, Sep 7, 2023 at 5:48 PM Stephen Hemminger
<stephen@networkplumber.org> wrote:
>
> On Thu,  7 Sep 2023 08:24:56 -0700
> Stephen Hemminger <stephen@networkplumber.org> wrote:
>
> >
> > +static __rte_always_inline
> > +void __rte_rand_put_state(struct rte_rand_state *state)
> > +{
> > +     if (state == &rand_states[RTE_MAX_LCORE])
> > +             rte_spinlock_unlock(&rte_rand_lock);
> > +}
>
> Conditional locking like this make clang lock analyzer unhappy though.

Ugly, but some macro can do the job...

diff --git a/lib/eal/common/rte_random.c b/lib/eal/common/rte_random.c
index 02b6b6b97b..3f2a4830fd 100644
--- a/lib/eal/common/rte_random.c
+++ b/lib/eal/common/rte_random.c
@@ -128,20 +128,22 @@ struct rte_rand_state *__rte_rand_get_state(void)
        idx = rte_lcore_id();

        /* last instance reserved for unregistered non-EAL threads */
-       if (unlikely(idx == LCORE_ID_ANY)) {
+       if (unlikely(idx == LCORE_ID_ANY))
                idx = RTE_MAX_LCORE;
-               rte_spinlock_lock(&rte_rand_lock);
-       }

        return &rand_states[idx];
 }

-static __rte_always_inline
-void __rte_rand_put_state(struct rte_rand_state *state)
-{
-       if (state == &rand_states[RTE_MAX_LCORE])
-               rte_spinlock_unlock(&rte_rand_lock);
-}
+#define PROTECT_NON_EAL_THREADS(...) do { \
+       unsigned int idx = rte_lcore_id(); \
+       if (idx == LCORE_ID_ANY) { \
+               rte_spinlock_lock(&rte_rand_lock); \
+               __VA_ARGS__ \
+               rte_spinlock_unlock(&rte_rand_lock); \
+       } else { \
+               __VA_ARGS__ \
+       } \
+} while (0)

 uint64_t
 rte_rand(void)
@@ -149,9 +151,10 @@ rte_rand(void)
        struct rte_rand_state *state;
        uint64_t res;

+       PROTECT_NON_EAL_THREADS(
        state = __rte_rand_get_state();
        res = __rte_rand_lfsr258(state);
-       __rte_rand_put_state(state);
+       );

        return res;
 }
@@ -168,6 +171,7 @@ rte_rand_max(uint64_t upper_bound)
        if (unlikely(upper_bound < 2))
                return 0;

+       PROTECT_NON_EAL_THREADS(
        state = __rte_rand_get_state();

        ones = rte_popcount64(upper_bound);
@@ -192,7 +196,7 @@ rte_rand_max(uint64_t upper_bound)
                        res = __rte_rand_lfsr258(state) & mask;
                } while (unlikely(res >= upper_bound));
        }
-       __rte_rand_put_state(state);
+       );

        return res;
 }
  
Mattias Rönnblom Sept. 8, 2023, 8:48 p.m. UTC | #3
On 2023-09-07 17:24, Stephen Hemminger wrote:
> Add missing locking so that if two non-EAL threads call rte_rand()
> they will not corrupt the per-thread state.
> 
> Fixes: 3f002f069612 ("eal: replace libc-based random generation with LFSR")

The API documentation clearly states that no MT safety guarantees are 
given for unregistered non-EAL threads. So this patch doesn't fix anything.

rte_rand() is MT safe for *registered* non-EAL threads.

> Signed-off-by: Stephen Hemminger <stephen@networkplumber.org>
> ---
>   lib/eal/common/rte_random.c | 54 ++++++++++++++++++++++++-------------
>   1 file changed, 36 insertions(+), 18 deletions(-)
> 
> diff --git a/lib/eal/common/rte_random.c b/lib/eal/common/rte_random.c
> index 812e5b4757b5..02b6b6b97bc0 100644
> --- a/lib/eal/common/rte_random.c
> +++ b/lib/eal/common/rte_random.c
> @@ -11,6 +11,7 @@
>   #include <rte_branch_prediction.h>
>   #include <rte_cycles.h>
>   #include <rte_lcore.h>
> +#include <rte_spinlock.h>
>   #include <rte_random.h>
>   
>   struct rte_rand_state {
> @@ -21,6 +22,9 @@ struct rte_rand_state {
>   	uint64_t z5;
>   } __rte_cache_aligned;
>   
> +/* Used for thread safety for non EAL threads. */
> +static rte_spinlock_t rte_rand_lock = RTE_SPINLOCK_INITIALIZER;
> +
>   /* One instance each for every lcore id-equipped thread, and one
>    * additional instance to be shared by all others threads (i.e., all
>    * unregistered non-EAL threads).
> @@ -124,20 +128,32 @@ struct rte_rand_state *__rte_rand_get_state(void)
>   	idx = rte_lcore_id();
>   
>   	/* last instance reserved for unregistered non-EAL threads */
> -	if (unlikely(idx == LCORE_ID_ANY))
> +	if (unlikely(idx == LCORE_ID_ANY)) {
>   		idx = RTE_MAX_LCORE;
> +		rte_spinlock_lock(&rte_rand_lock);

Non-EAL threads are very likely to be "regular" threads, which won't 
have a dedicated core all for themselves, and thus may well be preempted 
by the kernel. Such threads should not use spinlocks.

If a lock is to be added to achieve MT safety for parallel calls from 
unregistered non-EAL threads, it should be a regular mutex.

> +	}
>   
>   	return &rand_states[idx];
>   }
>   
> +static __rte_always_inline
> +void __rte_rand_put_state(struct rte_rand_state *state)
> +{
> +	if (state == &rand_states[RTE_MAX_LCORE])
> +		rte_spinlock_unlock(&rte_rand_lock);
> +}
> +
>   uint64_t
>   rte_rand(void)
>   {
>   	struct rte_rand_state *state;
> +	uint64_t res;
>   
>   	state = __rte_rand_get_state();
> +	res = __rte_rand_lfsr258(state);
> +	__rte_rand_put_state(state);
>   
> -	return __rte_rand_lfsr258(state);
> +	return res;
>   }
>   
>   uint64_t
> @@ -159,22 +175,24 @@ rte_rand_max(uint64_t upper_bound)
>   	/* Handle power-of-2 upper_bound as a special case, since it
>   	 * has no bias issues.
>   	 */
> -	if (unlikely(ones == 1))
> -		return __rte_rand_lfsr258(state) & (upper_bound - 1);
> -
> -	/* The approach to avoiding bias is to create a mask that
> -	 * stretches beyond the request value range, and up to the
> -	 * next power-of-2. In case the masked generated random value
> -	 * is equal to or greater than the upper bound, just discard
> -	 * the value and generate a new one.
> -	 */
> -
> -	leading_zeros = rte_clz64(upper_bound);
> -	mask >>= leading_zeros;
> -
> -	do {
> -		res = __rte_rand_lfsr258(state) & mask;
> -	} while (unlikely(res >= upper_bound));
> +	if (unlikely(ones == 1)) {
> +		res = __rte_rand_lfsr258(state) & (upper_bound - 1);
> +	} else {
> +		/* The approach to avoiding bias is to create a mask that
> +		 * stretches beyond the request value range, and up to the
> +		 * next power-of-2. In case the masked generated random value
> +		 * is equal to or greater than the upper bound, just discard
> +		 * the value and generate a new one.
> +		 */
> +
> +		leading_zeros = rte_clz64(upper_bound);
> +		mask >>= leading_zeros;
> +
> +		do {
> +			res = __rte_rand_lfsr258(state) & mask;
> +		} while (unlikely(res >= upper_bound));
> +	}
> +	__rte_rand_put_state(state);
>   
>   	return res;
>   }
  
Stephen Hemminger Sept. 8, 2023, 8:56 p.m. UTC | #4
On Fri, 8 Sep 2023 22:48:54 +0200
Mattias Rönnblom <hofors@lysator.liu.se> wrote:

> On 2023-09-07 17:24, Stephen Hemminger wrote:
> > Add missing locking so that if two non-EAL threads call rte_rand()
> > they will not corrupt the per-thread state.
> > 
> > Fixes: 3f002f069612 ("eal: replace libc-based random generation with LFSR")  
> 
> The API documentation clearly states that no MT safety guarantees are 
> given for unregistered non-EAL threads. So this patch doesn't fix anything.
> 
> rte_rand() is MT safe for *registered* non-EAL threads.

Reading the documentation, it only guarantees safety if registered.
We should add an otherwise clause rather than leaving case as undefined.
  
Mattias Rönnblom Sept. 9, 2023, 7 a.m. UTC | #5
On 2023-09-08 22:56, Stephen Hemminger wrote:
> On Fri, 8 Sep 2023 22:48:54 +0200
> Mattias Rönnblom <hofors@lysator.liu.se> wrote:
> 
>> On 2023-09-07 17:24, Stephen Hemminger wrote:
>>> Add missing locking so that if two non-EAL threads call rte_rand()
>>> they will not corrupt the per-thread state.
>>>
>>> Fixes: 3f002f069612 ("eal: replace libc-based random generation with LFSR")
>>
>> The API documentation clearly states that no MT safety guarantees are
>> given for unregistered non-EAL threads. So this patch doesn't fix anything.
>>
>> rte_rand() is MT safe for *registered* non-EAL threads.
> 
> Reading the documentation, it only guarantees safety if registered.
> We should add an otherwise clause rather than leaving case as undefined.

I agree. It is MT safe in case only single unregistered non-EAL thread 
uses the API (or multiple such threads, provided they are properly 
[externally] serialized).

"If called from EAL threads, registered non-EAL threads or a *single* 
unregistered non-EAL thread, this function is thread-safe. Multiple 
unregistered non-EAL threads may not safely call this function in 
parallel (i.e., must use external serialization)."

A lot of words, unfortunately. Maybe this is better:

"rte_xxx() is MT safe, with the exception it may not be called from 
multiple *unregistered* non-EAL threads in parallel."

Then of course comes the issue that nobody knows what is an registered 
non-EAL thread is. DPDK threading terminology is a big mess, so no wonder.
  
Stephen Hemminger Oct. 2, 2023, 4:10 p.m. UTC | #6
On Thu,  7 Sep 2023 08:24:56 -0700
Stephen Hemminger <stephen@networkplumber.org> wrote:

> Add missing locking so that if two non-EAL threads call rte_rand()
> they will not corrupt the per-thread state.
> 
> Fixes: 3f002f069612 ("eal: replace libc-based random generation with LFSR")
> Signed-off-by: Stephen Hemminger <stephen@networkplumber.org>

Drop this patch in favor of fixing the documentation and leaving
any thread issues up to application to sort out.
  

Patch

diff --git a/lib/eal/common/rte_random.c b/lib/eal/common/rte_random.c
index 812e5b4757b5..02b6b6b97bc0 100644
--- a/lib/eal/common/rte_random.c
+++ b/lib/eal/common/rte_random.c
@@ -11,6 +11,7 @@ 
 #include <rte_branch_prediction.h>
 #include <rte_cycles.h>
 #include <rte_lcore.h>
+#include <rte_spinlock.h>
 #include <rte_random.h>
 
 struct rte_rand_state {
@@ -21,6 +22,9 @@  struct rte_rand_state {
 	uint64_t z5;
 } __rte_cache_aligned;
 
+/* Used for thread safety for non EAL threads. */
+static rte_spinlock_t rte_rand_lock = RTE_SPINLOCK_INITIALIZER;
+
 /* One instance each for every lcore id-equipped thread, and one
  * additional instance to be shared by all others threads (i.e., all
  * unregistered non-EAL threads).
@@ -124,20 +128,32 @@  struct rte_rand_state *__rte_rand_get_state(void)
 	idx = rte_lcore_id();
 
 	/* last instance reserved for unregistered non-EAL threads */
-	if (unlikely(idx == LCORE_ID_ANY))
+	if (unlikely(idx == LCORE_ID_ANY)) {
 		idx = RTE_MAX_LCORE;
+		rte_spinlock_lock(&rte_rand_lock);
+	}
 
 	return &rand_states[idx];
 }
 
+static __rte_always_inline
+void __rte_rand_put_state(struct rte_rand_state *state)
+{
+	if (state == &rand_states[RTE_MAX_LCORE])
+		rte_spinlock_unlock(&rte_rand_lock);
+}
+
 uint64_t
 rte_rand(void)
 {
 	struct rte_rand_state *state;
+	uint64_t res;
 
 	state = __rte_rand_get_state();
+	res = __rte_rand_lfsr258(state);
+	__rte_rand_put_state(state);
 
-	return __rte_rand_lfsr258(state);
+	return res;
 }
 
 uint64_t
@@ -159,22 +175,24 @@  rte_rand_max(uint64_t upper_bound)
 	/* Handle power-of-2 upper_bound as a special case, since it
 	 * has no bias issues.
 	 */
-	if (unlikely(ones == 1))
-		return __rte_rand_lfsr258(state) & (upper_bound - 1);
-
-	/* The approach to avoiding bias is to create a mask that
-	 * stretches beyond the request value range, and up to the
-	 * next power-of-2. In case the masked generated random value
-	 * is equal to or greater than the upper bound, just discard
-	 * the value and generate a new one.
-	 */
-
-	leading_zeros = rte_clz64(upper_bound);
-	mask >>= leading_zeros;
-
-	do {
-		res = __rte_rand_lfsr258(state) & mask;
-	} while (unlikely(res >= upper_bound));
+	if (unlikely(ones == 1)) {
+		res = __rte_rand_lfsr258(state) & (upper_bound - 1);
+	} else {
+		/* The approach to avoiding bias is to create a mask that
+		 * stretches beyond the request value range, and up to the
+		 * next power-of-2. In case the masked generated random value
+		 * is equal to or greater than the upper bound, just discard
+		 * the value and generate a new one.
+		 */
+
+		leading_zeros = rte_clz64(upper_bound);
+		mask >>= leading_zeros;
+
+		do {
+			res = __rte_rand_lfsr258(state) & mask;
+		} while (unlikely(res >= upper_bound));
+	}
+	__rte_rand_put_state(state);
 
 	return res;
 }