Source-Changes-HG archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

[src/trunk]: src/sys/kern Use a temporary pool to consolidate entropy atomica...



details:   https://anonhg.NetBSD.org/src/rev/ad85b4a9149d
branches:  trunk
changeset: 971934:ad85b4a9149d
user:      riastradh <riastradh%NetBSD.org@localhost>
date:      Sun May 10 00:08:12 2020 +0000

description:
Use a temporary pool to consolidate entropy atomically.

There was a low-probability race with the entropy consolidation
logic: calls to entropy_extract at the same time as consolidation is
happening might witness partial contributions from the CPUs when
needed=256, say 64 bits at a time.

To avoid this, feed everything from the per-CPU pools into a
temporary pool, and then feed the temporary pool into the global pool
under the lock at the same time as we update needed.

diffstat:

 sys/kern/kern_entropy.c |  25 +++++++++++++++++--------
 1 files changed, 17 insertions(+), 8 deletions(-)

diffs (77 lines):

diff -r 27af61b3dbc7 -r ad85b4a9149d sys/kern/kern_entropy.c
--- a/sys/kern/kern_entropy.c   Sat May 09 22:00:48 2020 +0000
+++ b/sys/kern/kern_entropy.c   Sun May 10 00:08:12 2020 +0000
@@ -1,4 +1,4 @@
-/*     $NetBSD: kern_entropy.c,v 1.18 2020/05/09 06:12:32 riastradh Exp $      */
+/*     $NetBSD: kern_entropy.c,v 1.19 2020/05/10 00:08:12 riastradh Exp $      */
 
 /*-
  * Copyright (c) 2019 The NetBSD Foundation, Inc.
@@ -75,7 +75,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: kern_entropy.c,v 1.18 2020/05/09 06:12:32 riastradh Exp $");
+__KERNEL_RCSID(0, "$NetBSD: kern_entropy.c,v 1.19 2020/05/10 00:08:12 riastradh Exp $");
 
 #include <sys/param.h>
 #include <sys/types.h>
@@ -984,11 +984,14 @@
 {
        static const struct timeval interval = {.tv_sec = 60, .tv_usec = 0};
        static struct timeval lasttime; /* serialized by E->lock */
+       struct entpool pool;
+       uint8_t buf[ENTPOOL_CAPACITY];
        unsigned diff;
        uint64_t ticket;
 
-       /* Gather entropy on all CPUs.  */
-       ticket = xc_broadcast(0, &entropy_consolidate_xc, NULL, NULL);
+       /* Gather entropy on all CPUs into a temporary pool.  */
+       memset(&pool, 0, sizeof pool);
+       ticket = xc_broadcast(0, &entropy_consolidate_xc, &pool, NULL);
        xc_wait(ticket);
 
        /* Acquire the lock to notify waiters.  */
@@ -1000,6 +1003,11 @@
        /* Note when we last consolidated, i.e. now.  */
        E->timestamp = time_uptime;
 
+       /* Mix what we gathered into the global pool.  */
+       entpool_extract(&pool, buf, sizeof buf);
+       entpool_enter(&E->pool, buf, sizeof buf);
+       explicit_memset(&pool, 0, sizeof pool);
+
        /* Count the entropy that was gathered.  */
        diff = MIN(E->needed, E->pending);
        atomic_store_relaxed(&E->needed, E->needed - diff);
@@ -1024,8 +1032,9 @@
  *     into the global pool.
  */
 static void
-entropy_consolidate_xc(void *arg1 __unused, void *arg2 __unused)
+entropy_consolidate_xc(void *vpool, void *arg2 __unused)
 {
+       struct entpool *pool = vpool;
        struct entropy_cpu *ec;
        uint8_t buf[ENTPOOL_CAPACITY];
        uint32_t extra[7];
@@ -1063,15 +1072,15 @@
 
        /*
         * Copy over statistics, and enter the per-CPU extract and the
-        * extra timing into the global pool, under the global lock.
+        * extra timing into the temporary pool, under the global lock.
         */
        mutex_enter(&E->lock);
        extra[i++] = entropy_timer();
-       entpool_enter(&E->pool, buf, sizeof buf);
+       entpool_enter(pool, buf, sizeof buf);
        explicit_memset(buf, 0, sizeof buf);
        extra[i++] = entropy_timer();
        KASSERT(i == __arraycount(extra));
-       entpool_enter(&E->pool, extra, sizeof extra);
+       entpool_enter(pool, extra, sizeof extra);
        explicit_memset(extra, 0, sizeof extra);
        mutex_exit(&E->lock);
 }



Home | Main Index | Thread Index | Old Index