Source-Changes-HG archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

[src/trunk]: src/sys/arch/x86/x86 Align the TLB packet precisely on the stack...



details:   https://anonhg.NetBSD.org/src/rev/f7227f4464e0
branches:  trunk
changeset: 847277:f7227f4464e0
user:      ad <ad%NetBSD.org@localhost>
date:      Mon Dec 16 19:17:25 2019 +0000

description:
Align the TLB packet precisely on the stack, and do 7 INVLPG since it's
what fits in a single line.

diffstat:

 sys/arch/x86/x86/x86_tlb.c |  41 +++++++++++++++++------------------------
 1 files changed, 17 insertions(+), 24 deletions(-)

diffs (113 lines):

diff -r 42e11c5e5eb3 -r f7227f4464e0 sys/arch/x86/x86/x86_tlb.c
--- a/sys/arch/x86/x86/x86_tlb.c        Mon Dec 16 18:50:44 2019 +0000
+++ b/sys/arch/x86/x86/x86_tlb.c        Mon Dec 16 19:17:25 2019 +0000
@@ -1,4 +1,4 @@
-/*     $NetBSD: x86_tlb.c,v 1.12 2019/12/02 20:59:56 pgoyette Exp $    */
+/*     $NetBSD: x86_tlb.c,v 1.13 2019/12/16 19:17:25 ad Exp $  */
 
 /*-
  * Copyright (c) 2008-2019 The NetBSD Foundation, Inc.
@@ -40,7 +40,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: x86_tlb.c,v 1.12 2019/12/02 20:59:56 pgoyette Exp $");
+__KERNEL_RCSID(0, "$NetBSD: x86_tlb.c,v 1.13 2019/12/16 19:17:25 ad Exp $");
 
 #include <sys/param.h>
 #include <sys/kernel.h>
@@ -66,10 +66,10 @@
  * until the request is completed.  This keeps the cache line in the shared
  * state, and bus traffic to a minimum.
  *
- * On i386 the packet is 28 bytes in size.  On amd64 it's 52 bytes.
+ * On i386 the packet is 32 bytes in size.  On amd64 it's 60 bytes.
  */
 typedef struct {
-       uintptr_t               tp_va[6];
+       uintptr_t               tp_va[7];
        uint8_t                 tp_count;
        uint8_t                 tp_userpmap;
        uint8_t                 tp_global;
@@ -77,23 +77,14 @@
 } pmap_tlb_packet_t;
 
 /*
- * Padded packet stored on the initiator's stack.
- */
-typedef struct {
-       uint8_t                 ts_pad1[COHERENCY_UNIT];
-       pmap_tlb_packet_t       ts_tp;
-       uint8_t                 ts_pad2[COHERENCY_UNIT];
-} pmap_tlb_stackbuf_t;
-
-/*
  * No more than N separate invlpg.
  *
- * Statistically, a value of six is big enough to cover the requested number
+ * Statistically, a value of 7 is big enough to cover the requested number
  * of pages in ~ 95% of the TLB shootdowns we are getting. We therefore rarely
  * reach the limit, and increasing it can actually reduce the performance due
  * to the high cost of invlpg.
  */
-#define        TP_MAXVA                6       /* for individual mappings */
+#define        TP_MAXVA                7       /* for individual mappings */
 #define        TP_ALLVA                255     /* special: shoot all mappings */
 
 /*
@@ -355,8 +346,8 @@
 void
 pmap_tlb_shootnow(void)
 {
-       volatile pmap_tlb_packet_t *tp;
-       volatile pmap_tlb_stackbuf_t ts;
+       volatile pmap_tlb_packet_t *tp, *ts;
+       volatile uint8_t stackbuf[128];
        struct cpu_info *ci;
        kcpuset_t *target;
        u_int local, rcpucount;
@@ -405,11 +396,13 @@
         * against an interrupt on the current CPU trying the same.
         */
        KASSERT(rcpucount < ncpu);
-       ts.ts_tp = *tp;
-       KASSERT(!ts.ts_tp.tp_done);
+       KASSERT(sizeof(*ts) <= (sizeof(stackbuf) / 2));
+       ts = (void *)roundup2((uintptr_t)stackbuf, (sizeof(stackbuf) / 2));
+       *ts = *tp;
+       KASSERT(!ts->tp_done);
        while (atomic_cas_ptr(&pmap_tlb_packet, NULL,
-           __UNVOLATILE(&ts.ts_tp)) != NULL) {
-               KASSERT(pmap_tlb_packet != &ts.ts_tp);
+           __UNVOLATILE(ts)) != NULL) {
+               KASSERT(pmap_tlb_packet != ts);
                /*
                 * Don't bother with exponentional backoff, as the pointer
                 * is in a dedicated cache line and only updated twice per
@@ -439,7 +432,7 @@
         */
        pmap_tlb_pendcount = rcpucount;
        pmap_tlb_evcnt.ev_count++;
-       pmap_tlb_processpacket(&ts.ts_tp, target);
+       pmap_tlb_processpacket(ts, target);
 
        /*
         * Clear out the local CPU's buffer for the next user.  Once done,
@@ -461,7 +454,7 @@
         * perform local shootdown if needed, using our copy of the packet.
         */
        if (local) {
-               pmap_tlb_invalidate(&ts.ts_tp);
+               pmap_tlb_invalidate(ts);
        }
 
        /*
@@ -470,7 +463,7 @@
         * CPU out will update it and only we are reading it).  No memory
         * barrier required due to prior stores - yay x86.
         */
-       while (!ts.ts_tp.tp_done) {
+       while (!ts->tp_done) {
                x86_pause();
        }
 }



Home | Main Index | Thread Index | Old Index