tech-kern archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: RFC: PERCPU_LIST to fix PR kern/52515
> Date: Fri, 6 Oct 2017 02:58:20 +0000
> From: Taylor R Campbell <campbell+netbsd-tech-kern%mumble.net@localhost>
>
> I wonder whether the open-coded version would do better if it
> unconditionally loaded the percpu:
For concreteness, in case you'd like to test this, I've attached an
updated patch that yields the conditional move sequence I quoted.
Index: sys/sys/psref.h
===================================================================
RCS file: /cvsroot/src/sys/sys/psref.h,v
retrieving revision 1.2
diff -p -u -r1.2 psref.h
--- sys/sys/psref.h 16 Dec 2016 20:12:11 -0000 1.2
+++ sys/sys/psref.h 6 Oct 2017 03:08:58 -0000
@@ -69,7 +69,8 @@ struct psref_target {
* written only on the local CPU.
*/
struct psref {
- LIST_ENTRY(psref) psref_entry;
+ struct psref *psref_next;
+ struct psref **psref_prevp;
const struct psref_target *psref_target;
struct lwp *psref_lwp;
struct cpu_info *psref_cpu;
Index: sys/kern/subr_psref.c
===================================================================
RCS file: /cvsroot/src/sys/kern/subr_psref.c,v
retrieving revision 1.7
diff -p -u -r1.7 subr_psref.c
--- sys/kern/subr_psref.c 1 Jun 2017 02:45:13 -0000 1.7
+++ sys/kern/subr_psref.c 6 Oct 2017 03:08:58 -0000
@@ -78,8 +78,6 @@ __KERNEL_RCSID(0, "$NetBSD: subr_psref.c
#include <sys/queue.h>
#include <sys/xcall.h>
-LIST_HEAD(psref_head, psref);
-
static bool _psref_held(const struct psref_target *, struct psref_class *,
bool);
@@ -103,7 +101,7 @@ struct psref_class {
* Not exposed by the API.
*/
struct psref_cpu {
- struct psref_head pcpu_head;
+ struct psref *pcpu_first;
};
/*
@@ -135,7 +133,7 @@ psref_cpu_drained_p(void *p, void *cooki
const struct psref_cpu *pcpu = p;
bool *retp = cookie;
- if (!LIST_EMPTY(&pcpu->pcpu_head))
+ if (pcpu->pcpu_first != NULL)
*retp = false;
}
@@ -194,7 +192,9 @@ psref_check_duplication(struct psref_cpu
bool found = false;
struct psref *_psref;
- LIST_FOREACH(_psref, &pcpu->pcpu_head, psref_entry) {
+ for (_psref = pcpu->pcpu_first;
+ _psref != NULL;
+ _psref = _psref->psref_next) {
if (_psref == psref &&
_psref->psref_target == target) {
found = true;
@@ -250,7 +250,11 @@ psref_acquire(struct psref *psref, const
#endif
/* Record our reference. */
- LIST_INSERT_HEAD(&pcpu->pcpu_head, psref, psref_entry);
+ if ((psref->psref_next = pcpu->pcpu_first) != NULL)
+ psref->psref_next->psref_prevp = &psref->psref_next;
+ psref->psref_prevp = NULL; /* head of list */
+ pcpu->pcpu_first = psref;
+
psref->psref_target = target;
psref->psref_lwp = curlwp;
psref->psref_cpu = curcpu();
@@ -273,6 +277,7 @@ void
psref_release(struct psref *psref, const struct psref_target *target,
struct psref_class *class)
{
+ struct psref_cpu *pcpu;
int s;
KASSERTMSG((kpreempt_disabled() || cpu_softintr_p() ||
@@ -297,12 +302,31 @@ psref_release(struct psref *psref, const
/*
* Block interrupts and remove the psref from the current CPU's
- * list. No need to percpu_getref or get the head of the list,
- * and the caller guarantees that we are bound to a CPU anyway
- * (as does blocking interrupts).
+ * list.
*/
s = splraiseipl(class->prc_iplcookie);
- LIST_REMOVE(psref, psref_entry);
+ if (psref->psref_next != NULL)
+ psref->psref_next->psref_prevp = psref->psref_prevp;
+
+ /*
+ * If it's at the head of the list, we must modify the head
+ * explicitly. We can't use a back-pointer into the head
+ * because percpu_alloc may move the per-CPU objects around in
+ * memory.
+ */
+ pcpu = percpu_getref(class->prc_percpu);
+ KASSERTMSG(psref->psref_prevp == NULL || *psref->psref_prevp == psref,
+ "psref %p prevp %p points at %p",
+ psref, psref->psref_prevp, *psref->psref_prevp);
+ KASSERTMSG(psref->psref_prevp != NULL || pcpu->pcpu_first == psref,
+ "psref %p marked as first but psref_cpu %p on %d first is %p",
+ psref, pcpu, cpu_index(curcpu()), pcpu->pcpu_first);
+ *(psref->psref_prevp ? psref->psref_prevp : &pcpu->pcpu_first) =
+ psref->psref_next;
+ percpu_putref(class->prc_percpu);
+
+ psref->psref_prevp = (void *)1; /* poison */
+ psref->psref_next = (void *)2; /* poison */
splx(s);
/* If someone is waiting for users to drain, notify 'em. */
@@ -353,7 +377,11 @@ psref_copy(struct psref *pto, const stru
pcpu = percpu_getref(class->prc_percpu);
/* Record the new reference. */
- LIST_INSERT_HEAD(&pcpu->pcpu_head, pto, psref_entry);
+ if ((pto->psref_next = pcpu->pcpu_first) != NULL)
+ pto->psref_next->psref_prevp = &pto->psref_next;
+ pto->psref_prevp = NULL; /* head of list */
+ pcpu->pcpu_first = pto;
+
pto->psref_target = pfrom->psref_target;
pto->psref_lwp = curlwp;
pto->psref_cpu = curcpu();
@@ -474,7 +502,9 @@ _psref_held(const struct psref_target *t
pcpu = percpu_getref(class->prc_percpu);
/* Search through all the references on this CPU. */
- LIST_FOREACH(psref, &pcpu->pcpu_head, psref_entry) {
+ for (psref = pcpu->pcpu_first;
+ psref != NULL;
+ psref = psref->psref_next) {
/* Sanity-check the reference's CPU. */
KASSERTMSG((psref->psref_cpu == curcpu()),
"passive reference transferred from CPU %u to CPU %u",
Home |
Main Index |
Thread Index |
Old Index