Source-Changes-HG archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

[src/trunk]: src/sys/external/bsd/drm2/dist/drm/nouveau Fences may last longe...



details:   https://anonhg.NetBSD.org/src/rev/cbb2908b0d5a
branches:  trunk
changeset: 365589:cbb2908b0d5a
user:      riastradh <riastradh%NetBSD.org@localhost>
date:      Thu Aug 23 01:10:04 2018 +0000

description:
Fences may last longer than their channels.

- Use a reference count on the nouveau_fence_chan object.
- Acquire it with kpreemption disabled.
- Use xcall to wait for kpreempt-disabled sections to complete.

PR kern/53441

XXX pullup-7
XXX pullup-8

diffstat:

 sys/external/bsd/drm2/dist/drm/nouveau/nouveau_fence.c |  142 ++++++++++++++--
 sys/external/bsd/drm2/dist/drm/nouveau/nouveau_fence.h |    1 +
 2 files changed, 120 insertions(+), 23 deletions(-)

diffs (285 lines):

diff -r 01c367ed13e6 -r cbb2908b0d5a sys/external/bsd/drm2/dist/drm/nouveau/nouveau_fence.c
--- a/sys/external/bsd/drm2/dist/drm/nouveau/nouveau_fence.c    Thu Aug 23 01:06:50 2018 +0000
+++ b/sys/external/bsd/drm2/dist/drm/nouveau/nouveau_fence.c    Thu Aug 23 01:10:04 2018 +0000
@@ -1,4 +1,4 @@
-/*     $NetBSD: nouveau_fence.c,v 1.5 2018/08/23 01:06:50 riastradh Exp $      */
+/*     $NetBSD: nouveau_fence.c,v 1.6 2018/08/23 01:10:04 riastradh Exp $      */
 
 /*
  * Copyright (C) 2007 Ben Skeggs.
@@ -27,7 +27,10 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: nouveau_fence.c,v 1.5 2018/08/23 01:06:50 riastradh Exp $");
+__KERNEL_RCSID(0, "$NetBSD: nouveau_fence.c,v 1.6 2018/08/23 01:10:04 riastradh Exp $");
+
+#include <sys/types.h>
+#include <sys/xcall.h>
 
 #include <drm/drmP.h>
 
@@ -55,6 +58,60 @@
 };
 
 /*
+ * nouveau_fence_channel_acquire(fence)
+ *
+ *     Try to return the channel associated with fence.
+ */
+static struct nouveau_channel *
+nouveau_fence_channel_acquire(struct nouveau_fence *fence)
+{
+       struct nouveau_channel *chan;
+       struct nouveau_fence_chan *fctx;
+
+       /*
+        * Block cross-calls while we examine fence.  If we observe
+        * that fence->done is false, then the channel cannot be
+        * destroyed even by another CPU until after kpreempt_enable.
+        */
+       kpreempt_disable();
+       if (fence->done) {
+               chan = NULL;
+       } else {
+               chan = fence->channel;
+               fctx = chan->fence;
+               atomic_inc_uint(&fctx->refcnt);
+       }
+       kpreempt_enable();
+
+       return chan;
+}
+
+/*
+ * nouveau_fence_channel_release(channel)
+ *
+ *     Release the channel acquired with nouveau_fence_channel_acquire.
+ */
+static void
+nouveau_fence_channel_release(struct nouveau_channel *chan)
+{
+       struct nouveau_fence_chan *fctx = chan->fence;
+       unsigned old, new;
+
+       do {
+               old = fctx->refcnt;
+               if (old == 0) {
+                       spin_lock(&fctx->lock);
+                       if (atomic_dec_uint_nv(&fctx->refcnt) == 0)
+                               DRM_SPIN_WAKEUP_ALL(&fctx->waitqueue,
+                                   &fctx->lock);
+                       spin_unlock(&fctx->lock);
+                       return;
+               }
+               new = old - 1;
+       } while (atomic_cas_uint(&fctx->refcnt, old, new) != old);
+}
+
+/*
  * nouveau_fence_signal(fence)
  *
  *     Schedule all the work for fence's completion, and mark it done.
@@ -82,6 +139,11 @@
        list_del(&fence->head);
 }
 
+static void
+nouveau_fence_context_del_xc(void *a, void *b)
+{
+}
+
 /*
  * nouveau_fence_context_del(fctx)
  *
@@ -92,6 +154,7 @@
 nouveau_fence_context_del(struct nouveau_fence_chan *fctx)
 {
        struct nouveau_fence *fence, *fnext;
+       int ret __diagused;
 
        /* Signal all the fences in fctx.  */
        spin_lock(&fctx->lock);
@@ -104,6 +167,16 @@
        /* Wait for the workqueue to drain.  */
        flush_scheduled_work();
 
+       /* Wait for nouveau_fence_channel_acquire to complete on all CPUs.  */
+       xc_wait(xc_broadcast(0, nouveau_fence_context_del_xc, NULL, NULL));
+
+       /* Wait for any references to drain.  */
+       spin_lock(&fctx->lock);
+       DRM_SPIN_WAIT_NOINTR_UNTIL(ret, &fctx->waitqueue, &fctx->lock,
+           fctx->refcnt == 0);
+       BUG_ON(ret);
+       spin_unlock(&fctx->lock);
+
        /* Destroy the fence context.  */
        DRM_DESTROY_WAITQUEUE(&fctx->waitqueue);
        spin_lock_destroy(&fctx->lock);
@@ -122,6 +195,7 @@
        INIT_LIST_HEAD(&fctx->pending);
        spin_lock_init(&fctx->lock);
        DRM_INIT_WAITQUEUE(&fctx->waitqueue, "nvfnchan");
+       fctx->refcnt = 0;
 }
 
 /*
@@ -149,29 +223,37 @@
 nouveau_fence_work(struct nouveau_fence *fence,
                   void (*func)(void *), void *data)
 {
-       struct nouveau_channel *chan = fence->channel;
-       struct nouveau_fence_chan *fctx = chan->fence;
+       struct nouveau_channel *chan;
+       struct nouveau_fence_chan *fctx;
        struct fence_work *work = NULL;
 
+       if ((chan = nouveau_fence_channel_acquire(fence)) == NULL)
+               goto now0;
+       fctx = chan->fence;
+
        work = kmalloc(sizeof(*work), GFP_KERNEL);
        if (work == NULL) {
                WARN_ON(nouveau_fence_wait(fence, false, false));
-               func(data);
-               return;
+               goto now1;
        }
 
        spin_lock(&fctx->lock);
        if (fence->done) {
                spin_unlock(&fctx->lock);
-               kfree(work);
-               func(data);
-               return;
+               goto now2;
        }
        INIT_WORK(&work->base, nouveau_fence_work_handler);
        work->func = func;
        work->data = data;
        list_add(&work->head, &fence->work);
+       if (atomic_dec_uint_nv(&fctx->refcnt) == 0)
+               DRM_SPIN_WAKEUP_ALL(&fctx->waitqueue, &fctx->lock);
        spin_unlock(&fctx->lock);
+       return;
+
+now2:  kfree(work);
+now1:  nouveau_fence_channel_release(chan);
+now0:  func(data);
 }
 
 /*
@@ -265,14 +347,20 @@
 bool
 nouveau_fence_done(struct nouveau_fence *fence)
 {
-       struct nouveau_channel *chan = fence->channel;
-       struct nouveau_fence_chan *fctx = chan->fence;
+       struct nouveau_channel *chan;
+       struct nouveau_fence_chan *fctx;
        bool done;
 
+       if ((chan = nouveau_fence_channel_acquire(fence)) == NULL)
+               return true;
+
+       fctx = chan->fence;
        spin_lock(&fctx->lock);
        done = nouveau_fence_done_locked(fence, chan);
        spin_unlock(&fctx->lock);
 
+       nouveau_fence_channel_release(chan);
+
        return done;
 }
 
@@ -297,20 +385,22 @@
 }
 
 /*
- * nouveau_fence_wait_uevent(fence, intr)
+ * nouveau_fence_wait_uevent(fence, chan, intr)
  *
- *     Wait using a nouveau event for completion of fence.  Wait
- *     interruptibly iff intr is true.
+ *     Wait using a nouveau event for completion of fence on chan.
+ *     Wait interruptibly iff intr is true.
  */
 static int
-nouveau_fence_wait_uevent(struct nouveau_fence *fence, bool intr)
+nouveau_fence_wait_uevent(struct nouveau_fence *fence,
+    struct nouveau_channel *chan, bool intr)
 {
-       struct nouveau_channel *chan = fence->channel;
        struct nouveau_fifo *pfifo = nouveau_fifo(chan->drm->device);
        struct nouveau_fence_chan *fctx = chan->fence;
        struct nouveau_eventh *handler;
        int ret = 0;
 
+       BUG_ON(fence->channel != chan);
+
        ret = nouveau_event_new(pfifo->uevent, 0,
                                nouveau_fence_wait_uevent_handler,
                                fctx, &handler);
@@ -374,15 +464,19 @@
 int
 nouveau_fence_wait(struct nouveau_fence *fence, bool lazy, bool intr)
 {
-       struct nouveau_channel *chan = fence->channel;
-       struct nouveau_fence_priv *priv = chan->drm->fence;
+       struct nouveau_channel *chan;
+       struct nouveau_fence_priv *priv;
        unsigned long delay_usec = 1;
        int ret = 0;
 
+       if ((chan = nouveau_fence_channel_acquire(fence)) == NULL)
+               goto out0;
+
+       priv = chan->drm->fence;
        while (priv && priv->uevent && lazy && !nouveau_fence_done(fence)) {
-               ret = nouveau_fence_wait_uevent(fence, intr);
+               ret = nouveau_fence_wait_uevent(fence, chan, intr);
                if (ret < 0)
-                       return ret;
+                       goto out1;
        }
 
        while (!nouveau_fence_done(fence)) {
@@ -402,7 +496,8 @@
                }
        }
 
-       return ret;
+out1:  nouveau_fence_channel_release(chan);
+out0:  return ret;
 }
 
 int
@@ -412,13 +507,14 @@
        struct nouveau_channel *prev;
        int ret = 0;
 
-       prev = fence ? fence->channel : NULL;
-       if (prev) {
+       if (fence != NULL &&
+           (prev = nouveau_fence_channel_acquire(fence)) != NULL) {
                if (unlikely(prev != chan && !nouveau_fence_done(fence))) {
                        ret = fctx->sync(fence, prev, chan);
                        if (unlikely(ret))
                                ret = nouveau_fence_wait(fence, true, false);
                }
+               nouveau_fence_channel_release(prev);
        }
 
        return ret;
diff -r 01c367ed13e6 -r cbb2908b0d5a sys/external/bsd/drm2/dist/drm/nouveau/nouveau_fence.h
--- a/sys/external/bsd/drm2/dist/drm/nouveau/nouveau_fence.h    Thu Aug 23 01:06:50 2018 +0000
+++ b/sys/external/bsd/drm2/dist/drm/nouveau/nouveau_fence.h    Thu Aug 23 01:10:04 2018 +0000
@@ -46,6 +46,7 @@
 
        spinlock_t lock;
        drm_waitqueue_t waitqueue;
+       volatile unsigned refcnt;
        u32 sequence;
 };
 



Home | Main Index | Thread Index | Old Index