Source-Changes-HG archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

[src/trunk]: src/sys/kern style



details:   https://anonhg.NetBSD.org/src/rev/26be18080070
branches:  trunk
changeset: 449633:26be18080070
user:      maxv <maxv%NetBSD.org@localhost>
date:      Wed Mar 13 20:56:33 2019 +0000

description:
style

diffstat:

 sys/kern/subr_pool.c |  71 ++++++++++++++++++++++++++-------------------------
 1 files changed, 36 insertions(+), 35 deletions(-)

diffs (209 lines):

diff -r 637ccedd5a4f -r 26be18080070 sys/kern/subr_pool.c
--- a/sys/kern/subr_pool.c      Wed Mar 13 12:17:45 2019 +0000
+++ b/sys/kern/subr_pool.c      Wed Mar 13 20:56:33 2019 +0000
@@ -1,4 +1,4 @@
-/*     $NetBSD: subr_pool.c,v 1.235 2019/03/11 20:38:27 maxv Exp $     */
+/*     $NetBSD: subr_pool.c,v 1.236 2019/03/13 20:56:33 maxv Exp $     */
 
 /*
  * Copyright (c) 1997, 1999, 2000, 2002, 2007, 2008, 2010, 2014, 2015, 2018
@@ -33,7 +33,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: subr_pool.c,v 1.235 2019/03/11 20:38:27 maxv Exp $");
+__KERNEL_RCSID(0, "$NetBSD: subr_pool.c,v 1.236 2019/03/13 20:56:33 maxv Exp $");
 
 #ifdef _KERNEL_OPT
 #include "opt_ddb.h"
@@ -139,7 +139,7 @@
 int pool_inactive_time = 10;
 
 /* Next candidate for drainage (see pool_drain()) */
-static struct pool     *drainpp;
+static struct pool *drainpp;
 
 /* This lock protects both pool_head and drainpp. */
 static kmutex_t pool_head_lock;
@@ -383,16 +383,15 @@
 {
 
        /*
-        * we consider pool_item_header with smaller ph_page bigger.
-        * (this unnatural ordering is for the benefit of pr_find_pagehead.)
+        * We consider pool_item_header with smaller ph_page bigger. This
+        * unnatural ordering is for the benefit of pr_find_pagehead.
         */
-
        if (a->ph_page < b->ph_page)
-               return (1);
+               return 1;
        else if (a->ph_page > b->ph_page)
-               return (-1);
+               return -1;
        else
-               return (0);
+               return 0;
 }
 
 SPLAY_PROTOTYPE(phtree, pool_item_header, ph_node, phtree_compare);
@@ -592,7 +591,7 @@
                        palloc = &pool_allocator_kmem_fullpage;
                else if (palloc == &pool_allocator_nointr)
                        palloc = &pool_allocator_nointr_fullpage;
-       }               
+       }
 #endif /* POOL_SUBPAGE */
        if (!cold)
                mutex_enter(&pool_allocator_lock);
@@ -653,21 +652,20 @@
        pp->pr_freecheck = NULL;
        pool_redzone_init(pp, size);
 
+       /* Silently enforce '0 <= ioff < align'. */
+       ioff %= align;
+
        /*
-        * Decide whether to put the page header off page to avoid
-        * wasting too large a part of the page or too big item.
-        * Off-page page headers go on a hash table, so we can match
-        * a returned item with its header based on the page address.
-        * We use 1/16 of the page size and about 8 times of the item
-        * size as the threshold (XXX: tune)
+        * Decide whether to put the page header off page to avoid wasting too
+        * large a part of the page or too big item. Off-page page headers go
+        * on a hash table, so we can match a returned item with its header
+        * based on the page address. We use 1/16 of the page size and about 8
+        * times of the item size as the threshold. (XXX: tune)
         *
-        * However, we'll put the header into the page if we can put
-        * it without wasting any items.
-        *
-        * Silently enforce `0 <= ioff < align'.
+        * However, we'll put the header into the page if we can put it without
+        * wasting any items.
         */
-       pp->pr_itemoffset = ioff %= align;
-       /* See the comment below about reserved bytes. */
+       pp->pr_itemoffset = ioff;
        trysize = palloc->pa_pagesz - ((align - ioff) % align);
        phsize = ALIGN(sizeof(struct pool_item_header));
        if (pp->pr_roflags & PR_PHINPAGE ||
@@ -829,7 +827,7 @@
        else
                ph = pool_get(pp->pr_phpool, flags);
 
-       return (ph);
+       return ph;
 }
 
 /*
@@ -900,7 +898,7 @@
 
                mutex_exit(&pp->pr_lock);
                KASSERT((flags & (PR_NOWAIT|PR_LIMITFAIL)) != 0);
-               return (NULL);
+               return NULL;
        }
 
        /*
@@ -943,7 +941,7 @@
                        pp->pr_nfail++;
                        mutex_exit(&pp->pr_lock);
                        KASSERT((flags & (PR_WAITOK|PR_NOWAIT)) == PR_NOWAIT);
-                       return (NULL);
+                       return NULL;
                }
 
                /* Start the allocation process over. */
@@ -1126,6 +1124,9 @@
 static int
 pool_grow(struct pool *pp, int flags)
 {
+       struct pool_item_header *ph;
+       char *cp;
+
        /*
         * If there's a pool_grow in progress, wait for it to complete
         * and try again from the top.
@@ -1157,11 +1158,11 @@
        else
                pp->pr_flags |= PR_GROWINGNOWAIT;
 
-       char *cp = pool_allocator_alloc(pp, flags);
+       cp = pool_allocator_alloc(pp, flags);
        if (__predict_false(cp == NULL))
                goto out;
 
-       struct pool_item_header *ph = pool_alloc_item_header(pp, cp, flags);
+       ph = pool_alloc_item_header(pp, cp, flags);
        if (__predict_false(ph == NULL)) {
                pool_allocator_free(pp, cp);
                goto out;
@@ -1226,10 +1227,10 @@
 static void
 pool_prime_page(struct pool *pp, void *storage, struct pool_item_header *ph)
 {
+       const unsigned int align = pp->pr_align;
+       const unsigned int ioff = pp->pr_itemoffset;
        struct pool_item *pi;
        void *cp = storage;
-       const unsigned int align = pp->pr_align;
-       const unsigned int ioff = pp->pr_itemoffset;
        int n;
 
        KASSERT(mutex_owned(&pp->pr_lock));
@@ -1259,7 +1260,7 @@
                pp->pr_curcolor = 0;
 
        /*
-        * Adjust storage to apply aligment to `pr_itemoffset' in each item.
+        * Adjust storage to apply alignment to `pr_itemoffset' in each item.
         */
        if (ioff != 0)
                cp = (char *)cp + align - ioff;
@@ -1439,7 +1440,7 @@
                if (klock) {
                        KERNEL_UNLOCK_ONE(NULL);
                }
-               return (0);
+               return 0;
        }
 
        LIST_INIT(&pq);
@@ -1481,7 +1482,7 @@
                KERNEL_UNLOCK_ONE(NULL);
        }
 
-       return (rv);
+       return rv;
 }
 
 /*
@@ -1803,7 +1804,7 @@
 
 out:
        mutex_exit(&pp->pr_lock);
-       return (r);
+       return r;
 }
 
 /*
@@ -2501,7 +2502,7 @@
        pool_cache_destruct_object(pc, object);
 
        return false;
-} 
+}
 
 /*
  * pool_cache_put{,_paddr}:
@@ -2548,7 +2549,7 @@
                }
 
                /*
-                * Can't free to either group: try the slow path. 
+                * Can't free to either group: try the slow path.
                 * If put_slow() releases the object for us, it
                 * will return false.  Otherwise we need to retry.
                 */



Home | Main Index | Thread Index | Old Index