[BACK]Return to vfs_bio.c CVS log [TXT][DIR] Up to [local] / sys / kern

Annotation of sys/kern/vfs_bio.c, Revision 1.1

1.1     ! nbrk        1: /*     $OpenBSD: vfs_bio.c,v 1.99 2007/08/07 04:32:45 beck Exp $       */
        !             2: /*     $NetBSD: vfs_bio.c,v 1.44 1996/06/11 11:15:36 pk Exp $  */
        !             3:
        !             4: /*-
        !             5:  * Copyright (c) 1994 Christopher G. Demetriou
        !             6:  * Copyright (c) 1982, 1986, 1989, 1993
        !             7:  *     The Regents of the University of California.  All rights reserved.
        !             8:  * (c) UNIX System Laboratories, Inc.
        !             9:  * All or some portions of this file are derived from material licensed
        !            10:  * to the University of California by American Telephone and Telegraph
        !            11:  * Co. or Unix System Laboratories, Inc. and are reproduced herein with
        !            12:  * the permission of UNIX System Laboratories, Inc.
        !            13:  *
        !            14:  * Redistribution and use in source and binary forms, with or without
        !            15:  * modification, are permitted provided that the following conditions
        !            16:  * are met:
        !            17:  * 1. Redistributions of source code must retain the above copyright
        !            18:  *    notice, this list of conditions and the following disclaimer.
        !            19:  * 2. Redistributions in binary form must reproduce the above copyright
        !            20:  *    notice, this list of conditions and the following disclaimer in the
        !            21:  *    documentation and/or other materials provided with the distribution.
        !            22:  * 3. Neither the name of the University nor the names of its contributors
        !            23:  *    may be used to endorse or promote products derived from this software
        !            24:  *    without specific prior written permission.
        !            25:  *
        !            26:  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
        !            27:  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
        !            28:  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
        !            29:  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
        !            30:  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
        !            31:  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
        !            32:  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
        !            33:  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
        !            34:  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
        !            35:  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
        !            36:  * SUCH DAMAGE.
        !            37:  *
        !            38:  *     @(#)vfs_bio.c   8.6 (Berkeley) 1/11/94
        !            39:  */
        !            40:
        !            41: /*
        !            42:  * Some references:
        !            43:  *     Bach: The Design of the UNIX Operating System (Prentice Hall, 1986)
        !            44:  *     Leffler, et al.: The Design and Implementation of the 4.3BSD
        !            45:  *             UNIX Operating System (Addison Welley, 1989)
        !            46:  */
        !            47:
        !            48: #include <sys/param.h>
        !            49: #include <sys/systm.h>
        !            50: #include <sys/proc.h>
        !            51: #include <sys/buf.h>
        !            52: #include <sys/vnode.h>
        !            53: #include <sys/mount.h>
        !            54: #include <sys/malloc.h>
        !            55: #include <sys/pool.h>
        !            56: #include <sys/resourcevar.h>
        !            57: #include <sys/conf.h>
        !            58: #include <sys/kernel.h>
        !            59:
        !            60: #include <uvm/uvm_extern.h>
        !            61:
        !            62: #include <miscfs/specfs/specdev.h>
        !            63:
        !            64: /*
        !            65:  * Definitions for the buffer hash lists.
        !            66:  */
        !            67: #define        BUFHASH(dvp, lbn)       \
        !            68:        (&bufhashtbl[((long)(dvp) / sizeof(*(dvp)) + (int)(lbn)) & bufhash])
        !            69: LIST_HEAD(bufhashhdr, buf) *bufhashtbl, invalhash;
        !            70: u_long bufhash;
        !            71:
        !            72: /*
        !            73:  * Insq/Remq for the buffer hash lists.
        !            74:  */
        !            75: #define        binshash(bp, dp)        LIST_INSERT_HEAD(dp, bp, b_hash)
        !            76: #define        bremhash(bp)            LIST_REMOVE(bp, b_hash)
        !            77:
        !            78: /*
        !            79:  * Definitions for the buffer free lists.
        !            80:  */
        !            81: #define        BQUEUES         6               /* number of free buffer queues */
        !            82:
        !            83: #define        BQ_DIRTY        0               /* LRU queue with dirty buffers */
        !            84:
        !            85:
        !            86: TAILQ_HEAD(bqueues, buf) bufqueues[BQUEUES];
        !            87: int bqpages[BQUEUES];          /* pages allocated, per queue */
        !            88: int bqpagelow;
        !            89: int needbuffer;
        !            90: struct bio_ops bioops;
        !            91:
        !            92: /*
        !            93:  * Buffer pool for I/O buffers.
        !            94:  */
        !            95: struct pool bufpool;
        !            96: struct vm_map *buf_map;
        !            97: struct bufhead bufhead = LIST_HEAD_INITIALIZER(bufhead);
        !            98: struct buf *buf_get(size_t);
        !            99: struct buf *buf_stub(struct vnode *, daddr64_t);
        !           100: void buf_put(struct buf *);
        !           101:
        !           102: /*
        !           103:  * Insq/Remq for the buffer free lists.
        !           104:  */
        !           105: #define        binsheadfree(bp, dp)    TAILQ_INSERT_HEAD(dp, bp, b_freelist)
        !           106: #define        binstailfree(bp, dp)    TAILQ_INSERT_TAIL(dp, bp, b_freelist)
        !           107:
        !           108: struct buf *bio_doread(struct vnode *, daddr64_t, int, int);
        !           109: struct buf *getnewbuf(size_t, int, int, int *);
        !           110: void buf_init(struct buf *, int);
        !           111: void bread_cluster_callback(struct buf *);
        !           112:
        !           113: /*
        !           114:  * We keep a few counters to monitor the utilization of the buffer cache
        !           115:  *
        !           116:  *  numbufpages   - number of pages totally allocated.
        !           117:  *  numdirtypages - number of pages on BQ_DIRTY queue.
        !           118:  *  lodirtypages  - low water mark for buffer cleaning daemon.
        !           119:  *  hidirtypages  - high water mark for buffer cleaning daemon.
        !           120:  *  numfreepages  - number of pages on BQ_CLEAN and BQ_DIRTY queues. unused.
        !           121:  *  numcleanpages - number of pages on BQ_CLEAN queue.
        !           122:  *                 Used to track the need to speedup the cleaner and
        !           123:  *                 as a reserve for special processes like syncer.
        !           124:  *  maxcleanpages - the highest page count on BQ_CLEAN.
        !           125:  */
        !           126: long numbufpages;
        !           127: long numdirtypages;
        !           128: long lodirtypages;
        !           129: long hidirtypages;
        !           130: long numfreepages;
        !           131: long numcleanpages;
        !           132: long locleanpages;
        !           133: long hicleanpages;
        !           134: long maxcleanpages;
        !           135:
        !           136: struct proc *cleanerproc;
        !           137: int bd_req;                    /* Sleep point for cleaner daemon. */
        !           138:
        !           139: int size2cqueue(int *size);
        !           140:
        !           141: int
        !           142: size2cqueue(int *size)
        !           143: {
        !           144:        int i = 0, q;
        !           145:        int s = *size;
        !           146:        s -= 1;
        !           147:        while (s > 0) {
        !           148:                s = s >> 1;
        !           149:                i++;
        !           150:        }
        !           151:        if (i < PAGE_SHIFT) {
        !           152:                i = PAGE_SHIFT; /* < 4096 -> 4096 */
        !           153:         }
        !           154:        *size = 1 << i;
        !           155:        q = (i + 1 - PAGE_SHIFT); /* XXX 4096 is queue 1 */
        !           156:        if (q >= BQUEUES)
        !           157:                panic("queue %d > BQUEUES %d", q, BQUEUES);
        !           158:        if (q == 0)
        !           159:                panic("can't return dirty q");
        !           160:        return(q);
        !           161: }
        !           162:
        !           163: void
        !           164: bremfree(struct buf *bp)
        !           165: {
        !           166:        struct bqueues *dp = NULL;
        !           167:        int queue;
        !           168:
        !           169:        /*
        !           170:         * We only calculate the head of the freelist when removing
        !           171:         * the last element of the list as that is the only time that
        !           172:         * it is needed (e.g. to reset the tail pointer).
        !           173:         *
        !           174:         * NB: This makes an assumption about how tailq's are implemented.
        !           175:         */
        !           176:        if (TAILQ_NEXT(bp, b_freelist) == NULL) {
        !           177:                for (dp = bufqueues; dp < &bufqueues[BQUEUES]; dp++)
        !           178:                        if (dp->tqh_last == &TAILQ_NEXT(bp, b_freelist))
        !           179:                                break;
        !           180:                if (dp == &bufqueues[BQUEUES])
        !           181:                        panic("bremfree: lost tail");
        !           182:        }
        !           183:        numfreepages -= btoc(bp->b_bufsize);
        !           184:        if (!ISSET(bp->b_flags, B_DELWRI)) {
        !           185:                int qs = bp->b_bufsize;
        !           186:                queue = size2cqueue(&qs);
        !           187:                numcleanpages -= btoc(bp->b_bufsize);
        !           188:                bqpages[queue] -= btoc(bp->b_bufsize);
        !           189:        } else
        !           190:                numdirtypages -= btoc(bp->b_bufsize);
        !           191:        TAILQ_REMOVE(dp, bp, b_freelist);
        !           192: }
        !           193:
        !           194: void
        !           195: buf_init(struct buf *bp, int size)
        !           196: {
        !           197:        int npages, queue;
        !           198:
        !           199:        splassert(IPL_BIO);
        !           200:
        !           201:        npages = btoc(size);
        !           202:        bzero((char *)bp, sizeof *bp);
        !           203:        bp->b_vnbufs.le_next = NOLIST;
        !           204:        bp->b_freelist.tqe_next = NOLIST;
        !           205:        bp->b_synctime = time_uptime + 300;
        !           206:        bp->b_dev = NODEV;
        !           207:        queue = size2cqueue(&size);
        !           208:        LIST_INIT(&bp->b_dep);
        !           209:        numbufpages += npages;
        !           210:        numfreepages += npages;
        !           211:        numcleanpages += npages;
        !           212:        bqpages[queue] += npages;
        !           213:        if (maxcleanpages < numcleanpages)
        !           214:                maxcleanpages = numcleanpages;
        !           215: }
        !           216:
        !           217: /*
        !           218:  * This is a non-sleeping expanded equivalent of getblk() that allocates only
        !           219:  * the buffer structure, and not its contents.
        !           220:  */
        !           221: struct buf *
        !           222: buf_stub(struct vnode *vp, daddr64_t lblkno)
        !           223: {
        !           224:        struct buf *bp;
        !           225:        int s;
        !           226:
        !           227:        s = splbio();
        !           228:        bp = pool_get(&bufpool, PR_NOWAIT);
        !           229:        splx(s);
        !           230:
        !           231:        if (bp == NULL)
        !           232:                return (NULL);
        !           233:
        !           234:        bzero((char *)bp, sizeof *bp);
        !           235:        bp->b_vnbufs.le_next = NOLIST;
        !           236:        bp->b_freelist.tqe_next = NOLIST;
        !           237:        bp->b_synctime = time_uptime + 300;
        !           238:        bp->b_dev = NODEV;
        !           239:        bp->b_bufsize = 0;
        !           240:        bp->b_data = NULL;
        !           241:        bp->b_flags = B_BUSY;
        !           242:        bp->b_dev = NODEV;
        !           243:        bp->b_blkno = bp->b_lblkno = lblkno;
        !           244:        bp->b_iodone = NULL;
        !           245:        bp->b_error = 0;
        !           246:        bp->b_resid = 0;
        !           247:        bp->b_bcount = 0;
        !           248:        bp->b_dirtyoff = bp->b_dirtyend = 0;
        !           249:        bp->b_validoff = bp->b_validend = 0;
        !           250:
        !           251:        LIST_INIT(&bp->b_dep);
        !           252:
        !           253:        s = splbio();
        !           254:        LIST_INSERT_HEAD(&bufhead, bp, b_list);
        !           255:        bgetvp(vp, bp);
        !           256:        splx(s);
        !           257:
        !           258:        return (bp);
        !           259: }
        !           260:
        !           261: struct buf *
        !           262: buf_get(size_t size)
        !           263: {
        !           264:        struct bqueues *dp;
        !           265:        struct buf *bp;
        !           266:        int npages;
        !           267:        int queue, qs;
        !           268:        void *data;
        !           269:
        !           270:        splassert(IPL_BIO);
        !           271:
        !           272:        KASSERT(size > 0);
        !           273:
        !           274:        size = round_page(size);
        !           275:        qs = size;
        !           276:        queue = size2cqueue(&qs);
        !           277:        npages = btoc(qs);
        !           278:
        !           279:        if (numbufpages + npages > bufpages)
        !           280:                return (NULL);
        !           281:
        !           282:        bp = pool_get(&bufpool, PR_WAITOK);
        !           283:
        !           284:        data = (void *)uvm_km_alloc(buf_map, qs);
        !           285:        if (data == NULL) {
        !           286:                pool_put(&bufpool, bp);
        !           287:                return (NULL);
        !           288:        }
        !           289:        buf_init(bp, qs);
        !           290:        bp->b_flags = B_INVAL;
        !           291:        bp->b_bufsize = qs;
        !           292:        bp->b_data = data;
        !           293:        dp = &bufqueues[queue];
        !           294:        binsheadfree(bp, dp);
        !           295:        binshash(bp, &invalhash);
        !           296:        LIST_INSERT_HEAD(&bufhead, bp, b_list);
        !           297:
        !           298:        return (bp);
        !           299: }
        !           300:
        !           301: void
        !           302: buf_put(struct buf *bp)
        !           303: {
        !           304:        splassert(IPL_BIO);
        !           305: #ifdef DIAGNOSTIC
        !           306:        if (bp->b_data != NULL)
        !           307:                KASSERT(bp->b_bufsize > 0);
        !           308: #endif
        !           309: #ifdef QUEUE_MACRO_DEBUG
        !           310:        if (bp->b_freelist.tqe_next != NOLIST &&
        !           311:            bp->b_freelist.tqe_next != (void *)-1)
        !           312:                panic("buf_put: still on the free list");
        !           313:
        !           314:        if (bp->b_vnbufs.le_next != NOLIST &&
        !           315:            bp->b_vnbufs.le_next != (void *)-1)
        !           316:                panic("buf_put: still on the vnode list");
        !           317: #endif
        !           318: #ifdef DIAGNOSTIC
        !           319:        if (!LIST_EMPTY(&bp->b_dep))
        !           320:                panic("buf_put: b_dep is not empty");
        !           321: #endif
        !           322:        LIST_REMOVE(bp, b_list);
        !           323:
        !           324:        if (bp->b_data != NULL) {
        !           325:                bremhash(bp);
        !           326:                numbufpages -= btoc(bp->b_bufsize);
        !           327:                uvm_km_free(buf_map, (vaddr_t)bp->b_data, bp->b_bufsize);
        !           328:        }
        !           329:
        !           330:        pool_put(&bufpool, bp);
        !           331: }
        !           332:
        !           333: /*
        !           334:  * Initialize buffers and hash links for buffers.
        !           335:  */
        !           336: void
        !           337: bufinit(void)
        !           338: {
        !           339:        vaddr_t minaddr, maxaddr;
        !           340:        struct bqueues *dp;
        !           341:
        !           342:        pool_init(&bufpool, sizeof(struct buf), 0, 0, 0, "bufpl", NULL);
        !           343:        pool_setipl(&bufpool, IPL_BIO);
        !           344:        for (dp = bufqueues; dp < &bufqueues[BQUEUES]; dp++)
        !           345:                TAILQ_INIT(dp);
        !           346:        minaddr = vm_map_min(kernel_map);
        !           347:        buf_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr,
        !           348:            ptoa(bufpages), 0, FALSE, NULL);
        !           349:
        !           350:        /*
        !           351:         * XXX don't starve any one queue below 5% of the total number
        !           352:         * of buffer cache pages.
        !           353:         */
        !           354:        bqpagelow = bufpages / 20;
        !           355:
        !           356:        bufhashtbl = hashinit(bufpages / 4, M_CACHE, M_WAITOK, &bufhash);
        !           357:        hidirtypages = (bufpages / 4) * 3;
        !           358:        lodirtypages = bufpages / 2;
        !           359:
        !           360:        /*
        !           361:         * Reserve 5% of bufpages for syncer's needs,
        !           362:         * but not more than 25% and if possible
        !           363:         * not less than 2 * MAXBSIZE. locleanpages
        !           364:         * value must be not too small
        !           365:         */
        !           366:        hicleanpages = bufpages / 2;
        !           367:        locleanpages = hicleanpages / 2;
        !           368:        if (locleanpages < btoc(2 * MAXBSIZE))
        !           369:                locleanpages = btoc(2 * MAXBSIZE);
        !           370:        if (locleanpages > bufpages / 4)
        !           371:                locleanpages = bufpages / 4;
        !           372:
        !           373:        maxcleanpages = locleanpages;
        !           374: }
        !           375:
        !           376: struct buf *
        !           377: bio_doread(struct vnode *vp, daddr64_t blkno, int size, int async)
        !           378: {
        !           379:        struct buf *bp;
        !           380:
        !           381:        bp = getblk(vp, blkno, size, 0, 0);
        !           382:
        !           383:        /*
        !           384:         * If buffer does not have valid data, start a read.
        !           385:         * Note that if buffer is B_INVAL, getblk() won't return it.
        !           386:         * Therefore, it's valid if its I/O has completed or been delayed.
        !           387:         */
        !           388:        if (!ISSET(bp->b_flags, (B_DONE | B_DELWRI))) {
        !           389:                SET(bp->b_flags, B_READ | async);
        !           390:                VOP_STRATEGY(bp);
        !           391:
        !           392:                /* Pay for the read. */
        !           393:                curproc->p_stats->p_ru.ru_inblock++;            /* XXX */
        !           394:        } else if (async) {
        !           395:                brelse(bp);
        !           396:        }
        !           397:
        !           398:        return (bp);
        !           399: }
        !           400:
        !           401: /*
        !           402:  * Read a disk block.
        !           403:  * This algorithm described in Bach (p.54).
        !           404:  */
        !           405: int
        !           406: bread(struct vnode *vp, daddr64_t blkno, int size, struct ucred *cred,
        !           407:     struct buf **bpp)
        !           408: {
        !           409:        struct buf *bp;
        !           410:
        !           411:        /* Get buffer for block. */
        !           412:        bp = *bpp = bio_doread(vp, blkno, size, 0);
        !           413:
        !           414:        /* Wait for the read to complete, and return result. */
        !           415:        return (biowait(bp));
        !           416: }
        !           417:
        !           418: /*
        !           419:  * Read-ahead multiple disk blocks. The first is sync, the rest async.
        !           420:  * Trivial modification to the breada algorithm presented in Bach (p.55).
        !           421:  */
        !           422: int
        !           423: breadn(struct vnode *vp, daddr64_t blkno, int size, daddr64_t rablks[],
        !           424:     int rasizes[], int nrablks, struct ucred *cred, struct buf **bpp)
        !           425: {
        !           426:        struct buf *bp;
        !           427:        int i;
        !           428:
        !           429:        bp = *bpp = bio_doread(vp, blkno, size, 0);
        !           430:
        !           431:        /*
        !           432:         * For each of the read-ahead blocks, start a read, if necessary.
        !           433:         */
        !           434:        for (i = 0; i < nrablks; i++) {
        !           435:                /* If it's in the cache, just go on to next one. */
        !           436:                if (incore(vp, rablks[i]))
        !           437:                        continue;
        !           438:
        !           439:                /* Get a buffer for the read-ahead block */
        !           440:                (void) bio_doread(vp, rablks[i], rasizes[i], B_ASYNC);
        !           441:        }
        !           442:
        !           443:        /* Otherwise, we had to start a read for it; wait until it's valid. */
        !           444:        return (biowait(bp));
        !           445: }
        !           446:
        !           447: /*
        !           448:  * Called from interrupt context.
        !           449:  */
        !           450: void
        !           451: bread_cluster_callback(struct buf *bp)
        !           452: {
        !           453:        int i;
        !           454:        struct buf **xbpp;
        !           455:
        !           456:        xbpp = (struct buf **)bp->b_saveaddr;
        !           457:
        !           458:        for (i = 0; xbpp[i] != 0; i++) {
        !           459:                if (ISSET(bp->b_flags, B_ERROR))
        !           460:                        SET(xbpp[i]->b_flags, B_INVAL | B_ERROR);
        !           461:                biodone(xbpp[i]);
        !           462:        }
        !           463:
        !           464:        free(xbpp, M_TEMP);
        !           465:        bp->b_data = NULL;
        !           466:        buf_put(bp);
        !           467: }
        !           468:
        !           469: int
        !           470: bread_cluster(struct vnode *vp, daddr64_t blkno, int size, struct buf **rbpp)
        !           471: {
        !           472:        struct buf *bp, **xbpp;
        !           473:        int howmany, i, maxra, inc;
        !           474:        daddr64_t sblkno;
        !           475:        size_t spill;
        !           476:
        !           477:        *rbpp = bio_doread(vp, blkno, size, 0);
        !           478:
        !           479:        if (size != round_page(size))
        !           480:                return (biowait(*rbpp));
        !           481:
        !           482:        if (VOP_BMAP(vp, blkno + 1, NULL, &sblkno, &maxra))
        !           483:                return (biowait(*rbpp));
        !           484:
        !           485:        maxra++;
        !           486:        if (sblkno == -1 || maxra < 2)
        !           487:                return (biowait(*rbpp));
        !           488:
        !           489:        howmany = MAXPHYS / size;
        !           490:        if (howmany > maxra)
        !           491:                howmany = maxra;
        !           492:
        !           493:        xbpp = malloc((howmany + 1) * sizeof(struct buf *), M_TEMP, M_NOWAIT);
        !           494:        if (xbpp == NULL)
        !           495:                return (biowait(*rbpp));
        !           496:
        !           497:        for (i = 0; i < howmany; i++) {
        !           498:                if (incore(vp, blkno + i + 1)) {
        !           499:                        for (--i; i >= 0; i--) {
        !           500:                                SET(xbpp[i]->b_flags, B_INVAL);
        !           501:                                brelse(xbpp[i]);
        !           502:                        }
        !           503:                        free(xbpp, M_TEMP);
        !           504:                        return (biowait(*rbpp));
        !           505:                }
        !           506:                xbpp[i] = buf_stub(vp, blkno + i + 1);
        !           507:                if (xbpp[i] == NULL) {
        !           508:                        for (--i; i >= 0; i--) {
        !           509:                                SET(xbpp[i]->b_flags, B_INVAL);
        !           510:                                brelse(xbpp[i]);
        !           511:                        }
        !           512:                        free(xbpp, M_TEMP);
        !           513:                        return (biowait(*rbpp));
        !           514:                }
        !           515:        }
        !           516:
        !           517:        xbpp[howmany] = 0;
        !           518:
        !           519:        bp = getnewbuf(howmany * size, 0, 0, NULL);
        !           520:        if (bp == NULL) {
        !           521:                for (i = 0; i < howmany; i++) {
        !           522:                        SET(xbpp[i]->b_flags, B_INVAL);
        !           523:                        brelse(xbpp[i]);
        !           524:                }
        !           525:                free(xbpp, M_TEMP);
        !           526:                return (biowait(*rbpp));
        !           527:        }
        !           528:
        !           529:        inc = btodb(size);
        !           530:
        !           531:        for (i = 0; i < howmany; i++) {
        !           532:                SET(xbpp[i]->b_flags, B_READ | B_ASYNC);
        !           533:                binshash(xbpp[i], BUFHASH(vp, xbpp[i]->b_lblkno));
        !           534:                xbpp[i]->b_blkno = sblkno + (i * inc);
        !           535:                xbpp[i]->b_bufsize = xbpp[i]->b_bcount = size;
        !           536:                xbpp[i]->b_data = bp->b_data + (i * size);
        !           537:        }
        !           538:
        !           539:        bp->b_blkno = sblkno;
        !           540:        bp->b_lblkno = blkno + 1;
        !           541:        SET(bp->b_flags, B_READ | B_ASYNC | B_CALL);
        !           542:        bp->b_saveaddr = (void *)xbpp;
        !           543:        bp->b_iodone = bread_cluster_callback;
        !           544:        bp->b_vp = vp;
        !           545:        spill = bp->b_bufsize - bp->b_bcount;
        !           546:        if (spill) {
        !           547:                uvm_km_free(buf_map, (vaddr_t) bp->b_data + bp->b_bcount,
        !           548:                    spill);
        !           549:                numbufpages -= atop(spill);
        !           550:        }
        !           551:        VOP_STRATEGY(bp);
        !           552:        curproc->p_stats->p_ru.ru_inblock++;
        !           553:
        !           554:        return (biowait(*rbpp));
        !           555: }
        !           556:
        !           557: /*
        !           558:  * Block write.  Described in Bach (p.56)
        !           559:  */
        !           560: int
        !           561: bwrite(struct buf *bp)
        !           562: {
        !           563:        int rv, async, wasdelayed, s;
        !           564:        struct vnode *vp;
        !           565:        struct mount *mp;
        !           566:
        !           567:        vp = bp->b_vp;
        !           568:        if (vp != NULL)
        !           569:                mp = vp->v_type == VBLK? vp->v_specmountpoint : vp->v_mount;
        !           570:        else
        !           571:                mp = NULL;
        !           572:
        !           573:        /*
        !           574:         * Remember buffer type, to switch on it later.  If the write was
        !           575:         * synchronous, but the file system was mounted with MNT_ASYNC,
        !           576:         * convert it to a delayed write.
        !           577:         * XXX note that this relies on delayed tape writes being converted
        !           578:         * to async, not sync writes (which is safe, but ugly).
        !           579:         */
        !           580:        async = ISSET(bp->b_flags, B_ASYNC);
        !           581:        if (!async && mp && ISSET(mp->mnt_flag, MNT_ASYNC)) {
        !           582:                bdwrite(bp);
        !           583:                return (0);
        !           584:        }
        !           585:
        !           586:        /*
        !           587:         * Collect statistics on synchronous and asynchronous writes.
        !           588:         * Writes to block devices are charged to their associated
        !           589:         * filesystem (if any).
        !           590:         */
        !           591:        if (mp != NULL) {
        !           592:                if (async)
        !           593:                        mp->mnt_stat.f_asyncwrites++;
        !           594:                else
        !           595:                        mp->mnt_stat.f_syncwrites++;
        !           596:        }
        !           597:
        !           598:        wasdelayed = ISSET(bp->b_flags, B_DELWRI);
        !           599:        CLR(bp->b_flags, (B_READ | B_DONE | B_ERROR | B_DELWRI));
        !           600:
        !           601:        s = splbio();
        !           602:
        !           603:        /*
        !           604:         * If not synchronous, pay for the I/O operation and make
        !           605:         * sure the buf is on the correct vnode queue.  We have
        !           606:         * to do this now, because if we don't, the vnode may not
        !           607:         * be properly notified that its I/O has completed.
        !           608:         */
        !           609:        if (wasdelayed) {
        !           610:                reassignbuf(bp);
        !           611:        } else
        !           612:                curproc->p_stats->p_ru.ru_oublock++;
        !           613:
        !           614:
        !           615:        /* Initiate disk write.  Make sure the appropriate party is charged. */
        !           616:        bp->b_vp->v_numoutput++;
        !           617:        splx(s);
        !           618:        SET(bp->b_flags, B_WRITEINPROG);
        !           619:        VOP_STRATEGY(bp);
        !           620:
        !           621:        if (async)
        !           622:                return (0);
        !           623:
        !           624:        /*
        !           625:         * If I/O was synchronous, wait for it to complete.
        !           626:         */
        !           627:        rv = biowait(bp);
        !           628:
        !           629:        /* Release the buffer. */
        !           630:        brelse(bp);
        !           631:
        !           632:        return (rv);
        !           633: }
        !           634:
        !           635:
        !           636: /*
        !           637:  * Delayed write.
        !           638:  *
        !           639:  * The buffer is marked dirty, but is not queued for I/O.
        !           640:  * This routine should be used when the buffer is expected
        !           641:  * to be modified again soon, typically a small write that
        !           642:  * partially fills a buffer.
        !           643:  *
        !           644:  * NB: magnetic tapes cannot be delayed; they must be
        !           645:  * written in the order that the writes are requested.
        !           646:  *
        !           647:  * Described in Leffler, et al. (pp. 208-213).
        !           648:  */
        !           649: void
        !           650: bdwrite(struct buf *bp)
        !           651: {
        !           652:        int s;
        !           653:
        !           654:        /*
        !           655:         * If the block hasn't been seen before:
        !           656:         *      (1) Mark it as having been seen,
        !           657:         *      (2) Charge for the write.
        !           658:         *      (3) Make sure it's on its vnode's correct block list,
        !           659:         *      (4) If a buffer is rewritten, move it to end of dirty list
        !           660:         */
        !           661:        if (!ISSET(bp->b_flags, B_DELWRI)) {
        !           662:                SET(bp->b_flags, B_DELWRI);
        !           663:                bp->b_synctime = time_uptime + 35;
        !           664:                s = splbio();
        !           665:                reassignbuf(bp);
        !           666:                splx(s);
        !           667:                curproc->p_stats->p_ru.ru_oublock++;    /* XXX */
        !           668:        } else {
        !           669:                /*
        !           670:                 * see if this buffer has slacked through the syncer
        !           671:                 * and enforce an async write upon it.
        !           672:                 */
        !           673:                if (bp->b_synctime < time_uptime) {
        !           674:                        bawrite(bp);
        !           675:                        return;
        !           676:                }
        !           677:        }
        !           678:
        !           679:        /* If this is a tape block, write the block now. */
        !           680:        if (major(bp->b_dev) < nblkdev &&
        !           681:            bdevsw[major(bp->b_dev)].d_type == D_TAPE) {
        !           682:                bawrite(bp);
        !           683:                return;
        !           684:        }
        !           685:
        !           686:        /* Otherwise, the "write" is done, so mark and release the buffer. */
        !           687:        CLR(bp->b_flags, B_NEEDCOMMIT);
        !           688:        SET(bp->b_flags, B_DONE);
        !           689:        brelse(bp);
        !           690: }
        !           691:
        !           692: /*
        !           693:  * Asynchronous block write; just an asynchronous bwrite().
        !           694:  */
        !           695: void
        !           696: bawrite(struct buf *bp)
        !           697: {
        !           698:
        !           699:        SET(bp->b_flags, B_ASYNC);
        !           700:        VOP_BWRITE(bp);
        !           701: }
        !           702:
        !           703: /*
        !           704:  * Must be called at splbio()
        !           705:  */
        !           706: void
        !           707: buf_dirty(struct buf *bp)
        !           708: {
        !           709:        splassert(IPL_BIO);
        !           710:
        !           711:        if (ISSET(bp->b_flags, B_DELWRI) == 0) {
        !           712:                SET(bp->b_flags, B_DELWRI);
        !           713:                bp->b_synctime = time_uptime + 35;
        !           714:                reassignbuf(bp);
        !           715:        }
        !           716: }
        !           717:
        !           718: /*
        !           719:  * Must be called at splbio()
        !           720:  */
        !           721: void
        !           722: buf_undirty(struct buf *bp)
        !           723: {
        !           724:        splassert(IPL_BIO);
        !           725:
        !           726:        if (ISSET(bp->b_flags, B_DELWRI)) {
        !           727:                CLR(bp->b_flags, B_DELWRI);
        !           728:                reassignbuf(bp);
        !           729:        }
        !           730: }
        !           731:
        !           732: /*
        !           733:  * Release a buffer on to the free lists.
        !           734:  * Described in Bach (p. 46).
        !           735:  */
        !           736: void
        !           737: brelse(struct buf *bp)
        !           738: {
        !           739:        struct bqueues *bufq;
        !           740:        int s;
        !           741:
        !           742:        /* Block disk interrupts. */
        !           743:        s = splbio();
        !           744:
        !           745:        if (bp->b_data != NULL)
        !           746:                KASSERT(bp->b_bufsize > 0);
        !           747:
        !           748:        /*
        !           749:         * Determine which queue the buffer should be on, then put it there.
        !           750:         */
        !           751:
        !           752:        /* If it's not cacheable, or an error, mark it invalid. */
        !           753:        if (ISSET(bp->b_flags, (B_NOCACHE|B_ERROR)))
        !           754:                SET(bp->b_flags, B_INVAL);
        !           755:
        !           756:        if (ISSET(bp->b_flags, B_INVAL)) {
        !           757:                int queue, qs;
        !           758:
        !           759:                /*
        !           760:                 * If the buffer is invalid, place it in the clean queue, so it
        !           761:                 * can be reused.
        !           762:                 */
        !           763:                if (LIST_FIRST(&bp->b_dep) != NULL)
        !           764:                        buf_deallocate(bp);
        !           765:
        !           766:                if (ISSET(bp->b_flags, B_DELWRI)) {
        !           767:                        CLR(bp->b_flags, B_DELWRI);
        !           768:                }
        !           769:
        !           770:                if (bp->b_vp)
        !           771:                        brelvp(bp);
        !           772:
        !           773:                /*
        !           774:                 * If the buffer has no associated data, place it back in the
        !           775:                 * pool.
        !           776:                 */
        !           777:                if (bp->b_data == NULL) {
        !           778:                        buf_put(bp);
        !           779:                        splx(s);
        !           780:                        return;
        !           781:                }
        !           782:
        !           783:                qs = bp->b_bufsize;
        !           784:                queue = size2cqueue(&qs);
        !           785:                numcleanpages += btoc(bp->b_bufsize);
        !           786:                bqpages[queue] += btoc(bp->b_bufsize);
        !           787:                if (maxcleanpages < numcleanpages)
        !           788:                        maxcleanpages = numcleanpages;
        !           789:                binsheadfree(bp, &bufqueues[queue]);
        !           790:        } else {
        !           791:                /*
        !           792:                 * It has valid data.  Put it on the end of the appropriate
        !           793:                 * queue, so that it'll stick around for as long as possible.
        !           794:                 */
        !           795:                int queue, qs;
        !           796:                numfreepages += btoc(bp->b_bufsize);
        !           797:                qs = bp->b_bufsize;
        !           798:                queue = size2cqueue(&qs);
        !           799:
        !           800:                if (!ISSET(bp->b_flags, B_DELWRI)) {
        !           801:                        numcleanpages += btoc(bp->b_bufsize);
        !           802:                        bqpages[queue] += btoc(bp->b_bufsize);
        !           803:                        if (maxcleanpages < numcleanpages)
        !           804:                                maxcleanpages = numcleanpages;
        !           805:                        bufq = &bufqueues[queue];
        !           806:                } else {
        !           807:                        numdirtypages += btoc(bp->b_bufsize);
        !           808:                        bufq = &bufqueues[BQ_DIRTY];
        !           809:                }
        !           810:                if (ISSET(bp->b_flags, B_AGE)) {
        !           811:                        binsheadfree(bp, bufq);
        !           812:                        bp->b_synctime = time_uptime + 30;
        !           813:                } else {
        !           814:                        binstailfree(bp, bufq);
        !           815:                        bp->b_synctime = time_uptime + 300;
        !           816:                }
        !           817:        }
        !           818:
        !           819:        /* Unlock the buffer. */
        !           820:        CLR(bp->b_flags, (B_AGE | B_ASYNC | B_BUSY | B_NOCACHE | B_DEFERRED));
        !           821:
        !           822:        /* Wake up any processes waiting for any buffer to become free. */
        !           823:        if (needbuffer) {
        !           824:                needbuffer--;
        !           825:                wakeup_one(&needbuffer);
        !           826:        }
        !           827:
        !           828:        /* Wake up any processes waiting for _this_ buffer to become free. */
        !           829:        if (ISSET(bp->b_flags, B_WANTED)) {
        !           830:                CLR(bp->b_flags, B_WANTED);
        !           831:                wakeup(bp);
        !           832:        }
        !           833:
        !           834:        splx(s);
        !           835: }
        !           836:
        !           837: /*
        !           838:  * Determine if a block is in the cache. Just look on what would be its hash
        !           839:  * chain. If it's there, return a pointer to it, unless it's marked invalid.
        !           840:  */
        !           841: struct buf *
        !           842: incore(struct vnode *vp, daddr64_t blkno)
        !           843: {
        !           844:        struct buf *bp;
        !           845:
        !           846:        /* Search hash chain */
        !           847:        LIST_FOREACH(bp, BUFHASH(vp, blkno), b_hash) {
        !           848:                if (bp->b_lblkno == blkno && bp->b_vp == vp &&
        !           849:                    !ISSET(bp->b_flags, B_INVAL))
        !           850:                        return (bp);
        !           851:        }
        !           852:
        !           853:        return (NULL);
        !           854: }
        !           855:
        !           856: /*
        !           857:  * Get a block of requested size that is associated with
        !           858:  * a given vnode and block offset. If it is found in the
        !           859:  * block cache, mark it as having been found, make it busy
        !           860:  * and return it. Otherwise, return an empty block of the
        !           861:  * correct size. It is up to the caller to ensure that the
        !           862:  * cached blocks be of the correct size.
        !           863:  */
        !           864: struct buf *
        !           865: getblk(struct vnode *vp, daddr64_t blkno, int size, int slpflag, int slptimeo)
        !           866: {
        !           867:        struct bufhashhdr *bh;
        !           868:        struct buf *bp, *nb = NULL;
        !           869:        int s, error;
        !           870:
        !           871:        /*
        !           872:         * XXX
        !           873:         * The following is an inlined version of 'incore()', but with
        !           874:         * the 'invalid' test moved to after the 'busy' test.  It's
        !           875:         * necessary because there are some cases in which the NFS
        !           876:         * code sets B_INVAL prior to writing data to the server, but
        !           877:         * in which the buffers actually contain valid data.  In this
        !           878:         * case, we can't allow the system to allocate a new buffer for
        !           879:         * the block until the write is finished.
        !           880:         */
        !           881:        bh = BUFHASH(vp, blkno);
        !           882: start:
        !           883:        LIST_FOREACH(bp, BUFHASH(vp, blkno), b_hash) {
        !           884:                if (bp->b_lblkno != blkno || bp->b_vp != vp)
        !           885:                        continue;
        !           886:
        !           887:                s = splbio();
        !           888:                if (ISSET(bp->b_flags, B_BUSY)) {
        !           889:                        if (nb != NULL) {
        !           890:                                SET(nb->b_flags, B_INVAL);
        !           891:                                binshash(nb, &invalhash);
        !           892:                                brelse(nb);
        !           893:                                nb = NULL;
        !           894:                        }
        !           895:                        SET(bp->b_flags, B_WANTED);
        !           896:                        error = tsleep(bp, slpflag | (PRIBIO + 1), "getblk",
        !           897:                            slptimeo);
        !           898:                        splx(s);
        !           899:                        if (error)
        !           900:                                return (NULL);
        !           901:                        goto start;
        !           902:                }
        !           903:
        !           904:                if (!ISSET(bp->b_flags, B_INVAL)) {
        !           905:                        SET(bp->b_flags, (B_BUSY | B_CACHE));
        !           906:                        bremfree(bp);
        !           907:                        splx(s);
        !           908:                        break;
        !           909:                }
        !           910:                splx(s);
        !           911:        }
        !           912:        if (nb && bp) {
        !           913:                SET(nb->b_flags, B_INVAL);
        !           914:                binshash(nb, &invalhash);
        !           915:                brelse(nb);
        !           916:                nb = NULL;
        !           917:        }
        !           918:        if (bp == NULL && nb == NULL) {
        !           919:                nb = getnewbuf(size, slpflag, slptimeo, &error);
        !           920:                if (nb == NULL) {
        !           921:                        if (error == ERESTART || error == EINTR)
        !           922:                                return (NULL);
        !           923:                }
        !           924:                goto start;
        !           925:        }
        !           926:        if (nb) {
        !           927:                bp = nb;
        !           928:                binshash(bp, bh);
        !           929:                bp->b_blkno = bp->b_lblkno = blkno;
        !           930:                s = splbio();
        !           931:                bgetvp(vp, bp);
        !           932:                splx(s);
        !           933:        }
        !           934:        return (bp);
        !           935: }
        !           936:
        !           937: /*
        !           938:  * Get an empty, disassociated buffer of given size.
        !           939:  */
        !           940: struct buf *
        !           941: geteblk(int size)
        !           942: {
        !           943:        struct buf *bp;
        !           944:
        !           945:        while ((bp = getnewbuf(size, 0, 0, NULL)) == NULL)
        !           946:                ;
        !           947:        SET(bp->b_flags, B_INVAL);
        !           948:        binshash(bp, &invalhash);
        !           949:
        !           950:        return (bp);
        !           951: }
        !           952:
        !           953: /*
        !           954:  * Find a buffer which is available for use.
        !           955:  */
        !           956: struct buf *
        !           957: getnewbuf(size_t size, int slpflag, int slptimeo, int *ep)
        !           958: {
        !           959:        struct buf *bp;
        !           960:        int s, error, queue, qs;
        !           961:
        !           962: #if 0          /* we would really like this but sblock update kills it */
        !           963:        KASSERT(curproc != syncerproc && curproc != cleanerproc);
        !           964: #endif
        !           965:
        !           966:        s = splbio();
        !           967:        /*
        !           968:         * Wake up cleaner if we're getting low on pages.
        !           969:         */
        !           970:        if (numdirtypages >= hidirtypages || numcleanpages <= locleanpages)
        !           971:                wakeup(&bd_req);
        !           972:
        !           973:        /* we just ask. it can say no.. */
        !           974: getsome:
        !           975:        qs = size;
        !           976:        queue = size2cqueue(&qs);
        !           977:        bp = buf_get(qs); /* XXX use qs instead and no need in buf_get? */
        !           978:        if (bp == NULL) {
        !           979:                /*
        !           980:                 * No free ones, try to reuse a clean one of the same or
        !           981:                 * larger size.
        !           982:                 */
        !           983:                do {
        !           984:                        bp = TAILQ_FIRST(&bufqueues[queue]);
        !           985:                        queue++;
        !           986:                } while (bp == NULL && queue < BQUEUES);
        !           987:        }
        !           988:        if (bp == NULL) {
        !           989:                /* we couldn't reuse a free one, nothing of the right size */
        !           990:                /* XXX free 20 buffers per q - ugly hack  should really
        !           991:                 * reuse big ones without truncating. fix later
        !           992:                 */
        !           993:                int q, gotsome = 0;
        !           994:                int freemax = 20;
        !           995:                for (q = 1; q < BQUEUES; q++) {
        !           996:                        int i = freemax;
        !           997:                        while (bqpages[q] > bqpagelow
        !           998:                            && (bp = TAILQ_FIRST(&bufqueues[q]))
        !           999:                            && i--) {
        !          1000:                                gotsome++;
        !          1001:                                bremfree(bp);
        !          1002:                                if (LIST_FIRST(&bp->b_dep) != NULL)
        !          1003:                                        buf_deallocate(bp);
        !          1004:
        !          1005:                                if (ISSET(bp->b_flags, B_DELWRI)) {
        !          1006:                                        CLR(bp->b_flags, B_DELWRI);
        !          1007:                                }
        !          1008:
        !          1009:                                if (bp->b_vp)
        !          1010:                                        brelvp(bp);
        !          1011:
        !          1012:                                buf_put(bp);
        !          1013:                        }
        !          1014:                }
        !          1015:                if (gotsome)
        !          1016:                        goto getsome;
        !          1017:        }
        !          1018:        if (bp == NULL) {
        !          1019:                /* wait for a free buffer of any kind */
        !          1020:                needbuffer++;
        !          1021:                error = tsleep(&needbuffer, slpflag | (PRIBIO + 1),
        !          1022:                    "getnewbuf", slptimeo);
        !          1023:                if (ep != NULL) {
        !          1024:                        *ep = error;
        !          1025:                        if (error) {
        !          1026:                                splx(s);
        !          1027:                                return (NULL);
        !          1028:                        }
        !          1029:                }
        !          1030:                goto getsome;
        !          1031:        }
        !          1032:
        !          1033:        bremfree(bp);
        !          1034:        /* Buffer is no longer on free lists. */
        !          1035:        SET(bp->b_flags, B_BUSY);
        !          1036:
        !          1037: #ifdef DIAGNOSTIC
        !          1038:        if (ISSET(bp->b_flags, B_DELWRI))
        !          1039:                panic("Dirty buffer on BQ_CLEAN");
        !          1040: #endif
        !          1041:
        !          1042:        /* disassociate us from our vnode, if we had one... */
        !          1043:        if (bp->b_vp)
        !          1044:                brelvp(bp);
        !          1045:
        !          1046:        splx(s);
        !          1047:
        !          1048: #ifdef DIAGNOSTIC
        !          1049:        /* CLEAN buffers must have no dependencies */
        !          1050:        if (LIST_FIRST(&bp->b_dep) != NULL)
        !          1051:                panic("BQ_CLEAN has buffer with dependencies");
        !          1052: #endif
        !          1053:
        !          1054:        /* clear out various other fields */
        !          1055:        bp->b_flags = B_BUSY;
        !          1056:        bp->b_dev = NODEV;
        !          1057:        bp->b_blkno = bp->b_lblkno = 0;
        !          1058:        bp->b_iodone = NULL;
        !          1059:        bp->b_error = 0;
        !          1060:        bp->b_resid = 0;
        !          1061:        bp->b_bcount = size;
        !          1062:        bp->b_dirtyoff = bp->b_dirtyend = 0;
        !          1063:        bp->b_validoff = bp->b_validend = 0;
        !          1064:
        !          1065:        bremhash(bp);
        !          1066:        return (bp);
        !          1067: }
        !          1068:
        !          1069: /*
        !          1070:  * Buffer cleaning daemon.
        !          1071:  */
        !          1072: void
        !          1073: buf_daemon(struct proc *p)
        !          1074: {
        !          1075:        struct timeval starttime, timediff;
        !          1076:        struct buf *bp;
        !          1077:        int s;
        !          1078:
        !          1079:        cleanerproc = curproc;
        !          1080:
        !          1081:        s = splbio();
        !          1082:        for (;;) {
        !          1083:                if (!numdirtypages ||
        !          1084:                    (numdirtypages < hidirtypages && !needbuffer))
        !          1085:                        tsleep(&bd_req, PRIBIO - 7, "cleaner", 0);
        !          1086:
        !          1087:                getmicrouptime(&starttime);
        !          1088:
        !          1089:                while ((bp = TAILQ_FIRST(&bufqueues[BQ_DIRTY]))) {
        !          1090:                        struct timeval tv;
        !          1091:
        !          1092:                        if (numdirtypages < lodirtypages && !needbuffer)
        !          1093:                                break;
        !          1094:
        !          1095:                        bremfree(bp);
        !          1096:                        SET(bp->b_flags, B_BUSY);
        !          1097:                        splx(s);
        !          1098:
        !          1099:                        if (ISSET(bp->b_flags, B_INVAL)) {
        !          1100:                                brelse(bp);
        !          1101:                                s = splbio();
        !          1102:                                continue;
        !          1103:                        }
        !          1104: #ifdef DIAGNOSTIC
        !          1105:                        if (!ISSET(bp->b_flags, B_DELWRI))
        !          1106:                                panic("Clean buffer on BQ_DIRTY");
        !          1107: #endif
        !          1108:                        if (LIST_FIRST(&bp->b_dep) != NULL &&
        !          1109:                            !ISSET(bp->b_flags, B_DEFERRED) &&
        !          1110:                            buf_countdeps(bp, 0, 0)) {
        !          1111:                                SET(bp->b_flags, B_DEFERRED);
        !          1112:                                s = splbio();
        !          1113:                                numfreepages += btoc(bp->b_bufsize);
        !          1114:                                numdirtypages += btoc(bp->b_bufsize);
        !          1115:                                binstailfree(bp, &bufqueues[BQ_DIRTY]);
        !          1116:                                CLR(bp->b_flags, B_BUSY);
        !          1117:                                continue;
        !          1118:                        }
        !          1119:
        !          1120:                        bawrite(bp);
        !          1121:
        !          1122:                        /* Never allow processing to run for more than 1 sec */
        !          1123:                        getmicrouptime(&tv);
        !          1124:                        timersub(&tv, &starttime, &timediff);
        !          1125:                        if (timediff.tv_sec)
        !          1126:                                break;
        !          1127:
        !          1128:                        s = splbio();
        !          1129:                }
        !          1130:        }
        !          1131: }
        !          1132:
        !          1133: /*
        !          1134:  * Wait for operations on the buffer to complete.
        !          1135:  * When they do, extract and return the I/O's error value.
        !          1136:  */
        !          1137: int
        !          1138: biowait(struct buf *bp)
        !          1139: {
        !          1140:        int s;
        !          1141:
        !          1142:        s = splbio();
        !          1143:        while (!ISSET(bp->b_flags, B_DONE))
        !          1144:                tsleep(bp, PRIBIO + 1, "biowait", 0);
        !          1145:        splx(s);
        !          1146:
        !          1147:        /* check for interruption of I/O (e.g. via NFS), then errors. */
        !          1148:        if (ISSET(bp->b_flags, B_EINTR)) {
        !          1149:                CLR(bp->b_flags, B_EINTR);
        !          1150:                return (EINTR);
        !          1151:        }
        !          1152:
        !          1153:        if (ISSET(bp->b_flags, B_ERROR))
        !          1154:                return (bp->b_error ? bp->b_error : EIO);
        !          1155:        else
        !          1156:                return (0);
        !          1157: }
        !          1158:
        !          1159: /*
        !          1160:  * Mark I/O complete on a buffer.
        !          1161:  *
        !          1162:  * If a callback has been requested, e.g. the pageout
        !          1163:  * daemon, do so. Otherwise, awaken waiting processes.
        !          1164:  *
        !          1165:  * [ Leffler, et al., says on p.247:
        !          1166:  *     "This routine wakes up the blocked process, frees the buffer
        !          1167:  *     for an asynchronous write, or, for a request by the pagedaemon
        !          1168:  *     process, invokes a procedure specified in the buffer structure" ]
        !          1169:  *
        !          1170:  * In real life, the pagedaemon (or other system processes) wants
        !          1171:  * to do async stuff to, and doesn't want the buffer brelse()'d.
        !          1172:  * (for swap pager, that puts swap buffers on the free lists (!!!),
        !          1173:  * for the vn device, that puts malloc'd buffers on the free lists!)
        !          1174:  *
        !          1175:  * Must be called at splbio().
        !          1176:  */
        !          1177: void
        !          1178: biodone(struct buf *bp)
        !          1179: {
        !          1180:        splassert(IPL_BIO);
        !          1181:
        !          1182:        if (ISSET(bp->b_flags, B_DONE))
        !          1183:                panic("biodone already");
        !          1184:        SET(bp->b_flags, B_DONE);               /* note that it's done */
        !          1185:
        !          1186:        if (LIST_FIRST(&bp->b_dep) != NULL)
        !          1187:                buf_complete(bp);
        !          1188:
        !          1189:        if (!ISSET(bp->b_flags, B_READ)) {
        !          1190:                CLR(bp->b_flags, B_WRITEINPROG);
        !          1191:                vwakeup(bp->b_vp);
        !          1192:        }
        !          1193:
        !          1194:        if (ISSET(bp->b_flags, B_CALL)) {       /* if necessary, call out */
        !          1195:                CLR(bp->b_flags, B_CALL);       /* but note callout done */
        !          1196:                (*bp->b_iodone)(bp);
        !          1197:        } else {
        !          1198:                if (ISSET(bp->b_flags, B_ASYNC)) {/* if async, release it */
        !          1199:                        brelse(bp);
        !          1200:                } else {                        /* or just wakeup the buffer */
        !          1201:                        CLR(bp->b_flags, B_WANTED);
        !          1202:                        wakeup(bp);
        !          1203:                }
        !          1204:        }
        !          1205: }
        !          1206:
        !          1207: #if 1
        !          1208: void
        !          1209: vfs_bufstats(void) {
        !          1210:        return;
        !          1211: }
        !          1212: /* #ifdef DDB */
        !          1213: #else
        !          1214: /*
        !          1215:  * Print out statistics on the current allocation of the buffer pool.
        !          1216:  * Can be enabled to print out on every ``sync'' by setting "syncprt"
        !          1217:  * in vfs_syscalls.c using sysctl.
        !          1218:  */
        !          1219: void
        !          1220: vfs_bufstats(void)
        !          1221: {
        !          1222:        int s, i, j, count;
        !          1223:        struct buf *bp;
        !          1224:        struct bqueues *dp;
        !          1225:        int counts[MAXBSIZE/PAGE_SIZE+1];
        !          1226:        int totals[BQUEUES];
        !          1227:        long ptotals[BQUEUES];
        !          1228:        long pages;
        !          1229:        static char *bname[BQUEUES] = { "CLEAN", "DIRTY", "EMPTY" };
        !          1230:
        !          1231:        s = splbio();
        !          1232:        for (dp = bufqueues, i = 0; dp < &bufqueues[BQUEUES]; dp++, i++) {
        !          1233:                count = 0;
        !          1234:                pages = 0;
        !          1235:                for (j = 0; j <= MAXBSIZE/PAGE_SIZE; j++)
        !          1236:                        counts[j] = 0;
        !          1237:                TAILQ_FOREACH(bp, dp, b_freelist) {
        !          1238:                        counts[bp->b_bufsize/PAGE_SIZE]++;
        !          1239:                        count++;
        !          1240:                        pages += btoc(bp->b_bufsize);
        !          1241:                }
        !          1242:                totals[i] = count;
        !          1243:                ptotals[i] = pages;
        !          1244:                printf("%s: total-%d(%d pages)", bname[i], count, pages);
        !          1245:                for (j = 0; j <= MAXBSIZE/PAGE_SIZE; j++)
        !          1246:                        if (counts[j] != 0)
        !          1247:                                printf(", %d-%d", j * PAGE_SIZE, counts[j]);
        !          1248:                printf("\n");
        !          1249:        }
        !          1250:        if ((ptotals[BQ_CLEAN] + ptotals[BQ_DIRTY]) != numfreepages)
        !          1251:                printf("numfreepages counter wrong: %ld != %ld\n",
        !          1252:                    numfreepages, ptotals[BQ_CLEAN] + ptotals[BQ_DIRTY]);
        !          1253:        if (ptotals[BQ_CLEAN] != numcleanpages)
        !          1254:                printf("numcleanpages counter wrong: %ld != %ld\n",
        !          1255:                    numcleanpages, ptotals[<BQ_CLEAN]);
        !          1256:        else
        !          1257:                printf("numcleanpages: %ld\n", numcleanpages);
        !          1258:        if (numdirtypages != ptotals[BQ_DIRTY])
        !          1259:                printf("numdirtypages counter wrong: %ld != %ld\n",
        !          1260:                    numdirtypages, ptotals[BQ_DIRTY]);
        !          1261:        else
        !          1262:                printf("numdirtypages: %ld\n", numdirtypages);
        !          1263:
        !          1264:        printf("syncer eating up to %ld pages from %ld reserved\n",
        !          1265:            maxcleanpages - hicleanpages, locleanpages);
        !          1266:        splx(s);
        !          1267: }
        !          1268: #endif /* DEBUG */

CVSweb