Index: cam/ctl/ctl_backend_block.c =================================================================== --- cam/ctl/ctl_backend_block.c (revision 249436) +++ cam/ctl/ctl_backend_block.c (working copy) @@ -610,7 +610,7 @@ ctl_be_block_flush_file(struct ctl_be_block_lun *b vn_lock(be_lun->vn, lock_flags | LK_RETRY); - binuptime(&beio->ds_t0); + getbinuptime(&beio->ds_t0); devstat_start_transaction(beio->lun->disk_stats, &beio->ds_t0); error = VOP_FSYNC(be_lun->vn, MNT_WAIT, curthread); @@ -679,7 +679,7 @@ ctl_be_block_dispatch_file(struct ctl_be_block_lun if (beio->bio_cmd == BIO_READ) { vn_lock(be_lun->vn, LK_SHARED | LK_RETRY); - binuptime(&beio->ds_t0); + getbinuptime(&beio->ds_t0); devstat_start_transaction(beio->lun->disk_stats, &beio->ds_t0); /* @@ -721,7 +721,7 @@ ctl_be_block_dispatch_file(struct ctl_be_block_lun vn_lock(be_lun->vn, lock_flags | LK_RETRY); - binuptime(&beio->ds_t0); + getbinuptime(&beio->ds_t0); devstat_start_transaction(beio->lun->disk_stats, &beio->ds_t0); /* @@ -827,7 +827,7 @@ ctl_be_block_flush_dev(struct ctl_be_block_lun *be beio->num_bios_sent = 1; beio->send_complete = 1; - binuptime(&beio->ds_t0); + getbinuptime(&beio->ds_t0); devstat_start_transaction(be_lun->disk_stats, &beio->ds_t0); (*dev_data->csw->d_strategy)(bio); @@ -862,7 +862,7 @@ ctl_be_block_dispatch_dev(struct ctl_be_block_lun * XXX KDM need to accurately reflect the number of I/Os outstanding * to a device. */ - binuptime(&beio->ds_t0); + getbinuptime(&beio->ds_t0); devstat_start_transaction(be_lun->disk_stats, &beio->ds_t0); for (i = 0; i < beio->num_segs; i++) { Index: dev/xen/blkback/blkback.c =================================================================== --- dev/xen/blkback/blkback.c (revision 249436) +++ dev/xen/blkback/blkback.c (working copy) @@ -1248,7 +1248,7 @@ xbb_get_resources(struct xbb_softc *xbb, struct xb nreq->ring_req = ring_req; } - binuptime(&nreq->ds_t0); + getbinuptime(&nreq->ds_t0); devstat_start_transaction(xbb->xbb_stats_in, &nreq->ds_t0); STAILQ_INSERT_TAIL(&(*reqlist)->contig_req_list, nreq, links); (*reqlist)->num_children++; @@ -1538,7 +1538,7 @@ xbb_dispatch_io(struct xbb_softc *xbb, struct xbb_ } } - binuptime(&reqlist->ds_t0); + getbinuptime(&reqlist->ds_t0); devstat_start_transaction(xbb->xbb_stats, &reqlist->ds_t0); switch (reqlist->operation) { Index: geom/gate/g_gate.c =================================================================== --- geom/gate/g_gate.c (revision 249436) +++ geom/gate/g_gate.c (working copy) @@ -339,7 +339,7 @@ g_gate_guard(void *arg) struct bio *bp, *bp2; sc = arg; - binuptime(&curtime); + getbinuptime(&curtime); g_gate_hold(sc->sc_unit, NULL); mtx_lock(&sc->sc_queue_mtx); TAILQ_FOREACH_SAFE(bp, &sc->sc_inqueue.queue, bio_queue, bp2) { Index: geom/geom_io.c =================================================================== --- geom/geom_io.c (revision 249436) +++ geom/geom_io.c (working copy) @@ -473,10 +473,7 @@ g_io_request(struct bio *bp, struct g_consumer *cp ("Bio already on queue bp=%p", bp)); bp->bio_flags |= BIO_ONQUEUE; - if (g_collectstats) - binuptime(&bp->bio_t0); - else - getbinuptime(&bp->bio_t0); + getbinuptime(&bp->bio_t0); /* * The statistics collection is lockless, as such, but we Index: kern/subr_devstat.c =================================================================== --- kern/subr_devstat.c (revision 249436) +++ kern/subr_devstat.c (working copy) @@ -277,7 +277,7 @@ devstat_start_transaction(struct devstat *ds, stru if (now != NULL) ds->busy_from = *now; else - binuptime(&ds->busy_from); + getbinuptime(&ds->busy_from); } ds->start_count++; atomic_add_rel_int(&ds->sequence0, 1); @@ -294,7 +294,7 @@ devstat_start_transaction_bio(struct devstat *ds, if (ds == NULL) return; - binuptime(&bp->bio_t0); + getbinuptime(&bp->bio_t0); devstat_start_transaction(ds, &bp->bio_t0); DTRACE_DEVSTAT_BIO_START(); } @@ -338,7 +338,7 @@ devstat_end_transaction(struct devstat *ds, uint32 if (now == NULL) { now = &lnow; - binuptime(now); + getbinuptime(now); } atomic_add_acq_int(&ds->sequence1, 1);