--- sys/geom/geom_kern.c.prev 2010-06-16 19:47:35.000000000 +0300 +++ sys/geom/geom_kern.c 2010-06-16 19:47:23.000000000 +0300 @@ -139,7 +139,7 @@ g_event_procbody(void) thread_unlock(tp); for(;;) { g_run_events(); - tsleep(&g_wait_event, PRIBIO, "-", hz/10); + tsleep(&g_wait_event, PRIBIO, "-", hz); } } --- sys/dev/random/randomdev_soft.c.prev 2010-06-14 15:29:07.000000000 +0300 +++ sys/dev/random/randomdev_soft.c 2010-06-16 19:16:56.000000000 +0300 @@ -283,7 +283,7 @@ random_kthread(void *arg __unused) random_kthread_control = 0; /* Work done, so don't belabour the issue */ - pause("-", hz / 10); + pause("-", hz); } --- sys/dev/syscons/syscons.c.prev 2010-05-22 11:28:08.000000000 +0300 +++ sys/dev/syscons/syscons.c 2010-08-20 18:31:25.000000000 +0300 @@ -1812,7 +1812,7 @@ scrn_timer(void *arg) #endif if (again) - timeout(scrn_timer, sc, hz / 25); + timeout(scrn_timer, sc, hz / 10); splx(s); } --- sys/kern/kern_resource.c.prev 2010-07-19 05:19:38.000000000 +0300 +++ sys/kern/kern_resource.c 2010-08-20 18:49:05.000000000 +0300 @@ -644,7 +644,7 @@ } } if ((p->p_flag & P_WEXIT) == 0) - callout_reset(&p->p_limco, hz, lim_cb, p); + callout_reset(&p->p_limco, hz * 10, lim_cb, p); } int --- sys/netinet/tcp_subr.c.prev 2010-08-19 20:29:49.000000000 +0300 +++ sys/netinet/tcp_subr.c 2010-08-20 18:22:30.000000000 +0300 @@ -1571,7 +1571,7 @@ tcp_isn_tick(void *xtp) VNET_FOREACH(vnet_iter) { CURVNET_SET(vnet_iter); /* XXX appease INVARIANTS */ projected_offset = - V_isn_offset_old + ISN_BYTES_PER_SECOND / 100; + V_isn_offset_old + ISN_BYTES_PER_SECOND / 10; if (SEQ_GT(projected_offset, V_isn_offset)) V_isn_offset = projected_offset; @@ -1581,7 +1581,7 @@ tcp_isn_tick(void *xtp) } ISN_UNLOCK(); VNET_LIST_RUNLOCK_NOSLEEP(); - callout_reset(&isn_callout, hz/100, tcp_isn_tick, NULL); + callout_reset(&isn_callout, hz/10, tcp_isn_tick, NULL); } /* --- sys/kern/sched_ule.c 2010-08-19 20:29:49.000000000 +0300 +++ sys/kern/sched_ule.c 2010-08-26 13:30:10.000000000 +0300 @@ -1156,6 +1160,12 @@ sched_pickcpu(struct thread *td, int fla SCHED_STAT_INC(pickcpu_idle_affinity); return (ts->ts_cpu); } + if (ts->ts_cpu == self && + tdq->tdq_lowpri == ((4 * RQ_PPQ) + PI_SOFT) && + tdq->tdq_load == 1) { + SCHED_STAT_INC(pickcpu_idle_affinity); + return (ts->ts_cpu); + } if (SCHED_AFFINITY(ts, CG_SHARE_L2) && tdq->tdq_lowpri > pri) { SCHED_STAT_INC(pickcpu_affinity); return (ts->ts_cpu);