// Copyright 2009 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. // Time-related runtime and pieces of package time. package runtime import ( "internal/abi" "internal/runtime/atomic" "runtime/internal/sys" "unsafe" ) // A timer is a potentially repeating trigger for calling t.f(t.arg, t.seq). // Timers are allocated by client code, often as part of other data structures. // Each P has a heap of pointers to timers that it manages. // // A timer is expected to be used by only one client goroutine at a time, // but there will be concurrent access by the P managing that timer. // Timer accesses are protected by the lock t.mu, with a snapshot of // t's state bits published in t.astate to enable certain fast paths to make // decisions about a timer without acquiring the lock. type timer struct { // mu protects reads and writes to all fields, with exceptions noted below. mu mutex astate atomic.Uint8 // atomic copy of state bits at last unlock state uint8 // state bits isChan bool // timer has a channel; immutable; can be read without lock blocked uint32 // number of goroutines blocked on timer's channel // Timer wakes up at when, and then at when+period, ... (period > 0 only) // each time calling f(arg, seq, delay) in the timer goroutine, so f must be // a well-behaved function and not block. // // The arg and seq are client-specified opaque arguments passed back to f. // When used from netpoll, arg and seq have meanings defined by netpoll // and are completely opaque to this code; in that context, seq is a sequence // number to recognize and squech stale function invocations. // When used from package time, arg is a channel (for After, NewTicker) // or the function to call (for AfterFunc) and seq is unused (0). // // Package time does not know about seq, but if this is a channel timer (t.isChan == true), // this file uses t.seq as a sequence number to recognize and squelch // sends that correspond to an earlier (stale) timer configuration, // similar to its use in netpoll. In this usage (that is, when t.isChan == true), // writes to seq are protected by both t.mu and t.sendLock, // so reads are allowed when holding either of the two mutexes. // // The delay argument is nanotime() - t.when, meaning the delay in ns between // when the timer should have gone off and now. Normally that amount is // small enough not to matter, but for channel timers that are fed lazily, // the delay can be arbitrarily long; package time subtracts it out to make // it look like the send happened earlier than it actually did. // (No one looked at the channel since then, or the send would have // not happened so late, so no one can tell the difference.) when int64 period int64 f func(arg any, seq uintptr, delay int64) arg any seq uintptr // If non-nil, the timers containing t. ts *timers // sendLock protects sends on the timer's channel. // Not used for async (pre-Go 1.23) behavior when debug.asynctimerchan.Load() != 0. sendLock mutex } // init initializes a newly allocated timer t. // Any code that allocates a timer must call t.init before using it. // The arg and f can be set during init, or they can be nil in init // and set by a future call to t.modify. func (t *timer) init(f func(arg any, seq uintptr, delay int64), arg any) { lockInit(&t.mu, lockRankTimer) t.f = f t.arg = arg } // A timers is a per-P set of timers. type timers struct { // mu protects timers; timers are per-P, but the scheduler can // access the timers of another P, so we have to lock. mu mutex // heap is the set of timers, ordered by heap[i].when. // Must hold lock to access. heap []timerWhen // len is an atomic copy of len(heap). len atomic.Uint32 // zombies is the number of timers in the heap // that are marked for removal. zombies atomic.Int32 // raceCtx is the race context used while executing timer functions. raceCtx uintptr // minWhenHeap is the minimum heap[i].when value (= heap[0].when). // The wakeTime method uses minWhenHeap and minWhenModified // to determine the next wake time. // If minWhenHeap = 0, it means there are no timers in the heap. minWhenHeap atomic.Int64 // minWhenModified is a lower bound on the minimum // heap[i].when over timers with the timerModified bit set. // If minWhenModified = 0, it means there are no timerModified timers in the heap. minWhenModified atomic.Int64 } type timerWhen struct { timer *timer when int64 } func (ts *timers) lock() { lock(&ts.mu) } func (ts *timers) unlock() { // Update atomic copy of len(ts.heap). // We only update at unlock so that the len is always // the most recent unlocked length, not an ephemeral length. // This matters if we lock ts, delete the only timer from the heap, // add it back, and unlock. We want ts.len.Load to return 1 the // entire time, never 0. This is important for pidleput deciding // whether ts is empty. ts.len.Store(uint32(len(ts.heap))) unlock(&ts.mu) } // Timer state field. const ( // timerHeaped is set when the timer is stored in some P's heap. timerHeaped uint8 = 1 << iota // timerModified is set when t.when has been modified // but the heap's heap[i].when entry still needs to be updated. // That change waits until the heap in which // the timer appears can be locked and rearranged. // timerModified is only set when timerHeaped is also set. timerModified // timerZombie is set when the timer has been stopped // but is still present in some P's heap. // Only set when timerHeaped is also set. // It is possible for timerModified and timerZombie to both // be set, meaning that the timer was modified and then stopped. // A timer sending to a channel may be placed in timerZombie // to take it out of the heap even though the timer is not stopped, // as long as nothing is reading from the channel. timerZombie ) // timerDebug enables printing a textual debug trace of all timer operations to stderr. const timerDebug = false func (t *timer) trace(op string) { if timerDebug { t.trace1(op) } } func (t *timer) trace1(op string) { if !timerDebug { return } bits := [4]string{"h", "m", "z", "c"} for i := range 3 { if t.state&(1< 0 { // If timer should have triggered already (but nothing looked at it yet), // trigger now, so that a receive after the stop sees the "old" value // that should be there. // (It is possible to have t.blocked > 0 if there is a racing receive // in blockTimerChan, but timerHeaped not being set means // it hasn't run t.maybeAdd yet; in that case, running the // timer ourselves now is fine.) if now := nanotime(); t.when <= now { systemstack(func() { t.unlockAndRun(now) // resets t.when }) t.lock() } } } // stop stops the timer t. It may be on some other P, so we can't // actually remove it from the timers heap. We can only mark it as stopped. // It will be removed in due course by the P whose heap it is on. // Reports whether the timer was stopped before it was run. func (t *timer) stop() bool { async := debug.asynctimerchan.Load() != 0 if !async && t.isChan { lock(&t.sendLock) } t.lock() t.trace("stop") if async { t.maybeRunAsync() } if t.state&timerHeaped != 0 { t.state |= timerModified if t.state&timerZombie == 0 { t.state |= timerZombie t.ts.zombies.Add(1) } } pending := t.when > 0 t.when = 0 if !async && t.isChan { // Stop any future sends with stale values. // See timer.unlockAndRun. t.seq++ } t.unlock() if !async && t.isChan { unlock(&t.sendLock) if timerchandrain(t.hchan()) { pending = true } } return pending } // deleteMin removes timer 0 from ts. // ts must be locked. func (ts *timers) deleteMin() { assertLockHeld(&ts.mu) t := ts.heap[0].timer if t.ts != ts { throw("wrong timers") } t.ts = nil last := len(ts.heap) - 1 if last > 0 { ts.heap[0] = ts.heap[last] } ts.heap[last] = timerWhen{} ts.heap = ts.heap[:last] if last > 0 { ts.siftDown(0) } ts.updateMinWhenHeap() if last == 0 { // If there are no timers, then clearly there are no timerModified timers. ts.minWhenModified.Store(0) } } // modify modifies an existing timer. // This is called by the netpoll code or time.Ticker.Reset or time.Timer.Reset. // Reports whether the timer was modified before it was run. // If f == nil, then t.f, t.arg, and t.seq are not modified. func (t *timer) modify(when, period int64, f func(arg any, seq uintptr, delay int64), arg any, seq uintptr) bool { if when <= 0 { throw("timer when must be positive") } if period < 0 { throw("timer period must be non-negative") } async := debug.asynctimerchan.Load() != 0 if !async && t.isChan { lock(&t.sendLock) } t.lock() if async { t.maybeRunAsync() } t.trace("modify") t.period = period if f != nil { t.f = f t.arg = arg t.seq = seq } wake := false pending := t.when > 0 t.when = when if t.state&timerHeaped != 0 { t.state |= timerModified if t.state&timerZombie != 0 { // In the heap but marked for removal (by a Stop). // Unmark it, since it has been Reset and will be running again. t.ts.zombies.Add(-1) t.state &^= timerZombie } // The corresponding heap[i].when is updated later. // See comment in type timer above and in timers.adjust below. if min := t.ts.minWhenModified.Load(); min == 0 || when < min { wake = true // Force timerModified bit out to t.astate before updating t.minWhenModified, // to synchronize with t.ts.adjust. See comment in adjust. t.astate.Store(t.state) t.ts.updateMinWhenModified(when) } } add := t.needsAdd() if !async && t.isChan { // Stop any future sends with stale values. // See timer.unlockAndRun. t.seq++ } t.unlock() if !async && t.isChan { if timerchandrain(t.hchan()) { pending = true } unlock(&t.sendLock) } if add { t.maybeAdd() } if wake { wakeNetPoller(when) } return pending } // needsAdd reports whether t needs to be added to a timers heap. // t must be locked. func (t *timer) needsAdd() bool { assertLockHeld(&t.mu) need := t.state&timerHeaped == 0 && t.when > 0 && (!t.isChan || t.blocked > 0) if need { t.trace("needsAdd+") } else { t.trace("needsAdd-") } return need } // maybeAdd adds t to the local timers heap if it needs to be in a heap. // The caller must not hold t's lock nor any timers heap lock. // The caller probably just unlocked t, but that lock must be dropped // in order to acquire a ts.lock, to avoid lock inversions. // (timers.adjust holds ts.lock while acquiring each t's lock, // so we cannot hold any t's lock while acquiring ts.lock). // // Strictly speaking it *might* be okay to hold t.lock and // acquire ts.lock at the same time, because we know that // t is not in any ts.heap, so nothing holding a ts.lock would // be acquiring the t.lock at the same time, meaning there // isn't a possible deadlock. But it is easier and safer not to be // too clever and respect the static ordering. // (If we don't, we have to change the static lock checking of t and ts.) // // Concurrent calls to time.Timer.Reset or blockTimerChan // may result in concurrent calls to t.maybeAdd, // so we cannot assume that t is not in a heap on entry to t.maybeAdd. func (t *timer) maybeAdd() { // Note: Not holding any locks on entry to t.maybeAdd, // so the current g can be rescheduled to a different M and P // at any time, including between the ts := assignment and the // call to ts.lock. If a reschedule happened then, we would be // adding t to some other P's timers, perhaps even a P that the scheduler // has marked as idle with no timers, in which case the timer could // go unnoticed until long after t.when. // Calling acquirem instead of using getg().m makes sure that // we end up locking and inserting into the current P's timers. mp := acquirem() ts := &mp.p.ptr().timers ts.lock() ts.cleanHead() t.lock() t.trace("maybeAdd") when := int64(0) wake := false if t.needsAdd() { t.state |= timerHeaped when = t.when wakeTime := ts.wakeTime() wake = wakeTime == 0 || when < wakeTime ts.addHeap(t) } t.unlock() ts.unlock() releasem(mp) if wake { wakeNetPoller(when) } } // reset resets the time when a timer should fire. // If used for an inactive timer, the timer will become active. // Reports whether the timer was active and was stopped. func (t *timer) reset(when, period int64) bool { return t.modify(when, period, nil, nil, 0) } // cleanHead cleans up the head of the timer queue. This speeds up // programs that create and delete timers; leaving them in the heap // slows down heap operations. // The caller must have locked ts. func (ts *timers) cleanHead() { ts.trace("cleanHead") assertLockHeld(&ts.mu) gp := getg() for { if len(ts.heap) == 0 { return } // This loop can theoretically run for a while, and because // it is holding timersLock it cannot be preempted. // If someone is trying to preempt us, just return. // We can clean the timers later. if gp.preemptStop { return } // Delete zombies from tail of heap. It requires no heap adjustments at all, // and doing so increases the chances that when we swap out a zombie // in heap[0] for the tail of the heap, we'll get a non-zombie timer, // shortening this loop. n := len(ts.heap) if t := ts.heap[n-1].timer; t.astate.Load()&timerZombie != 0 { t.lock() if t.state&timerZombie != 0 { t.state &^= timerHeaped | timerZombie | timerModified t.ts = nil ts.zombies.Add(-1) ts.heap[n-1] = timerWhen{} ts.heap = ts.heap[:n-1] } t.unlock() continue } t := ts.heap[0].timer if t.ts != ts { throw("bad ts") } if t.astate.Load()&(timerModified|timerZombie) == 0 { // Fast path: head of timers does not need adjustment. return } t.lock() updated := t.updateHeap() t.unlock() if !updated { // Head of timers does not need adjustment. return } } } // take moves any timers from src into ts // and then clears the timer state from src, // because src is being destroyed. // The caller must not have locked either timers. // For now this is only called when the world is stopped. func (ts *timers) take(src *timers) { ts.trace("take") assertWorldStopped() if len(src.heap) > 0 { // The world is stopped, so we ignore the locking of ts and src here. // That would introduce a sched < timers lock ordering, // which we'd rather avoid in the static ranking. for _, tw := range src.heap { t := tw.timer t.ts = nil if t.state&timerZombie != 0 { t.state &^= timerHeaped | timerZombie | timerModified } else { t.state &^= timerModified ts.addHeap(t) } } src.heap = nil src.zombies.Store(0) src.minWhenHeap.Store(0) src.minWhenModified.Store(0) src.len.Store(0) ts.len.Store(uint32(len(ts.heap))) } } // adjust looks through the timers in ts.heap for // any timers that have been modified to run earlier, and puts them in // the correct place in the heap. While looking for those timers, // it also moves timers that have been modified to run later, // and removes deleted timers. The caller must have locked ts. func (ts *timers) adjust(now int64, force bool) { ts.trace("adjust") assertLockHeld(&ts.mu) // If we haven't yet reached the time of the earliest modified // timer, don't do anything. This speeds up programs that adjust // a lot of timers back and forth if the timers rarely expire. // We'll postpone looking through all the adjusted timers until // one would actually expire. if !force { first := ts.minWhenModified.Load() if first == 0 || first > now { if verifyTimers { ts.verify() } return } } // minWhenModified is a lower bound on the earliest t.when // among the timerModified timers. We want to make it more precise: // we are going to scan the heap and clean out all the timerModified bits, // at which point minWhenModified can be set to 0 (indicating none at all). // // Other P's can be calling ts.wakeTime concurrently, and we'd like to // keep ts.wakeTime returning an accurate value throughout this entire process. // // Setting minWhenModified = 0 *before* the scan could make wakeTime // return an incorrect value: if minWhenModified < minWhenHeap, then clearing // it to 0 will make wakeTime return minWhenHeap (too late) until the scan finishes. // To avoid that, we want to set minWhenModified to 0 *after* the scan. // // Setting minWhenModified = 0 *after* the scan could result in missing // concurrent timer modifications in other goroutines; those will lock // the specific timer, set the timerModified bit, and set t.when. // To avoid that, we want to set minWhenModified to 0 *before* the scan. // // The way out of this dilemma is to preserve wakeTime a different way. // wakeTime is min(minWhenHeap, minWhenModified), and minWhenHeap // is protected by ts.lock, which we hold, so we can modify it however we like // in service of keeping wakeTime accurate. // // So we can: // // 1. Set minWhenHeap = min(minWhenHeap, minWhenModified) // 2. Set minWhenModified = 0 // (Other goroutines may modify timers and update minWhenModified now.) // 3. Scan timers // 4. Set minWhenHeap = heap[0].when // // That order preserves a correct value of wakeTime throughout the entire // operation: // Step 1 “locks in” an accurate wakeTime even with minWhenModified cleared. // Step 2 makes sure concurrent t.when updates are not lost during the scan. // Step 3 processes all modified timer values, justifying minWhenModified = 0. // Step 4 corrects minWhenHeap to a precise value. // // The wakeTime method implementation reads minWhenModified *before* minWhenHeap, // so that if the minWhenModified is observed to be 0, that means the minWhenHeap that // follows will include the information that was zeroed out of it. // // Originally Step 3 locked every timer, which made sure any timer update that was // already in progress during Steps 1+2 completed and was observed by Step 3. // All that locking was too expensive, so now we do an atomic load of t.astate to // decide whether we need to do a full lock. To make sure that we still observe any // timer update already in progress during Steps 1+2, t.modify sets timerModified // in t.astate *before* calling t.updateMinWhenModified. That ensures that the // overwrite in Step 2 cannot lose an update: if it does overwrite an update, Step 3 // will see the timerModified and do a full lock. ts.minWhenHeap.Store(ts.wakeTime()) ts.minWhenModified.Store(0) changed := false for i := 0; i < len(ts.heap); i++ { tw := &ts.heap[i] t := tw.timer if t.ts != ts { throw("bad ts") } if t.astate.Load()&(timerModified|timerZombie) == 0 { // Does not need adjustment. continue } t.lock() switch { case t.state&timerHeaped == 0: badTimer() case t.state&timerZombie != 0: ts.zombies.Add(-1) t.state &^= timerHeaped | timerZombie | timerModified n := len(ts.heap) ts.heap[i] = ts.heap[n-1] ts.heap[n-1] = timerWhen{} ts.heap = ts.heap[:n-1] t.ts = nil i-- changed = true case t.state&timerModified != 0: tw.when = t.when t.state &^= timerModified changed = true } t.unlock() } if changed { ts.initHeap() } ts.updateMinWhenHeap() if verifyTimers { ts.verify() } } // wakeTime looks at ts's timers and returns the time when we // should wake up the netpoller. It returns 0 if there are no timers. // This function is invoked when dropping a P, so it must run without // any write barriers. // //go:nowritebarrierrec func (ts *timers) wakeTime() int64 { // Note that the order of these two loads matters: // adjust updates minWhen to make it safe to clear minNextWhen. // We read minWhen after reading minNextWhen so that // if we see a cleared minNextWhen, we are guaranteed to see // the updated minWhen. nextWhen := ts.minWhenModified.Load() when := ts.minWhenHeap.Load() if when == 0 || (nextWhen != 0 && nextWhen < when) { when = nextWhen } return when } // check runs any timers in ts that are ready. // If now is not 0 it is the current time. // It returns the passed time or the current time if now was passed as 0. // and the time when the next timer should run or 0 if there is no next timer, // and reports whether it ran any timers. // If the time when the next timer should run is not 0, // it is always larger than the returned time. // We pass now in and out to avoid extra calls of nanotime. // //go:yeswritebarrierrec func (ts *timers) check(now int64) (rnow, pollUntil int64, ran bool) { ts.trace("check") // If it's not yet time for the first timer, or the first adjusted // timer, then there is nothing to do. next := ts.wakeTime() if next == 0 { // No timers to run or adjust. return now, 0, false } if now == 0 { now = nanotime() } // If this is the local P, and there are a lot of deleted timers, // clear them out. We only do this for the local P to reduce // lock contention on timersLock. zombies := ts.zombies.Load() if zombies < 0 { badTimer() } force := ts == &getg().m.p.ptr().timers && int(zombies) > int(ts.len.Load())/4 if now < next && !force { // Next timer is not ready to run, and we don't need to clear deleted timers. return now, next, false } ts.lock() if len(ts.heap) > 0 { ts.adjust(now, false) for len(ts.heap) > 0 { // Note that runtimer may temporarily unlock ts. if tw := ts.run(now); tw != 0 { if tw > 0 { pollUntil = tw } break } ran = true } // Note: Delaying the forced adjustment until after the ts.run // (as opposed to calling ts.adjust(now, force) above) // is significantly faster under contention, such as in // package time's BenchmarkTimerAdjust10000, // though we do not fully understand why. force = ts == &getg().m.p.ptr().timers && int(ts.zombies.Load()) > int(ts.len.Load())/4 if force { ts.adjust(now, true) } } ts.unlock() return now, pollUntil, ran } // run examines the first timer in ts. If it is ready based on now, // it runs the timer and removes or updates it. // Returns 0 if it ran a timer, -1 if there are no more timers, or the time // when the first timer should run. // The caller must have locked ts. // If a timer is run, this will temporarily unlock ts. // //go:systemstack func (ts *timers) run(now int64) int64 { ts.trace("run") assertLockHeld(&ts.mu) Redo: if len(ts.heap) == 0 { return -1 } tw := ts.heap[0] t := tw.timer if t.ts != ts { throw("bad ts") } if t.astate.Load()&(timerModified|timerZombie) == 0 && tw.when > now { // Fast path: not ready to run. return tw.when } t.lock() if t.updateHeap() { t.unlock() goto Redo } if t.state&timerHeaped == 0 || t.state&timerModified != 0 { badTimer() } if t.when > now { // Not ready to run. t.unlock() return t.when } t.unlockAndRun(now) assertLockHeld(&ts.mu) // t is unlocked now, but not ts return 0 } // unlockAndRun unlocks and runs the timer t (which must be locked). // If t is in a timer set (t.ts != nil), the caller must also have locked the timer set, // and this call will temporarily unlock the timer set while running the timer function. // unlockAndRun returns with t unlocked and t.ts (re-)locked. // //go:systemstack func (t *timer) unlockAndRun(now int64) { t.trace("unlockAndRun") assertLockHeld(&t.mu) if t.ts != nil { assertLockHeld(&t.ts.mu) } if raceenabled { // Note that we are running on a system stack, // so there is no chance of getg().m being reassigned // out from under us while this function executes. tsLocal := &getg().m.p.ptr().timers if tsLocal.raceCtx == 0 { tsLocal.raceCtx = racegostart(abi.FuncPCABIInternal((*timers).run) + sys.PCQuantum) } raceacquirectx(tsLocal.raceCtx, unsafe.Pointer(t)) } if t.state&(timerModified|timerZombie) != 0 { badTimer() } f := t.f arg := t.arg seq := t.seq var next int64 delay := now - t.when if t.period > 0 { // Leave in heap but adjust next time to fire. next = t.when + t.period*(1+delay/t.period) if next < 0 { // check for overflow. next = maxWhen } } else { next = 0 } ts := t.ts t.when = next if t.state&timerHeaped != 0 { t.state |= timerModified if next == 0 { t.state |= timerZombie t.ts.zombies.Add(1) } t.updateHeap() } t.unlock() if raceenabled { // Temporarily use the current P's racectx for g0. gp := getg() if gp.racectx != 0 { throw("unexpected racectx") } gp.racectx = gp.m.p.ptr().timers.raceCtx } if ts != nil { ts.unlock() } async := debug.asynctimerchan.Load() != 0 if !async && t.isChan { // For a timer channel, we want to make sure that no stale sends // happen after a t.stop or t.modify, but we cannot hold t.mu // during the actual send (which f does) due to lock ordering. // It can happen that we are holding t's lock above, we decide // it's time to send a time value (by calling f), grab the parameters, // unlock above, and then a t.stop or t.modify changes the timer // and returns. At that point, the send needs not to happen after all. // The way we arrange for it not to happen is that t.stop and t.modify // both increment t.seq while holding both t.mu and t.sendLock. // We copied the seq value above while holding t.mu. // Now we can acquire t.sendLock (which will be held across the send) // and double-check that t.seq is still the seq value we saw above. // If not, the timer has been updated and we should skip the send. // We skip the send by reassigning f to a no-op function. lock(&t.sendLock) if t.seq != seq { f = func(any, uintptr, int64) {} } } f(arg, seq, delay) if !async && t.isChan { unlock(&t.sendLock) } if ts != nil { ts.lock() } if raceenabled { gp := getg() gp.racectx = 0 } } // verifyTimerHeap verifies that the timers is in a valid state. // This is only for debugging, and is only called if verifyTimers is true. // The caller must have locked ts. func (ts *timers) verify() { assertLockHeld(&ts.mu) for i, tw := range ts.heap { if i == 0 { // First timer has no parent. continue } // The heap is timerHeapN-ary. See siftupTimer and siftdownTimer. p := int(uint(i-1) / timerHeapN) if tw.when < ts.heap[p].when { print("bad timer heap at ", i, ": ", p, ": ", ts.heap[p].when, ", ", i, ": ", tw.when, "\n") throw("bad timer heap") } } if n := int(ts.len.Load()); len(ts.heap) != n { println("timer heap len", len(ts.heap), "!= atomic len", n) throw("bad timer heap len") } } // updateMinWhenHeap sets ts.minWhenHeap to ts.heap[0].when. // The caller must have locked ts or the world must be stopped. func (ts *timers) updateMinWhenHeap() { assertWorldStoppedOrLockHeld(&ts.mu) if len(ts.heap) == 0 { ts.minWhenHeap.Store(0) } else { ts.minWhenHeap.Store(ts.heap[0].when) } } // updateMinWhenModified updates ts.minWhenModified to be <= when. // ts need not be (and usually is not) locked. func (ts *timers) updateMinWhenModified(when int64) { for { old := ts.minWhenModified.Load() if old != 0 && old < when { return } if ts.minWhenModified.CompareAndSwap(old, when) { return } } } // timeSleepUntil returns the time when the next timer should fire. Returns // maxWhen if there are no timers. // This is only called by sysmon and checkdead. func timeSleepUntil() int64 { next := int64(maxWhen) // Prevent allp slice changes. This is like retake. lock(&allpLock) for _, pp := range allp { if pp == nil { // This can happen if procresize has grown // allp but not yet created new Ps. continue } if w := pp.timers.wakeTime(); w != 0 { next = min(next, w) } } unlock(&allpLock) return next } const timerHeapN = 4 // Heap maintenance algorithms. // These algorithms check for slice index errors manually. // Slice index error can happen if the program is using racy // access to timers. We don't want to panic here, because // it will cause the program to crash with a mysterious // "panic holding locks" message. Instead, we panic while not // holding a lock. // siftUp puts the timer at position i in the right place // in the heap by moving it up toward the top of the heap. func (ts *timers) siftUp(i int) { heap := ts.heap if i >= len(heap) { badTimer() } tw := heap[i] when := tw.when if when <= 0 { badTimer() } for i > 0 { p := int(uint(i-1) / timerHeapN) // parent if when >= heap[p].when { break } heap[i] = heap[p] i = p } if heap[i].timer != tw.timer { heap[i] = tw } } // siftDown puts the timer at position i in the right place // in the heap by moving it down toward the bottom of the heap. func (ts *timers) siftDown(i int) { heap := ts.heap n := len(heap) if i >= n { badTimer() } if i*timerHeapN+1 >= n { return } tw := heap[i] when := tw.when if when <= 0 { badTimer() } for { leftChild := i*timerHeapN + 1 if leftChild >= n { break } w := when c := -1 for j, tw := range heap[leftChild:min(leftChild+timerHeapN, n)] { if tw.when < w { w = tw.when c = leftChild + j } } if c < 0 { break } heap[i] = heap[c] i = c } if heap[i].timer != tw.timer { heap[i] = tw } } // initHeap reestablishes the heap order in the slice ts.heap. // It takes O(n) time for n=len(ts.heap), not the O(n log n) of n repeated add operations. func (ts *timers) initHeap() { // Last possible element that needs sifting down is parent of last element; // last element is len(t)-1; parent of last element is (len(t)-1-1)/timerHeapN. if len(ts.heap) <= 1 { return } for i := int(uint(len(ts.heap)-1-1) / timerHeapN); i >= 0; i-- { ts.siftDown(i) } } // badTimer is called if the timer data structures have been corrupted, // presumably due to racy use by the program. We panic here rather than // panicking due to invalid slice access while holding locks. // See issue #25686. func badTimer() { throw("timer data corruption") } // Timer channels. // maybeRunChan checks whether the timer needs to run // to send a value to its associated channel. If so, it does. // The timer must not be locked. func (t *timer) maybeRunChan() { if t.astate.Load()&timerHeaped != 0 { // If the timer is in the heap, the ordinary timer code // is in charge of sending when appropriate. return } t.lock() now := nanotime() if t.state&timerHeaped != 0 || t.when == 0 || t.when > now { t.trace("maybeRunChan-") // Timer in the heap, or not running at all, or not triggered. t.unlock() return } t.trace("maybeRunChan+") systemstack(func() { t.unlockAndRun(now) }) } // blockTimerChan is called when a channel op has decided to block on c. // The caller holds the channel lock for c and possibly other channels. // blockTimerChan makes sure that c is in a timer heap, // adding it if needed. func blockTimerChan(c *hchan) { t := c.timer t.lock() t.trace("blockTimerChan") if !t.isChan { badTimer() } t.blocked++ // If this is the first enqueue after a recent dequeue, // the timer may still be in the heap but marked as a zombie. // Unmark it in this case, if the timer is still pending. if t.state&timerHeaped != 0 && t.state&timerZombie != 0 && t.when > 0 { t.state &^= timerZombie t.ts.zombies.Add(-1) } // t.maybeAdd must be called with t unlocked, // because it needs to lock t.ts before t. // Then it will do nothing if t.needsAdd(state) is false. // Check that now before the unlock, // avoiding the extra lock-lock-unlock-unlock // inside maybeAdd when t does not need to be added. add := t.needsAdd() t.unlock() if add { t.maybeAdd() } } // unblockTimerChan is called when a channel op that was blocked on c // is no longer blocked. Every call to blockTimerChan must be paired with // a call to unblockTimerChan. // The caller holds the channel lock for c and possibly other channels. // unblockTimerChan removes c from the timer heap when nothing is // blocked on it anymore. func unblockTimerChan(c *hchan) { t := c.timer t.lock() t.trace("unblockTimerChan") if !t.isChan || t.blocked == 0 { badTimer() } t.blocked-- if t.blocked == 0 && t.state&timerHeaped != 0 && t.state&timerZombie == 0 { // Last goroutine that was blocked on this timer. // Mark for removal from heap but do not clear t.when, // so that we know what time it is still meant to trigger. t.state |= timerZombie t.ts.zombies.Add(1) } t.unlock() }