// Copyright 2023 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. package runtime import "unsafe" // A coro represents extra concurrency without extra parallelism, // as would be needed for a coroutine implementation. // The coro does not represent a specific coroutine, only the ability // to do coroutine-style control transfers. // It can be thought of as like a special channel that always has // a goroutine blocked on it. If another goroutine calls coroswitch(c), // the caller becomes the goroutine blocked in c, and the goroutine // formerly blocked in c starts running. // These switches continue until a call to coroexit(c), // which ends the use of the coro by releasing the blocked // goroutine in c and exiting the current goroutine. // // Coros are heap allocated and garbage collected, so that user code // can hold a pointer to a coro without causing potential dangling // pointer errors. type coro struct { gp guintptr f func(*coro) // State for validating thread-lock interactions. mp *m lockedExt uint32 // mp's external LockOSThread counter at coro creation time. lockedInt uint32 // mp's internal lockOSThread counter at coro creation time. } //go:linkname newcoro // newcoro creates a new coro containing a // goroutine blocked waiting to run f // and returns that coro. func newcoro(f func(*coro)) *coro { c := new(coro) c.f = f pc := getcallerpc() gp := getg() systemstack(func() { mp := gp.m start := corostart startfv := *(**funcval)(unsafe.Pointer(&start)) gp = newproc1(startfv, gp, pc, true, waitReasonCoroutine) // Scribble down locked thread state if needed and/or donate // thread-lock state to the new goroutine. if mp.lockedExt+mp.lockedInt != 0 { c.mp = mp c.lockedExt = mp.lockedExt c.lockedInt = mp.lockedInt } }) gp.coroarg = c c.gp.set(gp) return c } // corostart is the entry func for a new coroutine. // It runs the coroutine user function f passed to corostart // and then calls coroexit to remove the extra concurrency. func corostart() { gp := getg() c := gp.coroarg gp.coroarg = nil defer coroexit(c) c.f(c) } // coroexit is like coroswitch but closes the coro // and exits the current goroutine func coroexit(c *coro) { gp := getg() gp.coroarg = c gp.coroexit = true mcall(coroswitch_m) } //go:linkname coroswitch // coroswitch switches to the goroutine blocked on c // and then blocks the current goroutine on c. func coroswitch(c *coro) { gp := getg() gp.coroarg = c mcall(coroswitch_m) } // coroswitch_m is the implementation of coroswitch // that runs on the m stack. // // Note: Coroutine switches are expected to happen at // an order of magnitude (or more) higher frequency // than regular goroutine switches, so this path is heavily // optimized to remove unnecessary work. // The fast path here is three CAS: the one at the top on gp.atomicstatus, // the one in the middle to choose the next g, // and the one at the bottom on gnext.atomicstatus. // It is important not to add more atomic operations or other // expensive operations to the fast path. func coroswitch_m(gp *g) { c := gp.coroarg gp.coroarg = nil exit := gp.coroexit gp.coroexit = false mp := gp.m // Track and validate thread-lock interactions. // // The rules with thread-lock interactions are simple. When a coro goroutine is switched to, // the same thread must be used, and the locked state must match with the thread-lock state of // the goroutine which called newcoro. Thread-lock state consists of the thread and the number // of internal (cgo callback, etc.) and external (LockOSThread) thread locks. locked := gp.lockedm != 0 if c.mp != nil || locked { if mp != c.mp || mp.lockedInt != c.lockedInt || mp.lockedExt != c.lockedExt { print("coro: got thread ", unsafe.Pointer(mp), ", want ", unsafe.Pointer(c.mp), "\n") print("coro: got lock internal ", mp.lockedInt, ", want ", c.lockedInt, "\n") print("coro: got lock external ", mp.lockedExt, ", want ", c.lockedExt, "\n") throw("coro: OS thread locking must match locking at coroutine creation") } } // Acquire tracer for writing for the duration of this call. // // There's a lot of state manipulation performed with shortcuts // but we need to make sure the tracer can only observe the // start and end states to maintain a coherent model and avoid // emitting an event for every single transition. trace := traceAcquire() if locked { // Detach the goroutine from the thread; we'll attach to the goroutine we're // switching to before returning. gp.lockedm.set(nil) } if exit { // The M might have a non-zero OS thread lock count when we get here, gdestroy // will avoid destroying the M if the G isn't explicitly locked to it via lockedm, // which we cleared above. It's fine to gdestroy here also, even when locked to // the thread, because we'll be switching back to another goroutine anyway, which // will take back its thread-lock state before returning. gdestroy(gp) gp = nil } else { // If we can CAS ourselves directly from running to waiting, so do, // keeping the control transfer as lightweight as possible. gp.waitreason = waitReasonCoroutine if !gp.atomicstatus.CompareAndSwap(_Grunning, _Gwaiting) { // The CAS failed: use casgstatus, which will take care of // coordinating with the garbage collector about the state change. casgstatus(gp, _Grunning, _Gwaiting) } // Clear gp.m. setMNoWB(&gp.m, nil) } // The goroutine stored in c is the one to run next. // Swap it with ourselves. var gnext *g for { // Note: this is a racy load, but it will eventually // get the right value, and if it gets the wrong value, // the c.gp.cas will fail, so no harm done other than // a wasted loop iteration. // The cas will also sync c.gp's // memory enough that the next iteration of the racy load // should see the correct value. // We are avoiding the atomic load to keep this path // as lightweight as absolutely possible. // (The atomic load is free on x86 but not free elsewhere.) next := c.gp if next.ptr() == nil { throw("coroswitch on exited coro") } var self guintptr self.set(gp) if c.gp.cas(next, self) { gnext = next.ptr() break } } // Check if we're switching to ourselves. This case is able to break our // thread-lock invariants and an unbuffered channel implementation of // coroswitch would deadlock. It's clear that this case should just not // work. if gnext == gp { throw("coroswitch of a goroutine to itself") } // Emit the trace event after getting gnext but before changing curg. // GoSwitch expects that the current G is running and that we haven't // switched yet for correct status emission. if trace.ok() { trace.GoSwitch(gnext, exit) } // Start running next, without heavy scheduling machinery. // Set mp.curg and gnext.m and then update scheduling state // directly if possible. setGNoWB(&mp.curg, gnext) setMNoWB(&gnext.m, mp) // Synchronize with any out-standing goroutine profile. We're about to start // executing, and an invariant of the profiler is that we tryRecordGoroutineProfile // whenever a goroutine is about to start running. // // N.B. We must do this before transitioning to _Grunning but after installing gnext // in curg, so that we have a valid curg for allocation (tryRecordGoroutineProfile // may allocate). if goroutineProfile.active { tryRecordGoroutineProfile(gnext, nil, osyield) } if !gnext.atomicstatus.CompareAndSwap(_Gwaiting, _Grunning) { // The CAS failed: use casgstatus, which will take care of // coordinating with the garbage collector about the state change. casgstatus(gnext, _Gwaiting, _Grunnable) casgstatus(gnext, _Grunnable, _Grunning) } // Donate locked state. if locked { mp.lockedg.set(gnext) gnext.lockedm.set(mp) } // Release the trace locker. We've completed all the necessary transitions.. if trace.ok() { traceRelease(trace) } // Switch to gnext. Does not return. gogo(&gnext.sched) }