Source file src/runtime/mgc.go
1 // Copyright 2009 The Go Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style 3 // license that can be found in the LICENSE file. 4 5 // Garbage collector (GC). 6 // 7 // The GC runs concurrently with mutator threads, is type accurate (aka precise), allows multiple 8 // GC thread to run in parallel. It is a concurrent mark and sweep that uses a write barrier. It is 9 // non-generational and non-compacting. Allocation is done using size segregated per P allocation 10 // areas to minimize fragmentation while eliminating locks in the common case. 11 // 12 // The algorithm decomposes into several steps. 13 // This is a high level description of the algorithm being used. For an overview of GC a good 14 // place to start is Richard Jones' gchandbook.org. 15 // 16 // The algorithm's intellectual heritage includes Dijkstra's on-the-fly algorithm, see 17 // Edsger W. Dijkstra, Leslie Lamport, A. J. Martin, C. S. Scholten, and E. F. M. Steffens. 1978. 18 // On-the-fly garbage collection: an exercise in cooperation. Commun. ACM 21, 11 (November 1978), 19 // 966-975. 20 // For journal quality proofs that these steps are complete, correct, and terminate see 21 // Hudson, R., and Moss, J.E.B. Copying Garbage Collection without stopping the world. 22 // Concurrency and Computation: Practice and Experience 15(3-5), 2003. 23 // 24 // 1. GC performs sweep termination. 25 // 26 // a. Stop the world. This causes all Ps to reach a GC safe-point. 27 // 28 // b. Sweep any unswept spans. There will only be unswept spans if 29 // this GC cycle was forced before the expected time. 30 // 31 // 2. GC performs the mark phase. 32 // 33 // a. Prepare for the mark phase by setting gcphase to _GCmark 34 // (from _GCoff), enabling the write barrier, enabling mutator 35 // assists, and enqueueing root mark jobs. No objects may be 36 // scanned until all Ps have enabled the write barrier, which is 37 // accomplished using STW. 38 // 39 // b. Start the world. From this point, GC work is done by mark 40 // workers started by the scheduler and by assists performed as 41 // part of allocation. The write barrier shades both the 42 // overwritten pointer and the new pointer value for any pointer 43 // writes (see mbarrier.go for details). Newly allocated objects 44 // are immediately marked black. 45 // 46 // c. GC performs root marking jobs. This includes scanning all 47 // stacks, shading all globals, and shading any heap pointers in 48 // off-heap runtime data structures. Scanning a stack stops a 49 // goroutine, shades any pointers found on its stack, and then 50 // resumes the goroutine. 51 // 52 // d. GC drains the work queue of grey objects, scanning each grey 53 // object to black and shading all pointers found in the object 54 // (which in turn may add those pointers to the work queue). 55 // 56 // e. Because GC work is spread across local caches, GC uses a 57 // distributed termination algorithm to detect when there are no 58 // more root marking jobs or grey objects (see gcMarkDone). At this 59 // point, GC transitions to mark termination. 60 // 61 // 3. GC performs mark termination. 62 // 63 // a. Stop the world. 64 // 65 // b. Set gcphase to _GCmarktermination, and disable workers and 66 // assists. 67 // 68 // c. Perform housekeeping like flushing mcaches. 69 // 70 // 4. GC performs the sweep phase. 71 // 72 // a. Prepare for the sweep phase by setting gcphase to _GCoff, 73 // setting up sweep state and disabling the write barrier. 74 // 75 // b. Start the world. From this point on, newly allocated objects 76 // are white, and allocating sweeps spans before use if necessary. 77 // 78 // c. GC does concurrent sweeping in the background and in response 79 // to allocation. See description below. 80 // 81 // 5. When sufficient allocation has taken place, replay the sequence 82 // starting with 1 above. See discussion of GC rate below. 83 84 // Concurrent sweep. 85 // 86 // The sweep phase proceeds concurrently with normal program execution. 87 // The heap is swept span-by-span both lazily (when a goroutine needs another span) 88 // and concurrently in a background goroutine (this helps programs that are not CPU bound). 89 // At the end of STW mark termination all spans are marked as "needs sweeping". 90 // 91 // The background sweeper goroutine simply sweeps spans one-by-one. 92 // 93 // To avoid requesting more OS memory while there are unswept spans, when a 94 // goroutine needs another span, it first attempts to reclaim that much memory 95 // by sweeping. When a goroutine needs to allocate a new small-object span, it 96 // sweeps small-object spans for the same object size until it frees at least 97 // one object. When a goroutine needs to allocate large-object span from heap, 98 // it sweeps spans until it frees at least that many pages into heap. There is 99 // one case where this may not suffice: if a goroutine sweeps and frees two 100 // nonadjacent one-page spans to the heap, it will allocate a new two-page 101 // span, but there can still be other one-page unswept spans which could be 102 // combined into a two-page span. 103 // 104 // It's critical to ensure that no operations proceed on unswept spans (that would corrupt 105 // mark bits in GC bitmap). During GC all mcaches are flushed into the central cache, 106 // so they are empty. When a goroutine grabs a new span into mcache, it sweeps it. 107 // When a goroutine explicitly frees an object or sets a finalizer, it ensures that 108 // the span is swept (either by sweeping it, or by waiting for the concurrent sweep to finish). 109 // The finalizer goroutine is kicked off only when all spans are swept. 110 // When the next GC starts, it sweeps all not-yet-swept spans (if any). 111 112 // GC rate. 113 // Next GC is after we've allocated an extra amount of memory proportional to 114 // the amount already in use. The proportion is controlled by GOGC environment variable 115 // (100 by default). If GOGC=100 and we're using 4M, we'll GC again when we get to 8M 116 // (this mark is computed by the gcController.heapGoal method). This keeps the GC cost in 117 // linear proportion to the allocation cost. Adjusting GOGC just changes the linear constant 118 // (and also the amount of extra memory used). 119 120 // Oblets 121 // 122 // In order to prevent long pauses while scanning large objects and to 123 // improve parallelism, the garbage collector breaks up scan jobs for 124 // objects larger than maxObletBytes into "oblets" of at most 125 // maxObletBytes. When scanning encounters the beginning of a large 126 // object, it scans only the first oblet and enqueues the remaining 127 // oblets as new scan jobs. 128 129 package runtime 130 131 import ( 132 "internal/cpu" 133 "internal/runtime/atomic" 134 "unsafe" 135 ) 136 137 const ( 138 _DebugGC = 0 139 140 // concurrentSweep is a debug flag. Disabling this flag 141 // ensures all spans are swept while the world is stopped. 142 concurrentSweep = true 143 144 // debugScanConservative enables debug logging for stack 145 // frames that are scanned conservatively. 146 debugScanConservative = false 147 148 // sweepMinHeapDistance is a lower bound on the heap distance 149 // (in bytes) reserved for concurrent sweeping between GC 150 // cycles. 151 sweepMinHeapDistance = 1024 * 1024 152 ) 153 154 // heapObjectsCanMove always returns false in the current garbage collector. 155 // It exists for go4.org/unsafe/assume-no-moving-gc, which is an 156 // unfortunate idea that had an even more unfortunate implementation. 157 // Every time a new Go release happened, the package stopped building, 158 // and the authors had to add a new file with a new //go:build line, and 159 // then the entire ecosystem of packages with that as a dependency had to 160 // explicitly update to the new version. Many packages depend on 161 // assume-no-moving-gc transitively, through paths like 162 // inet.af/netaddr -> go4.org/intern -> assume-no-moving-gc. 163 // This was causing a significant amount of friction around each new 164 // release, so we added this bool for the package to //go:linkname 165 // instead. The bool is still unfortunate, but it's not as bad as 166 // breaking the ecosystem on every new release. 167 // 168 // If the Go garbage collector ever does move heap objects, we can set 169 // this to true to break all the programs using assume-no-moving-gc. 170 // 171 //go:linkname heapObjectsCanMove 172 func heapObjectsCanMove() bool { 173 return false 174 } 175 176 func gcinit() { 177 if unsafe.Sizeof(workbuf{}) != _WorkbufSize { 178 throw("size of Workbuf is suboptimal") 179 } 180 // No sweep on the first cycle. 181 sweep.active.state.Store(sweepDrainedMask) 182 183 // Initialize GC pacer state. 184 // Use the environment variable GOGC for the initial gcPercent value. 185 // Use the environment variable GOMEMLIMIT for the initial memoryLimit value. 186 gcController.init(readGOGC(), readGOMEMLIMIT()) 187 188 work.startSema = 1 189 work.markDoneSema = 1 190 lockInit(&work.sweepWaiters.lock, lockRankSweepWaiters) 191 lockInit(&work.assistQueue.lock, lockRankAssistQueue) 192 lockInit(&work.strongFromWeak.lock, lockRankStrongFromWeakQueue) 193 lockInit(&work.wbufSpans.lock, lockRankWbufSpans) 194 } 195 196 // gcenable is called after the bulk of the runtime initialization, 197 // just before we're about to start letting user code run. 198 // It kicks off the background sweeper goroutine, the background 199 // scavenger goroutine, and enables GC. 200 func gcenable() { 201 // Kick off sweeping and scavenging. 202 c := make(chan int, 2) 203 go bgsweep(c) 204 go bgscavenge(c) 205 <-c 206 <-c 207 memstats.enablegc = true // now that runtime is initialized, GC is okay 208 } 209 210 // Garbage collector phase. 211 // Indicates to write barrier and synchronization task to perform. 212 var gcphase uint32 213 214 // The compiler knows about this variable. 215 // If you change it, you must change builtin/runtime.go, too. 216 // If you change the first four bytes, you must also change the write 217 // barrier insertion code. 218 // 219 // writeBarrier should be an internal detail, 220 // but widely used packages access it using linkname. 221 // Notable members of the hall of shame include: 222 // - github.com/bytedance/sonic 223 // 224 // Do not remove or change the type signature. 225 // See go.dev/issue/67401. 226 // 227 //go:linkname writeBarrier 228 var writeBarrier struct { 229 enabled bool // compiler emits a check of this before calling write barrier 230 pad [3]byte // compiler uses 32-bit load for "enabled" field 231 alignme uint64 // guarantee alignment so that compiler can use a 32 or 64-bit load 232 } 233 234 // gcBlackenEnabled is 1 if mutator assists and background mark 235 // workers are allowed to blacken objects. This must only be set when 236 // gcphase == _GCmark. 237 var gcBlackenEnabled uint32 238 239 const ( 240 _GCoff = iota // GC not running; sweeping in background, write barrier disabled 241 _GCmark // GC marking roots and workbufs: allocate black, write barrier ENABLED 242 _GCmarktermination // GC mark termination: allocate black, P's help GC, write barrier ENABLED 243 ) 244 245 //go:nosplit 246 func setGCPhase(x uint32) { 247 atomic.Store(&gcphase, x) 248 writeBarrier.enabled = gcphase == _GCmark || gcphase == _GCmarktermination 249 } 250 251 // gcMarkWorkerMode represents the mode that a concurrent mark worker 252 // should operate in. 253 // 254 // Concurrent marking happens through four different mechanisms. One 255 // is mutator assists, which happen in response to allocations and are 256 // not scheduled. The other three are variations in the per-P mark 257 // workers and are distinguished by gcMarkWorkerMode. 258 type gcMarkWorkerMode int 259 260 const ( 261 // gcMarkWorkerNotWorker indicates that the next scheduled G is not 262 // starting work and the mode should be ignored. 263 gcMarkWorkerNotWorker gcMarkWorkerMode = iota 264 265 // gcMarkWorkerDedicatedMode indicates that the P of a mark 266 // worker is dedicated to running that mark worker. The mark 267 // worker should run without preemption. 268 gcMarkWorkerDedicatedMode 269 270 // gcMarkWorkerFractionalMode indicates that a P is currently 271 // running the "fractional" mark worker. The fractional worker 272 // is necessary when GOMAXPROCS*gcBackgroundUtilization is not 273 // an integer and using only dedicated workers would result in 274 // utilization too far from the target of gcBackgroundUtilization. 275 // The fractional worker should run until it is preempted and 276 // will be scheduled to pick up the fractional part of 277 // GOMAXPROCS*gcBackgroundUtilization. 278 gcMarkWorkerFractionalMode 279 280 // gcMarkWorkerIdleMode indicates that a P is running the mark 281 // worker because it has nothing else to do. The idle worker 282 // should run until it is preempted and account its time 283 // against gcController.idleMarkTime. 284 gcMarkWorkerIdleMode 285 ) 286 287 // gcMarkWorkerModeStrings are the strings labels of gcMarkWorkerModes 288 // to use in execution traces. 289 var gcMarkWorkerModeStrings = [...]string{ 290 "Not worker", 291 "GC (dedicated)", 292 "GC (fractional)", 293 "GC (idle)", 294 } 295 296 // pollFractionalWorkerExit reports whether a fractional mark worker 297 // should self-preempt. It assumes it is called from the fractional 298 // worker. 299 func pollFractionalWorkerExit() bool { 300 // This should be kept in sync with the fractional worker 301 // scheduler logic in findRunnableGCWorker. 302 now := nanotime() 303 delta := now - gcController.markStartTime 304 if delta <= 0 { 305 return true 306 } 307 p := getg().m.p.ptr() 308 selfTime := p.gcFractionalMarkTime + (now - p.gcMarkWorkerStartTime) 309 // Add some slack to the utilization goal so that the 310 // fractional worker isn't behind again the instant it exits. 311 return float64(selfTime)/float64(delta) > 1.2*gcController.fractionalUtilizationGoal 312 } 313 314 var work workType 315 316 type workType struct { 317 full lfstack // lock-free list of full blocks workbuf 318 _ cpu.CacheLinePad // prevents false-sharing between full and empty 319 empty lfstack // lock-free list of empty blocks workbuf 320 _ cpu.CacheLinePad // prevents false-sharing between empty and nproc/nwait 321 322 wbufSpans struct { 323 lock mutex 324 // free is a list of spans dedicated to workbufs, but 325 // that don't currently contain any workbufs. 326 free mSpanList 327 // busy is a list of all spans containing workbufs on 328 // one of the workbuf lists. 329 busy mSpanList 330 } 331 332 // Restore 64-bit alignment on 32-bit. 333 _ uint32 334 335 // bytesMarked is the number of bytes marked this cycle. This 336 // includes bytes blackened in scanned objects, noscan objects 337 // that go straight to black, objects allocated as black during 338 // the cycle, and permagrey objects scanned by markroot during 339 // the concurrent scan phase. 340 // 341 // This is updated atomically during the cycle. Updates may be batched 342 // arbitrarily, since the value is only read at the end of the cycle. 343 // 344 // Because of benign races during marking, this number may not 345 // be the exact number of marked bytes, but it should be very 346 // close. 347 // 348 // Put this field here because it needs 64-bit atomic access 349 // (and thus 8-byte alignment even on 32-bit architectures). 350 bytesMarked uint64 351 352 markrootNext uint32 // next markroot job 353 markrootJobs uint32 // number of markroot jobs 354 355 nproc uint32 356 tstart int64 357 nwait uint32 358 359 // Number of roots of various root types. Set by gcMarkRootPrepare. 360 // 361 // nStackRoots == len(stackRoots), but we have nStackRoots for 362 // consistency. 363 nDataRoots, nBSSRoots, nSpanRoots, nStackRoots int 364 365 // Base indexes of each root type. Set by gcMarkRootPrepare. 366 baseData, baseBSS, baseSpans, baseStacks, baseEnd uint32 367 368 // stackRoots is a snapshot of all of the Gs that existed 369 // before the beginning of concurrent marking. The backing 370 // store of this must not be modified because it might be 371 // shared with allgs. 372 stackRoots []*g 373 374 // Each type of GC state transition is protected by a lock. 375 // Since multiple threads can simultaneously detect the state 376 // transition condition, any thread that detects a transition 377 // condition must acquire the appropriate transition lock, 378 // re-check the transition condition and return if it no 379 // longer holds or perform the transition if it does. 380 // Likewise, any transition must invalidate the transition 381 // condition before releasing the lock. This ensures that each 382 // transition is performed by exactly one thread and threads 383 // that need the transition to happen block until it has 384 // happened. 385 // 386 // startSema protects the transition from "off" to mark or 387 // mark termination. 388 startSema uint32 389 // markDoneSema protects transitions from mark to mark termination. 390 markDoneSema uint32 391 392 bgMarkDone uint32 // cas to 1 when at a background mark completion point 393 // Background mark completion signaling 394 395 // mode is the concurrency mode of the current GC cycle. 396 mode gcMode 397 398 // userForced indicates the current GC cycle was forced by an 399 // explicit user call. 400 userForced bool 401 402 // initialHeapLive is the value of gcController.heapLive at the 403 // beginning of this GC cycle. 404 initialHeapLive uint64 405 406 // assistQueue is a queue of assists that are blocked because 407 // there was neither enough credit to steal or enough work to 408 // do. 409 assistQueue struct { 410 lock mutex 411 q gQueue 412 } 413 414 // sweepWaiters is a list of blocked goroutines to wake when 415 // we transition from mark termination to sweep. 416 sweepWaiters struct { 417 lock mutex 418 list gList 419 } 420 421 // strongFromWeak controls how the GC interacts with weak->strong 422 // pointer conversions. 423 strongFromWeak struct { 424 // block is a flag set during mark termination that prevents 425 // new weak->strong conversions from executing by blocking the 426 // goroutine and enqueuing it onto q. 427 // 428 // Mutated only by one goroutine at a time in gcMarkDone, 429 // with globally-synchronizing events like forEachP and 430 // stopTheWorld. 431 block bool 432 433 // q is a queue of goroutines that attempted to perform a 434 // weak->strong conversion during mark termination. 435 // 436 // Protected by lock. 437 lock mutex 438 q gQueue 439 } 440 441 // cycles is the number of completed GC cycles, where a GC 442 // cycle is sweep termination, mark, mark termination, and 443 // sweep. This differs from memstats.numgc, which is 444 // incremented at mark termination. 445 cycles atomic.Uint32 446 447 // Timing/utilization stats for this cycle. 448 stwprocs, maxprocs int32 449 tSweepTerm, tMark, tMarkTerm, tEnd int64 // nanotime() of phase start 450 451 // pauseNS is the total STW time this cycle, measured as the time between 452 // when stopping began (just before trying to stop Ps) and just after the 453 // world started again. 454 pauseNS int64 455 456 // debug.gctrace heap sizes for this cycle. 457 heap0, heap1, heap2 uint64 458 459 // Cumulative estimated CPU usage. 460 cpuStats 461 } 462 463 // GC runs a garbage collection and blocks the caller until the 464 // garbage collection is complete. It may also block the entire 465 // program. 466 func GC() { 467 // We consider a cycle to be: sweep termination, mark, mark 468 // termination, and sweep. This function shouldn't return 469 // until a full cycle has been completed, from beginning to 470 // end. Hence, we always want to finish up the current cycle 471 // and start a new one. That means: 472 // 473 // 1. In sweep termination, mark, or mark termination of cycle 474 // N, wait until mark termination N completes and transitions 475 // to sweep N. 476 // 477 // 2. In sweep N, help with sweep N. 478 // 479 // At this point we can begin a full cycle N+1. 480 // 481 // 3. Trigger cycle N+1 by starting sweep termination N+1. 482 // 483 // 4. Wait for mark termination N+1 to complete. 484 // 485 // 5. Help with sweep N+1 until it's done. 486 // 487 // This all has to be written to deal with the fact that the 488 // GC may move ahead on its own. For example, when we block 489 // until mark termination N, we may wake up in cycle N+2. 490 491 // Wait until the current sweep termination, mark, and mark 492 // termination complete. 493 n := work.cycles.Load() 494 gcWaitOnMark(n) 495 496 // We're now in sweep N or later. Trigger GC cycle N+1, which 497 // will first finish sweep N if necessary and then enter sweep 498 // termination N+1. 499 gcStart(gcTrigger{kind: gcTriggerCycle, n: n + 1}) 500 501 // Wait for mark termination N+1 to complete. 502 gcWaitOnMark(n + 1) 503 504 // Finish sweep N+1 before returning. We do this both to 505 // complete the cycle and because runtime.GC() is often used 506 // as part of tests and benchmarks to get the system into a 507 // relatively stable and isolated state. 508 for work.cycles.Load() == n+1 && sweepone() != ^uintptr(0) { 509 Gosched() 510 } 511 512 // Callers may assume that the heap profile reflects the 513 // just-completed cycle when this returns (historically this 514 // happened because this was a STW GC), but right now the 515 // profile still reflects mark termination N, not N+1. 516 // 517 // As soon as all of the sweep frees from cycle N+1 are done, 518 // we can go ahead and publish the heap profile. 519 // 520 // First, wait for sweeping to finish. (We know there are no 521 // more spans on the sweep queue, but we may be concurrently 522 // sweeping spans, so we have to wait.) 523 for work.cycles.Load() == n+1 && !isSweepDone() { 524 Gosched() 525 } 526 527 // Now we're really done with sweeping, so we can publish the 528 // stable heap profile. Only do this if we haven't already hit 529 // another mark termination. 530 mp := acquirem() 531 cycle := work.cycles.Load() 532 if cycle == n+1 || (gcphase == _GCmark && cycle == n+2) { 533 mProf_PostSweep() 534 } 535 releasem(mp) 536 } 537 538 // gcWaitOnMark blocks until GC finishes the Nth mark phase. If GC has 539 // already completed this mark phase, it returns immediately. 540 func gcWaitOnMark(n uint32) { 541 for { 542 // Disable phase transitions. 543 lock(&work.sweepWaiters.lock) 544 nMarks := work.cycles.Load() 545 if gcphase != _GCmark { 546 // We've already completed this cycle's mark. 547 nMarks++ 548 } 549 if nMarks > n { 550 // We're done. 551 unlock(&work.sweepWaiters.lock) 552 return 553 } 554 555 // Wait until sweep termination, mark, and mark 556 // termination of cycle N complete. 557 work.sweepWaiters.list.push(getg()) 558 goparkunlock(&work.sweepWaiters.lock, waitReasonWaitForGCCycle, traceBlockUntilGCEnds, 1) 559 } 560 } 561 562 // gcMode indicates how concurrent a GC cycle should be. 563 type gcMode int 564 565 const ( 566 gcBackgroundMode gcMode = iota // concurrent GC and sweep 567 gcForceMode // stop-the-world GC now, concurrent sweep 568 gcForceBlockMode // stop-the-world GC now and STW sweep (forced by user) 569 ) 570 571 // A gcTrigger is a predicate for starting a GC cycle. Specifically, 572 // it is an exit condition for the _GCoff phase. 573 type gcTrigger struct { 574 kind gcTriggerKind 575 now int64 // gcTriggerTime: current time 576 n uint32 // gcTriggerCycle: cycle number to start 577 } 578 579 type gcTriggerKind int 580 581 const ( 582 // gcTriggerHeap indicates that a cycle should be started when 583 // the heap size reaches the trigger heap size computed by the 584 // controller. 585 gcTriggerHeap gcTriggerKind = iota 586 587 // gcTriggerTime indicates that a cycle should be started when 588 // it's been more than forcegcperiod nanoseconds since the 589 // previous GC cycle. 590 gcTriggerTime 591 592 // gcTriggerCycle indicates that a cycle should be started if 593 // we have not yet started cycle number gcTrigger.n (relative 594 // to work.cycles). 595 gcTriggerCycle 596 ) 597 598 // test reports whether the trigger condition is satisfied, meaning 599 // that the exit condition for the _GCoff phase has been met. The exit 600 // condition should be tested when allocating. 601 func (t gcTrigger) test() bool { 602 if !memstats.enablegc || panicking.Load() != 0 || gcphase != _GCoff { 603 return false 604 } 605 switch t.kind { 606 case gcTriggerHeap: 607 trigger, _ := gcController.trigger() 608 return gcController.heapLive.Load() >= trigger 609 case gcTriggerTime: 610 if gcController.gcPercent.Load() < 0 { 611 return false 612 } 613 lastgc := int64(atomic.Load64(&memstats.last_gc_nanotime)) 614 return lastgc != 0 && t.now-lastgc > forcegcperiod 615 case gcTriggerCycle: 616 // t.n > work.cycles, but accounting for wraparound. 617 return int32(t.n-work.cycles.Load()) > 0 618 } 619 return true 620 } 621 622 // gcStart starts the GC. It transitions from _GCoff to _GCmark (if 623 // debug.gcstoptheworld == 0) or performs all of GC (if 624 // debug.gcstoptheworld != 0). 625 // 626 // This may return without performing this transition in some cases, 627 // such as when called on a system stack or with locks held. 628 func gcStart(trigger gcTrigger) { 629 // Since this is called from malloc and malloc is called in 630 // the guts of a number of libraries that might be holding 631 // locks, don't attempt to start GC in non-preemptible or 632 // potentially unstable situations. 633 mp := acquirem() 634 if gp := getg(); gp == mp.g0 || mp.locks > 1 || mp.preemptoff != "" { 635 releasem(mp) 636 return 637 } 638 releasem(mp) 639 mp = nil 640 641 if gp := getg(); gp.syncGroup != nil { 642 // Disassociate the G from its synctest bubble while allocating. 643 // This is less elegant than incrementing the group's active count, 644 // but avoids any contamination between GC and synctest. 645 sg := gp.syncGroup 646 gp.syncGroup = nil 647 defer func() { 648 gp.syncGroup = sg 649 }() 650 } 651 652 // Pick up the remaining unswept/not being swept spans concurrently 653 // 654 // This shouldn't happen if we're being invoked in background 655 // mode since proportional sweep should have just finished 656 // sweeping everything, but rounding errors, etc, may leave a 657 // few spans unswept. In forced mode, this is necessary since 658 // GC can be forced at any point in the sweeping cycle. 659 // 660 // We check the transition condition continuously here in case 661 // this G gets delayed in to the next GC cycle. 662 for trigger.test() && sweepone() != ^uintptr(0) { 663 } 664 665 // Perform GC initialization and the sweep termination 666 // transition. 667 semacquire(&work.startSema) 668 // Re-check transition condition under transition lock. 669 if !trigger.test() { 670 semrelease(&work.startSema) 671 return 672 } 673 674 // In gcstoptheworld debug mode, upgrade the mode accordingly. 675 // We do this after re-checking the transition condition so 676 // that multiple goroutines that detect the heap trigger don't 677 // start multiple STW GCs. 678 mode := gcBackgroundMode 679 if debug.gcstoptheworld == 1 { 680 mode = gcForceMode 681 } else if debug.gcstoptheworld == 2 { 682 mode = gcForceBlockMode 683 } 684 685 // Ok, we're doing it! Stop everybody else 686 semacquire(&gcsema) 687 semacquire(&worldsema) 688 689 // For stats, check if this GC was forced by the user. 690 // Update it under gcsema to avoid gctrace getting wrong values. 691 work.userForced = trigger.kind == gcTriggerCycle 692 693 trace := traceAcquire() 694 if trace.ok() { 695 trace.GCStart() 696 traceRelease(trace) 697 } 698 699 // Check that all Ps have finished deferred mcache flushes. 700 for _, p := range allp { 701 if fg := p.mcache.flushGen.Load(); fg != mheap_.sweepgen { 702 println("runtime: p", p.id, "flushGen", fg, "!= sweepgen", mheap_.sweepgen) 703 throw("p mcache not flushed") 704 } 705 } 706 707 gcBgMarkStartWorkers() 708 709 systemstack(gcResetMarkState) 710 711 work.stwprocs, work.maxprocs = gomaxprocs, gomaxprocs 712 if work.stwprocs > ncpu { 713 // This is used to compute CPU time of the STW phases, 714 // so it can't be more than ncpu, even if GOMAXPROCS is. 715 work.stwprocs = ncpu 716 } 717 work.heap0 = gcController.heapLive.Load() 718 work.pauseNS = 0 719 work.mode = mode 720 721 now := nanotime() 722 work.tSweepTerm = now 723 var stw worldStop 724 systemstack(func() { 725 stw = stopTheWorldWithSema(stwGCSweepTerm) 726 }) 727 728 // Accumulate fine-grained stopping time. 729 work.cpuStats.accumulateGCPauseTime(stw.stoppingCPUTime, 1) 730 731 // Finish sweep before we start concurrent scan. 732 systemstack(func() { 733 finishsweep_m() 734 }) 735 736 // clearpools before we start the GC. If we wait the memory will not be 737 // reclaimed until the next GC cycle. 738 clearpools() 739 740 work.cycles.Add(1) 741 742 // Assists and workers can start the moment we start 743 // the world. 744 gcController.startCycle(now, int(gomaxprocs), trigger) 745 746 // Notify the CPU limiter that assists may begin. 747 gcCPULimiter.startGCTransition(true, now) 748 749 // In STW mode, disable scheduling of user Gs. This may also 750 // disable scheduling of this goroutine, so it may block as 751 // soon as we start the world again. 752 if mode != gcBackgroundMode { 753 schedEnableUser(false) 754 } 755 756 // Enter concurrent mark phase and enable 757 // write barriers. 758 // 759 // Because the world is stopped, all Ps will 760 // observe that write barriers are enabled by 761 // the time we start the world and begin 762 // scanning. 763 // 764 // Write barriers must be enabled before assists are 765 // enabled because they must be enabled before 766 // any non-leaf heap objects are marked. Since 767 // allocations are blocked until assists can 768 // happen, we want to enable assists as early as 769 // possible. 770 setGCPhase(_GCmark) 771 772 gcBgMarkPrepare() // Must happen before assists are enabled. 773 gcMarkRootPrepare() 774 775 // Mark all active tinyalloc blocks. Since we're 776 // allocating from these, they need to be black like 777 // other allocations. The alternative is to blacken 778 // the tiny block on every allocation from it, which 779 // would slow down the tiny allocator. 780 gcMarkTinyAllocs() 781 782 // At this point all Ps have enabled the write 783 // barrier, thus maintaining the no white to 784 // black invariant. Enable mutator assists to 785 // put back-pressure on fast allocating 786 // mutators. 787 atomic.Store(&gcBlackenEnabled, 1) 788 789 // In STW mode, we could block the instant systemstack 790 // returns, so make sure we're not preemptible. 791 mp = acquirem() 792 793 // Update the CPU stats pause time. 794 // 795 // Use maxprocs instead of stwprocs here because the total time 796 // computed in the CPU stats is based on maxprocs, and we want them 797 // to be comparable. 798 work.cpuStats.accumulateGCPauseTime(nanotime()-stw.finishedStopping, work.maxprocs) 799 800 // Concurrent mark. 801 systemstack(func() { 802 now = startTheWorldWithSema(0, stw) 803 work.pauseNS += now - stw.startedStopping 804 work.tMark = now 805 806 // Release the CPU limiter. 807 gcCPULimiter.finishGCTransition(now) 808 }) 809 810 // Release the world sema before Gosched() in STW mode 811 // because we will need to reacquire it later but before 812 // this goroutine becomes runnable again, and we could 813 // self-deadlock otherwise. 814 semrelease(&worldsema) 815 releasem(mp) 816 817 // Make sure we block instead of returning to user code 818 // in STW mode. 819 if mode != gcBackgroundMode { 820 Gosched() 821 } 822 823 semrelease(&work.startSema) 824 } 825 826 // gcMarkDoneFlushed counts the number of P's with flushed work. 827 // 828 // Ideally this would be a captured local in gcMarkDone, but forEachP 829 // escapes its callback closure, so it can't capture anything. 830 // 831 // This is protected by markDoneSema. 832 var gcMarkDoneFlushed uint32 833 834 // gcDebugMarkDone contains fields used to debug/test mark termination. 835 var gcDebugMarkDone struct { 836 // spinAfterRaggedBarrier forces gcMarkDone to spin after it executes 837 // the ragged barrier. 838 spinAfterRaggedBarrier atomic.Bool 839 840 // restartedDueTo27993 indicates that we restarted mark termination 841 // due to the bug described in issue #27993. 842 // 843 // Protected by worldsema. 844 restartedDueTo27993 bool 845 } 846 847 // gcMarkDone transitions the GC from mark to mark termination if all 848 // reachable objects have been marked (that is, there are no grey 849 // objects and can be no more in the future). Otherwise, it flushes 850 // all local work to the global queues where it can be discovered by 851 // other workers. 852 // 853 // This should be called when all local mark work has been drained and 854 // there are no remaining workers. Specifically, when 855 // 856 // work.nwait == work.nproc && !gcMarkWorkAvailable(p) 857 // 858 // The calling context must be preemptible. 859 // 860 // Flushing local work is important because idle Ps may have local 861 // work queued. This is the only way to make that work visible and 862 // drive GC to completion. 863 // 864 // It is explicitly okay to have write barriers in this function. If 865 // it does transition to mark termination, then all reachable objects 866 // have been marked, so the write barrier cannot shade any more 867 // objects. 868 func gcMarkDone() { 869 // Ensure only one thread is running the ragged barrier at a 870 // time. 871 semacquire(&work.markDoneSema) 872 873 top: 874 // Re-check transition condition under transition lock. 875 // 876 // It's critical that this checks the global work queues are 877 // empty before performing the ragged barrier. Otherwise, 878 // there could be global work that a P could take after the P 879 // has passed the ragged barrier. 880 if !(gcphase == _GCmark && work.nwait == work.nproc && !gcMarkWorkAvailable(nil)) { 881 semrelease(&work.markDoneSema) 882 return 883 } 884 885 // forEachP needs worldsema to execute, and we'll need it to 886 // stop the world later, so acquire worldsema now. 887 semacquire(&worldsema) 888 889 // Prevent weak->strong conversions from generating additional 890 // GC work. forEachP will guarantee that it is observed globally. 891 work.strongFromWeak.block = true 892 893 // Flush all local buffers and collect flushedWork flags. 894 gcMarkDoneFlushed = 0 895 forEachP(waitReasonGCMarkTermination, func(pp *p) { 896 // Flush the write barrier buffer, since this may add 897 // work to the gcWork. 898 wbBufFlush1(pp) 899 900 // Flush the gcWork, since this may create global work 901 // and set the flushedWork flag. 902 // 903 // TODO(austin): Break up these workbufs to 904 // better distribute work. 905 pp.gcw.dispose() 906 // Collect the flushedWork flag. 907 if pp.gcw.flushedWork { 908 atomic.Xadd(&gcMarkDoneFlushed, 1) 909 pp.gcw.flushedWork = false 910 } 911 }) 912 913 if gcMarkDoneFlushed != 0 { 914 // More grey objects were discovered since the 915 // previous termination check, so there may be more 916 // work to do. Keep going. It's possible the 917 // transition condition became true again during the 918 // ragged barrier, so re-check it. 919 semrelease(&worldsema) 920 goto top 921 } 922 923 // For debugging/testing. 924 for gcDebugMarkDone.spinAfterRaggedBarrier.Load() { 925 } 926 927 // There was no global work, no local work, and no Ps 928 // communicated work since we took markDoneSema. Therefore 929 // there are no grey objects and no more objects can be 930 // shaded. Transition to mark termination. 931 now := nanotime() 932 work.tMarkTerm = now 933 getg().m.preemptoff = "gcing" 934 var stw worldStop 935 systemstack(func() { 936 stw = stopTheWorldWithSema(stwGCMarkTerm) 937 }) 938 // The gcphase is _GCmark, it will transition to _GCmarktermination 939 // below. The important thing is that the wb remains active until 940 // all marking is complete. This includes writes made by the GC. 941 942 // Accumulate fine-grained stopping time. 943 work.cpuStats.accumulateGCPauseTime(stw.stoppingCPUTime, 1) 944 945 // There is sometimes work left over when we enter mark termination due 946 // to write barriers performed after the completion barrier above. 947 // Detect this and resume concurrent mark. This is obviously 948 // unfortunate. 949 // 950 // See issue #27993 for details. 951 // 952 // Switch to the system stack to call wbBufFlush1, though in this case 953 // it doesn't matter because we're non-preemptible anyway. 954 restart := false 955 systemstack(func() { 956 for _, p := range allp { 957 wbBufFlush1(p) 958 if !p.gcw.empty() { 959 restart = true 960 break 961 } 962 } 963 }) 964 if restart { 965 gcDebugMarkDone.restartedDueTo27993 = true 966 967 getg().m.preemptoff = "" 968 systemstack(func() { 969 // Accumulate the time we were stopped before we had to start again. 970 work.cpuStats.accumulateGCPauseTime(nanotime()-stw.finishedStopping, work.maxprocs) 971 972 // Start the world again. 973 now := startTheWorldWithSema(0, stw) 974 work.pauseNS += now - stw.startedStopping 975 }) 976 semrelease(&worldsema) 977 goto top 978 } 979 980 gcComputeStartingStackSize() 981 982 // Disable assists and background workers. We must do 983 // this before waking blocked assists. 984 atomic.Store(&gcBlackenEnabled, 0) 985 986 // Notify the CPU limiter that GC assists will now cease. 987 gcCPULimiter.startGCTransition(false, now) 988 989 // Wake all blocked assists. These will run when we 990 // start the world again. 991 gcWakeAllAssists() 992 993 // Wake all blocked weak->strong conversions. These will run 994 // when we start the world again. 995 work.strongFromWeak.block = false 996 gcWakeAllStrongFromWeak() 997 998 // Likewise, release the transition lock. Blocked 999 // workers and assists will run when we start the 1000 // world again. 1001 semrelease(&work.markDoneSema) 1002 1003 // In STW mode, re-enable user goroutines. These will be 1004 // queued to run after we start the world. 1005 schedEnableUser(true) 1006 1007 // endCycle depends on all gcWork cache stats being flushed. 1008 // The termination algorithm above ensured that up to 1009 // allocations since the ragged barrier. 1010 gcController.endCycle(now, int(gomaxprocs), work.userForced) 1011 1012 // Perform mark termination. This will restart the world. 1013 gcMarkTermination(stw) 1014 } 1015 1016 // World must be stopped and mark assists and background workers must be 1017 // disabled. 1018 func gcMarkTermination(stw worldStop) { 1019 // Start marktermination (write barrier remains enabled for now). 1020 setGCPhase(_GCmarktermination) 1021 1022 work.heap1 = gcController.heapLive.Load() 1023 startTime := nanotime() 1024 1025 mp := acquirem() 1026 mp.preemptoff = "gcing" 1027 mp.traceback = 2 1028 curgp := mp.curg 1029 // N.B. The execution tracer is not aware of this status 1030 // transition and handles it specially based on the 1031 // wait reason. 1032 casGToWaitingForGC(curgp, _Grunning, waitReasonGarbageCollection) 1033 1034 // Run gc on the g0 stack. We do this so that the g stack 1035 // we're currently running on will no longer change. Cuts 1036 // the root set down a bit (g0 stacks are not scanned, and 1037 // we don't need to scan gc's internal state). We also 1038 // need to switch to g0 so we can shrink the stack. 1039 systemstack(func() { 1040 gcMark(startTime) 1041 // Must return immediately. 1042 // The outer function's stack may have moved 1043 // during gcMark (it shrinks stacks, including the 1044 // outer function's stack), so we must not refer 1045 // to any of its variables. Return back to the 1046 // non-system stack to pick up the new addresses 1047 // before continuing. 1048 }) 1049 1050 var stwSwept bool 1051 systemstack(func() { 1052 work.heap2 = work.bytesMarked 1053 if debug.gccheckmark > 0 { 1054 // Run a full non-parallel, stop-the-world 1055 // mark using checkmark bits, to check that we 1056 // didn't forget to mark anything during the 1057 // concurrent mark process. 1058 // 1059 // Turn off gcwaiting because that will force 1060 // gcDrain to return early if this goroutine 1061 // happens to have its preemption flag set. 1062 // This is fine because the world is stopped. 1063 // Restore it after we're done just to be safe. 1064 sched.gcwaiting.Store(false) 1065 startCheckmarks() 1066 gcResetMarkState() 1067 gcMarkRootPrepare() 1068 gcw := &getg().m.p.ptr().gcw 1069 gcDrain(gcw, 0) 1070 wbBufFlush1(getg().m.p.ptr()) 1071 gcw.dispose() 1072 endCheckmarks() 1073 sched.gcwaiting.Store(true) 1074 } 1075 1076 // marking is complete so we can turn the write barrier off 1077 setGCPhase(_GCoff) 1078 stwSwept = gcSweep(work.mode) 1079 }) 1080 1081 mp.traceback = 0 1082 casgstatus(curgp, _Gwaiting, _Grunning) 1083 1084 trace := traceAcquire() 1085 if trace.ok() { 1086 trace.GCDone() 1087 traceRelease(trace) 1088 } 1089 1090 // all done 1091 mp.preemptoff = "" 1092 1093 if gcphase != _GCoff { 1094 throw("gc done but gcphase != _GCoff") 1095 } 1096 1097 // Record heapInUse for scavenger. 1098 memstats.lastHeapInUse = gcController.heapInUse.load() 1099 1100 // Update GC trigger and pacing, as well as downstream consumers 1101 // of this pacing information, for the next cycle. 1102 systemstack(gcControllerCommit) 1103 1104 // Update timing memstats 1105 now := nanotime() 1106 sec, nsec, _ := time_now() 1107 unixNow := sec*1e9 + int64(nsec) 1108 work.pauseNS += now - stw.startedStopping 1109 work.tEnd = now 1110 atomic.Store64(&memstats.last_gc_unix, uint64(unixNow)) // must be Unix time to make sense to user 1111 atomic.Store64(&memstats.last_gc_nanotime, uint64(now)) // monotonic time for us 1112 memstats.pause_ns[memstats.numgc%uint32(len(memstats.pause_ns))] = uint64(work.pauseNS) 1113 memstats.pause_end[memstats.numgc%uint32(len(memstats.pause_end))] = uint64(unixNow) 1114 memstats.pause_total_ns += uint64(work.pauseNS) 1115 1116 // Accumulate CPU stats. 1117 // 1118 // Use maxprocs instead of stwprocs for GC pause time because the total time 1119 // computed in the CPU stats is based on maxprocs, and we want them to be 1120 // comparable. 1121 // 1122 // Pass gcMarkPhase=true to accumulate so we can get all the latest GC CPU stats 1123 // in there too. 1124 work.cpuStats.accumulateGCPauseTime(now-stw.finishedStopping, work.maxprocs) 1125 work.cpuStats.accumulate(now, true) 1126 1127 // Compute overall GC CPU utilization. 1128 // Omit idle marking time from the overall utilization here since it's "free". 1129 memstats.gc_cpu_fraction = float64(work.cpuStats.GCTotalTime-work.cpuStats.GCIdleTime) / float64(work.cpuStats.TotalTime) 1130 1131 // Reset assist time and background time stats. 1132 // 1133 // Do this now, instead of at the start of the next GC cycle, because 1134 // these two may keep accumulating even if the GC is not active. 1135 scavenge.assistTime.Store(0) 1136 scavenge.backgroundTime.Store(0) 1137 1138 // Reset idle time stat. 1139 sched.idleTime.Store(0) 1140 1141 if work.userForced { 1142 memstats.numforcedgc++ 1143 } 1144 1145 // Bump GC cycle count and wake goroutines waiting on sweep. 1146 lock(&work.sweepWaiters.lock) 1147 memstats.numgc++ 1148 injectglist(&work.sweepWaiters.list) 1149 unlock(&work.sweepWaiters.lock) 1150 1151 // Increment the scavenge generation now. 1152 // 1153 // This moment represents peak heap in use because we're 1154 // about to start sweeping. 1155 mheap_.pages.scav.index.nextGen() 1156 1157 // Release the CPU limiter. 1158 gcCPULimiter.finishGCTransition(now) 1159 1160 // Finish the current heap profiling cycle and start a new 1161 // heap profiling cycle. We do this before starting the world 1162 // so events don't leak into the wrong cycle. 1163 mProf_NextCycle() 1164 1165 // There may be stale spans in mcaches that need to be swept. 1166 // Those aren't tracked in any sweep lists, so we need to 1167 // count them against sweep completion until we ensure all 1168 // those spans have been forced out. 1169 // 1170 // If gcSweep fully swept the heap (for example if the sweep 1171 // is not concurrent due to a GODEBUG setting), then we expect 1172 // the sweepLocker to be invalid, since sweeping is done. 1173 // 1174 // N.B. Below we might duplicate some work from gcSweep; this is 1175 // fine as all that work is idempotent within a GC cycle, and 1176 // we're still holding worldsema so a new cycle can't start. 1177 sl := sweep.active.begin() 1178 if !stwSwept && !sl.valid { 1179 throw("failed to set sweep barrier") 1180 } else if stwSwept && sl.valid { 1181 throw("non-concurrent sweep failed to drain all sweep queues") 1182 } 1183 1184 systemstack(func() { 1185 // The memstats updated above must be updated with the world 1186 // stopped to ensure consistency of some values, such as 1187 // sched.idleTime and sched.totaltime. memstats also include 1188 // the pause time (work,pauseNS), forcing computation of the 1189 // total pause time before the pause actually ends. 1190 // 1191 // Here we reuse the same now for start the world so that the 1192 // time added to /sched/pauses/total/gc:seconds will be 1193 // consistent with the value in memstats. 1194 startTheWorldWithSema(now, stw) 1195 }) 1196 1197 // Flush the heap profile so we can start a new cycle next GC. 1198 // This is relatively expensive, so we don't do it with the 1199 // world stopped. 1200 mProf_Flush() 1201 1202 // Prepare workbufs for freeing by the sweeper. We do this 1203 // asynchronously because it can take non-trivial time. 1204 prepareFreeWorkbufs() 1205 1206 // Free stack spans. This must be done between GC cycles. 1207 systemstack(freeStackSpans) 1208 1209 // Ensure all mcaches are flushed. Each P will flush its own 1210 // mcache before allocating, but idle Ps may not. Since this 1211 // is necessary to sweep all spans, we need to ensure all 1212 // mcaches are flushed before we start the next GC cycle. 1213 // 1214 // While we're here, flush the page cache for idle Ps to avoid 1215 // having pages get stuck on them. These pages are hidden from 1216 // the scavenger, so in small idle heaps a significant amount 1217 // of additional memory might be held onto. 1218 // 1219 // Also, flush the pinner cache, to avoid leaking that memory 1220 // indefinitely. 1221 forEachP(waitReasonFlushProcCaches, func(pp *p) { 1222 pp.mcache.prepareForSweep() 1223 if pp.status == _Pidle { 1224 systemstack(func() { 1225 lock(&mheap_.lock) 1226 pp.pcache.flush(&mheap_.pages) 1227 unlock(&mheap_.lock) 1228 }) 1229 } 1230 pp.pinnerCache = nil 1231 }) 1232 if sl.valid { 1233 // Now that we've swept stale spans in mcaches, they don't 1234 // count against unswept spans. 1235 // 1236 // Note: this sweepLocker may not be valid if sweeping had 1237 // already completed during the STW. See the corresponding 1238 // begin() call that produced sl. 1239 sweep.active.end(sl) 1240 } 1241 1242 // Print gctrace before dropping worldsema. As soon as we drop 1243 // worldsema another cycle could start and smash the stats 1244 // we're trying to print. 1245 if debug.gctrace > 0 { 1246 util := int(memstats.gc_cpu_fraction * 100) 1247 1248 var sbuf [24]byte 1249 printlock() 1250 print("gc ", memstats.numgc, 1251 " @", string(itoaDiv(sbuf[:], uint64(work.tSweepTerm-runtimeInitTime)/1e6, 3)), "s ", 1252 util, "%: ") 1253 prev := work.tSweepTerm 1254 for i, ns := range []int64{work.tMark, work.tMarkTerm, work.tEnd} { 1255 if i != 0 { 1256 print("+") 1257 } 1258 print(string(fmtNSAsMS(sbuf[:], uint64(ns-prev)))) 1259 prev = ns 1260 } 1261 print(" ms clock, ") 1262 for i, ns := range []int64{ 1263 int64(work.stwprocs) * (work.tMark - work.tSweepTerm), 1264 gcController.assistTime.Load(), 1265 gcController.dedicatedMarkTime.Load() + gcController.fractionalMarkTime.Load(), 1266 gcController.idleMarkTime.Load(), 1267 int64(work.stwprocs) * (work.tEnd - work.tMarkTerm), 1268 } { 1269 if i == 2 || i == 3 { 1270 // Separate mark time components with /. 1271 print("/") 1272 } else if i != 0 { 1273 print("+") 1274 } 1275 print(string(fmtNSAsMS(sbuf[:], uint64(ns)))) 1276 } 1277 print(" ms cpu, ", 1278 work.heap0>>20, "->", work.heap1>>20, "->", work.heap2>>20, " MB, ", 1279 gcController.lastHeapGoal>>20, " MB goal, ", 1280 gcController.lastStackScan.Load()>>20, " MB stacks, ", 1281 gcController.globalsScan.Load()>>20, " MB globals, ", 1282 work.maxprocs, " P") 1283 if work.userForced { 1284 print(" (forced)") 1285 } 1286 print("\n") 1287 printunlock() 1288 } 1289 1290 // Set any arena chunks that were deferred to fault. 1291 lock(&userArenaState.lock) 1292 faultList := userArenaState.fault 1293 userArenaState.fault = nil 1294 unlock(&userArenaState.lock) 1295 for _, lc := range faultList { 1296 lc.mspan.setUserArenaChunkToFault() 1297 } 1298 1299 // Enable huge pages on some metadata if we cross a heap threshold. 1300 if gcController.heapGoal() > minHeapForMetadataHugePages { 1301 systemstack(func() { 1302 mheap_.enableMetadataHugePages() 1303 }) 1304 } 1305 1306 semrelease(&worldsema) 1307 semrelease(&gcsema) 1308 // Careful: another GC cycle may start now. 1309 1310 releasem(mp) 1311 mp = nil 1312 1313 // now that gc is done, kick off finalizer thread if needed 1314 if !concurrentSweep { 1315 // give the queued finalizers, if any, a chance to run 1316 Gosched() 1317 } 1318 } 1319 1320 // gcBgMarkStartWorkers prepares background mark worker goroutines. These 1321 // goroutines will not run until the mark phase, but they must be started while 1322 // the work is not stopped and from a regular G stack. The caller must hold 1323 // worldsema. 1324 func gcBgMarkStartWorkers() { 1325 // Background marking is performed by per-P G's. Ensure that each P has 1326 // a background GC G. 1327 // 1328 // Worker Gs don't exit if gomaxprocs is reduced. If it is raised 1329 // again, we can reuse the old workers; no need to create new workers. 1330 if gcBgMarkWorkerCount >= gomaxprocs { 1331 return 1332 } 1333 1334 // Increment mp.locks when allocating. We are called within gcStart, 1335 // and thus must not trigger another gcStart via an allocation. gcStart 1336 // bails when allocating with locks held, so simulate that for these 1337 // allocations. 1338 // 1339 // TODO(prattmic): cleanup gcStart to use a more explicit "in gcStart" 1340 // check for bailing. 1341 mp := acquirem() 1342 ready := make(chan struct{}, 1) 1343 releasem(mp) 1344 1345 for gcBgMarkWorkerCount < gomaxprocs { 1346 mp := acquirem() // See above, we allocate a closure here. 1347 go gcBgMarkWorker(ready) 1348 releasem(mp) 1349 1350 // N.B. we intentionally wait on each goroutine individually 1351 // rather than starting all in a batch and then waiting once 1352 // afterwards. By running one goroutine at a time, we can take 1353 // advantage of runnext to bounce back and forth between 1354 // workers and this goroutine. In an overloaded application, 1355 // this can reduce GC start latency by prioritizing these 1356 // goroutines rather than waiting on the end of the run queue. 1357 <-ready 1358 // The worker is now guaranteed to be added to the pool before 1359 // its P's next findRunnableGCWorker. 1360 1361 gcBgMarkWorkerCount++ 1362 } 1363 } 1364 1365 // gcBgMarkPrepare sets up state for background marking. 1366 // Mutator assists must not yet be enabled. 1367 func gcBgMarkPrepare() { 1368 // Background marking will stop when the work queues are empty 1369 // and there are no more workers (note that, since this is 1370 // concurrent, this may be a transient state, but mark 1371 // termination will clean it up). Between background workers 1372 // and assists, we don't really know how many workers there 1373 // will be, so we pretend to have an arbitrarily large number 1374 // of workers, almost all of which are "waiting". While a 1375 // worker is working it decrements nwait. If nproc == nwait, 1376 // there are no workers. 1377 work.nproc = ^uint32(0) 1378 work.nwait = ^uint32(0) 1379 } 1380 1381 // gcBgMarkWorkerNode is an entry in the gcBgMarkWorkerPool. It points to a single 1382 // gcBgMarkWorker goroutine. 1383 type gcBgMarkWorkerNode struct { 1384 // Unused workers are managed in a lock-free stack. This field must be first. 1385 node lfnode 1386 1387 // The g of this worker. 1388 gp guintptr 1389 1390 // Release this m on park. This is used to communicate with the unlock 1391 // function, which cannot access the G's stack. It is unused outside of 1392 // gcBgMarkWorker(). 1393 m muintptr 1394 } 1395 1396 func gcBgMarkWorker(ready chan struct{}) { 1397 gp := getg() 1398 1399 // We pass node to a gopark unlock function, so it can't be on 1400 // the stack (see gopark). Prevent deadlock from recursively 1401 // starting GC by disabling preemption. 1402 gp.m.preemptoff = "GC worker init" 1403 node := new(gcBgMarkWorkerNode) 1404 gp.m.preemptoff = "" 1405 1406 node.gp.set(gp) 1407 1408 node.m.set(acquirem()) 1409 1410 ready <- struct{}{} 1411 // After this point, the background mark worker is generally scheduled 1412 // cooperatively by gcController.findRunnableGCWorker. While performing 1413 // work on the P, preemption is disabled because we are working on 1414 // P-local work buffers. When the preempt flag is set, this puts itself 1415 // into _Gwaiting to be woken up by gcController.findRunnableGCWorker 1416 // at the appropriate time. 1417 // 1418 // When preemption is enabled (e.g., while in gcMarkDone), this worker 1419 // may be preempted and schedule as a _Grunnable G from a runq. That is 1420 // fine; it will eventually gopark again for further scheduling via 1421 // findRunnableGCWorker. 1422 // 1423 // Since we disable preemption before notifying ready, we guarantee that 1424 // this G will be in the worker pool for the next findRunnableGCWorker. 1425 // This isn't strictly necessary, but it reduces latency between 1426 // _GCmark starting and the workers starting. 1427 1428 for { 1429 // Go to sleep until woken by 1430 // gcController.findRunnableGCWorker. 1431 gopark(func(g *g, nodep unsafe.Pointer) bool { 1432 node := (*gcBgMarkWorkerNode)(nodep) 1433 1434 if mp := node.m.ptr(); mp != nil { 1435 // The worker G is no longer running; release 1436 // the M. 1437 // 1438 // N.B. it is _safe_ to release the M as soon 1439 // as we are no longer performing P-local mark 1440 // work. 1441 // 1442 // However, since we cooperatively stop work 1443 // when gp.preempt is set, if we releasem in 1444 // the loop then the following call to gopark 1445 // would immediately preempt the G. This is 1446 // also safe, but inefficient: the G must 1447 // schedule again only to enter gopark and park 1448 // again. Thus, we defer the release until 1449 // after parking the G. 1450 releasem(mp) 1451 } 1452 1453 // Release this G to the pool. 1454 gcBgMarkWorkerPool.push(&node.node) 1455 // Note that at this point, the G may immediately be 1456 // rescheduled and may be running. 1457 return true 1458 }, unsafe.Pointer(node), waitReasonGCWorkerIdle, traceBlockSystemGoroutine, 0) 1459 1460 // Preemption must not occur here, or another G might see 1461 // p.gcMarkWorkerMode. 1462 1463 // Disable preemption so we can use the gcw. If the 1464 // scheduler wants to preempt us, we'll stop draining, 1465 // dispose the gcw, and then preempt. 1466 node.m.set(acquirem()) 1467 pp := gp.m.p.ptr() // P can't change with preemption disabled. 1468 1469 if gcBlackenEnabled == 0 { 1470 println("worker mode", pp.gcMarkWorkerMode) 1471 throw("gcBgMarkWorker: blackening not enabled") 1472 } 1473 1474 if pp.gcMarkWorkerMode == gcMarkWorkerNotWorker { 1475 throw("gcBgMarkWorker: mode not set") 1476 } 1477 1478 startTime := nanotime() 1479 pp.gcMarkWorkerStartTime = startTime 1480 var trackLimiterEvent bool 1481 if pp.gcMarkWorkerMode == gcMarkWorkerIdleMode { 1482 trackLimiterEvent = pp.limiterEvent.start(limiterEventIdleMarkWork, startTime) 1483 } 1484 1485 decnwait := atomic.Xadd(&work.nwait, -1) 1486 if decnwait == work.nproc { 1487 println("runtime: work.nwait=", decnwait, "work.nproc=", work.nproc) 1488 throw("work.nwait was > work.nproc") 1489 } 1490 1491 systemstack(func() { 1492 // Mark our goroutine preemptible so its stack 1493 // can be scanned. This lets two mark workers 1494 // scan each other (otherwise, they would 1495 // deadlock). We must not modify anything on 1496 // the G stack. However, stack shrinking is 1497 // disabled for mark workers, so it is safe to 1498 // read from the G stack. 1499 // 1500 // N.B. The execution tracer is not aware of this status 1501 // transition and handles it specially based on the 1502 // wait reason. 1503 casGToWaitingForGC(gp, _Grunning, waitReasonGCWorkerActive) 1504 switch pp.gcMarkWorkerMode { 1505 default: 1506 throw("gcBgMarkWorker: unexpected gcMarkWorkerMode") 1507 case gcMarkWorkerDedicatedMode: 1508 gcDrainMarkWorkerDedicated(&pp.gcw, true) 1509 if gp.preempt { 1510 // We were preempted. This is 1511 // a useful signal to kick 1512 // everything out of the run 1513 // queue so it can run 1514 // somewhere else. 1515 if drainQ, n := runqdrain(pp); n > 0 { 1516 lock(&sched.lock) 1517 globrunqputbatch(&drainQ, int32(n)) 1518 unlock(&sched.lock) 1519 } 1520 } 1521 // Go back to draining, this time 1522 // without preemption. 1523 gcDrainMarkWorkerDedicated(&pp.gcw, false) 1524 case gcMarkWorkerFractionalMode: 1525 gcDrainMarkWorkerFractional(&pp.gcw) 1526 case gcMarkWorkerIdleMode: 1527 gcDrainMarkWorkerIdle(&pp.gcw) 1528 } 1529 casgstatus(gp, _Gwaiting, _Grunning) 1530 }) 1531 1532 // Account for time and mark us as stopped. 1533 now := nanotime() 1534 duration := now - startTime 1535 gcController.markWorkerStop(pp.gcMarkWorkerMode, duration) 1536 if trackLimiterEvent { 1537 pp.limiterEvent.stop(limiterEventIdleMarkWork, now) 1538 } 1539 if pp.gcMarkWorkerMode == gcMarkWorkerFractionalMode { 1540 atomic.Xaddint64(&pp.gcFractionalMarkTime, duration) 1541 } 1542 1543 // Was this the last worker and did we run out 1544 // of work? 1545 incnwait := atomic.Xadd(&work.nwait, +1) 1546 if incnwait > work.nproc { 1547 println("runtime: p.gcMarkWorkerMode=", pp.gcMarkWorkerMode, 1548 "work.nwait=", incnwait, "work.nproc=", work.nproc) 1549 throw("work.nwait > work.nproc") 1550 } 1551 1552 // We'll releasem after this point and thus this P may run 1553 // something else. We must clear the worker mode to avoid 1554 // attributing the mode to a different (non-worker) G in 1555 // tracev2.GoStart. 1556 pp.gcMarkWorkerMode = gcMarkWorkerNotWorker 1557 1558 // If this worker reached a background mark completion 1559 // point, signal the main GC goroutine. 1560 if incnwait == work.nproc && !gcMarkWorkAvailable(nil) { 1561 // We don't need the P-local buffers here, allow 1562 // preemption because we may schedule like a regular 1563 // goroutine in gcMarkDone (block on locks, etc). 1564 releasem(node.m.ptr()) 1565 node.m.set(nil) 1566 1567 gcMarkDone() 1568 } 1569 } 1570 } 1571 1572 // gcMarkWorkAvailable reports whether executing a mark worker 1573 // on p is potentially useful. p may be nil, in which case it only 1574 // checks the global sources of work. 1575 func gcMarkWorkAvailable(p *p) bool { 1576 if p != nil && !p.gcw.empty() { 1577 return true 1578 } 1579 if !work.full.empty() { 1580 return true // global work available 1581 } 1582 if work.markrootNext < work.markrootJobs { 1583 return true // root scan work available 1584 } 1585 return false 1586 } 1587 1588 // gcMark runs the mark (or, for concurrent GC, mark termination) 1589 // All gcWork caches must be empty. 1590 // STW is in effect at this point. 1591 func gcMark(startTime int64) { 1592 if gcphase != _GCmarktermination { 1593 throw("in gcMark expecting to see gcphase as _GCmarktermination") 1594 } 1595 work.tstart = startTime 1596 1597 // Check that there's no marking work remaining. 1598 if work.full != 0 || work.markrootNext < work.markrootJobs { 1599 print("runtime: full=", hex(work.full), " next=", work.markrootNext, " jobs=", work.markrootJobs, " nDataRoots=", work.nDataRoots, " nBSSRoots=", work.nBSSRoots, " nSpanRoots=", work.nSpanRoots, " nStackRoots=", work.nStackRoots, "\n") 1600 panic("non-empty mark queue after concurrent mark") 1601 } 1602 1603 if debug.gccheckmark > 0 { 1604 // This is expensive when there's a large number of 1605 // Gs, so only do it if checkmark is also enabled. 1606 gcMarkRootCheck() 1607 } 1608 1609 // Drop allg snapshot. allgs may have grown, in which case 1610 // this is the only reference to the old backing store and 1611 // there's no need to keep it around. 1612 work.stackRoots = nil 1613 1614 // Clear out buffers and double-check that all gcWork caches 1615 // are empty. This should be ensured by gcMarkDone before we 1616 // enter mark termination. 1617 // 1618 // TODO: We could clear out buffers just before mark if this 1619 // has a non-negligible impact on STW time. 1620 for _, p := range allp { 1621 // The write barrier may have buffered pointers since 1622 // the gcMarkDone barrier. However, since the barrier 1623 // ensured all reachable objects were marked, all of 1624 // these must be pointers to black objects. Hence we 1625 // can just discard the write barrier buffer. 1626 if debug.gccheckmark > 0 { 1627 // For debugging, flush the buffer and make 1628 // sure it really was all marked. 1629 wbBufFlush1(p) 1630 } else { 1631 p.wbBuf.reset() 1632 } 1633 1634 gcw := &p.gcw 1635 if !gcw.empty() { 1636 printlock() 1637 print("runtime: P ", p.id, " flushedWork ", gcw.flushedWork) 1638 if gcw.wbuf1 == nil { 1639 print(" wbuf1=<nil>") 1640 } else { 1641 print(" wbuf1.n=", gcw.wbuf1.nobj) 1642 } 1643 if gcw.wbuf2 == nil { 1644 print(" wbuf2=<nil>") 1645 } else { 1646 print(" wbuf2.n=", gcw.wbuf2.nobj) 1647 } 1648 print("\n") 1649 throw("P has cached GC work at end of mark termination") 1650 } 1651 // There may still be cached empty buffers, which we 1652 // need to flush since we're going to free them. Also, 1653 // there may be non-zero stats because we allocated 1654 // black after the gcMarkDone barrier. 1655 gcw.dispose() 1656 } 1657 1658 // Flush scanAlloc from each mcache since we're about to modify 1659 // heapScan directly. If we were to flush this later, then scanAlloc 1660 // might have incorrect information. 1661 // 1662 // Note that it's not important to retain this information; we know 1663 // exactly what heapScan is at this point via scanWork. 1664 for _, p := range allp { 1665 c := p.mcache 1666 if c == nil { 1667 continue 1668 } 1669 c.scanAlloc = 0 1670 } 1671 1672 // Reset controller state. 1673 gcController.resetLive(work.bytesMarked) 1674 } 1675 1676 // gcSweep must be called on the system stack because it acquires the heap 1677 // lock. See mheap for details. 1678 // 1679 // Returns true if the heap was fully swept by this function. 1680 // 1681 // The world must be stopped. 1682 // 1683 //go:systemstack 1684 func gcSweep(mode gcMode) bool { 1685 assertWorldStopped() 1686 1687 if gcphase != _GCoff { 1688 throw("gcSweep being done but phase is not GCoff") 1689 } 1690 1691 lock(&mheap_.lock) 1692 mheap_.sweepgen += 2 1693 sweep.active.reset() 1694 mheap_.pagesSwept.Store(0) 1695 mheap_.sweepArenas = mheap_.heapArenas 1696 mheap_.reclaimIndex.Store(0) 1697 mheap_.reclaimCredit.Store(0) 1698 unlock(&mheap_.lock) 1699 1700 sweep.centralIndex.clear() 1701 1702 if !concurrentSweep || mode == gcForceBlockMode { 1703 // Special case synchronous sweep. 1704 // Record that no proportional sweeping has to happen. 1705 lock(&mheap_.lock) 1706 mheap_.sweepPagesPerByte = 0 1707 unlock(&mheap_.lock) 1708 // Flush all mcaches. 1709 for _, pp := range allp { 1710 pp.mcache.prepareForSweep() 1711 } 1712 // Sweep all spans eagerly. 1713 for sweepone() != ^uintptr(0) { 1714 } 1715 // Free workbufs eagerly. 1716 prepareFreeWorkbufs() 1717 for freeSomeWbufs(false) { 1718 } 1719 // All "free" events for this mark/sweep cycle have 1720 // now happened, so we can make this profile cycle 1721 // available immediately. 1722 mProf_NextCycle() 1723 mProf_Flush() 1724 return true 1725 } 1726 1727 // Background sweep. 1728 lock(&sweep.lock) 1729 if sweep.parked { 1730 sweep.parked = false 1731 ready(sweep.g, 0, true) 1732 } 1733 unlock(&sweep.lock) 1734 return false 1735 } 1736 1737 // gcResetMarkState resets global state prior to marking (concurrent 1738 // or STW) and resets the stack scan state of all Gs. 1739 // 1740 // This is safe to do without the world stopped because any Gs created 1741 // during or after this will start out in the reset state. 1742 // 1743 // gcResetMarkState must be called on the system stack because it acquires 1744 // the heap lock. See mheap for details. 1745 // 1746 //go:systemstack 1747 func gcResetMarkState() { 1748 // This may be called during a concurrent phase, so lock to make sure 1749 // allgs doesn't change. 1750 forEachG(func(gp *g) { 1751 gp.gcscandone = false // set to true in gcphasework 1752 gp.gcAssistBytes = 0 1753 }) 1754 1755 // Clear page marks. This is just 1MB per 64GB of heap, so the 1756 // time here is pretty trivial. 1757 lock(&mheap_.lock) 1758 arenas := mheap_.heapArenas 1759 unlock(&mheap_.lock) 1760 for _, ai := range arenas { 1761 ha := mheap_.arenas[ai.l1()][ai.l2()] 1762 clear(ha.pageMarks[:]) 1763 } 1764 1765 work.bytesMarked = 0 1766 work.initialHeapLive = gcController.heapLive.Load() 1767 } 1768 1769 // Hooks for other packages 1770 1771 var poolcleanup func() 1772 var boringCaches []unsafe.Pointer // for crypto/internal/boring 1773 var uniqueMapCleanup chan struct{} // for unique 1774 1775 // sync_runtime_registerPoolCleanup should be an internal detail, 1776 // but widely used packages access it using linkname. 1777 // Notable members of the hall of shame include: 1778 // - github.com/bytedance/gopkg 1779 // - github.com/songzhibin97/gkit 1780 // 1781 // Do not remove or change the type signature. 1782 // See go.dev/issue/67401. 1783 // 1784 //go:linkname sync_runtime_registerPoolCleanup sync.runtime_registerPoolCleanup 1785 func sync_runtime_registerPoolCleanup(f func()) { 1786 poolcleanup = f 1787 } 1788 1789 //go:linkname boring_registerCache crypto/internal/boring/bcache.registerCache 1790 func boring_registerCache(p unsafe.Pointer) { 1791 boringCaches = append(boringCaches, p) 1792 } 1793 1794 //go:linkname unique_runtime_registerUniqueMapCleanup unique.runtime_registerUniqueMapCleanup 1795 func unique_runtime_registerUniqueMapCleanup(f func()) { 1796 // Create the channel on the system stack so it doesn't inherit the current G's 1797 // synctest bubble (if any). 1798 systemstack(func() { 1799 uniqueMapCleanup = make(chan struct{}, 1) 1800 }) 1801 // Start the goroutine in the runtime so it's counted as a system goroutine. 1802 go func(cleanup func()) { 1803 for { 1804 <-uniqueMapCleanup 1805 cleanup() 1806 } 1807 }(f) 1808 } 1809 1810 func clearpools() { 1811 // clear sync.Pools 1812 if poolcleanup != nil { 1813 poolcleanup() 1814 } 1815 1816 // clear boringcrypto caches 1817 for _, p := range boringCaches { 1818 atomicstorep(p, nil) 1819 } 1820 1821 // clear unique maps 1822 if uniqueMapCleanup != nil { 1823 select { 1824 case uniqueMapCleanup <- struct{}{}: 1825 default: 1826 } 1827 } 1828 1829 // Clear central sudog cache. 1830 // Leave per-P caches alone, they have strictly bounded size. 1831 // Disconnect cached list before dropping it on the floor, 1832 // so that a dangling ref to one entry does not pin all of them. 1833 lock(&sched.sudoglock) 1834 var sg, sgnext *sudog 1835 for sg = sched.sudogcache; sg != nil; sg = sgnext { 1836 sgnext = sg.next 1837 sg.next = nil 1838 } 1839 sched.sudogcache = nil 1840 unlock(&sched.sudoglock) 1841 1842 // Clear central defer pool. 1843 // Leave per-P pools alone, they have strictly bounded size. 1844 lock(&sched.deferlock) 1845 // disconnect cached list before dropping it on the floor, 1846 // so that a dangling ref to one entry does not pin all of them. 1847 var d, dlink *_defer 1848 for d = sched.deferpool; d != nil; d = dlink { 1849 dlink = d.link 1850 d.link = nil 1851 } 1852 sched.deferpool = nil 1853 unlock(&sched.deferlock) 1854 } 1855 1856 // Timing 1857 1858 // itoaDiv formats val/(10**dec) into buf. 1859 func itoaDiv(buf []byte, val uint64, dec int) []byte { 1860 i := len(buf) - 1 1861 idec := i - dec 1862 for val >= 10 || i >= idec { 1863 buf[i] = byte(val%10 + '0') 1864 i-- 1865 if i == idec { 1866 buf[i] = '.' 1867 i-- 1868 } 1869 val /= 10 1870 } 1871 buf[i] = byte(val + '0') 1872 return buf[i:] 1873 } 1874 1875 // fmtNSAsMS nicely formats ns nanoseconds as milliseconds. 1876 func fmtNSAsMS(buf []byte, ns uint64) []byte { 1877 if ns >= 10e6 { 1878 // Format as whole milliseconds. 1879 return itoaDiv(buf, ns/1e6, 0) 1880 } 1881 // Format two digits of precision, with at most three decimal places. 1882 x := ns / 1e3 1883 if x == 0 { 1884 buf[0] = '0' 1885 return buf[:1] 1886 } 1887 dec := 3 1888 for x >= 100 { 1889 x /= 10 1890 dec-- 1891 } 1892 return itoaDiv(buf, x, dec) 1893 } 1894 1895 // Helpers for testing GC. 1896 1897 // gcTestMoveStackOnNextCall causes the stack to be moved on a call 1898 // immediately following the call to this. It may not work correctly 1899 // if any other work appears after this call (such as returning). 1900 // Typically the following call should be marked go:noinline so it 1901 // performs a stack check. 1902 // 1903 // In rare cases this may not cause the stack to move, specifically if 1904 // there's a preemption between this call and the next. 1905 func gcTestMoveStackOnNextCall() { 1906 gp := getg() 1907 gp.stackguard0 = stackForceMove 1908 } 1909 1910 // gcTestIsReachable performs a GC and returns a bit set where bit i 1911 // is set if ptrs[i] is reachable. 1912 func gcTestIsReachable(ptrs ...unsafe.Pointer) (mask uint64) { 1913 // This takes the pointers as unsafe.Pointers in order to keep 1914 // them live long enough for us to attach specials. After 1915 // that, we drop our references to them. 1916 1917 if len(ptrs) > 64 { 1918 panic("too many pointers for uint64 mask") 1919 } 1920 1921 // Block GC while we attach specials and drop our references 1922 // to ptrs. Otherwise, if a GC is in progress, it could mark 1923 // them reachable via this function before we have a chance to 1924 // drop them. 1925 semacquire(&gcsema) 1926 1927 // Create reachability specials for ptrs. 1928 specials := make([]*specialReachable, len(ptrs)) 1929 for i, p := range ptrs { 1930 lock(&mheap_.speciallock) 1931 s := (*specialReachable)(mheap_.specialReachableAlloc.alloc()) 1932 unlock(&mheap_.speciallock) 1933 s.special.kind = _KindSpecialReachable 1934 if !addspecial(p, &s.special, false) { 1935 throw("already have a reachable special (duplicate pointer?)") 1936 } 1937 specials[i] = s 1938 // Make sure we don't retain ptrs. 1939 ptrs[i] = nil 1940 } 1941 1942 semrelease(&gcsema) 1943 1944 // Force a full GC and sweep. 1945 GC() 1946 1947 // Process specials. 1948 for i, s := range specials { 1949 if !s.done { 1950 printlock() 1951 println("runtime: object", i, "was not swept") 1952 throw("IsReachable failed") 1953 } 1954 if s.reachable { 1955 mask |= 1 << i 1956 } 1957 lock(&mheap_.speciallock) 1958 mheap_.specialReachableAlloc.free(unsafe.Pointer(s)) 1959 unlock(&mheap_.speciallock) 1960 } 1961 1962 return mask 1963 } 1964 1965 // gcTestPointerClass returns the category of what p points to, one of: 1966 // "heap", "stack", "data", "bss", "other". This is useful for checking 1967 // that a test is doing what it's intended to do. 1968 // 1969 // This is nosplit simply to avoid extra pointer shuffling that may 1970 // complicate a test. 1971 // 1972 //go:nosplit 1973 func gcTestPointerClass(p unsafe.Pointer) string { 1974 p2 := uintptr(noescape(p)) 1975 gp := getg() 1976 if gp.stack.lo <= p2 && p2 < gp.stack.hi { 1977 return "stack" 1978 } 1979 if base, _, _ := findObject(p2, 0, 0); base != 0 { 1980 return "heap" 1981 } 1982 for _, datap := range activeModules() { 1983 if datap.data <= p2 && p2 < datap.edata || datap.noptrdata <= p2 && p2 < datap.enoptrdata { 1984 return "data" 1985 } 1986 if datap.bss <= p2 && p2 < datap.ebss || datap.noptrbss <= p2 && p2 <= datap.enoptrbss { 1987 return "bss" 1988 } 1989 } 1990 KeepAlive(p) 1991 return "other" 1992 } 1993