- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 37 for raceacquireg (0.22 sec)
-
src/cmd/vendor/golang.org/x/sys/plan9/race0.go
// Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. //go:build plan9 && !race package plan9 import ( "unsafe" ) const raceenabled = false func raceAcquire(addr unsafe.Pointer) { } func raceReleaseMerge(addr unsafe.Pointer) { } func raceReadRange(addr unsafe.Pointer, len int) { } func raceWriteRange(addr unsafe.Pointer, len int) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:33 UTC 2023 - 450 bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/sys/unix/race0.go
//go:build aix || (darwin && !race) || (linux && !race) || (freebsd && !race) || netbsd || openbsd || solaris || dragonfly || zos package unix import ( "unsafe" ) const raceenabled = false func raceAcquire(addr unsafe.Pointer) { } func raceReleaseMerge(addr unsafe.Pointer) { } func raceReadRange(addr unsafe.Pointer, len int) { } func raceWriteRange(addr unsafe.Pointer, len int) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:33 UTC 2023 - 553 bytes - Viewed (0) -
src/runtime/debugcall.go
gp.schedlink.set(newg) }) // Switch to the new goroutine. mcall(func(gp *g) { // Get newg. newg := gp.schedlink.ptr() gp.schedlink = 0 // Park the calling goroutine. trace := traceAcquire() if trace.ok() { // Trace the event before the transition. It may take a // stack trace, but we won't own the stack after the // transition anymore. trace.GoPark(traceBlockDebugCall, 1) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 05 20:50:21 UTC 2024 - 7.1K bytes - Viewed (0) -
src/runtime/mcentral.go
// Deduct credit for this span allocation and sweep if necessary. spanBytes := uintptr(class_to_allocnpages[c.spanclass.sizeclass()]) * _PageSize deductSweepCredit(spanBytes, 0) traceDone := false trace := traceAcquire() if trace.ok() { trace.GCSweepStart() traceRelease(trace) } // If we sweep spanBudget spans without finding any free // space, just allocate a fresh span. This limits the amount
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0) -
src/runtime/tracecpu.go
// Like the runtime/pprof package, even if that bug didn't exist // we would still want to do a goroutine-level sleep in between // reads to avoid frequent wakeups. trace.cpuSleep.sleep(100_000_000) tl := traceAcquire() if !tl.ok() { // Tracing disabled. break } keepGoing := traceReadCPU(tl.gen) traceRelease(tl) if !keepGoing { break } } done <- struct{}{} }()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 17:03:35 UTC 2024 - 8.7K bytes - Viewed (0) -
src/runtime/cgocall.go
// used again if callback decide to make syscall. winsyscall := mp.winsyscall exitsyscall() getg().m.winsyscall = winsyscall // Note that raceacquire must be called only after exitsyscall has // wired this M to a P. if raceenabled { raceacquire(unsafe.Pointer(&racecgosync)) } // From the garbage collector's perspective, time can move // backwards in the sequence above. If there's a callback into
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:16:47 UTC 2024 - 24.2K bytes - Viewed (0) -
src/runtime/proc.go
// might call into the tracer, and the tracer is non-reentrant. trace := traceAcquire() if s == _Psyscall && p2.runSafePointFn == 1 && atomic.Cas(&p2.status, s, _Pidle) { if trace.ok() { // It's important that we traceRelease before we call handoffp, which may also traceAcquire. trace.ProcSteal(p2, false) traceRelease(trace) } p2.syscalltick++ handoffp(p2)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 207.5K bytes - Viewed (0) -
src/runtime/traceallocfree.go
w.varint(uint64(pageSize)) w.varint(uint64(minHeapAlign)) w.varint(uint64(fixedStack)) // Finish writing the batch. w.flush().end() // Start tracing. trace := traceAcquire() if !trace.ok() { throw("traceSnapshotMemory: tracing is not enabled") } // Write out all the heap spans and heap objects. for _, s := range mheap_.allspans { if s.state.get() == mSpanDead {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:32:51 UTC 2024 - 5.9K bytes - Viewed (0) -
src/runtime/syscall_windows.go
} func cbsLock() { lock(&cbs.lock) // compileCallback is used by goenvs prior to completion of schedinit. // raceacquire involves a racecallback to get the proc, which is not // safe prior to scheduler initialization. Thus avoid instrumentation // until then. if raceenabled && mainStarted { raceacquire(unsafe.Pointer(&cbs.lock)) } } func cbsUnlock() { if raceenabled && mainStarted {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:12:46 UTC 2024 - 16.6K bytes - Viewed (0) -
src/runtime/mgcsweep.go
print("mspan.sweep: state=", state, " sweepgen=", s.sweepgen, " mheap.sweepgen=", sweepgen, "\n") throw("mspan.sweep: bad span state") } trace := traceAcquire() if trace.ok() { trace.GCSweepSpan(s.npages * _PageSize) traceRelease(trace) } mheap_.pagesSwept.Add(int64(s.npages)) spc := s.spanclass size := s.elemsize
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:52:18 UTC 2024 - 32.9K bytes - Viewed (0)