- Sort Score
- Result 10 results
- Languages All
Results 11 - 17 of 17 for Preemptable (0.21 sec)
-
src/runtime/mgcsweep.go
// // Note that this condition may transition from false to true at any // time as the sweeper runs. It may transition from true to false if a // GC runs; to prevent that the caller must be non-preemptible or must // somehow block GC progress. func isSweepDone() bool { return sweep.active.isDone() } // Returns only when span s has been swept. // //go:nowritebarrier func (s *mspan) ensureSwept() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:52:18 UTC 2024 - 32.9K bytes - Viewed (0) -
src/runtime/proc.go
// holding the transient P, leaving that P in limbo and deadlocking the // STW. // // Callers passing a non-nil P must already be in non-preemptible // context, otherwise such preemption could occur on function entry to // startm. Callers passing a nil P may be preemptible, so we must // disable preemption before acquiring a P from pidleget below. mp := acquirem() if !lockheld { lock(&sched.lock) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 207.5K bytes - Viewed (0) -
src/runtime/mheap.go
} else { return uint(i) & (1<<arenaL2Bits - 1) } } // inheap reports whether b is a pointer into a (potentially dead) heap object. // It returns false for pointers into mSpanManual spans. // Non-preemptible because it is used by write barriers. // //go:nowritebarrier //go:nosplit func inheap(b uintptr) bool { return spanOfHeap(b) != nil }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 78K bytes - Viewed (0) -
src/cmd/compile/internal/liveness/plive.go
// grows the stack, but not themselves be a safe-point. func (lv *liveness) hasStackMap(v *ssa.Value) bool { if !v.Op.IsCall() { return false } // wbZero and wbCopy are write barriers and // deeply non-preemptible. They are unsafe points and // hence should not have liveness maps. if sym, ok := v.Aux.(*ssa.AuxCall); ok && (sym.Fn == ir.Syms.WBZero || sym.Fn == ir.Syms.WBMove) { return false } return true }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 15:22:22 UTC 2024 - 45.2K bytes - Viewed (0) -
src/internal/trace/order.go
// before we get here (after the transfer out) but that's OK: that new // P won't be in the ProcSyscall state anymore. // // Basically: while we have a preemptible P, don't advance, because we // *know* from the event that we're going to lose it at some point during // the syscall. We shouldn't advance until that happens. if curCtx.P != NoProc { pState, ok := o.pStates[curCtx.P]
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 03 14:56:25 UTC 2024 - 52.4K bytes - Viewed (0) -
cluster/gce/util.sh
gcloud="gcloud beta" fi local preemptible_minions=() if [[ "${PREEMPTIBLE_NODE}" == "true" ]]; then preemptible_minions+=(--preemptible --maintenance-policy TERMINATE) fi local local_ssds=() local_ssd_ext_count=0 if [[ -n "${NODE_LOCAL_SSDS_EXT:-}" ]]; then IFS=";" read -r -a ssdgroups <<< "${NODE_LOCAL_SSDS_EXT:-}"
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 21:47:17 UTC 2024 - 161.6K bytes - Viewed (0) -
src/cmd/internal/obj/x86/asm6.go
} } // Mark nonpreemptible instruction sequences. // The 2-instruction TLS access sequence // MOVQ TLS, BX // MOVQ 0(BX)(TLS*1), BX // is not async preemptible, as if it is preempted and resumed on // a different thread, the TLS address may become invalid. if !CanUse1InsnTLS(ctxt) { useTLS := func(p *obj.Prog) bool {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 15 15:44:14 UTC 2024 - 146.9K bytes - Viewed (0)