- Sort Score
- Result 10 results
- Languages All
Results 61 - 70 of 147 for preemption (0.75 sec)
-
manifests/charts/istiod-remote/values.yaml
# system-node-critical, it is better to configure this in order to make sure your Istio pods # will not be killed because of low priority class. # Refer to https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass # for more detail. priorityClassName: "" proxy: image: proxyv2 # This controls the 'policy' in the sidecar injector. autoInject: enabled
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Mon Jun 03 01:55:05 UTC 2024 - 21K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/op.go
zeroWidth bool // op never translates into any machine code. example: copy, which may sometimes translate to machine code, is not zero-width. unsafePoint bool // this op is an unsafe point, i.e. not safe for async preemption symEffect SymEffect // effect this op has on symbol in aux scale uint8 // amd64/386 indexed load scale } type inputInfo struct { idx int // index in Args array
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 22 15:29:10 UTC 2024 - 18.7K bytes - Viewed (0) -
manifests/charts/istio-control/istio-discovery/values.yaml
# system-node-critical, it is better to configure this in order to make sure your Istio pods # will not be killed because of low priority class. # Refer to https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass # for more detail. priorityClassName: "" proxy: image: proxyv2 # This controls the 'policy' in the sidecar injector. autoInject: enabled
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Mon Jun 03 01:55:05 UTC 2024 - 21.1K bytes - Viewed (0) -
src/runtime/gc_test.go
} } func TestGCTestMoveStackOnNextCall(t *testing.T) { t.Parallel() var onStack int // GCTestMoveStackOnNextCall can fail in rare cases if there's // a preemption. This won't happen many times in quick // succession, so just retry a few times. for retry := 0; retry < 5; retry++ { runtime.GCTestMoveStackOnNextCall()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jun 05 22:33:52 UTC 2024 - 17.6K bytes - Viewed (0) -
src/runtime/stubs.go
// // frameSize is the total size of the argument frame at stackArgs and must // therefore be >= stackArgsSize. It must include additional space for spilling // register arguments for stack growth and preemption. // // TODO(mknyszek): Once we don't need the additional spill space, remove frameSize, // since frameSize will be redundant with stackArgsSize. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 20.2K bytes - Viewed (0) -
src/runtime/arena.go
// Redirect allocations that don't fit into a chunk well directly // from the heap. if cap >= 0 { return newarray(typ, cap) } return newobject(typ) } // Prevent preemption as we set up the space for a new object. // // Act like we're allocating. mp := acquirem() if mp.mallocing != 0 { throw("malloc deadlock") } if mp.gsignal == getg() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:44:56 UTC 2024 - 37.9K bytes - Viewed (0) -
src/cmd/internal/obj/loong64/obj.go
} p.To.Type = obj.TYPE_REG p.To.Reg = REG_R20 // Mark the stack bound check and morestack call async nonpreemptible. // If we get preempted here, when resumed the preemption request is // cleared, but we'll still call morestack, which will double the stack // unnecessarily. See issue #35470. p = c.ctxt.StartUnsafePoint(p, c.newprog) var q *obj.Prog
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 21 19:22:18 UTC 2023 - 19.7K bytes - Viewed (0) -
src/cmd/internal/obj/arm/obj5.go
} p.To.Type = obj.TYPE_REG p.To.Reg = REG_R1 // Mark the stack bound check and morestack call async nonpreemptible. // If we get preempted here, when resumed the preemption request is // cleared, but we'll still call morestack, which will double the stack // unnecessarily. See issue #35470. p = c.ctxt.StartUnsafePoint(p, c.newprog) if framesize <= abi.StackSmall {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Nov 20 17:19:36 UTC 2023 - 21.4K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/writebarrier.go
mem = bEnd.NewValue3A(pos, OpStore, types.TypeMem, w.Aux, ptr, val, mem) } } // The last store becomes the WBend marker. This marker is used by the liveness // pass to determine what parts of the code are preemption-unsafe. // All subsequent memory operations use this memory, so we have to sacrifice the // previous last memory op to become this new value. bEnd.Values = append(bEnd.Values, last) last.Block = bEnd
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Sep 08 19:09:14 UTC 2023 - 23.5K bytes - Viewed (0) -
src/runtime/mheap.go
// then release mheap_.lock. allArenas []arenaIdx // sweepArenas is a snapshot of allArenas taken at the // beginning of the sweep cycle. This can be read safely by // simply blocking GC (by disabling preemption). sweepArenas []arenaIdx // markArenas is a snapshot of allArenas taken at the beginning // of the mark cycle. Because allArenas is append-only, neither
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 78K bytes - Viewed (0)