- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 161 for mp (0.18 sec)
-
src/runtime/os_netbsd_amd64.go
import ( "internal/abi" "unsafe" ) func lwp_mcontext_init(mc *mcontextt, stk unsafe.Pointer, mp *m, gp *g, fn uintptr) { // Machine dependent mcontext initialisation for LWP. mc.__gregs[_REG_RIP] = uint64(abi.FuncPCABI0(lwp_tramp)) mc.__gregs[_REG_RSP] = uint64(uintptr(stk)) mc.__gregs[_REG_R8] = uint64(uintptr(unsafe.Pointer(mp))) mc.__gregs[_REG_R9] = uint64(uintptr(unsafe.Pointer(gp))) mc.__gregs[_REG_R12] = uint64(fn)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 21 22:12:04 UTC 2021 - 614 bytes - Viewed (0) -
cmd/iam-store.go
func (mp MappedPolicy) toSlice() []string { var policies []string for _, policy := range strings.Split(mp.Policies, ",") { if strings.TrimSpace(policy) == "" { continue } policies = append(policies, policy) } return policies } func (mp MappedPolicy) policySet() set.StringSet { return set.CreateStringSet(mp.toSlice()...) }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Jun 13 22:26:38 UTC 2024 - 75.8K bytes - Viewed (0) -
src/runtime/tracecpu.go
// profiling buffer. It is called from a signal handler, so is limited in what // it can do. mp must be the thread that is currently stopped in a signal. func traceCPUSample(gp *g, mp *m, pp *p, stk []uintptr) { if !traceEnabled() { // Tracing is usually turned off; don't spend time acquiring the signal // lock unless it's active. return } if mp == nil { // Drop samples that don't have an identifiable thread. We can't render
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 17:03:35 UTC 2024 - 8.7K bytes - Viewed (0) -
pkg/config/model_test.go
func TestToMap(t *testing.T) { cases := []struct { input Spec mp map[string]any }{ // Istio type { input: &networking.VirtualService{Gateways: []string{"foobar"}}, mp: map[string]any{ "gateways": []any{"foobar"}, }, }, // Kubernetes type { input: &corev1.PodSpec{ServiceAccountName: "foobar"}, mp: map[string]any{ "serviceAccountName": "foobar", }, },
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Fri Apr 12 17:37:32 UTC 2024 - 7.9K bytes - Viewed (0) -
src/runtime/pinner.go
func (p *Pinner) Pin(pointer any) { if p.pinner == nil { // Check the pinner cache first. mp := acquirem() if pp := mp.p.ptr(); pp != nil { p.pinner = pp.pinnerCache pp.pinnerCache = nil } releasem(mp) if p.pinner == nil { // Didn't get anything from the pinner cache. p.pinner = new(pinner) p.refs = p.refStore[:0]
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 04 14:29:45 UTC 2024 - 11K bytes - Viewed (0) -
src/runtime/netpoll_windows.go
return gList{}, 0 } var entries [64]overlappedEntry var wait uint32 var toRun gList mp := getg().m if delay >= 1e15 { // An arbitrary cap on how long to wait for a timer. // 1e15 ns == ~11.5 days. delay = 1e15 } if delay > 0 && mp.waitIocpHandle != 0 { // GetQueuedCompletionStatusEx doesn't use a high resolution timer internally,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Mar 26 19:58:28 UTC 2024 - 9.4K bytes - Viewed (0) -
src/runtime/preempt_nonwindows.go
// Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. //go:build !windows package runtime //go:nosplit func osPreemptExtEnter(mp *m) {} //go:nosplit
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 28 18:17:57 UTC 2021 - 290 bytes - Viewed (0) -
src/runtime/os_netbsd_arm.go
import ( "internal/abi" "unsafe" ) func lwp_mcontext_init(mc *mcontextt, stk unsafe.Pointer, mp *m, gp *g, fn uintptr) { // Machine dependent mcontext initialisation for LWP. mc.__gregs[_REG_R15] = uint32(abi.FuncPCABI0(lwp_tramp)) mc.__gregs[_REG_R13] = uint32(uintptr(stk)) mc.__gregs[_REG_R0] = uint32(uintptr(unsafe.Pointer(mp))) mc.__gregs[_REG_R1] = uint32(uintptr(unsafe.Pointer(gp))) mc.__gregs[_REG_R2] = uint32(fn) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Dec 05 20:34:30 UTC 2023 - 1.1K bytes - Viewed (0) -
src/runtime/export_debug_test.go
// a signal handler. Add the go:nowritebarrierrec annotation and restructure // this to avoid write barriers. switch h.gp.atomicstatus.Load() { case _Grunning: if getg().m != h.mp { println("trap on wrong M", getg().m, h.mp) return false } // Save the signal context h.saveSigContext(ctxt) // Set PC to debugCallV2. ctxt.setsigpc(uint64(abi.FuncPCABIInternal(debugCallV2)))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 15:41:45 UTC 2024 - 5.1K bytes - Viewed (0) -
src/runtime/trace.go
// buffers before we snapshotted it to begin with. lock(&sched.lock) mToFlush := allm for mp := mToFlush; mp != nil; mp = mp.alllink { mp.trace.link = mp.alllink } for mp := sched.freem; mp != nil; mp = mp.freelink { mp.trace.link = mToFlush mToFlush = mp } unlock(&sched.lock) // Iterate over our snapshot, flushing every buffer until we're done. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:17:41 UTC 2024 - 37.1K bytes - Viewed (0)