- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 273 for wall (3.02 sec)
-
src/runtime/rand.go
//go:linkname rand func rand() uint64 { // Note: We avoid acquirem here so that in the fast path // there is just a getg, an inlined c.Next, and a return. // The performance difference on a 16-core AMD is // 3.7ns/call this way versus 4.3ns/call with acquirem (+16%). mp := getg().m c := &mp.chacha8 for { // Note: c.Next is marked nosplit, // so we don't need to use mp.locks // on the fast path, which is that the
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 31 14:32:47 UTC 2024 - 8K bytes - Viewed (0) -
src/os/exec_unix.go
} // If we can block until Wait4 will succeed immediately, do so. ready, err := p.blockUntilWaitable() if err != nil { return nil, err } if ready { // Mark the process done now, before the call to Wait4, // so that Process.pidSignal will not send a signal. p.pidDeactivate(statusDone) // Acquire a write lock on sigMu to wait for any // active call to the signal method to complete. p.sigMu.Lock()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 10 22:06:47 UTC 2024 - 4.1K bytes - Viewed (0) -
src/runtime/atomic_pointer.go
// atomicwb performs a write barrier before an atomic pointer write. // The caller should guard the call with "if writeBarrier.enabled". // // atomicwb should be an internal detail, // but widely used packages access it using linkname. // Notable members of the hall of shame include: // - github.com/bytedance/gopkg // - github.com/songzhibin97/gkit // // Do not remove or change the type signature.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 4K bytes - Viewed (0) -
src/sync/pool.go
// not a suitable use for a Pool, since the overhead does not amortize well in // that scenario. It is more efficient to have such objects implement their own // free list. // // A Pool must not be copied after first use. // // In the terminology of [the Go memory model], a call to Put(x) “synchronizes before” // a call to [Pool.Get] returning that same value x. // Similarly, a call to New returning x “synchronizes before”
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 9.4K bytes - Viewed (1) -
internal/grid/stream.go
import ( "context" "errors" ) // A Stream is a two-way stream. // All responses *must* be read by the caller. // If the call is canceled through the context, // the appropriate error will be returned. type Stream struct { // responses from the remote server. // Channel will be closed after error or when remote closes. // All responses *must* be read by the caller until either an error is returned or the channel is closed.
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri Jun 07 15:51:52 UTC 2024 - 3.1K bytes - Viewed (0) -
src/net/ipsock.go
// routed to an IPv6 socket - two separate sockets are required if // both address families are to be supported. // See inet6(4) for details. type ipStackCapabilities struct { sync.Once // guards following ipv4Enabled bool ipv6Enabled bool ipv4MappedIPv6Enabled bool } var ipStackCaps ipStackCapabilities // supportsIPv4 reports whether the platform supports IPv4 networking
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 9.4K bytes - Viewed (0) -
guava/src/com/google/common/hash/LittleEndianByteArray.java
| ((source[offset + 3] & 0xFF) << 24); } /** * Indicates that the loading of Unsafe was successful and the load and store operations will be * very efficient. May be useful for calling code to fall back on an alternative implementation * that is slower than Unsafe.get/store but faster than the pure-Java mask-and-shift. */ static boolean usingUnsafe() {
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 9.8K bytes - Viewed (0) -
android/guava/src/com/google/common/hash/LittleEndianByteArray.java
| ((source[offset + 3] & 0xFF) << 24); } /** * Indicates that the loading of Unsafe was successful and the load and store operations will be * very efficient. May be useful for calling code to fall back on an alternative implementation * that is slower than Unsafe.get/store but faster than the pure-Java mask-and-shift. */ static boolean usingUnsafe() {
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 9.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/initialize_variables_in_session_init.cc
if (init_func_op) { return init_func_op; } else if (!session_init_op.getInitializers().empty()) { // When the init function with type "restore_op" is not found, fall back to // taking the init function corresponding to the first symbol in the // initializers list to be backwards-compatible, before // tf_saved_model.initializer_type attribute was introduced.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 23 09:05:47 UTC 2024 - 7K bytes - Viewed (0) -
src/cmd/compile/internal/ssagen/nowb.go
} for !q.Empty() { fn := q.PopLeft().Func // Check fn. if fn.WBPos.IsKnown() { var err strings.Builder call := funcs[fn] for call.target != nil { fmt.Fprintf(&err, "\n\t%v: called by %v", base.FmtPos(call.lineno), call.target.Nname) call = funcs[call.target] } // Seeing this error in a failed CI run? It indicates that // a function in the runtime package marked nowritebarrierrec
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 15 17:29:46 UTC 2024 - 5.9K bytes - Viewed (0)