- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 73 for doPing (0.12 sec)
-
internal/grid/muxclient.go
// Next loop will catch it. case <-pingTimer: if !m.doPing(respHandler) { return } goto sendResp } case <-pingTimer: if !m.doPing(respHandler) { return } } } } // doPing checks last ping time and sends another ping. func (m *muxClient) doPing(respHandler chan<- Response) (ok bool) { m.respMu.Lock() if m.closed {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri Jun 07 15:51:52 UTC 2024 - 15.9K bytes - Viewed (0) -
src/cmd/compile/internal/liveness/plive.go
// treat "dead" writes as equivalent to reads during the analysis; // used only during liveness analysis for stack slot merging (doesn't // make sense for stackmap analysis). conservativeWrites bool } // Map maps from *ssa.Value to StackMapIndex. // Also keeps track of unsafe ssa.Values and ssa.Blocks. // (unsafe = can't be interrupted during GC.) type Map struct { Vals map[ssa.ID]objw.StackMapIndex
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 15:22:22 UTC 2024 - 45.2K bytes - Viewed (0) -
android/guava/src/com/google/common/cache/Striped64.java
* resizing the table, as well as populating slots with new Cells. * There is no need for a blocking lock; when the lock is not * available, threads try other slots (or the base). During these * retries, there is increased contention and reduced locality, * which is still better than alternatives. * * Per-thread hash codes are initialized to random values.
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
guava/src/com/google/common/cache/Striped64.java
* resizing the table, as well as populating slots with new Cells. * There is no need for a blocking lock; when the lock is not * available, threads try other slots (or the base). During these * retries, there is increased contention and reduced locality, * which is still better than alternatives. * * Per-thread hash codes are initialized to random values.
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
android/guava/src/com/google/common/hash/Striped64.java
* resizing the table, as well as populating slots with new Cells. * There is no need for a blocking lock; when the lock is not * available, threads try other slots (or the base). During these * retries, there is increased contention and reduced locality, * which is still better than alternatives. * * Per-thread hash codes are initialized to random values.
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
pkg/scheduler/framework/types.go
} // Match returns true if ClusterEvent is matched with the coming event. // If the ce.Resource is "*", there's no requirement for the coming event' Resource. // Contrarily, if the coming event's Resource is "*", the ce.Resource should only be "*". // // Note: we have a special case here when the coming event is a wildcard event, // it will force all Pods to move to activeQ/backoffQ,
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 19:28:17 UTC 2024 - 36.7K bytes - Viewed (0) -
pkg/kubelet/util/manager/watch_based_manager.go
// - we want to also handle the case when object is marked as immutable later // - Secrets and ConfigMaps are periodically fetched by volumemanager anyway // - doing that wouldn't provide visible scalability/performance gain - we // already have it from here // - doing that would require significant refactoring to reflector // we limit ourselves to just quickly stop the reflector here. if c.isImmutable(object) {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 06:25:43 UTC 2024 - 11.6K bytes - Viewed (0) -
guava-tests/test/com/google/common/util/concurrent/UninterruptiblesTest.java
thread.joinSuccessfully(LONG_DELAY_MS); assertInterrupted(); } public void testJoinTimeoutMultiInterruptExpired() { /* * We don't "need" to schedule a thread completion at all here, but by doing * so, we come the closest we can to testing that the wait time is * appropriately decreased on each progressive join() call. */ TimedThread thread = TimedThread.createWithDelay(LONG_DELAY_MS);
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Jun 10 16:06:39 UTC 2024 - 31.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/mlir_graph_optimization_pass.cc
import_config.upgrade_legacy = true; // Disable shape inference during import as some TensorFlow op fails during // shape inference with dynamic shaped operands. This in turn causes the // import to fail. Shape inference during import is going to be removed and // the shape inference pass is run early in the pass pipeline, shape inference // during import is not necessary. import_config.enable_shape_inference = false;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 22:19:26 UTC 2024 - 18.5K bytes - Viewed (0) -
pkg/kubelet/pleg/generic.go
// Cache for storing the runtime states required for syncing pods. cache kubecontainer.Cache // For testability. clock clock.Clock // Pods that failed to have their status retrieved during a relist. These pods will be // retried during the next relisting. podsToReinspect map[types.UID]*kubecontainer.Pod // Stop the Generic PLEG by closing the channel. stopCh chan struct{} // Locks the relisting of the Generic PLEG
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 06:25:43 UTC 2024 - 19.3K bytes - Viewed (0)