- Sort Score
- Result 10 results
- Languages All
Results 161 - 170 of 240 for seas (0.28 sec)
-
pkg/kubelet/pod_workers.go
} klog.V(4).InfoS("Processing pod event done", "pod", podRef, "podUID", podUID, "updateType", update.WorkType) } } // acknowledgeTerminating sets the terminating flag on the pod status once the pod worker sees // the termination state so that other components know no new containers will be started in this // pod. It then returns the status function, if any, that applies to this pod.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Apr 02 13:22:37 UTC 2024 - 74.8K bytes - Viewed (0) -
src/runtime/trace.go
// the transition from GoWaiting to GoRunnable. The trace will then be broken, because // future events will be emitted assuming the tracer sees GoRunnable. // // In short, what we really need here is to make sure that the next time *any goroutine* // hits a traceAcquire, it sees that the trace is enabled. // // Note also that stopping the world is necessary to make sure sweep-related events are
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:17:41 UTC 2024 - 37.1K bytes - Viewed (0) -
api/openapi-spec/v3/apis__flowcontrol.apiserver.k8s.io__v1beta3_openapi.json
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Mar 08 04:18:56 UTC 2024 - 232.7K bytes - Viewed (0) -
api/openapi-spec/v3/apis__flowcontrol.apiserver.k8s.io__v1_openapi.json
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Mar 08 04:18:56 UTC 2024 - 231.7K bytes - Viewed (0) -
guava-tests/benchmark/com/google/common/util/concurrent/MonitorBasedArrayBlockingQueue.java
* which we don't support anyway. */ return nextIndex >= 0; } /** * Checks whether nextIndex is valid; if so setting nextItem. Stops iterator when either hits * putIndex or sees null item. */ private void checkNext() { if (nextIndex == putIndex) { nextIndex = -1; nextItem = null; } else { nextItem = items[nextIndex];
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Wed Apr 19 19:24:36 UTC 2023 - 22.5K bytes - Viewed (0) -
src/log/slog/handler.go
defer state.free() if h.json { state.buf.WriteByte('{') } // Built-in attributes. They are not in a group. stateGroups := state.groups state.groups = nil // So ReplaceAttrs sees no groups instead of the pre groups. rep := h.opts.ReplaceAttr // time if !r.Time.IsZero() { key := TimeKey val := r.Time.Round(0) // strip monotonic to match Attr behavior if rep == nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Dec 05 18:18:13 UTC 2023 - 17.5K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache_test.cc
EXPECT_EQ(offset, 0); num_requests++; memset(buffer, 'x', n); notification.Notify(); // Wait for other thread to issue read. Env::Default()->SleepForMicroseconds(100000); // 0.1 secs TF_SetStatus(status, TF_OK, ""); return n; }; tf_gcs_filesystem::RamFileBlockCache cache(block_size, block_size, 0, fetcher);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Oct 15 03:16:57 UTC 2021 - 23.2K bytes - Viewed (0) -
src/cmd/cgo/out.go
// assignments, so it won't use much stack space, so // it's OK to not split the stack. Splitting the stack // can run into a bug in clang (as of 2018-11-09): // this is a leaf function, and when clang sees a leaf // function it won't emit the split stack prologue for // the function. However, if this function refers to a // non-split-stack function, which will happen if the
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Mar 29 16:41:10 UTC 2024 - 59.6K bytes - Viewed (0) -
guava/src/com/google/common/util/concurrent/RateLimiter.java
protected SleepingStopwatch() {} /* * We always hold the mutex when calling this. TODO(cpovirk): Is that important? Perhaps we need * to guarantee that each call to reserveEarliestAvailable, etc. sees a value >= the previous? * Also, is it OK that we don't hold the mutex when sleeping? */ protected abstract long readMicros(); protected abstract void sleepMicrosUninterruptibly(long micros);
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue Apr 04 09:45:04 UTC 2023 - 21.5K bytes - Viewed (0) -
tensorflow/c/kernels_experimental.cc
// because a race condition can happen between this and another thread that // turns off some variable's `copy_on_read_mode` after this thread enables // sparse access; when a later function sees `copy_on_read_mode` is off, it // will try to lock the variable again for updating `copy_on_read_mode` and // cause the deadlock, since the variable mutex is non-re-entrant. for (auto* var : vars) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 23 06:12:29 UTC 2024 - 30.9K bytes - Viewed (0)