- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 453 for incremented (0.17 sec)
-
src/runtime/proc.go
} func mspinning() { // startm's caller incremented nmspinning. Set the new M's spinning. getg().m.spinning = true } // Schedules some M to run the p (creates an M if necessary). // If p==nil, tries to get an idle P, if no idle P's does nothing. // May run with m.p==nil, so write barriers are not allowed. // If spinning is set, the caller has incremented nmspinning and must provide a
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 207.5K bytes - Viewed (0) -
src/runtime/mgc.go
lock mutex list gList } // cycles is the number of completed GC cycles, where a GC // cycle is sweep termination, mark, mark termination, and // sweep. This differs from memstats.numgc, which is // incremented at mark termination. cycles atomic.Uint32 // Timing/utilization stats for this cycle. stwprocs, maxprocs int32 tSweepTerm, tMark, tMarkTerm, tEnd int64 // nanotime() of phase start
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 62K bytes - Viewed (0) -
pkg/scheduler/internal/queue/scheduling_queue.go
podBackoffQ *heap.Heap // unschedulablePods holds pods that have been tried and determined unschedulable. unschedulablePods *UnschedulablePods // schedulingCycle represents sequence number of scheduling cycle and is incremented // when a pod is popped. schedulingCycle int64 // moveRequestCycle caches the sequence number of scheduling cycle when we // received a move request. Unschedulable pods in and before this scheduling
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 13:26:09 UTC 2024 - 61.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/sparsecore/embedding_pipelining.cc
loop_operands_ip1, /*flag_for_inlining=*/false); Value C_ip1 = cond_caller_ip1->getResults().front(); // Build the ReturnOp. This mirrors the construction of the operands with // 'i' values incremented. std::vector<Value> tmp_body_results; Append(tmp_body_results, loop_operands_im1); Append(tmp_body_results, loop_operands_i); Append(tmp_body_results, loop_operands_ip1);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 92.9K bytes - Viewed (0) -
src/runtime/mbitmap.go
s.freeindex = snelems return snelems } s.allocCache >>= uint(bitIndex + 1) sfreeindex = result + 1 if sfreeindex%64 == 0 && sfreeindex != snelems { // We just incremented s.freeindex so it isn't 0. // As each 1 in s.allocCache was encountered and used for allocation // it was shifted away. At this point s.allocCache contains all 0s. // Refill s.allocCache so that it corresponds
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 60K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/storage/testing/store_tests.go
cont := out.Continue // the second list call should try to get 2 more items from etcd // but since there is only one item left, that is all we should get with no continueValue // both read counters should be incremented for the singular calls they make in this case out = &example.PodList{} options = storage.ListOptions{ // ResourceVersion should be unset when setting continuation token. ResourceVersion: "",
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 11 12:45:33 UTC 2024 - 91.4K bytes - Viewed (0) -
pilot/pkg/xds/eds.go
// prevent memory leaks. if event == model.EventDelete { inboundServiceDeletes.Increment() s.Env.EndpointIndex.DeleteServiceShard(shard, hostname, namespace, false) } else { inboundServiceUpdates.Increment() } } // EDSUpdate computes destination address membership across all clusters and networks. // This is the main method implementing EDS.
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Tue Apr 02 15:58:06 UTC 2024 - 9.9K bytes - Viewed (0) -
src/cmd/go/internal/telemetrystats/telemetrystats.go
package telemetrystats import ( "cmd/go/internal/base" "cmd/go/internal/cfg" "cmd/go/internal/modload" "cmd/internal/telemetry" ) func Increment() { incrementConfig() incrementVersionCounters() } // incrementConfig increments counters for the configuration // the command is running in. func incrementConfig() { if !modload.WillBeEnabled() { telemetry.Inc("go/mode:gopath")
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:09:11 UTC 2024 - 1.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td
for `tf.EmptyTensorList` or the specified size for `tf.TensorListReserve`. Each push will be turned into `tf.XlaDynamicUpdateSlice` with the incremented size, and each pop will be turned into a `tf.Slice` and a copy of the buffer with decremented size. Each `tf.TensorListSetItem` will be turned into a `tf.XlaDynamicUpdateSlice` with unchanged size, and each `tf.TensorListGetItem` will be rewritten to a `tf.Slice`.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 99.6K bytes - Viewed (0) -
pkg/kubelet/pod_workers.go
return TerminatedPod } if s.IsTerminationRequested() { return TerminatingPod } return SyncPod } // mergeLastUpdate records the most recent state from a new update. Pod and MirrorPod are // incremented. KillPodOptions is accumulated. If RunningPod is set, Pod is synthetic and // will *not* be used as the last pod state unless no previous pod state exists (because
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Apr 02 13:22:37 UTC 2024 - 74.8K bytes - Viewed (0)