- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 36 for queuesa (0.28 sec)
-
pkg/scheduler/internal/queue/scheduling_queue_test.go
queue.flushBackoffQCompleted(logger) } moveClockForward = func(t *testing.T, logger klog.Logger, queue *PriorityQueue, _ *framework.QueuedPodInfo) { queue.clock.(*testingclock.FakeClock).Step(2 * time.Second) } flushUnschedulerQ = func(t *testing.T, logger klog.Logger, queue *PriorityQueue, _ *framework.QueuedPodInfo) {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 13:26:09 UTC 2024 - 146.9K bytes - Viewed (0) -
pkg/controller/garbagecollector/garbagecollector_test.go
} func (t *trackingWorkqueue[T]) queue(item T) { if _, queued := t.pendingMap[item]; queued { // fmt.Printf("already queued: %#v\n", item) return } t.pendingMap[item] = struct{}{} t.pendingList = append(t.pendingList, item) } func (t *trackingWorkqueue[T]) dequeue(item T) { if _, queued := t.pendingMap[item]; !queued { // fmt.Printf("not queued: %#v\n", item) return }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 04 18:33:12 UTC 2024 - 111.6K bytes - Viewed (0) -
guava/src/com/google/common/cache/LocalCache.java
* fraction of read operations. */ final AtomicInteger readCount = new AtomicInteger(); /** * A queue of elements currently in the map, ordered by write time. Elements are added to the * tail of the queue on write. */ @GuardedBy("this") final Queue<ReferenceEntry<K, V>> writeQueue; /**
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Sat May 18 03:24:34 UTC 2024 - 149.2K bytes - Viewed (0) -
android/guava/src/com/google/common/cache/LocalCache.java
* fraction of read operations. */ final AtomicInteger readCount = new AtomicInteger(); /** * A queue of elements currently in the map, ordered by write time. Elements are added to the * tail of the queue on write. */ @GuardedBy("this") final Queue<ReferenceEntry<K, V>> writeQueue; /**
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Sat May 18 03:24:34 UTC 2024 - 143.6K bytes - Viewed (0) -
src/runtime/proc.go
wakep() } // injectglist adds each runnable G on the list to some run queue, // and clears glist. If there is no current P, they are added to the // global queue, and up to npidle M's are started to run them. // Otherwise, for each idle P, this adds a G to the global queue // and starts an M. Any remaining G's are added to the current P's // local run queue. // This may temporarily acquire sched.lock. // Can run concurrently with GC.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 207.5K bytes - Viewed (0) -
cmd/metrics-v2.go
Help: "Total number of objects queued for replication in the last full minute", Type: gaugeMetric, } } func getClusterReplCurrQueuedBytesMD() MetricDescription { return MetricDescription{ Namespace: nodeMetricNamespace, Subsystem: replicationSubsystem, Name: currInQueueBytes, Help: "Total number of bytes queued for replication in the last full minute",
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Jun 13 22:26:54 UTC 2024 - 131.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/shape_inference.cc
<< ")\n"); if (queue_set_.insert(fn).second) queue_.push(fn); } // Enqueues callers on functions. void EnqueueCallers(func::FuncOp fn); // Returns the function at the front of the queue. func::FuncOp front() { return queue_.front(); } // Returns whether work queue is empty. bool EmptyQueue() const { return queue_.empty(); } // Returns function from the front of the work queue.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Jun 08 07:28:49 UTC 2024 - 134.1K bytes - Viewed (0) -
pkg/controller/job/job_controller_test.go
manager.queue.Add(key) manager.processNextWorkItem(context.TODO()) retries := manager.queue.NumRequeues(key) if retries != 1 { t.Fatalf("%s: expected exactly 1 retry, got %d", job.Name, retries) } // await for the actual requeue after processing of the pending queue is done awaitForQueueLen(ctx, t, manager, 1) // the queue is emptied on success fakePodControl.Err = nil
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 11 15:36:36 UTC 2024 - 229.2K bytes - Viewed (0) -
staging/src/k8s.io/api/flowcontrol/v1beta3/generated.pb.go
if wireType != 0 { return fmt.Errorf("proto: wrong wireType = %d for field Queues", wireType) } m.Queues = 0 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated } if iNdEx >= l { return io.ErrUnexpectedEOF } b := dAtA[iNdEx] iNdEx++ m.Queues |= int32(b&0x7F) << shift if b < 0x80 { break } }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Mar 01 06:06:37 UTC 2024 - 141.9K bytes - Viewed (0) -
staging/src/k8s.io/api/flowcontrol/v1/generated.pb.go
if wireType != 0 { return fmt.Errorf("proto: wrong wireType = %d for field Queues", wireType) } m.Queues = 0 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated } if iNdEx >= l { return io.ErrUnexpectedEOF } b := dAtA[iNdEx] iNdEx++ m.Queues |= int32(b&0x7F) << shift if b < 0x80 { break } }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Mar 01 06:06:37 UTC 2024 - 141.8K bytes - Viewed (0)