- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 102 for dequeue0 (0.17 sec)
-
staging/src/k8s.io/apiserver/pkg/util/flowcontrol/fairqueuing/queueset/fifo_list_test.go
arrival := []*request{{}, {}, {}, {}, {}, {}} list := newRequestFIFO() for i := range arrival { list.Enqueue(arrival[i]) } dequeued := make([]*request, 0) for list.Length() > 0 { req, _ := list.Dequeue() dequeued = append(dequeued, req) } verifyOrder(t, arrival, dequeued) } func TestFIFOWithEnqueueDequeueSomeRequestsRemainInQueue(t *testing.T) { list := newRequestFIFO()
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Jul 28 08:48:40 UTC 2022 - 7.8K bytes - Viewed (0) -
pkg/controller/deployment/progress.go
// requeueStuckDeployment checks whether the provided deployment needs to be synced for a progress // check. It returns the time after the deployment will be requeued for the progress check, 0 if it // will be requeued now, or -1 if it does not need to be requeued. func (dc *DeploymentController) requeueStuckDeployment(ctx context.Context, d *apps.Deployment, newStatus apps.DeploymentStatus) time.Duration { logger := klog.FromContext(ctx)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Mar 13 11:00:44 UTC 2023 - 8.9K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.h
// has reached its capacity, and will be unblocked when the queued tensors again // drop below the capacity due to dequeuing. // // Tensors are dequeued via the corresponding TF dequeue op. // TODO(hongm): Add support for `timeout_ms`. TF_CAPI_EXPORT extern void TF_EnqueueNamedTensor(TF_Session* session, int tensor_id,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 27 21:07:00 UTC 2023 - 15.1K bytes - Viewed (0) -
src/sync/poolqueue.go
// // This is implemented as a doubly-linked list queue of poolDequeues // where each dequeue is double the size of the previous one. Once a // dequeue fills up, this allocates a new one and only ever pushes to // the latest dequeue. Pops happen from the other end of the list and // once a dequeue is exhausted, it gets removed from the list. type poolChain struct {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Feb 26 18:12:29 UTC 2024 - 8.3K bytes - Viewed (0) -
src/runtime/lock_sema.go
if v == locked { if atomic.Casuintptr(&l.key, locked, 0) { break } } else { // Other M's are waiting for the lock. // Dequeue an M. mp = muintptr(v &^ locked).ptr() if atomic.Casuintptr(&l.key, v, uintptr(mp.nextwaitm)) { // Dequeued an M. Wake it. semawakeup(mp) break } } } gp.m.mLockProfile.recordUnlock(l) gp.m.locks-- if gp.m.locks < 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 6.8K bytes - Viewed (0) -
pilot/pkg/xds/pushqueue_test.go
done := make(chan *Connection, 1) go func() { con, _, _ := p.Dequeue() done <- con }() select { case ret := <-done: return ret case <-time.After(time.Millisecond * 500): return nil } } func ExpectTimeout(t *testing.T, p *PushQueue) { t.Helper() done := make(chan struct{}, 1) go func() { p.Dequeue() done <- struct{}{} }() select { case <-done:
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Tue Apr 30 00:26:45 UTC 2024 - 8.8K bytes - Viewed (0) -
src/runtime/chan.go
// The value ep sent by the sender is copied to the receiver sg. // The receiver is then woken up to go on its merry way. // Channel c must be empty and locked. send unlocks c with unlockf. // sg must already be dequeued from c. // ep must be non-nil and point to the heap or the caller's stack. func send(c *hchan, sg *sudog, ep unsafe.Pointer, unlockf func(), skip int) { if raceenabled { if c.dataqsiz == 0 { racesync(c, sg)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:16:50 UTC 2024 - 25.9K bytes - Viewed (0) -
src/runtime/sema.go
panic("semaRoot queue") } root.rotateLeft(s.parent) } } } // dequeue searches for and finds the first goroutine // in semaRoot blocked on addr. // If the sudog was being profiled, dequeue returns the time // at which it was woken up as now. Otherwise now is 0. // If there are additional entries in the wait list, dequeue // returns tailtime set to the last entry's acquiretime.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 19K bytes - Viewed (0) -
pkg/scheduler/framework/interface.go
// and Pods rejected by these plugins are requeued based on this extension point. // Failures from other extension points are regarded as temporal errors (e.g., network failure), // and the scheduler requeue Pods without this extension point - always requeue Pods to activeQ after backoff. // This is because such temporal errors cannot be resolved by specific cluster events,
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri May 31 15:52:16 UTC 2024 - 35.4K bytes - Viewed (0) -
src/runtime/select.go
if casi >= nsends { sg = c.sendq.dequeue() if sg != nil { goto recv } if c.qcount > 0 { goto bufrecv } if c.closed != 0 { goto rclose } } else { if raceenabled { racereadpc(c.raceaddr(), casePC(casi), chansendpc) } if c.closed != 0 { goto sclose } sg = c.recvq.dequeue() if sg != nil { goto send }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Mar 13 21:36:04 UTC 2024 - 15K bytes - Viewed (0)