- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 121 for dequeue1 (0.25 sec)
-
tensorflow/cc/training/queue_runner_test.cc
auto q1 = FIFOQueue(root.WithOpName(kQueueName1), {DataType::DT_INT32}, FIFOQueue::Capacity(3)); auto dequeue0 = QueueDequeue(root.WithOpName(kDequeueOp0), q0, {DataType::DT_INT32}); auto enqueue1 = QueueEnqueue(root.WithOpName(kEnqueueOp1), q1, {dequeue0[0]}); auto dequeue1 = QueueDequeue(root.WithOpName(kDequeueOp1), q1, {DataType::DT_INT32}); auto close1 = QueueClose(root.WithOpName(kCloseOp1), q1);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Sep 21 06:27:51 UTC 2019 - 14.7K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/util/flowcontrol/fairqueuing/queueset/fifo_list_test.go
arrival := []*request{{}, {}, {}, {}, {}, {}} list := newRequestFIFO() for i := range arrival { list.Enqueue(arrival[i]) } dequeued := make([]*request, 0) for list.Length() > 0 { req, _ := list.Dequeue() dequeued = append(dequeued, req) } verifyOrder(t, arrival, dequeued) } func TestFIFOWithEnqueueDequeueSomeRequestsRemainInQueue(t *testing.T) { list := newRequestFIFO()
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Jul 28 08:48:40 UTC 2022 - 7.8K bytes - Viewed (0) -
pkg/controller/deployment/progress.go
// requeueStuckDeployment checks whether the provided deployment needs to be synced for a progress // check. It returns the time after the deployment will be requeued for the progress check, 0 if it // will be requeued now, or -1 if it does not need to be requeued. func (dc *DeploymentController) requeueStuckDeployment(ctx context.Context, d *apps.Deployment, newStatus apps.DeploymentStatus) time.Duration { logger := klog.FromContext(ctx)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Mar 13 11:00:44 UTC 2023 - 8.9K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.h
// has reached its capacity, and will be unblocked when the queued tensors again // drop below the capacity due to dequeuing. // // Tensors are dequeued via the corresponding TF dequeue op. // TODO(hongm): Add support for `timeout_ms`. TF_CAPI_EXPORT extern void TF_EnqueueNamedTensor(TF_Session* session, int tensor_id,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 27 21:07:00 UTC 2023 - 15.1K bytes - Viewed (0) -
src/runtime/sema_test.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jun 21 19:37:22 UTC 2022 - 4.2K bytes - Viewed (0) -
src/runtime/lock_sema.go
if v == locked { if atomic.Casuintptr(&l.key, locked, 0) { break } } else { // Other M's are waiting for the lock. // Dequeue an M. mp = muintptr(v &^ locked).ptr() if atomic.Casuintptr(&l.key, v, uintptr(mp.nextwaitm)) { // Dequeued an M. Wake it. semawakeup(mp) break } } } gp.m.mLockProfile.recordUnlock(l) gp.m.locks-- if gp.m.locks < 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 6.8K bytes - Viewed (0) -
src/sync/poolqueue.go
// // This is implemented as a doubly-linked list queue of poolDequeues // where each dequeue is double the size of the previous one. Once a // dequeue fills up, this allocates a new one and only ever pushes to // the latest dequeue. Pops happen from the other end of the list and // once a dequeue is exhausted, it gets removed from the list. type poolChain struct {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Feb 26 18:12:29 UTC 2024 - 8.3K bytes - Viewed (0) -
pilot/pkg/xds/pushqueue_test.go
done := make(chan *Connection, 1) go func() { con, _, _ := p.Dequeue() done <- con }() select { case ret := <-done: return ret case <-time.After(time.Millisecond * 500): return nil } } func ExpectTimeout(t *testing.T, p *PushQueue) { t.Helper() done := make(chan struct{}, 1) go func() { p.Dequeue() done <- struct{}{} }() select { case <-done:
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Tue Apr 30 00:26:45 UTC 2024 - 8.8K bytes - Viewed (0) -
src/runtime/chan.go
// The value ep sent by the sender is copied to the receiver sg. // The receiver is then woken up to go on its merry way. // Channel c must be empty and locked. send unlocks c with unlockf. // sg must already be dequeued from c. // ep must be non-nil and point to the heap or the caller's stack. func send(c *hchan, sg *sudog, ep unsafe.Pointer, unlockf func(), skip int) { if raceenabled { if c.dataqsiz == 0 { racesync(c, sg)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:16:50 UTC 2024 - 25.9K bytes - Viewed (0) -
src/runtime/sema.go
panic("semaRoot queue") } root.rotateLeft(s.parent) } } } // dequeue searches for and finds the first goroutine // in semaRoot blocked on addr. // If the sudog was being profiled, dequeue returns the time // at which it was woken up as now. Otherwise now is 0. // If there are additional entries in the wait list, dequeue // returns tailtime set to the last entry's acquiretime.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 19K bytes - Viewed (0)