- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 108 for unlocked (0.08 sec)
-
internal/dsync/drwmutex.go
// to indicate we haven't sufficiently unlocked lockers to avoid deadlocks. // // Caller may use this as an indication to call again. return !checkFailedUnlocks(*locks, tolerance) } // Unlock unlocks the write lock. // // It is a run-time error if dm is not locked on entry to Unlock. func (dm *DRWMutex) Unlock(ctx context.Context) { dm.m.Lock() dm.cancelRefresh() dm.m.Unlock()
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Mon Sep 09 15:49:49 UTC 2024 - 20.4K bytes - Viewed (0) -
guava/src/com/google/common/util/concurrent/CycleDetectingLockFactory.java
* * <ul> * <li>for an unnested {@code lock()} and {@code unlock()}, a cycle detecting lock takes 38ns as * opposed to the 24ns taken by a plain lock. * <li>for nested locking, the cost increases with the depth of the nesting: * <ul> * <li>2 levels: average of 64ns per lock()/unlock() * <li>3 levels: average of 77ns per lock()/unlock() * <li>4 levels: average of 99ns per lock()/unlock()
Registered: Fri Nov 01 12:43:10 UTC 2024 - Last Modified: Fri Dec 15 19:31:54 UTC 2023 - 35.9K bytes - Viewed (0) -
doc/go_mem.html
Registered: Tue Nov 05 11:13:11 UTC 2024 - Last Modified: Mon Mar 04 15:54:42 UTC 2024 - 26.6K bytes - Viewed (0) -
internal/grid/muxclient.go
} if m.subroute != nil { msg.Flags |= FlagSubroute } ch := make(chan Response, 1) m.respMu.Lock() if m.closed { m.respMu.Unlock() return nil, ErrDisconnected } m.respWait = ch m.respMu.Unlock() ctx := m.ctx // Add deadline if none. if msg.DeadlineMS == 0 { msg.DeadlineMS = uint32(defaultSingleRequestTimeout / time.Millisecond)
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Fri Jun 07 15:51:52 UTC 2024 - 15.9K bytes - Viewed (0) -
guava-tests/test/com/google/common/util/concurrent/CycleDetectingLockFactoryTest.java
Lock lockB = factory.newReentrantLock(OtherOrder.SECOND); lockA.lock(); lockA.lock(); lockB.lock(); lockB.lock(); lockA.unlock(); lockA.unlock(); lockB.unlock(); lockB.unlock(); } public void testExplicitOrdering_acquiringMultipleLocksWithSameRank() { CycleDetectingLockFactory.WithExplicitOrdering<OtherOrder> factory =
Registered: Fri Nov 01 12:43:10 UTC 2024 - Last Modified: Fri Oct 18 22:10:29 UTC 2024 - 16.1K bytes - Viewed (0) -
android/guava-tests/test/com/google/common/util/concurrent/CycleDetectingLockFactoryTest.java
Lock lockB = factory.newReentrantLock(OtherOrder.SECOND); lockA.lock(); lockA.lock(); lockB.lock(); lockB.lock(); lockA.unlock(); lockA.unlock(); lockB.unlock(); lockB.unlock(); } public void testExplicitOrdering_acquiringMultipleLocksWithSameRank() { CycleDetectingLockFactory.WithExplicitOrdering<OtherOrder> factory =
Registered: Fri Nov 01 12:43:10 UTC 2024 - Last Modified: Fri Oct 18 22:10:29 UTC 2024 - 16.1K bytes - Viewed (0) -
internal/ringbuffer/ring_buffer.go
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Wed May 15 00:11:04 UTC 2024 - 13.3K bytes - Viewed (0) -
internal/dsync/dsync_test.go
dm.Unlock(context.Background()) } func TestSimpleLockUnlockMultipleTimes(t *testing.T) { dm := NewDRWMutex(ds, "test") dm.Lock(id, source) time.Sleep(time.Duration(10+(rand.Float32()*50)) * time.Millisecond) dm.Unlock(context.Background()) dm.Lock(id, source) time.Sleep(time.Duration(10+(rand.Float32()*50)) * time.Millisecond) dm.Unlock(context.Background()) dm.Lock(id, source)
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Wed Jun 19 14:35:19 UTC 2024 - 11.1K bytes - Viewed (0) -
cmd/local-locker.go
return concat(s, strconv.Itoa(idx)) } func (l *localLocker) Unlock(_ context.Context, args dsync.LockArgs) (reply bool, err error) { if len(args.Resources) > maxDeleteList { return false, fmt.Errorf("internal error: localLocker.Unlock called with more than %d resources", maxDeleteList) } l.mutex.Lock() defer l.mutex.Unlock() err = nil for _, resource := range args.Resources {
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Wed Jul 24 10:24:01 UTC 2024 - 10.7K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.h
// from a placeholder node "arg_tensor_enqueue_<tensor_id>". // // `tensor` is still owned by the caller. This call will be blocked if the queue // has reached its capacity, and will be unblocked when the queued tensors again // drop below the capacity due to dequeuing. // // Tensors are dequeued via the corresponding TF dequeue op. // TODO(hongm): Add support for `timeout_ms`.
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Thu Apr 27 21:07:00 UTC 2023 - 15.1K bytes - Viewed (0)