- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 221 for Unlock (0.87 sec)
-
cmd/lock-rest-client_test.go
} // Attempt all calls. _, err = lkClient.RLock(t.Context(), dsync.LockArgs{}) if err == nil { t.Fatal("Expected for Rlock to fail") } _, err = lkClient.Lock(t.Context(), dsync.LockArgs{}) if err == nil { t.Fatal("Expected for Lock to fail") } _, err = lkClient.RUnlock(t.Context(), dsync.LockArgs{}) if err == nil { t.Fatal("Expected for RUnlock to fail") }
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Wed Apr 09 14:28:39 UTC 2025 - 2K bytes - Viewed (0) -
internal/store/batch.go
func (b *Batch[I]) Add(item I) error { b.Lock() defer b.Unlock() if b.isFull() { if b.store == nil { return ErrBatchFull } // commit batch to store if err := b.commit(); err != nil { return err } } b.items = append(b.items, item) return nil } // Len returns the no of items in the batch func (b *Batch[_]) Len() int { b.Lock() defer b.Unlock() return len(b.items)
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Sep 06 23:06:30 UTC 2024 - 2.9K bytes - Viewed (0) -
cmd/bucket-replication-stats.go
return ActiveWorkerStat{} } r.wlock.RLock() defer r.wlock.RUnlock() w := r.workers.get() return ActiveWorkerStat{ Curr: w.Curr, Max: w.Max, Avg: w.Avg, } } func (r *ReplicationStats) collectWorkerMetrics(ctx context.Context) { if r == nil { return } for { select { case <-ctx.Done(): return case <-r.wTimer.C: r.wlock.Lock() r.workers.update()Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Tue Feb 18 16:25:55 UTC 2025 - 13.6K bytes - Viewed (0) -
src/main/java/jcifs/internal/smb2/persistent/PersistentHandleManager.java
lock.readLock().unlock(); } } /** * Get handle information by path * @param path the file path * @return the handle info or null if not found */ public HandleInfo getHandleByPath(String path) { lock.readLock().lock(); try { return handles.get(path); } finally { lock.readLock().unlock(); } }
Registered: Sat Dec 20 13:44:44 UTC 2025 - Last Modified: Sun Aug 24 00:49:49 UTC 2025 - 13K bytes - Viewed (0) -
cmd/site-replication-utils.go
}() } } wg.Wait() sm.Unlock() } sTimer.Reset(siteResyncSaveInterval) case <-ctx.Done(): return } } } // update overall site resync state func (sm *siteResyncMetrics) updateState(s SiteResyncStatus) error { if !globalSiteReplicationSys.isEnabled() { return nil } sm.Lock() defer sm.Unlock() switch s.Status { case ResyncStarted:
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 8.8K bytes - Viewed (0) -
internal/store/queuestore.go
} // Del - Deletes an entry from the store. func (store *QueueStore[_]) Del(key Key) error { store.Lock() defer store.Unlock() return store.del(key) } // Len returns the entry count. func (store *QueueStore[_]) Len() int { store.RLock() l := len(store.entries) defer store.RUnlock() return l } // lockless call func (store *QueueStore[_]) del(key Key) error {
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Sun Sep 28 20:59:21 UTC 2025 - 7.8K bytes - Viewed (0) -
cmd/namespace-lock_test.go
nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 1*time.Millisecond) }() time.Sleep(1 * time.Millisecond) // wait for goroutine to advance; ref=2 // Unlock the 1st lock; ref=1 after this line nsLk.unlock("volume", "path", false) // Taking another lockMapMutex here allows queuing up additional lockers. This should // not be required but makes reproduction much easier. nsLk.lockMapMutex.Lock()
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 3K bytes - Viewed (0) -
src/archive/zip/register.go
fw *flate.Writer } func (w *pooledFlateWriter) Write(p []byte) (n int, err error) { w.mu.Lock() defer w.mu.Unlock() if w.fw == nil { return 0, errors.New("Write after Close") } return w.fw.Write(p) } func (w *pooledFlateWriter) Close() error { w.mu.Lock() defer w.mu.Unlock() var err error if w.fw != nil { err = w.fw.Close() flateWriterPool.Put(w.fw) w.fw = nil }
Registered: Tue Dec 30 11:13:12 UTC 2025 - Last Modified: Fri Oct 13 18:36:46 UTC 2023 - 3.7K bytes - Viewed (0) -
internal/deadlineconn/deadlineconn.go
func (c *DeadlineConn) setReadDeadline() { // Do not set a Read deadline, if upstream wants to cancel all reads. if c.readDeadline <= 0 || c.abortReads.Load() || c.infReads.Load() { return } c.mu.Lock() defer c.mu.Unlock() if c.abortReads.Load() { return } now := time.Now() if now.Sub(c.readSetAt) > updateInterval { c.Conn.SetReadDeadline(now.Add(c.readDeadline + updateInterval)) c.readSetAt = now }
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Mon Dec 02 13:21:17 UTC 2024 - 5.1K bytes - Viewed (0) -
internal/pubsub/pubsub.go
} ps.Lock() defer ps.Unlock() sub := &Sub[T]{ch: subCh, types: Mask(mask.Mask()), filter: filter} ps.subs = append(ps.subs, sub) // We hold a lock, so we are safe to update combined := Mask(atomic.LoadUint64(&ps.types)) combined.Merge(Mask(mask.Mask())) atomic.StoreUint64(&ps.types, uint64(combined)) go func() { <-doneCh ps.Lock() defer ps.Unlock()
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Tue Feb 06 16:57:30 UTC 2024 - 5.2K bytes - Viewed (0)