- Sort Score
- Num 10 results
- Language All
Results 1 - 6 of 6 for getBlock (0.06 seconds)
-
cmd/namespace-lock.go
) // local lock servers var globalLockServer *localLocker // RWLocker - locker interface to introduce GetRLock, RUnlock. type RWLocker interface { GetLock(ctx context.Context, timeout *dynamicTimeout) (lkCtx LockContext, timedOutErr error) Unlock(lkCtx LockContext) GetRLock(ctx context.Context, timeout *dynamicTimeout) (lkCtx LockContext, timedOutErr error) RUnlock(lkCtx LockContext) }Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 9.2K bytes - Click Count (0) -
internal/lsync/lrwmutex_test.go
) func testSimpleWriteLock(t *testing.T, duration time.Duration) (locked bool) { ctx := t.Context() lrwm := NewLRWMutex() if !lrwm.GetRLock(ctx, "", "object1", time.Second) { panic("Failed to acquire read lock") } // fmt.Println("1st read lock acquired, waiting...") if !lrwm.GetRLock(ctx, "", "object1", time.Second) { panic("Failed to acquire read lock") } // fmt.Println("2nd read lock acquired, waiting...")
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 7.8K bytes - Click Count (0) -
internal/dsync/drwmutex_test.go
ctx1, cancel1 := context.WithCancel(t.Context()) if !drwm1.GetRLock(ctx1, cancel1, id, source, Options{Timeout: time.Second}) { panic("Failed to acquire read lock") } // fmt.Println("1st read lock acquired, waiting...") drwm2 := NewDRWMutex(ds, "simplelock") ctx2, cancel2 := context.WithCancel(t.Context()) if !drwm2.GetRLock(ctx2, cancel2, id, source, Options{Timeout: time.Second}) {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 9.4K bytes - Click Count (0) -
internal/lsync/lrwmutex.go
func (lm *LRWMutex) RLock() { const isWriteLock = false lm.lockLoop(context.Background(), lm.id, lm.source, 1<<63-1, isWriteLock) } // GetRLock tries to get a read lock on lm before the timeout occurs. func (lm *LRWMutex) GetRLock(ctx context.Context, id string, source string, timeout time.Duration) (locked bool) { const isWriteLock = false return lm.lockLoop(ctx, id, source, timeout, isWriteLock) }
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Jan 02 17:15:06 GMT 2022 - 4.8K bytes - Click Count (0) -
cmd/shared-lock.go
lockContext chan LockContext } func (ld sharedLock) backgroundRoutine(ctx context.Context, objAPI ObjectLayer, lockName string) { for { locker := objAPI.NewNSLock(minioMetaBucket, lockName) lkctx, err := locker.GetLock(ctx, sharedLockTimeout) if err != nil { continue } keepLock: for { select { case <-ctx.Done(): return case <-lkctx.Context().Done():
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Mon Feb 13 09:26:38 GMT 2023 - 2.3K bytes - Click Count (0) -
cmd/callhome.go
} func runCallhome(ctx context.Context, objAPI ObjectLayer) bool { // Make sure only 1 callhome is running on the cluster. locker := objAPI.NewNSLock(minioMetaBucket, "callhome/runCallhome.lock") lkctx, err := locker.GetLock(ctx, callhomeLeaderLockTimeout) if err != nil { // lock timedout means some other node is the leader, // cycle back return 'true' return true } ctx = lkctx.Context() defer locker.Unlock(lkctx)
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 5.3K bytes - Click Count (0)