- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 61 for Unlock (0.17 sec)
-
internal/dsync/drwmutex.go
} // Unlock unlocks the write lock. // // It is a run-time error if dm is not locked on entry to Unlock. func (dm *DRWMutex) Unlock(ctx context.Context) { dm.m.Lock() dm.cancelRefresh() dm.m.Unlock() restClnts, owner := dm.clnt.GetLockers() // create temp array on stack locks := make([]string, len(restClnts)) { dm.m.Lock() defer dm.m.Unlock()
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Feb 22 06:26:06 GMT 2024 - 19.7K bytes - Viewed (0) -
cmd/bucket-targets.go
sys.aMutex.Lock() defer sys.aMutex.Unlock() if v, ok := sys.arnErrsMap[arn]; !ok { sys.arnErrsMap[arn] = arnErrs{ updateInProgress: true, count: v.count + 1, bucket: bucket, } } } func (sys *BucketTargetSys) markRefreshDone(bucket, arn string) { sys.aMutex.Lock() defer sys.aMutex.Unlock() if v, ok := sys.arnErrsMap[arn]; ok {
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 20.9K bytes - Viewed (0) -
cmd/erasure-server-pool-rebalance.go
ok = true } } if ok { lock := z.serverPools[0].NewNSLock(minioMetaBucket, rebalMetaName) lkCtx, err := lock.GetLock(ctx, globalOperationTimeout) if err != nil { rebalanceLogIf(ctx, fmt.Errorf("failed to acquire write lock on %s/%s: %w", minioMetaBucket, rebalMetaName, err)) return err } defer lock.Unlock(lkCtx) ctx = lkCtx.Context()
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 26.5K bytes - Viewed (0) -
cmd/bucket-metadata-sys.go
func (sys *BucketMetadataSys) Count() int { sys.RLock() defer sys.RUnlock() return len(sys.metadataMap) } // Remove bucket metadata from memory. func (sys *BucketMetadataSys) Remove(buckets ...string) { sys.Lock() for _, bucket := range buckets { delete(sys.metadataMap, bucket) globalBucketMonitor.DeleteBucket(bucket) } sys.Unlock() }
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 19.2K bytes - Viewed (0) -
cmd/perf-tests.go
activeConnections uint64 sync.RWMutex } func (n *netPerfRX) Connect() { n.Lock() defer n.Unlock() n.activeConnections++ atomic.StoreUint64(&n.RX, 0) n.lastToConnect = time.Now() } func (n *netPerfRX) Disconnect() { n.Lock() defer n.Unlock() n.activeConnections-- if n.firstToDisconnect.IsZero() { n.RXSample = atomic.LoadUint64(&n.RX)
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Sun Jan 28 18:04:17 GMT 2024 - 11.3K bytes - Viewed (0) -
internal/dsync/dsync_test.go
dm.Unlock(context.Background()) dm.Lock(id, source) time.Sleep(time.Duration(10+(rand.Float32()*50)) * time.Millisecond) dm.Unlock(context.Background()) dm.Lock(id, source) time.Sleep(time.Duration(10+(rand.Float32()*50)) * time.Millisecond) dm.Unlock(context.Background()) dm.Lock(id, source) time.Sleep(time.Duration(10+(rand.Float32()*50)) * time.Millisecond)
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Sat Dec 24 03:49:07 GMT 2022 - 11K bytes - Viewed (0) -
cmd/erasure-object_test.go
t.Fatal(err) } erasureDisks := xl.getDisks() z.serverPools[0].erasureDisksMu.Lock() xl.getDisks = func() []StorageAPI { for i := range erasureDisks[:6] { erasureDisks[i] = newNaughtyDisk(erasureDisks[i], nil, errFaultyDisk) } return erasureDisks } z.serverPools[0].erasureDisksMu.Unlock() _, err = obj.DeleteObject(ctx, bucket, object, ObjectOptions{})
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Tue Jan 30 20:43:25 GMT 2024 - 36.8K bytes - Viewed (0) -
cmd/metacache-walk.go
return ctx.Err() } if opts.Limit > 0 && objsReturned >= opts.Limit { return nil } if s.walkMu != nil { s.walkMu.Lock() } entries, err := s.ListDir(ctx, "", opts.Bucket, current, -1) if s.walkMu != nil { s.walkMu.Unlock() } if err != nil { // Folder could have gone away in-between if err != errVolumeNotFound && err != errFileNotFound {
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Mon Apr 15 08:25:46 GMT 2024 - 12.4K bytes - Viewed (0) -
cmd/admin-heal-ops.go
func (ahs *allHealState) setDiskHealingStatus(ep Endpoint, healing bool) { ahs.Lock() defer ahs.Unlock() ahs.healLocalDisks[ep] = healing } func (ahs *allHealState) pushHealLocalDisks(healLocalDisks ...Endpoint) { ahs.Lock() defer ahs.Unlock() for _, ep := range healLocalDisks { ahs.healLocalDisks[ep] = false } } func (ahs *allHealState) periodicHealSeqsClean(ctx context.Context) {
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 25.1K bytes - Viewed (1) -
cmd/local-locker.go
} func (l *localLocker) RUnlock(_ context.Context, args dsync.LockArgs) (reply bool, err error) { if len(args.Resources) > 1 { return false, fmt.Errorf("internal error: localLocker.RUnlock called with more than one resource") } l.mutex.Lock() defer l.mutex.Unlock() var lri []lockRequesterInfo resource := args.Resources[0] if lri, reply = l.lockMap[resource]; !reply { // No lock is held on the given name
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Mon Feb 19 22:54:46 GMT 2024 - 10.6K bytes - Viewed (0)