- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 287 for Unlock (0.17 sec)
-
src/archive/zip/register.go
fw *flate.Writer } func (w *pooledFlateWriter) Write(p []byte) (n int, err error) { w.mu.Lock() defer w.mu.Unlock() if w.fw == nil { return 0, errors.New("Write after Close") } return w.fw.Write(p) } func (w *pooledFlateWriter) Close() error { w.mu.Lock() defer w.mu.Unlock() var err error if w.fw != nil { err = w.fw.Close() flateWriterPool.Put(w.fw) w.fw = nil }
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Fri Oct 13 18:36:46 GMT 2023 - 3.7K bytes - Viewed (0) -
internal/lock/lock_test.go
t.Fatal("Unexpected error", err) } } // Tests lock and unlock semantics. func TestLockAndUnlock(t *testing.T) { f, err := os.CreateTemp("", "lock") if err != nil { t.Fatal(err) } f.Close() defer func() { err = os.Remove(f.Name()) if err != nil { t.Fatal(err) } }() // lock the file l, err := LockedOpenFile(f.Name(), os.O_WRONLY, 0o600) if err != nil {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Mon Sep 19 18:05:16 GMT 2022 - 3.6K bytes - Viewed (0) -
android/guava/src/com/google/common/util/concurrent/CycleDetectingLockFactory.java
* * <ul> * <li>for an unnested {@code lock()} and {@code unlock()}, a cycle detecting lock takes 38ns as * opposed to the 24ns taken by a plain lock. * <li>for nested locking, the cost increases with the depth of the nesting: * <ul> * <li>2 levels: average of 64ns per lock()/unlock() * <li>3 levels: average of 77ns per lock()/unlock() * <li>4 levels: average of 99ns per lock()/unlock()
Java - Registered: Fri Apr 26 12:43:10 GMT 2024 - Last Modified: Fri Dec 15 19:31:54 GMT 2023 - 35.9K bytes - Viewed (0) -
cni/pkg/nodeagent/ztunnelserver.go
} func (c *connMgr) addConn(conn *ZtunnelConnection) { log.Debug("ztunnel connected") c.mu.Lock() defer c.mu.Unlock() c.connectionSet[conn] = struct{}{} c.latestConn = conn ztunnelConnected.RecordInt(int64(len(c.connectionSet))) } func (c *connMgr) LatestConn() *ZtunnelConnection { c.mu.Lock() defer c.mu.Unlock() return c.latestConn } func (c *connMgr) deleteConn(conn *ZtunnelConnection) {
Go - Registered: Wed May 01 22:53:12 GMT 2024 - Last Modified: Fri Apr 12 21:47:31 GMT 2024 - 12.4K bytes - Viewed (0) -
cni/pkg/nodeagent/pod_cache.go
func (p *podNetnsCache) Get(uid string) Netns { // lock current snapshot pod map p.mu.RLock() defer p.mu.RUnlock() if info, f := p.currentPodCache[uid]; f { return info.Netns } return nil } // make sure uid is in the cache, even if we don't have a netns func (p *podNetnsCache) Ensure(uid string) { p.mu.Lock() defer p.mu.Unlock() if _, ok := p.currentPodCache[uid]; !ok {
Go - Registered: Wed May 01 22:53:12 GMT 2024 - Last Modified: Fri Apr 12 21:47:31 GMT 2024 - 4.5K bytes - Viewed (0) -
cmd/erasure-object_test.go
t.Fatal(err) } erasureDisks := xl.getDisks() z.serverPools[0].erasureDisksMu.Lock() xl.getDisks = func() []StorageAPI { for i := range erasureDisks[:6] { erasureDisks[i] = newNaughtyDisk(erasureDisks[i], nil, errFaultyDisk) } return erasureDisks } z.serverPools[0].erasureDisksMu.Unlock() _, err = obj.DeleteObject(ctx, bucket, object, ObjectOptions{})
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Tue Jan 30 20:43:25 GMT 2024 - 36.8K bytes - Viewed (0) -
cmd/erasure-server-pool.go
defer func() { if unlockOnDefer { nsUnlocker() } }() // Acquire lock if !opts.NoLock { lock := z.NewNSLock(bucket, object) lkctx, err := lock.GetRLock(ctx, globalOperationTimeout) if err != nil { return nil, err } ctx = lkctx.Context() nsUnlocker = func() { lock.RUnlock(lkctx) } unlockOnDefer = true } checkPrecondFn := opts.CheckPrecondFn
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 03 20:08:20 GMT 2024 - 80.5K bytes - Viewed (0) -
internal/cachevalue/cache.go
if t.updating.TryLock() { go func() { defer t.updating.Unlock() t.update() }() } return v.v, v.e } // Get lock. Either we get it or we wait for it. t.updating.Lock() if time.Since(time.UnixMilli(t.lastUpdateMs.Load())) < ttl { // There is a new value, release lock and return it. v = t.valErr.Load() t.updating.Unlock() return v.v, v.e } t.update() v = t.valErr.Load()
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri Mar 01 16:00:42 GMT 2024 - 4.2K bytes - Viewed (0) -
cmd/metacache-bucket.go
logger.Info("bucketMetacache.findCache: bucket %s does not match this bucket %s", o.Bucket, b.bucket) debug.PrintStack() return metacache{} } // Grab a write lock, since we create one if we cannot find one. b.mu.Lock() defer b.mu.Unlock() // Check if exists already. if c, ok := b.caches[o.ID]; ok { c.lastHandout = time.Now() b.caches[o.ID] = c b.debugf("returning existing %v", o.ID) return c
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 6.6K bytes - Viewed (0) -
internal/logger/logonce.go
if err == nil { return } nerr := unwrapErrs(err) l.Lock() shouldLog := true prev, ok := l.IDMap[id] if !ok { l.IDMap[id] = onceErr{ Err: nerr, Count: 1, } } else if prev.Err.Error() == nerr.Error() { // if errors are equal do not log. prev.Count++ l.IDMap[id] = prev shouldLog = false } l.Unlock() if shouldLog { consoleLogIf(ctx, subsystem, err, errKind...) }
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 3.8K bytes - Viewed (0)