- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 58 for mu (0.15 sec)
-
internal/lsync/lrwmutex.go
const isWriteLock = false return lm.lockLoop(ctx, id, source, timeout, isWriteLock) } func (lm *LRWMutex) lock(id, source string, isWriteLock bool) (locked bool) { lm.mu.Lock() defer lm.mu.Unlock() lm.id = id lm.source = source if isWriteLock { if lm.ref == 0 && !lm.isWriteLock { lm.ref = 1 lm.isWriteLock = true locked = true } } else { if !lm.isWriteLock {
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Sun Jan 02 17:15:06 GMT 2022 - 4.8K bytes - Viewed (0) -
cni/pkg/nodeagent/ztunnelserver.go
latestConn *ZtunnelConnection mu sync.Mutex } func (c *connMgr) addConn(conn *ZtunnelConnection) { log.Debug("ztunnel connected") c.mu.Lock() defer c.mu.Unlock() c.connectionSet[conn] = struct{}{} c.latestConn = conn ztunnelConnected.RecordInt(int64(len(c.connectionSet))) } func (c *connMgr) LatestConn() *ZtunnelConnection { c.mu.Lock() defer c.mu.Unlock() return c.latestConn }
Go - Registered: Wed May 01 22:53:12 GMT 2024 - Last Modified: Fri Apr 12 21:47:31 GMT 2024 - 12.4K bytes - Viewed (0) -
cmd/metacache-manager.go
// Return a transient bucket for invalid or system buckets. m.mu.RLock() b, ok := m.buckets[bucket] if ok { m.mu.RUnlock() if b.bucket != bucket { logger.Info("getBucket: cached bucket %s does not match this bucket %s", b.bucket, bucket) debug.PrintStack() } return b } m.mu.RUnlock() m.mu.Lock() defer m.mu.Unlock() // See if someone else fetched it while we waited for the lock.
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Tue Oct 25 00:44:15 GMT 2022 - 5.2K bytes - Viewed (0) -
cni/pkg/nodeagent/pod_cache.go
func (p *podNetnsCache) Get(uid string) Netns { // lock current snapshot pod map p.mu.RLock() defer p.mu.RUnlock() if info, f := p.currentPodCache[uid]; f { return info.Netns } return nil } // make sure uid is in the cache, even if we don't have a netns func (p *podNetnsCache) Ensure(uid string) { p.mu.Lock() defer p.mu.Unlock() if _, ok := p.currentPodCache[uid]; !ok {
Go - Registered: Wed May 01 22:53:12 GMT 2024 - Last Modified: Fri Apr 12 21:47:31 GMT 2024 - 4.5K bytes - Viewed (0) -
cni/pkg/log/uds.go
continue } msg.Msg = strings.TrimSpace(msg.Msg) messages = append(messages, msg) } // Lock log message printing to prevent log messages from different CNI // processes interleave. l.mu.Lock() defer l.mu.Unlock() for _, m := range messages { // There is no fatal log from CNI plugin switch m.Level { case "debug": pluginLog.LogWithTime(log.DebugLevel, m.Msg, m.Time) case "info":
Go - Registered: Wed May 01 22:53:12 GMT 2024 - Last Modified: Wed Nov 29 01:05:12 GMT 2023 - 3.3K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.cc
bool downloaded_block = false; auto reconcile_state = MakeCleanup([this, &downloaded_block, &key, &block] { // Perform this action in a cleanup callback to avoid locking mu_ after // locking block->mu. if (downloaded_block) { absl::MutexLock l(&mu_); // Do not update state if the block is already to be evicted. if (block->timestamp != 0) { // Use capacity() instead of size() to account for all memory
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Jul 16 01:39:09 GMT 2020 - 11.1K bytes - Viewed (0) -
tensorflow/c/c_api_internal.h
}; struct TF_Graph { TF_Graph(); mutable tensorflow::mutex mu; tensorflow::Graph graph TF_GUARDED_BY(mu); // Runs shape inference. tensorflow::ShapeRefiner refiner TF_GUARDED_BY(mu); // Maps from name of an operation to the Node* in 'graph'. std::unordered_map<tensorflow::string, tensorflow::Node*> name_map TF_GUARDED_BY(mu); // The keys of this map are all the active sessions using this graph. Each
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Sat May 13 00:49:12 GMT 2023 - 7.6K bytes - Viewed (0) -
cmd/handler-api.go
} func (t *apiConfig) odirectEnabled() bool { t.mu.RLock() defer t.mu.RUnlock() return t.enableODirect } func (t *apiConfig) shouldGzipObjects() bool { t.mu.RLock() defer t.mu.RUnlock() return t.gzipObjects } func (t *apiConfig) permitRootAccess() bool { t.mu.RLock() defer t.mu.RUnlock() return t.rootAccess }
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Mon Apr 08 09:22:27 GMT 2024 - 10K bytes - Viewed (0) -
cmd/untar.go
type disconnectReader struct { r io.Reader mu sync.Mutex } func (d *disconnectReader) Read(p []byte) (n int, err error) { d.mu.Lock() defer d.mu.Unlock() if d.r != nil { return d.r.Read(p) } return 0, errors.New("reader closed") } func (d *disconnectReader) Close() error { d.mu.Lock() d.r = nil d.mu.Unlock() return nil }
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 6K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.h
/// the block-cache-wide mu_ instance variable. The state variable should only /// be accessed while holding the Block's mu lock. The data vector should only /// be accessed after state == FINISHED, and it should never be modified. /// /// In order to prevent deadlocks, never grab the block-cache-wide mu_ lock /// AFTER grabbing any block's mu lock. It is safe to grab mu without locking /// mu_. struct Block {
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Mon Aug 31 04:46:34 GMT 2020 - 10.6K bytes - Viewed (0)