- Sort Score
- Num 10 results
- Language All
Results 11 - 20 of 51 for mu (0.01 seconds)
-
cmd/handler-api.go
} } } func (t *apiConfig) odirectEnabled() bool { t.mu.RLock() defer t.mu.RUnlock() return t.enableODirect } func (t *apiConfig) shouldGzipObjects() bool { t.mu.RLock() defer t.mu.RUnlock() return t.gzipObjects } func (t *apiConfig) permitRootAccess() bool { t.mu.RLock() defer t.mu.RUnlock() return t.rootAccess }
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 10.4K bytes - Click Count (0) -
cmd/metacache-manager.go
// Return a transient bucket for invalid or system buckets. m.mu.RLock() b, ok := m.buckets[bucket] if ok { m.mu.RUnlock() if b.bucket != bucket { logger.Info("getBucket: cached bucket %s does not match this bucket %s", b.bucket, bucket) debug.PrintStack() } return b } m.mu.RUnlock() m.mu.Lock() defer m.mu.Unlock() // See if someone else fetched it while we waited for the lock.
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Oct 25 00:44:15 GMT 2022 - 5.2K bytes - Click Count (0) -
internal/s3select/json/reader.go
// reads are running. type syncReadCloser struct { rc io.ReadCloser mu sync.Mutex } func (pr *syncReadCloser) Read(p []byte) (n int, err error) { // This ensures that Close will block until Read has completed. // This allows another goroutine to close the reader. pr.mu.Lock() defer pr.mu.Unlock() if pr.rc == nil { return 0, io.EOF } return pr.rc.Read(p) }Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Feb 18 16:25:55 GMT 2025 - 3.2K bytes - Click Count (0) -
cmd/untar.go
type disconnectReader struct { r io.Reader mu sync.Mutex } func (d *disconnectReader) Read(p []byte) (n int, err error) { d.mu.Lock() defer d.mu.Unlock() if d.r != nil { return d.r.Read(p) } return 0, errors.New("reader closed") } func (d *disconnectReader) Close() error { d.mu.Lock() d.r = nil d.mu.Unlock() return nil }Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Feb 18 16:25:55 GMT 2025 - 6K bytes - Click Count (2) -
cmd/ilm-config.go
ExpirationWorkers: 100, TransitionWorkers: 100, }, } type ilmConfig struct { mu sync.RWMutex cfg ilm.Config } func (c *ilmConfig) getExpirationWorkers() int { c.mu.RLock() defer c.mu.RUnlock() return c.cfg.ExpirationWorkers } func (c *ilmConfig) getTransitionWorkers() int { c.mu.RLock() defer c.mu.RUnlock() return c.cfg.TransitionWorkers }Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Mar 05 02:50:24 GMT 2024 - 1.3K bytes - Click Count (0) -
tensorflow/c/c_api_internal.h
}; struct TF_Graph { TF_Graph(); mutable tensorflow::mutex mu; tensorflow::Graph graph TF_GUARDED_BY(mu); // Runs shape inference. tensorflow::ShapeRefiner refiner TF_GUARDED_BY(mu); // Maps from name of an operation to the Node* in 'graph'. std::unordered_map<tensorflow::string, tensorflow::Node*> name_map TF_GUARDED_BY(mu); // The keys of this map are all the active sessions using this graph. EachCreated: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Sat May 13 00:49:12 GMT 2023 - 7.6K bytes - Click Count (0) -
tensorflow/c/c_api.cc
void TF_DeleteGraph(TF_Graph* g) { if (g == nullptr) return; g->mu.lock(); g->delete_requested = true; const bool del = g->sessions.empty(); g->mu.unlock(); if (del) delete g; } TF_Operation* TF_GraphOperationByName(TF_Graph* graph, const char* oper_name) { mutex_lock l(graph->mu); auto iter = graph->name_map.find(oper_name); if (iter == graph->name_map.end()) {
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Sat Oct 04 05:55:32 GMT 2025 - 102.4K bytes - Click Count (0) -
internal/dsync/dsync_test.go
type PaddedMutex struct { *DRWMutex } b.RunParallel(func(pb *testing.PB) { mu := PaddedMutex{NewDRWMutex(ds, "")} for pb.Next() { mu.Lock(id, source) mu.Unlock(b.Context()) } }) } func benchmarkMutex(b *testing.B, slack, work bool) { b.ResetTimer() b.ReportAllocs() mu := NewDRWMutex(ds, "") if slack { b.SetParallelism(10) }Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 10.8K bytes - Click Count (0) -
cmd/erasure-common.go
// unformatted or simply not accessible for some reason. // // - Do not consume disks which are being healed // // - Future: skip busy disks return } mu.Lock() newDisks = append(newDisks, disks[i]) mu.Unlock() }() } wg.Wait() return newDisks } func (er erasureObjects) getOnlineLocalDisks() (newDisks []StorageAPI) { disks := er.getOnlineDisks()
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 2.3K bytes - Click Count (0) -
tensorflow/c/c_api_function.cc
} int TF_GraphNumFunctions(TF_Graph* g) { tensorflow::mutex_lock l(g->mu); return g->graph.flib_def().num_functions(); } int TF_GraphGetFunctions(TF_Graph* g, TF_Function** funcs, int max_func, TF_Status* status) { tensorflow::FunctionDefLibrary lib; { tensorflow::mutex_lock l(g->mu); lib = g->graph.flib_def().ToProto(); }
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Sat Oct 12 16:27:48 GMT 2024 - 13.7K bytes - Click Count (0)