- Sort Score
- Result 10 results
- Languages All
Results 91 - 100 of about 10,000 for acquired (0.32 sec)
-
okhttp/src/main/kotlin/okhttp3/internal/connection/RealRoutePlanner.kt
} else -> null } } // If the call's connection wasn't released, reuse it. We don't call connectionAcquired() here // because we already acquired it. if (connectionUser.candidateConnection() != null) { check(toClose == null) return ReusePlan(candidate) } // The call's connection was released. toClose?.closeQuietly()
Registered: Sun Jun 16 04:42:17 UTC 2024 - Last Modified: Sat Apr 20 17:03:43 UTC 2024 - 12K bytes - Viewed (0) -
internal/dsync/dsync_test.go
stopLockServers() os.Exit(code) } func TestSimpleLock(t *testing.T) { dm := NewDRWMutex(ds, "test") dm.Lock(id, source) // fmt.Println("Lock acquired, waiting...") time.Sleep(testDrwMutexRefreshCallTimeout) dm.Unlock(context.Background()) } func TestSimpleLockUnlockMultipleTimes(t *testing.T) { dm := NewDRWMutex(ds, "test") dm.Lock(id, source)
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Sat Dec 24 03:49:07 UTC 2022 - 11K bytes - Viewed (0) -
platforms/core-execution/persistent-cache/src/main/java/org/gradle/cache/internal/DefaultCacheCoordinator.java
return new BTreePersistentIndexedCache<>(cacheFile, keySerializer, valueSerializer); } /** * Called just after the file lock has been acquired. */ private void afterLockAcquire(FileLock fileLock) { assert this.fileLock == null; this.fileLock = fileLock; this.stateAtOpen = fileLock.getState();
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Wed May 01 12:21:15 UTC 2024 - 20.5K bytes - Viewed (0) -
src/runtime/tracecpu.go
trace.cpuLogRead[0] = newProfBuf(3, profBufWordCount, profBufTagCount) trace.cpuLogRead[1] = newProfBuf(3, profBufWordCount, profBufTagCount) // We must not acquire trace.signalLock outside of a signal handler: a // profiling signal may arrive at any time and try to acquire it, leading to // deadlock. Because we can't use that lock to protect updates to // trace.cpuLogWrite (only use of the structure it references), reads and
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 17:03:35 UTC 2024 - 8.7K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/telemetry/internal/counter/counter.go
switch { case !state.locked() && state.havePtr(): if !c.state.update(&state, state.incReader()) { continue } // Counter unlocked or counter shared; has an initialized count pointer; acquired shared lock. if c.ptr.count == nil { for !c.state.update(&state, state.addExtra(uint64(n))) { // keep trying - we already took the reader lock state = c.state.load() }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jun 04 16:19:04 UTC 2024 - 10.3K bytes - Viewed (0) -
src/runtime/tracestack.go
gp = mp.curg } // Double-check that we own the stack we're about to trace. if debug.traceCheckStackOwnership != 0 && gp != nil { status := readgstatus(gp) // If the scan bit is set, assume we're the ones that acquired it. if status&_Gscan == 0 { // Use the trace status to check this. There are a number of cases // where a running goroutine might be in _Gwaiting, and these cases
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 21 14:38:56 UTC 2024 - 11K bytes - Viewed (0) -
src/runtime/os_netbsd.go
var deadline int64 if ns >= 0 { deadline = nanotime() + ns } for { v := atomic.Load(&gp.m.waitsemacount) if v > 0 { if atomic.Cas(&gp.m.waitsemacount, v, v-1) { return 0 // semaphore acquired } continue } // Sleep until unparked by semawakeup or timeout. var tsp *timespec var ts timespec if ns >= 0 { wait := deadline - nanotime() if wait <= 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 10.1K bytes - Viewed (0) -
src/runtime/traceruntime.go
// //go:nosplit func traceAcquireEnabled() traceLocker { // Any time we acquire a traceLocker, we may flush a trace buffer. But // buffer flushes are rare. Record the lock edge even if it doesn't happen // this time. lockRankMayTraceFlush() // Prevent preemption. mp := acquirem() // Acquire the trace seqlock. This prevents traceAdvance from moving forward
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 25.7K bytes - Viewed (0) -
pkg/kubelet/util/manager/watch_based_manager.go
} func (c *objectCache) startRecycleIdleWatch() { c.lock.Lock() defer c.lock.Unlock() for key, item := range c.items { if item.stopIfIdle(c.clock.Now(), c.maxIdleTime) { klog.V(4).InfoS("Not acquired for long time, Stopped watching for changes", "objectKey", key, "maxIdleTime", c.maxIdleTime) } } } func (c *objectCache) shutdownWhenStopped(stopCh <-chan struct{}) { <-stopCh c.lock.Lock()
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 06:25:43 UTC 2024 - 11.6K bytes - Viewed (0) -
platforms/core-execution/persistent-cache/src/main/java/org/gradle/cache/internal/DefaultFileLockManager.java
* <br><br> * * Algorithm:<br> * 1. We first try to acquire a lock on the state region with retries, see {@link #lockStateRegion(LockMode)}.<br> * 2a. If we use exclusive lock, and we succeed in step 1., then we acquire an exclusive lock * on the information region and write our details (port and lock id) there, and then we release lock of information region.
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Fri Apr 26 16:02:32 UTC 2024 - 22.3K bytes - Viewed (0)