- Sort Score
- Result 10 results
- Languages All
Results 181 - 190 of 587 for blocked (0.16 sec)
-
android/guava-tests/test/com/google/common/util/concurrent/MoreExecutorsTest.java
} } /** Wait for the given thread to reach the {@link State#TIMED_WAITING} thread state. */ void awaitTimedWaiting(Thread thread) { while (true) { switch (thread.getState()) { case BLOCKED: case NEW: case RUNNABLE: case WAITING: Thread.yield(); break; case TIMED_WAITING: return; case TERMINATED: default:
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Wed Sep 06 17:04:31 UTC 2023 - 28.2K bytes - Viewed (0) -
subprojects/core/src/main/java/org/gradle/execution/plan/DefaultPlanExecutor.java
queue.add(planDetails); maybeStartWorkers(queue, executor); // Run the work from the source from this thread as well, given that it will be blocked waiting for that work to complete anyway WorkerLease currentWorkerLease = workerLeaseService.getCurrentWorkerLease(); MergedQueues thisPlanOnly = new MergedQueues(coordinationService, true);
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Fri Apr 05 16:29:26 UTC 2024 - 29.1K bytes - Viewed (0) -
guava-tests/test/com/google/common/util/concurrent/ServiceManagerTest.java
assertThat(logFormatter.format(record)).doesNotContain("NoOpService"); } } /** * Tests that a ServiceManager can be fully shut down if one of its failure listeners is slow or * even permanently blocked. */ public void testListenerDeadlock() throws InterruptedException { final CountDownLatch failEnter = new CountDownLatch(1); final CountDownLatch failLeave = new CountDownLatch(1);
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Oct 02 17:20:27 UTC 2023 - 23.9K bytes - Viewed (0) -
src/runtime/mprof.go
// SetBlockProfileRate controls the fraction of goroutine blocking events // that are reported in the blocking profile. The profiler aims to sample // an average of one blocking event per rate nanoseconds spent blocked. // // To include every blocking event in the profile, pass rate = 1. // To turn off profiling entirely, pass rate <= 0. func SetBlockProfileRate(rate int) { var r int64 if rate <= 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 53.3K bytes - Viewed (0) -
src/net/lookup_test.go
host string, ) ([]IPAddr, error) { select { case <-unblockLookup: default: // Start a concurrent LookupIPAddr for the same host while the caller is // still blocked, and sleep a little to give it time to be deduplicated // before we cancel (and unblock) the caller. // (If the timing doesn't quite work out, we'll end up testing sequential // calls instead of concurrent ones, but the test should still pass.)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 41.4K bytes - Viewed (0) -
android/guava-tests/test/com/google/common/util/concurrent/ServiceManagerTest.java
assertThat(logFormatter.format(record)).doesNotContain("NoOpService"); } } /** * Tests that a ServiceManager can be fully shut down if one of its failure listeners is slow or * even permanently blocked. */ public void testListenerDeadlock() throws InterruptedException { final CountDownLatch failEnter = new CountDownLatch(1); final CountDownLatch failLeave = new CountDownLatch(1);
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Oct 02 17:20:27 UTC 2023 - 23.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/passes.td
2. Get list of supported block configurations of the op. 3. Calculate random sparsity of the weight. 3.1. If sparsity level is below the encoding threshold, keep in dense. 3.2. If sparsity level is above the encoding threshold, go to 4. 4. Try to encode the weight with supported block configurations. If the weight was pruned with the same block config, the blocked sparsity level
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 20:30:06 UTC 2024 - 22.6K bytes - Viewed (0) -
src/sync/rwmutex_test.go
var wl RWMutex var rl Locker wlocked := make(chan bool, 1) rlocked := make(chan bool, 1) rl = wl.RLocker() n := 10 go func() { for i := 0; i < n; i++ { rl.Lock() rl.Lock() rlocked <- true wl.Lock() wlocked <- true } }() for i := 0; i < n; i++ { <-rlocked rl.Unlock() select { case <-wlocked: t.Fatal("RLocker() didn't read-lock it")
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Oct 29 17:13:13 UTC 2021 - 4.9K bytes - Viewed (0) -
src/sync/map.go
func (m *Map) Load(key any) (value any, ok bool) { read := m.loadReadOnly() e, ok := read.m[key] if !ok && read.amended { m.mu.Lock() // Avoid reporting a spurious miss if m.dirty got promoted while we were // blocked on m.mu. (If further loads of the same key will not miss, it's // not worth copying the dirty map for this key.) read = m.loadReadOnly() e, ok = read.m[key] if !ok && read.amended { e, ok = m.dirty[key]
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 15.6K bytes - Viewed (0) -
src/testing/testing_test.go
// this deadlocked after https://go.dev/cl/506755. block := make(chan struct{}) var ready, done sync.WaitGroup for i := 0; i < 2; i++ { ready.Add(1) done.Add(1) go t.Run("", func(*testing.T) { ready.Done() <-block done.Done() }) } ready.Wait() close(block) done.Wait() } func TestParentRun(t1 *testing.T) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 19:10:41 UTC 2024 - 19.3K bytes - Viewed (0)