Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 67 for Unlock1 (0.15 sec)

  1. pkg/scheduler/internal/queue/scheduling_queue.go

    func (p *PriorityQueue) Close() {
    	p.lock.Lock()
    	defer p.lock.Unlock()
    	close(p.stop)
    	p.closed = true
    	p.cond.Broadcast()
    }
    
    // DeleteNominatedPodIfExists deletes <pod> from nominatedPods.
    func (npm *nominator) DeleteNominatedPodIfExists(pod *v1.Pod) {
    	npm.lock.Lock()
    	npm.deleteNominatedPodIfExistsUnlocked(pod)
    	npm.lock.Unlock()
    }
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Jun 12 13:26:09 UTC 2024
    - 61.4K bytes
    - Viewed (0)
  2. pkg/controller/podautoscaler/horizontal.go

    		a.recommendationsLock.Lock()
    		delete(a.recommendations, key)
    		a.recommendationsLock.Unlock()
    
    		a.scaleUpEventsLock.Lock()
    		delete(a.scaleUpEvents, key)
    		a.scaleUpEventsLock.Unlock()
    
    		a.scaleDownEventsLock.Lock()
    		delete(a.scaleDownEvents, key)
    		a.scaleDownEventsLock.Unlock()
    
    		return true, nil
    	}
    	if err != nil {
    		return false, err
    	}
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Sat May 04 18:33:12 UTC 2024
    - 63.6K bytes
    - Viewed (0)
  3. cmd/erasure-server-pool.go

    						mu.Lock()
    						results[i] = info
    						mu.Unlock()
    					}
    				}()
    				// Start scanner. Blocks until done.
    				err := erObj.nsScanner(ctx, allBuckets, wantCycle, updates, healScanMode)
    				if err != nil {
    					scannerLogIf(ctx, err)
    					mu.Lock()
    					if firstErr == nil {
    						firstErr = err
    					}
    					// Cancel remaining...
    					cancel()
    					mu.Unlock()
    					return
    				}
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Thu May 30 11:58:12 UTC 2024
    - 82.5K bytes
    - Viewed (0)
  4. cmd/batch-handlers.go

    		binary.LittleEndian.PutUint16(data[0:2], format)
    		binary.LittleEndian.PutUint16(data[2:4], version)
    
    		buf, err := ri.MarshalMsg(data)
    		ri.mu.Unlock()
    		if err != nil {
    			return err
    		}
    		return saveConfig(ctx, api, getJobReportPath(job), buf)
    	}
    	ri.mu.Unlock()
    	return nil
    }
    
    // Note: to be used only with batch jobs that affect multiple versions through
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Tue Jun 11 03:13:30 UTC 2024
    - 56K bytes
    - Viewed (0)
  5. pkg/kubelet/pod_workers.go

    	p.podLock.Lock()
    	defer p.podLock.Unlock()
    	if status, ok := p.podSyncStatuses[uid]; ok {
    		return status.IsTerminated()
    	}
    	// if the pod is not known, we return false (pod worker is not aware of it)
    	return false
    }
    
    func (p *podWorkers) CouldHaveRunningContainers(uid types.UID) bool {
    	p.podLock.Lock()
    	defer p.podLock.Unlock()
    	if status, ok := p.podSyncStatuses[uid]; ok {
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Tue Apr 02 13:22:37 UTC 2024
    - 74.8K bytes
    - Viewed (0)
  6. src/runtime/pprof/pprof_test.go

    	mu.Lock()
    	go func() {
    		awaitBlockedGoroutine(t, "sync.Mutex.Lock", "blockMutex", 1)
    		mu.Unlock()
    	}()
    	// Note: Unlock releases mu before recording the mutex event,
    	// so it's theoretically possible for this to proceed and
    	// capture the profile before the event is recorded. As long
    	// as this is blocked before the unlock happens, it's okay.
    	mu.Lock()
    }
    
    func blockMutexN(t *testing.T, n int, d time.Duration) {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 18:42:28 UTC 2024
    - 68.8K bytes
    - Viewed (0)
  7. guava/src/com/google/common/collect/MapMakerInternalMap.java

        /** Cleanup collected entries when the lock is available. */
        void tryDrainReferenceQueues() {
          if (tryLock()) {
            try {
              maybeDrainReferenceQueues();
            } finally {
              unlock();
            }
          }
        }
    
        @GuardedBy("this")
        void drainKeyReferenceQueue(ReferenceQueue<K> keyReferenceQueue) {
          Reference<? extends K> ref;
          int i = 0;
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Sat May 18 03:24:34 UTC 2024
    - 90.8K bytes
    - Viewed (0)
  8. staging/src/k8s.io/apiserver/pkg/storage/cacher/cacher.go

    		c.stopLock.Unlock()
    		return
    	}
    	c.stopped = true
    	c.ready.stop()
    	c.stopLock.Unlock()
    	close(c.stopCh)
    	c.stopWg.Wait()
    }
    
    func forgetWatcher(c *Cacher, w *cacheWatcher, index int, scope namespacedName, triggerValue string, triggerSupported bool) func(bool) {
    	return func(drainWatcher bool) {
    		c.Lock()
    		defer c.Unlock()
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Jun 12 10:12:02 UTC 2024
    - 51.8K bytes
    - Viewed (0)
  9. pkg/proxy/nftables/proxier.go

    // called and the state is fully propagated to local cache.
    func (proxier *Proxier) OnServiceSynced() {
    	proxier.mu.Lock()
    	proxier.servicesSynced = true
    	proxier.setInitialized(proxier.endpointSlicesSynced)
    	proxier.mu.Unlock()
    
    	// Sync unconditionally - this is called once per lifetime.
    	proxier.syncProxyRules()
    }
    
    // OnEndpointSliceAdd is called whenever creation of a new endpoint slice object
    // is observed.
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Sat Jun 08 13:48:54 UTC 2024
    - 55.5K bytes
    - Viewed (0)
  10. src/runtime/mgcscavenge.go

    		// might be in the process of firing on some other P; essentially we're
    		// racing with it. That's totally OK. Double wake-ups are perfectly safe.
    		s.timer.stop()
    		unlock(&s.lock)
    	} else {
    		unlock(&s.lock)
    		slept = s.sleepStub(sleepTime)
    	}
    
    	// Stop here if we're cooling down from the controller.
    	if s.controllerCooldown > 0 {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:48:45 UTC 2024
    - 52.3K bytes
    - Viewed (0)
Back to top