- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 38 for Acquirem (0.28 sec)
-
src/runtime/mgc.go
// allocations. // // TODO(prattmic): cleanup gcStart to use a more explicit "in gcStart" // check for bailing. mp := acquirem() ready := make(chan struct{}, 1) releasem(mp) for gcBgMarkWorkerCount < gomaxprocs { mp := acquirem() // See above, we allocate a closure here. go gcBgMarkWorker(ready) releasem(mp) // N.B. we intentionally wait on each goroutine individually
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 62K bytes - Viewed (0) -
src/runtime/mheap.go
s.state.set(mSpanDead) h.freeMSpanLocked(s) } // scavengeAll acquires the heap lock (blocking any additional // manipulation of the page allocator) and iterates over the whole // heap, scavenging every free page available. // // Must run on the system stack because it acquires the heap lock. // //go:systemstack func (h *mheap) scavengeAll() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 78K bytes - Viewed (0) -
src/runtime/malloc.go
// // This function is idempotent. // // The heap lock must not be held over this operation, since it will briefly acquire // the heap lock. // // Must be called on the system stack because it acquires the heap lock. // //go:systemstack func (h *mheap) enableMetadataHugePages() { // Enable huge pages for page structure. h.pages.enableChunkHugePages()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
src/runtime/mprof.go
// can't record a stack trace. return } if skip > maxSkip { print("requested skip=", skip) throw("invalid skip value") } gp := getg() mp := acquirem() // we must not be preempted while accessing profstack nstk := 1 if tracefpunwindoff() || gp.m.hasCgoOnStack() { mp.profStack[0] = logicalStackSentinel if gp.m.curg == nil || gp.m.curg == gp {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 53.3K bytes - Viewed (0) -
src/main/java/jcifs/smb/SmbSessionImpl.java
throw new ClassCastException(); } /** * @return session increased usage count */ public SmbSessionImpl acquire () { long usage = this.usageCount.incrementAndGet(); if ( log.isTraceEnabled() ) { log.trace("Acquire session " + usage + " " + this); } if ( usage == 1 ) { synchronized ( this ) {
Registered: Wed Jun 12 15:45:55 UTC 2024 - Last Modified: Sun Nov 14 17:41:04 UTC 2021 - 49K bytes - Viewed (0) -
src/main/java/jcifs/smb/SmbTransportImpl.java
return this.transportContext; } /** * {@inheritDoc} * * @see jcifs.util.transport.Transport#acquire() */ @Override public SmbTransportImpl acquire () { return (SmbTransportImpl) super.acquire(); } /** * @return the server's encryption key */ @Override public byte[] getServerEncryptionKey () {
Registered: Wed Jun 12 15:45:55 UTC 2024 - Last Modified: Wed Jan 18 23:47:00 UTC 2023 - 67K bytes - Viewed (0) -
android/guava/src/com/google/common/util/concurrent/AbstractFuture.java
* cancelled without a call to {@code cancel}, such as by calling {@code * setFuture(cancelledFuture)}. * * <p>Beware of completing a future while holding a lock. Its listeners may do slow work or * acquire other locks, risking deadlocks. */ @CanIgnoreReturnValue @Override public boolean cancel(boolean mayInterruptIfRunning) { Object localValue = value; boolean rValue = false;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 63.1K bytes - Viewed (0) -
guava/src/com/google/common/util/concurrent/AbstractFuture.java
* cancelled without a call to {@code cancel}, such as by calling {@code * setFuture(cancelledFuture)}. * * <p>Beware of completing a future while holding a lock. Its listeners may do slow work or * acquire other locks, risking deadlocks. */ @CanIgnoreReturnValue @Override public boolean cancel(boolean mayInterruptIfRunning) { Object localValue = value; boolean rValue = false;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 62.8K bytes - Viewed (1) -
pkg/scheduler/internal/queue/scheduling_queue.go
p.movePodsToActiveOrBackoffQueue(logger, p.getUnschedulablePodsWithMatchingAffinityTerm(logger, newPod), AssignedPodUpdate, oldPod, newPod) } p.lock.Unlock() } // NOTE: this function assumes a lock has been acquired in the caller. // moveAllToActiveOrBackoffQueue moves all pods from unschedulablePods to activeQ or backoffQ. // This function adds all pods and then signals the condition variable to ensure that
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 13:26:09 UTC 2024 - 61.4K bytes - Viewed (0) -
cmd/erasure-server-pool.go
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu May 30 11:58:12 UTC 2024 - 82.5K bytes - Viewed (0)