Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 15 for north (0.04 sec)

  1. src/runtime/mgcscavenge.go

    	}
    }
    
    // scavenge scavenges nbytes worth of free pages, starting with the
    // highest address first. Successive calls continue from where it left
    // off until the heap is exhausted. force makes all memory available to
    // scavenge, ignoring huge page heuristics.
    //
    // Returns the amount of memory scavenged in bytes.
    //
    // scavenge always tries to scavenge nbytes worth of memory, and will
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:48:45 UTC 2024
    - 52.3K bytes
    - Viewed (0)
  2. src/testing/testing.go

    	// If that happens, we will misattribute the background race to some other
    	// test, or to no test at all — but that false-negative is so unlikely that it
    	// is not worth adding race-report noise for the common case where the test is
    	// completely suspended during the call to Parallel.
    	t.checkRaces()
    
    	if t.chatty != nil {
    		t.chatty.Updatef(t.name, "=== PAUSE %s\n", t.name)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:00:11 UTC 2024
    - 76.1K bytes
    - Viewed (0)
  3. pkg/scheduler/internal/queue/scheduling_queue.go

    	if rejectorPlugins.Len() == 0 {
    		logger.V(6).Info("Worth requeuing because no failed plugins", "pod", klog.KObj(pInfo.Pod))
    		return queueAfterBackoff
    	}
    
    	if event.IsWildCard() {
    		// If the wildcard event is special one as someone wants to force all Pods to move to activeQ/backoffQ.
    		// We return queueAfterBackoff in this case, while resetting all blocked plugins.
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Jun 12 13:26:09 UTC 2024
    - 61.4K bytes
    - Viewed (0)
  4. src/runtime/mbitmap.go

    	default:
    		for i := typ.Size_; i < dataSize; i += typ.Size_ {
    			src |= src0 << (i / goarch.PtrSize)
    			scanSize += typ.Size_
    		}
    	}
    
    	// Since we're never writing more than one uintptr's worth of bits, we're either going
    	// to do one or two writes.
    	dst := span.heapBits()
    	o := (x - span.base()) / goarch.PtrSize
    	i := o / ptrBits
    	j := o % ptrBits
    	if j+bits > ptrBits {
    		// Two writes.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 00:18:55 UTC 2024
    - 60K bytes
    - Viewed (0)
  5. gradle/verification-metadata.xml

                <trusting group="org.testcontainers"/>
                <trusting group="^com[.]diffplug($|([.].*))" regex="true"/>
                <trusting group="^org[.]jetbrains($|([.].*))" regex="true"/>
                <trusting group="^org[.]rnorth($|([.].*))" regex="true"/>
             </trusted-key>
             <trusted-key id="8A10792983023D5D14C93B488D7F1BEC1E2ECAE7" group="^com[.]fasterxml($|([.].*))" regex="true"/>
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Tue May 21 22:30:36 UTC 2024
    - 90.1K bytes
    - Viewed (0)
  6. pilot/pkg/networking/core/listener_test.go

    	// if proxy is not using Iptables and cannot bind to privileged ports (1-1023).
    	//
    	// Even if a user explicitly created a Sidecar config with an ingress listener for a privileged port,
    	// it is still not worth it creating such a listener if we already known that a proxy will end up
    	// rejecting it.
    	testPrivilegedPorts(t, func(t *testing.T, proxy *model.Proxy, port uint32) []*listener.Listener {
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Thu Jun 13 01:56:28 UTC 2024
    - 93.6K bytes
    - Viewed (0)
  7. src/crypto/tls/conn.go

    	// are included in the MAC. Otherwise, an attacker that could
    	// distinguish MAC failures from padding failures could mount an attack
    	// similar to POODLE in SSL 3.0: given a good ciphertext that uses a
    	// full block's worth of padding, replace the final block with another
    	// block. If the MAC check passed but the padding check failed, the
    	// last byte of that block decrypted to the block size.
    	//
    	// See also macAndPaddingGood logic below.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 03:10:12 UTC 2024
    - 51.8K bytes
    - Viewed (0)
  8. src/runtime/mgcmark.go

    		print("scanning async preempted goroutine ", gp.goid, " stack [", hex(gp.stack.lo), ",", hex(gp.stack.hi), ")\n")
    	}
    
    	// Scan the saved context register. This is effectively a live
    	// register that gets moved back and forth between the
    	// register and sched.ctxt without a write barrier.
    	if gp.sched.ctxt != nil {
    		scanblock(uintptr(unsafe.Pointer(&gp.sched.ctxt)), goarch.PtrSize, &oneptrmask[0], gcw, &state)
    	}
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Apr 18 21:25:11 UTC 2024
    - 52.5K bytes
    - Viewed (0)
  9. cmd/api-errors.go

    		HTTPStatusCode: http.StatusBadRequest,
    	},
    
    	// Generic Invalid-Request error. Should be used for response errors only for unlikely
    	// corner case errors for which introducing new APIErrorCode is not worth it. LogIf()
    	// should be used to log the error at the source of the error for debugging purposes.
    	ErrInvalidRequest: {
    		Code:           "InvalidRequest",
    		Description:    "Invalid Request",
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Thu Jun 13 22:26:38 UTC 2024
    - 92.1K bytes
    - Viewed (0)
  10. src/runtime/mgc.go

    		// rather than starting all in a batch and then waiting once
    		// afterwards. By running one goroutine at a time, we can take
    		// advantage of runnext to bounce back and forth between
    		// workers and this goroutine. In an overloaded application,
    		// this can reduce GC start latency by prioritizing these
    		// goroutines rather than waiting on the end of the run queue.
    		<-ready
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 62K bytes
    - Viewed (0)
Back to top