Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 40 of 251 for growing (0.42 sec)

  1. operator/pkg/tpath/tree.go

    		return nil
    	}
    	return setPathContext(nc.Parent, nc.Parent.Node, false) // note: tail recursive
    }
    
    // setValueContext writes the given value to the Node in the given PathContext.
    // If setting the value requires growing the final slice, grows it.
    func setValueContext(nc *PathContext, value any, merge bool) (bool, error) {
    	if nc.Parent == nil {
    		return false, nil
    	}
    
    	vv, mapFromString := tryToUnmarshalStringToYAML(value)
    
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Tue May 23 17:08:31 UTC 2023
    - 17.5K bytes
    - Viewed (0)
  2. src/cmd/vendor/golang.org/x/text/transform/transform.go

    	if (err == nil && pSrc == len(s)) ||
    		(err != nil && err != ErrShortDst && err != ErrShortSrc) {
    		return string(dst[:pDst]), pSrc, err
    	}
    
    	// Transform the remaining input, growing dst and src buffers as necessary.
    	for {
    		n := copy(src, s[pSrc:])
    		atEOF := pSrc+n == len(s)
    		nDst, nSrc, err := t.Transform(dst[pDst:], src[:n], atEOF)
    		pDst += nDst
    		pSrc += nSrc
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Jan 24 13:01:26 UTC 2024
    - 21.7K bytes
    - Viewed (0)
  3. src/runtime/mgcscavenge.go

    // RSS. As a result, we need to take into account for fragmentation internal to
    // spans. heapGoal / lastHeapGoal defines the ratio between the current heap goal
    // and the last heap goal, which tells us by how much the heap is growing and
    // shrinking. We estimate what the heap will grow to in terms of pages by taking
    // this ratio and multiplying it by heapInUse at the end of the last GC, which
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:48:45 UTC 2024
    - 52.3K bytes
    - Viewed (0)
  4. src/runtime/mgcpacer.go

    	if work > scanWorkExpected {
    		// We've already done more scan work than expected. Because our expectation
    		// is based on a steady-state scannable heap size, we assume this means our
    		// heap is growing. Compute a new heap goal that takes our existing runway
    		// computed for scanWorkExpected and extrapolates it to maxScanWork, the worst-case
    		// scan work. This keeps our assist ratio stable if the heap continues to grow.
    		//
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 55.4K bytes
    - Viewed (0)
  5. src/strings/strings.go

    	// larger chunks as the source of the write, as when the source
    	// is too large we are basically just thrashing the CPU D-cache.
    	// So if the result length is larger than an empirically-found
    	// limit (8KB), we stop growing the source string once the limit
    	// is reached and keep reusing the same source string - that
    	// should therefore be always resident in the L1 cache - until we
    	// have completed the construction of the result.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 03 16:48:16 UTC 2024
    - 31.2K bytes
    - Viewed (0)
  6. src/encoding/json/decode.go

    		break
    	}
    
    	i := 0
    	for {
    		// Look ahead for ] - can only happen on first iteration.
    		d.scanWhile(scanSkipSpace)
    		if d.opcode == scanEndArray {
    			break
    		}
    
    		// Expand slice length, growing the slice if necessary.
    		if v.Kind() == reflect.Slice {
    			if i >= v.Cap() {
    				v.Grow(1)
    			}
    			if i >= v.Len() {
    				v.SetLen(i + 1)
    			}
    		}
    
    		if i < v.Len() {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 00:18:55 UTC 2024
    - 35.3K bytes
    - Viewed (0)
  7. pkg/scheduler/schedule_one.go

    	if !sched.hasExtenderFilters() && !sched.hasScoring(fwk) {
    		numNodesToFind = 1
    	}
    
    	// Create feasible list with enough space to avoid growing it
    	// and allow assigning.
    	feasibleNodes := make([]*framework.NodeInfo, numNodesToFind)
    
    	if !fwk.HasFilterPlugins() {
    		for i := range feasibleNodes {
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Thu Jun 06 13:28:08 UTC 2024
    - 43.4K bytes
    - Viewed (0)
  8. src/runtime/export_test.go

    		// remains true, since otherwise the second readMetricsLocked below could
    		// allocate before it returns.
    		readMetricsLocked(samplesp, len, cap)
    
    		// Read memstats first. It's going to flush
    		// the mcaches which readMetrics does not do, so
    		// going the other way around may result in
    		// inconsistent statistics.
    		readmemstats_m(memStats)
    
    		// Read metrics again. We need to be sure we're on the
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:50:53 UTC 2024
    - 46.1K bytes
    - Viewed (0)
  9. src/runtime/mgcmark.go

    			// mark the object.
    			//
    			// Note that it's possible for findObject to
    			// fail if obj points to a just-allocated heap
    			// object because of a race with growing the
    			// heap. In this case, we know the object was
    			// just allocated and hence will be marked by
    			// allocation itself.
    			if obj, span, objIndex := findObject(obj, b, addr-b); obj != 0 {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Apr 18 21:25:11 UTC 2024
    - 52.5K bytes
    - Viewed (0)
  10. src/strings/strings_test.go

    	// 1.  Grow. This triggers two reallocations in Map.
    	maxRune := func(rune) rune { return unicode.MaxRune }
    	m := Map(maxRune, a)
    	expect := tenRunes(unicode.MaxRune)
    	if m != expect {
    		t.Errorf("growing: expected %q got %q", expect, m)
    	}
    
    	// 2. Shrink
    	minRune := func(rune) rune { return 'a' }
    	m = Map(minRune, tenRunes(unicode.MaxRune))
    	expect = a
    	if m != expect {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 03 12:58:37 UTC 2024
    - 53K bytes
    - Viewed (0)
Back to top