- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 251 for growing (0.42 sec)
-
operator/pkg/tpath/tree.go
return nil } return setPathContext(nc.Parent, nc.Parent.Node, false) // note: tail recursive } // setValueContext writes the given value to the Node in the given PathContext. // If setting the value requires growing the final slice, grows it. func setValueContext(nc *PathContext, value any, merge bool) (bool, error) { if nc.Parent == nil { return false, nil } vv, mapFromString := tryToUnmarshalStringToYAML(value)
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Tue May 23 17:08:31 UTC 2023 - 17.5K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/text/transform/transform.go
if (err == nil && pSrc == len(s)) || (err != nil && err != ErrShortDst && err != ErrShortSrc) { return string(dst[:pDst]), pSrc, err } // Transform the remaining input, growing dst and src buffers as necessary. for { n := copy(src, s[pSrc:]) atEOF := pSrc+n == len(s) nDst, nSrc, err := t.Transform(dst[pDst:], src[:n], atEOF) pDst += nDst pSrc += nSrc
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jan 24 13:01:26 UTC 2024 - 21.7K bytes - Viewed (0) -
src/runtime/mgcscavenge.go
// RSS. As a result, we need to take into account for fragmentation internal to // spans. heapGoal / lastHeapGoal defines the ratio between the current heap goal // and the last heap goal, which tells us by how much the heap is growing and // shrinking. We estimate what the heap will grow to in terms of pages by taking // this ratio and multiplying it by heapInUse at the end of the last GC, which
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:48:45 UTC 2024 - 52.3K bytes - Viewed (0) -
src/runtime/mgcpacer.go
if work > scanWorkExpected { // We've already done more scan work than expected. Because our expectation // is based on a steady-state scannable heap size, we assume this means our // heap is growing. Compute a new heap goal that takes our existing runway // computed for scanWorkExpected and extrapolates it to maxScanWork, the worst-case // scan work. This keeps our assist ratio stable if the heap continues to grow. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 55.4K bytes - Viewed (0) -
src/strings/strings.go
// larger chunks as the source of the write, as when the source // is too large we are basically just thrashing the CPU D-cache. // So if the result length is larger than an empirically-found // limit (8KB), we stop growing the source string once the limit // is reached and keep reusing the same source string - that // should therefore be always resident in the L1 cache - until we // have completed the construction of the result.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 16:48:16 UTC 2024 - 31.2K bytes - Viewed (0) -
src/encoding/json/decode.go
break } i := 0 for { // Look ahead for ] - can only happen on first iteration. d.scanWhile(scanSkipSpace) if d.opcode == scanEndArray { break } // Expand slice length, growing the slice if necessary. if v.Kind() == reflect.Slice { if i >= v.Cap() { v.Grow(1) } if i >= v.Len() { v.SetLen(i + 1) } } if i < v.Len() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 35.3K bytes - Viewed (0) -
pkg/scheduler/schedule_one.go
if !sched.hasExtenderFilters() && !sched.hasScoring(fwk) { numNodesToFind = 1 } // Create feasible list with enough space to avoid growing it // and allow assigning. feasibleNodes := make([]*framework.NodeInfo, numNodesToFind) if !fwk.HasFilterPlugins() { for i := range feasibleNodes {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Jun 06 13:28:08 UTC 2024 - 43.4K bytes - Viewed (0) -
src/runtime/export_test.go
// remains true, since otherwise the second readMetricsLocked below could // allocate before it returns. readMetricsLocked(samplesp, len, cap) // Read memstats first. It's going to flush // the mcaches which readMetrics does not do, so // going the other way around may result in // inconsistent statistics. readmemstats_m(memStats) // Read metrics again. We need to be sure we're on the
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:50:53 UTC 2024 - 46.1K bytes - Viewed (0) -
src/runtime/mgcmark.go
// mark the object. // // Note that it's possible for findObject to // fail if obj points to a just-allocated heap // object because of a race with growing the // heap. In this case, we know the object was // just allocated and hence will be marked by // allocation itself. if obj, span, objIndex := findObject(obj, b, addr-b); obj != 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 18 21:25:11 UTC 2024 - 52.5K bytes - Viewed (0) -
src/strings/strings_test.go
// 1. Grow. This triggers two reallocations in Map. maxRune := func(rune) rune { return unicode.MaxRune } m := Map(maxRune, a) expect := tenRunes(unicode.MaxRune) if m != expect { t.Errorf("growing: expected %q got %q", expect, m) } // 2. Shrink minRune := func(rune) rune { return 'a' } m = Map(minRune, tenRunes(unicode.MaxRune)) expect = a if m != expect {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 12:58:37 UTC 2024 - 53K bytes - Viewed (0)