- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 22 for vSlots (0.59 sec)
-
src/cmd/compile/internal/test/mergelocals_test.go
} } } func TestMergeLocalsIntegration(t *testing.T) { testenv.MustHaveGoBuild(t) // This test does a build of a specific canned package to // check whether merging of stack slots is taking place. // The idea is to do the compile with a trace option turned // on and then pick up on the frame offsets of specific // variables. // // Stack slot merging is a greedy algorithm, and there can
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 18 15:43:53 UTC 2024 - 4.7K bytes - Viewed (0) -
src/internal/coverage/cfile/apis.go
} // Implementation note: this function would be faster and simpler // if we could just zero out the entire counter array, but for the // moment we go through and zero out just the slots in the array // corresponding to the counter values. We do this to avoid the // following bad scenario: suppose that a user builds their Go // program with "-cover", and that program has a function (call it
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 09:57:47 UTC 2024 - 5.4K bytes - Viewed (0) -
src/cmd/vendor/github.com/google/pprof/internal/report/stacks.go
// Alternative names to display (with decreasing lengths) to make text fit. // Guaranteed to be non-empty. Display []string // Places holds the list of stack slots where this source occurs. // In particular, if [a,b] is an element in Places, // StackSet.Stacks[a].Sources[b] points to this source. // // No stack will be referenced twice in the Places slice for a given
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Feb 16 15:19:53 UTC 2024 - 5.6K bytes - Viewed (0) -
src/internal/stringslite/strings.go
c0 := substr[0] c1 := substr[1] i := 0 t := len(s) - n + 1 fails := 0 for i < t { if s[i] != c0 { // IndexByte is faster than bytealg.IndexString, so use it as long as // we're not getting lots of false positives. o := IndexByte(s[i+1:t], c0) if o < 0 { return -1 } i += o + 1 } if s[i+1] == c1 && s[i:i+n] == substr { return i } fails++ i++
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 04 01:23:42 UTC 2024 - 2.9K bytes - Viewed (0) -
src/runtime/pprof/vminfo_darwin_test.go
t.Logf("vmmap output: %s", out) if ee, ok := cmdErr.(*exec.ExitError); ok && len(ee.Stderr) > 0 { t.Logf("%v: %v\n%s", cmd, cmdErr, ee.Stderr) if testing.Short() && strings.Contains(string(ee.Stderr), "No process corpse slots currently available, waiting to get one") { t.Skipf("Skipping knwn flake in short test mode") } retryable = bytes.Contains(ee.Stderr, []byte("resource shortage")) } t.Logf("%v: %v\n", cmd, cmdErr)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 16 19:59:50 UTC 2024 - 5.5K bytes - Viewed (0) -
src/reflect/makefunc.go
// // This is similar to what reflectcallmove does in the runtime, except // that happens on the return path, whereas this happens on the call path. // // nosplit because pointers are being held in uintptr slots in args, so // having our stack scanned now could lead to accidentally freeing // memory. // //go:nosplit func moveMakeFuncArgPtrs(ctxt *makeFuncCtxt, args *abi.RegArgs) { for i, arg := range args.Ints {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 02 15:20:05 UTC 2024 - 5.9K bytes - Viewed (0) -
hack/verify-shellcheck.sh
disabled=( # this lint disallows non-constant source, which we use extensively without # any known bugs 1090 # this lint warns when shellcheck cannot find a sourced file # this wouldn't be a bad idea to warn on, but it fails on lots of path # dependent sourcing, so just disable enforcing it 1091 # this lint prefers command -v to which, they are not the same 2230 ) # comma separate for passing to shellcheck join_by() { local IFS="$1";
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Mar 01 06:07:05 UTC 2024 - 4.9K bytes - Viewed (0) -
tests/binary/binaries_test.go
cases := map[string]struct { minMb int64 maxMb int64 }{ // TODO: shrink the ranges here once the active work to reduce binary size is complete // For now, having two small a range will result in lots of "merge conflicts" "istioctl": {60, 85}, "pilot-agent": {20, 24}, // TODO(https://github.com/kubernetes/kubernetes/issues/101384) bump this down a bit? "pilot-discovery": {60, 85},
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Thu May 23 15:07:03 UTC 2024 - 4.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/split_merged_operands.cc
// to TFLite is actually broken here, we sort of hard-code the variable tensors // based on the actual ops using them, such as unidirectional_sequence_lstm. // // MLIRConverter also benefits from lots of typical compiler optimization like // merging same input values if they're identical. These optimizations are // desirable but not for those TFLite ops which have variable tensors as inputs.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 20:30:06 UTC 2024 - 5.9K bytes - Viewed (0) -
src/runtime/mwbbuf.go
// workbufs. // // This must not have write barriers because it is part of the write // barrier implementation. // // This and everything it calls must be nosplit because 1) the stack // contains untyped slots from gcWriteBarrier and 2) there must not be // a GC safe point between the write barrier test in the caller and // flushing the buffer. // // TODO: A "go:nosplitrec" annotation would be perfect for this. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0)