- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 24 for walkFn (0.33 sec)
-
cmd/batch-handlers.go
s3Type := r.Target.Type == BatchJobReplicateResourceS3 || r.Source.Type == BatchJobReplicateResourceS3 if err := api.Walk(ctx, r.Source.Bucket, r.Source.Prefix, walkCh, WalkOptions{ Marker: lastObject, Filter: selectObj, AskDisks: walkQuorum, }); err != nil { cancel() // Do not need to retry if we can't list objects on source.
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Tue Jun 11 03:13:30 UTC 2024 - 56K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/debug.go
// Populate "rval" with what we've computed. rval.Slots = state.slots rval.VarSlots = state.varSlots rval.Vars = state.vars rval.LocationLists = state.lists } // liveness walks the function in control flow order, calculating the start // and end state of each block. func (state *debugState) liveness() []*BlockDebug { blockLocs := make([]*BlockDebug, state.f.NumBlocks()) counterTime := int32(1)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 10 19:44:43 UTC 2024 - 58.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/flatbuffer_import.cc
std::vector<mlir::func::FuncOp> to_delete_funcs; module.walk([&](mlir::vhlo::ReduceOpV1 reduce_op) { auto body = symbol_table.lookup<mlir::func::FuncOp>( reduce_op->getAttr("body").cast<mlir::FlatSymbolRefAttr>().getValue()); InlineVhloOpRegion(reduce_op.getBody(), body); reduce_op->removeAttr("body"); to_delete_funcs.push_back(body); }); module.walk([&](mlir::vhlo::ReduceWindowOpV1 reduce_window_op) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 21 18:21:50 UTC 2024 - 66.8K bytes - Viewed (0) -
src/runtime/traceback.go
package runtime import ( "internal/abi" "internal/bytealg" "internal/goarch" "internal/stringslite" "runtime/internal/sys" "unsafe" ) // The code in this file implements stack trace walking for all architectures. // The most important fact about a given architecture is whether it uses a link register. // On systems with link registers, the prologue for a non-leaf function stores the
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 55.1K bytes - Viewed (0) -
staging/src/k8s.io/apiextensions-apiserver/test/integration/ratcheting_test.go
nil, // No need for selectable fields ) return func(new, old *unstructured.Unstructured) { _ = strategy.ValidateUpdate(context.TODO(), new, old) }, nil } // Recursively walks the provided directory and parses the YAML files into // unstructured objects. If there are more than one object in a single file, // they are all added to the returned slice.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue May 28 08:48:22 UTC 2024 - 59.5K bytes - Viewed (0) -
src/runtime/map.go
// // When the hashtable grows, we allocate a new array // of buckets twice as big. Buckets are incrementally // copied from the old bucket array to the new bucket array. // // Map iterators walk through the array of buckets and // return the keys in walk order (bucket #, then overflow // chain order, then bucket index). To maintain iteration // semantics, we never move keys within their bucket (if // we did, keys might be returned 0 or 2 times). When
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 57.6K bytes - Viewed (0) -
cmd/xl-storage.go
formatData []byte nrRequests uint64 major, minor uint32 fsType string immediatePurge chan string // mutex to prevent concurrent read operations overloading walks. rotational bool walkMu *sync.Mutex walkReadMu *sync.Mutex } // checkPathLength - returns error if given path name length more than 255 func checkPathLength(pathName string) error {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon Jun 10 15:51:27 UTC 2024 - 85.3K bytes - Viewed (2) -
src/cmd/compile/internal/ssa/rewrite.go
found: // At this point, r is the first value in a that we find by walking backwards. // if we return anything, r will be it. r := b // Keep going, counting the other a's that we find. They must all dominate r. na := 0 for d > 0 { for _, x := range a { if b == x.Block { na++ } } if na == len(a) { // Found all of a in a backwards walk. We can return r. return r }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 64.2K bytes - Viewed (0) -
src/cmd/go/go_test.go
func TestGoListStdDoesNotIncludeCommands(t *testing.T) { tooSlow(t, "walks all of GOROOT/src") tg := testgo(t) defer tg.cleanup() tg.parallel() tg.run("list", "std") tg.grepStdoutNot("cmd/", "go list std shows commands") } func TestGoListCmdOnlyShowsCommands(t *testing.T) { skipIfGccgo(t, "gccgo does not have GOROOT") tooSlow(t, "walks all of GOROOT/src/cmd") tg := testgo(t)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 81.1K bytes - Viewed (0) -
cmd/erasure-server-pool.go
return z.s3Peer.HealBucket(ctx, bucket, opts) } // Walk a bucket, optionally prefix recursively, until we have returned // all the contents of the provided bucket+prefix. // TODO: Note that most errors will result in a truncated listing. func (z *erasureServerPools) Walk(ctx context.Context, bucket, prefix string, results chan<- itemOrErr[ObjectInfo], opts WalkOptions) error {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu May 30 11:58:12 UTC 2024 - 82.5K bytes - Viewed (0)