- Sort Score
- Result 10 results
- Languages All
Results 191 - 200 of 596 for region (2.66 sec)
-
src/runtime/trace/annotation_test.go
func BenchmarkStartRegion(b *testing.B) { b.ReportAllocs() ctx, task := NewTask(context.Background(), "benchmark") defer task.End() b.RunParallel(func(pb *testing.PB) { for pb.Next() { StartRegion(ctx, "region").End() } }) } func BenchmarkNewTask(b *testing.B) { b.ReportAllocs() pctx, task := NewTask(context.Background(), "benchmark") defer task.End() b.RunParallel(func(pb *testing.PB) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 16:44:47 UTC 2024 - 721 bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/analysis/resource_alias_analysis.h
// set of unique IDs representing the possible resources it could alias. // // Note that this is not an inter-procedural or inter-regional analysis, i.e., // each function and region are handled separately and cross-function or cross- // region aliasing cannot be checked by this analysis. class ResourceAliasAnalysis : public detail::PerFunctionAggregateAnalysis<
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 15 09:04:13 UTC 2024 - 7.6K bytes - Viewed (0) -
src/internal/trace/event_test.go
_ = ev.Range() }) mustPanic(t, func() { _ = ev.Metric() }) mustPanic(t, func() { _ = ev.Log() }) mustPanic(t, func() { _ = ev.Task() }) mustPanic(t, func() { _ = ev.Region() }) mustPanic(t, func() { _ = ev.Label() }) mustPanic(t, func() { _ = ev.RangeAttributes() }) } func mustPanic(t *testing.T, f func()) { defer func() { if r := recover(); r == nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 784 bytes - Viewed (0) -
src/runtime/mpagealloc.go
// first-fit approach. // // Each entry in the radix tree is a summary that describes three properties of // a particular region of the address space: the number of contiguous free pages // at the start and end of the region it represents, and the maximum number of // contiguous free pages found anywhere in that region. // // Each level of the radix tree is stored as one contiguous array, which represents
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 39.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/sparsecore/embedding_pipelining.cc
// generalizing to multiple TPU clusters. Region* region = (*forward_pass_ops.begin())->getParentRegion(); StringAttr replication_attr = GetReplicationAttr(*forward_pass_ops.begin()); llvm::SmallVector<Operation*> checkset(forward_pass_ops.getArrayRef()); checkset.append(backward_pass_ops.begin(), backward_pass_ops.end()); for (Operation* op : checkset) { if (op->getParentRegion() != region) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 92.9K bytes - Viewed (0) -
staging/src/k8s.io/api/core/v1/well_known_labels.go
AnnotationPeerAdvertiseAddress = "kubernetes.io/peer-advertise-address" LabelTopologyZone = "topology.kubernetes.io/zone" LabelTopologyRegion = "topology.kubernetes.io/region" // These label have been deprecated since 1.17, but will be supported for // the foreseeable future, to accommodate things like long-lived PVs that // use them. New users should prefer the "topology.kubernetes.io/*"
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jul 19 00:36:22 UTC 2023 - 3.8K bytes - Viewed (0) -
src/cmd/trace/pprof.go
}) var lastTimestamp trace.Time var n int // Select only the outermost regions. for _, i := range intervals { if lastTimestamp <= i.start { intervals[n] = i // new non-overlapping region starts. lastTimestamp = i.end n++ } // Otherwise, skip because this region overlaps with a previous region. } gToIntervals[g] = intervals[:n] } return gToIntervals, nil }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 10.1K bytes - Viewed (0) -
src/runtime/mpagecache.go
return c.cache == 0 } // alloc allocates npages from the page cache and is the main entry // point for allocation. // // Returns a base address and the amount of scavenged memory in the // allocated region in bytes. // // Returns a base address of zero on failure, in which case the // amount of scavenged memory should be ignored. func (c *pageCache) alloc(npages uintptr) (uintptr, uintptr) { if c.cache == 0 { return 0, 0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Apr 19 14:30:00 UTC 2023 - 5.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/cluster_ops_by_policy.cc
mlir::LogicalResult PropagateValuesConstraints( mlir::Region ®ion, const ClusteringPolicySet &policies, ValuesConstraintSet &constraints, bool resolve, bool emit_remarks) { // Propagate constraints for all operations in the region. llvm::SmallVector<Operation *> worklist; region.walk([&](Operation *op) { worklist.emplace_back(op); }); // Propagate constraints only through operations inside the `region`.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 27.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/metrics/error_collector_inst.cc
if (absl::StartsWith(op_name, "tf.") || absl::StartsWith(op_name, "tfl.")) { loc_to_name_.emplace(op->getLoc(), op_name); } }; for (auto ®ion : module->getRegions()) { region.walk(collectOps); } pass_name_ = extract_pass_name(pass->getName().str()); error_collector_->Clear(); } void ErrorCollectorInstrumentation::runAfterPass(Pass *pass,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 25 01:48:36 UTC 2024 - 5.3K bytes - Viewed (0)