- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 250 for peak (0.04 sec)
-
cmd/bucket-replication-metrics.go
} func (rx *XferStats) merge(o XferStats) XferStats { curr := calcAvg(rx.curr(), o.curr(), rx.N, o.N) peak := rx.Peak if o.Peak > peak { peak = o.Peak } if curr > peak { peak = curr } return XferStats{ Avg: calcAvg(rx.Avg, o.Avg, rx.N, o.N), Peak: peak, Curr: curr, measure: rx.measure, N: rx.N + o.N, } }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Tue Feb 06 06:00:45 UTC 2024 - 14.2K bytes - Viewed (0) -
cmd/bucket-stats.go
totpeak = math.Max(math.Max(tot.Peak, totpeak), tot.Curr) lpeak = math.Max(math.Max(v[Large].Peak, lpeak), v[Large].Curr) speak = math.Max(math.Max(v[Small].Peak, speak), v[Small].Curr) if lpeak > 0 || speak > 0 { count++ } } if count > 0 { lrg := XferStats{ Avg: lavg / float64(count), Curr: lcurr / float64(count), Peak: lpeak, } sml := XferStats{
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Tue Feb 06 06:00:45 UTC 2024 - 13.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/remat/rematerializer.cc
return profile; } Rematerializer::MemSpec Rematerializer::GetPeakMemory( const RematSpec& remat) const { MemSpec peak; MapMem([&](const MemSpec& m) { peak = std::max(m, peak, BySize); }, remat); return peak; } int Rematerializer::FindBestRematPoint(const int begin, const int end, const int peak_loc) const {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 14 20:57:44 UTC 2023 - 13.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/remat/rematerializer.h
// (2) Allows for the fast simulation of changes to the peak memory requirement // under rematerialization of intermediate results in the graph // (3) Implements a greedy algorithm for finding rematerializations of // intermediate results in that graph to lower peak memory requirements. class Rematerializer { public: Rematerializer() = default;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 14 20:57:44 UTC 2023 - 12K bytes - Viewed (0) -
cmd/bucket-targets.go
} type latencyStat struct { lastmin lastMinuteLatency curr time.Duration avg time.Duration peak time.Duration N int64 } func (l *latencyStat) update(d time.Duration) { l.lastmin.add(d) l.N++ if d > l.peak { l.peak = d } l.curr = l.lastmin.getTotal().avg() l.avg = time.Duration((int64(l.avg)*(l.N-1) + int64(l.curr)) / l.N) }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Wed May 01 01:09:56 UTC 2024 - 20.9K bytes - Viewed (0) -
cmd/bucket-replication-metrics_gen.go
} case "av": z.Avg, err = dc.ReadFloat64() if err != nil { err = msgp.WrapError(err, "Avg") return } case "p": z.Peak, err = dc.ReadFloat64() if err != nil { err = msgp.WrapError(err, "Peak") return } case "n": z.N, err = dc.ReadInt64() if err != nil { err = msgp.WrapError(err, "N") return } default:
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Mar 21 17:21:35 UTC 2024 - 33.3K bytes - Viewed (0) -
src/main/java/org/codelibs/fess/app/web/api/admin/stats/ApiAdminStatsAction.java
final JvmThreadsObj jvmThreadsObj = new JvmThreadsObj(); jvmObj.threads = jvmThreadsObj; jvmThreadsObj.count = threads.getCount(); jvmThreadsObj.peak = threads.getPeakCount(); final Classes classes = jvmStats.getClasses(); final JvmClassesObj jvmClassesObj = new JvmClassesObj(); jvmObj.classes = jvmClassesObj;
Registered: Wed Jun 12 13:08:18 UTC 2024 - Last Modified: Thu Feb 22 01:37:57 UTC 2024 - 12.1K bytes - Viewed (0) -
src/runtime/mgcpacer_test.go
stackBytes: constant(8192), length: 50, checker: func(t *testing.T, c []gcCycleResult) { n := len(c) if peak := c[n-1].heapPeak; peak >= applyMemoryLimitHeapGoalHeadroom(512<<20) { t.Errorf("peak heap size reaches heap limit: %d", peak) } if n >= 25 { // At this alloc/scan rate, the pacer should be extremely close to the goal utilization.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 19 13:53:21 UTC 2023 - 39.3K bytes - Viewed (0) -
tensorflow/cc/saved_model/loader.cc
// not storing the rewritten subgraph for each signature. rewritten_options.config.mutable_experimental() ->set_disable_output_partition_graphs(true); // TODO(mrry): Consider specializing the session creation to reduce peak // RAM consumption by using `Session::Create(GraphDef&&)`. TF_RETURN_IF_ERROR(LoadSavedModelGeneric(rewritten_options, run_options, export_dir, tags, bundle));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 02 04:36:00 UTC 2024 - 23K bytes - Viewed (0) -
src/runtime/heapdump.go
func runtime_debug_WriteHeapDump(fd uintptr) { stw := stopTheWorld(stwWriteHeapDump) // Keep m on this G's stack instead of the system stack. // Both readmemstats_m and writeheapdump_m have pretty large // peak stack depths and we risk blowing the system stack. // This is safe because the world is stopped, so we don't // need to worry about anyone shrinking and therefore moving // our stack. var m MemStats systemstack(func() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 09 04:07:57 UTC 2024 - 17.6K bytes - Viewed (0)