- Sort Score
- Result 10 results
- Languages All
Results 71 - 80 of 84 for Thresholds (0.14 sec)
-
platforms/core-execution/workers/src/main/java/org/gradle/workers/internal/WorkerDaemonExpiration.java
* Simple implementation of memory based expiration. * * Use the maximum heap size of each daemon, not their actual memory usage. * Expire as much daemons as needed to free the requested memory under the threshold. */ private class SimpleMemoryExpirationSelector implements Transformer<List<WorkerDaemonClient>, List<WorkerDaemonClient>> { private final long memoryBytesToRelease; private long releasedBytes;
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Fri Sep 22 09:36:27 UTC 2023 - 5.1K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/stmtlines_test.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 03 17:05:14 UTC 2023 - 3.5K bytes - Viewed (0) -
src/runtime/sema_test.go
// nondeterministic, so this test checks that we get the // desired outcome in a significant majority of cases. // The actual ratio of direct handoffs is much higher // (>90%) but we use a lower threshold to minimize the // chances that unrelated changes in the runtime will // cause the test to fail or become flaky. if ok < iter*2/3 { t.Fatal("direct handoff < 2/3:", ok, iter) } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jun 21 19:37:22 UTC 2022 - 4.2K bytes - Viewed (0) -
internal/config/cache/cache.go
var configLock sync.RWMutex // Enabled - indicates if cache is enabled or not func (c *Config) Enabled() bool { return c.Enable && c.Endpoint != "" } // MatchesSize verifies if input 'size' falls under cacheable threshold func (c Config) MatchesSize(size int64) bool { configLock.RLock() defer configLock.RUnlock() return c.Enable && c.BlockSize > 0 && size <= c.BlockSize } // Update updates new cache frequency
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 5.6K bytes - Viewed (0) -
src/hash/crc64/crc64.go
helperTable = slicing8TableISO // For smaller sizes creating extended table takes too much time } else if len(p) >= 2048 { // According to the tests between various x86 and arm CPUs, 2k is a reasonable // threshold for now. This may change in the future. helperTable = makeSlicingBy8Table(tab) } else { break } // Update using slicing-by-8 for len(p) > 8 { crc ^= byteorder.LeUint64(p)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 18 22:36:41 UTC 2024 - 5.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions_drq.cc
call_op.emitRemark("Quantization is skipped for ") << call_op->getName().getStringRef().str() << " because it has " << current_num_elements << " elements which is fewer than the threshold(" << min_num_elements_for_weights_ << " elements)."; call_op->removeAttr(kQuantTraitAttrName); } } StringRef function_name =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/compile_mlir_util_test.cc
OpPassManager pass_manager; llvm::StringRef device_type = "XLA_CPU_JIT"; absl::string_view kInlinePass = "inline{default-pipeline=canonicalize " "inlining-threshold=4294967295 max-iterations=4 }"; CreateConvertMlirToXlaHloPipeline(pass_manager, device_type, /*enable_op_fallback=*/true,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 25 19:54:38 UTC 2024 - 9.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/tf_to_corert_pipeline.mlir
// RUN: tf-tfrt-opt -tf-executor-to-tfrt-pipeline="enable-optimizer=true tfrt-cost-threshold=1024" %s | FileCheck %s --dump-input=fail // CHECK: tfrt.cost_threshold = 1024 : i64 module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 462 : i32}} { // CHECK-LABEL: func @__forward_call_369 // CHECK-SAME: ([[in_chain:%.*]]: !tfrt.chain // CHECK-SAME: [[arg1:%.*]]: !tfrt_fallback.tf_tensor {tf._user_specified_name = "inputs"},
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 00:18:59 UTC 2024 - 7.7K bytes - Viewed (0) -
src/net/http/header_test.go
"Empty": {}, "Blank": {""}, "Double-Blank": {"", ""}, }, nil, "Blank: \r\nDouble-Blank: \r\nDouble-Blank: \r\n", }, // Tests header sorting when over the insertion sort threshold side: { Header{ "k1": {"1a", "1b"}, "k2": {"2a", "2b"}, "k3": {"3a", "3b"}, "k4": {"4a", "4b"}, "k5": {"5a", "5b"}, "k6": {"6a", "6b"}, "k7": {"7a", "7b"}, "k8": {"8a", "8b"},
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Sep 07 01:07:32 UTC 2022 - 6.1K bytes - Viewed (0) -
src/cmd/compile/internal/inline/inlheur/testdata/props/calls.go
// <endcallsites> // <endfuncpreamble> func T_call_scoring_in_noninlinable_func(x int, sl []int) int { if x == 101 { // Drive up the cost of inlining this funcfunc over the // regular threshold. for i := 0; i < 10; i++ { for j := 0; j < i; j++ { sl = append(sl, append(sl, append(sl, append(sl, x)...)...)...) sl = append(sl, sl[0], sl[1], sl[2]) x += calleeNoInline(x) } } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Nov 16 20:15:25 UTC 2023 - 7.1K bytes - Viewed (0)