- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 34 for yield4 (0.14 sec)
-
tensorflow/compiler/mlir/tf2xla/internal/passes/extract_outside_compilation.cc
builder.create<mlir::TF::YieldOp>(if_region.getLoc(), /*operands=*/ArrayRef<Value>{}); // Create empty else branch region. auto& else_branch = host_side_if.getElseBranch(); else_branch.push_back(new Block); builder.setInsertionPointToEnd(&else_branch.front()); builder.create<mlir::TF::YieldOp>(if_region.getLoc(),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 21:25:12 UTC 2024 - 68.3K bytes - Viewed (0) -
src/runtime/mprof.go
// profile that event by storing a pointer to the lock in its "to be profiled // upon unlock" field. If that field is already set, it uses the relative // magnitudes to weight a random choice between itself and the other lock, with // the loser's time being added to the "additional contention" field. Otherwise // if the M's call stack buffer is occupied, it does the comparison against that // sample's magnitude. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 53.3K bytes - Viewed (0) -
platforms/software/dependency-management/src/integTest/groovy/org/gradle/integtests/resolve/DependencySubstitutionRulesIntegrationTest.groovy
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Tue Apr 02 20:50:18 UTC 2024 - 52.8K bytes - Viewed (0) -
android/guava/src/com/google/common/base/CharMatcher.java
tmp[5 - i] = hex.charAt(c & 0xF); c = (char) (c >> 4); } return String.copyValueOf(tmp); } // Fast matchers /** A matcher for which precomputation will not yield any significant benefit. */ abstract static class FastMatcher extends CharMatcher { @Override public final CharMatcher precomputed() { return this; } @Override
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Feb 09 15:49:48 UTC 2024 - 53.7K bytes - Viewed (0) -
guava/src/com/google/common/cache/CacheBuilder.java
* usually have much noticeable impact. A value of one permits only one thread to modify the cache * at a time, but since read operations and cache loading computations can proceed concurrently, * this still yields higher concurrency than full synchronization. * * <p>Defaults to 4. <b>Note:</b>The default may change in the future. If you care about this * value, you should always choose it explicitly. *
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Thu Feb 15 16:12:13 UTC 2024 - 51.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/flatbuffer_import.cc
auto saved_pos = op_builder.saveInsertionPoint(); op_builder.setInsertionPointToEnd(®ion.front()); mlir::Operation* cloned_op = op_builder.clone(*op); // Add the yield operation. op_builder.create<mlir::TFL::YieldOp>(op_loc, cloned_op->getResults()); // Now emit into the function body again. op_builder.restoreInsertionPoint(saved_pos);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 21 18:21:50 UTC 2024 - 66.8K bytes - Viewed (0) -
guava/src/com/google/common/base/CharMatcher.java
tmp[5 - i] = hex.charAt(c & 0xF); c = (char) (c >> 4); } return String.copyValueOf(tmp); } // Fast matchers /** A matcher for which precomputation will not yield any significant benefit. */ abstract static class FastMatcher extends CharMatcher { @Override public final CharMatcher precomputed() { return this; } @Override
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Feb 09 15:49:48 UTC 2024 - 53.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/prepare-tf.mlir
%7 = "tf.FakeQuantWithMinMaxVarsPerChannel"(%arg1, %arg2, %arg3) {device = "", narrow_range = false, num_bits = 8 : i64} : (tensor<*xf32>, tensor<*xf32>, tensor<*xf32>) -> tensor<*xf32> "tfl.yield"(%7) : (tensor<*xf32>) -> () }) {device = "", narrow_range = false, num_bits = 8 : i64} : (tensor<3x4xf32>, tensor<4xf32>, tensor<4xf32>) -> tensor<3x4xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 07:26:59 UTC 2024 - 59.8K bytes - Viewed (0) -
src/go/types/expr.go
if key == nil { check.errorf(kv, InvalidLitField, "invalid field name %s in struct literal", kv.Key) continue } i := fieldIndex(utyp.fields, check.pkg, key.Name, false) if i < 0 { var alt Object if j := fieldIndex(fields, check.pkg, key.Name, true); j >= 0 { alt = fields[j] } msg := check.lookupError(base, key.Name, alt, true)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 02:09:54 UTC 2024 - 49.7K bytes - Viewed (0) -
pkg/scheduler/framework/runtime/framework.go
// filters fail, we don't run the second pass. // We consider only equal or higher priority pods in the first pass, because // those are the current "pod" must yield to them and not take a space opened // for running them. It is ok if the current "pod" take resources freed for // lower priority pods. // Requiring that the new pod is schedulable in both circumstances ensures that
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri May 17 09:07:27 UTC 2024 - 60.9K bytes - Viewed (0)