- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 146 for REWRITES (0.56 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/verify_quant_legalization.cc
// some generic types that are legal in MHLO. This pass legalizes TF types into // types that are legal in MHLO. For example, TF::Qint8Type is converted to i8. // Rewrites here should run before TF to MHLO op legalizations are run. #include <memory> #include "absl/log/log.h" #include "llvm/ADT/STLExtras.h" #include "llvm/Support/Casting.h"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 3.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/host_runtime/runtime_passes.td
include "mlir/Pass/PassBase.td" def TPURewritePass : Pass<"tf-tpu-rewrite", "mlir::ModuleOp"> { let summary = "Rewrites a `tf_device.cluster_func` on TPUs into TPU runtime operations."; let description = [{ This pass rewrites a `tf_device.cluster_func` operation into a sequence of `tf._TPUCompileMlir` and `tf.TPUExecute` operations. `tf._TPUCompileMlir` contains a MLIR module that is
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jan 10 18:58:57 UTC 2024 - 10.7K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_xla_computations_pass.h
See the License for the specific language governing permissions and limitations under the License. ==============================================================================*/ // Rewrites computations generated by the xla.compile() Python code into // XlaLaunch nodes. // // xla.compile() does two main things: // a) marks operators that make up an XLA computation with the attribute
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 06:59:07 UTC 2024 - 3.6K bytes - Viewed (0) -
pkg/test/framework/resource/matcher.go
i++ case '/': if cs == 0 && cp == 0 { a = append(a, s[:i]) s = s[i+1:] i = 0 continue } } i++ } return append(a, s) } // rewrite rewrites a subname to having only printable characters and no white // space. // From go/src/testing/match.go func rewrite(s string) string { b := []byte{} for _, r := range s { switch { case isSpace(r):
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Fri Feb 12 23:30:37 UTC 2021 - 3.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/quantize.cc
PopulateAllQuantizablePatterns(ctx, patterns); if (failed(applyPatternsAndFoldGreedily(module_op, std::move(patterns)))) { // There are cases where no rewrites happen even if a pattern matches, // causing this to result in a convergence failure. Consider this as a // best-effort. module_op.emitWarning("Failed to converge pattern at QuantizePass."); } }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 22 07:08:19 UTC 2024 - 5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.h
// graph rewrites. Returns false if the inference_type is DT_FLOAT or // `weight_quantization` flag is set. bool RunPropagationAndRewriteQuantizationPasses() const { return inference_type != tensorflow::DT_FLOAT && !weight_quantization; } // TODO: b/202075505 - make implicit weight type clearer // Whether run the passes and graph rewrites for dynamic range quantization.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 13 10:16:19 UTC 2024 - 10.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_communication.cc
auto new_results = GetValueWithToken(builder, llvm::to_vector(terminator->getOperands()), token, rewritten_operands, flatten_tuple); terminator->setOperands(new_results); } // Rewrites a `mhlo.if` op to receive and forward a `mhlo.token`. As If op does // not have any operands other than the predicate, hence we implicitly capture // the parent token. Also we use the same implicit token for use in the If op's
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 40.5K bytes - Viewed (0) -
tensorflow/compiler/jit/extract_outside_compilation_pass.h
const std::map<string, int>& host_compute_core, FunctionLibraryRuntime* flr, FunctionLibraryDefinition* fld, std::vector<string>* shape_inference_graphs, bool* has_outside_compilation); // Rewrites XLA computation in `clusters` to replace outside compilation nodes // with XlaHostCompute, and moves those outside compilations into `g`. If shapes // of outside compilation outputs cannot be determined now, we will store shape
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 5.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/partitioned_topological_sort.cc
// flatbuffer file. if (auto custom_op = dyn_cast<mlir::TFL::CustomOp>(*op)) { return custom_op.getCustomCode().starts_with("Flex"); } // We never see TFL::IfOps in the IR -- it is flatbuffer_export that rewrites // them from TF::IfOps. if (isa<TF::IfOp>(op)) { return false; } // We assume all other TF operators are flex delegated. return op->getDialect()->getNamespace() == "tf"; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Dec 19 15:05:28 UTC 2023 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/optimize.cc
ArrayRef<int64_t> values, Location location) { auto cst_attr = rewriter.getI64TensorAttr(values); return rewriter.create<TF::ConstOp>(location, cst_attr.getType(), cst_attr); } // Rewrites broadcast->reshape to a reshape->broadcast that reduces // the rank of the input and output of the broadcast. class SimplifyBroadcastReshape : public OpRewritePattern<BroadcastToOp> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.1K bytes - Viewed (0)