- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 668 for Dialect (0.22 sec)
-
tensorflow/compiler/mlir/lite/stablehlo/odml_converter/odml_converter_main.cc
#include "llvm/Support/raw_ostream.h" #include "mlir/Dialect/Arith/IR/Arith.h" // from @llvm-project #include "mlir/Dialect/Func/IR/FuncOps.h" // from @llvm-project #include "mlir/Support/LogicalResult.h" // from @llvm-project #include "mlir/Tools/mlir-opt/MlirOptMain.h" // from @llvm-project #include "stablehlo/dialect/ChloOps.h" // from @stablehlo #include "stablehlo/dialect/StablehloOps.h" // from @stablehlo
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 19:05:30 UTC 2024 - 2.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/testing/test_post_calibration_component.cc
#include "mlir/Support/LogicalResult.h" // from @llvm-project #include "mlir/Support/TypeID.h" // from @llvm-project #include "stablehlo/dialect/ChloOps.h" // from @stablehlo // IWYU pragma: keep #include "stablehlo/dialect/StablehloOps.h" // from @stablehlo // IWYU pragma: keep #include "stablehlo/dialect/VhloOps.h" // from @stablehlo // IWYU pragma: keep #include "tensorflow/compiler/mlir/lite/quantization/ir/QuantOps.h" // IWYU pragma: keep
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 12:53:33 UTC 2024 - 3.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/cluster_tf.h
// that is compatible with the given device_type. The MLIR should be in the TF // Executor Dialect for graph nodes and edges or TF Functional. It will convert // to TF Functional internally. Individual Op inside a node should be the // Tensorflow Dialect. The output MLIR is in the TF Functional Dialect. The // input MLIR should not have infeed and outfeed ops, which are unsupported via
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Oct 26 14:08:41 UTC 2023 - 1.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/convert_xla_call_module_op_to_bfloat16.cc
#include <string> #include "absl/status/status.h" #include "absl/status/statusor.h" #include "llvm/Support/raw_ostream.h" #include "mlir/Dialect/Func/IR/FuncOps.h" // from @llvm-project #include "mlir/Dialect/Quant/QuantOps.h" // from @llvm-project // IWYU pragma: keep #include "mlir/Dialect/Shape/IR/Shape.h" // from @llvm-project // IWYU pragma: keep #include "mlir/IR/Builders.h" // from @llvm-project
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 08:32:43 UTC 2024 - 5.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/ir/mlrt/mlrt_dialect.cc
bool would_be_cloned, mlir::IRMapping &mapping) const final { // All mlrt dialect ops can be inlined. return true; } }; } // namespace MlrtDialect::MlrtDialect(mlir::MLIRContext *context) : mlir::Dialect(/*name=*/"mlrt", context, mlir::TypeID::get<MlrtDialect>()) { addTypes<FutureType>(); addTypes<PromiseType>();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 3.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/launch_to_device_attribute.cc
limitations under the License. ==============================================================================*/ #include <memory> #include <vector> #include "mlir/Dialect/Func/IR/FuncOps.h" // from @llvm-project #include "mlir/IR/Attributes.h" // from @llvm-project #include "mlir/IR/Dialect.h" // from @llvm-project #include "mlir/IR/Operation.h" // from @llvm-project #include "mlir/IR/Visitors.h" // from @llvm-project
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/verify_quant_legalization.cc
#include <memory> #include "absl/log/log.h" #include "llvm/ADT/STLExtras.h" #include "llvm/Support/Casting.h" #include "mlir/Dialect/Func/IR/FuncOps.h" // from @llvm-project #include "mlir/Dialect/Quant/QuantOps.h" // from @llvm-project #include "mlir/Dialect/Quant/QuantTypes.h" // from @llvm-project #include "mlir/IR/Operation.h" // from @llvm-project #include "mlir/IR/OperationSupport.h" // from @llvm-project
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 3.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/compile_mlir_util.h
// Lowers MLIR module to XLA HLO inside an XlaComputation. The input module // should only contain operations in tf dialect. If the input module contains // operation in the tf_executor dialect, for example, returns an error. // Exception to this are tf_executor dialect ops that are optimized away through // canonicalization. // // Operations in tf dialect are lowered to XLA HLO through the following steps: // . Legalizes control flow operations.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 21 17:24:39 UTC 2024 - 10.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/post_quantize.td
==============================================================================*/ include "mlir/IR/OpBase.td" include "mlir/IR/PatternBase.td" include "mlir/Dialect/Func/IR/FuncOps.td" include "mlir/Dialect/Arith/IR/ArithOps.td" include "tensorflow/compiler/mlir/tensorflow/ir/tf_ops.td" include "mlir/Dialect/Arith/IR/ArithOps.td" include "tensorflow/compiler/mlir/lite/quantization/ir/QuantOps.td"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Dec 10 05:52:02 UTC 2023 - 1.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops.h
// after control dependences are raise to the standard form. #ifndef TENSORFLOW_COMPILER_MLIR_TENSORFLOW_IR_TF_OPS_H_ #define TENSORFLOW_COMPILER_MLIR_TENSORFLOW_IR_TF_OPS_H_ #include "mlir/Bytecode/BytecodeOpInterface.h" // from @llvm-project // IWYU pragma: keep #include "mlir/Dialect/Func/IR/FuncOps.h" // from @llvm-project #include "mlir/Dialect/Traits.h" // from @llvm-project
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 04 22:48:00 UTC 2024 - 2.8K bytes - Viewed (0)