- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 104 for HLO (0.02 sec)
-
tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/argument-sharding.mlir
// RUN: tf-mlir-translate -mlir-tf-to-hlo-text %s -tf-input-shapes=128,10:10,1024:128,1024 -tf-xla-emit-use-tuple-args -tf-xla-emit-return-tuple | FileCheck %s module attributes {tf.versions = {producer = 179 : i32}} { func.func @main(%arg0: tensor<128x10xf32> {mhlo.sharding = "\08\03\1A\02\01\02\22\02\00\01"}, %arg1: tensor<10x1024xf32> {mhlo.sharding = "\08\01\1A\01\01\22\01\00"}, %arg2: tensor<128x1024xf32> {mhlo.sharding = ""}) { func.return } }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 28 12:06:33 UTC 2022 - 1.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.h
#include "tensorflow/core/tpu/kernels/tpu_compile.pb.h" #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h" namespace tensorflow { namespace tf2xla { namespace v1 { // Compiles the given Tensorflow graph into xla::HLO. The result is in // compilation_result. If the input computation is in MLIR, it will be // converted to a Tensorflow graph. Otherwise, the graph compiler will be run. absl::Status CompileTensorflowGraphToHlo(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 08:08:57 UTC 2024 - 2.1K bytes - Viewed (0) -
tensorflow/compiler/jit/get_compiler_ir.cc
auto is_tfrt_tpu_supported_stage = [](IrExportStage stage) { return stage == IrExportStage::HLO || stage == IrExportStage::HLO_NO_METADATA || stage == IrExportStage::HLO_SERIALIZED; }; // TODO(b/238830423): support GetCompilerIr on TFRT TPU device for stages // that requires compilation from HLO to executable. if (device_type != DEVICE_CPU && stream == nullptr &&
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 06:59:07 UTC 2024 - 19K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/stablehlo_add.mlir
// RUN: tf-mlir-translate -mlir-tf-to-hlo-text %s -tf-input-shapes=: -tf-xla-emit-return-tuple | FileCheck %s // TODO(b/259459405): Remove this test along with the upstream refactoring to // avoid non TF inputs. // This is not a supported mode. module attributes {tf.versions = {producer = 179 : i32}} { func.func @main(%arg0: tensor<f32>, %arg1: tensor<f32>) -> tensor<f32> { %0 = "stablehlo.add"(%arg0, %arg1) : (tensor<f32>, tensor<f32>) -> tensor<f32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Nov 17 02:51:12 UTC 2022 - 892 bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tf_to_hlo_pipeline/sccp-post-shape-inference.mlir
// RUN: tf-opt -tf-to-hlo-pipeline %s | FileCheck %s // Verifies that constants generated post shape inference are propagated. // get_shape result in this test. module attributes {tf.versions = {producer = 179 : i32}} { // CHECK-LABEL: func @main func.func @main(%arg0: tensor<10x19xf32>, %arg1: tensor<19x10xf32> {mhlo.is_same_data_across_replicas = true}) -> tensor<?xi64> { %0 = "tf.Shape"(%arg0) : (tensor<10x19xf32>) -> tensor<2xi64>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jul 25 02:54:34 UTC 2023 - 1020 bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/tf_xla_passes.td
include "mlir/Pass/PassBase.td" def LegalizeTFCommunicationPass : Pass<"xla-legalize-tf-communication", "ModuleOp"> { let summary = "Legalize TF/XLA communication ops (TensorFlow dialect) to the HLO " "dialect"; let dependentDialects = ["mhlo::MhloDialect", "sparse_tensor::SparseTensorDialect"]; let constructor = "::mlir::mhlo::CreateLegalizeTFCommunicationPass()"; let description = [{
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 03 05:56:39 UTC 2023 - 1.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/replicate-tensor-list-init-ops.mlir
// RUN: tf-mlir-translate -mlir-tf-to-hlo-text %s -tf-input-shapes=: | FileCheck %s module attributes {tf.versions = {producer = 179 : i32}} { func.func @main() -> (tensor<300x?xf32>, tensor<300x?xf32>) { %elem_shape = "tf.Const"() {value = dense<-1> : tensor<i32>} : () -> tensor<i32> %size = "tf.Const"() {value = dense<300> : tensor<i32>} : () -> tensor<i32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Feb 03 09:30:08 UTC 2023 - 1.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/graph-resource.mlir
// RUN: tf-mlir-translate -mlir-tf-graph-to-hlo-text %s -tf-input-shapes=2:2 -tf-input-data-types=DT_FLOAT,DT_FLOAT -tf-xla-input-types=parameter,resource -tf-xla-emit-return-tuple | FileCheck %s module attributes {tf.versions = {producer = 511 : i32}} { func.func @main(%arg0: tensor<*xf32>, %arg1: tensor<*x!tf_type.resource>) { tf_executor.graph {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 28 12:06:33 UTC 2022 - 1.3K bytes - Viewed (0) -
tensorflow/compiler/jit/tests/device_compiler_serialize_options_test.cc
tensorflow::testing::TmpDir(), "my_test_prefix")); // Run again and these should all hit in the persistent cache despite having // altered the persistent cache entries' HLO modules (disabled strict // signature checks). listener()->ClearListenerHistory(); for (int b = 1; b < 4; ++b) { TF_ASSERT_OK(ExecuteWithBatch(graph, b)); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 15 15:13:38 UTC 2023 - 2.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/g3doc/dialects.md
# MLIR dialects ## Overview To separate different hardware and software targets, MLIR has “dialects”, including: * TensorFlow IR, which represents all things possible in TensorFlow graphs. * XLA HLO IR, which is designed to take advantage of XLA’s compilation abilities (with output to, among other things, TPUs). * An experimental affine dialect, which focuses on [polyhedral representations](https://en.wikipedia.org/wiki/Polytope_model)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Feb 21 01:37:38 UTC 2020 - 1.7K bytes - Viewed (0)