- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 135 for tpu0 (0.04 sec)
-
tensorflow/compiler/mlir/tensorflow/transforms/tpu_cluster_cleanup_attributes.cc
// device. Device attribute is used to infer the appropriate sharding // within TPUs for this op. // TODO(b/183598857): Use explicit sharding ops from the front-end. // For example, dequeue ops generated by // tensorflow/python/tpu/tpu_feed.py if (!tensorflow::IsTPUReplicatedCore(attr.getValue()) && !isa<tf_device::LaunchOp>(op)) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Oct 05 23:50:19 UTC 2022 - 3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir.h
std::vector<tpu::ShardingAndIndex>* arg_core_mapping, std::vector<std::vector<xla::Shape>>* per_core_arg_shapes); // Compiles a serialized MLIR module into XLA HLO, generates all accompanying // metadata and stores them in CompilationResult. absl::StatusOr<XlaCompilationResult> LegalizeWithMlirBridge( const tpu::MlirToHloArgs& computation, const tpu::TPUCompileMetadataProto& metadata, bool use_tuple_args,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 2.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/g3doc/enable_mlir_bridge.md
# Enable MLIR-Based new TPU Bridge **MLIR-Based new TPU Bridge is an experimental feature, tread lightly.** ## For TF 1.x-Based Models In tf.ConfigProto.Experimental, there is a knob controlling whether the new TPU Bridge is enabled or not. You can set it by using the following example code: ``` session_config = tf.ConfigProto( ...... experimental=tf.ConfigProto.Experimental( enable_mlir_bridge=True, ), ...... ) ```
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jan 13 23:12:13 UTC 2020 - 989 bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/executor_tpuv1_outline_island/case_op.mlir
"tf.TPUReplicateMetadata"() {_xla_compile_device_type = "TPU", _replication_info = "cluster", device = "device", num_replicas = 1, topology = "topology"} : () -> () %index = "tf.opA"(%arg0) {_xla_compile_device_type = "TPU", _replication_info = "cluster"} : (tensor<i1>) -> tensor<i32> %input = "tf.opB"(%arg0) {_xla_compile_device_type = "TPU", _replication_info = "cluster"} : (tensor<i1>) -> tensor<i32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 27 16:14:17 UTC 2022 - 2.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/executor_tpuv1_outline_island/executor_tpuv1_outline_tpu_island.mlir
tf_executor.graph { %1 = tf_executor.island { "tf.SomeTpuOp"() {_xla_compile_device_type = "TPU"} : () -> () "tf.SomeOtherTpuOp"() {_xla_compile_device_type = "TPU"} : () -> () tf_executor.yield } %2 = tf_executor.island { "tf.TPUReplicateMetadata"() {_xla_compile_device_type = "TPU", _replication_info = "replicate", device = "device", num_replicas = 1, topology = "topology"} : () -> ()
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Jun 04 03:54:58 UTC 2022 - 4.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo.cc
#include "tensorflow/core/platform/status.h" #include "tensorflow/core/protobuf/tpu/compile_metadata.pb.h" #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h" #include "tsl/platform/statusor.h" namespace tensorflow { namespace tf2xla { namespace internal { using metrics::IncrementTfMlirBridgeSecondPhaseCounter; using metrics::MlirBridgeSecondPhaseMetric; using tpu::MlirToHloArgs; absl::StatusOr<XlaCompilationResult> LegalizeTfToHlo(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 3.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/convert_to_legacy_compile_and_replicate_attributes.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 22:03:30 UTC 2024 - 6.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/mark_input_output_aliases.mlir
%1 = "tf.ReadVariableOp"(%arg1) : (!tf_res_f32) -> tensor<f32> %2 = "tf.ReadVariableOp"(%arg2) : (!tf_res_f32) -> tensor<f32> %device_output:2 = "tf_device.cluster_func"(%0, %1, %2) {_xla_compile_device_type = "TPU", _replication_info = "tpu", func = @device_func_0} : (tensor<i32>, tensor<f32>, tensor<f32>) -> (tensor<f32>, tensor<i32>) "tf.AssignVariableOp"(%arg1, %device_output#0) : (!tf_res_f32, tensor<f32>) -> ()
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 04:14:26 UTC 2024 - 6.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.h
absl::Status CompileTensorflowGraphToHlo( const std::variant<tpu::MlirToHloArgs, tpu::FunctionToHloArgs>& computation, const tpu::TPUCompileMetadataProto& metadata, bool use_tuple_args, XlaShapeLayoutHelpers::ShapeDeterminationFns shape_determination_funcs, const std::vector<tensorflow::TensorShape>& arg_shapes, std::vector<tpu::ShardingAndIndex>* arg_core_mapping,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 08:08:57 UTC 2024 - 2.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo.h
#include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h" #include "tsl/platform/statusor.h" namespace tensorflow { namespace tf2xla { namespace internal { // Legalize the given MLIR module to XLA HLO using a combination of the MLIR // Bridge and XlaBuilder absl::StatusOr<XlaCompilationResult> LegalizeTfToHlo( const tpu::MlirToHloArgs& computation,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 2K bytes - Viewed (0)