- Sort Score
- Result 10 results
- Languages All
Results 71 - 80 of 235 for tpu0 (0.04 sec)
-
tensorflow/compiler/mlir/tf2xla/api/v1/cluster_tf.cc
pm.addPass(mlir::tf_executor::CreateTFExecutorTPUV1IslandOutliningPass()); } // Run the TF XLA Bridge based on the input pipeline, which can be either TPU // bridge pipeline or non TPU bridge pipeline. tensorflow::Status RunTFXLABridge( ModuleOp module, llvm::function_ref<void(OpPassManager &pm)> pipeline_builder, llvm::StringRef module_name = llvm::StringRef(),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 28 22:25:18 UTC 2024 - 9.8K bytes - Viewed (0) -
.bazelrc
# ARM64 WHEEL test:linux_arm64_wheel_test_filters --test_tag_filters=-no_oss,-no_aarch64,-oss_excluded,-oss_serial,-gpu,-tpu,-benchmark-test,-v1only,-no_oss_py38,-no_oss_py39,-no_oss_py310
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 17:12:54 UTC 2024 - 52.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/host_runtime/lower_cluster_to_runtime_ops.cc
return absl::OkStatus(); } // TODO(b/305211853): Unify the CPU/TPU/GPU Execution Ops and thus these two // passes should merge together. void RegisterTPULowerClusterToRuntimeOpsPassPipeline() { static mlir::PassPipelineRegistration<StandardPipelineOptions> pipeline( "tfrt-lower-cluster-to-runtime-ops-tpu", "Run all the passes involved after the clustering transformations from "
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 17 18:52:57 UTC 2024 - 9.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/sparsecore/embedding_sequencing.cc
// For all other inputs that go from TPU op to TPU op, insert the // TPUOutput/Input pair. // Add TPUReplicatedInput/TPUReplicatedOutput pairs along each edge. // TODO(bfontain): Should be merged with the above loop. llvm::SetVector<Value> values_to_add_nodes; for (Operation* op : backward_pass_ops) { // TODO(bfontain): Should validate that all the TPU ops are in the same // replication region.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 39.4K bytes - Viewed (0) -
ci/official/envs/linux_x86_tpu
source ci/official/envs/linux_x86 TFCI_BAZEL_COMMON_ARGS="--repo_env=TF_PYTHON_VERSION=$TFCI_PYTHON_VERSION --config release_cpu_linux --config=tpu" TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_tpu TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow_tpu" TFCI_LIB_SUFFIX="-tpu-linux-x86_64" TFCI_WHL_BAZEL_TEST_ENABLE=0 TFCI_WHL_IMPORT_TEST_ENABLE=0 TFCI_WHL_SIZE_LIMIT=580M
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 27 21:16:27 UTC 2024 - 1.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/utils/test_metadata_config.h
#include <variant> #include <vector> #include "absl/strings/string_view.h" #include "tensorflow/core/framework/tensor_shape.h" #include "tensorflow/core/protobuf/tpu/compile_metadata.pb.h" namespace tensorflow { namespace tf2xla { namespace internal { // Fills in arg_shapes and metadata_proto with appropriate values based on the // input mlir module.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 20:04:11 UTC 2024 - 1.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tpu_resource_read_for_write.cc
getOperation().walk([&](tf_device::ClusterFuncOp cluster_func) { cluster_funcs.push_back(cluster_func); }); OpBuilder builder(&getContext()); // Add resource reads for resource writes from TPU cluster where for such // resources the TPU cluster does not read from. for (tf_device::ClusterFuncOp cluster_func : cluster_funcs) { builder.setInsertionPoint(cluster_func); SmallVector<Value, 4> read_operands;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 16:54:40 UTC 2024 - 5.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tensor_device_copy_conversion.cc
const StringAttr &op_device) { // In TFRT TPU, tensor transfer is handled specifically by D2H and // H2D transfer kernels. So fold the tf.Identity op if: // * the identity op is placed on TPU, and // * the arg to the identity op is produced by a TPUExecuteOp. if (op_device && op_device.getValue().contains("TPU")) { return true; } Operation *def_op = arg.getDefiningOp();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 4.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v2/BUILD
"//tensorflow/core:lib", "//tensorflow/core/platform:statusor", "//tensorflow/core/tpu:tpu_compile", "//tensorflow/core/tpu/kernels:tpu_compile_op_support", "//tensorflow/core/tpu/kernels:tpu_compile_proto_cc", "//tensorflow/core/tpu/kernels:tpu_util_hdrs", "@com_google_absl//absl/log", "@com_google_absl//absl/status", "@com_google_absl//absl/strings",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 23:04:51 UTC 2024 - 10.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/graphdef2mlir/batch_use_same_function/saved_model.pbtxt
#RUN: tf-mlir-translate --savedmodel-signaturedefs-to-mlir-lite -tf-savedmodel-tags=serve,tpu %p | FileCheck %s # Test importing a saved model with 2 signatures that are using a same # BatchFunction Op, which references to a same inference_func from graph_def # library. The result should be that both signatures uses the same # BatchFunction Op (the shared_name is the same) and the same copy of # inference_func. # CHECK: f = @inference_func[[post_fix:[^,]*]]
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 09 16:20:29 UTC 2022 - 2.6K bytes - Viewed (0)