- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 17 for tpu0 (0.04 sec)
-
tensorflow/compiler/mlir/g3doc/_includes/tf_passes.md
This pass expands outside compilation attributes to Identity/Cast ops at the head of TPU computation if it's only used by outside compiled ops. ### `-tf-tpu-identity-pruning` _Removes Identity/IdentityN ops from the TPU computation_ ### `-tf-tpu-merge-variables-with-execute` _Merges device variable reads and updates into TPU execute ops_
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Aug 02 02:26:39 UTC 2023 - 96.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tpu_cluster_formation.mlir
%3 = "tf.ReadVariableOp"(%0) {_xla_compile_device_type = "TPU", _replication_info = "a"} : (!tf_res) -> tensor<f32> %4 = "tf.ReadVariableOp"(%1) {_xla_compile_device_type = "TPU", _replication_info = "b"} : (!tf_res) -> tensor<f32> %5 = "tf.ReadVariableOp"(%2) {_xla_compile_device_type = "TPU", _replication_info = "c"} : (!tf_res) -> tensor<f32> %6 = "tf.Identity"(%3) {_xla_compile_device_type = "TPU", _replication_info = "a"} : (tensor<f32>) -> tensor<f32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 22:03:30 UTC 2024 - 53.9K bytes - Viewed (0) -
.bazelrc
# ARM64 WHEEL test:linux_arm64_wheel_test_filters --test_tag_filters=-no_oss,-no_aarch64,-oss_excluded,-oss_serial,-gpu,-tpu,-benchmark-test,-v1only,-no_oss_py38,-no_oss_py39,-no_oss_py310
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 17:12:54 UTC 2024 - 52.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops.td
let summary = "Op that compiles a computation in MLIR into a TPU program, and loads and executes it on a TPU device."; let description = [{ For the internal use of the TPU compiler. 'static_shapes' are tensors specifying the maximum dimension sizes for the tensors specified in `dynamic_operands`. 'args' are inputs to the TPU computation.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 04:08:35 UTC 2024 - 90.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/sparsecore/embedding_pipelining.cc
// For all other inputs that go from TPU op to TPU op, insert the // TPUOutput/Input pair. // Add TPUReplicatedInput/TPUReplicatedOutput pairs along each edge. // TODO(bfontain): Should be merged with the above loop. llvm::SetVector<Value> values_to_add_nodes; for (Operation* op : backward_pass_ops) { // TODO(bfontain): Should validate that all the TPU ops are in the same // replication region.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 92.9K bytes - Viewed (0) -
src/runtime/mbitmap.go
return } tp0 := s.typePointersOfType(typ, addr) tp1 := s.typePointersOf(addr, size) failed := false for { var addr0, addr1 uintptr tp0, addr0 = tp0.next(addr + size) tp1, addr1 = tp1.next(addr + size) if addr0 != addr1 { failed = true break } if addr0 == 0 { break } } if failed { tp0 := s.typePointersOfType(typ, addr)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 60K bytes - Viewed (0) -
tensorflow/compiler/jit/BUILD
]), alwayslink = 1, ) cc_library( name = "xla_tpu_jit", visibility = ["//visibility:public"], deps = if_libtpu([ "//tensorflow/core/tpu/graph_rewrite:tpu_rewrite_pass_registration", "//tensorflow/core/tpu/graph_rewrite:configure_tpu_embedding_rewrite_registration", "@local_xla//xla/stream_executor/tpu:tpu_transfer_manager", ]), alwayslink = 1, ) cc_library(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 00:41:19 UTC 2024 - 61.5K bytes - Viewed (0) -
hack/local-up-cluster.sh
ETCD_PID= fi } function print_color { message=$1 prefix=${2:+$2: } # add colon only if defined color=${3:-1} # default is red echo -n "$(tput bold)$(tput setaf "${color}")" echo "${prefix}${message}" echo -n "$(tput sgr0)" } function warning_log { print_color "$1" "W$(date "+%m%d %H:%M:%S")]" 1 } function start_etcd { echo "Starting etcd"
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 25 02:33:52 UTC 2024 - 53.3K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass_test.cc
// TODO(b/117085735): This form of clustering should be prevented. TEST(XlaCompilationTest, NOT_DontClusterSpreadingNodes) { // MatMulSource below creates data for nodes on GPU0 and GPU1 and is placed // on GPU0. However, it should not be clustered with the next node on // GPU0, because that will prevent the node on GPU1 from beginning its work as // soon as the data has been produced. // // This graph is:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 14 10:11:10 UTC 2024 - 79.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/passes/extract_outside_compilation.cc
// _XlaSendFromHost but the _XlaCompileMlir has not yet been created for device // cluster that contains the outside compiled ops. This placeholder should be // replaced by the TPU cluster _XlaCompileMlir in a subsequent pass. mlir::TF::_XlaCompileMlirPlaceholderProgramKeyOp CreateCompilationKeyPlaceholder(Location loc, OpBuilder& builder) { auto result_type =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 21:25:12 UTC 2024 - 68.3K bytes - Viewed (0)