- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 148 for tpu0 (0.04 sec)
-
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo.h
#include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h" #include "tsl/platform/statusor.h" namespace tensorflow { namespace tf2xla { namespace internal { // Legalize the given MLIR module to XLA HLO using a combination of the MLIR // Bridge and XlaBuilder absl::StatusOr<XlaCompilationResult> LegalizeTfToHlo( const tpu::MlirToHloArgs& computation,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir.cc
#include "tensorflow/core/platform/profile_utils/cpu_utils.h" #include "tensorflow/core/platform/types.h" #include "tensorflow/core/protobuf/tpu/compile_metadata.pb.h" #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h" #include "tensorflow/core/tpu/tpu_compile.h" #include "tsl/platform/error_logging.h" #include "tsl/platform/errors.h" #include "tsl/platform/statusor.h" namespace tensorflow {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 6.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/function/function.h
// If true, use ServingCoreSelector to pick TPU core. Otherwise, obtain core // location from assigned device name. // Currently we don't use core_selector for training use cases. bool tpu_use_core_selector = false; // If true, use BundledTransferToTpuOp to transfer variables and input tensors // to TPU. bool tpu_use_bundled_transfer = false; // If true, lower an TF op that's placed on TPU device to be executed with
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 17 04:50:20 UTC 2024 - 3.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/set_tpu_infeed_layout.cc
#include "xla/shape.h" #include "xla/stream_executor/tpu/c_api_conversions.h" #include "xla/stream_executor/tpu/tpu_api.h" #include "xla/translate/mhlo_to_hlo/type_to_shape.h" namespace mlir { static FailureOr<std::vector<int64_t>> GetTPUInfeedLayoutFromAPI( RankedTensorType t) { // Call the TPU API to determine the right infeed layout. Note that // this can fail if we're not running on a TPU-enabled node.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 6.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf.h
// client - The Xla Compilation client. absl::StatusOr<tensorflow::XlaCompilationResult> LegalizeMlirToHlo( const std::variant<tpu::MlirToHloArgs, tpu::FunctionToHloArgs>& computation, const tpu::TPUCompileMetadataProto& metadata, bool use_tuple_args, llvm::StringRef device_type, std::vector<std::unique_ptr<mlir::Pass>>& custom_legalization_passes,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 07:32:57 UTC 2024 - 2.9K bytes - Viewed (0) -
istioctl/pkg/metrics/metrics.go
return time.Duration(val) * time.Millisecond } func printHeader(writer io.Writer) { w := tabwriter.NewWriter(writer, 13, 1, 2, ' ', tabwriter.AlignRight) _, _ = fmt.Fprintf(w, "%40s\tTOTAL RPS\tERROR RPS\tP50 LATENCY\tP90 LATENCY\tP99 LATENCY\t\n", "WORKLOAD") _ = w.Flush() } func printMetrics(writer io.Writer, wm workloadMetrics) { w := tabwriter.NewWriter(writer, 13, 1, 2, ' ', tabwriter.AlignRight)
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Sat Apr 13 05:23:38 UTC 2024 - 8.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf.cc
return absl::OkStatus(); } } // namespace absl::StatusOr<tensorflow::XlaCompilationResult> LegalizeMlirToHlo( const std::variant<tpu::MlirToHloArgs, tpu::FunctionToHloArgs>& computation, const tpu::TPUCompileMetadataProto& metadata, bool use_tuple_args, llvm::StringRef device_type, std::vector<std::unique_ptr<mlir::Pass>>& custom_legalization_passes,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 00:40:46 UTC 2024 - 6.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir_test.cc
#include "tensorflow/core/lib/monitoring/cell_reader.h" #include "tensorflow/core/protobuf/tpu/compile_metadata.pb.h" #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h" #include "tsl/platform/statusor.h" namespace tensorflow { namespace tf2xla { namespace internal { namespace { using tpu::MlirToHloArgs; using tpu::ShardingAndIndex; using tpu::TPUCompileMetadataProto; static constexpr char kMlirModuleStr[] = R"(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 4.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_tpu_model_to_cpu.cc
#include "tensorflow/compiler/mlir/tensorflow/ir/tf_ops.h" #include "tensorflow/core/tpu/tpu_defs.h" namespace mlir { namespace quant { namespace { #include "tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_tpu_model_to_cpu.inc" // Convert a TPU model to be compatible on CPU by rewriting/removing TPU ops. class ConvertTpuModelToCpuPass
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tpu_validate_inputs.cc
if (!op->hasAttr(TF::kTpuReplicateAttr)) { op->emitOpError("TF2XLA TPU bridge input check: " + errormsg() + "missing _tpu_replicate attr"); return false; } auto opattr = op->getAttr(TF::kTpuReplicateAttr); if (opattr != attr) { op->emitOpError("TF2XLA TPU bridge input check: " + errormsg() + "invalid _tpu_replicate attr.")
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 07 06:51:01 UTC 2024 - 21.5K bytes - Viewed (0)