Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 30 of 148 for tpu0 (0.04 sec)

  1. tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo.h

    #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h"
    #include "tsl/platform/statusor.h"
    
    namespace tensorflow {
    namespace tf2xla {
    namespace internal {
    
    // Legalize the given MLIR module to XLA HLO using a combination of the MLIR
    // Bridge and XlaBuilder
    absl::StatusOr<XlaCompilationResult> LegalizeTfToHlo(
        const tpu::MlirToHloArgs& computation,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 20:29:34 UTC 2024
    - 2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir.cc

    #include "tensorflow/core/platform/profile_utils/cpu_utils.h"
    #include "tensorflow/core/platform/types.h"
    #include "tensorflow/core/protobuf/tpu/compile_metadata.pb.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h"
    #include "tensorflow/core/tpu/tpu_compile.h"
    #include "tsl/platform/error_logging.h"
    #include "tsl/platform/errors.h"
    #include "tsl/platform/statusor.h"
    
    namespace tensorflow {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 20:29:34 UTC 2024
    - 6.1K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tfrt/function/function.h

      // If true, use ServingCoreSelector to pick TPU core. Otherwise, obtain core
      // location from assigned device name.
      // Currently we don't use core_selector for training use cases.
      bool tpu_use_core_selector = false;
    
      // If true, use BundledTransferToTpuOp to transfer variables and input tensors
      // to TPU.
      bool tpu_use_bundled_transfer = false;
    
      // If true, lower an TF op that's placed on TPU device to be executed with
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 17 04:50:20 UTC 2024
    - 3.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/transforms/set_tpu_infeed_layout.cc

    #include "xla/shape.h"
    #include "xla/stream_executor/tpu/c_api_conversions.h"
    #include "xla/stream_executor/tpu/tpu_api.h"
    #include "xla/translate/mhlo_to_hlo/type_to_shape.h"
    
    namespace mlir {
    
    static FailureOr<std::vector<int64_t>> GetTPUInfeedLayoutFromAPI(
        RankedTensorType t) {
      // Call the TPU API to determine the right infeed layout. Note that
      // this can fail if we're not running on a TPU-enabled node.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.1K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf.h

    //  client - The Xla Compilation client.
    absl::StatusOr<tensorflow::XlaCompilationResult> LegalizeMlirToHlo(
        const std::variant<tpu::MlirToHloArgs, tpu::FunctionToHloArgs>& computation,
        const tpu::TPUCompileMetadataProto& metadata, bool use_tuple_args,
        llvm::StringRef device_type,
        std::vector<std::unique_ptr<mlir::Pass>>& custom_legalization_passes,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 13 07:32:57 UTC 2024
    - 2.9K bytes
    - Viewed (0)
  6. istioctl/pkg/metrics/metrics.go

    	return time.Duration(val) * time.Millisecond
    }
    
    func printHeader(writer io.Writer) {
    	w := tabwriter.NewWriter(writer, 13, 1, 2, ' ', tabwriter.AlignRight)
    	_, _ = fmt.Fprintf(w, "%40s\tTOTAL RPS\tERROR RPS\tP50 LATENCY\tP90 LATENCY\tP99 LATENCY\t\n", "WORKLOAD")
    	_ = w.Flush()
    }
    
    func printMetrics(writer io.Writer, wm workloadMetrics) {
    	w := tabwriter.NewWriter(writer, 13, 1, 2, ' ', tabwriter.AlignRight)
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Sat Apr 13 05:23:38 UTC 2024
    - 8.4K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf.cc

      return absl::OkStatus();
    }
    
    }  // namespace
    
    absl::StatusOr<tensorflow::XlaCompilationResult> LegalizeMlirToHlo(
        const std::variant<tpu::MlirToHloArgs, tpu::FunctionToHloArgs>& computation,
        const tpu::TPUCompileMetadataProto& metadata, bool use_tuple_args,
        llvm::StringRef device_type,
        std::vector<std::unique_ptr<mlir::Pass>>& custom_legalization_passes,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 29 00:40:46 UTC 2024
    - 6.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir_test.cc

    #include "tensorflow/core/lib/monitoring/cell_reader.h"
    #include "tensorflow/core/protobuf/tpu/compile_metadata.pb.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h"
    #include "tsl/platform/statusor.h"
    
    namespace tensorflow {
    namespace tf2xla {
    namespace internal {
    namespace {
    
    using tpu::MlirToHloArgs;
    using tpu::ShardingAndIndex;
    using tpu::TPUCompileMetadataProto;
    
    static constexpr char kMlirModuleStr[] = R"(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 20:29:34 UTC 2024
    - 4.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_tpu_model_to_cpu.cc

    #include "tensorflow/compiler/mlir/tensorflow/ir/tf_ops.h"
    #include "tensorflow/core/tpu/tpu_defs.h"
    
    namespace mlir {
    namespace quant {
    namespace {
    
    #include "tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_tpu_model_to_cpu.inc"
    
    // Convert a TPU model to be compatible on CPU by rewriting/removing TPU ops.
    class ConvertTpuModelToCpuPass
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/transforms/tpu_validate_inputs.cc

      if (!op->hasAttr(TF::kTpuReplicateAttr)) {
        op->emitOpError("TF2XLA TPU bridge input check: " + errormsg() +
                        "missing _tpu_replicate attr");
        return false;
      }
      auto opattr = op->getAttr(TF::kTpuReplicateAttr);
      if (opattr != attr) {
        op->emitOpError("TF2XLA TPU bridge input check: " + errormsg() +
                        "invalid _tpu_replicate attr.")
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 07 06:51:01 UTC 2024
    - 21.5K bytes
    - Viewed (0)
Back to top