Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 95 for tpu0 (0.2 sec)

  1. tensorflow/compiler/mlir/tensorflow/transforms/tpu_dynamic_layout_pass.cc

      void runOnFunction(
          func::FuncOp func,
          const TF::ResourceAliasAnalysis::Info& resource_alias_analysis);
    
      StringRef getArgument() const final { return "tf-tpu-dynamic-layout-pass"; }
    
      StringRef getDescription() const final {
        return "Inserts TPU layout ops to determine layout at run time.";
      }
    };
    
    // Checks if the input producer op is supported in this transform. Right now, we
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 12.7K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/xla_platform_info.cc

        return absl::OkStatus();
      }
    
      // TFRT-TPU is used if device type is `DEVICE_TPU` and platform_info does not
      // have `xla_device_metadata`. This is used for TFRT-TPU when
      // BuildXlaDeviceCompiler() is called in GetCompilerIr(). Currently only
      // lowering to HLO is needed there and xla::LocalClient doesn't support
      // building the executable for TFRT-TPU and hence, is set to nullptr here.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 17:23:27 UTC 2024
    - 17.4K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/ir/tf_ops.td

      let summary = "Op that compiles a computation in MLIR into a TPU program, and loads and executes it on a TPU device.";
    
      let description = [{
    For the internal use of the TPU compiler.
    
    'static_shapes' are tensors specifying the maximum dimension sizes for the tensors specified in `dynamic_operands`.
    'args' are inputs to the TPU computation.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 04:08:35 UTC 2024
    - 90.5K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/replace_stablehlo_ops_in_main_function_with_xla_call_module_ops.mlir

      // CHECK: %[[IDENTIFY:.*]] = "tf.Identity"(%[[SUBGRAPH_0]]#1) {device = ""} : (tensor<1024x3xf32>) -> tensor<1024x3xf32>
      // CHECK: %[[SUBGRAPH_1:.*]] = "tf.XlaCallModule"() <{Sout = [#tf_type.shape<1024x3>], {{.*}} ["CPU", "TPU"], {{.*}}}> {_entry_function = @_stablehlo_main_1
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 01:09:50 UTC 2024
    - 39.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/stablehlo/cc/weight_only_ptq.cc

                              WeightOnlyPtqComponent::kName, *function_aliases,
                              *ctx, *module));
    
      // Remove the `tpu` tag for exporting because the output quantized model is
      // essentially a CPU model.
      tags.erase("tpu");
    
      py_function_library.SaveExportedModel(
          dst_saved_model_path, post_calibrated_exported_model,
          src_saved_model_path, tags, signature_def_map);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 02:59:01 UTC 2024
    - 5.1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc

      return builder->create<mlir::TF::ConcatOp>(
          location, output_type, concat_dimension_op.getOutput(), inputs);
    }
    
    // For tile sharded inputs to TPU computation, inject split op between the
    // input values and TPU computation so that tiled input values are passed in
    // as inputs to TPU computations. If more than one dimension is sharded, then
    // a tree of connected split ops are added before tf_device.parallel_execute op.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 21:28:13 UTC 2024
    - 34K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/stablehlo/cc/static_range_ptq.cc

                              PostCalibrationComponent::kName, *function_aliases,
                              *ctx, *module));
    
      // Remove the `tpu` tag for exporting because the output quantized model is
      // essentially a CPU model.
      tags.erase("tpu");
    
      py_function_library.SaveExportedModel(
          dst_saved_model_path, post_calibrated_exported_model,
          src_saved_model_path, tags, signature_def_map);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 12:49:45 UTC 2024
    - 6K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/transforms/tpu_annotate_dynamic_shape_inputs.cc

    #include "tensorflow/compiler/mlir/tensorflow/ir/tf_ops.h"
    #include "tensorflow/compiler/mlir/tensorflow/utils/attribute_utils.h"
    #include "xla/mlir_hlo/mhlo/IR/hlo_ops.h"
    
    #define DEBUG_TYPE "tf-tpu-annotate-dynamic-shape-inputs"
    
    namespace mlir {
    namespace TFTPU {
    
    namespace {
    
    #define GEN_PASS_DEF_TPUANNOTATEDYNAMICSHAPEINPUTSPASS
    #include "tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.h.inc"
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tf2xla/internal/utils/BUILD

            "//tensorflow/compiler/tf2xla/kernels:xla_ops",
            "//tensorflow/core:framework",
            "//tensorflow/core:test_main",
            "//tensorflow/core/protobuf/tpu:compile_metadata_proto_cc",
            "//tensorflow/core/tpu/kernels/xla:host_compute_ops",
            "@com_google_absl//absl/status",
            "@com_google_absl//absl/strings:string_view",
            "@llvm-project//mlir:FuncDialect",
            "@llvm-project//mlir:IR",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 13 23:59:33 UTC 2024
    - 2.7K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tfrt/tests/mlrt/rewrite_ifrt_load_variable.mlir

    // RUN: tf-tfrt-opt -split-input-file -tf-mlrt-rewrite-ifrt-load-variable %s | FileCheck %s
    
    // Variable is used by both CPU and TPU
    //
    // CHECK-LABEL: func @serving_default(%arg0: tensor<1x3xf32>) -> tensor<1x1xf32>
    // CHECK-NEXT:    [[HANDLE:%.*]] = "tf.VarHandleOp"()
    // CHECK-NEXT:    [[ARRAYKEY:%.*]], [[FURTURE:%.*]] = "tf_mlrt.tf_ifrt_load_variable"([[HANDLE]])
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 21:35:32 UTC 2024
    - 1.7K bytes
    - Viewed (0)
Back to top