Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 684 for Advice (0.11 sec)

  1. tensorflow/compiler/mlir/tensorflow/transforms/colocate_tpu_copy_with_dynamic_shape.cc

          auto device = op->getAttrOfType<StringAttr>(kDevice);
          for (auto *operand : operands)
            propagateIfChanged(operand, operand->SetDevice(device));
        } else {
          // Propagate device through other ops. These ops might have their
          // own device annotation, but that's fine. We only care about
          // where the TPUExecute ops live.
          StringAttr device;
          for (const Device *d : results) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Aug 23 00:30:27 UTC 2023
    - 5.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/experimental/tac/transforms/target_annotation.cc

          *this, "device-specs",
          llvm::cl::desc(
              "comma separated list of device specs, like CPU, GPU, Hexagon."),
          llvm::cl::ZeroOrMore};
    
      void getDependentDialects(mlir::DialectRegistry& registry) const override {
        if (!module_) {
          for (const auto& device : device_specs_flag_) {
            auto* hardware = this->GetTargetHardware(device);
            if (hardware == nullptr) continue;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 19 19:32:06 UTC 2023
    - 5.9K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/experimental/tac/common/targets.h

      return name;
    }
    
    // Get the target annotation form the op.
    inline std::optional<std::string> GetTargetAnnotation(Operation* op) {
      auto device = op->getAttrOfType<StringAttr>(kDevice);
      if (device == nullptr || device.getValue().empty()) return std::nullopt;
    
      return GetCanonicalHardwareName(device.getValue().str());
    }
    
    // Get inference type attribute from the operation if available.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 06 03:08:33 UTC 2023
    - 4.7K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/tests/graphdef2mlir/graph-device-retval.pbtxt

    A. Unique TensorFlower <******@****.***> 1605121757 -0800
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Nov 11 19:14:04 UTC 2020
    - 1.5K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tf2xla/internal/passes/clustering_passes.h

    // Creates a pass that extracts outside compilation (Host ops inside device
    // cluster) at head/tail of Device cluster to run before/after XLA computation.
    std::unique_ptr<mlir::OperationPass<mlir::ModuleOp>>
    CreateExtractHeadTailOutsideCompilationPass();
    
    // Creates a pass that extract outside compilation (Host ops inside cevice
    // cluster) ops to a separate parallel_execute region to run on CPU.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 02:01:13 UTC 2024
    - 3.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/experimental/tac/execution_metadata_exporter.cc

        return std::nullopt;
    
      if (!HasValidHardwareTarget(op)) return std::nullopt;
    
      auto device = op->getAttrOfType<mlir::StringAttr>(mlir::TFL::tac::kDevice);
      if (device == nullptr) return std::nullopt;
    
      llvm::StringRef device_name_str = device.getValue();
      return device_name_str.str();
    }
    
    std::optional<std::vector<float>> GetPerDeviceCosts(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 06:11:34 UTC 2024
    - 7.5K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/g3doc/space_to_depth.md

        needed to apply space to depth optimization for a model that needs to load a
        pre-train checkpoint. This transform can be done on the host or TPU device
        based on the cost. As the size of the kernel is relatively small, this won't
        add additional cost to TPU device time. Below is the logic to transform the
        kernel of shape [7, 7, 3, 64] to [4, 4, 12, 84].
    
        ```python
        conv0 = tf.compat.v1.layers.Conv2D(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Oct 24 02:51:43 UTC 2020
    - 8.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/jit/xla_ops_on_regular_devices.cc

                              XlaCompileOnDemandOp);                               \
      REGISTER_KERNEL_BUILDER(Name("XlaSvd").Device(DEVICE),                       \
                              XlaCompileOnDemandOp);                               \
      REGISTER_KERNEL_BUILDER(Name("XlaDot").Device(DEVICE),                       \
                              XlaCompileOnDemandOp);                               \
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Aug 19 19:55:14 UTC 2022
    - 8.8K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/utils/device_util.cc

                                           mlir::Builder* builder) {
      // Parse GPU device compute capability from physical device description.
      static auto* r = new llvm::Regex("compute capability: ([0-9]+)\\.([0-9]+)");
    
      llvm::SmallVector<llvm::StringRef, 3> cc;
      if (r->match(device.attributes().physical_device_desc(), &cc)) {
        return mlir::TF::GpuDeviceMetadata::get(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/tests/convert_to_legacy_compile_and_replicate_attributes.mlir

        %outputs, %control = tf_executor.island wraps "tf.GuaranteeConst"(%arg1) {T = f32, device = ""} : (tensor<f32>) -> tensor<f32>
        %outputs_0, %control_1 = tf_executor.island wraps "tf.GuaranteeConst"(%arg2) {T = f32, device = ""} : (tensor<f32>) -> tensor<f32>
        %control_2 = tf_executor.island wraps "tf.NoOp"() {_pivot_for_cluster = "cluster", device = ""} : () -> ()
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 22:03:30 UTC 2024
    - 6.1K bytes
    - Viewed (0)
Back to top