Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 52 for opset (0.26 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/passes/insert_quantized_functions.cc

      Option<OpSet> op_set_{
          *this, "target-opset", llvm::cl::init(OpSet::TF),
          llvm::cl::desc("Choose target opset."),
          llvm::cl::values(
              clEnumValN(OpSet::TF, "TF",
                         "Uses TF ops that mimic quantization behavior"),
              clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"),
              clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 22 05:52:39 UTC 2024
    - 8.7K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions_drq.cc

     private:
      Option<OpSet> target_opset_{
          *this, "target-opset", llvm::cl::init(OpSet::TF),
          llvm::cl::desc("Choose target opset."),
          llvm::cl::values(
              clEnumValN(OpSet::TF, "TF",
                         "Uses TF ops that mimic quantization behavior"),
              clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"),
              clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.5K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/quantization_options.proto

      QuantizationMethod quantization_method = 6;
    
      reserved 1 to 4;
    }
    
    // List of supported opsets to deploy the quantized model.
    // The quantized model contains different set of ops depending on the opset.
    // NEXT ID: 5
    enum OpSet {
      OP_SET_UNSPECIFIED = 0;  // go/do-include-enum-unspecified
      // Uses TF ops that mimic quantization behavior. Used when the corresponding
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 19 06:31:19 UTC 2024
    - 9.2K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/tests/preprocess_op_weight_only.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-preprocess-op='target-opset=XLA quantization-method=weight_only enable-per-channel-quantization=false' | FileCheck --check-prefix PerTensor %s
    // RUN: tf-quant-opt %s -split-input-file -quant-preprocess-op='target-opset=XLA quantization-method=weight_only enable-per-channel-quantization=true' | FileCheck --check-prefix PerChannel %s
    
    module {
      // For XLA weight-only per-channel depthwise convolution, tensor shape should have
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 4.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions_drq.mlir

    // RUN: tf-quant-opt %s -quant-insert-quantized-functions='quantization-method=drq' | FileCheck %s
    // RUN: tf-quant-opt %s -quant-insert-quantized-functions='quantization-method=drq target-opset=UNIFORM_QUANTIZED' | FileCheck --check-prefix=UQ-CHECK %s
    
    // Empty module
    module {
      func.func @simple_fn(%arg0: tensor<*xf32>) -> tensor<*xf32> {
        func.return %arg0 : tensor<*xf32>
      }
    }
    
    // CHECK-NOT: func private @internal_calculate_quant_params
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Dec 01 12:06:54 UTC 2022
    - 1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/tests/fake_quant_e2e_xla.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-convert-fake-quant-to-qdq -quant-lift-quantizable-spots-as-functions='target-opset=XLA' -quant-insert-quantized-functions -quant-quantize-composite-functions='target-opset=XLA' -symbol-dce -inline -tf-shape-inference -canonicalize -quant-replace-cast-hacks-with-tf-xla-ops -cse -quant-optimize | FileCheck %s
    
    module attributes {tf.versions = {bad_consumers = [], min_consumer = 12 : i32, producer = 1219 : i32}, tf_saved_model.semantics} {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 7.2K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions_weight_only.mlir

    // RUN: tf-quant-opt %s -quant-insert-quantized-functions='quantization-method=weight_only target-opset=XLA' | FileCheck %s
    
    // Empty module
    module {
      func.func @simple_fn(%arg0: tensor<*xf32>) -> tensor<*xf32> {
        func.return %arg0 : tensor<*xf32>
      }
    }
    
    // CHECK-NOT: func private @internal_dequantize_f32
    // CHECK-NOT: func private @internal_conv3d_fn
    // CHECK-NOT: func private @internal_batch_matmul_fn
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 16 03:34:36 UTC 2023
    - 843 bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions.mlir

    // RUN: tf-quant-opt %s -quant-insert-quantized-functions | FileCheck %s
    // RUN: tf-quant-opt %s -quant-insert-quantized-functions='quantization-method=ptq target-opset=UNIFORM_QUANTIZED' --mlir-print-ir-after-all | FileCheck --check-prefix=UQ-CHECK %s
    
    // Empty module
    module {
      func.func @simple_fn(%arg0: tensor<*xf32>) -> tensor<*xf32> {
        func.return %arg0 : tensor<*xf32>
      }
    }
    
    // CHECK-NOT: func private @internal_rescale_fn
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 3.3K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/tests/lift_quantizable_spots_as_functions_xla_selective_quantization.mlir

    // RUN: tf-quant-opt %s -split-input-file -mlir-print-debuginfo -mlir-print-local-scope -quant-add-quantization-unit-loc -inline -quant-prepare-lifting -quant-lift-quantizable-spots-as-functions='target-opset=XLA' | FileCheck %s
    
    // This file test the selective quantiation feature in TF Quantizer. In the test
    // config, the op named "test_opt_out" will not be quantized.
    
    module attributes {tf.versions = {bad_consumers = [], min_consumer = 12 : i32, producer = 1269 : i32}} {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 6.8K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/tests/lift_quantizable_spots_as_functions_xla.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-lift-quantizable-spots-as-functions='target-opset=XLA' | FileCheck %s
    
    module attributes {tf.versions = {bad_consumers = [], min_consumer = 12 : i32, producer = 1269 : i32}, tf_saved_model.semantics} {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 8.3K bytes
    - Viewed (0)
Back to top