Search Options

Results per page
Sort
Preferred Languages
Advance

Results 101 - 110 of 178 for dequantize (0.2 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/propagate_quantize_type.mlir

    // CHECK: %[[IDENTITY:.*]] = "tf.Identity"(%cst_0) : (tensor<200x100x300xi8>) -> tensor<200x100x300xi8>
    // CHECK: %[[DEQUANTIZED:.*]] = "tf.PartitionedCall"(%[[IDENTITY]]) <{config = "", config_proto = "", executor_type = "", f = @composite_dequantize_uniform}> : (tensor<200x100x300xi8>) -> tensor<200x100x300xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 6.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/mark_for_compilation_pass.cc

    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 12:19:41 UTC 2024
    - 85.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/ir/tfl_ops.td

    // Quantization ops.
    //===----------------------------------------------------------------------===//
    def TFL_DequantizeOp: TFL_Op<"dequantize", [NoMemoryEffect]> {
      let summary = "Dequantize operator";
    
      let description = [{
        Converts quantized array of integers to floating-points according to the
        quantization parameters.
      }];
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 186K bytes
    - Viewed (0)
  4. tensorflow/compiler/aot/quantize.h

    Jake Harmon <******@****.***> 1694027275 -0700
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 1.4K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/transforms/prepare_quantize.cc

      // ranges.
      bool SetInputNodesQuantizationParams(func::FuncOp func);
    
      // The function might contain more stats ops than required, and it will
      // introduce requantize if the calibration stats have conflicts. This method
      // tries to remove all the redundant stats ops.
      bool RemoveRedundantStats(func::FuncOp func);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 17.6K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_quantize.cc

      // Whether the func contains Quantize ops. This is used to determine whether
      // to use the quantization parameters from the fixed output range property.
      bool ContainsQuantizeOps(func::FuncOp func);
    
      QuantizationSpecs quant_specs_;
    
      Option<bool> enable_post_training_quantize_{
          *this, "post-training-quantize", llvm::cl::init(false),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 17.2K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/ops.mlir

    func.func @testDequantize(tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x f32> {
    ^bb0(%arg0: tensor<? x !quant.uniform<i8:f32, 0.1>>):
      // CHECK: "tfl.dequantize"(%arg0) : (tensor<?x!quant.uniform<i8:f32, 1.000000e-01>>) -> tensor<?xf32>
      %0 = "tfl.dequantize"(%arg0): (tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x f32>
      func.return %0 : tensor<? x f32>
    }
    
    // CHECK-LABEL: testLogicalNot
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h

    // Stores information about how to quantize a user-specified custom operation.
    // CustomOpInfo contains info of its corresponding CustomOp registered in the
    // CustomOpMap. 'quantizable_input_indices' is used to determine which indices
    // of the CustomOp are quantizable. 'is_weight_only' is used specify whether the
    // custom op is quantized only for storage and dequantized at runtime.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 4.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/quantization/quantization_context.h

    struct RequantizeState {
      // Sometimes, we have to "requantize" the quantization result to satisfy all
      // the constraints. The "requantize" can happen either on the input or output
      // of the quantization result.
      enum RequantizePosition {
        NO_REQUANTIZE,
        ON_INPUT,
        ON_OUTPUT
      } pos = NO_REQUANTIZE;
    
      // Quantization parameters will be used to add the requantize ops.
      QuantParams params;
    };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 08 01:38:03 UTC 2024
    - 9.9K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      let summary = "Quantizes then dequantizes a tensor.";
    
      let description = [{
    This is almost identical to QuantizeAndDequantizeV2, except that it returns a
    gradient of 1 for inputs that are within the quantization range, or 0 otherwise.
      }];
    
      let arguments = (ins
        Arg<TF_FloatTensor, [{Tensor to quantize and then dequantize.}]>:$input,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
Back to top