Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 14 for dequantize (0.16 sec)

  1. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights_test.cc

        const auto float_graph = model_->subgraphs()->Get(subgraph_idx);
        // The output graph should have an extra tensor from the added dequantize
        // op.
        ASSERT_EQ(quantized_graph->tensors()->size(),
                  float_graph->tensors()->size() + 1);
        // Check that a dequantize op exists.
        int32_t dequant_input_idx = -1;
        int32_t dequant_output_idx = -1;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 32.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.td

           (TF_ConstOp (GetI64ScalarElementsAttr<-1>)))),
      (TF_SoftmaxCrossEntropyWithLogitsOp $features, $adjusted_labels)]>;
    
    //===----------------------------------------------------------------------===//
    // Dequantize op patterns.
    //===----------------------------------------------------------------------===//
    
    def DequantizeHalfRange : NativeCodeCall<
      "DequantizeHalfRange(&$_builder, $0)">;
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 04 13:30:42 UTC 2024
    - 24.7K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

      func.return %0 : tensor<8x8x8x8xf32>
    
      // CHECK-LABEL: fakeQuantArgsFalse
      // CHECK: "tfl.quantize"(%arg0) <{qtype = tensor<8x8x8x8x!quant.uniform<u8:f32, 0.0011764706057660721:85>>}>
      // CHECK: %1 = "tfl.dequantize"(%0) : (tensor<8x8x8x8x!quant.uniform<u8:f32, 0.0011764706057660721:85>>) -> tensor<8x8x8x8xf32>
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tf_tfl_passes.cc

        const mlir::TFL::PassConfig& pass_config,
        mlir::OpPassManager* pass_manager) {
      // This pass wraps all the tf.FakeQuant ops in a custom op so they are not
      // folded before being converted to tfl.quantize and tfl.dequantize ops.
      auto wrapped_ops = mlir::TFL::AllTfFakeQuantOps();
      pass_manager->addNestedPass<mlir::func::FuncOp>(
          mlir::TFL::CreateRaiseCustomOpsPass(wrapped_ops));
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 18:45:51 UTC 2024
    - 25.5K bytes
    - Viewed (0)
  5. RELEASE.md

        ([CVE-2022-21728](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-21728))
    *   Fixes a heap OOB access in `Dequantize`
        ([CVE-2022-21726](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-21726))
    *   Fixes an integer overflow in shape inference for `Dequantize`
        ([CVE-2022-21727](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-21727))
    *   Fixes a heap OOB access in `FractionalAvgPoolGrad`
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/quantization/lite/quantize_model_test.cc

      // The original model reshape->custom->custom->squeeze.
      ASSERT_THAT(*float_graph->operators(), SizeIs(4));
      // The resulting model should be:
      // reshape->dequantize->custom->custom->quantize->squeeze.
      ASSERT_THAT(subgraph->operators, SizeIs(6));
      const std::vector<BuiltinOperator> op_codes = {
          BuiltinOperator_RESHAPE,  BuiltinOperator_DEQUANTIZE,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 73.9K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/ir/tfl_ops.td

    // Quantization ops.
    //===----------------------------------------------------------------------===//
    def TFL_DequantizeOp: TFL_Op<"dequantize", [NoMemoryEffect]> {
      let summary = "Dequantize operator";
    
      let description = [{
        Converts quantized array of integers to floating-points according to the
        quantization parameters.
      }];
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 186K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tests/ops.mlir

    func.func @testDequantize(tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x f32> {
    ^bb0(%arg0: tensor<? x !quant.uniform<i8:f32, 0.1>>):
      // CHECK: "tfl.dequantize"(%arg0) : (tensor<?x!quant.uniform<i8:f32, 1.000000e-01>>) -> tensor<?xf32>
      %0 = "tfl.dequantize"(%arg0): (tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x f32>
      func.return %0 : tensor<? x f32>
    }
    
    // CHECK-LABEL: testLogicalNot
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h

    // Stores information about how to quantize a user-specified custom operation.
    // CustomOpInfo contains info of its corresponding CustomOp registered in the
    // CustomOpMap. 'quantizable_input_indices' is used to determine which indices
    // of the CustomOp are quantizable. 'is_weight_only' is used specify whether the
    // custom op is quantized only for storage and dequantized at runtime.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 4.2K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      let summary = "Quantizes then dequantizes a tensor.";
    
      let description = [{
    This is almost identical to QuantizeAndDequantizeV2, except that it returns a
    gradient of 1 for inputs that are within the quantization range, or 0 otherwise.
      }];
    
      let arguments = (ins
        Arg<TF_FloatTensor, [{Tensor to quantize and then dequantize.}]>:$input,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
Back to top