- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 14 for dequantize (0.16 sec)
-
tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights_test.cc
const auto float_graph = model_->subgraphs()->Get(subgraph_idx); // The output graph should have an extra tensor from the added dequantize // op. ASSERT_EQ(quantized_graph->tensors()->size(), float_graph->tensors()->size() + 1); // Check that a dequantize op exists. int32_t dequant_input_idx = -1; int32_t dequant_output_idx = -1;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 32.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.td
(TF_ConstOp (GetI64ScalarElementsAttr<-1>)))), (TF_SoftmaxCrossEntropyWithLogitsOp $features, $adjusted_labels)]>; //===----------------------------------------------------------------------===// // Dequantize op patterns. //===----------------------------------------------------------------------===// def DequantizeHalfRange : NativeCodeCall< "DequantizeHalfRange(&$_builder, $0)">;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 04 13:30:42 UTC 2024 - 24.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir
func.return %0 : tensor<8x8x8x8xf32> // CHECK-LABEL: fakeQuantArgsFalse // CHECK: "tfl.quantize"(%arg0) <{qtype = tensor<8x8x8x8x!quant.uniform<u8:f32, 0.0011764706057660721:85>>}> // CHECK: %1 = "tfl.dequantize"(%0) : (tensor<8x8x8x8x!quant.uniform<u8:f32, 0.0011764706057660721:85>>) -> tensor<8x8x8x8xf32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 05 01:54:33 UTC 2024 - 153.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tf_tfl_passes.cc
const mlir::TFL::PassConfig& pass_config, mlir::OpPassManager* pass_manager) { // This pass wraps all the tf.FakeQuant ops in a custom op so they are not // folded before being converted to tfl.quantize and tfl.dequantize ops. auto wrapped_ops = mlir::TFL::AllTfFakeQuantOps(); pass_manager->addNestedPass<mlir::func::FuncOp>( mlir::TFL::CreateRaiseCustomOpsPass(wrapped_ops));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 18:45:51 UTC 2024 - 25.5K bytes - Viewed (0) -
RELEASE.md
([CVE-2022-21728](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-21728)) * Fixes a heap OOB access in `Dequantize` ([CVE-2022-21726](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-21726)) * Fixes an integer overflow in shape inference for `Dequantize` ([CVE-2022-21727](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-21727)) * Fixes a heap OOB access in `FractionalAvgPoolGrad`
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 730.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_model_test.cc
// The original model reshape->custom->custom->squeeze. ASSERT_THAT(*float_graph->operators(), SizeIs(4)); // The resulting model should be: // reshape->dequantize->custom->custom->quantize->squeeze. ASSERT_THAT(subgraph->operators, SizeIs(6)); const std::vector<BuiltinOperator> op_codes = { BuiltinOperator_RESHAPE, BuiltinOperator_DEQUANTIZE,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 73.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/ir/tfl_ops.td
// Quantization ops. //===----------------------------------------------------------------------===// def TFL_DequantizeOp: TFL_Op<"dequantize", [NoMemoryEffect]> { let summary = "Dequantize operator"; let description = [{ Converts quantized array of integers to floating-points according to the quantization parameters. }];
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 19:09:08 UTC 2024 - 186K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/ops.mlir
func.func @testDequantize(tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x f32> { ^bb0(%arg0: tensor<? x !quant.uniform<i8:f32, 0.1>>): // CHECK: "tfl.dequantize"(%arg0) : (tensor<?x!quant.uniform<i8:f32, 1.000000e-01>>) -> tensor<?xf32> %0 = "tfl.dequantize"(%arg0): (tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x f32> func.return %0 : tensor<? x f32> } // CHECK-LABEL: testLogicalNot
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 19:09:08 UTC 2024 - 189.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h
// Stores information about how to quantize a user-specified custom operation. // CustomOpInfo contains info of its corresponding CustomOp registered in the // CustomOpMap. 'quantizable_input_indices' is used to determine which indices // of the CustomOp are quantizable. 'is_weight_only' is used specify whether the // custom op is quantized only for storage and dequantized at runtime.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 4.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
let summary = "Quantizes then dequantizes a tensor."; let description = [{ This is almost identical to QuantizeAndDequantizeV2, except that it returns a gradient of 1 for inputs that are within the quantization range, or 0 otherwise. }]; let arguments = (ins Arg<TF_FloatTensor, [{Tensor to quantize and then dequantize.}]>:$input,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0)