- Sort Score
- Result 10 results
- Languages All
Results 101 - 110 of 178 for dequantize (0.2 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/tests/propagate_quantize_type.mlir
// CHECK: %[[IDENTITY:.*]] = "tf.Identity"(%cst_0) : (tensor<200x100x300xi8>) -> tensor<200x100x300xi8> // CHECK: %[[DEQUANTIZED:.*]] = "tf.PartitionedCall"(%[[IDENTITY]]) <{config = "", config_proto = "", executor_type = "", f = @composite_dequantize_uniform}> : (tensor<200x100x300xi8>) -> tensor<200x100x300xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 6.6K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass.cc
"DataFormatDimMap", "DataFormatVecPermute", "DepthToSpace", "DepthwiseConv2dNative", "DepthwiseConv2dNativeBackpropFilter", "DepthwiseConv2dNativeBackpropInput", "Dequantize", "Diag", "DynamicInfeedEnqueueTupleOp", "DynamicInfeedDequeueTupleOp", "DynamicStitch", "DynamicPartition", "Einsum", "EmptyTensorList",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 12:19:41 UTC 2024 - 85.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/ir/tfl_ops.td
// Quantization ops. //===----------------------------------------------------------------------===// def TFL_DequantizeOp: TFL_Op<"dequantize", [NoMemoryEffect]> { let summary = "Dequantize operator"; let description = [{ Converts quantized array of integers to floating-points according to the quantization parameters. }];
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 19:09:08 UTC 2024 - 186K bytes - Viewed (0) -
tensorflow/compiler/aot/quantize.h
Jake Harmon <******@****.***> 1694027275 -0700
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 1.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/prepare_quantize.cc
// ranges. bool SetInputNodesQuantizationParams(func::FuncOp func); // The function might contain more stats ops than required, and it will // introduce requantize if the calibration stats have conflicts. This method // tries to remove all the redundant stats ops. bool RemoveRedundantStats(func::FuncOp func);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_quantize.cc
// Whether the func contains Quantize ops. This is used to determine whether // to use the quantization parameters from the fixed output range property. bool ContainsQuantizeOps(func::FuncOp func); QuantizationSpecs quant_specs_; Option<bool> enable_post_training_quantize_{ *this, "post-training-quantize", llvm::cl::init(false),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/ops.mlir
func.func @testDequantize(tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x f32> { ^bb0(%arg0: tensor<? x !quant.uniform<i8:f32, 0.1>>): // CHECK: "tfl.dequantize"(%arg0) : (tensor<?x!quant.uniform<i8:f32, 1.000000e-01>>) -> tensor<?xf32> %0 = "tfl.dequantize"(%arg0): (tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x f32> func.return %0 : tensor<? x f32> } // CHECK-LABEL: testLogicalNot
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 19:09:08 UTC 2024 - 189.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h
// Stores information about how to quantize a user-specified custom operation. // CustomOpInfo contains info of its corresponding CustomOp registered in the // CustomOpMap. 'quantizable_input_indices' is used to determine which indices // of the CustomOp are quantizable. 'is_weight_only' is used specify whether the // custom op is quantized only for storage and dequantized at runtime.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 4.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/quantization_context.h
struct RequantizeState { // Sometimes, we have to "requantize" the quantization result to satisfy all // the constraints. The "requantize" can happen either on the input or output // of the quantization result. enum RequantizePosition { NO_REQUANTIZE, ON_INPUT, ON_OUTPUT } pos = NO_REQUANTIZE; // Quantization parameters will be used to add the requantize ops. QuantParams params; };
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 08 01:38:03 UTC 2024 - 9.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
let summary = "Quantizes then dequantizes a tensor."; let description = [{ This is almost identical to QuantizeAndDequantizeV2, except that it returns a gradient of 1 for inputs that are within the quantization range, or 0 otherwise. }]; let arguments = (ins Arg<TF_FloatTensor, [{Tensor to quantize and then dequantize.}]>:$input,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0)