Search Options

Results per page
Sort
Preferred Languages
Advance

Results 81 - 90 of 120 for Quantile (0.17 sec)

  1. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.cc

        // asymmetric range. For a state tensor, assigning correct quantization
        // parameters is sufficient, and for constants with asymmetric range it's
        // not correctly quantized by legacy quantizer so call the new Quantize.
        return Quantize(real_value, tensor_type);
      } else if (width == 16) {
        if (const auto uniform_type = dyn_cast<UniformQuantizedType>(q_type)) {
          const auto quantized_values =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 02:10:16 UTC 2024
    - 43.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/transforms/prepare_patterns.td

    class UsedBy<string op> : Constraint<
      CPred<"llvm::isa<mlir::TFL::" # op # "Op>(*$0.getUsers().begin())">>;
    
    // When the op is passing-through, the output types of the quantized ops need
    // to be updated as well. Since the quantize op manages its own type by the
    // "qtype" attribute, we should update the type shape in this attribute.
    def ReorderTransposeDequantQuant :
          Pat<(TF_TransposeOp:$old_value
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 00:40:15 UTC 2024
    - 10.5K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/modify_io_nodes.cc

            returned_type = quant::ConvertSignedQuantizedToUnsigned(
                dequantize_input.getType(), dequantize_op.getLoc());
            // replace the dequantize op by a quantize op
            TypeAttr type_attr = TypeAttr::get(returned_type);
            auto quantize_op = builder.create<QuantizeOp>(
                dequantize_op.getLoc(), returned_type, dequantize_input, type_attr);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.9K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/passes/quantization_patterns.cc

        // `stablehlo.convolution` assumes the following format:
        // [b, 0, 1, f]x[0, 1, i, o]->[b, 0, 1, f]
        // `stablehlo.dot_general` can take various formats. We only per-channel
        // quantize non-batch ops.
        // `stablehlo.dot_general` legalizable to `tfl.fully_connected` has a
        // filter rank of 2 with the last dimension as the channel dimension.
        const int64_t quantization_dimension =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 06:04:36 UTC 2024
    - 41.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tf_to_tfl_flatbuffer.cc

              &q_builder, input_model, quantized_type, use_updated_hybrid_scheme,
              ::tflite::optimize::QuantizerType::OLD_QUANTIZER) != kTfLiteOk) {
        return absl::InvalidArgumentError(
            "Quantize weights transformation failed.");
      }
      const uint8_t* q_buffer = q_builder.GetBufferPointer();
      *result =
          std::string(reinterpret_cast<const char*>(q_buffer), q_builder.GetSize());
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 23.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/experimental/tac/tests/get-alternative-subgraph.mlir

    // CHECK:           %[[VAL_20:.*]] = "tfl.quantize"(%[[VAL_19]]) <{qtype = tensor<1x384x128x!quant.uniform<i8:f32, 3.000000e-01:-3>>}> {tac.device = "GPU", tac.inference_type = "FLOAT"} : (tensor<1x384x128xf32>) -> tensor<1x384x128x!quant.uniform<i8:f32, 3.000000e-01:-3>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 20.1K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/optimize_batch_matmul.mlir

    // CHECK-NOT: "tfl.batch_matmul"
    func.func @Batchmatmul2FullyconnectedQDQ(%arg0: tensor<4x128x2xf32>, %arg1: tensor<2x1xf32>) -> (tensor<4x128x1xf32>) {
      %0 = arith.constant dense<[[1.0], [2.0]]> : tensor<2x1xf32>
      %1 = "tfl.quantize"(%0) {qtype = tensor<2x1x!quant.uniform<i8:f32, 0.024986599940879671:92>>} : (tensor<2x1xf32>) -> tensor<2x1x!quant.uniform<i8:f32, 0.024986599940879671:92>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 9K bytes
    - Viewed (0)
  8. tensorflow/compiler/aot/compile.cc

    #include "absl/base/call_once.h"
    #include "llvm-c/Target.h"
    #include "llvm/Support/ManagedStatic.h"
    #include "tensorflow/compiler/aot/codegen.h"
    #include "tensorflow/compiler/aot/flags.h"
    #include "tensorflow/compiler/aot/quantize.h"
    #include "tensorflow/compiler/tf2xla/tf2xla.h"
    #include "tensorflow/compiler/tf2xla/tf2xla_util.h"
    #include "xla/client/client_library.h"
    #include "xla/client/compile_only_client.h"
    #include "xla/client/xla_computation.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 15 08:28:57 UTC 2024
    - 11.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_xla.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-lift-quantizable-spots-as-functions -quant-quantize='target-opset=XLA' -verify-each=false | FileCheck %s
    
    func.func private @conv(%input: tensor<1x3x4x3xf32> {tf._user_specified_name = "input_tensor"}) -> tensor<*xf32> attributes {tf._construction_context = "kEagerRuntime", tf._input_shapes = [#tf_type.shape<1x3x4x3>]} {
      %weight = arith.constant dense_resource<__elided__> : tensor<2x3x3x2xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 19:32:28 UTC 2024
    - 11.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/quantization/lite/quantize_model_test.cc

        readonly_model_ = input_model_->GetModel();
        model_ = UnPackFlatBufferModel(*readonly_model_);
      }
    };
    
    TEST_F(QuantizeLSTM2Test, VerifyLSTM) {
      // Quantize model.
      auto status = QuantizeModelAllOperators(
          &model_, TensorType_FLOAT32, TensorType_FLOAT32,
          /*allow_float=*/false, TensorType_INT8, output_buffer_);
      ASSERT_THAT(status, Eq(kTfLiteOk));
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 73.9K bytes
    - Viewed (0)
Back to top