Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 15 of 15 for quantization (0.15 sec)

  1. tensorflow/compiler/mlir/lite/transforms/prepare_tf.cc

    #include "mlir/Transforms/GreedyPatternRewriteDriver.h"  // from @llvm-project
    #include "tensorflow/compiler/mlir/lite/ir/tfl_ops.h"
    #include "tensorflow/compiler/mlir/lite/quantization/ir/QuantOps.h"
    #include "tensorflow/compiler/mlir/lite/stablehlo/transforms/passes.h"
    #include "tensorflow/compiler/mlir/lite/transforms/dilated_conv.h"
    #include "tensorflow/compiler/mlir/lite/transforms/passes.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 21:49:50 UTC 2024
    - 64.6K bytes
    - Viewed (0)
  2. .bazelrc

    # CROSS-COMPILE ARM64 PYCPP
    build:cross_compile_linux_arm64_pycpp_test --config=linux_arm64_pycpp_test
    # Tests that fail only when cross-compiled
    build:cross_compile_linux_arm64_pycpp_test -//tensorflow/compiler/mlir/quantization/stablehlo:convert_tf_quant_to_mhlo_int_test
    # MACOS ARM64 PYCPP
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 17:12:54 UTC 2024
    - 52.9K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/optimize_patterns.td

         TFL_AF_None),
        $x,
        TFL_AF_None),
      (TFL_HardSwishOp $x)>;
    
    // Matching HardSwish with extra FakeQuant. These FakeQuant ops were due to
    // incorrect placement in the quantization aware training.
    def MatchHardSwishQuantized : Pat<
      (TFL_MulOp (TFL_DequantizeOp (TFL_QuantizeOp
        (TFL_MulOp
         $x, (TFL_DequantizeOp (TFL_QuantizeOp (TFL_AddOp
              $x,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 16 20:31:41 UTC 2024
    - 66.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc

        auto quant_to_float =
            rewriter.create<DivOp>(op.getLoc(), float_diff, quant_diff);
    
        auto float_to_quant =
            rewriter.create<DivOp>(op.getLoc(), quant_diff, float_diff);
    
        // During quantization, the quantized min/max values may not line up
        // perfectly with the specified min/max. Nudge them into the right range.
        auto min_scaled =
            rewriter.create<DivOp>(op.getLoc(), float_min, quant_to_float);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 74.9K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composit_functions_debugging.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-insert-quantized-functions -quant-quantize-composite-functions='target-opset=XLA enable-per-channel-quantization=true' | FileCheck --check-prefix=XLA_PerChannel %s
    
    module {
      func.func @conv_with_int_per_layer(%arg0: tensor<*xf32>) -> tensor<*xf32> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Nov 06 01:23:21 UTC 2023
    - 80.5K bytes
    - Viewed (0)
Back to top