Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 203 for biases (0.13 sec)

  1. tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_per_channel_4bit.pbtxt

          }
        }
      }
    }
    node {
      name: "BoxPredictor_4/ClassPredictor/biases/read"
      op: "Identity"
      input: "BoxPredictor_4/ClassPredictor/biases"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
      attr {
        key: "_class"
        value {
          list {
            s: "loc:@BoxPredictor_4/ClassPredictor/biases"
          }
        }
      }
    }
    node {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 18.1K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/transforms/default_quant_params.cc

     private:
      // Whether the value is used as a bias input of another op. Here we assume
      // bias is used immediately by the user. This assumption is always correct
      // after constant folding.
      bool UsedAsBias(Value value) {
        for (auto &use : value.getUses()) {
          auto biases = TFL::GetOpQuantSpec(use.getOwner())->biases_params;
          if (biases.find(use.getOperandNumber()) != biases.end()) return true;
        }
        return false;
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 9.4K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_per_channel.pbtxt

          }
        }
      }
    }
    node {
      name: "BoxPredictor_4/ClassPredictor/biases/read"
      op: "Identity"
      input: "BoxPredictor_4/ClassPredictor/biases"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
      attr {
        key: "_class"
        value {
          list {
            s: "loc:@BoxPredictor_4/ClassPredictor/biases"
          }
        }
      }
    }
    node {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 18.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_without_identity.pbtxt

          }
        }
      }
    }
    node {
      name: "BoxPredictor_4/ClassPredictor/biases/read"
      op: "Identity"
      input: "BoxPredictor_4/ClassPredictor/biases"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
      attr {
        key: "_class"
        value {
          list {
            s: "loc:@BoxPredictor_4/ClassPredictor/biases"
          }
        }
      }
    }
    node {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 13.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_without_identity_4bit.pbtxt

          }
        }
      }
    }
    node {
      name: "BoxPredictor_4/ClassPredictor/biases/read"
      op: "Identity"
      input: "BoxPredictor_4/ClassPredictor/biases"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
      attr {
        key: "_class"
        value {
          list {
            s: "loc:@BoxPredictor_4/ClassPredictor/biases"
          }
        }
      }
    }
    node {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 13.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_driver.cc

              GetQuantScaleSpec(user);
          const BiasParamsMap biases = spec->biases_params;
    
          // The quantization parameters of a `weight` shouldn't be determined by
          // other values. So any constants which are not bias, an operand of an
          // op with same scale requirements, and haven't been quantized are
          // weights.
          if (!biases.contains(operand_num) &&
              !scale_spec->has_same_scale_requirement &&
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 38.1K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/experimental/tac/transforms/fold_constants_to_subgraph.cc

            return;
    
          // We only fold int32/int64 for Const and i32 for QConst if not specify
          // all constants flag. (Since they're more like "configs" or i32 biases.)
          // We will fold every const ops (and q_const ops) if we speicfy the
          // fold_all_constants_flag.
          if (!fold_all_constants_flag_) {
            if (!IsConstOrQConstInt(op)) return;
          }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.h

    struct OpQuantSpec {
      // Maps the operand index of a bias input to its quantization specifications,
      // including the non-bias operand indexes and the method retrieving
      // quantization parameters from list of parameters of the non-bias operands.
      // This map is empty if the op doesn't have a bias operand.
      BiasParamsMap biases_params;
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 20:30:06 UTC 2024
    - 41.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/tests/prepare-quantize.mlir

      %bias = arith.constant dense<1.0> : tensor<32xf32>
      %input = "tfl.dequantize"(%arg0) : (tensor<1x224x224x3x!quant.uniform<u8:f32, 1.5>>) -> tensor<1x224x224x3xf32>
      %weight = "tfl.dequantize"(%arg1) : (tensor<32x3x3x3x!quant.uniform<u8<1:255>:f32:3, {1.0,2.0,3.0}>>) -> tensor<32x3x3x3xf32>
      %conv = "tfl.conv_2d"(%input, %weight, %bias) {dilation_h_factor = 1 : i32, dilation_w_factor = 1 : i32,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 67.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/transforms/prepare_quantize_helper.h

          // TODO(b/172517537): make this work with non-PTQ case.
          if (llvm::isa<func::ConstantOp, arith::ConstantOp, TFL::ConstOp>(
                  input.getDefiningOp())) {
            // Tensors with derived scale are biases, and handled in propagation.
            if (tensor_property.use_derived_scale) continue;
            // For weights, use quantization scale inferred from the values.
            if (failed(processConstantOp(op, input.getDefiningOp(), index,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 28K bytes
    - Viewed (0)
Back to top