- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 106 for Bias (0.14 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/optimize.td
def optimizeConsecutiveConv#OpsTuple[0]#OpsTuple[1] : Pat< (OpsTuple[1] (OpsTuple[0] $input, $zp_offset, $broadcast_dims_1), $bias, $broadcast_dims_2), (OpsTuple[0] $input, (OpsTuple[2] $zp_offset, $bias, $broadcast_dims_2), $broadcast_dims_1), [ (IsNull $broadcast_dims_1), (IsNull $broadcast_dims_2), (TensorOf<[AnyInteger]> $input),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Feb 24 02:26:47 UTC 2024 - 2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/pipelines/process_nchw_tensor.mlir
// CHECK: return %[[TRANSPOSE_1]] // ----- // Tests that a `add(convolution(%activation, %weight), %bias)` pattern with the // activation tensor of NCHW format and non-constant bias is converted to NHWC // convolution, but without the deferred transpose for `stablehlo.add`. // Transpose ops are inserted to the activation and output of
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 18 20:32:46 UTC 2024 - 12.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization.td
// Specify this trait if the bias-th input of the op is a bias input, which // needs a scale based on the scales of op1 and op2. class AccumulatorUniformScale<int bias, int op1, int op2> : NativeOpTrait< !strconcat("quant::AccumulatorUniformScale<", !interleave([bias, op1, op2], ", "), ">::Impl")>;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 07:39:40 UTC 2024 - 8.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/quantize_patterns.td
def FoldQuantWeightsIntoTposeConv : Pat< (TFL_TransposeConvOp $output_shape, (TFL_DequantizeOp $quant_weights), $quant_input, $bias, $padding, $stride_h, $stride_w, $faf), (TFL_TransposeConvOp $output_shape, $quant_weights, $quant_input, $bias, $padding, $stride_h, $stride_w, $faf),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 23:10:13 UTC 2024 - 2.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_xla.mlir
%weight = arith.constant dense_resource<__elided__> : tensor<2x3x3x2xf32> %bias = arith.constant dense<[7.11401462, 7.05456924]> : tensor<2xf32> %q_input= "quantfork.qcast"(%input) : (tensor<1x3x4x3xf32>) -> tensor<1x3x4x3x!quant.uniform<i8:f32, 0.58810077742034317:-128>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 19:32:28 UTC 2024 - 11.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_driver.h
// constant is used by multiple ops as a bias, duplicate constants and // let each op assign its own quantization parameter for bias. // - Adds all the non-bias constants (weights) to a set for looking up // later. // - Adds all per-channel weights to a set for looking up later. void PreprocessConstantOps();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 20 11:42:17 UTC 2024 - 16.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/python/integration_test/quantize_model_test_base.py
bias_fn: Optional[ops.Operation] = None, activation_fn: Optional[ops.Operation] = None, ) -> module.Module: class MatmulModel(module.Module): """A simple model with a single matmul. Bias and activation function are optional. """ def __init__( self, weight_shape: Sequence[int], ) -> None: """Initializes a MatmulModel. Args:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 18.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/lift_as_function_call.td
"\""# func_name #"\", $0...)", returns>; // Add the second argument to the first argument, which is expected to be an // argument list. // bias(einsum(inputs), bias) --> einsum_with_bias(AppendToVector(inputs, bias)) // Since inputs is a vector in case of einsum, we cannot use ArgumentList here. def AppendToVector : NativeCodeCall<"AppendToVector($0, $1)">;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 25 00:32:20 UTC 2024 - 3.4K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad_test.cc
A.reset(A_raw); } // Bias float Bias_vals[] = {2.0f, 3.0f}; int64_t Bias_dims[] = {2}; AbstractTensorHandlePtr Bias; { AbstractTensorHandle* Bias_raw; status_ = TestTensorHandleWithDims<float, TF_FLOAT>( immediate_execution_ctx_.get(), Bias_vals, Bias_dims, 1, &Bias_raw); ASSERT_EQ(errors::OK, status_.code()) << status_.message(); Bias.reset(Bias_raw); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 28 13:53:47 UTC 2024 - 8.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/transforms/device_transform_patterns.cc
PatternRewriter& rewriter) { auto bias = op->getOperand(bias_idx); if (!mlir::isa<NoneType>(bias.getType())) return failure(); // Proceed to create a zero bias. auto output = op->getResult(0); auto output_type = mlir::dyn_cast_or_null<RankedTensorType>(output.getType()); if (!output_type) return failure(); // bias should be a vector sized of the last output dim.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 25.4K bytes - Viewed (0)