- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 15 for 1xi8 (0.05 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/tests/bridge/optimize.mlir
} // ----- // CHECK-LABEL: func @convolution_add_add_i8 func.func @convolution_add_add_i8( %lhs: tensor<?x3x2x1xi8>, %rhs: tensor<2x1x1x1xi8>, %zp_offset: tensor<?x2x2x1xi8>, %bias: tensor<1xi8> ) -> tensor<?x2x2x1xi8> { // CHECK-DAG: %[[conv:.*]] = mhlo.convolution // CHECK-DAG: %[[combined:.*]] = chlo.broadcast_add %[[zp_offset:.*]], %[[bias:.*]]
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Feb 24 02:26:47 UTC 2024 - 10.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/tests/compose-uniform-quantized-type.mlir
%11 = stablehlo.reshape %2 : (tensor<1x1x1x1xi8>) -> tensor<1xi8> %12 = stablehlo.broadcast_in_dim %11, dims = [0] : (tensor<1xi8>) -> tensor<1x3x3x4xi8> %13 = stablehlo.convert %12 : (tensor<1x3x3x4xi8>) -> tensor<1x3x3x4xf32> %14 = stablehlo.convert %3 : (tensor<3x3x4x4xi8>) -> tensor<3x3x4x4xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 17:10:32 UTC 2024 - 37K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/constants.mlir
func.return %0 : tensor<4xf64> } func.func @i8() -> tensor<4xi8> { // CHECK-LABEL: @i8 // CHECK: value = dense<[1, 2, 3, 4]> : tensor<4xi8> %0 = "tfl.pseudo_const" () { value = dense<[1, 2, 3, 4]> : tensor<4xi8> } : () -> tensor<4xi8> func.return %0 : tensor<4xi8> } func.func @i16() -> tensor<4xi16> { // CHECK-LABEL: @i16
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 12.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/constants_offset.mlir
func.return %0 : tensor<4xf64> } func.func @i8() -> tensor<4xi8> { // CHECK-LABEL: @i8 // CHECK: value = dense<[1, 2, 3, 4]> : tensor<4xi8> %0 = "tfl.pseudo_const" () { value = dense<[1, 2, 3, 4]> : tensor<4xi8> } : () -> tensor<4xi8> func.return %0 : tensor<4xi8> } func.func @i16() -> tensor<4xi16> { // CHECK-LABEL: @i16
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 12.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-quant.mlir
// CHECK: %[[CONVERT_1:.*]] = mhlo.bitcast_convert %[[QUANTIZE]] : (tensor<2x!quant.uniform<i8:f32, 1.000000e+00:3>>) -> tensor<2xi8> // CHECK: %[[CONVERT_2:.*]] = mhlo.bitcast_convert %[[CONVERT_1]] : (tensor<2xi8>) -> tensor<2x!quant.uniform<i8:f32, 1.000000e+00:3>> // CHECK: %[[DEQUANTIZE:.*]] = mhlo.uniform_dequantize %[[CONVERT_2]] : (tensor<2x!quant.uniform<i8:f32, 1.000000e+00:3>>) -> tensor<2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 01:25:29 UTC 2024 - 37.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/const-fold.mlir
} // CHECK-LABEL: @cast_ui8_to_i8 func.func @cast_ui8_to_i8() -> tensor<4xi8> { %cst = arith.constant dense<[0, 255, 127, 128]> : tensor<4xui8> %0 = "tfl.cast"(%cst) : (tensor<4xui8>) -> tensor<4xi8> func.return %0 : tensor<4xi8> // CHECK: %[[CST:.*]] = arith.constant dense<[0, -1, 127, -128]> : tensor<4xi8> // CHECK: return %[[CST]] } // CHECK-LABEL: @cast_i8_to_i32
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 45.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library.mlir
%i8 = "tf.Cast"(%round) : (tensor<*xf32>) -> tensor<*xi8> func.return %i8 : tensor<*xi8> } func.func @dequantize_i8(%input : tensor<*xi8>, %scale : tensor<*xf32>, %zp : tensor<*xi32>) -> tensor<*xf32> { // Use identity op to avoid the weight being constant-folded. %identity = "tf.Identity"(%input) : (tensor<*xi8>) -> tensor<*xi8> %input_i32 = "tf.Cast"(%identity) : (tensor<*xi8>) -> tensor<*xi32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jan 08 01:16:10 UTC 2024 - 30.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_xla.mlir
func.return %11 : tensor<*xf32> } // CHECK: %[[maxpool_i8:.*]] = "tf.MaxPool" // CHECK-SAME: (tensor<*xi8>) -> tensor<*xi8> // CHECK: %[[reshape_i8:.*]] = "tf.Reshape"(%[[maxpool_i8]] // CHECK-SAME: (tensor<*xi8>, tensor<2xi32>) -> tensor<*xi8> // CHECK: %[[scast:.*]] = "quantfork.scast"(%[[reshape_i8]] // CHECK: %[[matmul:.*]] = "tf.PartitionedCall"(%[[scast]] // CHECK-SAME: f = @composite_matmul_fn_1
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 19:32:28 UTC 2024 - 11.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/attrs_and_constraints_test.cc
// Argument has rank 2, but its dimensions are unknown. constexpr absl::string_view kArgumentWithUnknownDims = R"mlir( func.func @unknown_dims_arg(%arg: tensor<?x?xi8>) -> tensor<?x?xi8> { return %arg : tensor<?x?xi8> } )mlir"; OwningOpRef<ModuleOp> module_op = ParseModuleOpString(kArgumentWithUnknownDims); ASSERT_TRUE(module_op);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 17:10:32 UTC 2024 - 22.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/ir/QuantOps.td
// operate on quantized values. // // Examples from storage to quantized type: // i8 -> !quant<"uniform[i8:f32]{1.0}"> // tensor<4xi8> -> tensor<4x!quant<"uniform[i8:f32]{1.0}">> // vector<4xi8> -> vector<4x!quant<"uniform[i8:f32]{1.0}">> def Quantization_StorageCastOp : Quantization_Op<"scast", [Pure]> { let arguments = (ins quant_RealOrStorageValueType:$arg);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jan 09 03:10:59 UTC 2024 - 10.2K bytes - Viewed (0)