- Sort Score
- Result 10 results
- Languages All
Results 61 - 70 of 108 for I8 (0.02 sec)
-
tensorflow/compiler/mlir/tfr/tests/decompose.mlir
func.func @quantized_tensor(%arg0: tensor<1x10x!quant.uniform<i8:f32, 0.0038396485615521669:-128>>) -> tensor<1x10x!quant.uniform<i8:f32, 3.906250e-03:-128>> { %0 = "tf.Intermediate"(%arg0) : (tensor<1x10x!quant.uniform<i8:f32, 0.0038396485615521669:-128>>) -> tensor<1x10x!quant.uniform<i8:f32, 3.906250e-03:-128>> func.return %0 : tensor<1x10x!quant.uniform<i8:f32, 3.906250e-03:-128>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 16.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_without_identity.pbtxt
attr { key: "Tshape" value { type: DT_INT32 } } } # MLIR-LABEL: func @main # MLIR-SAME: (%[[ARG_0:[a-z0-9]+]]: tensor<1x1x1x256x!quant.uniform<i8:f32, 0.21632751372549019:27>>) -> tensor<1x6x31x!quant.uniform<i8:f32, 0.09363494573854933:22>> # MLIR-SAME: control_outputs = "" # MLIR-SAME: inputs = "input" # MLIR-SAME: outputs = "output"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 13.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/tfl_while_outline.mlir
%cst_0 = arith.constant dense<1> : tensor<256x256xi8> %cst_1 = "tfl.pseudo_qconst"() {qtype = tensor<256x256x!quant.uniform<i8:f32, 1.000000e+00>>, value = dense<1> : tensor<256x256xi8>} : () -> tensor<256x256x!quant.uniform<i8:f32, 1.000000e+00>> %cst_2 = arith.constant dense<0> : tensor<i32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 13.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/ops/stablehlo_op_quant_spec_test.cc
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 04 07:19:09 UTC 2024 - 14.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/uniform_quantized_types.h
#include "mlir/Support/LLVM.h" // from @llvm-project namespace mlir { namespace quant { // Creates a `UniformQuantizedType` with the given `scale` and `zero_point` // values. The produced type has f32 as its expressed type and i8 as its // storage type. The available values use the full range of the storage value, // i.e. [-128, 127]. Assumes asymmetric quantization, meaning the zero point // value can be a non-zero value.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.7K bytes - Viewed (0) -
src/internal/types/testdata/spec/assignability.go
func _[ Int8 ~int8, Int16 ~int16, Int32 ~int32, Int64 ~int64, Int8_16 ~int8 | ~int16, ]( i8 Int8, i16 Int16, i32 Int32, i64 Int64, i8_16 Int8_16, ) { b = 42 b = 42.0 // etc. i8 = -1 << 7 i8 = 1<<7 - 1 i16 = -1 << 15 i16 = 1<<15 - 1 i32 = -1 << 31 i32 = 1<<31 - 1 i64 = -1 << 63 i64 = 1<<63 - 1
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Nov 09 17:24:42 UTC 2023 - 5.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/ir/QuantOps.td
// operate on quantized values. // // Examples from storage to quantized type: // i8 -> !quant<"uniform[i8:f32]{1.0}"> // tensor<4xi8> -> tensor<4x!quant<"uniform[i8:f32]{1.0}">> // vector<4xi8> -> vector<4x!quant<"uniform[i8:f32]{1.0}">> def Quantization_StorageCastOp : Quantization_Op<"scast", [Pure]> { let arguments = (ins quant_RealOrStorageValueType:$arg);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jan 09 03:10:59 UTC 2024 - 10.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_without_identity_4bit.pbtxt
attr { key: "Tshape" value { type: DT_INT32 } } } # MLIR-LABEL: func @main # MLIR-SAME: (%[[ARG_0:[a-z0-9]+]]: tensor<1x1x1x256x!quant.uniform<i8:f32, 0.21632751372549019:27>>) -> tensor<1x6x31x!quant.uniform<i8:f32, 0.09363494573854933:22>> # MLIR-SAME: control_outputs = "" # MLIR-SAME: inputs = "input" # MLIR-SAME: outputs = "output"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 13.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_per_channel.pbtxt
attr { key: "Tshape" value { type: DT_INT32 } } } # MLIR-LABEL: func @main # MLIR-SAME: (%[[ARG_0:[a-z0-9]+]]: tensor<1x1x1x256x!quant.uniform<i8:f32, 0.21632751372549019:27>>) -> tensor<1x6x31x!quant.uniform<i8:f32, 0.09363494573854933:22>> # MLIR-SAME: control_outputs = "" # MLIR-SAME: inputs = "input" # MLIR-SAME: outputs = "output"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 18.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/prepare_lifting.mlir
%0 = "quantfork.qcast"(%cst_1) : (tensor<2x3x3x2xf32>) -> tensor<2x3x3x2x!quant.uniform<i8<-127:127>:f32:3, {0.003937007874015748,0.003937007874015748}>> %1 = "quantfork.dcast"(%0) : (tensor<2x3x3x2x!quant.uniform<i8<-127:127>:f32:3, {0.003937007874015748,0.003937007874015748}>>) -> tensor<2x3x3x2xf32> %2 = "quantfork.qcast"(%arg0) : (tensor<1x3x4x3xf32>) -> tensor<1x3x4x3x!quant.uniform<i8:f32, 0.0011764706057660721:-43>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 14 03:24:59 UTC 2024 - 33.3K bytes - Viewed (0)