- Sort Score
- Result 10 results
- Languages All
Results 131 - 140 of 226 for f32 (0.02 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/prepare_quantize/prepare_quantize_per_channel.mlir
%cst_min = stablehlo.constant dense<0.0> : tensor<f32> %cst_max = stablehlo.constant dense<6.0> : tensor<f32> %7 = "stablehlo.clamp"(%cst_min, %6, %cst_max) {device = ""} : (tensor<f32>, tensor<1x2x2x2xf32>, tensor<f32>) -> tensor<1x2x2x2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 26 07:48:15 UTC 2024 - 8.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/execution_metadata_exporter_test.cc
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 06:11:34 UTC 2024 - 6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/quantization_options.proto
// conversion, then dequantized during inference. // Activation: f32, Weight: qi8, Bias: f32 WEIGHT_ONLY = 1; // Apply default dynamic range quantization. Quantized tensor value's // ranges are determined during graph runtime. // Activation: f32, Weight: qi8, Bias: f32 POST_TRAINING_QUANTIZATION_DYNAMIC_RANGE = 2;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 22 02:20:05 UTC 2023 - 3.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/ir/testdata/test.mlir
%cpu = corert.get_op_handler %ch "cpu" %0 = corert.executeop(%cpu) "tf.Relu"(%arg0) { T = f32 } : 1 %arg1 = tfrt_fallback_async.corert_tensorhandle_to_fallback_tensor %arg1_th {_tfrt_cost = 1 : i64, device = "/CPU:0"} : (!corert.tensorhandle) -> (!tfrt_fallback.tf_tensor) %1 = tfrt_fallback_async.executeop key(0) cost(100) device("/CPU:0") "tf.Relu"(%arg1) { T = f32 } : 1 tfrt.return
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 25 11:03:04 UTC 2022 - 496 bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize_batch_matmul.td
def FuseTransposeFCLhsToBatchMatmul : Pat< (TFL_BatchMatMulOp (TFL_FullyConnectedOp:$fc_output (TFL_TransposeOp TensorOf<[F32]>:$fc_lhs, (Arith_ConstantOp:$perm_value $p0)), TensorOf<[F32]>:$fc_rhs, $bias, $TFL_AF_None, $TFL_FCWO_Default, $keep_num_dims, $asymmetric_quantize_inputs_fc ), $bmm_rhs, ConstBoolAttrTrue, $transpose_rhs,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 09 23:44:09 UTC 2023 - 2.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/replicate-tensor-list-init-ops.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Feb 03 09:30:08 UTC 2023 - 1.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/ir/QuantizeUtils.h
/// Examples: /// 1. realValue is a primitive value attribute: /// (realValue: FloatAttr, quantizedElementType: UniformQuantizedType[i8:f32]) /// -> (IntegerAttr, outConvertedType: i8) /// 2. realValue is an elements attribute: /// (realValue: DenseElementsAttr[tensor<2x2xf32>], /// quantizedElementType: UniformQuantizedType[i8:f32]) /// -> (DenseElementsAttr[tensor<2x2xi8>], outConvertedType: tensor<2x2xi8>) Attribute quantizeAttr(Attribute realValue,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jul 29 18:55:28 UTC 2022 - 3.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/device_assignment.mlir
// CHECK: device = "gpu" %1 = "tf.MatMul"(%arg0, %0) {T = f32, _output_shapes = ["tfshape$dim { size: 3 } dim { size: 3 }"], device = "", transpose_a = false, transpose_b = false} : (tensor<3x1xf32>, tensor<1x3xf32>) -> tensor<3x3xf32> // CHECK: device = "cpu" %2 = "tf.Relu"(%1) {T = f32, _output_shapes = ["tfshape$dim { size: 3 } dim { size: 3 }"], device = "cpu"} : (tensor<3x3xf32>) -> tensor<3x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 24 05:47:26 UTC 2022 - 924 bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_layout_assignment_gpu_cc_70.mlir
%y, %batch_mean, %batch_var, %reserve_1, %reserve_2, %reserve_3 = "tf.FusedBatchNormV3"(%arg0, %arg1, %arg1, %arg1, %arg1) { data_format = "NHWC", epsilon = 1.001 : f32, exponential_avg_factor = 1.0 : f32, is_training = true } : (tensor<1x28x28x64xf32>, tensor<64xf32>, tensor<64xf32>, tensor<64xf32>, tensor<64xf32>)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 21 08:41:18 UTC 2022 - 8.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/convert_func_to_bfloat16.cc
}); } }; // This helper function makes legality check easier. Both convert ops in the // patterns below are considered legal: // - `BitcastConvertOp` (i32 -> f32) + `ConvertOp` (f32 -> bf16) // - `ConvertOp` (bf16 -> f32) -> `BitcastConvertOp` (f32 -> i32) template <typename ConvertOp, typename OtherConvertOp> bool IsConvertOpLegal(ConvertOp convert_op, BFloat16TypeConverter& converter) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 9.3K bytes - Viewed (0)