- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 22 for f16 (0.07 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/passes/quantize_weight.cc
// Create new ConstantOp-ConvertOp-Operation sequences. At this moment, // old ConstantOp is guaranteed to have one F32->F16 convert op regardless // of its number of users. rewriter.setInsertionPointAfter(op); // create new F16 constant op in that location ConstantOp new_const = rewriter.create<ConstantOp>( op->getLoc(), new_result_type, new_value_attr);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 9.9K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/opGen.go
{1, 9223372034707292160}, // F0 F1 F2 F3 F4 F5 F6 F7 F8 F9 F10 F11 F12 F13 F14 F15 F16 F17 F18 F19 F20 F21 F22 F23 F24 F25 F26 F27 F28 F29 F30 F31 }, outputs: []outputInfo{ {0, 9223372034707292160}, // F0 F1 F2 F3 F4 F5 F6 F7 F8 F9 F10 F11 F12 F13 F14 F15 F16 F17 F18 F19 F20 F21 F22 F23 F24 F25 F26 F27 F28 F29 F30 F31
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 1M bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/optimize_no_verify.mlir
// CHECK-LABEL: fuseScalarAddIntoConv2dHalf func.func @fuseScalarAddIntoConv2dHalf(%arg0: tensor<256x32x32x3xf16>, %arg1: tensor<16x3x3x3xf16>) -> tensor<256x8x7x16xf16> { %cst = arith.constant dense<1.5> : tensor<f16> %cst_0 = arith.constant dense<[1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0, 10.0, 11.0, 12.0, 13.0, 14.0, 15.0, 16.0]> : tensor<16xf16>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 5.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/validators.h
return tensorType.getElementType().isBF16(); } // Returns true iff the given value is a f16 tensor. inline bool TFTypeIsHalfTensor(Value value) { auto tensorType = mlir::dyn_cast<TensorType>(value.getType()); if (!tensorType) return false; return tensorType.getElementType().isF16(); } // Returns true iff the given value is a f16 or bf16 tensor. inline bool TFTypeIsBFloat16OrHalfTensor(Value value) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 4.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/const-fold.mlir
%5 = "tfl.mul"(%0, %1) {fused_activation_function = "NONE"} : (tensor< f16>, tensor< f16>) -> tensor< f16> %6 = "tfl.mul"(%0, %3) {fused_activation_function = "NONE"} : (tensor< f16>, tensor<4xf16>) -> tensor<4xf16> %7 = "tfl.mul"(%2, %1) {fused_activation_function = "NONE"} : (tensor<4xf16>, tensor< f16>) -> tensor<4xf16>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 45.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/constants_offset.mlir
func.return %0 : tensor<4xcomplex<f64>> } func.func @f16() -> tensor<4xf16> { // CHECK-LABEL: @f16 // CHECK: value = dense<[1.000000e+00, 2.000000e+00, 3.000000e+00, 4.000000e+00]> : tensor<4xf16> %0 = "tfl.pseudo_const"() { value = dense<[1.0, 2.0, 3.0, 4.0]> : tensor<4xf16> } : () -> tensor<4xf16>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 12.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/constants.mlir
func.return %0 : tensor<4xcomplex<f64>> } func.func @f16() -> tensor<4xf16> { // CHECK-LABEL: @f16 // CHECK: value = dense<[1.000000e+00, 2.000000e+00, 3.000000e+00, 4.000000e+00]> : tensor<4xf16> %0 = "tfl.pseudo_const"() { value = dense<[1.0, 2.0, 3.0, 4.0]> : tensor<4xf16> } : () -> tensor<4xf16>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 12.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize_op_order.cc
get_num_elements(input_type) <= get_num_elements(output_type)) { return failure(); } Type input_element_type = getElementTypeOrSelf(dequantize_op.getInput()); // Most passthrough ops do not support F16. if (input_element_type.isF16()) { return failure(); } // Set the output type of the dequantize op and push it down. dequantize_op.getOutput().setType(output_type);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_patterns.td
include "mhlo/IR/hlo_ops.td" def SignedIntTensor : TensorOf<[I1, I8, I16, I32, I64]>; def UnsignedIntTensor : TensorOf<[UI8, UI16, UI32, UI64]>; // IEEE compliant floating point tensors. def IEEEFloatTensor : TensorOf<[F16, F32, F64]>; //===----------------------------------------------------------------------===// // BatchNorm op patterns. //===----------------------------------------------------------------------===//
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 06 18:46:23 UTC 2024 - 34.8K bytes - Viewed (0) -
src/crypto/tls/testdata/Client-TLSv12-RenegotiationRejected
00000000 14 03 03 00 01 01 16 03 03 00 20 2e d7 9e b0 66 |.......... ....f| 00000010 f1 39 84 a9 d9 93 17 bd 94 5a 31 49 bb eb f0 a0 |.9.......Z1I....| 00000020 7b af d0 3b ae 1a 5d f6 46 31 36 |{..;..].F16| >>> Flow 5 (client to server) 00000000 17 03 03 00 16 cd 39 a7 64 3b 6a de 14 e0 26 ea |......9.d;j...&.| 00000010 66 b2 73 b1 8e b0 e3 a9 94 62 4f |f.s......bO| >>> Flow 6 (server to client)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:33:38 UTC 2024 - 7K bytes - Viewed (0)