Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 22 for f16 (0.07 sec)

  1. tensorflow/compiler/mlir/quantization/stablehlo/passes/quantize_weight.cc

          // Create new ConstantOp-ConvertOp-Operation sequences. At this moment,
          // old ConstantOp is guaranteed to have one F32->F16 convert op regardless
          // of its number of users.
          rewriter.setInsertionPointAfter(op);
          // create new F16 constant op in that location
          ConstantOp new_const = rewriter.create<ConstantOp>(
              op->getLoc(), new_result_type, new_value_attr);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 9.9K bytes
    - Viewed (0)
  2. src/cmd/compile/internal/ssa/opGen.go

    				{1, 9223372034707292160}, // F0 F1 F2 F3 F4 F5 F6 F7 F8 F9 F10 F11 F12 F13 F14 F15 F16 F17 F18 F19 F20 F21 F22 F23 F24 F25 F26 F27 F28 F29 F30 F31
    			},
    			outputs: []outputInfo{
    				{0, 9223372034707292160}, // F0 F1 F2 F3 F4 F5 F6 F7 F8 F9 F10 F11 F12 F13 F14 F15 F16 F17 F18 F19 F20 F21 F22 F23 F24 F25 F26 F27 F28 F29 F30 F31
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 15:49:20 UTC 2024
    - 1M bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/tests/optimize_no_verify.mlir

    // CHECK-LABEL: fuseScalarAddIntoConv2dHalf
    func.func @fuseScalarAddIntoConv2dHalf(%arg0: tensor<256x32x32x3xf16>, %arg1: tensor<16x3x3x3xf16>) -> tensor<256x8x7x16xf16> {
      %cst = arith.constant dense<1.5> : tensor<f16>
      %cst_0 = arith.constant dense<[1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0, 10.0, 11.0, 12.0, 13.0, 14.0, 15.0, 16.0]> : tensor<16xf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 5.8K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/utils/validators.h

      return tensorType.getElementType().isBF16();
    }
    
    // Returns true iff the given value is a f16 tensor.
    inline bool TFTypeIsHalfTensor(Value value) {
      auto tensorType = mlir::dyn_cast<TensorType>(value.getType());
      if (!tensorType) return false;
      return tensorType.getElementType().isF16();
    }
    
    // Returns true iff the given value is a f16 or bf16 tensor.
    inline bool TFTypeIsBFloat16OrHalfTensor(Value value) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 4.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/const-fold.mlir

      %5 = "tfl.mul"(%0, %1) {fused_activation_function = "NONE"} : (tensor<  f16>, tensor<  f16>) -> tensor<  f16>
      %6 = "tfl.mul"(%0, %3) {fused_activation_function = "NONE"} : (tensor<  f16>, tensor<4xf16>) -> tensor<4xf16>
      %7 = "tfl.mul"(%2, %1) {fused_activation_function = "NONE"} : (tensor<4xf16>, tensor<  f16>) -> tensor<4xf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 45.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/constants_offset.mlir

      func.return %0 : tensor<4xcomplex<f64>>
    }
    
    func.func @f16() -> tensor<4xf16> {
      // CHECK-LABEL: @f16
      // CHECK: value = dense<[1.000000e+00, 2.000000e+00, 3.000000e+00, 4.000000e+00]> : tensor<4xf16>
      %0 = "tfl.pseudo_const"() { value = dense<[1.0, 2.0, 3.0, 4.0]> : tensor<4xf16> } : () -> tensor<4xf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 12.1K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/constants.mlir

      func.return %0 : tensor<4xcomplex<f64>>
    }
    
    func.func @f16() -> tensor<4xf16> {
      // CHECK-LABEL: @f16
      // CHECK: value = dense<[1.000000e+00, 2.000000e+00, 3.000000e+00, 4.000000e+00]> : tensor<4xf16>
      %0 = "tfl.pseudo_const"() { value = dense<[1.0, 2.0, 3.0, 4.0]> : tensor<4xf16> } : () -> tensor<4xf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 12.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/transforms/optimize_op_order.cc

            get_num_elements(input_type) <= get_num_elements(output_type)) {
          return failure();
        }
        Type input_element_type = getElementTypeOrSelf(dequantize_op.getInput());
        // Most passthrough ops do not support F16.
        if (input_element_type.isF16()) {
          return failure();
        }
    
        // Set the output type of the dequantize op and push it down.
        dequantize_op.getOutput().setType(output_type);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_patterns.td

    include "mhlo/IR/hlo_ops.td"
    
    def SignedIntTensor : TensorOf<[I1, I8, I16, I32, I64]>;
    def UnsignedIntTensor : TensorOf<[UI8, UI16, UI32, UI64]>;
    
    // IEEE compliant floating point tensors.
    def IEEEFloatTensor : TensorOf<[F16, F32, F64]>;
    
    //===----------------------------------------------------------------------===//
    // BatchNorm op patterns.
    //===----------------------------------------------------------------------===//
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 34.8K bytes
    - Viewed (0)
  10. src/crypto/tls/testdata/Client-TLSv12-RenegotiationRejected

    00000000  14 03 03 00 01 01 16 03  03 00 20 2e d7 9e b0 66  |.......... ....f|
    00000010  f1 39 84 a9 d9 93 17 bd  94 5a 31 49 bb eb f0 a0  |.9.......Z1I....|
    00000020  7b af d0 3b ae 1a 5d f6  46 31 36                 |{..;..].F16|
    >>> Flow 5 (client to server)
    00000000  17 03 03 00 16 cd 39 a7  64 3b 6a de 14 e0 26 ea  |......9.d;j...&.|
    00000010  66 b2 73 b1 8e b0 e3 a9  94 62 4f                 |f.s......bO|
    >>> Flow 6 (server to client)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 22:33:38 UTC 2024
    - 7K bytes
    - Viewed (0)
Back to top