Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 41 for ctanh (0.24 sec)

  1. tensorflow/compiler/mlir/lite/tests/prepare-quantize-post-training.mlir

          %cst_2, %cst_2, %cst_2,
          %cst_7, %cst_7, %cst_7, %cst_7,
          %cst_2, %cst_2,
          %recurrent_stats, %cell_stats,
          %cst_2, %cst_2, %cst_2, %cst_2) {cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", proj_clip = 0.000000e+00 : f32, time_major = false}
        : ( tensor<1x28x28xf32>,
            tensor<20x28xf32>, tensor<20x28xf32>, tensor<20x28xf32>, tensor<20x28xf32>,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 52.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/stablehlo/transforms/composite_lowering_patterns.td

              (variadic $inputs),
              ConstantStrAttr<StrAttr, "aten.gelu.default">, $attrs, $_, $_),
            (TFL_GeluOp $inputs, ConstBoolAttrTrue),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 18:45:51 UTC 2024
    - 4.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/lstm.json

              "intermediates": [
                25, 26, 27, 28, 29
              ],
              "builtin_options_type": "UnidirectionalSequenceLSTMOptions",
              "builtin_options": {
                "fused_activation_function": "TANH",
                "cell_clip": 50.0
              },
              "mutating_variable_inputs": [
                false,
                false, false, false, false,
                false, false, false, false,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 01 06:25:50 UTC 2024
    - 9.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/prepare-composite-functions-tf.mlir

    // CHECK-SAME:      <{cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", kernel_type = #tfl<lstm_kernel_type_attr FULL>, proj_clip = 0.000000e+00 : f32}> ({
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 122.1K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/transforms/legalize_tf.cc

        // Populate attributes.
        SmallVector<NamedAttribute, 4> attributes;
        // Activation will always be tanh.
        attributes.push_back(rewriter.getNamedAttr("fused_activation_function",
                                                   rewriter.getStringAttr("TANH")));
        // cell_clip.
        attributes.push_back(
            rewriter.getNamedAttr("cell_clip", rewriter.getF32FloatAttr(0.0)));
        // proj_clip.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 20 20:06:54 UTC 2024
    - 45.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/stablehlo/tests/optimize.mlir

        %4 = mhlo.exponential %3 : tensor<13x13x13xf32>
        %5 = mhlo.tanh %3 : tensor<13x13x13xf32>
        return %4, %5 : tensor<13x13x13xf32>, tensor<13x13x13xf32>
    
    // CHECK:      "mhlo.pad"(%arg0, %0) <{
    // CHECK-SAME:     edge_padding_high = dense<2> : tensor<3xi64>,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 15:32:52 UTC 2024
    - 22.7K bytes
    - Viewed (0)
  7. tensorflow/cc/gradients/math_grad.cc

    REGISTER_GRADIENT_OP("Acosh", AcoshGrad);
    
    Status AtanhGrad(const Scope& scope, const Operation& op,
                     const std::vector<Output>& grad_inputs,
                     std::vector<Output>* grad_outputs) {
      // y = atanh(x)
      // dy/dx = 1 / (1 - x^2)
      auto one = Cast(scope, Const(scope, 1.0), op.input(0).type());
      auto dydx = Reciprocal(scope, Sub(scope, one, Square(scope, op.input(0))));
      // grad(x) = grad(y) * conj(dy/dx)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Aug 25 18:20:20 UTC 2023
    - 50.7K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.h

      // inconsistent with Conv 1x1 which always performs per channel quantization.
      bool disable_per_channel_for_dense_layers = false;
    
      // Whether to use fixed output ranges of the activation ops (tanh, sigmoid,
      // etc.) and not infer weight constants.
      // If this option is set, quantization emulation ops should be placed after
      // the ops in the input graph. This flag should be set to false for
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 13 10:16:19 UTC 2024
    - 10.8K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/utils/lstm_utils.cc

          input_cell_state_, input_layer_norm_coefficients_,
          forget_layer_norm_coefficients_, cell_layer_norm_coefficients_,
          output_layer_norm_coefficients_, builder_.getStringAttr("TANH"),
          builder_.getF32FloatAttr(10.0), builder_.getF32FloatAttr(0.0),
          mlir::TFL::LSTMKernelTypeAttr::get(builder_.getContext(),
                                             mlir::TFL::LSTMKernelType::FULL),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 36.2K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/tests/ops.mlir

    ^bb0(%arg0: tensor<? x f32>):
      // CHECK: "tfl.tanh"(%arg0)
      %0 = "tfl.tanh"(%arg0): (tensor<? x f32>) -> tensor<? x f32>
      func.return %0 : tensor<? x f32>
    }
    
    // CHECK-LABEL: testTanhWithQI8
    func.func @testTanhWithQI8(%arg0: tensor<? x !quant.uniform<i8:f32, 0.1>>) -> tensor<? x !quant.uniform<i8:f32, 0.1>> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
Back to top