Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 20 for Sigmoid (0.15 sec)

  1. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.h

      // inconsistent with Conv 1x1 which always performs per channel quantization.
      bool disable_per_channel_for_dense_layers = false;
    
      // Whether to use fixed output ranges of the activation ops (tanh, sigmoid,
      // etc.) and not infer weight constants.
      // If this option is set, quantization emulation ops should be placed after
      // the ops in the input graph. This flag should be set to false for
      // post-training quantization.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 13 10:16:19 UTC 2024
    - 10.8K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_patterns.td

                (CastValueToI64 $old, $shape),
                MHLO_RngDistributionValue<"NORMAL">),
              [(IsShapedTensor $shape)]>;
    
    //===----------------------------------------------------------------------===//
    // Sigmoid grad op.
    //===----------------------------------------------------------------------===//
    
    // TODO(hinsu): Handle unranked inputs by broadcasting constant one to the
    // shape of $l instead of having it as a constant.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 34.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/legalize_patterns.td

    // =============================================================================
    // Training OPs
    // =============================================================================
    
    // `grad = dy * y * (1 - y)`, where `y = sigmoid(x)`
    def LegalizeSigmoidGrad : Pat<(TF_SigmoidGradOp $y, $dy),
      (TFL_MulOp $dy, (TFL_MulOp $y, (TFL_SubOp
            (Arith_ConstantOp ConstantAttr<RankedF32ElementsAttr<[]>, "1.0f">),
            $y, TFL_AF_None),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 04 13:30:42 UTC 2024
    - 28.5K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.h

    // be checked first if present.
    // TODO: b/323478683: Consider deprecating this.
    struct OpQuantScaleSpec {
      // Whether this op has a fixed range requirement (e.g. sigmoid)
      bool has_fixed_output_range = false;
      // Whether this op should have same operand and result scales (e.g. concat)
      bool has_same_scale_requirement = false;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 20:30:06 UTC 2024
    - 41.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

    // CHECK-LABEL: argmin
    // CHECK:  "tfl.arg_min"(%arg0, %arg1) : (tensor<3xi32>, tensor<i32>) -> tensor<i32>
    }
    
    func.func @sigmoid(%arg0: tensor<?x88xf32>) -> tensor<?x88xf32> {
      %0 = "tf.Sigmoid"(%arg0) : (tensor<?x88xf32>) -> tensor<?x88xf32>
      func.return %0 : tensor<?x88xf32>
    // CHECK-LABEL: sigmoid
    // CHECK:  "tfl.logistic"(%arg0) : (tensor<?x88xf32>) -> tensor<?x88xf32>
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/transforms/optimize_patterns.td

            $filter_width, $filter_height, ActFnAttr),
        [(HasOneUse $pool_out)]>;
    }
    
    // TODO(hinsu): Also fuse ops corresponding to SIGN_BIT fused
    // activation functions.
    // Currently we're not fusing tanh, sigmoid, hard_swish and other activations
    // those cannot be simply translated into clamping.
    foreach actFnPair = [[TFL_ReluOp, TFL_AF_Relu],
                         [TFL_Relu6Op, TFL_AF_Relu6],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 16 20:31:41 UTC 2024
    - 66.4K bytes
    - Viewed (0)
  7. RELEASE.md

        recurrent activation function for GRU from `hard_sigmoid` to `sigmoid`, and
        `reset_after` to True in 2.0. Historically recurrent activation is
        `hard_sigmoid` since it is fast than 'sigmoid'. With new unified backend
        between CPU and GPU mode, since the CuDNN kernel is using sigmoid, we change
        the default for CPU mode to sigmoid as well. With that, the default GRU will
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf.mlir

      %0 = "tf.Neg"(%arg0) : (tensor<?xf32>) -> tensor<?xf32>
      func.return %0 : tensor<?xf32>
    }
    
    // -----
    
    // CHECK-LABEL: @sigmoid
    func.func @sigmoid(%arg0: tensor<2xf32>) -> tensor<2xf32> {
      // CHECK: mhlo.logistic
      %0 = "tf.Sigmoid"(%arg0) : (tensor<2xf32>) -> tensor<2xf32>
      func.return %0 : tensor<2xf32>
    }
    
    // -----
    
    // CHECK-LABEL: @sigmoid_complex
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 335.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/jit/mark_for_compilation_pass.cc

                "Floor", "IsFinite", "IsInf", "IsNan", "Inv", "Reciprocal", "Log",
                "Log1p", "Invert", "LogicalNot", "Ndtri", "Neg", "Rint", "Round",
                "Rsqrt", "Sigmoid", "Sign", "Sinh", "Softplus", "Softsign", "Sqrt",
                "Square", "Tan", "Tanh", "Real", "Imag", "Erf", "Erfc", "Erfinv",
                "Lgamma", "Digamma",
                // Binary
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 12:19:41 UTC 2024
    - 85.3K bytes
    - Viewed (0)
  10. src/crypto/rsa/rsa.go

    	}
    
    	var (
    		err  error
    		m, c *bigmod.Nat
    		N    *bigmod.Modulus
    		t0   = bigmod.NewNat()
    	)
    	if priv.Precomputed.n == nil {
    		N, err = bigmod.NewModulusFromBig(priv.N)
    		if err != nil {
    			return nil, ErrDecryption
    		}
    		c, err = bigmod.NewNat().SetBytes(ciphertext, N)
    		if err != nil {
    			return nil, ErrDecryption
    		}
    		m = bigmod.NewNat().Exp(c, priv.D.Bytes(), N)
    	} else {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 00:11:18 UTC 2024
    - 23.4K bytes
    - Viewed (0)
Back to top