Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 26 for mul (0.03 sec)

  1. src/crypto/elliptic/params.go

    	}
    	i := new(big.Int).Lsh(h, 1)
    	i.Mul(i, i)
    	j := new(big.Int).Mul(h, i)
    
    	s1 := new(big.Int).Mul(y1, z2)
    	s1.Mul(s1, z2z2)
    	s1.Mod(s1, curve.P)
    	s2 := new(big.Int).Mul(y2, z1)
    	s2.Mul(s2, z1z1)
    	s2.Mod(s2, curve.P)
    	r := new(big.Int).Sub(s2, s1)
    	if r.Sign() == -1 {
    		r.Add(r, curve.P)
    	}
    	yEqual := r.Sign() == 0
    	if xEqual && yEqual {
    		return curve.doubleJacobian(x1, y1, z1)
    	}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Apr 16 17:46:09 UTC 2024
    - 9.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/legalize-tf-no-runtime-verification.mlir

    // CHECK-LABEL: broadcast_to_bf16
    // CHECK:  [[CST:%.*]] = arith.constant dense<1.000000e+00> : tensor<3x3xbf16>
    // CHECK:  [[MUL:%.*]] = tfl.mul(%arg0, [[CST]]) <{fused_activation_function = "NONE"}> : (tensor<3xbf16>, tensor<3x3xbf16>) -> tensor<3x3xbf16>
    // CHECK:  return [[MUL]] : tensor<3x3xbf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 648 bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/math.mlir

      %1 = "tfl.squared_difference"(%arg0, %0) {fused_activation_function = "NONE"} : (tensor<4xf32>, tensor<4xf32>) -> tensor<4xf32> loc("squared_difference")
      %2 = "tfl.mul"(%arg0, %1) {fused_activation_function = "NONE"} : (tensor<4xf32>, tensor<4xf32>) -> tensor<4xf32> loc("mul")
      %3 = "tfl.div"(%2, %1) {fused_activation_function = "NONE"} : (tensor<4xf32>, tensor<4xf32>) -> tensor<4xf32> loc("div")
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 1.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/stablehlo/odml_converter/transforms/outline_composites.cc

      rewriter.clearInsertionPoint();
      return new_func;
    }
    
    // Outlines non-approximate GELU into a stablehlo composite.
    //
    //    -> mul 1/sqrt(2) -> erf -> add 1 ->
    // in                                    mul
    //    ---------> mul 0.5 --------------->
    //
    // This pattern assumes all binary ewise ops with one constant argument
    // have that constant argument as the second operand. It works by
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 9.6K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/optimize_no_verify.mlir

      %1 = "tfl.mul"(%0, %cst_2) {fused_activation_function = "NONE"} : (tensor<1x256xbf16>, tensor<32x1x256xbf16>) -> tensor<32x1x256xbf16>
      func.return %1 : tensor<32x1x256xbf16>
    
    // CHECK:  %[[V0:.*]] = "tfl.fully_connected"(%arg0, {{.*}}) <{{{.*}}}> : (tensor<1x10368xbf16>, tensor<256x10368xbf16>, none) -> tensor<1x256xbf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 5.8K bytes
    - Viewed (0)
  6. src/crypto/ecdsa/ecdsa_legacy.go

    			kInv = new(big.Int).ModInverse(k, N)
    
    			r, _ = c.ScalarBaseMult(k.Bytes())
    			r.Mod(r, N)
    			if r.Sign() != 0 {
    				break
    			}
    		}
    
    		e := hashToInt(hash, c)
    		s = new(big.Int).Mul(priv.D, r)
    		s.Add(s, e)
    		s.Mul(s, kInv)
    		s.Mod(s, N) // N != 0
    		if s.Sign() != 0 {
    			break
    		}
    	}
    
    	return encodeSignature(r.Bytes(), s.Bytes())
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 00:11:18 UTC 2024
    - 4.8K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/stablehlo/tests/fold_broadcast.mlir

      %0 = "mhlo.broadcast_in_dim"(%cst0) <{broadcast_dimensions = dense<3> : tensor<1xi64>}> : (tensor<4xi32>) -> tensor<1x1x2x4xi32>
      // CHECK: %[[MUL:.*]] = mhlo.multiply %[[BROADCAST]], %[[ARG]] : tensor<1x1x2x4xi32>
      %1 = mhlo.multiply %0, %arg0 : tensor<1x1x2x4xi32>
      // CHECK:      return %[[MUL]] : tensor<1x1x2x4xi32>
      func.return %1 : tensor<1x1x2x4xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 15:32:52 UTC 2024
    - 4.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/experimental/tac/tests/e2e/device-transform-nnapi.mlir

        func.return %0 : tensor<4xf32>
        // CHECK:  [[VAL_0:%.*]] = tfl.sub %arg0, %arg1 {fused_activation_function = "NONE"} : tensor<4xf32>
        // CHECK:  [[VAL_1:%.*]] = tfl.mul [[VAL_0]], [[VAL_0]] {fused_activation_function = "NONE"} : tensor<4xf32
      }
    
      // CHECK-LABEL: pack
      func.func @pack(%arg0: tensor<1xf32>, %arg1: tensor<1xf32>) -> tensor<2x1xf32> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 1.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/convert_to_legacy_compile_and_replicate_attributes.mlir

        %outputs_11, %control_12 = tf_executor.island wraps "tf.Mul"(%outputs, %outputs_9) {_replication_info = "cluster", _xla_compile_device_type = "TPU", device = ""} : (tensor<f32>, tensor<*xf32>) -> tensor<*xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 22:03:30 UTC 2024
    - 6.1K bytes
    - Viewed (0)
  10. tensorflow/c/experimental/gradients/nn_grad.cc

      AbstractTensorHandle* expand_dims_outputs;
      TF_RETURN_IF_ERROR(
          ops::ExpandDims(ctx, vec, dim.get(), &expand_dims_outputs, "ExpandDims"));
      TF_RETURN_IF_ERROR(
          ops::Mul(ctx, expand_dims_outputs, mat, &outputs[0], "Mul"));
      expand_dims_outputs->Unref();
      return absl::OkStatus();
    }
    
    class SparseSoftmaxCrossEntropyWithLogitsGradientFunction
        : public GradientFunction {
     public:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 09 06:38:45 UTC 2024
    - 5.7K bytes
    - Viewed (0)
Back to top