- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 60 for MULTIPLY (0.21 sec)
-
src/crypto/internal/bigmod/nat.go
out.montgomeryMul(out, out, m) // Select x^k in constant time from the table. k := uint((b >> j) & 0b1111) for i := range table { tmp.assign(ctEq(k, uint(i+1)), table[i]) } // Multiply by x^k, discarding the result if k = 0. tmp.montgomeryMul(out, tmp, m) out.assign(not(ctEq(k, 0)), tmp) } } return out.montgomeryReduction(m) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 13 18:57:38 UTC 2024 - 24K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/replace_stablehlo_ops_in_main_function_with_xla_call_module_ops.cc
// // Consider a sequence of op: // // ``` // node 1: %0 = stablehlo.constant // node 2: %1 = stablehlo.constant // node 3: %2 = stablehlo.add %0, %1 // node 4: %3 = stablehlo.multiply %2, %1 // node 5: return %3 // ``` // // In Backward Liveliness analysis, the liveliness for each node above becomes: // live_in[5] = use[5] U (live_out[5] - def[5])
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 21K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/tests/legalize_hlo.mlir
%0 = "mhlo.broadcast_in_dim"(%arg0) <{broadcast_dimensions = dense<[0, 1]> : tensor<2xi64>}> : (tensor<1x1xf32>) -> tensor<1x1000xf32> %1 = mhlo.multiply %0, %arg1 : tensor<1x1000xf32> %2 = mhlo.multiply %arg1, %0 : tensor<1x1000xf32> func.return %1, %2 : tensor<1x1000xf32>, tensor<1x1000xf32> } // CHECK-LABEL: func @broadcast_mul_chlo(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 07:26:59 UTC 2024 - 340.2K bytes - Viewed (0) -
android/guava-tests/test/com/google/common/net/InetAddressesTest.java
checkBigIntegerConversion("127.255.255.255", BigInteger.valueOf(Integer.MAX_VALUE)); checkBigIntegerConversion( "255.255.255.254", BigInteger.valueOf(Integer.MAX_VALUE).multiply(BigInteger.valueOf(2))); checkBigIntegerConversion( "255.255.255.255", BigInteger.ONE.shiftLeft(32).subtract(BigInteger.ONE)); } public void testFromIpv6BigIntegerValid() {
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri May 24 16:44:05 UTC 2024 - 35.3K bytes - Viewed (0) -
src/math/big/nat.go
// // x = xh*b + x0 (0 <= x0 < b) // y = yh*b + y0 (0 <= y0 < b) // b = 1<<(_W*k) ("base" of digits xi, yi) // k := karatsubaLen(n, karatsubaThreshold) // k <= n // multiply x0 and y0 via Karatsuba x0 := x[0:k] // x0 is not normalized y0 := y[0:k] // y0 is not normalized z = z.make(max(6*k, m+n)) // enough space for karatsuba of x0*y0 and full result of x*y
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 13 21:31:58 UTC 2024 - 31.7K bytes - Viewed (0) -
src/testing/benchmark.go
} // Order of operations matters. // For very fast benchmarks, prevIters ~= prevns. // If you divide first, you get 0 or 1, // which can hide an order of magnitude in execution time. // So multiply first, then divide. n = goalns * prevIters / prevns // Run more iterations than we think we'll need (1.2x). n += n / 5 // Don't grow too fast in case we had timing errors previously. n = min(n, 100*last)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 23.9K bytes - Viewed (0) -
docs/en/docs/async.md
* **Deep Learning**: this is a sub-field of Machine Learning, so, the same applies. It's just that there is not a single spreadsheet of numbers to multiply, but a huge set of them, and in many cases, you use a special processor to build and / or use those models. ### Concurrency + Parallelism: Web + Machine Learning
Registered: Mon Jun 17 08:32:26 UTC 2024 - Last Modified: Mon May 20 00:24:48 UTC 2024 - 23K bytes - Viewed (0) -
src/cmd/asm/internal/asm/asm.go
p.firstProg = prog } else { p.lastProg.Link = prog } p.lastProg = prog if doLabel { p.pc++ for _, label := range p.pendingLabels { if p.labels[label] != nil { p.errorf("label %q multiply defined", label) return } p.labels[label] = prog } p.pendingLabels = p.pendingLabels[0:0] } prog.Pc = p.pc if *flags.Debug { fmt.Println(p.lineNum, prog) } if testOut != nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 02:04:54 UTC 2024 - 25.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.h
bool is_signed, bool narrow_range, bool legacy_float_scale = false, bool use_fake_quant_num_bits = false); // Returns the quantized type of a bias input, given the quantized types of // other operands which are multiply-accumulated (the bias is added to the // accumulated value). quant::QuantizedType GetUniformQuantizedTypeForBias( const std::vector<quant::QuantizedType>& op_types, int adjusted_quant_dim,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 20:30:06 UTC 2024 - 41.7K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/PPC64.rules
(Sub32F ...) => (FSUBS ...) (Sub64F ...) => (FSUB ...) (Min(32|64)F x y) && buildcfg.GOPPC64 >= 9 => (XSMINJDP x y) (Max(32|64)F x y) && buildcfg.GOPPC64 >= 9 => (XSMAXJDP x y) // Combine 64 bit integer multiply and adds (ADD l:(MULLD x y) z) && buildcfg.GOPPC64 >= 9 && l.Uses == 1 && clobber(l) => (MADDLD x y z) (Mod16 x y) => (Mod32 (SignExt16to32 x) (SignExt16to32 y)) (Mod16u x y) => (Mod32u (ZeroExt16to32 x) (ZeroExt16to32 y))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 53.2K bytes - Viewed (0)