- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 136 for Shift2 (0.33 sec)
-
src/cmd/compile/internal/ssa/_gen/PPC64.rules
(Rsh16x(64|32|16|8) x y) && shiftIsBounded(v) => (SRAD (MOVHreg x) y) (Rsh8x(64|32|16|8) x y) && shiftIsBounded(v) => (SRAD (MOVBreg x) y) // Unbounded shifts. Go shifts saturate to 0 or -1 when shifting beyond the number of // bits in a type, PPC64 shifts do not (see the ISA for details). // // Note, y is always non-negative. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 53.2K bytes - Viewed (0) -
src/syscall/mksyscall.pl
if($ARGV[0] eq "-b32") { $_32bit = "big-endian"; shift; } elsif($ARGV[0] eq "-l32") { $_32bit = "little-endian"; shift; } if($ARGV[0] eq "-plan9") { $plan9 = 1; shift; } if($ARGV[0] eq "-darwin") { $darwin = 1; $libc = 1; shift; } if($ARGV[0] eq "-openbsd") { $openbsd = 1; shift; } if($ARGV[0] eq "-netbsd") { $netbsd = 1; shift; } if($ARGV[0] eq "-dragonfly") { $dragonfly = 1;
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 24 17:15:02 UTC 2024 - 10.3K bytes - Viewed (0) -
src/runtime/slice.go
case isPowerOfTwo(et.Size_): var shift uintptr if goarch.PtrSize == 8 { // Mask shift for better code generation. shift = uintptr(sys.TrailingZeros64(uint64(et.Size_))) & 63 } else { shift = uintptr(sys.TrailingZeros32(uint32(et.Size_))) & 31 } lenmem = uintptr(oldLen) << shift newlenmem = uintptr(newLen) << shift capmem = roundupsize(uintptr(newcap)<<shift, noscan)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 12.2K bytes - Viewed (0) -
src/crypto/md5/md5block_ppc64x.s
#define ROUND1(a, b, c, d, index, const, shift) \ ADD $const, index, R9; \ ADD R9, a; \ AND b, c, R9; \ ANDN b, d, R31; \ OR R9, R31, R9; \ ADD R9, a; \ ROTLW $shift, a; \ ADD b, a; #define ROUND2(a, b, c, d, index, const, shift) \ ADD $const, index, R9; \ ADD R9, a; \ AND b, d, R31; \ ANDN d, c, R9; \ OR R9, R31; \ ADD R31, a; \ ROTLW $shift, a; \ ADD b, a;
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 20 18:05:32 UTC 2024 - 5.3K bytes - Viewed (0) -
staging/src/k8s.io/apiextensions-apiserver/test/integration/fieldselector_test.go
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon May 06 15:53:03 UTC 2024 - 27K bytes - Viewed (0) -
src/cmd/vet/main.go
"golang.org/x/tools/go/analysis/passes/lostcancel" "golang.org/x/tools/go/analysis/passes/nilfunc" "golang.org/x/tools/go/analysis/passes/printf" "golang.org/x/tools/go/analysis/passes/shift" "golang.org/x/tools/go/analysis/passes/sigchanyzer" "golang.org/x/tools/go/analysis/passes/slog" "golang.org/x/tools/go/analysis/passes/stdmethods" "golang.org/x/tools/go/analysis/passes/stdversion"
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 14 19:41:17 UTC 2024 - 2.9K bytes - Viewed (0) -
src/crypto/internal/edwards25519/field/fe.go
} // Bits 0:51 (bytes 0:8, bits 0:64, shift 0, mask 51). v.l0 = byteorder.LeUint64(x[0:8]) v.l0 &= maskLow51Bits // Bits 51:102 (bytes 6:14, bits 48:112, shift 3, mask 51). v.l1 = byteorder.LeUint64(x[6:14]) >> 3 v.l1 &= maskLow51Bits // Bits 102:153 (bytes 12:20, bits 96:160, shift 6, mask 51). v.l2 = byteorder.LeUint64(x[12:20]) >> 6 v.l2 &= maskLow51Bits
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 13 18:57:38 UTC 2024 - 11.8K bytes - Viewed (0) -
src/internal/runtime/atomic/atomic_mips64x.s
MOVBU val+8(FP), R2 // Align ptr down to 4 bytes so we can use 32-bit load/store. MOVV $~3, R3 AND R1, R3 // Compute val shift. #ifdef GOARCH_mips64 // Big endian. ptr = ptr ^ 3 XOR $3, R1 #endif // R4 = ((ptr & 3) * 8) AND $3, R1, R4 SLLV $3, R4 // Shift val for aligned ptr. R2 = val << R4 SLLV R4, R2 SYNC LL (R3), R4 OR R2, R4 SC R4, (R3) BEQ R4, -4(PC) SYNC
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 11 21:29:34 UTC 2024 - 7.2K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/generic.rules
// Non-constant rotate detection. // We use shiftIsBounded to make sure that neither of the shifts are >64. // Note: these rules are subtle when the shift amounts are 0/64, as Go shifts // are different from most native shifts. But it works out.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 16 22:21:05 UTC 2024 - 135.3K bytes - Viewed (0) -
src/math/big/floatexample_test.go
// z = 1002.718282 (0x.faadf854p+10, prec = 32, acc = Below) } func ExampleFloat_shift() { // Implement Float "shift" by modifying the (binary) exponents directly. for s := -5; s <= 5; s++ { x := big.NewFloat(0.5) x.SetMantExp(x, x.MantExp(nil)+s) // shift x by s fmt.Println(x) } // Output: // 0.015625 // 0.03125 // 0.0625 // 0.125 // 0.25 // 0.5 // 1 // 2 // 4
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Jun 06 15:46:54 UTC 2024 - 4.1K bytes - Viewed (0)