- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 48 for b_len (0.04 sec)
-
src/cmd/asm/internal/asm/operand_test.go
{"_seek<>(SB)", "_seek<>(SB)"}, {"a2+16(FP)", "a2+16(FP)"}, {"addr2+24(FP)", "addr2+24(FP)"}, {"asmcgocall<>(SB)", "asmcgocall<>(SB)"}, {"b+24(FP)", "b+24(FP)"}, {"b_len+32(FP)", "b_len+32(FP)"}, {"racecall<>(SB)", "racecall<>(SB)"}, {"rcv_name+20(FP)", "rcv_name+20(FP)"}, {"retoffset+28(FP)", "retoffset+28(FP)"}, {"runtime·_GetStdHandle(SB)", "runtime._GetStdHandle(SB)"},
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 29 18:31:05 UTC 2023 - 23.9K bytes - Viewed (0) -
src/math/big/arith_ppc64x.s
MOVD y+48(FP), R4 // R4 = y = c MOVD z_len+8(FP), R11 // R11 = z_len CMP R11, $0 // If z_len is zero, return BEQ done // We will process the first iteration out of the loop so we capture // the value of c. In the subsequent iterations, we will rely on the // value of CA set here. MOVD 0(R8), R20 // R20 = x[i] ADD $-1, R11 // R11 = z_len - 1 ADDC R20, R4, R6 // R6 = x[i] + c
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 18:17:17 UTC 2024 - 16.8K bytes - Viewed (0) -
src/math/big/arith_386.s
MOVL y+24(FP), AX // c = y MOVL z_len+4(FP), BP MOVL $0, BX // i = 0 JMP E4 L4: MOVL (SI)(BX*4), DX SUBL AX, DX MOVL DX, (DI)(BX*4) SBBL AX, AX // save CF NEGL AX ADDL $1, BX // i++ E4: CMPL BX, BP // i < n JL L4 MOVL AX, c+28(FP) RET // func shlVU(z, x []Word, s uint) (c Word) TEXT ·shlVU(SB),NOSPLIT,$0 MOVL z_len+4(FP), BX // i = z SUBL $1, BX // i--
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:27 UTC 2023 - 4K bytes - Viewed (0) -
src/math/big/arith_arm.s
// arithmetic operations on vectors implemented in arith.go. // func addVV(z, x, y []Word) (c Word) TEXT ·addVV(SB),NOSPLIT,$0 ADD.S $0, R0 // clear carry flag MOVW z+0(FP), R1 MOVW z_len+4(FP), R4 MOVW x+12(FP), R2 MOVW y+24(FP), R3 ADD R4<<2, R1, R4 B E1 L1: MOVW.P 4(R2), R5 MOVW.P 4(R3), R6 ADC.S R6, R5 MOVW.P R5, 4(R1) E1: TEQ R1, R4 BNE L1 MOVW $0, R0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:27 UTC 2023 - 4K bytes - Viewed (0) -
src/crypto/internal/bigmod/nat_ppc64x.s
TEXT ·addMulVVW1024(SB), $0-32 MOVD $4, R6 // R6 = z_len/4 JMP addMulVVWx<>(SB) // func addMulVVW1536(z, x *uint, y uint) (c uint) TEXT ·addMulVVW1536(SB), $0-32 MOVD $6, R6 // R6 = z_len/4 JMP addMulVVWx<>(SB) // func addMulVVW2048(z, x *uint, y uint) (c uint) TEXT ·addMulVVW2048(SB), $0-32 MOVD $8, R6 // R6 = z_len/4 JMP addMulVVWx<>(SB) // This local function expects to be called only by
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Jan 25 19:32:43 UTC 2024 - 1.9K bytes - Viewed (0) -
src/math/big/arith_amd64.s
// It is restored with ADDQ Rx, Rx: if Rx was -1 the carry is set, otherwise it is cleared. // This is faster than using rotate instructions. // func addVV(z, x, y []Word) (c Word) TEXT ·addVV(SB),NOSPLIT,$0 MOVQ z_len+8(FP), DI MOVQ x+24(FP), R8 MOVQ y+48(FP), R9 MOVQ z+0(FP), R10 MOVQ $0, CX // c = 0 MOVQ $0, SI // i = 0 // s/JL/JMP/ below to disable the unrolled loop SUBQ $4, DI // n -= 4
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:27 UTC 2023 - 9.1K bytes - Viewed (0) -
src/math/big/arith_arm64.s
// TODO: Consider re-implementing using Advanced SIMD // once the assembler supports those instructions. // func addVV(z, x, y []Word) (c Word) TEXT ·addVV(SB),NOSPLIT,$0 MOVD z_len+8(FP), R0 MOVD x+24(FP), R8 MOVD y+48(FP), R9 MOVD z+0(FP), R10 ADDS $0, R0 // clear carry flag TBZ $0, R0, two MOVD.P 8(R8), R11 MOVD.P 8(R9), R15 ADCS R15, R11 MOVD.P R11, 8(R10) SUB $1, R0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:27 UTC 2023 - 11.8K bytes - Viewed (0) -
src/math/big/arith_s390x.s
MOVD R2, 0(R1) // MOVD $·addVV_vec(SB), 0(R1) BR ·addVV_vec(SB) GLOBL addvectorfacility+0x00(SB), NOPTR, $8 DATA addvectorfacility+0x00(SB)/8, $·addVV_check(SB) TEXT ·addVV_vec(SB), NOSPLIT, $0 MOVD z_len+8(FP), R3 MOVD x+24(FP), R8 MOVD y+48(FP), R9 MOVD z+0(FP), R2 MOVD $0, R4 // c = 0 MOVD $0, R0 // make sure it's zero MOVD $0, R10 // i = 0 // s/JL/JMP/ below to disable the unrolled loop
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:27 UTC 2023 - 20.3K bytes - Viewed (0) -
src/crypto/md5/md5block_ppc64x.s
ADD R9, a; \ ORN d, b, R31; \ XOR c, R31; \ ADD R31, a; \ ROTLW $shift, a; \ ADD b, a; TEXT ·block(SB),NOSPLIT,$0-32 MOVD dig+0(FP), R10 MOVD p+8(FP), R6 MOVD p_len+16(FP), R5 // We assume p_len >= 64 SRD $6, R5 MOVD R5, CTR MOVWZ 0(R10), R22 MOVWZ 4(R10), R3 MOVWZ 8(R10), R4 MOVWZ 12(R10), R5 loop: MOVD R22, R14 MOVD R3, R15 MOVD R4, R16
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 20 18:05:32 UTC 2024 - 5.3K bytes - Viewed (0) -
src/math/big/rat.go
Emin = 1 - Ebias Emax = Ebias ) // TODO(adonovan): specialize common degenerate cases: 1.0, integers. alen := a.bitLen() if alen == 0 { return 0, true } blen := b.bitLen() if blen == 0 { panic("division by zero") } // 1. Left-shift A or B such that quotient A/B is in [1<<Msize1, 1<<(Msize2+1) // (Msize2 bits if A < B when they are left-aligned, Msize2+1 bits if A >= B).
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 11:59:09 UTC 2023 - 13.5K bytes - Viewed (0)