Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 48 for b_len (0.04 sec)

  1. src/cmd/asm/internal/asm/operand_test.go

    	{"_seek<>(SB)", "_seek<>(SB)"},
    	{"a2+16(FP)", "a2+16(FP)"},
    	{"addr2+24(FP)", "addr2+24(FP)"},
    	{"asmcgocall<>(SB)", "asmcgocall<>(SB)"},
    	{"b+24(FP)", "b+24(FP)"},
    	{"b_len+32(FP)", "b_len+32(FP)"},
    	{"racecall<>(SB)", "racecall<>(SB)"},
    	{"rcv_name+20(FP)", "rcv_name+20(FP)"},
    	{"retoffset+28(FP)", "retoffset+28(FP)"},
    	{"runtime·_GetStdHandle(SB)", "runtime._GetStdHandle(SB)"},
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Aug 29 18:31:05 UTC 2023
    - 23.9K bytes
    - Viewed (0)
  2. src/math/big/arith_ppc64x.s

    	MOVD y+48(FP), R4	// R4 = y = c
    	MOVD z_len+8(FP), R11	// R11 = z_len
    
    	CMP   R11, $0		// If z_len is zero, return
    	BEQ   done
    
    	// We will process the first iteration out of the loop so we capture
    	// the value of c. In the subsequent iterations, we will rely on the
    	// value of CA set here.
    	MOVD  0(R8), R20	// R20 = x[i]
    	ADD   $-1, R11		// R11 = z_len - 1
    	ADDC  R20, R4, R6	// R6 = x[i] + c
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 18:17:17 UTC 2024
    - 16.8K bytes
    - Viewed (0)
  3. src/math/big/arith_386.s

    	MOVL y+24(FP), AX	// c = y
    	MOVL z_len+4(FP), BP
    	MOVL $0, BX		// i = 0
    	JMP E4
    
    L4:	MOVL (SI)(BX*4), DX
    	SUBL AX, DX
    	MOVL DX, (DI)(BX*4)
    	SBBL AX, AX		// save CF
    	NEGL AX
    	ADDL $1, BX		// i++
    
    E4:	CMPL BX, BP		// i < n
    	JL L4
    
    	MOVL AX, c+28(FP)
    	RET
    
    
    // func shlVU(z, x []Word, s uint) (c Word)
    TEXT ·shlVU(SB),NOSPLIT,$0
    	MOVL z_len+4(FP), BX	// i = z
    	SUBL $1, BX		// i--
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 23:33:27 UTC 2023
    - 4K bytes
    - Viewed (0)
  4. src/math/big/arith_arm.s

    // arithmetic operations on vectors implemented in arith.go.
    
    // func addVV(z, x, y []Word) (c Word)
    TEXT ·addVV(SB),NOSPLIT,$0
    	ADD.S	$0, R0		// clear carry flag
    	MOVW	z+0(FP), R1
    	MOVW	z_len+4(FP), R4
    	MOVW	x+12(FP), R2
    	MOVW	y+24(FP), R3
    	ADD	R4<<2, R1, R4
    	B E1
    L1:
    	MOVW.P	4(R2), R5
    	MOVW.P	4(R3), R6
    	ADC.S	R6, R5
    	MOVW.P	R5, 4(R1)
    E1:
    	TEQ	R1, R4
    	BNE L1
    
    	MOVW	$0, R0
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 23:33:27 UTC 2023
    - 4K bytes
    - Viewed (0)
  5. src/crypto/internal/bigmod/nat_ppc64x.s

    TEXT ·addMulVVW1024(SB), $0-32
    	MOVD	$4, R6 // R6 = z_len/4
    	JMP		addMulVVWx<>(SB)
    
    // func addMulVVW1536(z, x *uint, y uint) (c uint)
    TEXT ·addMulVVW1536(SB), $0-32
    	MOVD	$6, R6 // R6 = z_len/4
    	JMP		addMulVVWx<>(SB)
    
    // func addMulVVW2048(z, x *uint, y uint) (c uint)
    TEXT ·addMulVVW2048(SB), $0-32
    	MOVD	$8, R6 // R6 = z_len/4
    	JMP		addMulVVWx<>(SB)
    
    // This local function expects to be called only by
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Jan 25 19:32:43 UTC 2024
    - 1.9K bytes
    - Viewed (0)
  6. src/math/big/arith_amd64.s

    // It is restored with ADDQ Rx, Rx: if Rx was -1 the carry is set, otherwise it is cleared.
    // This is faster than using rotate instructions.
    
    // func addVV(z, x, y []Word) (c Word)
    TEXT ·addVV(SB),NOSPLIT,$0
    	MOVQ z_len+8(FP), DI
    	MOVQ x+24(FP), R8
    	MOVQ y+48(FP), R9
    	MOVQ z+0(FP), R10
    
    	MOVQ $0, CX		// c = 0
    	MOVQ $0, SI		// i = 0
    
    	// s/JL/JMP/ below to disable the unrolled loop
    	SUBQ $4, DI		// n -= 4
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 23:33:27 UTC 2023
    - 9.1K bytes
    - Viewed (0)
  7. src/math/big/arith_arm64.s

    // TODO: Consider re-implementing using Advanced SIMD
    // once the assembler supports those instructions.
    
    // func addVV(z, x, y []Word) (c Word)
    TEXT ·addVV(SB),NOSPLIT,$0
    	MOVD	z_len+8(FP), R0
    	MOVD	x+24(FP), R8
    	MOVD	y+48(FP), R9
    	MOVD	z+0(FP), R10
    	ADDS	$0, R0		// clear carry flag
    	TBZ	$0, R0, two
    	MOVD.P	8(R8), R11
    	MOVD.P	8(R9), R15
    	ADCS	R15, R11
    	MOVD.P	R11, 8(R10)
    	SUB	$1, R0
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 23:33:27 UTC 2023
    - 11.8K bytes
    - Viewed (0)
  8. src/math/big/arith_s390x.s

    	MOVD R2, 0(R1)
    
    	// MOVD	$·addVV_vec(SB), 0(R1)
    	BR ·addVV_vec(SB)
    
    GLOBL addvectorfacility+0x00(SB), NOPTR, $8
    DATA addvectorfacility+0x00(SB)/8, $·addVV_check(SB)
    
    TEXT ·addVV_vec(SB), NOSPLIT, $0
    	MOVD z_len+8(FP), R3
    	MOVD x+24(FP), R8
    	MOVD y+48(FP), R9
    	MOVD z+0(FP), R2
    
    	MOVD $0, R4  // c = 0
    	MOVD $0, R0  // make sure it's zero
    	MOVD $0, R10 // i = 0
    
    	// s/JL/JMP/ below to disable the unrolled loop
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 23:33:27 UTC 2023
    - 20.3K bytes
    - Viewed (0)
  9. src/crypto/md5/md5block_ppc64x.s

    	ADD	R9, a; \
    	ORN     d, b, R31; \
    	XOR	c, R31; \
    	ADD	R31, a; \
    	ROTLW	$shift, a; \
    	ADD	b, a;
    
    
    TEXT ·block(SB),NOSPLIT,$0-32
    	MOVD	dig+0(FP), R10
    	MOVD	p+8(FP), R6
    	MOVD	p_len+16(FP), R5
    
    	// We assume p_len >= 64
    	SRD 	$6, R5
    	MOVD	R5, CTR
    
    	MOVWZ	0(R10), R22
    	MOVWZ	4(R10), R3
    	MOVWZ	8(R10), R4
    	MOVWZ	12(R10), R5
    
    loop:
    	MOVD	R22, R14
    	MOVD	R3, R15
    	MOVD	R4, R16
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon May 20 18:05:32 UTC 2024
    - 5.3K bytes
    - Viewed (0)
  10. src/math/big/rat.go

    		Emin  = 1 - Ebias
    		Emax  = Ebias
    	)
    
    	// TODO(adonovan): specialize common degenerate cases: 1.0, integers.
    	alen := a.bitLen()
    	if alen == 0 {
    		return 0, true
    	}
    	blen := b.bitLen()
    	if blen == 0 {
    		panic("division by zero")
    	}
    
    	// 1. Left-shift A or B such that quotient A/B is in [1<<Msize1, 1<<(Msize2+1)
    	// (Msize2 bits if A < B when they are left-aligned, Msize2+1 bits if A >= B).
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 11:59:09 UTC 2023
    - 13.5K bytes
    - Viewed (0)
Back to top