Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 8 of 8 for vspltish (0.18 sec)

  1. src/internal/bytealg/index_ppc64x.s

    	CMP      R9, LASTBYTE // compare with last
    	BGE      index2to16   // 2 <= len(string) <= 16
    	MOVD     $0xff00, R21 // Mask for later
    	MTVSRD   R21, V25     // Move to Vreg
    	VSPLTH   $3, V25, V31 // Splat mask
    	VSPLTH   $0, V0, V1   // Splat 1st 2 bytes of sep
    	VSPLTISB $0, V10      // Clear V10
    
    	// First case: 2 byte separator
    	// V1: 2 byte separator splatted
    	// V2: 16 bytes at addr
    	// V4: 16 bytes at addr+1
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Apr 21 16:47:45 UTC 2023
    - 31.6K bytes
    - Viewed (0)
  2. src/cmd/internal/obj/ppc64/a.out.go

    	AVCMPNEB
    	AVCMPNEBCC
    	AVCMPNEH
    	AVCMPNEHCC
    	AVCMPNEW
    	AVCMPNEWCC
    	AVPERM
    	AVPERMXOR
    	AVPERMR
    	AVBPERMQ
    	AVBPERMD
    	AVSEL
    	AVSPLTB
    	AVSPLTH
    	AVSPLTW
    	AVSPLTISB
    	AVSPLTISH
    	AVSPLTISW
    	AVCIPH
    	AVCIPHER
    	AVCIPHERLAST
    	AVNCIPH
    	AVNCIPHER
    	AVNCIPHERLAST
    	AVSBOX
    	AVSHASIGMA
    	AVSHASIGMAW
    	AVSHASIGMAD
    	AVMRGEW
    	AVMRGOW
    	AVCLZLSBB
    	AVCTZLSBB
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Apr 01 18:50:29 UTC 2024
    - 16K bytes
    - Viewed (0)
  3. src/crypto/aes/gcm_ppc64x.s

    	MOVD   $0x30, R10
    	LXVD2X (HTBL)(R0), VH // Load H
    
    	VSPLTISB $-16, XC2           // 0xf0
    	VSPLTISB $1, T0              // one
    	VADDUBM  XC2, XC2, XC2       // 0xe0
    	VXOR     ZERO, ZERO, ZERO
    	VOR      XC2, T0, XC2        // 0xe1
    	VSLDOI   $15, XC2, ZERO, XC2 // 0xe1...
    	VSLDOI   $1, ZERO, T0, T1    // ...1
    	VADDUBM  XC2, XC2, XC2       // 0xc2...
    	VSPLTISB $7, T2
    	VOR      XC2, T1, XC2        // 0xc2....01
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 04 17:29:44 UTC 2024
    - 27.1K bytes
    - Viewed (0)
  4. src/hash/crc32/crc32_ppc64le.s

    	MOVD	$80,off80
    	MOVD	$96,off96
    	MOVD	$112,off112
    	MOVD	$0,R15
    
    	MOVD	R3,R10	// save initial crc
    
    	NOR	R3,R3,R3  // ^crc
    	MOVWZ	R3,R3	// 32 bits
    	VXOR	zeroes,zeroes,zeroes  // clear the V reg
    	VSPLTISW $-1,V0
    	VSLDOI	$4,V29,V0,mask_32bit
    	VSLDOI	$8,V29,V0,mask_64bit
    
    	VXOR	V8,V8,V8
    	MTVSRD	R3,VS40	// crc initial value VS40 = V8
    
    #ifdef REFLECT
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon May 06 12:09:50 UTC 2024
    - 13.1K bytes
    - Viewed (0)
  5. src/crypto/aes/asm_ppc64x.s

    	MOVD	$4, CNT                          // li 7,4
    	STXVD2X	IN0, (R0+OUTENC)
    	STXVD2X	IN0, (R0+OUTDEC)
    	ADD	$16, OUTENC, OUTENC
    	ADD	$-16, OUTDEC, OUTDEC
    	VSPLTISB	$8, KEY                  // vspltisb 3,8
    	MOVD	CNT, CTR                         // mtctr 7
    	VSUBUBM	MASK, KEY, MASK                  // vsububm 5,5,3
    
    loop192:
    	VPERM	IN1, IN1, MASK, KEY // vperm 3,2,2,5
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon May 20 18:05:32 UTC 2024
    - 18.6K bytes
    - Viewed (0)
  6. src/crypto/sha512/sha512block_ppc64x.s

    	MOVD	$0x0e0, R_x0e0
    	MOVD	$0x0f0, R_x0f0
    	MOVD	$0x100, R_x100
    	MOVD	$0x110, R_x110
    
    
    #ifdef GOARCH_ppc64le
    	// Generate the mask used with VPERM for LE
    	MOVWZ	$8, TEMP
    	LVSL	(TEMP)(R0), LEMASK
    	VSPLTISB	$0x0F, KI
    	VXOR	KI, LEMASK, LEMASK
    #endif
    
    	LXVD2X	(CTX)(R_x000), VS32	// v0 = vs32
    	LXVD2X	(CTX)(R_x010), VS34	// v2 = vs34
    	LXVD2X	(CTX)(R_x020), VS36	// v4 = vs36
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 04 17:29:44 UTC 2024
    - 15.8K bytes
    - Viewed (0)
  7. src/crypto/sha256/sha256block_ppc64x.s

    	SLD	$6, LEN
    	ADD	INP, LEN, END
    
    	CMP	INP, END
    	BEQ	end
    
    	MOVD	$·kcon(SB), TBL_STRT
    	MOVD	$0x10, R_x010
    
    #ifdef GOARCH_ppc64le
    	MOVWZ	$8, TEMP
    	LVSL	(TEMP)(R0), LEMASK
    	VSPLTISB	$0x0F, KI
    	VXOR	KI, LEMASK, LEMASK
    #endif
    
    	LXVW4X	(CTX)(R_x000), V0
    	LXVW4X	(CTX)(R_x010), V4
    
    	// unpack the input values into vector registers
    	VSLDOI	$4, V0, V0, V1
    	VSLDOI	$8, V0, V0, V2
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 04 17:29:44 UTC 2024
    - 14.4K bytes
    - Viewed (0)
  8. src/cmd/internal/notsha256/sha256block_ppc64x.s

    	SLD	$6, LEN
    	ADD	INP, LEN, END
    
    	CMP	INP, END
    	BEQ	end
    
    	MOVD	$·kcon(SB), TBL_STRT
    	MOVD	$0x10, R_x010
    
    #ifdef GOARCH_ppc64le
    	MOVWZ	$8, TEMP
    	LVSL	(TEMP)(R0), LEMASK
    	VSPLTISB	$0x0F, KI
    	VXOR	KI, LEMASK, LEMASK
    #endif
    
    	LXVW4X	(CTX)(R_x000), V0
    	LXVW4X	(CTX)(R_x010), V4
    
    	// unpack the input values into vector registers
    	VSLDOI	$4, V0, V0, V1
    	VSLDOI	$8, V0, V0, V2
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 23:33:27 UTC 2023
    - 14.5K bytes
    - Viewed (0)
Back to top