- Sort Score
- Result 10 results
- Languages All
Results 1 - 8 of 8 for vspltish (0.18 sec)
-
src/internal/bytealg/index_ppc64x.s
CMP R9, LASTBYTE // compare with last BGE index2to16 // 2 <= len(string) <= 16 MOVD $0xff00, R21 // Mask for later MTVSRD R21, V25 // Move to Vreg VSPLTH $3, V25, V31 // Splat mask VSPLTH $0, V0, V1 // Splat 1st 2 bytes of sep VSPLTISB $0, V10 // Clear V10 // First case: 2 byte separator // V1: 2 byte separator splatted // V2: 16 bytes at addr // V4: 16 bytes at addr+1
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 21 16:47:45 UTC 2023 - 31.6K bytes - Viewed (0) -
src/cmd/internal/obj/ppc64/a.out.go
AVCMPNEB AVCMPNEBCC AVCMPNEH AVCMPNEHCC AVCMPNEW AVCMPNEWCC AVPERM AVPERMXOR AVPERMR AVBPERMQ AVBPERMD AVSEL AVSPLTB AVSPLTH AVSPLTW AVSPLTISB AVSPLTISH AVSPLTISW AVCIPH AVCIPHER AVCIPHERLAST AVNCIPH AVNCIPHER AVNCIPHERLAST AVSBOX AVSHASIGMA AVSHASIGMAW AVSHASIGMAD AVMRGEW AVMRGOW AVCLZLSBB AVCTZLSBB
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 01 18:50:29 UTC 2024 - 16K bytes - Viewed (0) -
src/crypto/aes/gcm_ppc64x.s
MOVD $0x30, R10 LXVD2X (HTBL)(R0), VH // Load H VSPLTISB $-16, XC2 // 0xf0 VSPLTISB $1, T0 // one VADDUBM XC2, XC2, XC2 // 0xe0 VXOR ZERO, ZERO, ZERO VOR XC2, T0, XC2 // 0xe1 VSLDOI $15, XC2, ZERO, XC2 // 0xe1... VSLDOI $1, ZERO, T0, T1 // ...1 VADDUBM XC2, XC2, XC2 // 0xc2... VSPLTISB $7, T2 VOR XC2, T1, XC2 // 0xc2....01
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:29:44 UTC 2024 - 27.1K bytes - Viewed (0) -
src/hash/crc32/crc32_ppc64le.s
MOVD $80,off80 MOVD $96,off96 MOVD $112,off112 MOVD $0,R15 MOVD R3,R10 // save initial crc NOR R3,R3,R3 // ^crc MOVWZ R3,R3 // 32 bits VXOR zeroes,zeroes,zeroes // clear the V reg VSPLTISW $-1,V0 VSLDOI $4,V29,V0,mask_32bit VSLDOI $8,V29,V0,mask_64bit VXOR V8,V8,V8 MTVSRD R3,VS40 // crc initial value VS40 = V8 #ifdef REFLECT
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 06 12:09:50 UTC 2024 - 13.1K bytes - Viewed (0) -
src/crypto/aes/asm_ppc64x.s
MOVD $4, CNT // li 7,4 STXVD2X IN0, (R0+OUTENC) STXVD2X IN0, (R0+OUTDEC) ADD $16, OUTENC, OUTENC ADD $-16, OUTDEC, OUTDEC VSPLTISB $8, KEY // vspltisb 3,8 MOVD CNT, CTR // mtctr 7 VSUBUBM MASK, KEY, MASK // vsububm 5,5,3 loop192: VPERM IN1, IN1, MASK, KEY // vperm 3,2,2,5
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 20 18:05:32 UTC 2024 - 18.6K bytes - Viewed (0) -
src/crypto/sha512/sha512block_ppc64x.s
MOVD $0x0e0, R_x0e0 MOVD $0x0f0, R_x0f0 MOVD $0x100, R_x100 MOVD $0x110, R_x110 #ifdef GOARCH_ppc64le // Generate the mask used with VPERM for LE MOVWZ $8, TEMP LVSL (TEMP)(R0), LEMASK VSPLTISB $0x0F, KI VXOR KI, LEMASK, LEMASK #endif LXVD2X (CTX)(R_x000), VS32 // v0 = vs32 LXVD2X (CTX)(R_x010), VS34 // v2 = vs34 LXVD2X (CTX)(R_x020), VS36 // v4 = vs36
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:29:44 UTC 2024 - 15.8K bytes - Viewed (0) -
src/crypto/sha256/sha256block_ppc64x.s
SLD $6, LEN ADD INP, LEN, END CMP INP, END BEQ end MOVD $·kcon(SB), TBL_STRT MOVD $0x10, R_x010 #ifdef GOARCH_ppc64le MOVWZ $8, TEMP LVSL (TEMP)(R0), LEMASK VSPLTISB $0x0F, KI VXOR KI, LEMASK, LEMASK #endif LXVW4X (CTX)(R_x000), V0 LXVW4X (CTX)(R_x010), V4 // unpack the input values into vector registers VSLDOI $4, V0, V0, V1 VSLDOI $8, V0, V0, V2
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:29:44 UTC 2024 - 14.4K bytes - Viewed (0) -
src/cmd/internal/notsha256/sha256block_ppc64x.s
SLD $6, LEN ADD INP, LEN, END CMP INP, END BEQ end MOVD $·kcon(SB), TBL_STRT MOVD $0x10, R_x010 #ifdef GOARCH_ppc64le MOVWZ $8, TEMP LVSL (TEMP)(R0), LEMASK VSPLTISB $0x0F, KI VXOR KI, LEMASK, LEMASK #endif LXVW4X (CTX)(R_x000), V0 LXVW4X (CTX)(R_x010), V4 // unpack the input values into vector registers VSLDOI $4, V0, V0, V1 VSLDOI $8, V0, V0, V2
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:27 UTC 2023 - 14.5K bytes - Viewed (0)