Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for xmm1 (0.08 sec)

  1. src/cmd/asm/internal/asm/testdata/amd64enc_extra.s

    	// "VMOVQ r/m64, xmm1"/6E vs "VMOVQ xmm2/m64, xmm1"/7E with mem operand.
    	VMOVQ (AX), X20           // 62e1fd086e20 or 62e1fe087e20
    	VMOVQ 7(DX), X20          // 62e1fd086ea207000000 or 62e1fe087ea207000000
    	VMOVQ -15(R11)(CX*1), X20 // 62c1fd086ea40bf1ffffff or 62c1fe087ea40bf1ffffff
    	VMOVQ (SP)(AX*2), X20     // 62e1fd086e2444 or 62e1fe087e2444
    	// "VMOVQ xmm1, r/m64"/7E vs "VMOVQ xmm1, xmm2/m64"/D6 with mem operand.
    Registered: Tue Sep 09 11:13:09 UTC 2025
    - Last Modified: Thu Feb 20 11:20:03 UTC 2025
    - 57.7K bytes
    - Viewed (0)
  2. src/cmd/asm/internal/lex/lex_test.go

    		"\n.\n.MOVBLZX.(.8.*.4.).(.R12.).,.AX.\n.ADDB.AX.,.DX.\n",
    	},
    	{
    		"nested multiline macro",
    		lines(
    			"#define KEYROUND(xmm, load, off, r1, r2, index) \\",
    			"\tMOVBLZX	(BP)(DX*4),	R8 \\",
    			"\tload((off+1), r2) \\",
    			"\tMOVB	R8,		(off*4)(R12) \\",
    			"\tPINSRW	$index, (BP)(R8*4), xmm",
    			"#define LOAD(off, reg) \\",
    			"\tMOVBLZX	(off*4)(R12),	reg \\",
    			"\tADDB	reg,		DX",
    			"KEYROUND(X0, LOAD, 8, AX, BX, 0)",
    Registered: Tue Sep 09 11:13:09 UTC 2025
    - Last Modified: Tue Aug 29 07:48:38 UTC 2023
    - 5.8K bytes
    - Viewed (0)
  3. src/cmd/asm/internal/asm/parse.go

    //
    // For 386/AMD64 register list specifies 4VNNIW-style multi-source operand.
    // For range of 4 elements, Intel manual uses "+3" notation, for example:
    //
    //	VP4DPWSSDS zmm1{k1}{z}, zmm2+3, m128
    //
    // Given asm line:
    //
    //	VP4DPWSSDS Z5, [Z10-Z13], (AX)
    //
    // zmm2 is Z10, and Z13 is the only valid value for it (Z10+3).
    // Only simple ranges are accepted, like [Z0-Z3].
    Registered: Tue Sep 09 11:13:09 UTC 2025
    - Last Modified: Fri Feb 14 15:13:11 UTC 2025
    - 37.3K bytes
    - Viewed (0)
  4. lib/fips140/v1.0.0.zip

    H.lo·H^2.lo VPMSUMD IN1, H2L, XL1 // H^2.lo·H^2.lo VPMSUMD IN, H2, XM // H.hi·H^2.lo+H.lo·H^2.hi VPMSUMD IN1, H2, XM1 // H^2.hi·H^2.lo+H^2.lo·H^2.hi VPMSUMD IN, H2H, XH // H.hi·H^2.hi VPMSUMD IN1, H2H, XH1 // H^2.hi·H^2.hi VPMSUMD XL, XC2, T2 // 1st reduction phase VPMSUMD XL1, XC2, HH // 1st reduction phase VSLDOI $8, XM, ZERO, T0 VSLDOI $8, ZERO, XM, T1 VSLDOI $8, XM1, ZERO, HL VSLDOI $8, ZERO, XM1, H VXOR XL, T0, XL VXOR XH, T1, XH VXOR XL1, HL, XL1 VXOR XH1, H, XH1 VSLDOI $8, XL, XL, XL VSLDOI $8, XL1,...
    Registered: Tue Sep 09 11:13:09 UTC 2025
    - Last Modified: Wed Jan 29 15:10:35 UTC 2025
    - 635K bytes
    - Viewed (0)
Back to top