- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 18 for cmpi (1.63 sec)
-
tensorflow/compiler/mlir/tfr/integration/tfr_decompose_ctx_test.cc
constexpr char tfr_raw_text[] = R"( tfr.func @tf__my_add_n(%values: !tfr.tensor_list, %n: i64 {tfr.name="N"}) -> !tfr.tensor { %index = arith.constant 0 : index %cst = arith.constant 1 : i64 %eq = arith.cmpi "eq", %n, %cst : i64 %v1 = tfr.get_element %values[%index] : (!tfr.tensor_list, index) -> !tfr.tensor %res = scf.if %eq -> !tfr.tensor { scf.yield %v1 : !tfr.tensor } else {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 5.8K bytes - Viewed (0) -
src/internal/bytealg/compare_ppc64x.s
RET PCALIGN $16 cmp4: // 4 - 7B CMP R9,$4 BLT cmp2 ANDCC $3,R9,R9 _LWBEX (R0)(R5),R10 _LWBEX (R0)(R6),R11 _LWBEX (R9)(R5),R12 _LWBEX (R9)(R6),R14 RLDIMI $32,R10,$0,R12 RLDIMI $32,R11,$0,R14 CMPU R12,R14 BR cmp0 PCALIGN $16 cmp2: // 2 - 3B CMP R9,$2 BLT cmp1 ANDCC $1,R9,R9 _LHBEX (R0)(R5),R10 _LHBEX (R0)(R6),R11 _LHBEX (R9)(R5),R12
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Aug 28 17:33:20 UTC 2023 - 6.7K bytes - Viewed (0) -
src/internal/bytealg/indexbyte_amd64.s
// Shuffle X0 around so that each byte contains // the character we're looking for. MOVD AX, X0 PUNPCKLBW X0, X0 PUNPCKLBW X0, X0 PSHUFL $0, X0, X0 CMPQ BX, $16 JLT small MOVQ SI, DI CMPQ BX, $32 JA avx2 sse: LEAQ -16(SI)(BX*1), AX // AX = address of last 16 bytes JMP sseloopentry PCALIGN $16 sseloop: // Move the next 16-byte chunk of the data into X1.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Nov 01 19:06:01 UTC 2023 - 3.1K bytes - Viewed (0) -
src/cmp/cmp.go
// Copyright 2023 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. // Package cmp provides types and functions related to comparing // ordered values. package cmp // Ordered is a constraint that permits any ordered type: any type // that supports the operators < <= >= >. // If future releases of Go add new ordered types,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 19 16:31:02 UTC 2024 - 2K bytes - Viewed (0) -
src/internal/bytealg/count_ppc64x.s
#else tail: // Count the last 0 - 31 bytes. CMP R4, $16 BLT tail_8 MOVD (R3), R12 MOVD 8(R3), R14 CMPB R12, R5, R12 CMPB R14, R5, R14 POPCNTD R12, R12 POPCNTD R14, R14 ADD R12, R18, R18 ADD R14, R18, R18 ADD $16, R3, R3 ADD $-16, R4, R4 tail_8: // Count the remaining 0 - 15 bytes. CMP R4, $8 BLT tail_4 MOVD (R3), R12 CMPB R12, R5, R12 POPCNTD R12, R12 ADD R12, R18, R18
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Aug 14 20:30:44 UTC 2023 - 3.6K bytes - Viewed (0) -
src/internal/bytealg/equal_amd64.s
// b in DI // count in BX // Output: // result in AX TEXT memeqbody<>(SB),NOSPLIT,$0-0 CMPQ BX, $8 JB small CMPQ BX, $64 JB bigloop #ifndef hasAVX2 CMPB internal∕cpu·X86+const_offsetX86HasAVX2(SB), $1 JE hugeloop_avx2 // 64 bytes at a time using xmm registers PCALIGN $16 hugeloop: CMPQ BX, $64 JB bigloop MOVOU (SI), X0 MOVOU (DI), X1 MOVOU 16(SI), X2 MOVOU 16(DI), X3
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Nov 17 16:34:40 UTC 2023 - 2.8K bytes - Viewed (0) -
src/syscall/asm_linux_386.s
TEXT ·rawVforkSyscall(SB),NOSPLIT|NOFRAME,$0-24 MOVL trap+0(FP), AX // syscall entry MOVL a1+4(FP), BX MOVL a2+8(FP), CX MOVL a3+12(FP), DX POPL SI // preserve return address INVOKE_SYSCALL PUSHL SI CMPL AX, $0xfffff001 JLS ok MOVL $-1, r1+16(FP) NEGL AX MOVL AX, err+20(FP) RET ok: MOVL AX, r1+16(FP) MOVL $0, err+20(FP) RET
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Sep 07 19:11:15 UTC 2023 - 3.1K bytes - Viewed (0) -
src/runtime/sys_windows_386.s
TEXT runtime·asmstdcall(SB),NOSPLIT,$0 MOVL fn+0(FP), BX MOVL SP, BP // save stack pointer // SetLastError(0). MOVL $0, 0x34(FS) MOVL libcall_n(BX), CX // Fast version, do not store args on the stack. CMPL CX, $0 JE docall // Copy args to the stack. MOVL CX, AX SALL $2, AX SUBL AX, SP // room for args MOVL SP, DI MOVL libcall_args(BX), SI CLD REP; MOVSL docall:
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Sep 21 15:56:43 UTC 2023 - 6.5K bytes - Viewed (0) -
src/internal/bytealg/count_amd64.s
#include "textflag.h" TEXT ·Count(SB),NOSPLIT,$0-40 #ifndef hasPOPCNT CMPB internal∕cpu·X86+const_offsetX86HasPOPCNT(SB), $1 JEQ 2(PC) JMP ·countGeneric(SB) #endif MOVQ b_base+0(FP), SI MOVQ b_len+8(FP), BX MOVB c+24(FP), AL LEAQ ret+32(FP), R8 JMP countbody<>(SB) TEXT ·CountString(SB),NOSPLIT,$0-32 #ifndef hasPOPCNT CMPB internal∕cpu·X86+const_offsetX86HasPOPCNT(SB), $1 JEQ 2(PC)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Oct 06 20:54:43 UTC 2023 - 4.7K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/looprotate.go
// to loops with a check-loop-condition-at-end. // This helps loops avoid extra unnecessary jumps. // // loop: // CMPQ ... // JGE exit // ... // JMP loop // exit: // // JMP entry // loop: // ... // entry: // CMPQ ... // JLT loop func loopRotate(f *Func) { loopnest := f.loopnest() if loopnest.hasIrreducible { return } if len(loopnest.loops) == 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 15 15:44:14 UTC 2024 - 3K bytes - Viewed (0)