Search Options

Results per page
Sort
Preferred Languages
Advance

Results 131 - 140 of 148 for MASK (0.07 sec)

  1. src/runtime/traceback.go

    		}
    		bits := *(*uint8)(add(liveInfo, uintptr(liveIdx)+uintptr(slotIdx/8)))
    		return bits&(1<<(slotIdx%8)) != 0
    	}
    
    	print1 := func(off, sz, slotIdx uint8) {
    		x := readUnaligned64(add(argp, uintptr(off)))
    		// mask out irrelevant bits
    		if sz < 8 {
    			shift := 64 - sz*8
    			if goarch.BigEndian {
    				x = x >> shift
    			} else {
    				x = x << shift >> shift
    			}
    		}
    		print(hex(x))
    		if !isLive(off, slotIdx) {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 55.1K bytes
    - Viewed (0)
  2. src/cmd/compile/internal/ssa/_gen/ARM64.rules

    // On ARM64 the bitmask can fit into an instruction.
    (SRLconst [c] (SLLconst [c] x)) && 0 < c && c < 64 => (ANDconst [1<<uint(64-c)-1] x) // mask out high bits
    (SLLconst [c] (SRLconst [c] x)) && 0 < c && c < 64 => (ANDconst [^(1<<uint(c)-1)] x) // mask out low bits
    
    // Special case setting bit as 1. An example is math.Copysign(c,-1)
    (ORconst [c1] (ANDconst [c2] x)) && c2|c1 == ^0  => (ORconst [c1] x)
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 15:49:20 UTC 2024
    - 113.1K bytes
    - Viewed (0)
  3. src/cmd/compile/internal/ssa/_gen/PPC64.rules

    (RotateLeft(32|64) ...) => ((ROTLW|ROTL) ...)
    
    // Constant rotate generation
    (ROTLW  x (MOVDconst [c])) => (ROTLWconst  x [c&31])
    (ROTL   x (MOVDconst [c])) => (ROTLconst   x [c&63])
    
    // Combine rotate and mask operations
    (ANDconst [m] (ROTLWconst [r] x)) && isPPC64WordRotateMask(m) => (RLWINM [encodePPC64RotateMask(r,m,32)] x)
    (AND (MOVDconst [m]) (ROTLWconst [r] x)) && isPPC64WordRotateMask(m) => (RLWINM [encodePPC64RotateMask(r,m,32)] x)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Jun 07 19:02:52 UTC 2024
    - 53.2K bytes
    - Viewed (0)
  4. src/runtime/asm_amd64.s

    	AESENC	runtime·aeskeysched+16(SB), X0
    	AESENC	runtime·aeskeysched+32(SB), X0
    	MOVQ	X0, AX	// return X0
    	RET
    noaes:
    	JMP	runtime·memhash64Fallback<ABIInternal>(SB)
    
    // simple mask to get rid of data in the high part of the register.
    DATA masks<>+0x00(SB)/8, $0x0000000000000000
    DATA masks<>+0x08(SB)/8, $0x0000000000000000
    DATA masks<>+0x10(SB)/8, $0x00000000000000ff
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Sat May 11 20:38:24 UTC 2024
    - 60.4K bytes
    - Viewed (0)
  5. src/cmd/internal/obj/arm64/asm7.go

    				}
    				o1 |= uint32(p.From.Index&0x1f) << 16
    			}
    		}
    		o1 |= uint32(p.To.Offset)
    		// cmd/asm/internal/arch/arm64.go:ARM64RegisterListOffset
    		// add opcode(bit 12-15) for vld1, mask it off if it's not vld1
    		o1 = c.maskOpvldvst(p, o1)
    		o1 |= uint32(r&31) << 5
    
    	case 82: /* vmov/vdup Rn, Vd.<T> */
    		rf := int(p.From.Reg)
    		rt := int(p.To.Reg)
    		o1 = 7<<25 | 3<<10
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 15:44:14 UTC 2024
    - 201.1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc

        Value n_plus_y = rewriter.create<AddOp>(loc, iotaN, y);
    
        // GatherOp is happy about letting us index out of bounds values, but those
        // values will be undefined. So we mask them later. Set up the boolean
        // expression that tells us which entries, in the output shape, are out of
        // bounds and thus become the padding_value.
        Value x_in_bounds = rewriter.create<AndOp>(
            loc,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 20:00:43 UTC 2024
    - 291.8K bytes
    - Viewed (0)
  7. RELEASE.md

     * If you use masked losses with Keras the loss values may be different in TensorFlow `2.12` compared to previous versions.
     * In cases where the mask was previously ignored, you will now get an error if you pass a mask with an incompatible shape.
    
    ### Major Features and Improvements
    
    `tf.keras`:
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
  8. src/cmd/vendor/github.com/ianlancetaylor/demangle/demangle.go

    	// the parsing of the AST, only the conversion of the AST
    	// to a string.
    	LLVMStyle
    )
    
    // maxLengthShift is how we shift the MaxLength value.
    const maxLengthShift = 16
    
    // maxLengthMask is a mask for the maxLength value.
    const maxLengthMask = 0x1f << maxLengthShift
    
    // MaxLength returns an Option that limits the maximum length of a
    // demangled string. The maximum length is expressed as a power of 2,
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 31 19:48:28 UTC 2024
    - 94.1K bytes
    - Viewed (0)
  9. CHANGELOG/CHANGELOG-1.27.md

    - Fixed a bug that caused the `apiserver` to panic when trying to allocate a Service with a dynamic `ClusterIP` and was configured with Service CIDRs with a /28 mask for IPv4 and a /124 mask for IPv6 ([#115322](https://github.com/kubernetes/kubernetes/pull/115322), [@aojea](https://github.com/aojea))
    - Fixed a bug where Kubernetes would apply a default StorageClass to a PersistentVolumeClaim,
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Tue Jun 11 23:01:06 UTC 2024
    - 455.3K bytes
    - Viewed (0)
  10. src/crypto/x509/x509_test.go

    	block, _ := pem.Decode([]byte(badIPMaskPEM))
    	_, err := ParseCertificate(block.Bytes)
    	if err == nil {
    		t.Fatalf("unexpected success")
    	}
    
    	const expected = "contained invalid mask"
    	if !strings.Contains(err.Error(), expected) {
    		t.Fatalf("expected %q in error but got: %s", expected, err)
    	}
    }
    
    const additionalGeneralSubtreePEM = `
    -----BEGIN CERTIFICATE-----
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 21:00:16 UTC 2024
    - 163.4K bytes
    - Viewed (0)
Back to top