- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 217 for Fmask (0.07 sec)
-
src/runtime/signal_unix.go
// After this is called the thread can receive signals. func minitSignalMask() { nmask := getg().m.sigmask for i := range sigtable { if !blockableSig(uint32(i)) { sigdelset(&nmask, i) } } sigprocmask(_SIG_SETMASK, &nmask, nil) } // unminitSignals is called from dropm, via unminit, to undo the // effect of calling minit on a non-Go thread. // //go:nosplit func unminitSignals() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 10 16:04:54 UTC 2024 - 45K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewrite.go
} else if nbits == 32 { mb = bits.LeadingZeros32(uint32(mask)) me = 32 - bits.TrailingZeros32(uint32(mask)) mbn = bits.LeadingZeros32(^uint32(mask)) men = 32 - bits.TrailingZeros32(^uint32(mask)) } else { mb = bits.LeadingZeros64(uint64(mask)) me = 64 - bits.TrailingZeros64(uint64(mask)) mbn = bits.LeadingZeros64(^uint64(mask)) men = 64 - bits.TrailingZeros64(^uint64(mask)) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 64.2K bytes - Viewed (0) -
src/cmd/internal/obj/riscv/obj.go
ACPOPW & obj.AMask: rIIEncoding, ACTZ & obj.AMask: rIIEncoding, ACTZW & obj.AMask: rIIEncoding, AMAX & obj.AMask: rIIIEncoding, AMAXU & obj.AMask: rIIIEncoding, AMIN & obj.AMask: rIIIEncoding, AMINU & obj.AMask: rIIIEncoding, AORN & obj.AMask: rIIIEncoding, ASEXTB & obj.AMask: rIIEncoding, ASEXTH & obj.AMask: rIIEncoding, AXNOR & obj.AMask: rIIIEncoding,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sun Apr 07 03:32:27 UTC 2024 - 77K bytes - Viewed (0) -
src/sync/atomic/type.go
// And atomically performs a bitwise AND operation on x using the bitmask // provided as mask and returns the old value. func (x *Int32) And(mask int32) (old int32) { return AndInt32(&x.v, mask) } // Or atomically performs a bitwise OR operation on x using the bitmask // provided as mask and returns the old value. func (x *Int32) Or(mask int32) (old int32) { return OrInt32(&x.v, mask) } // An Int64 is an atomic int64. The zero value is zero.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 8.5K bytes - Viewed (0) -
src/net/ip.go
} } return true } // Mask returns the result of masking the IP address ip with mask. func (ip IP) Mask(mask IPMask) IP { if len(mask) == IPv6len && len(ip) == IPv4len && allFF(mask[:12]) { mask = mask[12:] } if len(mask) == IPv4len && len(ip) == IPv6len && bytealg.Equal(ip[:12], v4InV6Prefix) { ip = ip[12:] } n := len(ip) if n != len(mask) { return nil } out := make(IP, n)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 18 03:13:26 UTC 2024 - 13.9K bytes - Viewed (0) -
src/sync/atomic/atomic_test.go
i int32 after int32 } x.before = magic32 x.after = magic32 x.i = -1 j := x.i for mask := int32(1); mask != 0; mask <<= 1 { old := x.i k := AndInt32(&x.i, ^mask) j &= ^mask if x.i != j || k != old { t.Fatalf("mask=%d i=%d j=%d k=%d old=%d", mask, x.i, j, k, old) } } if x.before != magic32 || x.after != magic32 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 71.4K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/CompactHashSet.java
int hash = smearedHash(object); int mask = hashTableMask(); int tableIndex = hash & mask; int next = CompactHashing.tableGet(requireTable(), tableIndex); if (next == UNSET) { // uninitialized bucket if (newSize > mask) { // Resize and add new entry mask = resizeTable(mask, CompactHashing.newCapacity(mask), hash, newEntryIndex); } else {
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue May 28 18:11:09 UTC 2024 - 24K bytes - Viewed (0) -
cmd/kubeadm/app/apis/kubeadm/validation/validation_test.go
}{ // dual-stack: {"dual IPv4 only, but mask too small. Default node-mask", "10.0.0.16/29", nil, false}, {"dual IPv4 only, but mask too small. Configured node-mask", "10.0.0.16/24", []kubeadmapi.Arg{{Name: "node-cidr-mask-size-ipv4", Value: "23"}}, false}, {"dual IPv6 only, but mask too small. Default node-mask", "2001:db8::1/112", nil, false},
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed May 01 16:01:49 UTC 2024 - 46.1K bytes - Viewed (0) -
src/sync/atomic/doc.go
func AndUint32(addr *uint32, mask uint32) (old uint32) // AndInt64 atomically performs a bitwise AND operation on *addr using the bitmask provided as mask // and returns the old value. // Consider using the more ergonomic and less error-prone [Int64.And] instead. func AndInt64(addr *int64, mask int64) (old int64) // AndUint64 atomically performs a bitwise AND operation on *addr using the bitmask provided as mask // and returns the old.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 11.7K bytes - Viewed (0) -
src/cmd/internal/obj/x86/asm6.go
if pjc == 0 { return c } var toPad int32 fj, fjSize := fusedJump(p) mask := int32(pjc - 1) if fj { if (c&mask)+int32(fjSize) >= int32(pjc) { toPad = int32(pjc) - (c & mask) } } else if isJump(p) { if (c&mask)+int32(p.Isize) >= int32(pjc) { toPad = int32(pjc) - (c & mask) } } if toPad <= 0 { return c } return noppad(ctxt, s, c, toPad)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 15 15:44:14 UTC 2024 - 146.9K bytes - Viewed (0)