- Sort Score
- Result 10 results
- Languages All
Results 101 - 110 of 518 for MASK (0.04 sec)
-
src/cmd/compile/internal/ssa/regalloc.go
} // allocReg chooses a register from the set of registers in mask. // If there is no unused register, a Value will be kicked out of // a register to make room. func (s *regAllocState) allocReg(mask regMask, v *Value) register { if v.OnWasmStack { return noRegister } mask &= s.allocatable mask &^= s.nospill if mask == 0 { s.f.Fatalf("no register available for %s", v.LongString()) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 21 17:49:56 UTC 2023 - 87.2K bytes - Viewed (0) -
pkg/kubelet/cm/memorymanager/policy_static.go
hints := map[string][]topologymanager.TopologyHint{} bitmask.IterateBitMasks(numaNodes, func(mask bitmask.BitMask) { maskBits := mask.GetBits() singleNUMAHint := len(maskBits) == 1 totalFreeSize := map[v1.ResourceName]uint64{} totalAllocatableSize := map[v1.ResourceName]uint64{} // calculate total free and allocatable memory for the node mask for _, nodeID := range maskBits { for resourceName := range requestedResources {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sun Nov 12 07:34:55 UTC 2023 - 34K bytes - Viewed (0) -
pkg/volume/volume_linux.go
// is not what we want; thus, we skip chmod for symlinks. if info.Mode()&os.ModeSymlink != 0 { return nil } mask := rwMask if readonly { mask = roMask } if info.IsDir() { mask |= os.ModeSetgid mask |= execMask } err = os.Chmod(filename, info.Mode()|mask) if err != nil { klog.ErrorS(err, "chmod failed", "path", filename) } return nil }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Apr 03 19:34:37 UTC 2023 - 6.5K bytes - Viewed (0) -
okhttp/src/main/kotlin/okhttp3/internal/ws/WebSocketProtocol.kt
internal const val B0_FLAG_RSV3 = 16 /** Byte 0 mask for the frame opcode. */ internal const val B0_MASK_OPCODE = 15 /** Flag in the opcode which indicates a control frame. */ internal const val OPCODE_FLAG_CONTROL = 8 /** * Byte 1 flag for whether the payload data is masked. * * If this flag is set, the next four
Registered: Sun Jun 16 04:42:17 UTC 2024 - Last Modified: Mon Jan 08 01:13:22 UTC 2024 - 4.8K bytes - Viewed (0) -
src/math/sqrt.go
return x case x < 0: return NaN() } ix := Float64bits(x) // normalize x exp := int((ix >> shift) & mask) if exp == 0 { // subnormal x for ix&(1<<shift) == 0 { ix <<= 1 exp-- } exp++ } exp -= bias // unbias exponent ix &^= mask << shift ix |= 1 << shift if exp&1 == 1 { // odd exp, double x to make it even ix <<= 1 }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Aug 15 17:07:57 UTC 2022 - 4.8K bytes - Viewed (0) -
src/cmd/internal/obj/riscv/cpu.go
// ITypeImmMask is a mask including only the immediate portion of // I-type instructions. ITypeImmMask = 0xfff00000 // JTypeImmMask is a mask including only the immediate portion of // J-type instructions. JTypeImmMask = 0xfffff000 // STypeImmMask is a mask including only the immediate portion of // S-type instructions. STypeImmMask = 0xfe000f80 // UTypeImmMask is a mask including only the immediate portion of
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Mar 20 14:19:33 UTC 2024 - 13.1K bytes - Viewed (0) -
src/runtime/mpagecache.go
func (c *pageCache) allocN(npages uintptr) (uintptr, uintptr) { i := findBitRange64(c.cache, uint(npages)) if i >= 64 { return 0, 0 } mask := ((uint64(1) << npages) - 1) << i scav := sys.OnesCount64(c.scav & mask) c.cache &^= mask // mark in-use bits c.scav &^= mask // clear scavenged bits return c.base + uintptr(i*pageSize), uintptr(scav) * pageSize } // flush empties out unallocated free pages in the given cache
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Apr 19 14:30:00 UTC 2023 - 5.6K bytes - Viewed (0) -
src/runtime/race.go
//go:linkname abigen_sync_atomic_OrInt64 sync/atomic.OrInt64 func abigen_sync_atomic_OrInt64(addr *int64, mask int64) (old int64) //go:linkname abigen_sync_atomic_OrUint64 sync/atomic.OrUint64 func abigen_sync_atomic_OrUint64(addr *uint64, mask uint64) (old uint64) //go:linkname abigen_sync_atomic_OrUintptr sync/atomic.OrUintptr func abigen_sync_atomic_OrUintptr(addr *uintptr, mask uintptr) (old uintptr)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 20.4K bytes - Viewed (0) -
src/os/signal/doc.go
the SIGPROF signal in particular. The non-Go code should not change the signal mask on any threads created by the Go runtime. If the non-Go code starts new threads itself, those threads may set the signal mask as they please. If the non-Go code starts a new thread, changes the signal mask, and then invokes a Go function in that thread, the Go runtime will
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 18:11:00 UTC 2024 - 11K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/arch/arm/armasm/decode.go
// Considering only the top 4 bits, the conditional instructions use mask=0, value=0, // while the unconditional instructions use mask=f, value=f. // Prepare a version of x with the condition cleared to 0 in conditional instructions // and then assume mask=f during matching. const condMask = 0xf0000000 xNoCond := x if x&condMask != condMask { xNoCond &^= condMask }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 22 17:16:14 UTC 2022 - 12.6K bytes - Viewed (0)