- Sort Score
- Result 10 results
- Languages All
Results 71 - 80 of 451 for aligned (0.11 sec)
-
src/runtime/tagptr_64bit.go
// get to really high addresses and panic if it does. addrBits = 48 // In addition to the 16 bits taken from the top, we can take 3 from the // bottom, because node must be pointer-aligned, giving a total of 19 bits // of count. tagBits = 64 - addrBits + 3 // On AIX, 64-bit addresses are split into 36-bit segment number and 28-bit
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 18 20:22:50 UTC 2023 - 3.2K bytes - Viewed (0) -
pkg/kubelet/cm/cpumanager/policy_options.go
// Flag to evenly distribute CPUs across NUMA nodes in cases where more // than one NUMA node is required to satisfy the allocation. DistributeCPUsAcrossNUMA bool // Flag to ensure CPUs are considered aligned at socket boundary rather than // NUMA boundary AlignBySocket bool } // NewStaticPolicyOptions creates a StaticPolicyOptions struct from the user configuration.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Sep 27 13:02:15 UTC 2023 - 5.1K bytes - Viewed (0) -
src/math/rand/example_test.go
// Typically a non-fixed seed should be used, such as time.Now().UnixNano(). // Using a fixed seed will produce the same output on every run. r := rand.New(rand.NewSource(99)) // The tabwriter here helps us generate aligned output. w := tabwriter.NewWriter(os.Stdout, 1, 1, 1, ' ', 0) defer w.Flush() show := func(name string, v1, v2, v3 any) { fmt.Fprintf(w, "%s\t%v\t%v\t%v\n", name, v1, v2, v3) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Oct 26 16:24:57 UTC 2022 - 4.2K bytes - Viewed (0) -
src/runtime/memmove_amd64.s
// Algorithm: // 1. Unaligned save of the tail's 128 bytes // 2. Unaligned save of the head's 32 bytes // 3. Destination-aligned copying of body (128 bytes per iteration) // 4. Put head on the new place // 5. Put the tail on the new place // It can be important to satisfy processor's pipeline requirements for // small sizes as the cost of unaligned memory region copying is
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sun Apr 10 15:52:08 UTC 2022 - 12.5K bytes - Viewed (0) -
src/image/internal/imageutil/gen.go
} const pre = `// Code generated by go run gen.go; DO NOT EDIT. package imageutil import ( "image" ) // DrawYCbCr draws the YCbCr source image on the RGBA destination image with // r.Min in dst aligned with sp in src. It reports whether the draw was // successful. If it returns false, no dst pixels were changed. // // This function assumes that r is entirely within dst's bounds and the
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 28 18:17:57 UTC 2021 - 4.3K bytes - Viewed (0) -
src/runtime/pinner.go
func (s *mspan) refreshPinnerBits() { p := s.getPinnerBits() if p == nil { return } hasPins := false bytes := alignUp(s.pinnerBitSize(), 8) // Iterate over each 8-byte chunk and check for pins. Note that // newPinnerBits guarantees that pinnerBits will be 8-byte aligned, so we // don't have to worry about edge cases, irrelevant bits will simply be // zero.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 04 14:29:45 UTC 2024 - 11K bytes - Viewed (0) -
src/cmd/internal/objfile/elf.go
if p.Type == elf.PT_LOAD && p.Flags&elf.PF_X != 0 { // The memory mapping that contains the segment // starts at an aligned address. Apparently this // is what pprof expects, as it uses this and the // start address of the mapping to compute PC // delta. return p.Vaddr - p.Vaddr%p.Align, nil } } return 0, fmt.Errorf("unknown load address") } func (f *elfFile) dwarf() (*dwarf.Data, error) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 07 20:44:50 UTC 2023 - 3.9K bytes - Viewed (0) -
src/runtime/stubs.go
// // *ptr is uninitialized memory (e.g., memory that's being reused // for a new allocation) and hence contains only "junk". // // memclrNoHeapPointers ensures that if ptr is pointer-aligned, and n // is a multiple of the pointer size, then any pointer-aligned, // pointer-sized portion is cleared atomically. Despite the function // name, this is necessary because this function is the underlying
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 20.2K bytes - Viewed (0) -
src/runtime/mpagealloc_32bit.go
print("runtime: base = ", hex(base), ", limit = ", hex(limit), "\n") throw("sysGrow bounds not aligned to pallocChunkBytes") } // Walk up the tree and update the summary slices. for l := len(p.summary) - 1; l >= 0; l-- { // Figure out what part of the summary array this new address space needs. // Note that we need to align the ranges to the block width (1<<levelBits[l])
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 20 20:08:25 UTC 2023 - 4.6K bytes - Viewed (0) -
src/runtime/syscall_windows.go
// caller reserved spill space. p.dstSpill = alignUp(p.dstSpill, uintptr(t.Align_)) p.dstSpill += t.Size_ } else { // Register assignment failed. // Undo the work and stack assign. p.parts = oldParts // The Go ABI aligns arguments. p.dstStackSize = alignUp(p.dstStackSize, uintptr(t.Align_)) // Copy just the size of the argument. Note that this
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:12:46 UTC 2024 - 16.6K bytes - Viewed (0)