Search Options

Results per page
Sort
Preferred Languages
Advance

Results 71 - 80 of 451 for aligned (0.11 sec)

  1. src/runtime/tagptr_64bit.go

    	// get to really high addresses and panic if it does.
    	addrBits = 48
    
    	// In addition to the 16 bits taken from the top, we can take 3 from the
    	// bottom, because node must be pointer-aligned, giving a total of 19 bits
    	// of count.
    	tagBits = 64 - addrBits + 3
    
    	// On AIX, 64-bit addresses are split into 36-bit segment number and 28-bit
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Apr 18 20:22:50 UTC 2023
    - 3.2K bytes
    - Viewed (0)
  2. pkg/kubelet/cm/cpumanager/policy_options.go

    	// Flag to evenly distribute CPUs across NUMA nodes in cases where more
    	// than one NUMA node is required to satisfy the allocation.
    	DistributeCPUsAcrossNUMA bool
    	// Flag to ensure CPUs are considered aligned at socket boundary rather than
    	// NUMA boundary
    	AlignBySocket bool
    }
    
    // NewStaticPolicyOptions creates a StaticPolicyOptions struct from the user configuration.
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Sep 27 13:02:15 UTC 2023
    - 5.1K bytes
    - Viewed (0)
  3. src/math/rand/example_test.go

    	// Typically a non-fixed seed should be used, such as time.Now().UnixNano().
    	// Using a fixed seed will produce the same output on every run.
    	r := rand.New(rand.NewSource(99))
    
    	// The tabwriter here helps us generate aligned output.
    	w := tabwriter.NewWriter(os.Stdout, 1, 1, 1, ' ', 0)
    	defer w.Flush()
    	show := func(name string, v1, v2, v3 any) {
    		fmt.Fprintf(w, "%s\t%v\t%v\t%v\n", name, v1, v2, v3)
    	}
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Oct 26 16:24:57 UTC 2022
    - 4.2K bytes
    - Viewed (0)
  4. src/runtime/memmove_amd64.s

    	// Algorithm:
    	// 1. Unaligned save of the tail's 128 bytes
    	// 2. Unaligned save of the head's 32  bytes
    	// 3. Destination-aligned copying of body (128 bytes per iteration)
    	// 4. Put head on the new place
    	// 5. Put the tail on the new place
    	// It can be important to satisfy processor's pipeline requirements for
    	// small sizes as the cost of unaligned memory region copying is
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Sun Apr 10 15:52:08 UTC 2022
    - 12.5K bytes
    - Viewed (0)
  5. src/image/internal/imageutil/gen.go

    }
    
    const pre = `// Code generated by go run gen.go; DO NOT EDIT.
    
    package imageutil
    
    import (
    	"image"
    )
    
    // DrawYCbCr draws the YCbCr source image on the RGBA destination image with
    // r.Min in dst aligned with sp in src. It reports whether the draw was
    // successful. If it returns false, no dst pixels were changed.
    //
    // This function assumes that r is entirely within dst's bounds and the
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 28 18:17:57 UTC 2021
    - 4.3K bytes
    - Viewed (0)
  6. src/runtime/pinner.go

    func (s *mspan) refreshPinnerBits() {
    	p := s.getPinnerBits()
    	if p == nil {
    		return
    	}
    
    	hasPins := false
    	bytes := alignUp(s.pinnerBitSize(), 8)
    
    	// Iterate over each 8-byte chunk and check for pins. Note that
    	// newPinnerBits guarantees that pinnerBits will be 8-byte aligned, so we
    	// don't have to worry about edge cases, irrelevant bits will simply be
    	// zero.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Apr 04 14:29:45 UTC 2024
    - 11K bytes
    - Viewed (0)
  7. src/cmd/internal/objfile/elf.go

    		if p.Type == elf.PT_LOAD && p.Flags&elf.PF_X != 0 {
    			// The memory mapping that contains the segment
    			// starts at an aligned address. Apparently this
    			// is what pprof expects, as it uses this and the
    			// start address of the mapping to compute PC
    			// delta.
    			return p.Vaddr - p.Vaddr%p.Align, nil
    		}
    	}
    	return 0, fmt.Errorf("unknown load address")
    }
    
    func (f *elfFile) dwarf() (*dwarf.Data, error) {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Apr 07 20:44:50 UTC 2023
    - 3.9K bytes
    - Viewed (0)
  8. src/runtime/stubs.go

    //
    // *ptr is uninitialized memory (e.g., memory that's being reused
    // for a new allocation) and hence contains only "junk".
    //
    // memclrNoHeapPointers ensures that if ptr is pointer-aligned, and n
    // is a multiple of the pointer size, then any pointer-aligned,
    // pointer-sized portion is cleared atomically. Despite the function
    // name, this is necessary because this function is the underlying
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 20.2K bytes
    - Viewed (0)
  9. src/runtime/mpagealloc_32bit.go

    		print("runtime: base = ", hex(base), ", limit = ", hex(limit), "\n")
    		throw("sysGrow bounds not aligned to pallocChunkBytes")
    	}
    
    	// Walk up the tree and update the summary slices.
    	for l := len(p.summary) - 1; l >= 0; l-- {
    		// Figure out what part of the summary array this new address space needs.
    		// Note that we need to align the ranges to the block width (1<<levelBits[l])
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Apr 20 20:08:25 UTC 2023
    - 4.6K bytes
    - Viewed (0)
  10. src/runtime/syscall_windows.go

    		// caller reserved spill space.
    		p.dstSpill = alignUp(p.dstSpill, uintptr(t.Align_))
    		p.dstSpill += t.Size_
    	} else {
    		// Register assignment failed.
    		// Undo the work and stack assign.
    		p.parts = oldParts
    
    		// The Go ABI aligns arguments.
    		p.dstStackSize = alignUp(p.dstStackSize, uintptr(t.Align_))
    
    		// Copy just the size of the argument. Note that this
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 20:12:46 UTC 2024
    - 16.6K bytes
    - Viewed (0)
Back to top