Search Options

Results per page
Sort
Preferred Languages
Advance

Results 91 - 100 of 510 for aligned (0.19 sec)

  1. platforms/software/dependency-management/src/main/java/org/gradle/api/internal/artifacts/ivyservice/resolveengine/graph/builder/PotentialEdge.java

            // We need to check if the target version exists. For this, we have to try to get metadata for the aligned version.
            // If it's there, it means we can align, otherwise, we must NOT add the edge, or resolution would fail
            ComponentGraphResolveState metadata = version.getResolveStateOrNull();
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Tue Oct 10 21:10:11 UTC 2023
    - 4.3K bytes
    - Viewed (0)
  2. src/runtime/malloc.go

    		p = alignUp(p, align)
    		p2 := sysReserve(unsafe.Pointer(p), size)
    		if p != uintptr(p2) {
    			// Must have raced. Try again.
    			sysFreeOS(p2, size)
    			if retries++; retries == 100 {
    				throw("failed to allocate aligned heap memory; too many retries")
    			}
    			goto retry
    		}
    		// Success.
    		return p2, size
    	default:
    		// Trim off the unaligned parts.
    		pAligned := alignUp(p, align)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 59.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/xla_tensor.cc

    }
    
    // The pointer tag, OR-ed into the XlaTensor's address to distinguish it from
    // device-side tensors, which are either CPU or GPU memory pointers. This works
    // because we're guaranteed that CPU and GPU pointers are aligned to > 1 bits.
    namespace {
    constexpr uintptr_t kTag = 0x1ULL;
    }
    
    /*static*/ XlaTensor* XlaTensor::FromOpaquePointer(void* ptr) {
      uintptr_t value = reinterpret_cast<uintptr_t>(ptr);
      if (value & kTag) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 4.5K bytes
    - Viewed (0)
  4. src/internal/abi/abi.go

    //
    // This method is a helper for dealing with the endianness of different CPU
    // architectures, since sub-word-sized arguments in big endian architectures
    // need to be "aligned" to the upper edge of the register to be interpreted
    // by the CPU correctly.
    func (r *RegArgs) IntRegArgAddr(reg int, argSize uintptr) unsafe.Pointer {
    	if argSize > goarch.PtrSize || argSize == 0 || argSize&(argSize-1) != 0 {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Jan 23 15:51:32 UTC 2023
    - 3.1K bytes
    - Viewed (0)
  5. src/internal/runtime/atomic/atomic_mips64x.s

    TEXT ·Or8(SB), NOSPLIT, $0-9
    	MOVV	ptr+0(FP), R1
    	MOVBU	val+8(FP), R2
    	// Align ptr down to 4 bytes so we can use 32-bit load/store.
    	MOVV	$~3, R3
    	AND	R1, R3
    	// Compute val shift.
    #ifdef GOARCH_mips64
    	// Big endian.  ptr = ptr ^ 3
    	XOR	$3, R1
    #endif
    	// R4 = ((ptr & 3) * 8)
    	AND	$3, R1, R4
    	SLLV	$3, R4
    	// Shift val for aligned ptr. R2 = val << R4
    	SLLV	R4, R2
    
    	SYNC
    	LL	(R3), R4
    	OR	R2, R4
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Sat May 11 21:29:34 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  6. src/runtime/tagptr_64bit.go

    	// get to really high addresses and panic if it does.
    	addrBits = 48
    
    	// In addition to the 16 bits taken from the top, we can take 3 from the
    	// bottom, because node must be pointer-aligned, giving a total of 19 bits
    	// of count.
    	tagBits = 64 - addrBits + 3
    
    	// On AIX, 64-bit addresses are split into 36-bit segment number and 28-bit
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Apr 18 20:22:50 UTC 2023
    - 3.2K bytes
    - Viewed (0)
  7. code_of_conduct.md

    licensing terms applying to the Project developments.
    
    Project maintainers have the right and responsibility to remove, edit, or
    reject comments, commits, code, wiki edits, issues, and other contributions
    that are not aligned to this Code of Conduct, or to ban temporarily or
    permanently any contributor for other behaviors that they deem inappropriate,
    threatening, offensive, or harmful. However, these actions shall respect the
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Mon Jan 20 18:38:58 UTC 2020
    - 3.5K bytes
    - Viewed (0)
  8. pkg/kubelet/cm/cpumanager/policy_options.go

    	// Flag to evenly distribute CPUs across NUMA nodes in cases where more
    	// than one NUMA node is required to satisfy the allocation.
    	DistributeCPUsAcrossNUMA bool
    	// Flag to ensure CPUs are considered aligned at socket boundary rather than
    	// NUMA boundary
    	AlignBySocket bool
    }
    
    // NewStaticPolicyOptions creates a StaticPolicyOptions struct from the user configuration.
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Sep 27 13:02:15 UTC 2023
    - 5.1K bytes
    - Viewed (0)
  9. src/math/rand/example_test.go

    	// Typically a non-fixed seed should be used, such as time.Now().UnixNano().
    	// Using a fixed seed will produce the same output on every run.
    	r := rand.New(rand.NewSource(99))
    
    	// The tabwriter here helps us generate aligned output.
    	w := tabwriter.NewWriter(os.Stdout, 1, 1, 1, ' ', 0)
    	defer w.Flush()
    	show := func(name string, v1, v2, v3 any) {
    		fmt.Fprintf(w, "%s\t%v\t%v\t%v\n", name, v1, v2, v3)
    	}
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Oct 26 16:24:57 UTC 2022
    - 4.2K bytes
    - Viewed (0)
  10. src/cmd/internal/obj/s390x/rotate.go

    // will be returned.
    func (r RotateParams) OutMerge(mask uint64) *RotateParams {
    	mask &= r.OutMask()
    	if mask == 0 {
    		return nil
    	}
    
    	// normalize the mask so that the set bits are left aligned
    	o := bits.LeadingZeros64(^mask)
    	mask = bits.RotateLeft64(mask, o)
    	z := bits.LeadingZeros64(mask)
    	mask = bits.RotateLeft64(mask, z)
    
    	// check that the normalized mask is contiguous
    	l := bits.LeadingZeros64(^mask)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Dec 21 19:19:04 UTC 2020
    - 3.6K bytes
    - Viewed (0)
Back to top