- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 218 for aligned2 (0.18 sec)
-
src/runtime/defs_linux_arm64.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 19:05:10 UTC 2023 - 3.6K bytes - Viewed (0) -
platforms/software/dependency-management/src/main/java/org/gradle/api/internal/artifacts/ivyservice/resolveengine/graph/builder/PotentialEdge.java
ComponentState version = resolveState.getModule(toSelector.getModuleIdentifier()).getVersion(toModuleVersionId, toComponent); // We need to check if the target version exists. For this, we have to try to get metadata for the aligned version. // If it's there, it means we can align, otherwise, we must NOT add the edge, or resolution would fail ComponentGraphResolveState metadata = version.getResolveStateOrNull();
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Tue Oct 10 21:10:11 UTC 2023 - 4.3K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_tensor.cc
} // The pointer tag, OR-ed into the XlaTensor's address to distinguish it from // device-side tensors, which are either CPU or GPU memory pointers. This works // because we're guaranteed that CPU and GPU pointers are aligned to > 1 bits. namespace { constexpr uintptr_t kTag = 0x1ULL; } /*static*/ XlaTensor* XlaTensor::FromOpaquePointer(void* ptr) { uintptr_t value = reinterpret_cast<uintptr_t>(ptr); if (value & kTag) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 4.5K bytes - Viewed (0) -
src/internal/abi/abi.go
// // This method is a helper for dealing with the endianness of different CPU // architectures, since sub-word-sized arguments in big endian architectures // need to be "aligned" to the upper edge of the register to be interpreted // by the CPU correctly. func (r *RegArgs) IntRegArgAddr(reg int, argSize uintptr) unsafe.Pointer { if argSize > goarch.PtrSize || argSize == 0 || argSize&(argSize-1) != 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jan 23 15:51:32 UTC 2023 - 3.1K bytes - Viewed (0) -
src/runtime/tagptr_64bit.go
// get to really high addresses and panic if it does. addrBits = 48 // In addition to the 16 bits taken from the top, we can take 3 from the // bottom, because node must be pointer-aligned, giving a total of 19 bits // of count. tagBits = 64 - addrBits + 3 // On AIX, 64-bit addresses are split into 36-bit segment number and 28-bit
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 18 20:22:50 UTC 2023 - 3.2K bytes - Viewed (0) -
src/internal/runtime/atomic/atomic_mips64x.s
MOVV $~3, R3 AND R1, R3 // Compute val shift. #ifdef GOARCH_mips64 // Big endian. ptr = ptr ^ 3 XOR $3, R1 #endif // R4 = ((ptr & 3) * 8) AND $3, R1, R4 SLLV $3, R4 // Shift val for aligned ptr. R2 = val << R4 SLLV R4, R2 SYNC LL (R3), R4 OR R2, R4 SC R4, (R3) BEQ R4, -4(PC) SYNC RET // void And8(byte volatile*, byte); TEXT ·And8(SB), NOSPLIT, $0-9
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 11 21:29:34 UTC 2024 - 7.2K bytes - Viewed (0) -
src/math/rand/example_test.go
// Typically a non-fixed seed should be used, such as time.Now().UnixNano(). // Using a fixed seed will produce the same output on every run. r := rand.New(rand.NewSource(99)) // The tabwriter here helps us generate aligned output. w := tabwriter.NewWriter(os.Stdout, 1, 1, 1, ' ', 0) defer w.Flush() show := func(name string, v1, v2, v3 any) { fmt.Fprintf(w, "%s\t%v\t%v\t%v\n", name, v1, v2, v3) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Oct 26 16:24:57 UTC 2022 - 4.2K bytes - Viewed (0) -
pkg/kubelet/cm/cpumanager/policy_options.go
// Flag to evenly distribute CPUs across NUMA nodes in cases where more // than one NUMA node is required to satisfy the allocation. DistributeCPUsAcrossNUMA bool // Flag to ensure CPUs are considered aligned at socket boundary rather than // NUMA boundary AlignBySocket bool } // NewStaticPolicyOptions creates a StaticPolicyOptions struct from the user configuration.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Sep 27 13:02:15 UTC 2023 - 5.1K bytes - Viewed (0) -
src/runtime/pinner.go
return pinState{bytep, byteVal, mask} } func (s *mspan) pinnerBitSize() uintptr { return divRoundUp(uintptr(s.nelems)*2, 8) } // newPinnerBits returns a pointer to 8 byte aligned bytes to be used for this // span's pinner bits. newPinnerBits is used to mark objects that are pinned. // They are copied when the span is swept. func (s *mspan) newPinnerBits() *pinnerBits {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 04 14:29:45 UTC 2024 - 11K bytes - Viewed (0) -
src/runtime/syscall_windows.go
// passed as two words (little endian); and // structs are pushed on the stack. In // fastcall, arguments larger than the word // size are passed by reference. On arm, // 8-byte aligned arguments round up to the // next even register and can be split across // registers and the stack. panic("compileCallback: argument size is larger than uintptr") }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:12:46 UTC 2024 - 16.6K bytes - Viewed (0)