- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 192 for aligned2 (0.22 sec)
-
src/runtime/defs_linux_arm64.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 19:05:10 UTC 2023 - 3.6K bytes - Viewed (0) -
platforms/software/dependency-management/src/main/java/org/gradle/api/internal/artifacts/ivyservice/resolveengine/graph/builder/PotentialEdge.java
ComponentState version = resolveState.getModule(toSelector.getModuleIdentifier()).getVersion(toModuleVersionId, toComponent); // We need to check if the target version exists. For this, we have to try to get metadata for the aligned version. // If it's there, it means we can align, otherwise, we must NOT add the edge, or resolution would fail ComponentGraphResolveState metadata = version.getResolveStateOrNull();
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Tue Oct 10 21:10:11 UTC 2023 - 4.3K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_tensor.cc
} // The pointer tag, OR-ed into the XlaTensor's address to distinguish it from // device-side tensors, which are either CPU or GPU memory pointers. This works // because we're guaranteed that CPU and GPU pointers are aligned to > 1 bits. namespace { constexpr uintptr_t kTag = 0x1ULL; } /*static*/ XlaTensor* XlaTensor::FromOpaquePointer(void* ptr) { uintptr_t value = reinterpret_cast<uintptr_t>(ptr); if (value & kTag) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 4.5K bytes - Viewed (0) -
src/internal/runtime/atomic/atomic_mips64x.s
MOVV $~3, R3 AND R1, R3 // Compute val shift. #ifdef GOARCH_mips64 // Big endian. ptr = ptr ^ 3 XOR $3, R1 #endif // R4 = ((ptr & 3) * 8) AND $3, R1, R4 SLLV $3, R4 // Shift val for aligned ptr. R2 = val << R4 SLLV R4, R2 SYNC LL (R3), R4 OR R2, R4 SC R4, (R3) BEQ R4, -4(PC) SYNC RET // void And8(byte volatile*, byte); TEXT ·And8(SB), NOSPLIT, $0-9
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 11 21:29:34 UTC 2024 - 7.2K bytes - Viewed (0) -
pkg/kubelet/cm/cpumanager/policy_options.go
// Flag to evenly distribute CPUs across NUMA nodes in cases where more // than one NUMA node is required to satisfy the allocation. DistributeCPUsAcrossNUMA bool // Flag to ensure CPUs are considered aligned at socket boundary rather than // NUMA boundary AlignBySocket bool } // NewStaticPolicyOptions creates a StaticPolicyOptions struct from the user configuration.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Sep 27 13:02:15 UTC 2023 - 5.1K bytes - Viewed (0) -
src/runtime/pinner.go
return pinState{bytep, byteVal, mask} } func (s *mspan) pinnerBitSize() uintptr { return divRoundUp(uintptr(s.nelems)*2, 8) } // newPinnerBits returns a pointer to 8 byte aligned bytes to be used for this // span's pinner bits. newPinnerBits is used to mark objects that are pinned. // They are copied when the span is swept. func (s *mspan) newPinnerBits() *pinnerBits {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 04 14:29:45 UTC 2024 - 11K bytes - Viewed (0) -
src/runtime/syscall_windows.go
// passed as two words (little endian); and // structs are pushed on the stack. In // fastcall, arguments larger than the word // size are passed by reference. On arm, // 8-byte aligned arguments round up to the // next even register and can be split across // registers and the stack. panic("compileCallback: argument size is larger than uintptr") }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:12:46 UTC 2024 - 16.6K bytes - Viewed (0) -
platforms/software/dependency-management/src/integTest/groovy/org/gradle/integtests/resolve/alignment/ForcingUsingStrictlyPlatformAlignmentTest.groovy
} @RequiredFeature(feature = GradleMetadataResolveRunner.REPOSITORY_TYPE, value = "maven") @Issue("nebula-plugins/gradle-nebula-integration#51") def "force to higher patch version should bring the rest of aligned group up"() { given: "repository simulating Jackson situation" { path 'com.amazonaws:aws-java-sdk-core:1.11.438 -> org:cbor:2.6.7'
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Tue Oct 24 06:54:47 UTC 2023 - 21.6K bytes - Viewed (0) -
src/sync/atomic/doc.go
// atomic functions (types [Int64] and [Uint64] are automatically aligned). // The first word in an allocated struct, array, or slice; in a global // variable; or in a local variable (because the subject of all atomic operations // will escape to the heap) can be relied upon to be 64-bit aligned. // SwapInt32 atomically stores new into *addr and returns the previous *addr value.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 11.7K bytes - Viewed (0) -
tensorflow/compiler/aot/codegen_test_h.golden
// ((unknown): f32[1,2], (unknown): s64[3,4], (unknown): f32[1], (unknown): f32[1], (unknown): s32[5]) -> (u32[5,6], f32[1], s32[5]) // // Memory stats: // arg bytes total: 392 // arg bytes aligned: 576 // temp bytes total: 171 // temp bytes aligned: 512 class MyClass final : public tensorflow::XlaCompiledCpuFunction { public: // Number of input arguments for the compiled computation. static constexpr size_t kNumArgs = 5;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 01:20:01 UTC 2024 - 16.6K bytes - Viewed (0)