- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 35 for xoffset (0.25 sec)
-
src/runtime/malloc.go
// index 0 in the heap arena map. // // On amd64, the address space is 48 bits, sign extended to 64 // bits. This offset lets us handle "negative" addresses (or // high addresses if viewed as unsigned). // // On aix/ppc64, this offset allows to keep the heapAddrBits to // 48. Otherwise, it would be 60 in order to handle mmap addresses // (in range 0x0a00000000000000 - 0x0afffffffffffff). But in this
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
src/runtime/pprof/pprof_test.go
fFunc := runtime.FuncForPC(uintptr(abi.FuncPCABIInternal(f))) if fFunc == nil || fFunc.Entry() == 0 { panic("failed to locate function entry") } for offset := 0; offset < maxBytes; offset++ { innerPC := fFunc.Entry() + uintptr(offset) inner := runtime.FuncForPC(innerPC) if inner == nil { // No function known for this PC value. // It might simply be misaligned, so keep searching. continue }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 18:42:28 UTC 2024 - 68.8K bytes - Viewed (0) -
src/cmd/cgo/gcc.go
needsUnsafe := false // Walk backward so that in C.f1(C.f2()) we rewrite C.f2 first. for _, call := range f.Calls { if call.Done { continue } start := f.offset(call.Call.Pos()) end := f.offset(call.Call.End()) str, nu := p.rewriteCall(f, call) if str != "" { f.Edit.Replace(start, end, str) if nu { needsUnsafe = true } } } return needsUnsafe }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 20 15:50:06 UTC 2024 - 97K bytes - Viewed (0) -
src/sync/atomic/atomic_test.go
} if o := reflect.TypeOf(&signed).Elem().Field(1).Offset; o != 8 { t.Fatalf("Int64 offset = %d, want 8", o) } if p := reflect.ValueOf(&signed).Elem().Field(1).Addr().Pointer(); p&7 != 0 { t.Fatalf("Int64 pointer = %#x, want 8-aligned", p) } var unsigned struct { _ uint32 i Uint64 } if o := reflect.TypeOf(&unsigned).Elem().Field(1).Offset; o != 8 { t.Fatalf("Uint64 offset = %d, want 8", o)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 71.4K bytes - Viewed (0) -
src/runtime/traceback.go
if p == nil { return } liveInfo := funcdata(f, abi.FUNCDATA_ArgLiveInfo) liveIdx := pcdatavalue(f, abi.PCDATA_ArgLiveIndex, pc) startOffset := uint8(0xff) // smallest offset that needs liveness info (slots with a lower offset is always live) if liveInfo != nil { startOffset = *(*uint8)(liveInfo) } isLive := func(off, slotIdx uint8) bool { if liveInfo == nil || liveIdx <= 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 55.1K bytes - Viewed (0) -
.bazelrc
build:rbe_linux_cpu --copt=-Wno-error=unused-command-line-argument # This was added in clang-16 by https://reviews.llvm.org/D133574. # Can be removed once upb is updated, since a type definition is used within # offset of in the current version of ubp. # See https://github.com/protocolbuffers/upb/blob/9effcbcb27f0a665f9f345030188c0b291e32482/upb/upb.c#L183. build:rbe_linux_cpu --copt=-Wno-gnu-offsetof-extensions
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 17:12:54 UTC 2024 - 52.9K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/PPC64.rules
// Indexed ops generate indexed load or store instructions for all GOPPC64 values. // Non-indexed ops generate DS-form loads and stores when the offset fits in 16 bits, // and on power8 and power9, a multiple of 4 is required for MOVW and MOVD ops. // On power10, prefixed loads and stores can be used for offsets > 16 bits and <= 32 bits. // and support for PC relative addressing must be available if relocation is needed.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 53.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/prepare-tf.mlir
// mean * scale * rsqrt(variance + epsilon) // CHECK: %[[MUL3:.*]] = "tf.Mul"(%[[ARG3:.*]], %[[MUL1]]) // offset - mean * scale * rsqrt(variance + epsilon) // CHECK: %[[SUB:.*]] = "tf.Sub"(%[[ARG2:.*]], %[[MUL3]]) // x * scale * rsqrt(variance + epsilon) + // offset - mean * scale * rsqrt(variance + epsilon) // CHECK: %[[ADD2:.*]] = "tf.Add"(%[[MUL2]], %[[SUB]])
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 07:26:59 UTC 2024 - 59.8K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/ARM64Ops.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 58.8K bytes - Viewed (0) -
cmd/erasure-object.go
if length < 0 { length = fi.Size - startOffset } // Reply back invalid range if the input offset and length fall out of range. if startOffset > fi.Size || startOffset+length > fi.Size { return InvalidRange{startOffset, length, fi.Size} } // Get start part index and offset. partIndex, partOffset, err := fi.ObjectToPartOffset(ctx, startOffset) if err != nil {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon Jun 10 15:51:27 UTC 2024 - 78.6K bytes - Viewed (0)