- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 523 for offset_ (0.17 sec)
-
pkg/ctrlz/assets/static/js/popper-1.12.9.min.js
tion` is deprecated, use `modifier.fn`!');var i=t['function']||t.fn;t.enabled&&e(i)&&(o.offsets.popper=c(o.offsets.popper),o.offsets.reference=c(o.offsets.reference),o=i(o,t))}),o}function N(){if(!this.state.isDestroyed){var e={instance:this,styles:{},arrowStyles:{},attributes:{},flipped:!1,offsets:{}};e.offsets.reference=O(this.state,this.popper,this.reference),e.placement=v(this.options.placement,e.offsets.reference,this.popper,this.reference,this.options.modifiers.flip.boundariesElement,this....
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Tue May 23 17:08:31 UTC 2023 - 18.7K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/expand_calls.go
if et == types.TINT64 { tHi = x.typs.Int32 } tLo = x.typs.UInt32 return } // offsetFrom creates an offset from a pointer, simplifying chained offsets and offsets from SP func (x *expandState) offsetFrom(b *Block, from *Value, offset int64, pt *types.Type) *Value { ft := from.Type if offset == 0 { if ft == pt { return from }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 28 05:13:40 UTC 2023 - 31.9K bytes - Viewed (0) -
src/runtime/type.go
// and treat the offset as an identifier. We use negative offsets that // do not overlap with any compile-time module offsets. // // Entries are created by reflect.addReflectOff. var reflectOffs struct { lock mutex next int32 m map[int32]unsafe.Pointer minv map[unsafe.Pointer]int32 } func reflectOffsLock() { lock(&reflectOffs.lock) if raceenabled {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 12.7K bytes - Viewed (0) -
cmd/erasure-decode_test.go
} // To generate random offset/length. r := rand.New(rand.NewSource(UTCNow().UnixNano())) buf := &bytes.Buffer{} // Verify erasure.Decode() for random offsets and lengths. for i := 0; i < iterations; i++ { offset := r.Int63n(length) readLen := r.Int63n(length - offset) expected := data[offset : offset+readLen] // Get the checksums of the current part.
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Tue Jan 30 20:43:25 UTC 2024 - 21.1K bytes - Viewed (0) -
src/cmd/internal/goobj/objfile.go
return int(r.h.Offsets[BlkHashed64def+1]-r.h.Offsets[BlkHashed64def]) / SymSize } func (r *Reader) NHasheddef() int { return int(r.h.Offsets[BlkHasheddef+1]-r.h.Offsets[BlkHasheddef]) / SymSize } func (r *Reader) NNonpkgdef() int { return int(r.h.Offsets[BlkNonpkgdef+1]-r.h.Offsets[BlkNonpkgdef]) / SymSize } func (r *Reader) NNonpkgref() int {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 15 19:57:43 UTC 2024 - 23.8K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/tools/go/analysis/passes/asmdecl/asmdecl.go
vars: make(map[string]*asmVar), varByOffset: make(map[int]*asmVar), } offset = 0 addParams(decl.Type.Params.List, false) if decl.Type.Results != nil && len(decl.Type.Results.List) > 0 { offset += -offset & (arch.maxAlign - 1) addParams(decl.Type.Results.List, true) } fn.size = offset m[arch.name] = fn } return m }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 02:38:00 UTC 2024 - 22.8K bytes - Viewed (0) -
src/cmd/compile/internal/types/size.go
} for i, m := range methods { m.Offset = int64(i) * int64(PtrSize) } t.SetAllMethods(methods) } // calcStructOffset computes the offsets of a sequence of fields, // starting at the given offset. It returns the resulting offset and // maximum field alignment. func calcStructOffset(t *Type, fields []*Field, offset int64) int64 { for _, f := range fields { CalcSize(f.Type)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 15K bytes - Viewed (0) -
src/internal/zstd/block.go
r.repeatedOffset2 = r.repeatedOffset1 r.repeatedOffset1 = offset } else { if literal == 0 { offset++ } switch offset { case 1: offset = r.repeatedOffset1 case 2: offset = r.repeatedOffset2 r.repeatedOffset2 = r.repeatedOffset1 r.repeatedOffset1 = offset case 3: offset = r.repeatedOffset3 r.repeatedOffset3 = r.repeatedOffset2
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Sep 28 17:57:43 UTC 2023 - 10.2K bytes - Viewed (0) -
src/compress/flate/huffman_bit_writer.go
} // Write the offset offset := t.offset() offsetCode := offsetCode(offset) w.writeCode(oeCodes[offsetCode]) extraOffsetBits := uint(offsetExtraBits[offsetCode]) if extraOffsetBits > 0 { extraOffset := int32(offset - offsetBase[offsetCode]) w.writeBits(extraOffset, extraOffsetBits) } } } // huffOffset is a static offset encoder used for huffman only encoding.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Sep 29 22:59:14 UTC 2022 - 18.4K bytes - Viewed (0) -
src/cmd/link/internal/ld/pcln.go
func (state *pclntab) generatePctab(ctxt *Link, funcs []loader.Sym) { ldr := ctxt.loader // Pctab offsets of 0 are considered invalid in the runtime. We respect // that by just padding a single byte at the beginning of runtime.pctab, // that way no real offsets can be zero. size := int64(1) // Walk the functions, finding offset to store each pcdata. seen := make(map[loader.Sym]struct{}) saveOffset := func(pcSym loader.Sym) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Feb 21 22:16:54 UTC 2024 - 29.6K bytes - Viewed (0)