- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 2,632 for i$ (0.04 sec)
-
src/internal/weak/pointer_test.go
wt := make([]weak.Pointer[T], 10) for i := range bt { bt[i] = new(T) wt[i] = weak.Make(bt[i]) } for i := range bt { st := wt[i].Strong() if st != bt[i] { t.Fatalf("weak pointer is not the same as strong pointer: %p vs. %p", st, bt[i]) } if wp := weak.Make(st); wp != wt[i] { t.Fatalf("new weak pointer not equal to existing weak pointer: %v vs. %v", wp, wt[i]) } if i == 0 { continue }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 18 21:25:11 UTC 2024 - 3.1K bytes - Viewed (0) -
test/codegen/compare_and_branch.go
// s390x:"CGIJ\t[$]12, R[0-9]+, [$]127, " for i := doNotOptimize; i < 128; i++ { dummy() } // s390x:"CGIJ\t[$]10, R[0-9]+, [$]-128, " for i := doNotOptimize; i > -129; i-- { dummy() } // s390x:"CGIJ\t[$]2, R[0-9]+, [$]127, " for i := doNotOptimize; i >= 128; i++ { dummy() } // s390x:"CGIJ\t[$]4, R[0-9]+, [$]-128, " for i := doNotOptimize; i <= -129; i-- { dummy() } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Sep 12 21:01:50 UTC 2023 - 4.5K bytes - Viewed (0) -
src/maps/maps_test.go
} for i := 0; i < 8; i++ { if m2[i] != m[i] { t.Errorf("m2[%d] = %d, want %d", i, m2[i], m[i]) } } } func TestCloneWithMapAssign(t *testing.T) { var m = make(map[int]int) const N = 25 for i := 0; i < N; i++ { m[i] = i } m2 := Clone(m) if len(m2) != N { t.Errorf("len2(m2) = %d, want %d", len(m2), N) } for i := 0; i < N; i++ { if m2[i] != m[i] {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Dec 05 17:05:56 UTC 2023 - 5.6K bytes - Viewed (0) -
internal/bucket/bandwidth/monitor_gen_test.go
b.ReportAllocs() b.ResetTimer() for i := 0; i < b.N; i++ { v.MarshalMsg(nil) } } func BenchmarkAppendMsgBucketBandwidthReport(b *testing.B) { v := BucketBandwidthReport{} bts := make([]byte, 0, v.Msgsize()) bts, _ = v.MarshalMsg(bts[0:0]) b.SetBytes(int64(len(bts))) b.ReportAllocs() b.ResetTimer() for i := 0; i < b.N; i++ { bts, _ = v.MarshalMsg(bts[0:0]) } }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon Feb 19 22:54:46 UTC 2024 - 4.5K bytes - Viewed (0) -
cmd/tier-last-day-stats_gen_test.go
v := DailyAllTierStats{} b.ReportAllocs() b.ResetTimer() for i := 0; i < b.N; i++ { v.MarshalMsg(nil) } } func BenchmarkAppendMsgDailyAllTierStats(b *testing.B) { v := DailyAllTierStats{} bts := make([]byte, 0, v.Msgsize()) bts, _ = v.MarshalMsg(bts[0:0]) b.SetBytes(int64(len(bts))) b.ReportAllocs() b.ResetTimer() for i := 0; i < b.N; i++ { bts, _ = v.MarshalMsg(bts[0:0]) } }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon Feb 19 22:54:46 UTC 2024 - 4.6K bytes - Viewed (0) -
pkg/kube/krt/index.go
// Index maintains a simple index over an informer type Index[I any, K comparable] struct { mu sync.RWMutex objects map[K]sets.Set[Key[I]] c Collection[I] extract func(o I) []K } // Lookup finds all objects matching a given key func (i *Index[I, K]) Lookup(k K) []I { i.mu.RLock() defer i.mu.RUnlock() var res []I for obj := range i.objects[k] { item := i.c.GetKey(obj) if item == nil {
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Tue May 14 04:53:45 UTC 2024 - 2.8K bytes - Viewed (0) -
src/regexp/syntax/prog.go
} // skipNop follows any no-op or capturing instructions. func (p *Prog) skipNop(pc uint32) *Inst { i := &p.Inst[pc] for i.Op == InstNop || i.Op == InstCapture { i = &p.Inst[i.Out] } return i } // op returns i.Op but merges all the Rune special cases into InstRune func (i *Inst) op() InstOp { op := i.Op switch op { case InstRune1, InstRuneAny, InstRuneAnyNotNL: op = InstRune } return op
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Feb 26 20:50:01 UTC 2024 - 7.7K bytes - Viewed (0) -
src/net/parse.go
n = n*10 + int(s[i]-'0') if n >= big { return big, i, false } } if i == 0 { return 0, 0, false } return n, i, true } // Hexadecimal to integer. // Returns number, characters consumed, success. func xtoi(s string) (n int, i int, ok bool) { n = 0 for i = 0; i < len(s); i++ { if '0' <= s[i] && s[i] <= '9' { n *= 16 n += int(s[i] - '0')
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 06 14:00:54 UTC 2024 - 5.7K bytes - Viewed (0) -
src/internal/runtime/atomic/bench_test.go
sink = &x for i := 0; i < b.N; i++ { atomic.Store(&x, 0) } } func BenchmarkAnd8(b *testing.B) { var x [512]uint8 // give byte its own cache line sink = &x for i := 0; i < b.N; i++ { atomic.And8(&x[255], uint8(i)) } } func BenchmarkAnd(b *testing.B) { var x [128]uint32 // give x its own cache line sink = &x for i := 0; i < b.N; i++ { atomic.And(&x[63], uint32(i)) } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 3.2K bytes - Viewed (0) -
internal/ringbuffer/ring_buffer_benchmark_test.go
} }() b.ResetTimer() for i := 0; i < b.N; i++ { rb.Write(data) } } func BenchmarkRingBuffer_AsyncWrite(b *testing.B) { rb := New(1024) data := []byte(strings.Repeat("a", 512)) buf := make([]byte, 512) go func() { for { rb.Write(data) } }() b.ResetTimer() for i := 0; i < b.N; i++ { rb.Read(buf) } } func BenchmarkRingBuffer_AsyncWriteBlocking(b *testing.B) {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Wed May 15 00:11:04 UTC 2024 - 1.7K bytes - Viewed (0)