- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 35 for newSource (0.05 sec)
-
cmd/erasure-server-pool-rebalance.go
go func() { // Update rebalance.bin periodically once every 5-10s, chosen randomly // to avoid multiple pool leaders herding to update around the same // time. r := rand.New(rand.NewSource(time.Now().UnixNano())) randSleepFor := func() time.Duration { return 5*time.Second + time.Duration(float64(5*time.Second)*r.Float64()) } timer := time.NewTimer(randSleepFor()) defer timer.Stop()
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Thu Sep 04 20:47:24 UTC 2025 - 28.9K bytes - Viewed (0) -
cmd/xl-storage-format-v2_test.go
t.Run(test.name, func(t *testing.T) { // Run multiple times, shuffling the input order. for i := range int64(50) { t.Run(fmt.Sprint(i), func(t *testing.T) { rng := rand.New(rand.NewSource(i)) rng.Shuffle(len(test.input), func(i, j int) { test.input[i], test.input[j] = test.input[j], test.input[i] }) got := mergeXLV2Versions(test.quorum, true, 0, test.input...)
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 37.8K bytes - Viewed (0) -
cmd/erasure-decode_test.go
if err != nil { t.Fatal(err) } if n != length { t.Errorf("erasureCreateFile returned %d, expected %d", n, length) } // To generate random offset/length. r := rand.New(rand.NewSource(UTCNow().UnixNano())) buf := &bytes.Buffer{} // Verify erasure.Decode() for random offsets and lengths. for range iterations { offset := r.Int63n(length) readLen := r.Int63n(length - offset)
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 21K bytes - Viewed (0) -
cmd/object-api-utils_test.go
b.ResetTimer() b.ReportAllocs() for b.Loop() { concat(data...) } }) } func BenchmarkConcatImplementation(b *testing.B) { data := make([]string, 2) rng := rand.New(rand.NewSource(0)) for i := range 2 { var tmp [16]byte rng.Read(tmp[:]) data[i] = hex.EncodeToString(tmp[:]) } b.ResetTimer() benchmark(b, data) } func BenchmarkPathJoinOld(b *testing.B) {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 24.1K bytes - Viewed (0) -
cmd/iam.go
globalServerConfigMu.RLock() s := globalServerConfig globalServerConfigMu.RUnlock() sys.Lock() sys.iamRefreshInterval = iamRefreshInterval sys.Unlock() r := rand.New(rand.NewSource(time.Now().UnixNano())) var ( openidInit bool ldapInit bool authNInit bool authZInit bool ) stsTLSConfig, err := xtls.Lookup(s[config.IdentityTLSSubSys][config.Default])
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 75.3K bytes - Viewed (0) -
cmd/xl-storage-disk-id-check.go
if skipIfSuccessBefore <= 0 { skipIfSuccessBefore = globalDriveConfig.GetMaxTimeout() } } t := time.NewTicker(checkEvery) defer t.Stop() fn := mustGetUUID() rng := rand.New(rand.NewSource(time.Now().UnixNano())) monitor := func() bool { if contextCanceled(ctx) { return false } if p.health.status.Load() != diskHealthOK { return true }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Apr 25 05:41:04 UTC 2025 - 34.5K bytes - Viewed (0) -
cmd/erasure-sets.go
// the set topology, this monitoring happens at a given monitoring interval. func (s *erasureSets) monitorAndConnectEndpoints(ctx context.Context, monitorInterval time.Duration) { r := rand.New(rand.NewSource(time.Now().UnixNano())) time.Sleep(time.Duration(r.Float64() * float64(time.Second))) // Pre-emptively connect the disks if possible. s.connectDisks(false) monitor := time.NewTimer(monitorInterval)
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 37K bytes - Viewed (1) -
cmd/erasure-server-pool-decom.go
go func() { // Resume decommissioning of pools, but wait 3 minutes for cluster to stabilize. if err := sleepContext(ctx, 3*time.Minute); err != nil { return } r := rand.New(rand.NewSource(time.Now().UnixNano())) for { if err := z.Decommission(ctx, poolIndices...); err != nil { if errors.Is(err, errDecommissionAlreadyRunning) { // A previous decommission running found restart it.
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 42.1K bytes - Viewed (1) -
cmd/data-scanner.go
) // initDataScanner will start the scanner in the background. func initDataScanner(ctx context.Context, objAPI ObjectLayer) { go func() { r := rand.New(rand.NewSource(time.Now().UnixNano())) // Run the data scanner in a loop for { runDataScanner(ctx, objAPI) duration := max(time.Duration(r.Float64()*float64(scannerCycle.Load())),
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 45.5K bytes - Viewed (0) -
cmd/notification.go
// // The first call to return a non-nil error will be // collected in errs slice and returned by Wait(). func (g *NotificationGroup) Go(ctx context.Context, f func() error, index int, addr xnet.Host) { r := rand.New(rand.NewSource(time.Now().UnixNano())) g.workers.Take() go func() { defer g.workers.Give() g.errs[index] = NotificationPeerErr{ Host: addr, } retryCount := g.retryCount
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 45.9K bytes - Viewed (0)