- Sort Score
- Num 10 results
- Language All
Results 1 - 4 of 4 for resyncID (0.28 seconds)
-
cmd/bucket-replication.go
} p.resyncer.Lock() p.resyncer.statusMap[bucket] = meta p.resyncer.Unlock() tgts := meta.cloneTgtStats() for arn, st := range tgts { switch st.ResyncStatus { case ResyncFailed, ResyncStarted, ResyncPending: go p.resyncer.resyncBucket(ctx, objAPI, true, resyncOpts{ bucket: bucket, arn: arn, resyncID: st.ResyncID,
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 118.2K bytes - Click Count (0) -
cmd/site-replication.go
OpType: "start", ResyncID: rs.ResyncID, Buckets: res.Buckets, } if len(res.Buckets) > 0 { res.ErrDetail = "partial failure in starting site resync" } if len(buckets) != 0 && len(res.Buckets) == len(buckets) { return res, fmt.Errorf("all buckets resync failed") } return res, nil } // cancelResync stops an ongoing site level resync for the peer specified.
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 184.8K bytes - Click Count (1) -
CHANGELOG/CHANGELOG-1.4.md
If this happens to you, you can wait at most 10 minutes for the replication controller to start a resync, the extra pods will then be deleted. Or, you can manually trigger a resync by change the replicas in the spec of the replication controller. ### kubectl delete: < v1.4.0 client vs >=v1.4.0 cluster
Created: Fri Apr 03 09:05:14 GMT 2026 - Last Modified: Thu Dec 24 02:28:26 GMT 2020 - 133.5K bytes - Click Count (0) -
CHANGELOG/CHANGELOG-1.36.md
- Client-go: Informer resync processing improved handling of Resync handling. This reduces contention on store locks between incoming events and handler updates, which may result in observable timing differences of handler invocations. This behavior is guarded by an AtomicFIFO feature gate. This...
Created: Fri Apr 03 09:05:14 GMT 2026 - Last Modified: Thu Mar 19 23:38:00 GMT 2026 - 142.1K bytes - Click Count (0)