- Sort Score
- Num 10 results
- Language All
Results 1 - 3 of 3 for resyncID (0.27 seconds)
-
cmd/bucket-replication.go
} p.resyncer.Lock() p.resyncer.statusMap[bucket] = meta p.resyncer.Unlock() tgts := meta.cloneTgtStats() for arn, st := range tgts { switch st.ResyncStatus { case ResyncFailed, ResyncStarted, ResyncPending: go p.resyncer.resyncBucket(ctx, objAPI, true, resyncOpts{ bucket: bucket, arn: arn, resyncID: st.ResyncID,
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 118.2K bytes - Click Count (0) -
cmd/site-replication.go
OpType: "start", ResyncID: rs.ResyncID, Buckets: res.Buckets, } if len(res.Buckets) > 0 { res.ErrDetail = "partial failure in starting site resync" } if len(buckets) != 0 && len(res.Buckets) == len(buckets) { return res, fmt.Errorf("all buckets resync failed") } return res, nil } // cancelResync stops an ongoing site level resync for the peer specified.
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 184.8K bytes - Click Count (1) -
CHANGELOG/CHANGELOG-1.4.md
If this happens to you, you can wait at most 10 minutes for the replication controller to start a resync, the extra pods will then be deleted. Or, you can manually trigger a resync by change the replicas in the spec of the replication controller. ### kubectl delete: < v1.4.0 client vs >=v1.4.0 cluster
Created: Fri Dec 26 09:05:12 GMT 2025 - Last Modified: Thu Dec 24 02:28:26 GMT 2020 - 133.5K bytes - Click Count (0)