- Sort Score
- Num 10 results
- Language All
Results 11 - 20 of 25 for resyncID (0.07 seconds)
-
docs/site-replication/run-multi-site-minio-idp.sh
echo "expected 'bucket2' delete and 'newbucket2' creation to have replicated, exiting..." exit_1 fi # force a resync after removing all site replication ./mc admin replicate rm --all --force minio1 ./mc rb minio2 --force --dangerous ./mc admin replicate add minio1 minio2 ./mc admin replicate resync start minio1 minio2 sleep 30 ./mc ls -r --versions minio1/newbucket >/tmp/minio1.txt
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Sep 24 08:03:58 GMT 2024 - 12.1K bytes - Click Count (0) -
docs/site-replication/run-multi-site-ldap.sh
echo "expected 'bucket2' delete and 'newbucket2' creation to have replicated, exiting..." exit_1 fi # force a resync after removing all site replication ./mc admin replicate rm --all --force minio1 ./mc rb minio2 --force --dangerous ./mc admin replicate add minio1 minio2 ./mc admin replicate resync start minio1 minio2 sleep 30 ./mc ls -r --versions minio1/newbucket >/tmp/minio1.txt
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Sep 24 08:03:58 GMT 2024 - 10.1K bytes - Click Count (0) -
docs/bucket/replication/README.md
replicated objects to the DR cluster need to be re-replicated, `mc replicate resync start alias/bucket --remote-bucket <arn>` can be used to initiate a reset. This would initiate a re-sync between the two clusters by walking the bucket namespace and replicating eligible objects that satisfy the existing objects replication rule specified in the replication config. The status of the resync operation can be viewed with `mc replicate resync status alias/bucket --remote-bucket <arn>`. Note that Ex...
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 18.3K bytes - Click Count (0) -
docs/bucket/replication/setup_2site_existing_replication.sh
./mc replicate add sitea/bucket/ \ --remote-bucket http://minio:minio123@127.0.0.1:9004/bucket remote_arn=$(./mc replicate ls sitea/bucket --json | jq -r .rule.Destination.Bucket) sleep 1 ./mc replicate resync start sitea/bucket/ --remote-bucket "${remote_arn}" sleep 30s ## sleep for 30s idea is that we give 300ms per object. ./mc ls -r --versions sitea/bucket >/tmp/sitea.txt ./mc ls -r --versions siteb/bucket >/tmp/siteb.txtCreated: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Sep 06 09:42:21 GMT 2024 - 5.6K bytes - Click Count (0) -
cmd/globals.go
// Allocated etcd endpoint for config and bucket DNS. globalEtcdClient *etcd.Client // Cluster replication manager. globalSiteReplicationSys SiteReplicationSys // Cluster replication resync metrics globalSiteResyncMetrics *siteResyncMetrics // Is set to true when Bucket federation is requested // and is 'true' when etcdConfig.PathPrefix is empty globalBucketFederation bool
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Sep 03 18:23:41 GMT 2024 - 16.2K bytes - Click Count (1) -
cmd/mrf.go
defer rc.Close() var data [4]byte n, err := rc.Read(data[:]) if err != nil { return err } if n != len(data) { return errors.New("heal mrf: no data") } // Read resync meta header switch binary.LittleEndian.Uint16(data[0:2]) { case healMRFMetaFormat: default: return fmt.Errorf("heal mrf: unknown format: %d", binary.LittleEndian.Uint16(data[0:2])) }
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue May 27 15:19:03 GMT 2025 - 6.5K bytes - Click Count (0) -
cmd/admin-router.go
adminRouter.Methods(http.MethodPut).Path(adminVersion + "/site-replication/peer/remove").HandlerFunc(adminMiddleware(adminAPI.SRPeerRemove)) adminRouter.Methods(http.MethodPut).Path(adminVersion+"/site-replication/resync/op").HandlerFunc(adminMiddleware(adminAPI.SiteReplicationResyncOp)).Queries("operation", "{operation:.*}")
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Oct 10 18:57:03 GMT 2025 - 26.7K bytes - Click Count (0) -
cmd/admin-handlers-site-replication.go
adminLogIf(ctx, err) writeErrorResponseJSON(ctx, w, toAdminAPIErr(ctx, err), r.URL) return } } // SiteReplicationResyncOp - PUT /minio/admin/v3/site-replication/resync/op func (a adminAPIHandlers) SiteReplicationResyncOp(w http.ResponseWriter, r *http.Request) { ctx := r.Context() objectAPI, _ := validateAdminReq(ctx, w, r, policy.SiteReplicationResyncAction) if objectAPI == nil {
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 19.4K bytes - Click Count (0) -
CHANGELOG/CHANGELOG-1.4.md
If this happens to you, you can wait at most 10 minutes for the replication controller to start a resync, the extra pods will then be deleted. Or, you can manually trigger a resync by change the replicas in the spec of the replication controller. ### kubectl delete: < v1.4.0 client vs >=v1.4.0 cluster
Created: Fri Dec 26 09:05:12 GMT 2025 - Last Modified: Thu Dec 24 02:28:26 GMT 2020 - 133.5K bytes - Click Count (0) -
cmd/server-main.go
// Initialize bucket metadata sub-system. bootstrapTrace("globalBucketMetadataSys.Init", func() { globalBucketMetadataSys.Init(GlobalContext, buckets, newObject) }) // initialize replication resync state. bootstrapTrace("initResync", func() { globalReplicationPool.Get().initResync(GlobalContext, buckets, newObject) }) // Initialize site replication manager after bucket metadata
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue May 27 15:18:36 GMT 2025 - 35.9K bytes - Click Count (4)