- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 724 for resync (0.05 seconds)
-
cmd/site-replication-utils.go
const ( siteResyncPrefix = bucketMetaPrefix + "/site-replication/resync" ) type resyncState struct { resyncID string LastSaved time.Time } //msgp:ignore siteResyncMetrics type siteResyncMetrics struct { sync.RWMutex // resyncStatus maps resync ID to resync status for peer resyncStatus map[string]SiteResyncStatus // map peer deployment ID to resync ID peerResyncMap map[string]resyncState }Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 8.8K bytes - Click Count (0) -
cmd/bucket-replication-utils.go
arn string resyncID string resyncBefore time.Time } // ResyncStatusType status of resync operation type ResyncStatusType int const ( // NoResync - no resync in progress NoResync ResyncStatusType = iota // ResyncPending - resync pending ResyncPending // ResyncCanceled - resync canceled ResyncCanceled // ResyncStarted - resync in progress ResyncStarted
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 26.1K bytes - Click Count (0) -
docs/bucket/replication/DESIGN.md
replicate objects one at a time so as not to impede server load. Ideally, resync should not be initiated for multiple buckets simultaneously - progress of the syncing can be monitored by looking at `mc replicate resync status alias/bucket --remote-bucket <arn>`. In the event that resync operation failed to replicate some versions, they would be picked up by the healing mechanism in-built as part of the scanner. If the resync operation reports a failed status or in the event of a cluster restart while...
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 14.7K bytes - Click Count (0) -
cmd/bucket-replication.go
Error: errStr, Bytes: sz, } } // delete resync metadata from replication resync state in memory func (p *ReplicationPool) deleteResyncMetadata(ctx context.Context, bucket string) { if p == nil { return } p.resyncer.Lock() delete(p.resyncer.statusMap, bucket) defer p.resyncer.Unlock() globalSiteResyncMetrics.deleteBucket(bucket) }
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 118.2K bytes - Click Count (0) -
cmd/site-replication.go
} } return pi, true } // startResync initiates resync of data to peerSite specified. The overall site resync status // is maintained in .minio.sys/buckets/site-replication/resync/<deployment-id.meta>, while collecting // individual bucket resync status in .minio.sys/buckets/<bucket-name>/replication/resync.bin
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 184.8K bytes - Click Count (1) -
android/guava/src/com/google/common/util/concurrent/SmoothRateLimiter.java
*/ abstract double coolDownIntervalMicros(); /** Updates {@code storedPermits} and {@code nextFreeTicketMicros} based on the current time. */ void resync(long nowMicros) { // if nextFreeTicket is in the past, resync to now if (nowMicros > nextFreeTicketMicros) { double newPermits = (nowMicros - nextFreeTicketMicros) / coolDownIntervalMicros();
Created: Fri Apr 03 12:43:13 GMT 2026 - Last Modified: Wed May 14 19:40:47 GMT 2025 - 19.3K bytes - Click Count (0) -
docs/site-replication/run-multi-site-oidc.sh
echo "expected 'bucket2' delete and 'newbucket2' creation to have replicated, exiting..." exit_1 fi # force a resync after removing all site replication ./mc admin replicate rm --all --force minio1 ./mc rb minio2 --force --dangerous ./mc admin replicate add minio1 minio2 ./mc admin replicate resync start minio1 minio2 sleep 30 ./mc ls -r --versions minio1/newbucket >/tmp/minio1.txt
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Sep 24 08:03:58 GMT 2024 - 8.5K bytes - Click Count (0) -
docs/site-replication/run-multi-site-minio-idp.sh
echo "expected 'bucket2' delete and 'newbucket2' creation to have replicated, exiting..." exit_1 fi # force a resync after removing all site replication ./mc admin replicate rm --all --force minio1 ./mc rb minio2 --force --dangerous ./mc admin replicate add minio1 minio2 ./mc admin replicate resync start minio1 minio2 sleep 30 ./mc ls -r --versions minio1/newbucket >/tmp/minio1.txt
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Sep 24 08:03:58 GMT 2024 - 12.1K bytes - Click Count (0) -
docs/site-replication/run-multi-site-ldap.sh
echo "expected 'bucket2' delete and 'newbucket2' creation to have replicated, exiting..." exit_1 fi # force a resync after removing all site replication ./mc admin replicate rm --all --force minio1 ./mc rb minio2 --force --dangerous ./mc admin replicate add minio1 minio2 ./mc admin replicate resync start minio1 minio2 sleep 30 ./mc ls -r --versions minio1/newbucket >/tmp/minio1.txt
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Sep 24 08:03:58 GMT 2024 - 10.1K bytes - Click Count (0) -
cmd/bucket-replication-handlers.go
Bucket: bucket, Err: fmt.Errorf("replication resync status not available for %s (%s)", arn, err.Error()), }), r.URL) return } var rinfo ResyncTargetsInfo for tarn, st := range brs.TargetsMap { if arn != "" && tarn != arn { continue } rinfo.Targets = append(rinfo.Targets, ResyncTarget{ Arn: tarn, ResetID: st.ResyncID, StartTime: st.StartTime,
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Thu Aug 15 12:04:40 GMT 2024 - 23.3K bytes - Click Count (0)