- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 70 for goOffline (0.06 sec)
-
docs/metrics/prometheus/grafana/node/minio-node.json
"hide": false, "instant": true, "legendFormat": ".", "range": false, "refId": "B" } ], "title": "Total Online/Offline Drives", "type": "gauge" }, { "datasource": { "type": "prometheus", "uid": "${DS_PROMETHEUS}" }, "description": "", "fieldConfig": {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Mon Aug 04 01:46:49 UTC 2025 - 22.5K bytes - Viewed (0) -
cmd/bootstrap-peer-server.go
logger.Info(fmt.Sprintf("Waiting for at least %d remote servers with valid configuration to be online", len(clnts)/2)) if len(offlineEndpoints) > 0 { logger.Info(fmt.Sprintf("Following servers are currently offline or unreachable %s", offlineEndpoints)) } if len(incorrectConfigs) > 0 { logger.Info(fmt.Sprintf("Following servers have mismatching configuration %s", incorrectConfigs)) }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue Oct 01 22:13:18 UTC 2024 - 8.4K bytes - Viewed (0) -
cmd/metrics-realtime.go
_, ok := disks[d.Endpoint] if !ok { continue } } if d.State != madmin.DriveStateOk && d.State != madmin.DriveStateUnformatted { metrics[d.Endpoint] = madmin.DiskMetric{NDisks: 1, Offline: 1} continue } var dm madmin.DiskMetric dm.NDisks = 1 if d.Healing { dm.Healing++ } if d.Metrics != nil { dm.LifeTimeOps = make(map[string]uint64, len(d.Metrics.APICalls))
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sat Jun 01 05:16:24 UTC 2024 - 6.3K bytes - Viewed (0) -
cmd/mrf.go
wg sync.WaitGroup } func newMRFState() mrfState { return mrfState{ opCh: make(chan PartialOperation, mrfOpsQueueSize), } } // Add a partial S3 operation (put/delete) when one or more disks are offline. func (m *mrfState) addPartialOp(op PartialOperation) { if m == nil { return } if atomic.LoadInt32(&m.closed) == 1 { return } m.wg.Add(1) defer m.wg.Done()
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue May 27 15:19:03 UTC 2025 - 6.5K bytes - Viewed (0) -
cmd/bucket-replication.go
return } if globalBucketTargetSys.isOffline(tgt.EndpointURL()) { replLogOnceIf(ctx, fmt.Errorf("remote target is offline for bucket:%s arn:%s", dobj.Bucket, tgt.ARN), "replication-target-offline-delete-"+tgt.ARN) sendEvent(eventArgs{ BucketName: dobj.Bucket, Object: ObjectInfo{ Bucket: dobj.Bucket, Name: dobj.ObjectName,
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 118K bytes - Viewed (0) -
cmd/prepare-storage.go
if err != nil { var drivesNotFound int for _, format := range formatConfigs { if format != nil { continue } drivesNotFound++ } return nil, fmt.Errorf("%w (offline-drives=%d/%d)", err, drivesNotFound, len(formatConfigs)) } if format.ID == "" { return nil, errors.New("deployment ID missing from disk format, unable to start the server") } return format, nil }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Jul 12 20:51:54 UTC 2024 - 11.1K bytes - Viewed (1) -
cmd/erasure-healing-common.go
} inconsistent := 0 for i, meta := range partsMetadata { if !meta.IsValid() { // Since for majority of the cases erasure.Index matches with erasure.Distribution we can // consider the offline disks as consistent. continue } if !meta.Deleted { if len(meta.Erasure.Distribution) != len(onlineDisks) { // Erasure distribution seems to have lesser
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sun Mar 30 00:56:02 UTC 2025 - 12K bytes - Viewed (0) -
cmd/erasure-heal_test.go
package cmd import ( "bytes" "crypto/rand" "io" "os" "testing" ) var erasureHealTests = []struct { dataBlocks, disks int // number of offline disks is also number of staleDisks for // erasure reconstruction in this test offDisks int // bad disks are online disks which return errors badDisks, badStaleDisks int blocksize, size int64
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 7.9K bytes - Viewed (0) -
android/guava/src/com/google/common/hash/Hashing.java
* {@code consistentHash} will handle it. If, however, you are dividing traffic among * servers {@code alpha}, {@code bravo}, and {@code charlie} and you occasionally need to * take each of the servers offline, {@code consistentHash} will be a poor fit: It provides * no way for you to specify which of the three buckets is disappearing. Thus, if your
Registered: Fri Sep 05 12:43:10 UTC 2025 - Last Modified: Thu Jul 17 15:26:41 UTC 2025 - 29.8K bytes - Viewed (0) -
cmd/erasure-decode.go
} else if bitrotHeal == 1 { return newBuf, errFileCorrupt } return newBuf, nil } // If we cannot decode, just return read quorum error. return nil, fmt.Errorf("%w (offline-disks=%d/%d)", errErasureReadQuorum, disksNotFound, len(p.readers)) } // Decode reads from readers, reconstructs data if needed and writes the data to the writer.
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Thu Aug 29 01:40:52 UTC 2024 - 9.5K bytes - Viewed (0)