- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 26 for pool_id (0.18 sec)
-
cmd/erasure-server-pool.go
Maintenance: opts.Maintenance, SetID: setIdx, PoolID: poolIdx, Healthy: erasureSetUpCount[poolIdx][setIdx].online >= poolWriteQuorums[poolIdx], HealthyRead: erasureSetUpCount[poolIdx][setIdx].online >= poolReadQuorums[poolIdx], HealthyDrives: erasureSetUpCount[poolIdx][setIdx].online, HealingDrives: erasureSetUpCount[poolIdx][setIdx].healing, ReadQuorum: poolReadQuorums[poolIdx],
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 03 20:08:20 GMT 2024 - 80.5K bytes - Viewed (0) -
cmd/erasure-server-pool-rebalance.go
z.rebalMu.Lock() z.rebalMeta.PoolStats[poolIdx].Info.Status = status z.rebalMeta.PoolStats[poolIdx].Info.EndTime = now z.rebalMu.Unlock() case <-timer.C: traceMsg = fmt.Sprintf("saved at %s", time.Now()) } stopFn := globalRebalanceMetrics.log(rebalanceMetricSaveMetadata, poolIdx, traceMsg) err := z.saveRebalanceStats(GlobalContext, poolIdx, rebalSaveStats) stopFn(err)
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri Apr 26 19:29:28 GMT 2024 - 27.2K bytes - Viewed (0) -
cmd/naughty-disk_test.go
defer d.mu.Unlock() d.callNR++ if err, ok := d.errors[d.callNR]; ok { return err } if d.defaultErr != nil { return d.defaultErr } return nil } func (d *naughtyDisk) GetDiskLoc() (poolIdx, setIdx, diskIdx int) { return -1, -1, -1 } func (d *naughtyDisk) GetDiskID() (string, error) { return d.disk.GetDiskID() } func (d *naughtyDisk) SetDiskID(id string) { d.disk.SetDiskID(id)
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Tue Apr 23 17:15:52 GMT 2024 - 9.3K bytes - Viewed (0) -
cmd/xl-storage-disk-id-check.go
} weSleep := func() bool { return scannerIdleMode.Load() == 0 } return p.storage.NSScanner(ctx, cache, updates, scanMode, weSleep) } func (p *xlStorageDiskIDCheck) GetDiskLoc() (poolIdx, setIdx, diskIdx int) { return p.storage.GetDiskLoc() } func (p *xlStorageDiskIDCheck) Close() error { p.diskCancel() return p.storage.Close() }
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Sun Apr 28 17:53:50 GMT 2024 - 33K bytes - Viewed (0) -
cmd/metrics-v2.go
Description: getClusterHealthStatusMD(), Value: float64(health), }) for _, h := range result.ESHealth { labels := map[string]string{ "pool": strconv.Itoa(h.PoolID), "set": strconv.Itoa(h.SetID), } metrics = append(metrics, MetricV2{ Description: getClusterErasureSetReadQuorumMD(), VariableLabels: labels, Value: float64(h.ReadQuorum),
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 25 22:01:31 GMT 2024 - 132.6K bytes - Viewed (0) -
cmd/metrics-v3-cluster-erasure-set.go
erasureSetWriteQuorum = "write_quorum" erasureSetOnlineDrivesCount = "online_drives_count" erasureSetHealingDrivesCount = "healing_drives_count" erasureSetHealth = "health" ) const ( poolIDL = "pool_id" setIDL = "set_id" ) var ( erasureSetOverallWriteQuorumMD = NewGaugeMD(erasureSetOverallWriteQuorum, "Overall write quorum across pools and sets")
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Sun Mar 10 09:15:15 GMT 2024 - 3.1K bytes - Viewed (0) -
cmd/peer-s3-client.go
return err }, idx) } errs := g.Wait() var poolErrs []error for poolIdx := 0; poolIdx < sys.poolsCount; poolIdx++ { perPoolErrs := make([]error, 0, len(sys.peerClients)) for i, client := range sys.peerClients { if slices.Contains(client.GetPools(), poolIdx) { perPoolErrs = append(perPoolErrs, errs[i]) } } quorum := len(perPoolErrs) / 2
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 14.8K bytes - Viewed (0) -
cmd/test-utils_test.go
return certOut.Bytes(), keyOut.Bytes(), nil } func mustGetPoolEndpoints(poolIdx int, args ...string) EndpointServerPools { drivesPerSet := len(args) setCount := 1 if len(args) >= 16 { drivesPerSet = 16 setCount = len(args) / 16 } endpoints := mustGetNewEndpoints(poolIdx, drivesPerSet, args...) return []PoolEndpoints{{ SetCount: setCount, DrivesPerSet: drivesPerSet,
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 03 17:26:51 GMT 2024 - 76.2K bytes - Viewed (0) -
cmd/peer-rest-server.go
objAPI := newObjectLayerFn() if objAPI == nil { return np, grid.NewRemoteErr(errServerNotInitialized) } pools, ok := objAPI.(*erasureServerPools) if !ok { return np, grid.NewRemoteErr(errors.New("not a pooled setup")) } pools.StopRebalance() return } func (s *peerRESTServer) LoadRebalanceMetaHandler(mss *grid.MSS) (np grid.NoPayload, nerr *grid.RemoteErr) { objAPI := newObjectLayerFn()
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 51.8K bytes - Viewed (0) -
internal/grid/handlers.go
func (h *SingleHandler[Req, Resp]) PutResponse(r Resp) { h.recycleResp(r) } // AllowCallRequestPool indicates it is safe to reuse the request // on the client side, meaning the request is recycled/pooled when a request is sent. // CAREFUL: This should only be used when there are no pointers, slices that aren't freshly constructed. func (h *SingleHandler[Req, Resp]) AllowCallRequestPool(b bool) *SingleHandler[Req, Resp] {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Tue Apr 23 17:15:52 GMT 2024 - 27.1K bytes - Viewed (0)