- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 57 for poolEps (0.06 seconds)
-
cmd/storage-rest_test.go
endpoint.PoolIdx = 0 endpoint.SetIdx = 0 endpoint.DiskIdx = 0 poolEps := []PoolEndpoints{{ Endpoints: Endpoints{endpoint}, }} poolEps[0].SetCount = 1 poolEps[0].DrivesPerSet = 1 // Register handlers on newly created servers registerStorageRESTHandlers(tg.Mux[0], poolEps, tg.Managers[0]) registerStorageRESTHandlers(tg.Mux[1], poolEps, tg.Managers[1]) storage := globalLocalSetDrives[0][0][0]Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Wed Apr 09 14:28:39 GMT 2025 - 11.4K bytes - Click Count (0) -
cmd/admin-handlers-pools.go
} vars := mux.Vars(r) v := vars["pool"] byID := vars["by-id"] == "true" pools := strings.Split(v, ",") poolIndices := make([]int, 0, len(pools)) for _, pool := range pools { var idx int if byID { var err error idx, err = strconv.Atoi(pool) if err != nil { // We didn't find any matching pools, invalid input writeErrorResponseJSON(ctx, w, toAdminAPIErr(ctx, errInvalidArgument), r.URL)
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 11.1K bytes - Click Count (0) -
cmd/rebalance-admin.go
} if !ps.Participating { continue } // for participating pools, total bytes to be rebalanced by this pool is given by, // pf_c = (f_i + x)/c_i, // pf_c - percentage free space across pools, f_i - ith pool's free space, c_i - ith pool's capacity // i.e. x = c_i*pfc -f_i totalBytesToRebal := float64(ps.InitCapacity)*meta.PercentFreeGoal - float64(ps.InitFreeSpace) elapsed := time.Since(ps.Info.StartTime)Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 3.8K bytes - Click Count (0) -
.github/workflows/mint/minio-pools.yaml
Harshavardhana <******@****.***> 1699046298 -0700
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Nov 03 21:18:18 GMT 2023 - 2.3K bytes - Click Count (0) -
docs/distributed/DESIGN.md
Refer to the sizing guide with details on the default parity count chosen for different erasure stripe sizes [here](https://github.com/minio/minio/blob/master/docs/distributed/SIZING.md) MinIO places new objects in server pools based on proportionate free space, per pool. Following pseudo code demonstrates this behavior. ```go func getAvailablePoolIdx(ctx context.Context) int {
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Wed Feb 26 09:25:50 GMT 2025 - 8K bytes - Click Count (2) -
cmd/erasure-server-pool.go
}) if err != nil { return nil, err } if deploymentID == "" { // all pools should have same deployment ID deploymentID = formats[i].ID } // Validate if users brought different DeploymentID pools. if deploymentID != formats[i].ID { return nil, fmt.Errorf("all pools must have same deployment ID - expected %s, got %s for pool(%s)", deploymentID, formats[i].ID, humanize.Ordinal(i+1)) }Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 89.2K bytes - Click Count (0) -
cmd/erasure-server-pool-decom_test.go
} var nmeta1 poolMeta nmeta1.Version = poolMetaVersion nmeta1.Pools = append(nmeta1.Pools, meta.Pools...) for i, pool := range nmeta1.Pools { if i == 0 { nmeta1.Pools[i] = PoolStatus{ CmdLine: pool.CmdLine, ID: i, LastUpdate: UTCNow(), Decommission: &PoolDecommissionInfo{ Complete: true, }, } } } var nmeta2 poolMeta nmeta2.Version = poolMetaVersion
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 4.8K bytes - Click Count (0) -
internal/config/server.go
type ServerConfigV1 struct { ServerConfigVersion ServerConfigCommon Pools [][]string `yaml:"pools"` } // ServerConfig represents a MinIO configuration file type ServerConfig struct { ServerConfigVersion ServerConfigCommon Pools []struct { Args []string `yaml:"args"` SetDriveCount uint64 `yaml:"set-drive-count"` } `yaml:"pools"`Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri May 03 15:54:03 GMT 2024 - 1.9K bytes - Click Count (0) -
cmd/erasure-server-pool-decom.go
p.Pools[idx].LastUpdate = UTCNow() p.Pools[idx].Decommission.StartTime = time.Time{} p.Pools[idx].Decommission.Complete = false p.Pools[idx].Decommission.Failed = true p.Pools[idx].Decommission.Canceled = false return true } return false } func (p *poolMeta) DecommissionCancel(idx int) bool { if p.Pools[idx].Decommission != nil && !p.Pools[idx].Decommission.Canceled {
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 42.2K bytes - Click Count (1) -
cmd/erasure-server-pool-decom_gen.go
if err != nil { err = msgp.WrapError(err, "Pools") return } if cap(z.Pools) >= int(zb0002) { z.Pools = (z.Pools)[:zb0002] } else { z.Pools = make([]PoolStatus, zb0002) } for za0001 := range z.Pools { err = z.Pools[za0001].DecodeMsg(dc) if err != nil { err = msgp.WrapError(err, "Pools", za0001) return } } default:Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 26.7K bytes - Click Count (0)