- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 22 for poolEps (0.09 sec)
-
cmd/rebalance-admin.go
} if !ps.Participating { continue } // for participating pools, total bytes to be rebalanced by this pool is given by, // pf_c = (f_i + x)/c_i, // pf_c - percentage free space across pools, f_i - ith pool's free space, c_i - ith pool's capacity // i.e. x = c_i*pfc -f_i totalBytesToRebal := float64(ps.InitCapacity)*meta.PercentFreeGoal - float64(ps.InitFreeSpace) elapsed := time.Since(ps.Info.StartTime)
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Fri Dec 22 00:56:43 UTC 2023 - 3.8K bytes - Viewed (0) -
.github/workflows/mint/minio-pools.yaml
Harshavardhana <******@****.***> 1699046298 -0700
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Fri Nov 03 21:18:18 UTC 2023 - 2.3K bytes - Viewed (0) -
docs/distributed/DESIGN.md
Refer to the sizing guide with details on the default parity count chosen for different erasure stripe sizes [here](https://github.com/minio/minio/blob/master/docs/distributed/SIZING.md) MinIO places new objects in server pools based on proportionate free space, per pool. Following pseudo code demonstrates this behavior. ```go func getAvailablePoolIdx(ctx context.Context) int {
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Tue Aug 15 23:04:20 UTC 2023 - 8K bytes - Viewed (0) -
docs/distributed/DECOMMISSION.md
# Decommissioning Decommissiong is a mechanism in MinIO to drain older pools (usually with old hardware) and migrate the content from such pools to a newer pools (usually better hardware). Decommissioning spreads the data across all pools - for example, if you decommission `pool1`, all the data from `pool1` spreads across `pool2` and `pool3`. ## Features
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Mon Jul 11 14:59:49 UTC 2022 - 8.3K bytes - Viewed (0) -
cmd/erasure-server-pool-decom_test.go
} var nmeta1 poolMeta nmeta1.Version = poolMetaVersion nmeta1.Pools = append(nmeta1.Pools, meta.Pools...) for i, pool := range nmeta1.Pools { if i == 0 { nmeta1.Pools[i] = PoolStatus{ CmdLine: pool.CmdLine, ID: i, LastUpdate: UTCNow(), Decommission: &PoolDecommissionInfo{ Complete: true, }, } } } var nmeta2 poolMeta nmeta2.Version = poolMetaVersion
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Mon Jul 01 14:38:46 UTC 2024 - 4.8K bytes - Viewed (0) -
internal/config/server.go
type ServerConfigV1 struct { ServerConfigVersion ServerConfigCommon Pools [][]string `yaml:"pools"` } // ServerConfig represents a MinIO configuration file type ServerConfig struct { ServerConfigVersion ServerConfigCommon Pools []struct { Args []string `yaml:"args"` SetDriveCount uint64 `yaml:"set-drive-count"` } `yaml:"pools"`
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Fri May 03 15:54:03 UTC 2024 - 1.9K bytes - Viewed (0) -
cmd/testdata/config/2.yaml
version: v1 address: ':9000' console-address: ':9001' certs-dir: '/home/user/.minio/certs/' pools: # Specify the nodes and drives with pools - - 'https://server-example-pool1:9000/mnt/disk{1...4}/' - 'https://server1-pool1:9000/mnt/disk{1...4}/' - 'https://server3-pool1:9000/mnt/disk{1...4}/' - 'https://server4-pool1:9000/mnt/disk{1...4}/' - - 'https://server-example-pool2:9000/mnt/disk{1...4}/'
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Thu Dec 07 09:33:56 UTC 2023 - 869 bytes - Viewed (0) -
docs/distributed/CONFIG.md
address: ":8022" ssh-private-key: "/home/user/.ssh/id_rsa" ``` If you are using the config `v1` YAML you should migrate your `pools:` field values to the following format `v1` format ```yaml pools: # Specify the nodes and drives with pools - - "https://server-example-pool1:9000/mnt/disk{1...4}/" - "https://server{1...2}-pool1:9000/mnt/disk{1...4}/"
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Tue Jun 25 02:30:18 UTC 2024 - 4.2K bytes - Viewed (0) -
cmd/testdata/config/invalid.yaml
version: address: ':9000' console-address: ':9001' certs-dir: '/home/user/.minio/certs/' pools: # Specify the nodes and drives with pools - - 'https://server-example-pool1:9000/mnt/disk{1...4}/' - 'https://server1-pool1:9000/mnt/disk{1...4}/' - 'https://server3-pool1:9000/mnt/disk{1...4}/' - 'https://server4-pool1:9000/mnt/disk{1...4}/' - - 'https://server-example-pool2:9000/mnt/disk{1...4}/'
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Thu Dec 07 09:33:56 UTC 2023 - 866 bytes - Viewed (0) -
cmd/server-main_test.go
t.Error("expected success, got failure", err) } if err == nil { if len(sctx.Layout.pools) != 2 { t.Error("expected parsed pools to be 2, not", len(sctx.Layout.pools)) } if sctx.Layout.pools[0].cmdline != testcase.hash { t.Error("expected hash", testcase.hash, "got", sctx.Layout.pools[0].cmdline) } } }) } } // Tests initializing new object layer.
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Thu Dec 07 09:33:56 UTC 2023 - 3.1K bytes - Viewed (0)