- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 97 for polos (0.18 sec)
-
cmd/admin-handlers-pools.go
} vars := mux.Vars(r) v := vars["pool"] byID := vars["by-id"] == "true" pools := strings.Split(v, ",") poolIndices := make([]int, 0, len(pools)) for _, pool := range pools { var idx int if byID { var err error idx, err = strconv.Atoi(pool) if err != nil { // We didn't find any matching pools, invalid input writeErrorResponseJSON(ctx, w, toAdminAPIErr(ctx, errInvalidArgument), r.URL)
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 10.9K bytes - Viewed (0) -
.github/workflows/mint/minio-pools.yaml
Harshavardhana <******@****.***> 1699046298 -0700
Others - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri Nov 03 21:18:18 GMT 2023 - 2.3K bytes - Viewed (0) -
cmd/erasure-server-pool-decom.go
p.Pools[idx].LastUpdate = UTCNow() p.Pools[idx].Decommission.StartTime = time.Time{} p.Pools[idx].Decommission.Complete = false p.Pools[idx].Decommission.Failed = true p.Pools[idx].Decommission.Canceled = false return true } return false } func (p *poolMeta) DecommissionCancel(idx int) bool { if p.Pools[idx].Decommission != nil && !p.Pools[idx].Decommission.Canceled {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Wed May 01 15:18:21 GMT 2024 - 41.5K bytes - Viewed (1) -
cmd/rebalance-admin.go
type rebalanceAdminStatus struct { ID string // identifies the ongoing rebalance operation by a uuid Pools []rebalancePoolStatus `json:"pools"` // contains all pools, including inactive StoppedAt time.Time `json:"stoppedAt,omitempty"` } func rebalanceStatus(ctx context.Context, z *erasureServerPools) (r rebalanceAdminStatus, err error) {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri Dec 22 00:56:43 GMT 2023 - 3.8K bytes - Viewed (0) -
cmd/peer-s3-client.go
GetPools() []int } type localPeerS3Client struct { node Node pools []int } func (l *localPeerS3Client) GetHost() string { return l.node.Host } func (l *localPeerS3Client) SetPools(p []int) { l.pools = make([]int, len(p)) copy(l.pools, p) } func (l localPeerS3Client) GetPools() []int { return l.pools }
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 14.8K bytes - Viewed (0) -
cmd/erasure-server-pool-decom_test.go
} var nmeta1 poolMeta nmeta1.Version = poolMetaVersion nmeta1.Pools = append(nmeta1.Pools, meta.Pools...) for i, pool := range nmeta1.Pools { if i == 0 { nmeta1.Pools[i] = PoolStatus{ CmdLine: pool.CmdLine, ID: i, LastUpdate: UTCNow(), Decommission: &PoolDecommissionInfo{ Complete: true, }, } } } var nmeta2 poolMeta nmeta2.Version = poolMetaVersion
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Mon Jul 03 16:47:40 GMT 2023 - 4.8K bytes - Viewed (0) -
.github/workflows/mint.yml
run: | ${GITHUB_WORKSPACE}/.github/workflows/run-mint.sh "compress-encrypt" "minio" "minio123" "${{ steps.vars.outputs.sha_short }}" - name: multiple pools run: | ${GITHUB_WORKSPACE}/.github/workflows/run-mint.sh "pools" "minio" "minio123" "${{ steps.vars.outputs.sha_short }}" - name: standalone erasure run: |
Others - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Mar 28 23:44:49 GMT 2024 - 2.6K bytes - Viewed (0) -
cmd/testdata/config/invalid.yaml
version: address: ':9000' console-address: ':9001' certs-dir: '/home/user/.minio/certs/' pools: # Specify the nodes and drives with pools - - 'https://server-example-pool1:9000/mnt/disk{1...4}/' - 'https://server1-pool1:9000/mnt/disk{1...4}/' - 'https://server3-pool1:9000/mnt/disk{1...4}/' - 'https://server4-pool1:9000/mnt/disk{1...4}/' - - 'https://server-example-pool2:9000/mnt/disk{1...4}/'
Others - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Dec 07 09:33:56 GMT 2023 - 866 bytes - Viewed (0) -
cmd/testdata/config/invalid-types.yaml
version: v1 address: ':9000' console-address: ':9001' certs-dir: '/home/user/.minio/certs/' pools: # Specify the nodes and drives with pools - - '/mnt/disk{1...4}/' - 'https://server1-pool1:9000/mnt/disk{1...4}/' - 'https://server3-pool1:9000/mnt/disk{1...4}/' - 'https://server4-pool1:9000/mnt/disk{1...4}/' - - 'https://server-example-pool2:9000/mnt/disk{1...4}/'
Others - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Dec 07 09:33:56 GMT 2023 - 836 bytes - Viewed (0) -
docs/distributed/CONFIG.md
address: ":8022" ssh-private-key: "/home/user/.ssh/id_rsa" ``` If you are using the config `v1` YAML you should migrate your `pools:` field values to the following format `v1` format ```yaml pools: # Specify the nodes and drives with pools - - "https://server-example-pool1:9000/mnt/disk{1...4}/" - "https://server{1...2}-pool1:9000/mnt/disk{1...4}/"
Plain Text - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 03 15:54:03 GMT 2024 - 4.2K bytes - Viewed (0)