Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 22 for poolEps (0.09 sec)

  1. cmd/rebalance-admin.go

    		}
    		if !ps.Participating {
    			continue
    		}
    		// for participating pools, total bytes to be rebalanced by this pool is given by,
    		// pf_c = (f_i + x)/c_i,
    		// pf_c - percentage free space across pools, f_i - ith pool's free space, c_i - ith pool's capacity
    		// i.e. x = c_i*pfc -f_i
    		totalBytesToRebal := float64(ps.InitCapacity)*meta.PercentFreeGoal - float64(ps.InitFreeSpace)
    		elapsed := time.Since(ps.Info.StartTime)
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Fri Dec 22 00:56:43 UTC 2023
    - 3.8K bytes
    - Viewed (0)
  2. .github/workflows/mint/minio-pools.yaml

    Harshavardhana <******@****.***> 1699046298 -0700
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Fri Nov 03 21:18:18 UTC 2023
    - 2.3K bytes
    - Viewed (0)
  3. docs/distributed/DESIGN.md

    Refer to the sizing guide with details on the default parity count chosen for different erasure stripe sizes [here](https://github.com/minio/minio/blob/master/docs/distributed/SIZING.md)
    
    MinIO places new objects in server pools based on proportionate free space, per pool. Following pseudo code demonstrates this behavior.
    
    ```go
    func getAvailablePoolIdx(ctx context.Context) int {
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Tue Aug 15 23:04:20 UTC 2023
    - 8K bytes
    - Viewed (0)
  4. docs/distributed/DECOMMISSION.md

    # Decommissioning
    
    Decommissiong is a mechanism in MinIO to drain older pools (usually with old hardware) and migrate the content from such pools to a newer pools (usually better hardware). Decommissioning spreads the data across all pools - for example, if you decommission `pool1`, all the data from `pool1` spreads across `pool2` and `pool3`.
    
    ## Features
    
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Mon Jul 11 14:59:49 UTC 2022
    - 8.3K bytes
    - Viewed (0)
  5. cmd/erasure-server-pool-decom_test.go

    	}
    
    	var nmeta1 poolMeta
    	nmeta1.Version = poolMetaVersion
    	nmeta1.Pools = append(nmeta1.Pools, meta.Pools...)
    	for i, pool := range nmeta1.Pools {
    		if i == 0 {
    			nmeta1.Pools[i] = PoolStatus{
    				CmdLine:    pool.CmdLine,
    				ID:         i,
    				LastUpdate: UTCNow(),
    				Decommission: &PoolDecommissionInfo{
    					Complete: true,
    				},
    			}
    		}
    	}
    
    	var nmeta2 poolMeta
    	nmeta2.Version = poolMetaVersion
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Mon Jul 01 14:38:46 UTC 2024
    - 4.8K bytes
    - Viewed (0)
  6. internal/config/server.go

    type ServerConfigV1 struct {
    	ServerConfigVersion
    	ServerConfigCommon
    	Pools [][]string `yaml:"pools"`
    }
    
    // ServerConfig represents a MinIO configuration file
    type ServerConfig struct {
    	ServerConfigVersion
    	ServerConfigCommon
    	Pools []struct {
    		Args          []string `yaml:"args"`
    		SetDriveCount uint64   `yaml:"set-drive-count"`
    	} `yaml:"pools"`
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Fri May 03 15:54:03 UTC 2024
    - 1.9K bytes
    - Viewed (0)
  7. cmd/testdata/config/2.yaml

    version: v1
    address: ':9000'
    console-address: ':9001'
    certs-dir: '/home/user/.minio/certs/'
    pools: # Specify the nodes and drives with pools
      -
            - 'https://server-example-pool1:9000/mnt/disk{1...4}/'
            - 'https://server1-pool1:9000/mnt/disk{1...4}/'
            - 'https://server3-pool1:9000/mnt/disk{1...4}/'
            - 'https://server4-pool1:9000/mnt/disk{1...4}/'
      -
            - 'https://server-example-pool2:9000/mnt/disk{1...4}/'
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Thu Dec 07 09:33:56 UTC 2023
    - 869 bytes
    - Viewed (0)
  8. docs/distributed/CONFIG.md

        address: ":8022"
        ssh-private-key: "/home/user/.ssh/id_rsa"
    ```
    
    If you are using the config `v1` YAML you should migrate your `pools:` field values to the following format
    
    `v1` format
    ```yaml
    pools: # Specify the nodes and drives with pools
      -
        - "https://server-example-pool1:9000/mnt/disk{1...4}/"
        - "https://server{1...2}-pool1:9000/mnt/disk{1...4}/"
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Tue Jun 25 02:30:18 UTC 2024
    - 4.2K bytes
    - Viewed (0)
  9. cmd/testdata/config/invalid.yaml

    version:
    address: ':9000'
    console-address: ':9001'
    certs-dir: '/home/user/.minio/certs/'
    pools: # Specify the nodes and drives with pools
      -
            - 'https://server-example-pool1:9000/mnt/disk{1...4}/'
            - 'https://server1-pool1:9000/mnt/disk{1...4}/'
            - 'https://server3-pool1:9000/mnt/disk{1...4}/'
            - 'https://server4-pool1:9000/mnt/disk{1...4}/'
      -
            - 'https://server-example-pool2:9000/mnt/disk{1...4}/'
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Thu Dec 07 09:33:56 UTC 2023
    - 866 bytes
    - Viewed (0)
  10. cmd/server-main_test.go

    				t.Error("expected success, got failure", err)
    			}
    			if err == nil {
    				if len(sctx.Layout.pools) != 2 {
    					t.Error("expected parsed pools to be 2, not", len(sctx.Layout.pools))
    				}
    				if sctx.Layout.pools[0].cmdline != testcase.hash {
    					t.Error("expected hash", testcase.hash, "got", sctx.Layout.pools[0].cmdline)
    				}
    			}
    		})
    	}
    }
    
    // Tests initializing new object layer.
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Thu Dec 07 09:33:56 UTC 2023
    - 3.1K bytes
    - Viewed (0)
Back to top