Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 265 for Drives (0.04 sec)

  1. cmd/metrics-v3-cluster-health.go

    		"Count of offline drives in the cluster")
    	healthDrivesOnlineCountMD = NewGaugeMD(healthDrivesOnlineCount,
    		"Count of online drives in the cluster")
    	healthDrivesCountMD = NewGaugeMD(healthDrivesCount,
    		"Count of all drives in the cluster")
    )
    
    // loadClusterHealthDriveMetrics - `MetricsLoaderFn` for cluster storage drive metrics
    // such as online, offline and total drives.
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Sun Mar 10 09:15:15 UTC 2024
    - 3.9K bytes
    - Viewed (0)
  2. cmd/prepare-storage.go

    	if err == nil {
    		return storageDisks, format, nil
    	}
    
    	tries++ // tried already once
    
    	// Wait on each try for an update.
    	ticker := time.NewTicker(1 * time.Second)
    	defer ticker.Stop()
    
    	for {
    		// Only log once every 10 iterations, then reset the tries count.
    		verbose = tries >= 10
    		if verbose {
    			tries = 1
    		}
    
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Jul 12 20:51:54 UTC 2024
    - 11.1K bytes
    - Viewed (1)
  3. docs/multi-tenancy/README.md

    ### 1.2 Host Multiple Tenants on Multiple Drives (Erasure Code)
    
    Use the following commands to host 3 tenants on multiple drives:
    
    ```sh
    minio server --address :9001 /disk{1...4}/data/tenant1
    minio server --address :9002 /disk{1...4}/data/tenant2
    minio server --address :9003 /disk{1...4}/data/tenant3
    ```
    
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Tue Aug 12 18:20:36 UTC 2025
    - 3.1K bytes
    - Viewed (0)
  4. cmd/erasure-errors.go

    import "errors"
    
    // errErasureReadQuorum - did not meet read quorum.
    var errErasureReadQuorum = errors.New("Read failed. Insufficient number of drives online")
    
    // errErasureWriteQuorum - did not meet write quorum.
    var errErasureWriteQuorum = errors.New("Write failed. Insufficient number of drives online")
    
    // errNoHealRequired - returned when healing is attempted on a previously healed disks.
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Thu Aug 04 23:10:08 UTC 2022
    - 1.2K bytes
    - Viewed (0)
  5. cmd/global-heal.go

    		if len(disks) == healing {
    			// All drives in this erasure set were reformatted for some reasons, abort healing and mark it as successful
    			healingLogIf(ctx, errors.New("all drives are in healing state, aborting.."))
    			return nil
    		}
    
    		disks = disks[:len(disks)-healing] // healing drives are always at the end of the list
    
    		if len(disks) < er.setDriveCount/2 {
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Apr 04 13:49:12 UTC 2025
    - 16.2K bytes
    - Viewed (0)
  6. buildscripts/verify-healing-empty-erasure-set.sh

    	timeout 15m /tmp/mc ready myminio || fail
    
    	# Wait for all drives to be online and formatted
    	while [ $(/tmp/mc admin info --json myminio | jq '.info.servers[].drives[].state | select(. != "ok")' | wc -l) -gt 0 ]; do sleep 1; done
    	# Wait for all drives to be healed
    	while [ $(/tmp/mc admin info --json myminio | jq '.info.servers[].drives[].healing | select(. != null) | select(. == true)' | wc -l) -gt 0 ]; do sleep 1; done
    
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Jul 12 20:51:54 UTC 2024
    - 3.7K bytes
    - Viewed (0)
  7. cmd/fmt-gen.go

    				if deploymentID != "" {
    					newFormat.ID = deploymentID
    				}
    				drive := pool.Endpoints[i*setDriveCount+j]
    				fmtBytes, err := json.Marshal(newFormat)
    				if err != nil {
    					//nolint:gocritic
    					log.Fatalf("failed to marshal format.json for %s: %v", drive.String(), err)
    				}
    				fmtJSON := filepath.Join(drive.Host, drive.Path, minioMetaBucket, "format.json")
    				embedFileInZip(fmtZipW, fmtJSON, fmtBytes, 0o600)
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 3.7K bytes
    - Viewed (0)
  8. docs/metrics/prometheus/list.md

    | `minio_node_drive_latency_us`          | Average last minute latency in µs for drive API storage operations. |
    | `minio_node_drive_offline_total`       | Total drives offline in this node.                                  |
    | `minio_node_drive_online_total`        | Total drives online in this node.                                   |
    | `minio_node_drive_total`               | Total drives in this node.                                          |
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Tue Aug 12 18:20:36 UTC 2025
    - 43.4K bytes
    - Viewed (0)
  9. docs/throttle/README.md

    If you have traditional spinning (hdd) drives, some applications with high concurrency might require MinIO cluster to be tuned such that to avoid random I/O on the drives. The way to convert high concurrent I/O into a sequential I/O is by reducing the number of concurrent operations allowed per cluster. This allows MinIO cluster to be operationally resilient to such workloads, while also making sure the drives are at optimal efficiency and responsive.
    
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Aug 16 08:43:49 UTC 2024
    - 1.5K bytes
    - Viewed (1)
  10. docs/logging/README.md

    - Additionally in the case of the erasure coded setup `tags.objectLocation` provides per object details about
      - Pool number the object operation was performed on.
      - Set number the object operation was performed on.
      - The list of drives participating in this operation belong to the set.
    
    ```json
    {
      "version": "1",
      "deploymentid": "90e81272-45d9-4fe8-9c45-c9a7322bf4b5",
      "time": "2024-05-09T07:38:10.449688982Z",
      "event": "",
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Tue Aug 12 18:20:36 UTC 2025
    - 10.5K bytes
    - Viewed (0)
Back to top