Search Options

Display Count
Sort
Preferred Language
Advanced Search

Results 141 - 150 of 420 for diskID (0.06 seconds)

  1. cmd/metrics-v3-system-drive.go

    		return nil
    	}
    
    	for _, disk := range driveMetrics.storageInfo.Disks {
    		labels := []string{
    			driveL, disk.DrivePath,
    			poolIndexL, strconv.Itoa(disk.PoolIndex),
    			setIndexL, strconv.Itoa(disk.SetIndex),
    			driveIndexL, strconv.Itoa(disk.DiskIndex),
    		}
    
    		m.setDriveBasicMetrics(disk, labels)
    		if dm, found := driveMetrics.ioStats[disk.DrivePath]; found {
    			m.setDriveIOStatMetrics(dm, labels)
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Sun Mar 30 00:56:02 GMT 2025
    - 7.8K bytes
    - Click Count (0)
  2. cmd/object-api-utils.go

    	for _, disk := range di {
    		if disk == nil || disk.Total == 0 {
    			// Disk offline, no inodes or something else is wrong.
    			continue
    		}
    		nDisks++
    		total += disk.Total
    		available += disk.Total - disk.Used
    	}
    
    	if nDisks < len(di)/2 || nDisks <= 0 {
    		var errs []error
    		for index, disk := range di {
    			switch {
    			case disk == nil:
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Wed Jun 25 15:08:54 GMT 2025
    - 37.3K bytes
    - Click Count (0)
  3. cmd/erasure-server-pool-rebalance.go

    		}
    		totalCap += disk.TotalSpace
    		totalFree += disk.AvailableSpace
    
    		diskStats[disk.PoolIndex].AvailableSpace += disk.AvailableSpace
    		diskStats[disk.PoolIndex].TotalSpace += disk.TotalSpace
    	}
    	r.PercentFreeGoal = float64(totalFree) / float64(totalCap)
    
    	now := time.Now()
    	for idx := range z.serverPools {
    		r.PoolStats[idx] = &rebalanceStats{
    			Buckets:           make([]string, len(buckets)),
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Sun Sep 28 20:59:21 GMT 2025
    - 28.7K bytes
    - Click Count (0)
  4. docs/debugging/reorder-disks/main.go

    		format, err := getFormatJSON(disk.path)
    		if err != nil {
    			log.Printf("Unable to read format.json from `%s`, error: %v\n", disk.path, err)
    			continue
    		}
    		foundDiskLoc, err := getDiskLocation(format)
    		if err != nil {
    			log.Printf("Unable to get disk location of `%s`, error: %v\n", disk.path, err)
    			continue
    		}
    		actualDisksName[foundDiskLoc] = disk.path
    	}
    
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Fri May 24 23:05:23 GMT 2024
    - 5.4K bytes
    - Click Count (0)
  5. cmd/erasure-metadata.go

    }
    
    func writeAllMetadataWithRevert(ctx context.Context, disks []StorageAPI, origbucket, bucket, prefix string, files []FileInfo, quorum int, revert bool) ([]StorageAPI, error) {
    	g := errgroup.WithNErrs(len(disks))
    
    	// Start writing `xl.meta` to all disks in parallel.
    	for index := range disks {
    		g.Go(func() error {
    			if disks[index] == nil {
    				return errDiskNotFound
    			}
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Sun Sep 28 20:59:21 GMT 2025
    - 21.3K bytes
    - Click Count (0)
  6. cmd/peer-s3-server.go

    	globalLocalDrivesMu.RUnlock()
    
    	// Initialize sync waitgroup.
    	g := errgroup.WithNErrs(len(localDrives))
    
    	// Disk states slices
    	beforeState := make([]string, len(localDrives))
    	afterState := make([]string, len(localDrives))
    
    	// Make a volume entry on all underlying storage disks.
    	for index := range localDrives {
    		g.Go(func() (serr error) {
    			if localDrives[index] == nil {
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Fri Aug 29 02:39:48 GMT 2025
    - 8.2K bytes
    - Click Count (0)
  7. docs/resiliency/resiliency-tests.sh

    	echo
    	echo -e "${GREEN}Running test_resiliency_failure_with_too_many_disks_offline ...${NC}"
    	# There are 8 disks on each node with EC:4 and two erasure sets.
    	# We should be able to safely suspend one disk per set from each server.
    	# suspending one additional disk from each set should cause failures
    	docker exec resiliency-minio1-1 /bin/sh -c "mv /data1/.minio.sys /data1/.minio.bkp && touch /data1/.minio.sys"
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Sat Dec 21 04:24:45 GMT 2024
    - 20.5K bytes
    - Click Count (0)
  8. cmd/mrf.go

    	healMRFMetaFormat    = 1
    	healMRFMetaVersionV1 = 1
    )
    
    // PartialOperation is a successful upload/delete of an object
    // but not written in all disks (having quorum)
    type PartialOperation struct {
    	Bucket              string
    	Object              string
    	VersionID           string
    	Versions            []byte
    	SetIndex, PoolIndex int
    	Queued              time.Time
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Tue May 27 15:19:03 GMT 2025
    - 6.5K bytes
    - Click Count (0)
  9. cmd/test-utils_test.go

    func initTestServerWithBackend(ctx context.Context, t TestErrHandler, testServer TestServer, objLayer ObjectLayer, disks []string) TestServer {
    	// Test Server needs to start before formatting of disks.
    	// Get credential.
    	credentials := globalActiveCred
    	if !globalReplicationPool.IsSet() {
    		globalReplicationPool.Set(nil)
    	}
    	testServer.Obj = objLayer
    	testServer.rawDiskPaths = disks
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Fri Aug 29 02:39:48 GMT 2025
    - 77K bytes
    - Click Count (0)
  10. cmd/metrics-v3-cache.go

    		objLayer := newObjectLayerFn()
    		if objLayer == nil {
    			return v, err
    		}
    
    		storageInfo := objLayer.LocalStorageInfo(GlobalContext, true)
    		onlineDrives, offlineDrives := getOnlineOfflineDisksStats(storageInfo.Disks)
    		totalDrives := onlineDrives.Merge(offlineDrives)
    
    		v = storageMetrics{
    			storageInfo:   storageInfo,
    			onlineDrives:  onlineDrives.Sum(),
    			offlineDrives: offlineDrives.Sum(),
    Created: Sun Apr 05 19:28:12 GMT 2026
    - Last Modified: Sun Sep 28 20:59:21 GMT 2025
    - 8.1K bytes
    - Click Count (0)
Back to Top