- Sort Score
- Num 10 results
- Language All
Results 141 - 150 of 420 for diskID (0.25 seconds)
-
cmd/metrics-v3-system-drive.go
return nil } for _, disk := range driveMetrics.storageInfo.Disks { labels := []string{ driveL, disk.DrivePath, poolIndexL, strconv.Itoa(disk.PoolIndex), setIndexL, strconv.Itoa(disk.SetIndex), driveIndexL, strconv.Itoa(disk.DiskIndex), } m.setDriveBasicMetrics(disk, labels) if dm, found := driveMetrics.ioStats[disk.DrivePath]; found { m.setDriveIOStatMetrics(dm, labels)Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Mar 30 00:56:02 GMT 2025 - 7.8K bytes - Click Count (0) -
cmd/object-api-utils.go
for _, disk := range di { if disk == nil || disk.Total == 0 { // Disk offline, no inodes or something else is wrong. continue } nDisks++ total += disk.Total available += disk.Total - disk.Used } if nDisks < len(di)/2 || nDisks <= 0 { var errs []error for index, disk := range di { switch { case disk == nil:Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Wed Jun 25 15:08:54 GMT 2025 - 37.3K bytes - Click Count (0) -
cmd/erasure-server-pool-rebalance.go
} totalCap += disk.TotalSpace totalFree += disk.AvailableSpace diskStats[disk.PoolIndex].AvailableSpace += disk.AvailableSpace diskStats[disk.PoolIndex].TotalSpace += disk.TotalSpace } r.PercentFreeGoal = float64(totalFree) / float64(totalCap) now := time.Now() for idx := range z.serverPools { r.PoolStats[idx] = &rebalanceStats{ Buckets: make([]string, len(buckets)),
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 28.7K bytes - Click Count (0) -
docs/debugging/reorder-disks/main.go
format, err := getFormatJSON(disk.path) if err != nil { log.Printf("Unable to read format.json from `%s`, error: %v\n", disk.path, err) continue } foundDiskLoc, err := getDiskLocation(format) if err != nil { log.Printf("Unable to get disk location of `%s`, error: %v\n", disk.path, err) continue } actualDisksName[foundDiskLoc] = disk.path }
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri May 24 23:05:23 GMT 2024 - 5.4K bytes - Click Count (0) -
cmd/erasure-metadata.go
} func writeAllMetadataWithRevert(ctx context.Context, disks []StorageAPI, origbucket, bucket, prefix string, files []FileInfo, quorum int, revert bool) ([]StorageAPI, error) { g := errgroup.WithNErrs(len(disks)) // Start writing `xl.meta` to all disks in parallel. for index := range disks { g.Go(func() error { if disks[index] == nil { return errDiskNotFound }
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 21.3K bytes - Click Count (0) -
cmd/peer-s3-server.go
globalLocalDrivesMu.RUnlock() // Initialize sync waitgroup. g := errgroup.WithNErrs(len(localDrives)) // Disk states slices beforeState := make([]string, len(localDrives)) afterState := make([]string, len(localDrives)) // Make a volume entry on all underlying storage disks. for index := range localDrives { g.Go(func() (serr error) { if localDrives[index] == nil {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 8.2K bytes - Click Count (0) -
docs/resiliency/resiliency-tests.sh
echo echo -e "${GREEN}Running test_resiliency_failure_with_too_many_disks_offline ...${NC}" # There are 8 disks on each node with EC:4 and two erasure sets. # We should be able to safely suspend one disk per set from each server. # suspending one additional disk from each set should cause failures docker exec resiliency-minio1-1 /bin/sh -c "mv /data1/.minio.sys /data1/.minio.bkp && touch /data1/.minio.sys"Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sat Dec 21 04:24:45 GMT 2024 - 20.5K bytes - Click Count (0) -
cmd/mrf.go
healMRFMetaFormat = 1 healMRFMetaVersionV1 = 1 ) // PartialOperation is a successful upload/delete of an object // but not written in all disks (having quorum) type PartialOperation struct { Bucket string Object string VersionID string Versions []byte SetIndex, PoolIndex int Queued time.TimeCreated: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue May 27 15:19:03 GMT 2025 - 6.5K bytes - Click Count (0) -
cmd/test-utils_test.go
func initTestServerWithBackend(ctx context.Context, t TestErrHandler, testServer TestServer, objLayer ObjectLayer, disks []string) TestServer { // Test Server needs to start before formatting of disks. // Get credential. credentials := globalActiveCred if !globalReplicationPool.IsSet() { globalReplicationPool.Set(nil) } testServer.Obj = objLayer testServer.rawDiskPaths = disks
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 77K bytes - Click Count (0) -
cmd/metrics-v3-cache.go
objLayer := newObjectLayerFn() if objLayer == nil { return v, err } storageInfo := objLayer.LocalStorageInfo(GlobalContext, true) onlineDrives, offlineDrives := getOnlineOfflineDisksStats(storageInfo.Disks) totalDrives := onlineDrives.Merge(offlineDrives) v = storageMetrics{ storageInfo: storageInfo, onlineDrives: onlineDrives.Sum(), offlineDrives: offlineDrives.Sum(),
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 8.1K bytes - Click Count (0)