- Sort Score
- Num 10 results
- Language All
Results 141 - 150 of 408 for diskID (0.04 seconds)
-
docs/debugging/reorder-disks/main.go
format, err := getFormatJSON(disk.path) if err != nil { log.Printf("Unable to read format.json from `%s`, error: %v\n", disk.path, err) continue } foundDiskLoc, err := getDiskLocation(format) if err != nil { log.Printf("Unable to get disk location of `%s`, error: %v\n", disk.path, err) continue } actualDisksName[foundDiskLoc] = disk.path }
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri May 24 23:05:23 GMT 2024 - 5.4K bytes - Click Count (0) -
cmd/storage-interface.go
IsOnline() bool // Returns the last time this disk (re)-connected LastConn() time.Time // Indicates if disk is local or not. IsLocal() bool // Returns hostname if disk is remote. Hostname() string // Returns the entire endpoint. Endpoint() Endpoint // Close the disk, mark it purposefully closed, only implemented for remote disks. Close() error // Returns the unique 'uuid' of this disk.
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Apr 25 05:41:04 GMT 2025 - 5.3K bytes - Click Count (0) -
cmd/erasure-server-pool-rebalance.go
} totalCap += disk.TotalSpace totalFree += disk.AvailableSpace diskStats[disk.PoolIndex].AvailableSpace += disk.AvailableSpace diskStats[disk.PoolIndex].TotalSpace += disk.TotalSpace } r.PercentFreeGoal = float64(totalFree) / float64(totalCap) now := time.Now() for idx := range z.serverPools { r.PoolStats[idx] = &rebalanceStats{ Buckets: make([]string, len(buckets)),
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 28.7K bytes - Click Count (0) -
cmd/erasure-metadata.go
} func writeAllMetadataWithRevert(ctx context.Context, disks []StorageAPI, origbucket, bucket, prefix string, files []FileInfo, quorum int, revert bool) ([]StorageAPI, error) { g := errgroup.WithNErrs(len(disks)) // Start writing `xl.meta` to all disks in parallel. for index := range disks { g.Go(func() error { if disks[index] == nil { return errDiskNotFound }
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 21.3K bytes - Click Count (0) -
cmd/metrics-v3-system-drive.go
return nil } for _, disk := range driveMetrics.storageInfo.Disks { labels := []string{ driveL, disk.DrivePath, poolIndexL, strconv.Itoa(disk.PoolIndex), setIndexL, strconv.Itoa(disk.SetIndex), driveIndexL, strconv.Itoa(disk.DiskIndex), } m.setDriveBasicMetrics(disk, labels) if dm, found := driveMetrics.ioStats[disk.DrivePath]; found { m.setDriveIOStatMetrics(dm, labels)Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Mar 30 00:56:02 GMT 2025 - 7.8K bytes - Click Count (0) -
cmd/peer-s3-server.go
globalLocalDrivesMu.RUnlock() // Initialize sync waitgroup. g := errgroup.WithNErrs(len(localDrives)) // Disk states slices beforeState := make([]string, len(localDrives)) afterState := make([]string, len(localDrives)) // Make a volume entry on all underlying storage disks. for index := range localDrives { g.Go(func() (serr error) { if localDrives[index] == nil {
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 8.2K bytes - Click Count (0) -
docs/resiliency/resiliency-tests.sh
echo echo -e "${GREEN}Running test_resiliency_failure_with_too_many_disks_offline ...${NC}" # There are 8 disks on each node with EC:4 and two erasure sets. # We should be able to safely suspend one disk per set from each server. # suspending one additional disk from each set should cause failures docker exec resiliency-minio1-1 /bin/sh -c "mv /data1/.minio.sys /data1/.minio.bkp && touch /data1/.minio.sys"Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sat Dec 21 04:24:45 GMT 2024 - 20.5K bytes - Click Count (0) -
cmd/mrf.go
healMRFMetaFormat = 1 healMRFMetaVersionV1 = 1 ) // PartialOperation is a successful upload/delete of an object // but not written in all disks (having quorum) type PartialOperation struct { Bucket string Object string VersionID string Versions []byte SetIndex, PoolIndex int Queued time.TimeCreated: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue May 27 15:19:03 GMT 2025 - 6.5K bytes - Click Count (0) -
cmd/test-utils_test.go
func initTestServerWithBackend(ctx context.Context, t TestErrHandler, testServer TestServer, objLayer ObjectLayer, disks []string) TestServer { // Test Server needs to start before formatting of disks. // Get credential. credentials := globalActiveCred if !globalReplicationPool.IsSet() { globalReplicationPool.Set(nil) } testServer.Obj = objLayer testServer.rawDiskPaths = disks
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 77K bytes - Click Count (0) -
cmd/data-scanner-metric.go
} // currentPathUpdater provides a lightweight update function for keeping track of // current objects for each disk. // Returns a function that can be used to update the current object // and a function to call to when processing finished. func (p *scannerMetrics) currentPathUpdater(disk, initial string) (update func(path string), done func()) { initialPtr := unsafe.Pointer(&initial) tracker := ¤tPathTracker{
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Wed Oct 01 06:06:01 GMT 2025 - 9.4K bytes - Click Count (0)