- Sort Score
- Result 10 results
- Languages All
Results 131 - 140 of 924 for Franke (0.04 sec)
-
cmd/global-heal.go
indexed := make(map[string][]madmin.Disk) for _, disk := range si.Disks { setIdx := fmt.Sprintf("%d-%d", disk.PoolIndex, disk.SetIndex) indexed[setIdx] = append(indexed[setIdx], disk) } for id, disks := range indexed { ss := madmin.SetStatus{ ID: id, SetIndex: disks[0].SetIndex, PoolIndex: disks[0].PoolIndex, } for _, disk := range disks { ss.Disks = append(ss.Disks, disk)
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Apr 04 13:49:12 UTC 2025 - 16.2K bytes - Viewed (0) -
cmd/metrics.go
float64(dur), ) for k, v := range bgSeq.getScannedItemsMap() { ch <- prometheus.MustNewConstMetric( prometheus.NewDesc( prometheus.BuildFQName(healMetricsNamespace, "objects", "scanned"), "Objects scanned since uptime", []string{"type"}, nil), prometheus.CounterValue, float64(v), string(k), ) } for k, v := range bgSeq.getHealedItemsMap() {
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Oct 24 04:06:31 UTC 2025 - 16.7K bytes - Viewed (0) -
internal/store/queuestore_test.go
if err != nil { t.Fatal("Failed to create a queue store ", err) } // Put 10 items for range 10 { if _, err := store.Put(testItem); err != nil { t.Fatal("Failed to put to queue store ", err) } } itemKeys := store.List() // Get 10 items. if len(itemKeys) == 10 { for _, key := range itemKeys { item, eErr := store.Get(key) if eErr != nil {Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 10.2K bytes - Viewed (0) -
cmd/metacache-entries.go
for i, entry := range m { if entry.name != "" { n++ if first == nil { first = &m[i] } } } return first, n } // names will return all names in order. // Since this allocates it should not be used in critical functions. func (m metaCacheEntries) names() []string { res := make([]string, 0, len(m)) for _, obj := range m { res = append(res, obj.name) }Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Thu Aug 08 15:29:58 UTC 2024 - 24.1K bytes - Viewed (0) -
cmd/local-locker.go
} if len(args.UID) == 0 { for _, resource := range args.Resources { lris, ok := l.lockMap[resource] if !ok { continue } // Collect uids, so we don't mutate while we delete uids := make([]string, 0, len(lris)) for _, lri := range lris { uids = append(uids, lri.UID) } // Delete collected uids: for _, uid := range uids { lris, ok := l.lockMap[resource] if !ok {
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Sun Sep 28 20:59:21 UTC 2025 - 12K bytes - Viewed (0) -
internal/logger/logger.go
// as "{path#}/src/" into trimStrings for _, goPathString := range goPathList { trimStrings = append(trimStrings, filepath.Join(goPathString, "src")+string(filepath.Separator)) } for _, goRootString := range goRootList { trimStrings = append(trimStrings, filepath.Join(goRootString, "src")+string(filepath.Separator)) } for _, defaultgoPathString := range defaultgoPathList {
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 12.4K bytes - Viewed (0) -
docs/debugging/reorder-disks/main.go
if err != nil { return nil, err } exp := argP.Expand() if node == "" { for index, e := range exp { result = append(result, localDisk{index: index, path: strings.Join(e, "")}) } } else { for index, e := range exp { u, err := url.Parse(strings.Join(e, "")) if err != nil { return nil, err } if strings.Contains(u.Host, node) {Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 5.4K bytes - Viewed (0) -
clause/set.go
func Assignments(values map[string]interface{}) Set { keys := make([]string, 0, len(values)) for key := range values { keys = append(keys, key) } sort.Strings(keys) assignments := make([]Assignment, len(keys)) for idx, key := range keys { assignments[idx] = Assignment{Column: Column{Name: key}, Value: values[key]} } return assignments }
Registered: Sun Dec 28 09:35:17 UTC 2025 - Last Modified: Tue Sep 09 01:34:33 UTC 2025 - 1.7K bytes - Viewed (0) -
cmd/erasure-multipart.go
var wg sync.WaitGroup for _, disk := range er.getLocalDisks() { if disk != nil { wg.Add(1) go func(disk StorageAPI) { defer wg.Done() er.cleanupStaleUploadsOnDisk(ctx, disk) }(disk) } } wg.Wait() } func (er erasureObjects) deleteAll(ctx context.Context, bucket, prefix string) { var wg sync.WaitGroup for _, disk := range er.getDisks() { if disk == nil { continue
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Oct 24 04:05:31 UTC 2025 - 47.1K bytes - Viewed (0) -
cmd/erasure-server-pool-decom.go
decomStarted: decomStarted, } } specifiedPools := make(map[string]int) for idx, pool := range pools { specifiedPools[pool.endpoints.CmdLine] = idx } var update bool // Check if specified pools need to be removed from decommissioned pool. for k := range specifiedPools { pi, ok := rememberedPools[k] if !ok {
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Sun Sep 28 20:59:21 UTC 2025 - 42.2K bytes - Viewed (1)