- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 269 for defer (0.01 sec)
-
cmd/server-startup-msg_test.go
} } // Test printing server common message. func TestPrintServerCommonMessage(t *testing.T) { ctx, cancel := context.WithCancel(t.Context()) defer cancel() obj, fsDir, err := prepareFS(ctx) if err != nil { t.Fatal(err) } defer os.RemoveAll(fsDir) if err = newTestConfig(globalMinioDefaultRegion, obj); err != nil { t.Fatal(err) } apiEndpoints := []string{"http://127.0.0.1:9000"}
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Wed Apr 09 14:28:39 UTC 2025 - 3K bytes - Viewed (0) -
cmd/generic-handlers.go
} defer logger.AuditLog(r.Context(), w, r, mustGetClaimsFromToken(r)) writeErrorResponse(r.Context(), w, errorCodes.ToAPIErr(ErrUnsupportedMetadata), r.URL) return } if isHTTPHeaderSizeTooLarge(r.Header) { if ok { tc.FuncName = "handler.ValidRequest" tc.ResponseRecorder.LogErrBody = true }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 20.7K bytes - Viewed (1) -
cmd/prepare-storage.go
return func(endpoint Endpoint, err error, once bool) { reqInfo := (&logger.ReqInfo{}).AppendTags("endpoint", endpoint.String()) ctx := logger.SetReqInfo(GlobalContext, reqInfo) mutex.Lock() defer mutex.Unlock() m, ok := printOnce[endpoint] if !ok { m = make(map[string]int) printOnce[endpoint] = m if once { m[err.Error()]++ peersLogAlwaysIf(ctx, err) return } }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Jul 12 20:51:54 UTC 2024 - 11.1K bytes - Viewed (1) -
cmd/erasure-server-pool-rebalance.go
z.rebalMu.Lock() defer z.rebalMu.Unlock() r := z.rebalMeta if r == nil { return } r.PoolStats[poolIdx].update(bucket, fi) } const ( rebalMetaName = "rebalance.bin" rebalMetaFmt = 1 rebalMetaVer = 1 ) func (z *erasureServerPools) nextRebalBucket(poolIdx int) (string, bool) { z.rebalMu.RLock() defer z.rebalMu.RUnlock() r := z.rebalMeta
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Thu Sep 04 20:47:24 UTC 2025 - 28.9K bytes - Viewed (0) -
cmd/storage-rest-server.go
write([]byte{1}) write([]byte(err.Error())) } else { write([]byte{0}) } xioutil.SafeClose(doneCh) return } defer xioutil.SafeClose(doneCh) // Initiate ticker after body has been read. ticker := time.NewTicker(time.Second * 10) defer ticker.Stop() for { select { case <-ticker.C: // The done() might have been called // concurrently, check for it before we
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue May 27 15:19:03 UTC 2025 - 45.7K bytes - Viewed (0) -
src/arena/arena_test.go
a := arena.NewArena() defer a.Free() tt := arena.New[T1](a) tt.n = 1 ts := arena.MakeSlice[T1](a, 99, 100) if len(ts) != 99 { t.Errorf("Slice() len = %d, want 99", len(ts)) } if cap(ts) != 100 { t.Errorf("Slice() cap = %d, want 100", cap(ts)) } ts[1].n = 42 } func TestSmokeLarge(t *testing.T) { a := arena.NewArena() defer a.Free() for i := 0; i < 10*64; i++ {
Registered: Tue Sep 09 11:13:09 UTC 2025 - Last Modified: Wed Oct 12 20:23:36 UTC 2022 - 742 bytes - Viewed (0) -
cmd/bucket-replication.go
s.Lock() defer s.Unlock() m := s.statusMap[opts.bucket] st := m.TargetsMap[opts.arn] st.LastUpdate = UTCNow() st.ResyncStatus = status m.TargetsMap[opts.arn] = st m.LastUpdate = UTCNow() s.statusMap[opts.bucket] = m ctx, cancel := context.WithTimeout(context.Background(), time.Second) defer cancel() saveResyncStatus(ctx, opts.bucket, m, objAPI) }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 118K bytes - Viewed (0) -
cmd/iam-store.go
cache := store.rlock() defer store.runlock() u, ok := cache.iamUsersMap[user] if !ok { // Check the sts map u, ok = cache.iamSTSAccountsMap[user] } return u, ok } // GetMappedPolicy - fetches mapped policy from memory. func (store *IAMStoreSys) GetMappedPolicy(name string, isGroup bool) (MappedPolicy, bool) { cache := store.rlock() defer store.runlock() if isGroup {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 86.7K bytes - Viewed (0) -
cmd/erasure.go
go func() { // Add jitter to the update time so multiple sets don't sync up. updateTime := 30*time.Second + time.Duration(float64(10*time.Second)*rand.Float64()) t := time.NewTicker(updateTime) defer t.Stop() defer saverWg.Done() var lastSave time.Time for { select { case <-t.C: if cache.Info.LastUpdate.Equal(lastSave) { continue }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 16.1K bytes - Viewed (0) -
cmd/callhome.go
// cycle back return 'true' return true } ctx = lkctx.Context() defer locker.Unlock(lkctx) // Perform callhome once and then keep running it at regular intervals. performCallhome(ctx) callhomeTimer := time.NewTimer(globalCallhomeConfig.FrequencyDur()) defer callhomeTimer.Stop() for { if !globalCallhomeConfig.Enabled() { // Stop the processing as callhome got disabled
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 5.3K bytes - Viewed (0)