- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 269 for defer (0.01 sec)
-
cmd/metacache-set.go
rpc := mc.rpc cancel := mc.cancel defer func() { o.debugln(color.Green("saveMetaCacheStream:")+"err:", err) if err != nil && !errors.Is(err, io.EOF) { go mc.setErr(err.Error()) cancel() } }() defer cancel() // Save continuous updates go func() { var err error ticker := time.NewTicker(10 * time.Second) defer ticker.Stop() var exit bool for !exit {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 30.7K bytes - Viewed (0) -
cmd/sftp-server-driver.go
} cctx, cancel := context.WithCancel(context.Background()) defer cancel() if prefix == "" { // if all objects are not deleted yet this call may fail. return clnt.RemoveBucket(cctx, bucket) } objectsCh := make(chan minio.ObjectInfo) // Send object names that are needed to be removed to objectsCh go func() { defer xioutil.SafeClose(objectsCh) opts := minio.ListObjectsOptions{
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Mon Feb 10 16:35:49 UTC 2025 - 11.6K bytes - Viewed (0) -
cmd/update.go
if err != nil { return nil, nil, err } } else { return nil, nil, fmt.Errorf("unsupported protocol scheme: %s", u.Scheme) } defer xhttp.DrainBody(reader) b := bytebufferpool.Get() bc := bytebufferpool.Get() defer func() { b.Reset() bc.Reset() bytebufferpool.Put(b) bytebufferpool.Put(bc) }() w, err := zstd.NewWriter(bc) if err != nil {Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue Aug 12 18:20:36 UTC 2025 - 18.9K bytes - Viewed (0) -
cmd/acl-handlers.go
// only responds success if the ACL is private. func (api objectAPIHandlers) PutBucketACLHandler(w http.ResponseWriter, r *http.Request) { ctx := newContext(r, w, "PutBucketACL") defer logger.AuditLog(ctx, w, r, mustGetClaimsFromToken(r)) vars := mux.Vars(r) bucket := vars["bucket"] objAPI := api.ObjectAPI() if objAPI == nil {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 8.4K bytes - Viewed (0) -
cmd/erasure-healing.go
scanMode := opts.ScanMode storageDisks := er.getDisks() storageEndpoints := er.getEndpoints() defer func() { er.auditHealObject(ctx, bucket, object, versionID, result, err) }() if globalTrace.NumSubscribers(madmin.TraceHealing) > 0 { startTime := time.Now() defer func() { healTrace(healingMetricObject, startTime, bucket, object, &opts, err, &result) }() }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 34.6K bytes - Viewed (0) -
cmd/event-notification.go
func (evnot *EventNotifier) RemoveNotification(bucketName string) { evnot.Lock() defer evnot.Unlock() delete(evnot.bucketRulesMap, bucketName) } // RemoveAllBucketTargets - closes and removes all notification targets. func (evnot *EventNotifier) RemoveAllBucketTargets() { evnot.Lock() defer evnot.Unlock() targetIDSet := event.NewTargetIDSet() for k := range evnot.targetList.TargetMap() {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Jun 21 22:22:24 UTC 2024 - 7.7K bytes - Viewed (0) -
cmd/server-main_test.go
// Tests initializing new object layer. func TestNewObjectLayer(t *testing.T) { ctx, cancel := context.WithCancel(t.Context()) defer cancel() // Tests for ErasureSD object layer. nDisks := 1 disks, err := getRandomDisks(nDisks) if err != nil { t.Fatal("Failed to create drives for the backend") } defer removeRoots(disks) obj, err := newObjectLayer(ctx, mustGetPoolEndpoints(0, disks...)) if err != nil {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 3.1K bytes - Viewed (0) -
cmd/data-usage-cache.go
} buf := bytebufferpool.Get() defer func() { <-maxConcurrentScannerSaves buf.Reset() bytebufferpool.Put(buf) }() if err := d.serializeTo(buf); err != nil { return err } save := func(name string, timeout time.Duration) error { // Abandon if more than a minute, so we don't hold up scanner. ctx, cancel := context.WithTimeout(ctx, timeout) defer cancel()
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 34.7K bytes - Viewed (0) -
cmd/metacache-server-pool.go
listCtx, cancelList := context.WithCancel(ctx) defer cancelList() for _, pool := range z.serverPools { for _, set := range pool.sets { wg.Add(1) innerResults := make(chan metaCacheEntry, 100) inputs = append(inputs, innerResults) go func(i int, set *erasureObjects) { defer wg.Done() err := set.listPath(listCtx, o, innerResults) mu.Lock() defer mu.Unlock() if err == nil {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Mon May 26 07:06:43 UTC 2025 - 12.9K bytes - Viewed (0) -
cmd/ilm-config.go
} func (c *ilmConfig) getExpirationWorkers() int { c.mu.RLock() defer c.mu.RUnlock() return c.cfg.ExpirationWorkers } func (c *ilmConfig) getTransitionWorkers() int { c.mu.RLock() defer c.mu.RUnlock() return c.cfg.TransitionWorkers } func (c *ilmConfig) update(cfg ilm.Config) { c.mu.Lock() defer c.mu.Unlock() c.cfg = cfg
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue Mar 05 02:50:24 UTC 2024 - 1.3K bytes - Viewed (0)