- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 469 for Done (0.14 sec)
-
internal/s3select/sql/stringfuncs.go
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Tue Jun 01 21:59:40 GMT 2021 - 4.2K bytes - Viewed (0) -
cmd/erasure-server-pool-rebalance.go
if quit { return } timer.Reset(randSleepFor()) } }() rebalanceLogEvent(ctx, "Pool %d rebalancing is started", poolIdx+1) for { select { case <-ctx.Done(): doneCh <- ctx.Err() return default: } bucket, ok := z.nextRebalBucket(poolIdx) if !ok { // no more buckets to rebalance or target free_space/capacity reached break }
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri Apr 26 19:29:28 GMT 2024 - 27.2K bytes - Viewed (0) -
cmd/storage-errors.go
var errMoreData = StorageErr("more data was sent than what was advertised") // indicates readDirFn to return without further applying the fn() var errDoneForNow = errors.New("done for now") // errSkipFile returned by the fn() for readDirFn() when it needs // to proceed to next entry. var errSkipFile = errors.New("skip this file")
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 6.4K bytes - Viewed (0) -
cmd/data-scanner_test.go
workers := []chan expiryOp{make(chan expiryOp)} es.workers.Store(&workers) globalExpiryState = es var wg sync.WaitGroup wg.Add(1) expired := make([]ObjectToDelete, 0, 5) go func() { defer wg.Done() workers := globalExpiryState.workers.Load() for t := range (*workers)[0] { if t, ok := t.(newerNoncurrentTask); ok { expired = append(expired, t.versions...) } } }() lc := lifecycle.Lifecycle{
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 03 11:18:58 GMT 2024 - 6.9K bytes - Viewed (0) -
cmd/untar.go
return err } wg.Add(1) go func(name string, fi fs.FileInfo, b []byte) { rc := disconnectReader{r: bytes.NewReader(b)} defer func() { rc.Close() <-asyncWriters wg.Done() //nolint:staticcheck // SA6002 we are fine with the tiny alloc poolBuf128k.Put(b) }() if err := putObject(&rc, fi, name); err != nil { if o.ignoreErrs { s3LogIf(ctx, err)
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 6K bytes - Viewed (0) -
cmd/bucket-targets.go
latency: l, } } cancel() sys.hMutex.Lock() sys.hc = m sys.hMutex.Unlock() } hcTimer.Reset(defaultHealthCheckDuration) case <-ctx.Done(): return } } } // periodically rebuild the healthCheck map from list of targets to clear // out stale endpoints func (sys *BucketTargetSys) reloadHealthCheckers(ctx context.Context) {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Wed May 01 01:09:56 GMT 2024 - 20.9K bytes - Viewed (0) -
docs/metrics/prometheus/README.md
metrics_path: /minio/v2/metrics/bucket scheme: http static_configs: - targets: ['localhost:9000'] ``` ##### Node (optional) Optionally you can also collect per node metrics. This needs to be done on a per server instance. The scrape configurations should use all the servers under `targets` so that graphing systems like grafana can visualize them for all the nodes ```yaml scrape_configs: - job_name: minio-job
Plain Text - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri Apr 12 15:49:30 GMT 2024 - 7.1K bytes - Viewed (0) -
internal/grid/stream.go
select { case s.Requests <- b: return nil case <-s.ctx.Done(): return context.Cause(s.ctx) } } // Results returns the results from the remote server one by one. // If any error is returned by the callback, the stream will be canceled. // If the context is canceled, the stream will be canceled. func (s *Stream) Results(next func(b []byte) error) (err error) { done := false defer func() { if s.cancel != nil {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Wed Feb 28 18:05:18 GMT 2024 - 2.8K bytes - Viewed (0) -
internal/event/target/webhook.go
id: event.TargetID{ID: id, Name: "webhook"}, args: args, loggerOnce: loggerOnce, transport: transport, store: queueStore, cancel: cancel, cancelCh: ctx.Done(), } // Calculate the webhook addr with the port number format target.addr = args.Endpoint.Host
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Mon Nov 20 22:40:07 GMT 2023 - 8.5K bytes - Viewed (0) -
buildscripts/verify-healing-with-root-disks.sh
done for i in $(seq 1 4); do "${MINIO[@]}" --address ":$((start_port + i))" ${args[@]} 2>&1 >"${WORK_DIR}/server$i.log" & done # Wait until all nodes return 403 for i in $(seq 1 4); do while [ "$(curl -m 1 -s -o /dev/null -w "%{http_code}" http://localhost:$((start_port + i)))" -ne "403" ]; do echo -n "." sleep 1 done done } # Prepare fake disks with losetup
Shell Script - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 26 05:07:25 GMT 2023 - 2.2K bytes - Viewed (0)