- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 293 for Workers (0.18 sec)
-
cmd/data-scanner_test.go
es := newExpiryState(context.Background(), objAPI, 0) workers := []chan expiryOp{make(chan expiryOp)} es.workers.Store(&workers) globalExpiryState = es var wg sync.WaitGroup wg.Add(1) expired := make([]ObjectToDelete, 0, 5) go func() { defer wg.Done() workers := globalExpiryState.workers.Load() for t := range (*workers)[0] { if t, ok := t.(newerNoncurrentTask); ok {
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Sat Mar 02 05:11:03 GMT 2024 - 3.9K bytes - Viewed (0) -
internal/config/errors.go
"Invalid value for replication workers", "", "MINIO_API_REPLICATION_WORKERS: should be > 0", ) ErrInvalidTransitionWorkersValue = newErrFn( "Invalid value for transition workers", "", "MINIO_API_TRANSITION_WORKERS: should be >= GOMAXPROCS/2", ) ErrInvalidBatchKeyRotationWorkersWait = newErrFn( "Invalid value for batch key rotation workers wait", "Please input a non-negative duration",
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Mon Mar 18 22:25:32 GMT 2024 - 9.1K bytes - Viewed (0) -
internal/config/api/api.go
apiRemoteTransportDeadline = "remote_transport_deadline" apiListQuorum = "list_quorum" apiReplicationPriority = "replication_priority" apiReplicationMaxWorkers = "replication_max_workers" apiTransitionWorkers = "transition_workers" apiStaleUploadsCleanupInterval = "stale_uploads_cleanup_interval" apiStaleUploadsExpiry = "stale_uploads_expiry" apiDeleteCleanupInterval = "delete_cleanup_interval"
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Thu Mar 28 01:08:07 GMT 2024 - 11.1K bytes - Viewed (1) -
cmd/background-heal-ops.go
} func newHealRoutine() *healRoutine { workers := runtime.GOMAXPROCS(0) / 2 if envHealWorkers := env.Get("_MINIO_HEAL_WORKERS", ""); envHealWorkers != "" { if numHealers, err := strconv.Atoi(envHealWorkers); err != nil { bugLogIf(context.Background(), fmt.Errorf("invalid _MINIO_HEAL_WORKERS value: %w", err)) } else { workers = numHealers } } if workers == 0 { workers = 4 }
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 4.6K bytes - Viewed (0) -
internal/event/targetlist.go
return } } }() } wk.Wait() } var startOnce sync.Once // Init initialize target send workers. func (list *TargetList) Init(workers int) *TargetList { startOnce.Do(func() { go list.startSendWorkers(workers) }) return list } // NewTargetList - creates TargetList. func NewTargetList(ctx context.Context) *TargetList { list := &TargetList{
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 9.2K bytes - Viewed (0) -
internal/config/heal/heal.go
const ( Bitrot = "bitrotscan" Sleep = "max_sleep" IOCount = "max_io" DriveWorkers = "drive_workers" EnvBitrot = "MINIO_HEAL_BITROTSCAN" EnvSleep = "MINIO_HEAL_MAX_SLEEP" EnvIOCount = "MINIO_HEAL_MAX_IO" EnvDriveWorkers = "MINIO_HEAL_DRIVE_WORKERS" ) var configMutex sync.RWMutex // Config represents the heal settings. type Config struct {
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Thu Apr 11 01:10:30 GMT 2024 - 4.9K bytes - Viewed (0) -
cmd/bucket-replication-stats.go
sync.RWMutex // mutex for Cache mostRecentStatsMu sync.Mutex // mutex for mostRecentStats wlock sync.RWMutex // mutex for active workers movingAvgTicker *time.Ticker // Ticker for calculating moving averages wTimer *time.Ticker // ticker for calculating active workers qTimer *time.Ticker // ticker for calculating queue stats } func (r *ReplicationStats) trackEWMA() { for { select {
Go - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Thu Feb 22 06:26:06 GMT 2024 - 13.4K bytes - Viewed (0) -
docs/en/docs/deployment/concepts.md
Here are some possible combinations and strategies: * **Gunicorn** managing **Uvicorn workers** * Gunicorn would be the **process manager** listening on the **IP** and **port**, the replication would be by having **multiple Uvicorn worker processes**. * **Uvicorn** managing **Uvicorn workers**
Plain Text - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Thu Apr 18 19:53:19 GMT 2024 - 18K bytes - Viewed (0) -
internal/logger/target/http/http.go
// If the channel reaches above half capacity // we spawn more workers. The workers spawned // from this main worker routine will exit // once the channel drops below half capacity // and when it's been at least 30 seconds since // we launched a new worker. if mainWorker && len(h.logCh) > cap(h.logCh)/2 { nWorkers := atomic.LoadInt64(&h.workers) if nWorkers < h.maxWorkers {
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Mon Mar 25 16:44:20 GMT 2024 - 14.9K bytes - Viewed (0) -
docs/config/README.md
Each node is responsible of healing its local drives; Each drive will have multiple heal workers which is the quarter of the number of CPU cores of the node or the quarter of the configured nr_requests of the drive (https://www.kernel.org/doc/Documentation/block/queue-sysfs.txt). It is also possible to provide a custom number of workers by using this command: `mc admin config set alias/ heal drive_workers=100` . ``` ~ mc admin config set alias/ heal KEY:
Plain Text - Registered: Sun Apr 28 19:28:10 GMT 2024 - Last Modified: Mon Sep 11 21:48:54 GMT 2023 - 17.7K bytes - Viewed (0)