- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 28 for Corkery (0.3 sec)
-
docs/en/docs/deployment/concepts.md
Here are some possible combinations and strategies: * **Gunicorn** managing **Uvicorn workers** * Gunicorn would be the **process manager** listening on the **IP** and **port**, the replication would be by having **multiple Uvicorn worker processes**. * **Uvicorn** managing **Uvicorn workers**
Plain Text - Registered: Sun May 05 07:19:11 GMT 2024 - Last Modified: Thu May 02 22:37:31 GMT 2024 - 18K bytes - Viewed (0) -
cmd/bucket-replication.go
for len(p.workers) < n { input := make(chan ReplicationWorkerOperation, 10000) p.workers = append(p.workers, input) go p.AddWorker(input, &p.activeWorkers) } for len(p.workers) > n { worker := p.workers[len(p.workers)-1] p.workers = p.workers[:len(p.workers)-1] xioutil.SafeClose(worker) } } // ResizeWorkerPriority sets replication failed workers pool size
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Wed May 01 01:09:56 GMT 2024 - 112.2K bytes - Viewed (1) -
docs/en/docs/deployment/docker.md
So, in this case, you **would not** want to have a process manager like Gunicorn with Uvicorn workers, or Uvicorn using its own Uvicorn workers. You would want to have just a **single Uvicorn process** per container (but probably multiple containers).
Plain Text - Registered: Sun May 05 07:19:11 GMT 2024 - Last Modified: Thu May 02 22:37:31 GMT 2024 - 34K bytes - Viewed (0) -
.teamcity/test-buckets.json
"ivy", "testing-native", "maven", "integ-test" ] }, { "parallelizationMethod":{ "name":"TestDistribution" }, "subprojects":[ "workers", "model-core", "language-groovy", "build-init", "logging", "plugins-groovy", "kotlin-dsl", "plugins-java", "samples", "plugin-development",
Json - Registered: Wed May 01 11:36:15 GMT 2024 - Last Modified: Wed May 01 00:36:47 GMT 2024 - 50.1K bytes - Viewed (0) -
cmd/batch-handlers.go
// newBatchJobPool creates a pool of job manifest workers of specified size func newBatchJobPool(ctx context.Context, o ObjectLayer, workers int) *BatchJobPool { jpool := &BatchJobPool{ ctx: ctx, objLayer: o, jobCh: make(chan *BatchJobRequest, 10000), workerKillCh: make(chan struct{}, workers), jobCancelers: make(map[string]context.CancelFunc), }
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Sun May 05 16:56:21 GMT 2024 - 55.2K bytes - Viewed (0) -
cmd/erasure-server-pool-decom.go
const envDecomWorkers = "_MINIO_DECOMMISSION_WORKERS" workerSize, err := env.GetInt(envDecomWorkers, len(pool.sets)) if err != nil { decomLogIf(ctx, fmt.Errorf("invalid workers value err: %v, defaulting to %d", err, len(pool.sets))) workerSize = len(pool.sets) } // Each decom worker needs one List() goroutine/worker // add that many extra workers. workerSize += len(pool.sets)
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Wed May 01 15:18:21 GMT 2024 - 41.5K bytes - Viewed (1) -
cmd/data-scanner_test.go
es := newExpiryState(context.Background(), objAPI, 0) workers := []chan expiryOp{make(chan expiryOp)} es.workers.Store(&workers) globalExpiryState = es var wg sync.WaitGroup wg.Add(1) expired := make([]ObjectToDelete, 0, 5) go func() { defer wg.Done() workers := globalExpiryState.workers.Load() for t := range (*workers)[0] { if t, ok := t.(newerNoncurrentTask); ok {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 03 11:18:58 GMT 2024 - 6.9K bytes - Viewed (0) -
docs/en/docs/deployment/manually.md
```console $ pip install "hypercorn[trio]" ---> 100% ``` </div> ### Run with Trio Then you can pass the command line option `--worker-class` with the value `trio`: <div class="termy"> ```console $ hypercorn main:app --worker-class trio ``` </div> And that will start Hypercorn with your app using Trio as the backend.
Plain Text - Registered: Sun May 05 07:19:11 GMT 2024 - Last Modified: Thu May 02 22:37:31 GMT 2024 - 9.2K bytes - Viewed (0) -
cmd/handler-api.go
} t.replicationPriority = cfg.ReplicationPriority t.replicationMaxWorkers = cfg.ReplicationMaxWorkers // N B api.transition_workers will be deprecated if globalTransitionState != nil { globalTransitionState.UpdateWorkers(cfg.TransitionWorkers) } t.transitionWorkers = cfg.TransitionWorkers t.staleUploadsExpiry = cfg.StaleUploadsExpiry
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 03 20:08:20 GMT 2024 - 10K bytes - Viewed (0) -
okhttp/src/main/kotlin/okhttp3/internal/concurrent/TaskRunner.kt
import okhttp3.internal.assertHeld import okhttp3.internal.concurrent.TaskRunner.Companion.INSTANCE import okhttp3.internal.okHttpName import okhttp3.internal.threadFactory /** * A set of worker threads that are shared among a set of task queues. * * Use [INSTANCE] for a task runner that uses daemon threads. There is not currently a shared * instance for non-daemon threads. *
Plain Text - Registered: Fri May 03 11:42:14 GMT 2024 - Last Modified: Mon Apr 29 00:33:04 GMT 2024 - 10.6K bytes - Viewed (0)