- Sort Score
- Num 10 results
- Language All
Results 171 - 180 of 265 for grand (0.02 seconds)
-
docs/de/docs/tutorial/schema-extra-example.md
Aus diesem Grund verwendeten Versionen von FastAPI vor 0.99.0 immer noch Versionen von OpenAPI vor 3.1.0. /// ### Pydantic- und FastAPI-`examples` { #pydantic-and-fastapi-examples }Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Thu Mar 19 17:58:09 GMT 2026 - 10.4K bytes - Click Count (0) -
docs/distributed/DESIGN.md
```go func getAvailablePoolIdx(ctx context.Context) int { serverPools := z.getServerPoolsAvailableSpace(ctx) total := serverPools.TotalAvailable() // choose when we reach this many choose := rand.Uint64() % total atTotal := uint64(0) for _, pool := range serverPools { atTotal += pool.Available if atTotal > choose && pool.Available > 0 {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Wed Feb 26 09:25:50 GMT 2025 - 8K bytes - Click Count (2) -
cmd/erasure-healing_test.go
data := make([]byte, 1*humanize.MiByte) length := int64(len(data)) _, err = rand.Read(data) if err != nil { t.Fatal(err) } oi1, err := obj.PutObject(ctx, bucket, object, mustGetPutObjReader(t, bytes.NewReader(data), length, "", ""), ObjectOptions{}) if err != nil { t.Fatal(err) } // 2nd version. _, _ = rand.Read(data)
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 48.5K bytes - Click Count (0) -
cmd/erasure-sets.go
// the set topology, this monitoring happens at a given monitoring interval. func (s *erasureSets) monitorAndConnectEndpoints(ctx context.Context, monitorInterval time.Duration) { r := rand.New(rand.NewSource(time.Now().UnixNano())) time.Sleep(time.Duration(r.Float64() * float64(time.Second))) // Pre-emptively connect the disks if possible. s.connectDisks(false)
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 37K bytes - Click Count (1) -
cmd/xl-storage-disk-id-check.go
if skipIfSuccessBefore <= 0 { skipIfSuccessBefore = globalDriveConfig.GetMaxTimeout() } } t := time.NewTicker(checkEvery) defer t.Stop() fn := mustGetUUID() rng := rand.New(rand.NewSource(time.Now().UnixNano())) monitor := func() bool { if contextCanceled(ctx) { return false } if p.health.status.Load() != diskHealthOK { return true }
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Apr 25 05:41:04 GMT 2025 - 34.5K bytes - Click Count (0) -
cmd/erasure-server-pool-decom.go
go func() { // Resume decommissioning of pools, but wait 3 minutes for cluster to stabilize. if err := sleepContext(ctx, 3*time.Minute); err != nil { return } r := rand.New(rand.NewSource(time.Now().UnixNano())) for { if err := z.Decommission(ctx, poolIndices...); err != nil { if errors.Is(err, errDecommissionAlreadyRunning) { // A previous decommission running found restart it.
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 42.2K bytes - Click Count (1) -
cmd/iam.go
globalServerConfigMu.RLock() s := globalServerConfig globalServerConfigMu.RUnlock() sys.Lock() sys.iamRefreshInterval = iamRefreshInterval sys.Unlock() r := rand.New(rand.NewSource(time.Now().UnixNano())) var ( openidInit bool ldapInit bool authNInit bool authZInit bool ) stsTLSConfig, err := xtls.Lookup(s[config.IdentityTLSSubSys][config.Default])
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Wed Oct 15 17:00:45 GMT 2025 - 76.5K bytes - Click Count (0) -
docs/fr/docs/deployment/concepts.md
### Multiples processus - Workers { #multiple-processes-workers } Si vous avez plus de clients que ce qu'un seul processus peut gérer (par exemple si la machine virtuelle n'est pas très grande) et que vous avez **plusieurs cœurs** dans le CPU du serveur, alors vous pouvez avoir **plusieurs processus** exécutant la même application simultanément, et distribuer toutes les requêtes entre eux.Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Thu Mar 19 18:37:13 GMT 2026 - 21.4K bytes - Click Count (0) -
docs/es/docs/advanced/events.md
Por eso, declaramos la función manejadora del evento con `def` estándar en vez de `async def`. /// ### `startup` y `shutdown` juntos { #startup-and-shutdown-together } Hay una gran posibilidad de que la lógica para tu *startup* y *shutdown* esté conectada, podrías querer iniciar algo y luego finalizarlo, adquirir un recurso y luego liberarlo, etc.Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Thu Mar 19 18:15:55 GMT 2026 - 8.4K bytes - Click Count (0) -
docs/es/docs/tutorial/dependencies/index.md
Esto será especialmente útil cuando lo uses en una **gran code base** donde uses **las mismas dependencias** una y otra vez en **muchas *path operations***. ## Usar `async` o no usar `async` { #to-async-or-not-to-async }Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Thu Mar 19 18:15:55 GMT 2026 - 10K bytes - Click Count (0)