- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 852 for limiter (0.36 sec)
-
internal/bucket/bandwidth/monitor.go
m.tlock.RLock() defer m.tlock.RUnlock() return m.bucketsThrottle[opts] } // SetBandwidthLimit sets the bandwidth limit for a bucket func (m *Monitor) SetBandwidthLimit(bucket, arn string, limit int64) { m.tlock.Lock() defer m.tlock.Unlock() limitBytes := limit / int64(m.NodeCount) throttle, ok := m.bucketsThrottle[BucketOptions{Name: bucket, ReplicationARN: arn}] if !ok { throttle = &bucketThrottle{}
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon Feb 19 22:54:46 UTC 2024 - 6K bytes - Viewed (0) -
src/runtime/export_test.go
return l } func (l *GCCPULimiter) Fill() uint64 { return l.limiter.bucket.fill } func (l *GCCPULimiter) Capacity() uint64 { return l.limiter.bucket.capacity } func (l *GCCPULimiter) Overflow() uint64 { return l.limiter.overflow } func (l *GCCPULimiter) Limiting() bool { return l.limiter.limiting() } func (l *GCCPULimiter) NeedUpdate(now int64) bool {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:50:53 UTC 2024 - 46.1K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/util/flowcontrol/apf_controller.go
if limiter := pl.Spec.Limited; limiter != nil { return limiter.NominalConcurrencyShares, limiter.LendablePercent, limiter.BorrowingLimitPercent } limiter := pl.Spec.Exempt var nominalConcurrencyShares int32 if limiter.NominalConcurrencyShares != nil { nominalConcurrencyShares = *limiter.NominalConcurrencyShares }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 04 18:33:12 UTC 2024 - 48.8K bytes - Viewed (0) -
src/runtime/metrics/description.go
"allocs-by-size and frees-by-size.", Kind: KindUint64, Cumulative: true, }, { Name: "/gc/limiter/last-enabled:gc-cycle", Description: "GC cycle the last time the GC CPU limiter was enabled. " + "This metric is useful for diagnosing the root cause of an out-of-memory " + "error, because the limiter trades memory for CPU time when the GC's CPU " +
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Dec 06 17:59:12 UTC 2023 - 19.6K bytes - Viewed (0) -
pkg/controller/garbagecollector/garbagecollector_test.go
if e, a := len(t.pendingList), t.limiter.Len(); e != a { panic(fmt.Errorf("pendingList != limiter.Len(): %d / %d", e, a)) } return len(t.pendingList) } func (t *trackingWorkqueue[T]) ShutDown() { t.limiter.ShutDown() } func (t *trackingWorkqueue[T]) ShutDownWithDrain() { t.limiter.ShutDownWithDrain() } func (t *trackingWorkqueue[T]) ShuttingDown() bool { return t.limiter.ShuttingDown() }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 04 18:33:12 UTC 2024 - 111.6K bytes - Viewed (0) -
pilot/pkg/serviceregistry/kube/controller/namespacecontroller.go
CACertNamespaceConfigMap = "istio-ca-root-cert" // maxRetries is the number of times a namespace will be retried before it is dropped out of the queue. // With the current rate-limiter in use (5ms*2^(maxRetries-1)) the following numbers represent the // sequence of delays between successive queuing of a namespace. // // 5ms, 10ms, 20ms, 40ms, 80ms maxRetries = 5 )
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Wed Feb 28 16:41:38 UTC 2024 - 5.4K bytes - Viewed (0) -
pkg/controller/certificates/certificate_controller.go
// 10 qps, 100 bucket size. This is only for retry speed and its only the overall factor (not per item) &workqueue.TypedBucketRateLimiter[string]{Limiter: rate.NewLimiter(rate.Limit(10), 100)}, ), workqueue.TypedRateLimitingQueueConfig[string]{ Name: "certificate", }, ), handler: handler, } // Manage the addition/update of certificate requests
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 04 18:33:12 UTC 2024 - 6.4K bytes - Viewed (0) -
src/runtime/metrics/doc.go
runtime, only their block. Each block is already accounted for in allocs-by-size and frees-by-size. /gc/limiter/last-enabled:gc-cycle GC cycle the last time the GC CPU limiter was enabled. This metric is useful for diagnosing the root cause of an out-of-memory error, because the limiter trades memory for CPU time when the GC's CPU time gets too high. This is most likely
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:58:43 UTC 2024 - 20K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/storage/storagebackend/factory/etcd3.go
if prober != nil { prober.Close() clientErr = fmt.Errorf("server is shutting down") } }() // limit to a request every half of the configured timeout with a maximum burst of one // rate limited requests will receive the last request sent error (note: not the last received response) limiter := rate.NewLimiter(rate.Every(timeout/2), 1) // initial state is the clientErr
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Apr 24 07:56:39 UTC 2024 - 15.1K bytes - Viewed (0) -
pilot/pkg/autoregistration/controller.go
// cleanup is to delete the workload entry // queue contains workloadEntry that need to be unregistered queue controllers.Queue // cleanupLimit rate limit's auto registered WorkloadEntry cleanup calls to k8s cleanupLimit *rate.Limiter // cleanupQueue delays the cleanup of auto registered WorkloadEntries to allow for grace period cleanupQueue queue.Delayed adsConnections *adsConnections
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Tue Apr 16 00:00:36 UTC 2024 - 26.4K bytes - Viewed (0)