- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 18 for queue (0.15 sec)
-
cni/pkg/repair/repaircontroller.go
}) c.queue = controllers.NewQueue("repair pods", controllers.WithReconciler(c.Reconcile), controllers.WithMaxAttempts(5)) c.pods.AddEventHandler(controllers.ObjectHandler(c.queue.AddObject)) return c, nil } func (c *Controller) Run(stop <-chan struct{}) { kube.WaitForCacheSync("repair controller", stop, c.pods.HasSynced) c.queue.Run(stop) c.pods.ShutdownHandlers() }
Go - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Sat Feb 10 00:31:55 GMT 2024 - 10.4K bytes - Viewed (0) -
cni/pkg/repair/repair_test.go
c, err := NewRepairController(tt.client, tt.config) assert.NoError(t, err) t.Cleanup(func() { assert.NoError(t, c.queue.WaitForClose(time.Second)) }) stop := test.NewStop(t) tt.client.RunAndWait(stop) go c.Run(stop) kube.WaitForCacheSync("test", stop, c.queue.HasSynced) assert.EventuallyEqual(t, func() map[string]string { havePods := c.pods.List(metav1.NamespaceAll, klabels.Everything())
Go - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Tue Oct 24 03:31:28 GMT 2023 - 10.6K bytes - Viewed (0) -
architecture/networking/pilot.md
The Push Queue is mostly a normal queue, but it has some special logic to merge push requests for each given proxy. This results in each proxy having 0 or 1 outstanding push requests; if additional updates come in the existing push request is just expanded. Another job polls this queue and triggers each client to start a push. ```mermaid graph TD
Plain Text - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Wed Feb 07 17:53:24 GMT 2024 - 19.1K bytes - Viewed (0) -
common-protos/k8s.io/api/autoscaling/v2beta2/generated.proto
// +optional optional string apiVersion = 3; } // ExternalMetricSource indicates how to scale on a metric not associated with // any Kubernetes object (for example length of queue in cloud // messaging service, or QPS from loadbalancer running outside of cluster). message ExternalMetricSource { // metric identifies the target metric by name and selector optional MetricIdentifier metric = 1;
Plain Text - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Mon Mar 11 18:43:24 GMT 2024 - 21K bytes - Viewed (0) -
common-protos/k8s.io/api/autoscaling/v2beta1/generated.proto
// API version of the referent // +optional optional string apiVersion = 3; } // ExternalMetricSource indicates how to scale on a metric not associated with // any Kubernetes object (for example length of queue in cloud // messaging service, or QPS from loadbalancer running outside of cluster). // Exactly one "target" type should be set. message ExternalMetricSource { // metricName is the name of the metric in question.
Plain Text - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Mon Mar 11 18:43:24 GMT 2024 - 21.2K bytes - Viewed (0) -
cni/pkg/nodeagent/ztunnelserver.go
so that we can remove stale entries when the ztunnel pod is deleted or when the ztunnel pod is restarted in the same pod (remove old entries when the same uid connects again, but with different boot id?) save a queue of what needs to be sent to the ztunnel pod and send it one by one when it connects. when a new ztunnel connects with different uid, only propagate deletes to older ztunnels. */ type connMgr struct {
Go - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Fri Apr 12 21:47:31 GMT 2024 - 12.4K bytes - Viewed (0) -
cni/pkg/nodeagent/informers.go
systemNamespace string queue controllers.Queue pods kclient.Client[*corev1.Pod] namespaces kclient.Client[*corev1.Namespace] } func setupHandlers(ctx context.Context, kubeClient kube.Client, dataplane MeshDataplane, systemNamespace string) *InformerHandlers { s := &InformerHandlers{ctx: ctx, dataplane: dataplane, systemNamespace: systemNamespace} s.queue = controllers.NewQueue("ambient",
Go - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Thu Feb 08 01:03:24 GMT 2024 - 9.9K bytes - Viewed (0) -
manifests/charts/istiod-remote/templates/crd-all.gen.yaml
type: string http1MaxPendingRequests: description: Maximum number of requests that will be queued while waiting for a ready connection pool connection. format: int32 type: integer
Others - Registered: Wed Mar 20 22:53:08 GMT 2024 - Last Modified: Fri Mar 15 18:46:49 GMT 2024 - 570.3K bytes - Viewed (0) -
common-protos/k8s.io/api/flowcontrol/v1beta2/generated.proto
} // LimitResponse defines how to handle requests that can not be executed right now. // +union message LimitResponse { // `type` is "Queue" or "Reject". // "Queue" means that requests that can not be executed upon arrival // are held in a queue until they can be executed or a queuing limit // is reached. // "Reject" means that requests that can not be executed upon arrival // are rejected.
Plain Text - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Mon Mar 11 18:43:24 GMT 2024 - 19.4K bytes - Viewed (0) -
common-protos/k8s.io/api/flowcontrol/v1alpha1/generated.proto
} // LimitResponse defines how to handle requests that can not be executed right now. // +union message LimitResponse { // `type` is "Queue" or "Reject". // "Queue" means that requests that can not be executed upon arrival // are held in a queue until they can be executed or a queuing limit // is reached. // "Reject" means that requests that can not be executed upon arrival // are rejected.
Plain Text - Registered: Wed Apr 24 22:53:08 GMT 2024 - Last Modified: Mon Mar 11 18:43:24 GMT 2024 - 19.4K bytes - Viewed (0)