- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 31 for pod_1 (0.15 sec)
-
pkg/controller/endpointslice/endpointslice_controller_test.go
func TestSyncServicePodSelection(t *testing.T) { client, esController := newController(t, []string{"node-1"}, time.Duration(0)) ns := metav1.NamespaceDefault pod1 := newPod(1, ns, true, 0, false) esController.podStore.Add(pod1) // ensure this pod will not match the selector pod2 := newPod(2, ns, true, 0, false) pod2.Labels["foo"] = "boo" esController.podStore.Add(pod2)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 08:33:32 UTC 2024 - 65.5K bytes - Viewed (0) -
pkg/controller/job/job_controller_test.go
pod1 := newPod("pod1", job1) pod1.OwnerReferences = nil informer.Core().V1().Pods().Informer().GetIndexer().Add(pod1) // Labels changed on orphan. Expect newly matching controllers to queue. prev := *pod1 prev.Labels = map[string]string{"foo2": "bar2"} bumpResourceVersion(pod1) jm.updatePod(logger, &prev, pod1) verifyEmptyQueueAndAwaitForQueueLen(ctx, t, jm, 2)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 11 15:36:36 UTC 2024 - 229.2K bytes - Viewed (0) -
docs/distributed/decom-encrypted-sse-s3.sh
policy_count=$(./mc admin policy list myminio/ | wc -l) kill $pid (minio server http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_1.log) & pid_1=$! (minio server --address ":9001" http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_2.log) & pid_2=$! ./mc ready myminio
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon May 27 19:17:46 UTC 2024 - 4.5K bytes - Viewed (0) -
docs/distributed/decom.sh
./mc ls -r --versions mytier/tiered/ >tiered_ns_versions.txt kill $pid (minio server http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_1.log) & pid_1=$! (minio server --address ":9001" http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_2.log) & pid_2=$! ./mc ready myminio
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon May 27 19:17:46 UTC 2024 - 6.4K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/storage/cacher/watch_cache_test.go
func TestWatchCacheBasic(t *testing.T) { store := newTestWatchCache(2, &cache.Indexers{}) defer store.Stop() // Test Add/Update/Delete. pod1 := makeTestPod("pod", 1) if err := store.Add(pod1); err != nil { t.Errorf("unexpected error: %v", err) } if item, ok, _ := store.Get(pod1); !ok { t.Errorf("didn't find pod") } else { expected := makeTestStoreElement(makeTestPod("pod", 1))
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 09:20:10 UTC 2024 - 35.4K bytes - Viewed (0) -
docs/distributed/decom-encrypted-kes.sh
policy_count=$(./mc admin policy list myminio/ | wc -l) kill $pid (minio server http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_1.log) & pid_1=$! (minio server --address ":9001" http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_2.log) & pid_2=$! ./mc ready myminio
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon May 27 19:17:46 UTC 2024 - 7.1K bytes - Viewed (0) -
pkg/kubelet/cm/devicemanager/manager_test.go
as.Nil(err) err = testManager.Allocate(pod2, &pod2.Spec.Containers[0]) as.Nil(err) // when pod is in activePods, GetDeviceRunContainerOptions should return runContainerOpts, err := testManager.GetDeviceRunContainerOptions(pod1, &pod1.Spec.Containers[0]) as.Nil(err) as.Equal(len(runContainerOpts.Devices), 3) as.Equal(len(runContainerOpts.Mounts), 2)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 06:25:43 UTC 2024 - 65K bytes - Viewed (0) -
pkg/scheduler/framework/plugins/noderesources/fit_test.go
"queue-on-target-pod-some-resource-scale-down": { pod: st.MakePod().Name("pod1").Req(map[v1.ResourceName]string{v1.ResourceCPU: "1"}).Obj(), oldObj: st.MakePod().Name("pod1").Req(map[v1.ResourceName]string{v1.ResourceCPU: "2"}).Obj(), newObj: st.MakePod().Name("pod1").Req(map[v1.ResourceName]string{v1.ResourceCPU: "1"}).Obj(), enableInPlacePodVerticalScaling: true,
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 13:26:09 UTC 2024 - 57.4K bytes - Viewed (0) -
pkg/scheduler/scheduler_test.go
}, }, waitSchedulingPods: []*v1.Pod{ st.MakePod().Name("pod1").UID("pod1").SchedulerName(testSchedulerProfile1).Obj(), st.MakePod().Name("pod2").UID("pod2").SchedulerName(testSchedulerProfile1).Obj(), st.MakePod().Name("pod3").UID("pod3").SchedulerName(testSchedulerProfile1).Obj(), }, expectPodNamesInWaitingPods: []string{"pod1", "pod2", "pod3"}, }, {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri May 17 09:07:27 UTC 2024 - 42K bytes - Viewed (0) -
pkg/kubelet/kubelet_test.go
}, }, }, { name: "No CPU and memory, resource allocation exists", pod: podWithUIDNameNsSpec("11", "pod11", "foo", *emptyPodSpec), existingPodAllocation: podWithUIDNameNsSpec("11", "pod11", "foo", *emptyPodSpec), expectedPodResourceAllocation: state.PodResourceAllocation{ "11": map[string]v1.ResourceList{
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Jun 14 16:09:17 UTC 2024 - 106.9K bytes - Viewed (0)