- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 66 for pod_1 (0.1 sec)
-
docs/distributed/decom.sh
./mc ls -r --versions mytier/tiered/ >tiered_ns_versions.txt kill $pid (minio server http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_1.log) & pid_1=$! (minio server --address ":9001" http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_2.log) & pid_2=$! ./mc ready myminio
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon May 27 19:17:46 UTC 2024 - 6.4K bytes - Viewed (0) -
pkg/controller/volume/attachdetach/cache/desired_state_of_world_test.go
if len(volumesToAttach) != 0 { t.Fatalf("len(volumesToAttach) Expected: <0> Actual: <%v>", len(volumesToAttach)) } } // Populates data struct with pod1/volume/node and pod2/volume/node. // Calls DeleteNode() to delete the pod1/volume/node. // Verifies volume still exists, and one volumes to attach. func Test_DeletePod_Positive_2PodsExistNodeExistsVolumesExist(t *testing.T) { // Arrange
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Mar 01 10:42:15 UTC 2024 - 35.9K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/storage/cacher/watch_cache_test.go
func TestWatchCacheBasic(t *testing.T) { store := newTestWatchCache(2, &cache.Indexers{}) defer store.Stop() // Test Add/Update/Delete. pod1 := makeTestPod("pod", 1) if err := store.Add(pod1); err != nil { t.Errorf("unexpected error: %v", err) } if item, ok, _ := store.Get(pod1); !ok { t.Errorf("didn't find pod") } else { expected := makeTestStoreElement(makeTestPod("pod", 1))
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 09:20:10 UTC 2024 - 35.4K bytes - Viewed (0) -
docs/distributed/decom-encrypted-kes.sh
policy_count=$(./mc admin policy list myminio/ | wc -l) kill $pid (minio server http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_1.log) & pid_1=$! (minio server --address ":9001" http://localhost:9000/tmp/xl/{1...10}/disk{0...1} http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/expanded_2.log) & pid_2=$! ./mc ready myminio
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon May 27 19:17:46 UTC 2024 - 7.1K bytes - Viewed (0) -
pkg/kubelet/cm/devicemanager/manager_test.go
as.Nil(err) err = testManager.Allocate(pod2, &pod2.Spec.Containers[0]) as.Nil(err) // when pod is in activePods, GetDeviceRunContainerOptions should return runContainerOpts, err := testManager.GetDeviceRunContainerOptions(pod1, &pod1.Spec.Containers[0]) as.Nil(err) as.Equal(len(runContainerOpts.Devices), 3) as.Equal(len(runContainerOpts.Mounts), 2)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 06:25:43 UTC 2024 - 65K bytes - Viewed (0) -
pkg/controller/endpoint/endpoints_controller_test.go
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Feb 26 06:51:56 UTC 2024 - 87.7K bytes - Viewed (0) -
pkg/scheduler/framework/plugins/noderesources/fit_test.go
"queue-on-target-pod-some-resource-scale-down": { pod: st.MakePod().Name("pod1").Req(map[v1.ResourceName]string{v1.ResourceCPU: "1"}).Obj(), oldObj: st.MakePod().Name("pod1").Req(map[v1.ResourceName]string{v1.ResourceCPU: "2"}).Obj(), newObj: st.MakePod().Name("pod1").Req(map[v1.ResourceName]string{v1.ResourceCPU: "1"}).Obj(), enableInPlacePodVerticalScaling: true,
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 13:26:09 UTC 2024 - 57.4K bytes - Viewed (0) -
pkg/scheduler/extender_test.go
podsInNodeList: []*v1.Pod{ st.MakePod().Name("pod1").UID("uid1").Obj(), st.MakePod().Name("pod2").UID("uid2").Obj(), st.MakePod().Name("pod3").UID("uid3").Obj(), st.MakePod().Name("pod4").UID("uid4").Obj(), }, nodeInfos: nil, want: map[string]*extenderv1.Victims{ "node1": { Pods: []*v1.Pod{ st.MakePod().Name("pod1").UID("uid1").Obj(),
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Feb 26 19:07:19 UTC 2024 - 16.7K bytes - Viewed (0) -
pkg/scheduler/scheduler_test.go
}, }, waitSchedulingPods: []*v1.Pod{ st.MakePod().Name("pod1").UID("pod1").SchedulerName(testSchedulerProfile1).Obj(), st.MakePod().Name("pod2").UID("pod2").SchedulerName(testSchedulerProfile1).Obj(), st.MakePod().Name("pod3").UID("pod3").SchedulerName(testSchedulerProfile1).Obj(), }, expectPodNamesInWaitingPods: []string{"pod1", "pod2", "pod3"}, }, {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri May 17 09:07:27 UTC 2024 - 42K bytes - Viewed (0) -
pkg/controller/deployment/deployment_controller_test.go
rs1 := newReplicaSet(foo, "foo-1", 1) rs2 := newReplicaSet(foo, "foo-1", 1) pod1 := generatePodFromRS(rs1) pod2 := generatePodFromRS(rs2) f.dLister = append(f.dLister, foo) // Let's pretend this is a different pod. The gist is that the pod lister needs to // return a non-empty list. f.podLister = append(f.podLister, pod1, pod2) c, _, err := f.newController(ctx) if err != nil {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 04 18:33:12 UTC 2024 - 32.5K bytes - Viewed (0)