- Sort Score
- Result 10 results
- Languages All
Results 91 - 100 of 1,918 for Nodes (0.07 sec)
-
docs/distributed/README.md
**NOTE:**
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Jan 18 07:03:17 UTC 2024 - 8.8K bytes - Viewed (0) -
pkg/kubelet/cm/devicemanager/manager.go
return false } // If one or the other of nodes[i] or nodes[j] is the fake NUMA node -1 (they can't both be) if nodes[i] == nodeWithoutTopology { return false } if nodes[j] == nodeWithoutTopology { return true } // Otherwise both nodes[i] and nodes[j] are real NUMA nodes that are not in the 'hint's' affinity list. return perNodeDevices[nodes[i]].Len() < perNodeDevices[nodes[j]].Len()
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Apr 15 12:01:56 UTC 2024 - 43K bytes - Viewed (0) -
pkg/scheduler/framework/extender.go
// expected to be a subset of the supplied list. // The failedNodes and failedAndUnresolvableNodes optionally contains the list // of failed nodes and failure reasons, except nodes in the latter are // unresolvable. Filter(pod *v1.Pod, nodes []*NodeInfo) (filteredNodes []*NodeInfo, failedNodesMap extenderv1.FailedNodesMap, failedAndUnresolvable extenderv1.FailedNodesMap, err error)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Feb 26 19:07:19 UTC 2024 - 3.3K bytes - Viewed (0) -
pkg/controller/volume/attachdetach/attach_detach_controller_test.go
t.Fatalf("Error getting list of nodes %v", err) } if i > 100 { t.Fatalf("Time out while waiting for the node informer sync: found %d nodes, expected %d nodes", len(nodeList), nodesNum) } time.Sleep(100 * time.Millisecond) nodeList, err = informerFactory.Core().V1().Nodes().Lister().List(labels.Everything()) i++ } i = 0 podList, err := informerFactory.Core().V1().Pods().Lister().List(labels.Everything())
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Apr 18 11:00:37 UTC 2024 - 24.2K bytes - Viewed (0) -
pkg/kubelet/apis/podresources/server_v1_test.go
DeviceIds: []string{"dev0", "dev1"}, Topology: &podresourcesapi.TopologyInfo{Nodes: []*podresourcesapi.NUMANode{{ID: numaID}}}, }, } cpus := []int64{12, 23, 30} memory := []*podresourcesapi.ContainerMemory{ { MemoryType: "memory", Size_: 1073741824, Topology: &podresourcesapi.TopologyInfo{Nodes: []*podresourcesapi.NUMANode{{ID: numaID}}}, }, { MemoryType: "hugepages-1Gi",
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Apr 24 18:25:29 UTC 2024 - 45.9K bytes - Viewed (0) -
subprojects/core/src/main/java/org/gradle/execution/plan/TaskNodeFactory.java
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Tue Mar 12 02:21:10 UTC 2024 - 5.9K bytes - Viewed (0) -
pkg/scheduler/schedule_one_test.go
}}, } scache.AddNode(logger, &node) nodes = append(nodes, &node) objects = append(objects, &node) } // Create expected failure reasons for all the nodes. Hopefully they will get rolled up into a non-spammy summary. failedNodeStatues := framework.NodeToStatusMap{} for _, node := range nodes { failedNodeStatues[node.Name] = framework.NewStatus( framework.Unschedulable,
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 06:20:55 UTC 2024 - 128.1K bytes - Viewed (0) -
pkg/scheduler/schedule_one.go
} } nodes := allNodes if !preRes.AllNodes() { nodes = make([]*framework.NodeInfo, 0, len(preRes.NodeNames)) for nodeName := range preRes.NodeNames { // PreRes may return nodeName(s) which do not exist; we verify // node exists in the Snapshot. if nodeInfo, err := sched.nodeInfoSnapshot.Get(nodeName); err == nil { nodes = append(nodes, nodeInfo) } } }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Jun 06 13:28:08 UTC 2024 - 43.4K bytes - Viewed (0) -
android/guava/src/com/google/common/graph/AbstractBaseGraph.java
return InvalidatableSet.of( set, () -> nodes().contains(node), () -> String.format(NODE_REMOVED_FROM_GRAPH, node)); } protected final <T> Set<T> nodePairInvalidatableSet(Set<T> set, N nodeU, N nodeV) { return InvalidatableSet.of( set, () -> nodes().contains(nodeU) && nodes().contains(nodeV), () -> String.format(NODE_PAIR_REMOVED_FROM_GRAPH, nodeU, nodeV));
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Jan 22 17:29:38 UTC 2024 - 7.2K bytes - Viewed (0) -
cluster/gce/config-default.sh
# Useful for scheduling heapster in large clusters with nodes of small size. HEAPSTER_MACHINE_TYPE="${HEAPSTER_MACHINE_TYPE:-}" # Optional: Additional nodes would be created if their type and number is specified. # NUM_NODES would be lowered respectively. # Useful for running cluster-level addons that needs more resources than would fit # on small nodes, like network plugins. NUM_ADDITIONAL_NODES="${NUM_ADDITIONAL_NODES:-}"
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat Mar 16 20:16:32 UTC 2024 - 26.9K bytes - Viewed (0)