- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 37 for pmap (0.04 sec)
-
pilot/pkg/serviceregistry/kube/controller/pod.go
} func getPortMap(pod *v1.Pod) map[string]uint32 { pmap := map[string]uint32{} for _, c := range pod.Spec.Containers { for _, port := range c.Ports { if port.Name == "" || port.Protocol != v1.ProtocolTCP { continue } // First port wins, per Kubernetes (https://github.com/kubernetes/kubernetes/issues/54213) if _, f := pmap[port.Name]; !f { pmap[port.Name] = uint32(port.ContainerPort) }
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Thu May 23 21:07:03 UTC 2024 - 10.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/ir/mlrt/tf_mlrt_ops.td
}]; } def MapFnOp : TensorflowMlrt_Op<"map_fn", [AttrSizedOperandSegments, Pure]> { let summary = "The Parallel Map for tf_mlrt dialect"; let description = [{ The Pmap executes body function in parallel for all ranges up to $max_iterations. The pseudo code: for(int i = 0; i < $max_iterations; i++) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 20:44:15 UTC 2024 - 13.6K bytes - Viewed (0) -
src/runtime/map_faststr.go
"unsafe" ) func mapaccess1_faststr(t *maptype, h *hmap, ky string) unsafe.Pointer { if raceenabled && h != nil { callerpc := getcallerpc() racereadpc(unsafe.Pointer(h), callerpc, abi.FuncPCABIInternal(mapaccess1_faststr)) } if h == nil || h.count == 0 { return unsafe.Pointer(&zeroVal[0]) } if h.flags&hashWriting != 0 { fatal("concurrent map read and map write") } key := stringStructOf(&ky) if h.B == 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 15.3K bytes - Viewed (0) -
src/runtime/map_fast64.go
) func mapaccess1_fast64(t *maptype, h *hmap, key uint64) unsafe.Pointer { if raceenabled && h != nil { callerpc := getcallerpc() racereadpc(unsafe.Pointer(h), callerpc, abi.FuncPCABIInternal(mapaccess1_fast64)) } if h == nil || h.count == 0 { return unsafe.Pointer(&zeroVal[0]) } if h.flags&hashWriting != 0 { fatal("concurrent map read and map write") } var b *bmap if h.B == 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 14.1K bytes - Viewed (0) -
src/runtime/map_fast32.go
) func mapaccess1_fast32(t *maptype, h *hmap, key uint32) unsafe.Pointer { if raceenabled && h != nil { callerpc := getcallerpc() racereadpc(unsafe.Pointer(h), callerpc, abi.FuncPCABIInternal(mapaccess1_fast32)) } if h == nil || h.count == 0 { return unsafe.Pointer(&zeroVal[0]) } if h.flags&hashWriting != 0 { fatal("concurrent map read and map write") } var b *bmap if h.B == 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 13.9K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/tools/go/types/typeutil/map.go
// == cannot be used to check for equivalence, and thus we cannot // simply use a Go map. // // Just as with map[K]V, a nil *Map is a valid empty map. // // Not thread-safe. type Map struct { hasher Hasher // shared by many Maps table map[uint32][]entry // maps hash to bucket; entry.key==nil means unused length int // number of map entries } // entry is an entry (key/value association) in a hash bucket.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 02:38:00 UTC 2024 - 13.9K bytes - Viewed (0) -
src/sync/map.go
// The zero Map is empty and ready for use. A Map must not be copied after first use. // // In the terminology of [the Go memory model], Map arranges that a write operation // “synchronizes before” any read operation that observes the effect of the write, where // read and write operations are defined as follows. // [Map.Load], [Map.LoadAndDelete], [Map.LoadOrStore], [Map.Swap], [Map.CompareAndSwap], // and [Map.CompareAndDelete] are read operations;
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 15.6K bytes - Viewed (0) -
pilot/pkg/model/endpointshards.go
// Shardz returns a full deep copy of the global map of shards. This should be used only for testing // and debugging, as the cloning is expensive. func (e *EndpointIndex) Shardz() map[string]map[string]*EndpointShards { e.mu.RLock() defer e.mu.RUnlock() out := make(map[string]map[string]*EndpointShards, len(e.shardsBySvc)) for svcKey, v := range e.shardsBySvc { out[svcKey] = make(map[string]*EndpointShards, len(v))
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Fri Jun 14 04:34:37 UTC 2024 - 15.6K bytes - Viewed (0) -
src/cmd/compile/internal/typecheck/_builtin/runtime.go
func makemap64(mapType *byte, hint int64, mapbuf *any) (hmap map[any]any) func makemap(mapType *byte, hint int, mapbuf *any) (hmap map[any]any) func makemap_small() (hmap map[any]any) func mapaccess1(mapType *byte, hmap map[any]any, key *any) (val *any) func mapaccess1_fast32(mapType *byte, hmap map[any]any, key uint32) (val *any) func mapaccess1_fast64(mapType *byte, hmap map[any]any, key uint64) (val *any)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 21 21:08:03 UTC 2024 - 10.6K bytes - Viewed (0) -
src/internal/trace/summary.go
// (starting from goid). gmap := make(map[GoID]struct{}) gmap[goid] = struct{}{} for i := 0; i < 2; i++ { // Copy the map. gmap1 := make(map[GoID]struct{}) for g := range gmap { gmap1[g] = struct{}{} } for _, edge := range unblockEdges { if _, ok := gmap[edge.operand]; ok { gmap1[edge.operator] = struct{}{} } } gmap = gmap1 } return gmap }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 20.7K bytes - Viewed (0)