- Sort Score
- Result 10 results
- Languages All
Results 81 - 90 of 307 for mask6 (0.1 sec)
-
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_n_z.cc
// - All strides have a known value equal to 1 // - No masks are set (or masks can be applied by transforming the inputs to // Slice) // Verifies that, // // - begin, end and strides operands are 1D and they have the same number of // elements. Here, the number of elements should be less than 32 to support // 32-bit mask attributes. // - None of the strides values are zero.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 22:07:10 UTC 2024 - 170.8K bytes - Viewed (0) -
src/crypto/x509/parser.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:00:16 UTC 2024 - 38.5K bytes - Viewed (0) -
src/net/interface_windows.go
} switch sa := sa.(type) { case *syscall.SockaddrInet4: ifat = append(ifat, &IPNet{IP: IPv4(sa.Addr[0], sa.Addr[1], sa.Addr[2], sa.Addr[3]), Mask: CIDRMask(int(puni.OnLinkPrefixLength), 8*IPv4len)}) case *syscall.SockaddrInet6: ifa := &IPNet{IP: make(IP, IPv6len), Mask: CIDRMask(int(puni.OnLinkPrefixLength), 8*IPv6len)} copy(ifa.IP, sa.Addr[:]) ifat = append(ifat, ifa) } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Mar 15 10:25:02 UTC 2024 - 5.5K bytes - Viewed (0) -
src/runtime/map_fast64.go
m := bucketMask(h.B) b = (*bmap)(add(h.buckets, (hash&m)*uintptr(t.BucketSize))) if c := h.oldbuckets; c != nil { if !h.sameSizeGrow() { // There used to be half as many buckets; mask down one more power of two. m >>= 1 } oldb := (*bmap)(add(c, (hash&m)*uintptr(t.BucketSize))) if !evacuated(oldb) { b = oldb } } } for ; b != nil; b = b.overflow(t) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 14.1K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/LinkedHashMultimap.java
this.hashTable = hashTable; int mask = hashTable.length - 1; for (ValueSetLink<K, V> entry = firstEntry; entry != this; entry = entry.getSuccessorInValueSet()) { ValueEntry<K, V> valueEntry = (ValueEntry<K, V>) entry; int bucket = valueEntry.smearedValueHash & mask; valueEntry.nextInValueBucket = hashTable[bucket];
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Thu Feb 22 21:19:52 UTC 2024 - 23.6K bytes - Viewed (0) -
guava/src/com/google/common/collect/LinkedHashMultimap.java
this.hashTable = hashTable; int mask = hashTable.length - 1; for (ValueSetLink<K, V> entry = firstEntry; entry != this; entry = entry.getSuccessorInValueSet()) { ValueEntry<K, V> valueEntry = (ValueEntry<K, V>) entry; int bucket = valueEntry.smearedValueHash & mask; valueEntry.nextInValueBucket = hashTable[bucket];
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Thu Feb 22 21:19:52 UTC 2024 - 24.3K bytes - Viewed (0) -
src/runtime/map_fast32.go
m := bucketMask(h.B) b = (*bmap)(add(h.buckets, (hash&m)*uintptr(t.BucketSize))) if c := h.oldbuckets; c != nil { if !h.sameSizeGrow() { // There used to be half as many buckets; mask down one more power of two. m >>= 1 } oldb := (*bmap)(add(c, (hash&m)*uintptr(t.BucketSize))) if !evacuated(oldb) { b = oldb } } } for ; b != nil; b = b.overflow(t) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 13.9K bytes - Viewed (0) -
pkg/controller/nodeipam/ipam/range_allocator.go
// clusterCIDR's Mask applied (this means that clusterCIDR contains // serviceCIDR) or vice versa (which means that serviceCIDR contains // clusterCIDR). for idx, cidr := range r.clusterCIDRs { // if they don't overlap then ignore the filtering if !cidr.Contains(serviceCIDR.IP.Mask(cidr.Mask)) && !serviceCIDR.Contains(cidr.IP.Mask(serviceCIDR.Mask)) { continue }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Apr 24 10:06:15 UTC 2024 - 16.2K bytes - Viewed (0) -
src/crypto/internal/edwards25519/field/_asm/fe_amd64_asm.go
out = r.hi lo = r.lo SHLQ(Imm(64-51), r.lo, r.hi) r.lo, r.hi = nil, nil // make sure the uint128 is unusable return } // maskAndAdd sets r = r&mask + c*i. func maskAndAdd(r, mask, c GPVirtual, i uint64) { ANDQ(mask, r) if i != 1 { IMUL3Q(Imm(i), c, c) } ADDQ(c, r) } func mustAddr(c Component) Op { b, err := c.Resolve() if err != nil { panic(err) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:29:44 UTC 2024 - 7.2K bytes - Viewed (0) -
src/internal/runtime/atomic/atomic_arm.go
uaddr := uintptr(unsafe.Pointer(addr)) addr32 := (*uint32)(unsafe.Pointer(uaddr &^ 3)) word := uint32(v) << ((uaddr & 3) * 8) // little endian mask := uint32(0xFF) << ((uaddr & 3) * 8) // little endian word |= ^mask for { old := *addr32 if Cas(addr32, old, old&word) { return } } } //go:nosplit func Or(addr *uint32, v uint32) { for { old := *addr
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 15 19:57:43 UTC 2024 - 4.8K bytes - Viewed (0)