- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 113 for maskOf (0.09 sec)
-
src/net/interface_plan9.go
} if ip.To16() != nil && ip.To4() == nil { // IPv6 address mask = CIDRMask(pfxlen, 8*IPv6len) } addrs = append(addrs, &IPNet{IP: ip, Mask: mask}) } } return addrs, nil } // interfaceMulticastAddrTable returns addresses for a specific // interface. func interfaceMulticastAddrTable(ifi *Interface) ([]Addr, error) { return nil, nil
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 06 14:00:54 UTC 2024 - 4.7K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/cel/library/cost.go
// So we double the cost of parsing the string. cost := uint64(math.Ceil(float64(actualSize(args[0])) * 2 * common.StringTraversalCostFactor)) return &cost } case "masked", "prefixLength", "family", "isUnspecified", "isLoopback", "isLinkLocalMulticast", "isLinkLocalUnicast": // IP and CIDR accessors are nominal cost. cost := uint64(1) return &cost case "containsIP":
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Apr 23 17:22:44 UTC 2024 - 20.6K bytes - Viewed (0) -
src/net/netip/inlining_test.go
"Addr.Zone", "Addr.v4", "Addr.v6", "Addr.v6u16", "Addr.withoutZone", "AddrPortFrom", "AddrPort.Addr", "AddrPort.Port", "AddrPort.IsValid", "Prefix.IsSingleIP", "Prefix.Masked", "Prefix.IsValid", "PrefixFrom", "Prefix.Addr", "Prefix.Bits", "AddrFrom4", "IPv6LinkLocalAllNodes", "IPv6Unspecified", "MustParseAddr", "MustParseAddrPort", "MustParsePrefix",
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jun 04 17:10:01 UTC 2024 - 2K bytes - Viewed (0) -
src/crypto/internal/mlkem768/mlkem768.go
u[i] = ringDecodeAndDecompress10(b) } b := (*[encodingSize4]byte)(c[encodingSize10*k:]) v := ringDecodeAndDecompress4(b) var mask nttElement // s⊺ ◦ NTT(u) for i := range dx.s { mask = polyAdd(mask, nttMul(dx.s[i], ntt(u[i]))) } w := polySub(v, inverseNTT(mask)) return ringCompressAndEncode1(nil, w) } // fieldElement is an integer modulo q, an element of ℤ_q. It is always reduced.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 13 18:57:38 UTC 2024 - 28.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/prepare_tf.cc
padded_val.push_back(padding_val[i]); if (mask) *mask |= 1 << i; } } LogicalResult matchAndRewrite(Operation *op, PatternRewriter &rewriter) const override { TF::StridedSliceOp strided_slice_op = llvm::cast<TF::StridedSliceOp>(op); // Handle ellipsis mask. if (strided_slice_op.getEllipsisMask() != 0) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 21:49:50 UTC 2024 - 64.6K bytes - Viewed (0) -
guava/src/com/google/common/hash/Striped64.java
abstract class Striped64 extends Number { /* * This class maintains a lazily-initialized table of atomically * updated variables, plus an extra "base" field. The table size * is a power of two. Indexing uses masked per-thread hash codes. * Nearly all declarations in this class are package-private, * accessed directly by subclasses. * * Table entries are of class Cell; a variant of AtomicLong padded
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/RegularImmutableMap.java
return null; } if (hashTableObject instanceof byte[]) { byte[] hashTable = (byte[]) hashTableObject; int mask = hashTable.length - 1; for (int h = Hashing.smear(key.hashCode()); ; h++) { h &= mask; int keyIndex = hashTable[h] & BYTE_MASK; // unsigned read if (keyIndex == BYTE_MASK) { // -1 signed becomes 255 unsigned return null;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Apr 15 22:32:14 UTC 2024 - 22.7K bytes - Viewed (0) -
cmd/listen-notification-handlers.go
} pattern := event.NewPattern(prefix, suffix) var eventNames []event.Name var mask pubsub.Mask for _, s := range values[peerRESTListenEvents] { eventName, err := event.ParseName(s) if err != nil { writeErrorResponse(ctx, w, toAPIError(ctx, err), r.URL) return } mask.MergeMaskable(eventName) eventNames = append(eventNames, eventName) } if bucketName != "" {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc
ReducedPrecisionSupport mask = ReducedPrecisionSupport::None; if (toco_flags.quantize_to_float16()) { mask |= ReducedPrecisionSupport::Float16Inference; } if (toco_flags.allow_bfloat16()) { mask |= ReducedPrecisionSupport::Bfloat16Inference; } if (toco_flags.accumulation_type() == toco::IODataType::FLOAT16) { mask |= ReducedPrecisionSupport::Float16Accumulation;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun May 12 12:39:37 UTC 2024 - 17.3K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/server/options/encryptionconfig/config.go
// check if resource is masked by *.group rule anyResourceInGroup := schema.GroupResource{Group: gr.Group, Resource: "*"} if _, masked := resourceToPrefixTransformer[anyResourceInGroup]; masked { // an earlier rule already configured a transformer for *.group, masking this rule // return error since this is not allowed
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu May 16 16:56:39 UTC 2024 - 41.2K bytes - Viewed (0)