- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 32 for bitMasks (0.46 sec)
-
pkg/kubelet/cm/cpumanager/policy_static_test.go
topo: topoDualSocketHT, stAssignments: state.ContainerCPUAssignments{}, stDefaultCPUSet: cpuset.New(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11), numRequested: 2, socketMask: func() bitmask.BitMask { mask, _ := bitmask.NewBitMask(1) return mask }(), expCSet: cpuset.New(1, 7), }, { description: "Request 8 CPUs, BitMask on Socket 0",
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Apr 24 18:25:29 UTC 2024 - 40.8K bytes - Viewed (0) -
src/sync/atomic/type.go
// And atomically performs a bitwise AND operation on x using the bitmask // provided as mask and returns the old value. func (x *Int32) And(mask int32) (old int32) { return AndInt32(&x.v, mask) } // Or atomically performs a bitwise OR operation on x using the bitmask // provided as mask and returns the old value. func (x *Int32) Or(mask int32) (old int32) { return OrInt32(&x.v, mask) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 8.5K bytes - Viewed (0) -
src/sync/atomic/doc.go
// AndInt32 atomically performs a bitwise AND operation on *addr using the bitmask provided as mask // and returns the old value. // Consider using the more ergonomic and less error-prone [Int32.And] instead. func AndInt32(addr *int32, mask int32) (old int32) // AndUint32 atomically performs a bitwise AND operation on *addr using the bitmask provided as mask // and returns the old value.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 11.7K bytes - Viewed (0) -
pkg/kubelet/cm/cpumanager/topology_hints_test.go
cpuSetAcrossSocket, _ := cpuset.Parse("0-28,40-57") m0001, _ := bitmask.NewBitMask(0) m0011, _ := bitmask.NewBitMask(0, 1) m0101, _ := bitmask.NewBitMask(0, 2) m1001, _ := bitmask.NewBitMask(0, 3) m0111, _ := bitmask.NewBitMask(0, 1, 2) m1011, _ := bitmask.NewBitMask(0, 1, 3) m1101, _ := bitmask.NewBitMask(0, 2, 3) m1111, _ := bitmask.NewBitMask(0, 1, 2, 3) testCases := []struct { description string
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Apr 24 18:25:29 UTC 2024 - 19K bytes - Viewed (0) -
pkg/kubelet/cm/devicemanager/topology_hints.go
minAffinitySize := len(m.numaNodes) // Iterate through all combinations of NUMA Nodes and build hints from them. hints := []topologymanager.TopologyHint{} bitmask.IterateBitMasks(m.numaNodes, func(mask bitmask.BitMask) { // First, update minAffinitySize for the current request size. devicesInMask := 0 for _, device := range m.allDevices[resource] { if mask.AnySet(m.getNUMANodeIds(device.Topology)) {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat Jan 27 02:10:25 UTC 2024 - 9.9K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/cel/mutation/common/constants.go
import ( "github.com/google/cel-go/common/types/traits" ) // RootTypeReferenceName is the root reference that all type names should start with. const RootTypeReferenceName = "Object" // ObjectTraits is the bitmask that represents traits that an object should have.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jan 24 00:01:35 UTC 2024 - 897 bytes - Viewed (0) -
src/crypto/rsa/pss.go
// stop. var bitMask byte = 0xff >> (8*emLen - emBits) if em[0] & ^bitMask != 0 { return ErrVerification } // 7. Let dbMask = MGF(H, emLen - hLen - 1). // // 8. Let DB = maskedDB \xor dbMask. mgf1XOR(db, hash, h) // 9. Set the leftmost 8 * emLen - emBits bits of the leftmost octet in DB // to zero. db[0] &= bitMask
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:11:18 UTC 2024 - 11K bytes - Viewed (0) -
src/runtime/map_test.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 33.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.h
// A serialized "QuantizationInfo" object to specify value ranges for some of // the tensors with known names. std::string serialized_quant_stats = ""; // A bitmask to encode support for reduced precision inference in the model. tflite::optimize::ReducedPrecisionSupport support_mask = tflite::optimize::ReducedPrecisionSupport::None;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 13 10:16:19 UTC 2024 - 10.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/passes/xla_broadcast.cc
DenseIntElementsAttr& shape) { Type type = val_bcast.getType(); Type elem_type = getElementTypeOrSelf(type); // Xla's all_reduce legalizer bitcasts to 32 bits, so only // element types size <= 4 bytes are supported. if (elem_type.isBF16() || elem_type.isF16() || elem_type.isTF32() || elem_type.isF32()) { zero = builder.getFloatAttr(elem_type, 0);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 13 18:52:07 UTC 2024 - 13.9K bytes - Viewed (0)