- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 12 for Boundaries (0.29 sec)
-
tensorflow/compiler/mlir/lite/transforms/legalize_patterns.td
def LegalizeReal : Pat<(TF_RealOp $arg), (TFL_RealOp $arg)>; def LegalizeImag : Pat<(TF_ImagOp $arg), (TFL_ImagOp $arg)>; def LegalizeBucketize : Pat< (TF_BucketizeOp $input, F32ArrayAttr:$boundaries), (TFL_BucketizeOp $input, $boundaries)>; def LegalizeRandomUniform : Pat< (TF_RandomUniformOp $shape, $seed, $seed2), (TFL_RandomUniformOp $shape, (convertIntAttrTo64Bit $seed), (convertIntAttrTo64Bit $seed2))>;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 04 13:30:42 UTC 2024 - 28.5K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/ImmutableSet.java
* is unnecessary; for example, {@code copyOf(copyOf(anArrayList))} will copy the data only once. * This reduces the expense of habitually making defensive copies at API boundaries. However, the * precise conditions for skipping the copy operation are undefined. * * @throws NullPointerException if any of {@code elements} is null * @since 7.0 (source-compatible since 2.0) */
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Sun Jun 02 13:36:19 UTC 2024 - 22.5K bytes - Viewed (0) -
src/cmd/trace/main.go
func splitTrace(parsed *parsedTrace) ([]traceviewer.Range, error) { // TODO(mknyszek): Split traces by generation by doing a quick first pass over the // trace to identify all the generation boundaries. s, c := traceviewer.SplittingTraceConsumer(100 << 20) // 100 MiB if err := generateTrace(parsed, defaultGenOpts(), c); err != nil { return nil, err } return s.Ranges, nil }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 10.4K bytes - Viewed (0) -
internal/ioutil/ioutil.go
const DirectioAlignSize = 4096 // CopyAligned - copies from reader to writer using the aligned input // buffer, it is expected that input buffer is page aligned to // 4K page boundaries. Without passing aligned buffer may cause // this function to return error. // // This code is similar in spirit to io.Copy but it is only to be // used with DIRECT I/O based file descriptor and it is expected that
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Wed May 22 23:07:14 UTC 2024 - 10.2K bytes - Viewed (0) -
pilot/pkg/serviceregistry/serviceentry/controller.go
for _, i := range instances { // Insert all instances for this IP for services within the same namespace. This ensures we // match Kubernetes logic where Services do not cross namespace boundaries and avoids // possibility of other namespaces inserting service instances into namespaces they do not // control. if node.Metadata.Namespace == "" || i.Service.Attributes.Namespace == node.Metadata.Namespace {
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Wed May 29 15:31:09 UTC 2024 - 36.8K bytes - Viewed (0) -
src/runtime/string.go
// actual system page size is larger than this value. const pageSize = 4096 offset := 0 ptr := unsafe.Pointer(s) // IndexByteString uses wide reads, so we need to be careful // with page boundaries. Call IndexByteString on // [ptr, endOfPage) interval. safeLen := int(pageSize - uintptr(ptr)%pageSize) for { t := *(*string)(unsafe.Pointer(&stringStruct{ptr, safeLen})) // Check one page at a time.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 13.4K bytes - Viewed (0) -
src/runtime/race.go
throw("raceinit: race build must use cgo") } racecall(&__tsan_init, uintptr(unsafe.Pointer(&gctx)), uintptr(unsafe.Pointer(&pctx)), abi.FuncPCABI0(racecallbackthunk), 0) // Round data segment to page boundaries, because it's used in mmap(). start := ^uintptr(0) end := uintptr(0) if start > firstmoduledata.noptrdata { start = firstmoduledata.noptrdata } if start > firstmoduledata.data {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 20.4K bytes - Viewed (0) -
src/runtime/export_test.go
func (s *ScavengeIndex) AllocRange(base, limit uintptr) { sc, ec := chunkIndex(base), chunkIndex(limit-1) si, ei := chunkPageIndex(base), chunkPageIndex(limit-1) if sc == ec { // The range doesn't cross any chunk boundaries. s.i.alloc(sc, ei+1-si) } else { // The range crosses at least one chunk boundary. s.i.alloc(sc, pallocChunkPages-si) for c := sc + 1; c < ec; c++ { s.i.alloc(c, pallocChunkPages) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:50:53 UTC 2024 - 46.1K bytes - Viewed (0) -
src/internal/trace/gc.go
// box-filtered utilization function, which we call the // "windowed mutator utilization function". The resulting // function is continuous and piecewise linear (unless // window==0, which we handle elsewhere), where the boundaries // between segments occur when either edge of the window // encounters a change in the instantaneous mutator // utilization function. Hence, the minimum of this function
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 26K bytes - Viewed (0) -
src/time/format_test.go
"(2006)(002)(01) (15):(04):(05)", "(2006)(002)(04) (15):(04):(05)", } func TestNextStdChunk(t *testing.T) { // Most bugs in Parse or Format boil down to problems with // the exact detection of format chunk boundaries in the // helper function nextStdChunk (here called as NextStdChunk). // This test checks nextStdChunk's behavior directly, // instead of needing to test it only indirectly through Parse/Format.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:58:29 UTC 2024 - 36.4K bytes - Viewed (0)