- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 23 for lineFor (0.38 sec)
-
src/cmd/compile/internal/ssa/regalloc.go
// Copyright 2015 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. // Register allocation. // // We use a version of a linear scan register allocator. We treat the // whole function as a single long basic block and run through // it using a greedy register allocator. Then all merge edges // (those targeting a block with len(Preds)>1) are processed to
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 21 17:49:56 UTC 2023 - 87.2K bytes - Viewed (0) -
src/cmd/internal/testdir/testdir_test.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Mar 21 20:08:06 UTC 2024 - 57.5K bytes - Viewed (0) -
src/runtime/mgc.go
// (100 by default). If GOGC=100 and we're using 4M, we'll GC again when we get to 8M // (this mark is computed by the gcController.heapGoal method). This keeps the GC cost in // linear proportion to the allocation cost. Adjusting GOGC just changes the linear constant // (and also the amount of extra memory used). // Oblets // // In order to prevent long pauses while scanning large objects and to
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 62K bytes - Viewed (0) -
src/runtime/malloc.go
// iOS versions. // TODO(mknyszek): Once iOS <14 is deprecated, promote ios/arm64 // to a 48-bit address space like every other arm64 platform. // // WebAssembly currently has a limit of 4GB linear memory. heapAddrBits = (_64bit*(1-goarch.IsWasm)*(1-goos.IsIos*goarch.IsArm64))*48 + (1-_64bit+goarch.IsWasm)*(32-(goarch.IsMips+goarch.IsMipsle)) + 40*goos.IsIos*goarch.IsArm64
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/Iterators.java
if (topMetaIterator == null) { return false; } iterator = topMetaIterator.next(); if (iterator instanceof ConcatenatedIterator) { // Instead of taking linear time in the number of nested concatenations, unpack // them into the queue @SuppressWarnings("unchecked") ConcatenatedIterator<T> topConcat = (ConcatenatedIterator<T>) iterator;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri May 03 14:46:32 UTC 2024 - 50.2K bytes - Viewed (0) -
guava/src/com/google/common/collect/Iterators.java
if (topMetaIterator == null) { return false; } iterator = topMetaIterator.next(); if (iterator instanceof ConcatenatedIterator) { // Instead of taking linear time in the number of nested concatenations, unpack // them into the queue @SuppressWarnings("unchecked") ConcatenatedIterator<T> topConcat = (ConcatenatedIterator<T>) iterator;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri May 03 14:46:32 UTC 2024 - 50.2K bytes - Viewed (0) -
src/net/http/request.go
} } // patIndex returns the index of name in the list of named wildcards of the // request's pattern, or -1 if there is no such name. func (r *Request) patIndex(name string) int { // The linear search seems expensive compared to a map, but just creating the map // takes a lot of time, and most patterns will just have a couple of wildcards. if r.pat == nil { return -1 } i := 0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 49.4K bytes - Viewed (0) -
src/runtime/mgcscavenge.go
chunks []atomicScavChunkData min, max atomic.Uintptr minHeapIdx atomic.Uintptr // searchAddr* is the maximum address (in the offset address space, so we have a linear // view of the address space; see mranges.go:offAddr) containing memory available to // scavenge. It is a hint to the find operation to avoid O(n^2) behavior in repeated lookups. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:48:45 UTC 2024 - 52.3K bytes - Viewed (0) -
cmd/xl-storage-format-v2.go
} // Add space at the end. // Will have -1 modtime, so it will be inserted there. x.versions = append(x.versions, xlMetaV2ShallowVersion{header: xlMetaV2VersionHeader{ModTime: -1}}) // Linear search, we likely have to insert at front. for i, existing := range x.versions { if existing.header.ModTime <= modTime { // Insert at current idx. First move current back. copy(x.versions[i+1:], x.versions[i:])
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Wed May 29 19:14:09 UTC 2024 - 63.6K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass.cc
}).status()); // Check that the conclusion made above (that iterating over the graph once in // post order gives a maximal clustering) holds. Once the linear time // post-order scheme has been battle tested we can move this to happen only in // debug builds. VLOG(2) << "Checking idempotence"; TF_ASSIGN_OR_RETURN(bool changed,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 12:19:41 UTC 2024 - 85.3K bytes - Viewed (0)