- Sort Score
- Result 10 results
- Languages All
Results 11 - 17 of 17 for newElems (0.13 sec)
-
src/runtime/traceallocfree.go
// Write out allocated objects if it's a heap span. if s.state.get() != mSpanInUse { continue } // Find all allocated objects. abits := s.allocBitsForIndex(0) for i := uintptr(0); i < uintptr(s.nelems); i++ { if abits.index < uintptr(s.freeindex) || abits.isMarked() { x := s.base() + i*s.elemsize trace.HeapObjectExists(x, s.typePointersOfUnchecked(x).typ) } abits.advance() } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:32:51 UTC 2024 - 5.9K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/registry/generic/registry/store.go
if err != nil { return nil, err } newItems, err := meta.ExtractList(listObj) if err != nil { return nil, err } items = append(items, newItems...) for i := 0; i < len(newItems); i++ { select { case toProcess <- newItems[i]: case <-workersExited:
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Jan 19 23:22:44 UTC 2024 - 60.8K bytes - Viewed (0) -
pkg/apis/resource/validation/validation.go
oldSet := sets.New(oldClaim.Status.ReservedFor...) newSet := sets.New(resourceClaim.Status.ReservedFor...) newItems := newSet.Difference(oldSet) if len(newItems) > 0 { allErrs = append(allErrs, field.Forbidden(fldPath.Child("reservedFor"), "new entries may not be added while `deallocationRequested` or `deletionTimestamp` are set")) } } }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Jun 07 09:18:10 UTC 2024 - 28.3K bytes - Viewed (0) -
src/runtime/heapdump.go
size := s.elemsize n := (s.npages << _PageShift) / size if n > uintptr(len(freemark)) { throw("freemark array doesn't have enough entries") } for freeIndex := uint16(0); freeIndex < s.nelems; freeIndex++ { if s.isFree(uintptr(freeIndex)) { freemark[freeIndex] = true } } for j := uintptr(0); j < n; j, p = j+1, p+size { if freemark[j] { freemark[j] = false continue
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 09 04:07:57 UTC 2024 - 17.6K bytes - Viewed (0) -
src/runtime/mbitmap.go
// faster if profiling warrants it. func (s *mspan) nextFreeIndex() uint16 { sfreeindex := s.freeindex snelems := s.nelems if sfreeindex == snelems { return sfreeindex } if sfreeindex > snelems { throw("s.freeindex > s.nelems") } aCache := s.allocCache bitIndex := sys.TrailingZeros64(aCache) for bitIndex == 64 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 60K bytes - Viewed (0) -
tensorflow/c/c_api.cc
const tensorflow::TensorShape& shape) { static char empty; int64_t nelems = 1; std::vector<int64_t> dims; dims.reserve(shape.dims()); for (int i = 0; i < shape.dims(); ++i) { dims.push_back(shape.dim_size(i)); nelems *= shape.dim_size(i); } CHECK_EQ(nelems, 0); return TF_NewTensor( dtype, reinterpret_cast<const int64_t*>(dims.data()), shape.dims(),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 03:35:10 UTC 2024 - 102.3K bytes - Viewed (0) -
src/runtime/export_test.go
systemstack(func() { lock(&mheap_.lock) mheap_.spanalloc.free(unsafe.Pointer(s)) unlock(&mheap_.lock) }) } func MSpanCountAlloc(ms *MSpan, bits []byte) int { s := (*mspan)(ms) s.nelems = uint16(len(bits) * 8) s.gcmarkBits = (*gcBits)(unsafe.Pointer(&bits[0])) result := s.countAlloc() s.gcmarkBits = nil return result } const ( TimeHistSubBucketBits = timeHistSubBucketBits
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:50:53 UTC 2024 - 46.1K bytes - Viewed (0)