- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 397 for nalloc (0.1 sec)
-
test/chan/select2.go
runtime.ReadMemStats(memstats) alloc := memstats.Alloc // second time shouldn't increase footprint by much go sender(c, 100000) receiver(c, dummy, 100000) runtime.GC() runtime.ReadMemStats(memstats) // Be careful to avoid wraparound. if memstats.Alloc > alloc && memstats.Alloc-alloc > 1.1e5 { println("BUG: too much memory for 100,000 selects:", memstats.Alloc-alloc) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 02 13:43:18 UTC 2016 - 1K bytes - Viewed (0) -
src/runtime/cgo/gcc_linux_arm64.c
{ uintptr *pbounds; /* The memory sanitizer distributed with versions of clang before 3.8 has a bug: if you call mmap before malloc, mmap may return an address that is later overwritten by the msan library. Avoid this problem by forcing a call to malloc here, before we ever call malloc. This is only required for the memory sanitizer, so it's unfortunate that we always run it. It should be possible
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 22:06:46 UTC 2023 - 2.2K bytes - Viewed (0) -
test/escape4.go
package foo var p *int func alloc(x int) *int { // ERROR "can inline alloc" "moved to heap: x" return &x } var f func() func f1() { p = alloc(2) // ERROR "inlining call to alloc" "moved to heap: x" // Escape analysis used to miss inlined code in closures. func() { // ERROR "can inline f1.func1" p = alloc(3) // ERROR "inlining call to alloc"
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Sep 14 19:43:26 UTC 2023 - 1.4K bytes - Viewed (0) -
src/runtime/mcentral.go
// Copyright 2009 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. // Central free lists. // // See malloc.go for an overview. // // The mcentral doesn't actually contain the list of free objects; the mspan does. // Each mcentral is two lists of mspans: those with free objects (c->nonempty) // and those that are completely allocated (c->empty). package runtime
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0) -
src/runtime/pprof/pprof.go
// Pprof will ignore, but useful for people s := memStats fmt.Fprintf(w, "\n# runtime.MemStats\n") fmt.Fprintf(w, "# Alloc = %d\n", s.Alloc) fmt.Fprintf(w, "# TotalAlloc = %d\n", s.TotalAlloc) fmt.Fprintf(w, "# Sys = %d\n", s.Sys) fmt.Fprintf(w, "# Lookups = %d\n", s.Lookups) fmt.Fprintf(w, "# Mallocs = %d\n", s.Mallocs) fmt.Fprintf(w, "# Frees = %d\n", s.Frees) fmt.Fprintf(w, "# HeapAlloc = %d\n", s.HeapAlloc)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:52:17 UTC 2024 - 30.6K bytes - Viewed (0) -
pkg/registry/core/service/allocator/storage/storage.go
var once sync.Once return &Etcd{ alloc: alloc, storage: storage, baseKey: baseKey, resource: config.GroupResource, destroyFn: func() { once.Do(d) }, }, nil } // Allocate attempts to allocate the item. func (e *Etcd) Allocate(offset int) (bool, error) { e.lock.Lock() defer e.lock.Unlock() err := e.tryUpdate(func() error { ok, err := e.alloc.Allocate(offset) if err != nil {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Aug 24 09:23:05 UTC 2023 - 6.5K bytes - Viewed (0) -
src/cmd/cgo/internal/testsanitizers/testdata/tsan4.go
// Copyright 2016 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. package main // Check that calls to C.malloc/C.free do not trigger TSAN false // positive reports. // #cgo CFLAGS: -fsanitize=thread // #cgo LDFLAGS: -fsanitize=thread // #include <stdlib.h> import "C" import ( "runtime" "sync" ) func main() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 12 11:59:56 UTC 2023 - 634 bytes - Viewed (0) -
src/runtime/mprof.go
// The following complex 3-stage scheme of stats accumulation // is required to obtain a consistent picture of mallocs and frees // for some point in time. // The problem is that mallocs come in real time, while frees // come only after a GC during concurrent sweeping. So if we would // naively count them, we would get a skew toward mallocs. // // Hence, we delay information to get consistent snapshots as
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 53.3K bytes - Viewed (0) -
src/main/java/jcifs/smb1/smb1/Trans2QueryFSInformationResponse.java
static final int SMB_FS_FULL_SIZE_INFORMATION = 1007; class SmbInfoAllocation implements AllocInfo { long alloc; // Also handles SmbQueryFSSizeInfo long free; int sectPerAlloc; int bytesPerSect; public long getCapacity() { return alloc * sectPerAlloc * bytesPerSect; } public long getFree() { return free * sectPerAlloc * bytesPerSect;
Registered: Wed Jun 12 15:45:55 UTC 2024 - Last Modified: Fri Mar 22 21:10:40 UTC 2019 - 5.1K bytes - Viewed (0) -
src/runtime/arena.go
} // Prevent preemption as we set up the space for a new object. // // Act like we're allocating. mp := acquirem() if mp.mallocing != 0 { throw("malloc deadlock") } if mp.gsignal == getg() { throw("malloc during signal") } mp.mallocing = 1 var ptr unsafe.Pointer if !typ.Pointers() { // Allocate pointer-less objects from the tail end of the chunk.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:44:56 UTC 2024 - 37.9K bytes - Viewed (0)