- Sort Score
- Result 10 results
- Languages All
Results 81 - 90 of 363 for Malloc (0.14 sec)
-
src/net/rpc/server_test.go
func TestCountMallocs(t *testing.T) { if testing.Short() { t.Skip("skipping malloc count in short mode") } if runtime.GOMAXPROCS(0) > 1 { t.Skip("skipping; GOMAXPROCS>1") } fmt.Printf("mallocs per rpc round trip: %v\n", countMallocs(dialDirect, t)) } func TestCountMallocsOverHTTP(t *testing.T) { if testing.Short() { t.Skip("skipping malloc count in short mode") } if runtime.GOMAXPROCS(0) > 1 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jul 18 05:23:29 UTC 2023 - 19K bytes - Viewed (0) -
src/crypto/internal/boring/rsa.go
} if C._goboringcrypto_EVP_PKEY_CTX_set_rsa_mgf1_md(ctx, mgfMD) == 0 { return pkey, ctx, fail("EVP_PKEY_set_rsa_mgf1_md") } // ctx takes ownership of label, so malloc a copy for BoringCrypto to free. clabel := (*C.uint8_t)(C._goboringcrypto_OPENSSL_malloc(C.size_t(len(label)))) if clabel == nil { return pkey, ctx, fail("OPENSSL_malloc") }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Mar 26 23:38:03 UTC 2024 - 12K bytes - Viewed (0) -
src/runtime/iface.go
// Crash reliably, rather than only when we need to grow // the hash table. if getg().m.mallocing != 0 { throw("malloc deadlock") } t := itabTable if t.count >= 3*(t.size/4) { // 75% load factor // Grow hash table. // t2 = new(itabTableType) + some additional entries // We lie and tell malloc we want pointer-free memory because // all the pointed-to values are not in the heap.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 22.5K bytes - Viewed (0) -
src/cmd/vendor/github.com/google/pprof/profile/legacy_profile.go
return true } return false } var allocRxStr = strings.Join([]string{ // POSIX entry points. `calloc`, `cfree`, `malloc`, `free`, `memalign`, `do_memalign`, `(__)?posix_memalign`, `pvalloc`, `valloc`, `realloc`, // TC malloc. `tcmalloc::.*`, `tc_calloc`, `tc_cfree`, `tc_malloc`, `tc_free`, `tc_memalign`,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 22 18:58:12 UTC 2022 - 32.8K bytes - Viewed (0) -
src/runtime/race.go
} //go:nosplit func racewriterangepc(addr unsafe.Pointer, sz, callpc, pc uintptr) { gp := getg() if gp != gp.m.curg { // The call is coming from manual instrumentation of Go code running on g0/gsignal. // Not interesting. return } if callpc != 0 { racefuncenter(callpc) } racewriterangepc1(uintptr(addr), sz, pc) if callpc != 0 { racefuncexit() } } //go:nosplit
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 20.4K bytes - Viewed (0) -
src/runtime/mstats.go
// Size is the maximum byte size of an object in this // size class. Size uint32 // Mallocs is the cumulative count of heap objects // allocated in this size class. The cumulative bytes // of allocation is Size*Mallocs. The number of live // objects in this size class is Mallocs - Frees. Mallocs uint64 // Frees is the cumulative count of heap objects freed // in this size class. Frees uint64
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 08 21:03:13 UTC 2024 - 34.2K bytes - Viewed (0) -
src/cmd/compile/internal/types2/call.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 19:19:55 UTC 2024 - 31.5K bytes - Viewed (0) -
src/syscall/exec_plan9.go
// // In the child, this function must not acquire any locks, because // they might have been locked at the time of the fork. This means // no rescheduling, no malloc calls, and no new stack segments. // The calls to RawSyscall are okay because they are assembly // functions that do not grow the stack. // //go:norace
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Feb 26 21:03:59 UTC 2024 - 13.3K bytes - Viewed (0) -
src/go/types/call.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 19:19:55 UTC 2024 - 33.5K bytes - Viewed (0) -
src/runtime/mheap.go
func (s spanAllocType) manual() bool { return s != spanAllocHeap } // alloc allocates a new span of npage pages from the GC'd heap. // // spanclass indicates the span's size class and scannability. // // Returns a span that has been fully initialized. span.needzero indicates // whether the span has been zeroed. Note that it may not be. func (h *mheap) alloc(npages uintptr, spanclass spanClass) *mspan {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 78K bytes - Viewed (0)