- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 115 for flushed (0.15 sec)
-
src/runtime/tracestring.go
s = s[:maxTraceStringLen] } lock(&t.lock) w := unsafeTraceWriter(gen, t.buf) // Ensure we have a place to write to. var flushed bool w, flushed = w.ensure(2 + 2*traceBytesPerNumber + len(s) /* traceEvStrings + traceEvString + ID + len + string data */) if flushed { // Annotate the batch as containing strings. w.byte(byte(traceEvStrings)) } // Write out the string. w.byte(byte(traceEvString))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 17:03:35 UTC 2024 - 2.4K bytes - Viewed (0) -
src/runtime/mgcwork.go
bytesMarked uint64 // Heap scan work performed on this gcWork. This is aggregated into // gcController by dispose and may also be flushed by callers. // Other types of scan work are flushed immediately. heapScanWork int64 // flushedWork indicates that a non-empty work buffer was // flushed to the global work list since the last gcMarkDone // termination check. Specifically, this indicates that this
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 12.9K bytes - Viewed (0) -
src/runtime/tracetype.go
// bound is pretty loose, but avoids counting // lots of varint sizes. // // Add 1 because we might also write a traceAllocFreeTypesBatch byte. var flushed bool w, flushed = w.ensure(1 + maxBytes) if flushed { // Annotate the batch as containing types. w.byte(byte(traceAllocFreeTypesBatch)) } // Emit type. w.varint(uint64(node.id)) w.varint(uint64(uintptr(unsafe.Pointer(typ))))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:31:27 UTC 2024 - 2.3K bytes - Viewed (0) -
src/net/http/httptest/recorder.go
if rw.HeaderMap == nil { rw.HeaderMap = make(http.Header) } rw.snapHeader = rw.HeaderMap.Clone() } // Flush implements [http.Flusher]. To test whether Flush was // called, see rw.Flushed. func (rw *ResponseRecorder) Flush() { if !rw.wroteHeader { rw.WriteHeader(200) } rw.Flushed = true } // Result returns the response generated by the handler. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jan 10 03:29:50 UTC 2024 - 7K bytes - Viewed (0) -
src/runtime/traceallocfree.go
func traceSnapshotMemory(gen uintptr) { assertWorldStopped() // Write a batch containing information that'll be necessary to // interpret the events. var flushed bool w := unsafeTraceExpWriter(gen, nil, traceExperimentAllocFree) w, flushed = w.ensure(1 + 4*traceBytesPerNumber) if flushed { // Annotate the batch as containing additional info. w.byte(byte(traceAllocFreeInfoBatch)) } // Emit info.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:32:51 UTC 2024 - 5.9K bytes - Viewed (0) -
src/runtime/tracecpu.go
} // Write out a trace event. w := unsafeTraceWriter(gen, trace.cpuBuf[gen%2]) // Ensure we have a place to write to. var flushed bool w, flushed = w.ensure(2 + 5*traceBytesPerNumber /* traceEvCPUSamples + traceEvCPUSample + timestamp + g + m + p + stack ID */) if flushed { // Annotate the batch as containing strings. w.byte(byte(traceEvCPUSamples)) } // Add the stack to the table.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 17:03:35 UTC 2024 - 8.7K bytes - Viewed (0) -
pkg/proxy/conntrack/conntrack.go
err := ct.exec(parameters...) if err != nil && !strings.Contains(err.Error(), noConnectionToDelete) { // TODO: Better handling for deletion failure. When failure occur, stale udp connection may not get flushed. // These stale udp connection will keep black hole traffic. Making this a best effort operation for now, since it // is expensive to baby-sit all udp connections to kubernetes services.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Jan 15 18:09:05 UTC 2024 - 5.5K bytes - Viewed (0) -
src/runtime/traceexp.go
return traceExpWriter{traceWriter{traceLocker: traceLocker{gen: gen}, traceBuf: buf}, exp} } // ensure makes sure that at least maxSize bytes are available to write. // // Returns whether the buffer was flushed. func (w traceExpWriter) ensure(maxSize int) (traceExpWriter, bool) { refill := w.traceBuf == nil || !w.available(maxSize) if refill { w.traceWriter = w.traceWriter.refill(w.exp) } return w, refill }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:47:01 UTC 2024 - 2.4K bytes - Viewed (0) -
src/compress/gzip/gzip.go
n, z.err = z.compressor.Write(p) return n, z.err } // Flush flushes any pending compressed data to the underlying writer. // // It is useful mainly in compressed network protocols, to ensure that // a remote reader has enough data to reconstruct a packet. Flush does // not return until the data has been written. If the underlying // writer returns an error, Flush returns that error. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 26 13:32:40 UTC 2024 - 6.2K bytes - Viewed (0) -
src/runtime/tracestack.go
// Estimate the size of this record. This // bound is pretty loose, but avoids counting // lots of varint sizes. // // Add 1 because we might also write traceEvStacks. var flushed bool w, flushed = w.ensure(1 + maxBytes) if flushed { w.byte(byte(traceEvStacks)) } // Emit stack event. w.byte(byte(traceEvStack)) w.varint(uint64(node.id)) w.varint(uint64(len(frames)))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 21 14:38:56 UTC 2024 - 11K bytes - Viewed (0)