- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 72 for flushed (0.43 sec)
-
src/runtime/tracestring.go
s = s[:maxTraceStringLen] } lock(&t.lock) w := unsafeTraceWriter(gen, t.buf) // Ensure we have a place to write to. var flushed bool w, flushed = w.ensure(2 + 2*traceBytesPerNumber + len(s) /* traceEvStrings + traceEvString + ID + len + string data */) if flushed { // Annotate the batch as containing strings. w.byte(byte(traceEvStrings)) } // Write out the string. w.byte(byte(traceEvString))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 17:03:35 UTC 2024 - 2.4K bytes - Viewed (0) -
src/runtime/tracetype.go
// bound is pretty loose, but avoids counting // lots of varint sizes. // // Add 1 because we might also write a traceAllocFreeTypesBatch byte. var flushed bool w, flushed = w.ensure(1 + maxBytes) if flushed { // Annotate the batch as containing types. w.byte(byte(traceAllocFreeTypesBatch)) } // Emit type. w.varint(uint64(node.id)) w.varint(uint64(uintptr(unsafe.Pointer(typ))))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:31:27 UTC 2024 - 2.3K bytes - Viewed (0) -
src/runtime/traceallocfree.go
func traceSnapshotMemory(gen uintptr) { assertWorldStopped() // Write a batch containing information that'll be necessary to // interpret the events. var flushed bool w := unsafeTraceExpWriter(gen, nil, traceExperimentAllocFree) w, flushed = w.ensure(1 + 4*traceBytesPerNumber) if flushed { // Annotate the batch as containing additional info. w.byte(byte(traceAllocFreeInfoBatch)) } // Emit info.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:32:51 UTC 2024 - 5.9K bytes - Viewed (0) -
src/runtime/tracecpu.go
} // Write out a trace event. w := unsafeTraceWriter(gen, trace.cpuBuf[gen%2]) // Ensure we have a place to write to. var flushed bool w, flushed = w.ensure(2 + 5*traceBytesPerNumber /* traceEvCPUSamples + traceEvCPUSample + timestamp + g + m + p + stack ID */) if flushed { // Annotate the batch as containing strings. w.byte(byte(traceEvCPUSamples)) } // Add the stack to the table.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 17:03:35 UTC 2024 - 8.7K bytes - Viewed (0) -
src/runtime/traceexp.go
return traceExpWriter{traceWriter{traceLocker: traceLocker{gen: gen}, traceBuf: buf}, exp} } // ensure makes sure that at least maxSize bytes are available to write. // // Returns whether the buffer was flushed. func (w traceExpWriter) ensure(maxSize int) (traceExpWriter, bool) { refill := w.traceBuf == nil || !w.available(maxSize) if refill { w.traceWriter = w.traceWriter.refill(w.exp) } return w, refill }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:47:01 UTC 2024 - 2.4K bytes - Viewed (0) -
src/compress/gzip/gzip.go
n, z.err = z.compressor.Write(p) return n, z.err } // Flush flushes any pending compressed data to the underlying writer. // // It is useful mainly in compressed network protocols, to ensure that // a remote reader has enough data to reconstruct a packet. Flush does // not return until the data has been written. If the underlying // writer returns an error, Flush returns that error. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 26 13:32:40 UTC 2024 - 6.2K bytes - Viewed (0) -
src/runtime/tracestack.go
// Estimate the size of this record. This // bound is pretty loose, but avoids counting // lots of varint sizes. // // Add 1 because we might also write traceEvStacks. var flushed bool w, flushed = w.ensure(1 + maxBytes) if flushed { w.byte(byte(traceEvStacks)) } // Emit stack event. w.byte(byte(traceEvStack)) w.varint(uint64(node.id)) w.varint(uint64(len(frames)))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 21 14:38:56 UTC 2024 - 11K bytes - Viewed (0) -
src/runtime/trace.go
// will be flushed in time for the new generation) or it will have flushed its // buffers before we snapshotted it to begin with. lock(&sched.lock) mToFlush := allm for mp := mToFlush; mp != nil; mp = mp.alllink { mp.trace.link = mp.alllink } for mp := sched.freem; mp != nil; mp = mp.freelink { mp.trace.link = mToFlush mToFlush = mp } unlock(&sched.lock)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:17:41 UTC 2024 - 37.1K bytes - Viewed (0) -
platforms/core-runtime/launcher/src/integTest/groovy/org/gradle/launcher/daemon/AbstractDaemonLifecycleSpec.groovy
} void doDaemonContext(gradleHandle, Closure assertions) { // poll here since even though the daemon has been marked as busy in the registry, the context may not have been // flushed to the log yet. DaemonContext context poll(5) { context = DaemonContextParser.parseFromString(gradleHandle.standardOutput, GradleVersion.current()) } context.with(assertions)
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Apr 11 13:32:55 UTC 2024 - 7.1K bytes - Viewed (0) -
internal/logger/logrotate.go
// Write is blocking otherwise. func (w *Writer) Write(p []byte) (n int, err error) { return w.pw.Write(p) } // Close closes the writer. // Any accepted writes will be flushed. Any new writes will be rejected. // Once Close() exits, files are synchronized to disk. func (w *Writer) Close() error { w.pw.CloseWithError(nil) if w.f != nil { if err := w.closeCurrentFile(); err != nil {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 5.8K bytes - Viewed (0)