Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 349 for span1 (0.08 sec)

  1. src/runtime/mcentral.go

    	// unswept set and pushes spans that are still in-use on the
    	// swept set. Likewise, allocating an in-use span pushes it
    	// on the swept set.
    	//
    	// Some parts of the sweeper can sweep arbitrary spans, and hence
    	// can't remove them from the unswept set, but will add the span
    	// to the appropriate swept list. As a result, the parts of the
    	// sweeper and mcentral that do consume from the unswept list may
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  2. tests/integration/telemetry/tracing/tracing.go

    		t.Logf("got span %+v, want span %+v", got, want)
    		return false
    	}
    	if len(got.ChildSpans) < len(want.ChildSpans) {
    		t.Logf("got %d child spans from, want %d child spans, maybe trace has not be fully reported",
    			len(got.ChildSpans), len(want.ChildSpans))
    		return false
    	} else if len(got.ChildSpans) > len(want.ChildSpans) {
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Wed May 15 19:05:09 UTC 2024
    - 6.5K bytes
    - Viewed (0)
  3. src/runtime/mheap.go

    	if span.list != list {
    		print("runtime: failed mSpanList.remove span.npages=", span.npages,
    			" span=", span, " prev=", span.prev, " span.list=", span.list, " list=", list, "\n")
    		throw("mSpanList.remove")
    	}
    	if list.first == span {
    		list.first = span.next
    	} else {
    		span.prev.next = span.next
    	}
    	if list.last == span {
    		list.last = span.prev
    	} else {
    		span.next.prev = span.prev
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 22:31:00 UTC 2024
    - 78K bytes
    - Viewed (0)
  4. src/runtime/mgcsweep.go

    //   can free a whole span if none of the objects are marked, but that
    //   isn't its goal. This can be driven either synchronously by
    //   mcentral.cacheSpan for mcentral spans, or asynchronously by
    //   sweepone, which looks at all the mcentral lists.
    //
    // * The span reclaimer looks for spans that contain no marked objects
    //   and frees whole spans. This is a separate algorithm because
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:52:18 UTC 2024
    - 32.9K bytes
    - Viewed (0)
  5. src/runtime/mspanset.go

    	// this block. This number is used to help determine when a block
    	// may be safely recycled.
    	popped atomic.Uint32
    
    	// spans is the set of spans in this block.
    	spans [spanSetBlockEntries]atomicMSpanPointer
    }
    
    // push adds span s to buffer b. push is safe to call concurrently
    // with other push and pop operations.
    func (b *spanSet) push(s *mspan) {
    	// Obtain our slot.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 13.1K bytes
    - Viewed (0)
  6. platforms/core-runtime/logging/src/main/java/org/gradle/internal/logging/sink/ProgressLogEventGenerator.java

            }
    
            private StyledTextOutputEvent styledTextEvent(long timestamp, StyledTextOutputEvent.Span... spans) {
                return new StyledTextOutputEvent(timestamp, category, LogLevel.LIFECYCLE, buildOperationIdentifier, Arrays.asList(spans));
            }
    
            private void doOutput(RenderableOutputEvent event) {
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Tue May 28 13:09:37 UTC 2024
    - 7.7K bytes
    - Viewed (0)
  7. src/runtime/mcache.go

    		c = pp.mcache
    	}
    	return c
    }
    
    // refill acquires a new span of span class spc for c. This span will
    // have at least one free object. The current span in c must be full.
    //
    // Must run in a non-preemptible context since otherwise the owner of
    // c could change.
    func (c *mcache) refill(spc spanClass) {
    	// Return the current cached span to the central lists.
    	s := c.alloc[spc]
    
    	if s.allocCount != s.nelems {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 10K bytes
    - Viewed (0)
  8. src/runtime/traceallocfree.go

    		throw("traceSnapshotMemory: tracing is not enabled")
    	}
    
    	// Write out all the heap spans and heap objects.
    	for _, s := range mheap_.allspans {
    		if s.state.get() == mSpanDead {
    			continue
    		}
    		// It's some kind of span, so trace that it exists.
    		trace.SpanExists(s)
    
    		// Write out allocated objects if it's a heap span.
    		if s.state.get() != mSpanInUse {
    			continue
    		}
    
    		// Find all allocated objects.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 20:32:51 UTC 2024
    - 5.9K bytes
    - Viewed (0)
  9. tests/integration/telemetry/tracing/zipkin/server_tracing_test.go

    	"istio.io/istio/tests/integration/telemetry/tracing"
    )
    
    // TestServerTracing exercises the trace generation features of Istio, based on the Envoy Trace driver for zipkin.
    // The test verifies that all expected spans (a client span and a server span for each service call in the sample bookinfo app)
    // are generated and that they are all a part of the same distributed trace with correct hierarchy and name.
    //
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Mon Apr 08 22:02:59 UTC 2024
    - 2.3K bytes
    - Viewed (0)
  10. src/runtime/mbitmap.go

    	}
    	return tp
    }
    
    // objBase returns the base pointer for the object containing addr in span.
    //
    // Assumes that addr points into a valid part of span (span.base() <= addr < span.limit).
    //
    //go:nosplit
    func (span *mspan) objBase(addr uintptr) uintptr {
    	return span.base() + span.objIndex(addr)*span.elemsize
    }
    
    // bulkBarrierPreWrite executes a write barrier
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 00:18:55 UTC 2024
    - 60K bytes
    - Viewed (0)
Back to top