Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 40 of 70 for sched (0.41 sec)

  1. src/runtime/asm_amd64.s

    	LEAQ	fn+0(FP), AX
    	SUBQ	SP, AX
    
    	// Restore g->sched (== m->curg->sched) from saved values.
    	get_tls(CX)
    	MOVQ	g(CX), SI
    	MOVQ	SP, DI
    	ADDQ	AX, DI
    	MOVQ	-8(DI), BX
    	MOVQ	BX, (g_sched+gobuf_pc)(SI)
    	MOVQ	DI, (g_sched+gobuf_sp)(SI)
    
    	// Switch back to m->g0's stack and restore m->g0->sched.sp.
    	// (Unlike m->curg, the g0 goroutine never uses sched.pc,
    	// so we do not have to restore it.)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Sat May 11 20:38:24 UTC 2024
    - 60.4K bytes
    - Viewed (0)
  2. src/runtime/align_runtime_test.go

    	unsafe.Pointer(&ncgocall),
    	unsafe.Pointer(&test_z64),
    	unsafe.Pointer(&blockprofilerate),
    	unsafe.Pointer(&mutexprofilerate),
    	unsafe.Pointer(&gcController),
    	unsafe.Pointer(&memstats),
    	unsafe.Pointer(&sched),
    	unsafe.Pointer(&ticks),
    	unsafe.Pointer(&work),
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 1.8K bytes
    - Viewed (0)
  3. pkg/scheduler/eventhandlers_test.go

    			defer cancel()
    			sched := &Scheduler{
    				Cache:           cache.New(ctx, ttl),
    				SchedulingQueue: queue.NewTestQueue(ctx, nil),
    				logger:          logger,
    			}
    			sched.addPodToCache(tt.oldObj)
    			sched.updatePodInCache(tt.oldObj, tt.newObj)
    
    			if tt.oldObj.(*v1.Pod).UID != tt.newObj.(*v1.Pod).UID {
    				if pod, err := sched.Cache.GetPod(tt.oldObj.(*v1.Pod)); err == nil {
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Jan 10 14:38:54 UTC 2024
    - 21.5K bytes
    - Viewed (0)
  4. src/cmd/trace/main.go

    Go 1.7 does not require the binary argument.
    
    Supported profile types are:
        - net: network blocking profile
        - sync: synchronization blocking profile
        - syscall: syscall blocking profile
        - sched: scheduler latency profile
    
    Flags:
    	-http=addr: HTTP service address (e.g., ':6060')
    	-pprof=type: print a pprof-like profile instead
    	-d=int: print debug info such as parsed events (1 for high-level, 2 for low-level)
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 17 18:48:18 UTC 2024
    - 10.4K bytes
    - Viewed (0)
  5. src/runtime/trace.go

    	// buffers before we snapshotted it to begin with.
    	lock(&sched.lock)
    	mToFlush := allm
    	for mp := mToFlush; mp != nil; mp = mp.alllink {
    		mp.trace.link = mp.alllink
    	}
    	for mp := sched.freem; mp != nil; mp = mp.freelink {
    		mp.trace.link = mToFlush
    		mToFlush = mp
    	}
    	unlock(&sched.lock)
    
    	// Iterate over our snapshot, flushing every buffer until we're done.
    	//
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 21:17:41 UTC 2024
    - 37.1K bytes
    - Viewed (0)
  6. src/runtime/mgcmark.go

    // cached stacks around isn't a problem.
    func markrootFreeGStacks() {
    	// Take list of dead Gs with stacks.
    	lock(&sched.gFree.lock)
    	list := sched.gFree.stack
    	sched.gFree.stack = gList{}
    	unlock(&sched.gFree.lock)
    	if list.empty() {
    		return
    	}
    
    	// Free stacks.
    	q := gQueue{list.head, list.head}
    	for gp := list.head.ptr(); gp != nil; gp = gp.schedlink.ptr() {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Apr 18 21:25:11 UTC 2024
    - 52.5K bytes
    - Viewed (0)
  7. src/runtime/debugcall.go

    			// transition anymore.
    			trace.GoSched()
    		}
    		casgstatus(gp, _Grunning, _Grunnable)
    		if trace.ok() {
    			traceRelease(trace)
    		}
    		dropg()
    		lock(&sched.lock)
    		globrunqput(gp)
    		unlock(&sched.lock)
    
    		trace = traceAcquire()
    		casgstatus(callingG, _Gwaiting, _Grunnable)
    		if trace.ok() {
    			trace.GoUnpark(callingG, 0)
    			traceRelease(trace)
    		}
    		execute(callingG, true)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Apr 05 20:50:21 UTC 2024
    - 7.1K bytes
    - Viewed (0)
  8. src/runtime/runtime2.go

    	_defer    *_defer // innermost defer
    	m         *m      // current m; offset known to arm liblink
    	sched     gobuf
    	syscallsp uintptr // if status==Gsyscall, syscallsp = sched.sp to use during gc
    	syscallpc uintptr // if status==Gsyscall, syscallpc = sched.pc to use during gc
    	syscallbp uintptr // if status==Gsyscall, syscallbp = sched.bp to use in fpTraceback
    	stktopsp  uintptr // expected sp at top of stack, to check in traceback
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:57:37 UTC 2024
    - 47.9K bytes
    - Viewed (0)
  9. src/runtime/mgc.go

    	lock(&sched.sudoglock)
    	var sg, sgnext *sudog
    	for sg = sched.sudogcache; sg != nil; sg = sgnext {
    		sgnext = sg.next
    		sg.next = nil
    	}
    	sched.sudogcache = nil
    	unlock(&sched.sudoglock)
    
    	// Clear central defer pool.
    	// Leave per-P pools alone, they have strictly bounded size.
    	lock(&sched.deferlock)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 62K bytes
    - Viewed (0)
  10. pkg/scheduler/extender_test.go

    				t.Fatal(err)
    			}
    
    			sched := &Scheduler{
    				Cache:                    cache,
    				nodeInfoSnapshot:         emptySnapshot,
    				percentageOfNodesToScore: schedulerapi.DefaultPercentageOfNodesToScore,
    				Extenders:                extenders,
    				logger:                   logger,
    			}
    			sched.applyDefaultHandlers()
    
    			podIgnored := &v1.Pod{}
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Mon Feb 26 19:07:19 UTC 2024
    - 16.7K bytes
    - Viewed (0)
Back to top