Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 824 for getZ (0.23 sec)

  1. src/runtime/print.go

    func printlock() {
    	mp := getg().m
    	mp.locks++ // do not reschedule between printlock++ and lock(&debuglock).
    	mp.printlock++
    	if mp.printlock == 1 {
    		lock(&debuglock)
    	}
    	mp.locks-- // now we know debuglock is held and holding up mp.locks for us.
    }
    
    func printunlock() {
    	mp := getg().m
    	mp.printlock--
    	if mp.printlock == 0 {
    		unlock(&debuglock)
    	}
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Jan 20 03:27:26 UTC 2023
    - 5.9K bytes
    - Viewed (0)
  2. src/runtime/export_test.go

    	if !ok {
    		// We were unable to get the requested reservation.
    		// Release what we did get and fail.
    		sysFreeOS(got, physPageSize)
    	}
    	return
    }
    
    func GetNextArenaHint() uintptr {
    	return mheap_.arenaHints.addr
    }
    
    type G = g
    
    type Sudog = sudog
    
    func Getg() *G {
    	return getg()
    }
    
    func Goid() uint64 {
    	return getg().goid
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:50:53 UTC 2024
    - 46.1K bytes
    - Viewed (0)
  3. src/cmd/compile/internal/test/inl_test.go

    			"(*mspan).base",
    			"(*mspan).markBitsForBase",
    			"(*mspan).markBitsForIndex",
    			"(*mspan).writeUserArenaHeapBits",
    			"(*muintptr).set",
    			"(*puintptr).set",
    			"(*wbBuf).get1",
    			"(*wbBuf).get2",
    
    			// Trace-related ones.
    			"traceLocker.ok",
    			"traceEnabled",
    		},
    		"runtime/internal/sys": {},
    		"runtime/internal/math": {
    			"MulUintptr",
    		},
    		"bytes": {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Apr 09 04:07:57 UTC 2024
    - 10.7K bytes
    - Viewed (0)
  4. src/test/java/org/codelibs/core/beans/util/BeanUtilTest.java

            final Map<String, Object> dest = newHashMap();
            BeanUtil.copyBeanToMap(src, dest);
            assertThat(dest.get("aaa"), is((Object) "aaa"));
            assertThat(dest.get("bbb"), is(nullValue()));
            assertThat(dest.get("ccc"), is((Object) "ccc"));
            assertThat(dest.get("ddd"), is(nullValue()));
        }
    
        /**
         * @throws Exception
         */
        @Test
    Registered: Wed Jun 12 12:50:12 UTC 2024
    - Last Modified: Thu Mar 07 01:59:08 UTC 2024
    - 34.5K bytes
    - Viewed (0)
  5. src/runtime/rand.go

    //
    //go:nosplit
    //go:linkname rand
    func rand() uint64 {
    	// Note: We avoid acquirem here so that in the fast path
    	// there is just a getg, an inlined c.Next, and a return.
    	// The performance difference on a 16-core AMD is
    	// 3.7ns/call this way versus 4.3ns/call with acquirem (+16%).
    	mp := getg().m
    	c := &mp.chacha8
    	for {
    		// Note: c.Next is marked nosplit,
    		// so we don't need to use mp.locks
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 31 14:32:47 UTC 2024
    - 8K bytes
    - Viewed (0)
  6. src/runtime/coro.go

    // and then calls coroexit to remove the extra concurrency.
    func corostart() {
    	gp := getg()
    	c := gp.coroarg
    	gp.coroarg = nil
    
    	defer coroexit(c)
    	c.f(c)
    }
    
    // coroexit is like coroswitch but closes the coro
    // and exits the current goroutine
    func coroexit(c *coro) {
    	gp := getg()
    	gp.coroarg = c
    	gp.coroexit = true
    	mcall(coroswitch_m)
    }
    
    //go:linkname coroswitch
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Jun 07 19:09:18 UTC 2024
    - 7.4K bytes
    - Viewed (0)
  7. src/runtime/HACKING.md

    so their memory remains type stable. As a result, the runtime can
    avoid write barriers in the depths of the scheduler.
    
    `getg()` and `getg().m.curg`
    ----------------------------
    
    To get the current user `g`, use `getg().m.curg`.
    
    `getg()` alone returns the current `g`, but when executing on the
    system or signal stacks, this will return the current M's "g0" or
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 13.9K bytes
    - Viewed (0)
  8. src/runtime/debug.go

    //
    // Ensure mayMoreStackPreempt can be called for all ABIs.
    //
    //go:nosplit
    //go:linkname mayMoreStackPreempt
    func mayMoreStackPreempt() {
    	// Don't do anything on the g0 or gsignal stack.
    	gp := getg()
    	if gp == gp.m.g0 || gp == gp.m.gsignal {
    		return
    	}
    	// Force a preemption, unless the stack is already poisoned.
    	if gp.stackguard0 < stackPoisonMin {
    		gp.stackguard0 = stackPreempt
    	}
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Sat May 11 20:38:24 UTC 2024
    - 4.2K bytes
    - Viewed (0)
  9. src/runtime/proc.go

    //   - fortio.org/log
    //
    // Do not remove or change the type signature.
    // See go.dev/issue/67401.
    //
    //go:linkname getm
    func getm() uintptr {
    	return uintptr(unsafe.Pointer(getg().m))
    }
    
    var (
    	// Locking linked list of extra M's, via mp.schedlink. Must be accessed
    	// only via lockextra/unlockextra.
    	//
    	// Can't be atomic.Pointer[m] because we use an invalid pointer as a
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 207.5K bytes
    - Viewed (0)
  10. src/runtime/tracestack.go

    		// symbolizer.
    		pcBuf[0] = logicalStackSentinel
    		if getg() == gp {
    			nstk += callers(skip+1, pcBuf[1:])
    		} else if gp != nil {
    			nstk += gcallers(gp, skip, pcBuf[1:])
    		}
    	} else {
    		// Fast path: Unwind using frame pointers.
    		pcBuf[0] = uintptr(skip)
    		if getg() == gp {
    			nstk += fpTracebackPCs(unsafe.Pointer(getfp()), pcBuf[1:])
    		} else if gp != nil {
    			// Three cases:
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue May 21 14:38:56 UTC 2024
    - 11K bytes
    - Viewed (0)
Back to top