Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 13 for GETCALLERPC (0.21 sec)

  1. src/runtime/panic.go

    	panic(boundsError{x: int64(x), signed: true, y: y, code: boundsSlice3Acap})
    }
    func goPanicSlice3AcapU(x uint, y int) {
    	panicCheck1(getcallerpc(), "slice bounds out of range")
    	panic(boundsError{x: int64(x), signed: false, y: y, code: boundsSlice3Acap})
    }
    
    // failures in the comparisons for s[:x:y], 0 <= x <= y
    func goPanicSlice3B(x int, y int) {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 43.8K bytes
    - Viewed (0)
  2. src/runtime/slice.go

    			// type et. See the comment on bulkBarrierPreWrite.
    			bulkBarrierPreWriteSrcOnly(uintptr(to), uintptr(from), copymem, et)
    		}
    	}
    
    	if raceenabled {
    		callerpc := getcallerpc()
    		pc := abi.FuncPCABIInternal(makeslicecopy)
    		racereadrangepc(from, copymem, callerpc, pc)
    	}
    	if msanenabled {
    		msanread(from, copymem)
    	}
    	if asanenabled {
    		asanread(from, copymem)
    	}
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 12.2K bytes
    - Viewed (0)
  3. src/runtime/stubs.go

    func publicationBarrier()
    
    // getcallerpc returns the program counter (PC) of its caller's caller.
    // getcallersp returns the stack pointer (SP) of its caller's caller.
    // The implementation may be a compiler intrinsic; there is not
    // necessarily code implementing this on every platform.
    //
    // For example:
    //
    //	func f(arg1, arg2, arg3 int) {
    //		pc := getcallerpc()
    //		sp := getcallersp()
    //	}
    //
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 20.2K bytes
    - Viewed (0)
  4. src/runtime/mbarrier.go

    //
    //go:linkname reflect_typedmemmove reflect.typedmemmove
    func reflect_typedmemmove(typ *_type, dst, src unsafe.Pointer) {
    	if raceenabled {
    		raceWriteObjectPC(typ, dst, getcallerpc(), abi.FuncPCABIInternal(reflect_typedmemmove))
    		raceReadObjectPC(typ, src, getcallerpc(), abi.FuncPCABIInternal(reflect_typedmemmove))
    	}
    	if msanenabled {
    		msanwrite(dst, typ.Size_)
    		msanread(src, typ.Size_)
    	}
    	if asanenabled {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 15.7K bytes
    - Viewed (0)
  5. src/runtime/map.go

    // NOTE: The returned pointer may keep the whole map live, so don't
    // hold onto it for very long.
    func mapaccess1(t *maptype, h *hmap, key unsafe.Pointer) unsafe.Pointer {
    	if raceenabled && h != nil {
    		callerpc := getcallerpc()
    		pc := abi.FuncPCABIInternal(mapaccess1)
    		racereadpc(unsafe.Pointer(h), callerpc, pc)
    		raceReadObjectPC(t.Key, key, callerpc, pc)
    	}
    	if msanenabled && h != nil {
    		msanread(key, t.Key.Size_)
    	}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 57.6K bytes
    - Viewed (0)
  6. src/runtime/mprof.go

    		prof.stack[1] = abi.FuncPCABIInternal(_LostContendedRuntimeLock) + sys.PCQuantum
    		prof.stack[2] = 0
    		return
    	}
    
    	var nstk int
    	gp := getg()
    	sp := getcallersp()
    	pc := getcallerpc()
    	systemstack(func() {
    		var u unwinder
    		u.initAt(pc, sp, 0, gp, unwindSilentErrors|unwindJumpStack)
    		nstk = 1 + tracebackPCs(&u, skip, prof.stack[1:])
    	})
    	if nstk < len(prof.stack) {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:57:37 UTC 2024
    - 53.3K bytes
    - Viewed (0)
  7. src/runtime/coro.go

    }
    
    //go:linkname newcoro
    
    // newcoro creates a new coro containing a
    // goroutine blocked waiting to run f
    // and returns that coro.
    func newcoro(f func(*coro)) *coro {
    	c := new(coro)
    	c.f = f
    	pc := getcallerpc()
    	gp := getg()
    	systemstack(func() {
    		mp := gp.m
    		start := corostart
    		startfv := *(**funcval)(unsafe.Pointer(&start))
    		gp = newproc1(startfv, gp, pc, true, waitReasonCoroutine)
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Jun 07 19:09:18 UTC 2024
    - 7.4K bytes
    - Viewed (0)
  8. src/runtime/iface.go

    // be used as the second word of an interface value.
    func convT(t *_type, v unsafe.Pointer) unsafe.Pointer {
    	if raceenabled {
    		raceReadObjectPC(t, v, getcallerpc(), abi.FuncPCABIInternal(convT))
    	}
    	if msanenabled {
    		msanread(v, t.Size_)
    	}
    	if asanenabled {
    		asanread(v, t.Size_)
    	}
    	x := mallocgc(t.Size_, t, true)
    	typedmemmove(t, x, v)
    	return x
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 22.5K bytes
    - Viewed (0)
  9. src/runtime/traceback.go

    	traceback1(pc, sp, lr, gp, 0)
    }
    
    // tracebacktrap is like traceback but expects that the PC and SP were obtained
    // from a trap, not from gp->sched or gp->syscallpc/gp->syscallsp or getcallerpc/getcallersp.
    // Because they are from a trap instead of from a saved pair,
    // the initial PC must not be rewound to the previous instruction.
    // (All the saved pairs record a PC that is a return address, so we
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 55.1K bytes
    - Viewed (0)
  10. src/runtime/proc.go

    		// the stack, but put it in gp.stack before mstart,
    		// so the logic above hasn't set osStack yet.
    		osStack = true
    	}
    	mexit(osStack)
    }
    
    // The go:noinline is to guarantee the getcallerpc/getcallersp below are safe,
    // so that we can set up g0.sched to return to the call of mstart1 above.
    //
    //go:noinline
    func mstart1() {
    	gp := getg()
    
    	if gp != gp.m.g0 {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 207.5K bytes
    - Viewed (0)
Back to top