- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 161 for getPkg (0.38 sec)
-
src/runtime/lock_sema.go
} func notetsleep(n *note, ns int64) bool { gp := getg() if gp != gp.m.g0 { throw("notetsleep not on g0") } semacreate(gp.m) return notetsleep_internal(n, ns, nil, 0) } // same as runtimeĀ·notetsleep, but called on user g (not g0) // calls only nosplit functions between entersyscallblock/exitsyscall. func notetsleepg(n *note, ns int64) bool { gp := getg() if gp == gp.m.g0 { throw("notetsleepg on g0")
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 6.8K bytes - Viewed (0) -
src/runtime/race.go
//go:nosplit func racerelease(addr unsafe.Pointer) { racereleaseg(getg(), addr) } //go:nosplit func racereleaseg(gp *g, addr unsafe.Pointer) { if getg().raceignore != 0 || !isvalidaddr(addr) { return } racecall(&__tsan_release, gp.racectx, uintptr(addr), 0, 0) } //go:nosplit func racereleaseacquire(addr unsafe.Pointer) { racereleaseacquireg(getg(), addr) } //go:nosplit
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 20.4K bytes - Viewed (0) -
src/cmd/compile/internal/wasm/ssa.go
s.Prog(wasm.AI32WrapI64) } return } reg := v.Reg() getReg(s, reg) if reg != wasm.REG_SP { s.Prog(wasm.AI32WrapI64) } } func getValue64(s *ssagen.State, v *ssa.Value) { if v.OnWasmStack { s.OnWasmStackSkipped-- ssaGenValueOnStack(s, v, true) return } reg := v.Reg() getReg(s, reg) if reg == wasm.REG_SP { s.Prog(wasm.AI64ExtendI32U) } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Feb 24 00:21:13 UTC 2023 - 17.6K bytes - Viewed (0) -
tools/bug-report/pkg/bugreport/bugreport.go
}() clog, _, _, err := getLog(runner, resources, config, namespace, pod, "") appendGlobalErr(err) writeFile(filepath.Join(archive.CniPath(tempDir, pod), "cni.log"), clog, config.DryRun) log.Infof("Done with CNI logs %v", pod) }() } // getLog fetches the logs for the given namespace/pod/container and returns the log text and stats for it.
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Fri Apr 05 20:57:29 UTC 2024 - 20.7K bytes - Viewed (0) -
src/runtime/os_linux.go
// procid. We need this for asynchronous preemption and it's // useful in debuggers. getg().m.procid = uint64(gettid()) } // Called from dropm to undo the effect of an minit. // //go:nosplit func unminit() { unminitSignals() getg().m.procid = 0 } // Called from exitm, but not from drop, to undo the effect of thread-owned
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 25.9K bytes - Viewed (0) -
src/runtime/os_solaris.go
func asmsysvicall6() // declared for vet; do NOT call //go:nosplit func sysvicall0(fn *libcFunc) uintptr { // Leave caller's PC/SP around for traceback. gp := getg() var mp *m if gp != nil { mp = gp.m } if mp != nil && mp.libcallsp == 0 { mp.libcallg.set(gp) mp.libcallpc = getcallerpc() // sp must be the last, because once async cpu profiler finds
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jun 06 18:49:01 UTC 2023 - 6.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/prepare_quantize.cc
PatternRewriter& rewriter) const override { auto preceding_qcast = q_op.getArg().getDefiningOp<quantfork::QuantizeCastOp>(); if (!preceding_qcast) return failure(); auto new_qcast = rewriter.create<quantfork::QuantizeCastOp>( q_op.getLoc(), q_op.getType(), preceding_qcast.getArg()); new_qcast->setAttr(kVolatileOpAttrName, rewriter.getUnitAttr()); q_op->replaceAllUsesWith(new_qcast);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 03 05:11:03 UTC 2024 - 8.1K bytes - Viewed (0) -
src/runtime/lock_wasip1.go
// observe this. throw("self deadlock") } gp := getg() if gp.m.locks < 0 { throw("lock count") } gp.m.locks++ l.key = mutex_locked } func unlock(l *mutex) { unlockWithRank(l) } func unlock2(l *mutex) { if l.key == mutex_unlocked { throw("unlock of unlocked lock") } gp := getg() gp.m.locks-- if gp.m.locks < 0 { throw("lock count") }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 21 21:02:20 UTC 2023 - 2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/analysis/resource_alias_analysis.cc
// can backtracked to. Such results will be called "function passthrough". If // the result cannot be backtracked to a region argument, returns // std::nullopt. std::optional<int> GetArg(int result_index) const { if (auto arg = mlir::dyn_cast<BlockArgument>(GetValue(result_index))) if (arg.getParentBlock() == ®ion_->front()) return arg.getArgNumber(); return std::nullopt; } private:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 15 09:04:13 UTC 2024 - 28.2K bytes - Viewed (0) -
src/main/java/org/codelibs/core/log/JclLoggerAdapter.java
protected final String sourceClass; protected final Log logger; public JclLoggerAdapter(final Class<?> clazz) { sourceClass = clazz.getName(); logger = LogFactory.getLog(clazz); } @Override public boolean isFatalEnabled() { return logger.isFatalEnabled(); } @Override public void fatal(final String message) { logger.fatal(message);
Registered: Wed Jun 12 12:50:12 UTC 2024 - Last Modified: Thu Mar 07 01:59:08 UTC 2024 - 2.6K bytes - Viewed (0)