- Sort Score
- Result 10 results
- Languages All
Results 131 - 140 of 173 for getPkg (0.14 sec)
-
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_driver.cc
for (int i = 0; i < op->getNumOperands(); ++i) { if (auto dq_op = dyn_cast_or_null<quantfork::DequantizeCastOp>( op->getOperand(i).getDefiningOp())) { auto type = mlir::cast<TensorType>(dq_op.getArg().getType()).getElementType(); if (auto per_axis_qtype = mlir::dyn_cast_or_null<quant::UniformQuantizedPerAxisType>( QuantizedType::getQuantizedElementType(type))) { return true;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 38.1K bytes - Viewed (0) -
src/cmd/compile/internal/typecheck/func.go
} if t.NumResults() == 1 { n.SetType(l.Type().Result(0).Type) if n.Op() == ir.OCALLFUNC && n.Fun.Op() == ir.ONAME { if sym := n.Fun.(*ir.Name).Sym(); types.RuntimeSymName(sym) == "getg" { // Emit code for runtime.getg() directly instead of calling function. // Most such rewrites (for example the similar one for math.Sqrt) should be done in walk,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Mar 06 15:23:18 UTC 2024 - 21.1K bytes - Viewed (0) -
src/runtime/cgocheck.go
} if !cgoIsGoPointer(src) { return } if cgoIsGoPointer(unsafe.Pointer(dst)) { return } // If we are running on the system stack then dst might be an // address on the stack, which is OK. gp := getg() if gp == gp.m.g0 || gp == gp.m.gsignal { return } // Allocating memory can write to various mfixalloc structs // that look like they are non-Go memory. if gp.m.mallocing != 0 { return }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 09 04:07:57 UTC 2024 - 7.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/ir/tfr_ops.td
let extraClassDeclaration = [{ // Return element type of the input tensor type. Only available when the // input is a MLIR built-in tensor type. Attribute getInputElementType() { if (auto ty = getArg().getType().dyn_cast<TensorType>()) { return TypeAttr::get(ty.getElementType()); } return {}; } }]; let hasCanonicalizer = 1; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 22 10:54:29 UTC 2024 - 17.4K bytes - Viewed (0) -
src/runtime/stubs.go
//go:linkname add //go:nosplit func add(p unsafe.Pointer, x uintptr) unsafe.Pointer { return unsafe.Pointer(uintptr(p) + x) } // getg returns the pointer to the current g. // The compiler rewrites calls to this function into instructions // that fetch the g directly (from TLS or from the dedicated register). func getg() *g // mcall switches from the g to the g0 stack and invokes fn(g), // where g is the goroutine that made the call.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 20.2K bytes - Viewed (0) -
src/runtime/malloc.go
} else if size&1 == 0 { align = 2 } else { align = 1 } } return persistentalloc(size, align, &memstats.other_sys) } if inittrace.active && inittrace.id == getg().goid { // Init functions are executed sequentially in a single goroutine. inittrace.allocs += 1 } } // assistG is the G to charge for this allocation, or nil if
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
src/runtime/mbitmap.go
// another stack. Either way, no need for barriers. // This will also catch if dst is in a freed span, // though that should never have. return } buf := &getg().m.p.ptr().wbBuf // Double-check that the bitmaps generated in the two possible paths match. const doubleCheck = false if doubleCheck { doubleCheckTypePointersOfType(s, typ, dst, size) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 60K bytes - Viewed (0) -
src/runtime/traceruntime.go
// buffer flushes are rare. Record the lock edge even if it doesn't happen // this time. lockRankMayTraceFlush() // Check if we're already locked. If so, return an invalid traceLocker. if getg().m.trace.seqlock.Load()%2 == 1 { return traceLocker{} } return traceAcquireEnabled() } // ok returns true if the traceLocker is valid (i.e. tracing is enabled). //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 25.7K bytes - Viewed (0) -
src/runtime/runtime.go
// writeErrData is the common parts of writeErr{,Str}. // //go:nosplit func writeErrData(data *byte, n int32) { write(2, unsafe.Pointer(data), n) // If crashing, print a copy to the SetCrashOutput fd. gp := getg() if gp != nil && gp.m.dying > 0 || gp == nil && panicking.Load() > 0 { if fd := crashFD.Load(); fd != ^uintptr(0) { write(fd, unsafe.Pointer(data), n) } } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:16:47 UTC 2024 - 9.9K bytes - Viewed (0) -
src/runtime/netpoll_solaris.go
libc_port_associate, libc_port_dissociate, libc_port_getn, libc_port_alert libcFunc netpollWakeSig atomic.Uint32 // used to avoid duplicate calls of netpollBreak ) func errno() int32 { return *getg().m.perrno } func port_create() int32 { return int32(sysvicall0(&libc_port_create)) } func port_associate(port, source int32, object uintptr, events uint32, user uintptr) int32 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 11.2K bytes - Viewed (0)