- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 17 for stksize (0.24 sec)
-
src/cmd/compile/internal/ssagen/ssa.go
callArgs = append(callArgs, closure) stksize += int64(types.PtrSize) argStart += int64(types.PtrSize) if dextra != nil { // Extra token of type any for deferproc ACArgs = append(ACArgs, types.Types[types.TINTER]) callArgs = append(callArgs, dextra) stksize += 2 * int64(types.PtrSize) argStart += 2 * int64(types.PtrSize) } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 10 19:44:43 UTC 2024 - 284.9K bytes - Viewed (0) -
src/cmd/link/internal/ld/data.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jun 12 15:10:50 UTC 2024 - 100.5K bytes - Viewed (1) -
src/cmd/link/internal/ld/deadcode.go
off += 4 * arch.PtrSize case abi.Pointer: // reflect.ptrType off += arch.PtrSize case abi.Func: // reflect.funcType off += arch.PtrSize // 4 bytes, pointer aligned case abi.Slice: // reflect.sliceType off += arch.PtrSize case abi.Array: // reflect.arrayType off += 3 * arch.PtrSize case abi.Chan: // reflect.chanType off += 2 * arch.PtrSize case abi.Map: // reflect.mapType
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 14:52:41 UTC 2024 - 19K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/ImmutableSet.java
* can hold setSize elements with the desired load factor. Always returns at least setSize + 2. */ @VisibleForTesting static int chooseTableSize(int setSize) { setSize = Math.max(setSize, 2); // Correct the size for open addressing to match desired load factor. if (setSize < CUTOFF) { // Round up to the next highest power of 2.
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Sun Jun 02 13:36:19 UTC 2024 - 22.5K bytes - Viewed (0) -
src/runtime/mfinal.go
if (unsafe.Sizeof(finalizer{}) != 5*goarch.PtrSize || unsafe.Offsetof(finalizer{}.fn) != 0 || unsafe.Offsetof(finalizer{}.arg) != goarch.PtrSize || unsafe.Offsetof(finalizer{}.nret) != 2*goarch.PtrSize || unsafe.Offsetof(finalizer{}.fint) != 3*goarch.PtrSize || unsafe.Offsetof(finalizer{}.ot) != 4*goarch.PtrSize) { throw("finalizer out of sync") }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 01:56:56 UTC 2024 - 19K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/debug.go
listSym.WriteInt(ctxt, listSym.Size, ctxt.Arch.PtrSize, ^0) listSym.WriteAddr(ctxt, listSym.Size, ctxt.Arch.PtrSize, startPC, 0) } // Re-read list, translating its address from block/value ID to PC. for i := 0; i < len(list); { begin := getPC(decodeValue(ctxt, readPtr(ctxt, list[i:]))) end := getPC(decodeValue(ctxt, readPtr(ctxt, list[i+ctxt.Arch.PtrSize:])))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 10 19:44:43 UTC 2024 - 58.4K bytes - Viewed (0) -
src/runtime/runtime1.go
throw("bad h") } if unsafe.Sizeof(i) != 4 { throw("bad i") } if unsafe.Sizeof(j) != 8 { throw("bad j") } if unsafe.Sizeof(k) != goarch.PtrSize { throw("bad k") } if unsafe.Sizeof(l) != goarch.PtrSize { throw("bad l") } if unsafe.Sizeof(x1) != 1 { throw("bad unsafe.Sizeof x1") } if unsafe.Offsetof(y1.y) != 1 { throw("bad offsetof y1.y") }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:52:17 UTC 2024 - 19.3K bytes - Viewed (0) -
src/cmd/compile/internal/liveness/plive.go
var argsSymTmp, liveSymTmp obj.LSym args := bitvec.New(int32(maxArgs / int64(types.PtrSize))) aoff := objw.Uint32(&argsSymTmp, 0, uint32(len(lv.stackMaps))) // number of bitmaps aoff = objw.Uint32(&argsSymTmp, aoff, uint32(args.N)) // number of bits in each bitmap locals := bitvec.New(int32(maxLocals / int64(types.PtrSize))) loff := objw.Uint32(&liveSymTmp, 0, uint32(len(lv.stackMaps))) // number of bitmaps
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 15:22:22 UTC 2024 - 45.2K bytes - Viewed (0) -
tensorflow/compiler/jit/shape_inference.cc
DataType dtype = n->output_type(0); AddNodeAttr("dtype", dtype, &const_def); TensorProto value; value.set_dtype(dtype); value.mutable_tensor_shape()->add_dim()->set_size( shape_proto.dim_size()); for (const auto& dim : shape_proto.dim()) { if (dtype == DT_INT32) { value.add_int_val(dim.size()); } else {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 00:41:19 UTC 2024 - 13K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
llvm::to_vector<num_dims>(input_type.getShape())); SmallVector<int64_t, num_dims> ksize, strides; GetI64ArrayAttrValues(op.getKsize(), &ksize); GetI64ArrayAttrValues(op.getStrides(), &strides); Operation *result_op = AvgPoolDivideByCount<OpTy, num_dims>( reduce.getResult(0), input_shape, ksize, strides, op, init, rewriter); // Convert back if we enlarged the element type's bitwidth.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0)