Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 17 for stksize (0.24 sec)

  1. src/cmd/compile/internal/ssagen/ssa.go

    			callArgs = append(callArgs, closure)
    			stksize += int64(types.PtrSize)
    			argStart += int64(types.PtrSize)
    			if dextra != nil {
    				// Extra token of type any for deferproc
    				ACArgs = append(ACArgs, types.Types[types.TINTER])
    				callArgs = append(callArgs, dextra)
    				stksize += 2 * int64(types.PtrSize)
    				argStart += 2 * int64(types.PtrSize)
    			}
    		}
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Jun 10 19:44:43 UTC 2024
    - 284.9K bytes
    - Viewed (0)
  2. src/cmd/link/internal/ld/data.go

    		return
    	}
    
    	ptrsize := int64(p.ctxt.Arch.PtrSize)
    	typData := ldr.Data(typ)
    	nptr := decodetypePtrdata(p.ctxt.Arch, typData) / ptrsize
    
    	if debugGCProg {
    		fmt.Fprintf(os.Stderr, "gcprog sym: %s at %d (ptr=%d+%d)\n", ldr.SymName(s), ldr.SymValue(s), ldr.SymValue(s)/ptrsize, nptr)
    	}
    
    	sval := ldr.SymValue(s)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Jun 12 15:10:50 UTC 2024
    - 100.5K bytes
    - Viewed (1)
  3. src/cmd/link/internal/ld/deadcode.go

    		off += 4 * arch.PtrSize
    	case abi.Pointer: // reflect.ptrType
    		off += arch.PtrSize
    	case abi.Func: // reflect.funcType
    		off += arch.PtrSize // 4 bytes, pointer aligned
    	case abi.Slice: // reflect.sliceType
    		off += arch.PtrSize
    	case abi.Array: // reflect.arrayType
    		off += 3 * arch.PtrSize
    	case abi.Chan: // reflect.chanType
    		off += 2 * arch.PtrSize
    	case abi.Map: // reflect.mapType
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Jun 07 14:52:41 UTC 2024
    - 19K bytes
    - Viewed (0)
  4. android/guava/src/com/google/common/collect/ImmutableSet.java

       * can hold setSize elements with the desired load factor. Always returns at least setSize + 2.
       */
      @VisibleForTesting
      static int chooseTableSize(int setSize) {
        setSize = Math.max(setSize, 2);
        // Correct the size for open addressing to match desired load factor.
        if (setSize < CUTOFF) {
          // Round up to the next highest power of 2.
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Sun Jun 02 13:36:19 UTC 2024
    - 22.5K bytes
    - Viewed (0)
  5. src/runtime/mfinal.go

    				if (unsafe.Sizeof(finalizer{}) != 5*goarch.PtrSize ||
    					unsafe.Offsetof(finalizer{}.fn) != 0 ||
    					unsafe.Offsetof(finalizer{}.arg) != goarch.PtrSize ||
    					unsafe.Offsetof(finalizer{}.nret) != 2*goarch.PtrSize ||
    					unsafe.Offsetof(finalizer{}.fint) != 3*goarch.PtrSize ||
    					unsafe.Offsetof(finalizer{}.ot) != 4*goarch.PtrSize) {
    					throw("finalizer out of sync")
    				}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Jun 07 01:56:56 UTC 2024
    - 19K bytes
    - Viewed (0)
  6. src/cmd/compile/internal/ssa/debug.go

    		listSym.WriteInt(ctxt, listSym.Size, ctxt.Arch.PtrSize, ^0)
    		listSym.WriteAddr(ctxt, listSym.Size, ctxt.Arch.PtrSize, startPC, 0)
    	}
    
    	// Re-read list, translating its address from block/value ID to PC.
    	for i := 0; i < len(list); {
    		begin := getPC(decodeValue(ctxt, readPtr(ctxt, list[i:])))
    		end := getPC(decodeValue(ctxt, readPtr(ctxt, list[i+ctxt.Arch.PtrSize:])))
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Jun 10 19:44:43 UTC 2024
    - 58.4K bytes
    - Viewed (0)
  7. src/runtime/runtime1.go

    		throw("bad h")
    	}
    	if unsafe.Sizeof(i) != 4 {
    		throw("bad i")
    	}
    	if unsafe.Sizeof(j) != 8 {
    		throw("bad j")
    	}
    	if unsafe.Sizeof(k) != goarch.PtrSize {
    		throw("bad k")
    	}
    	if unsafe.Sizeof(l) != goarch.PtrSize {
    		throw("bad l")
    	}
    	if unsafe.Sizeof(x1) != 1 {
    		throw("bad unsafe.Sizeof x1")
    	}
    	if unsafe.Offsetof(y1.y) != 1 {
    		throw("bad offsetof y1.y")
    	}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:52:17 UTC 2024
    - 19.3K bytes
    - Viewed (0)
  8. src/cmd/compile/internal/liveness/plive.go

    	var argsSymTmp, liveSymTmp obj.LSym
    
    	args := bitvec.New(int32(maxArgs / int64(types.PtrSize)))
    	aoff := objw.Uint32(&argsSymTmp, 0, uint32(len(lv.stackMaps))) // number of bitmaps
    	aoff = objw.Uint32(&argsSymTmp, aoff, uint32(args.N))          // number of bits in each bitmap
    
    	locals := bitvec.New(int32(maxLocals / int64(types.PtrSize)))
    	loff := objw.Uint32(&liveSymTmp, 0, uint32(len(lv.stackMaps))) // number of bitmaps
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Jun 07 15:22:22 UTC 2024
    - 45.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/jit/shape_inference.cc

              DataType dtype = n->output_type(0);
              AddNodeAttr("dtype", dtype, &const_def);
              TensorProto value;
              value.set_dtype(dtype);
              value.mutable_tensor_shape()->add_dim()->set_size(
                  shape_proto.dim_size());
              for (const auto& dim : shape_proto.dim()) {
                if (dtype == DT_INT32) {
                  value.add_int_val(dim.size());
                } else {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 00:41:19 UTC 2024
    - 13K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc

            llvm::to_vector<num_dims>(input_type.getShape()));
        SmallVector<int64_t, num_dims> ksize, strides;
        GetI64ArrayAttrValues(op.getKsize(), &ksize);
        GetI64ArrayAttrValues(op.getStrides(), &strides);
    
        Operation *result_op = AvgPoolDivideByCount<OpTy, num_dims>(
            reduce.getResult(0), input_shape, ksize, strides, op, init, rewriter);
    
        // Convert back if we enlarged the element type's bitwidth.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 20:00:43 UTC 2024
    - 291.8K bytes
    - Viewed (0)
Back to top