Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 88 for fallocate (0.35 sec)

  1. src/runtime/map.go

    	// For hint < 0 overLoadFactor returns false since hint < bucketCnt.
    	B := uint8(0)
    	for overLoadFactor(hint, B) {
    		B++
    	}
    	h.B = B
    
    	// allocate initial hash table
    	// if B == 0, the buckets field is allocated lazily later (in mapassign)
    	// If hint is large zeroing this memory could take a while.
    	if h.B != 0 {
    		var nextOverflow *bmap
    		h.buckets, nextOverflow = makeBucketArray(t, h.B, nil)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 57.6K bytes
    - Viewed (0)
  2. src/runtime/mgcscavenge.go

    // The background scavenger and heap-growth scavenger only release memory in chunks
    // that have not been densely-allocated for at least 1 full GC cycle. The reason
    // behind this is likelihood of reuse: the Go heap is allocated in a first-fit order
    // and by the end of the GC mark phase, the heap tends to be densely packed. Releasing
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:48:45 UTC 2024
    - 52.3K bytes
    - Viewed (0)
  3. src/runtime/mgc.go

    // goroutine needs another span, it first attempts to reclaim that much memory
    // by sweeping. When a goroutine needs to allocate a new small-object span, it
    // sweeps small-object spans for the same object size until it frees at least
    // one object. When a goroutine needs to allocate large-object span from heap,
    // it sweeps spans until it frees at least that many pages into heap. There is
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 62K bytes
    - Viewed (0)
  4. tensorflow/c/c_api.h

        int max_values, TF_Status* status);
    
    // Gets the TF_Tensor valued attribute of `attr_name` of `oper`.
    //
    // Allocates a new TF_Tensor which the caller is expected to take
    // ownership of (and can deallocate using TF_DeleteTensor).
    TF_CAPI_EXPORT extern void TF_OperationGetAttrTensor(TF_Operation* oper,
                                                         const char* attr_name,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Oct 26 21:08:15 UTC 2023
    - 82.3K bytes
    - Viewed (0)
  5. tensorflow/c/kernels_test.cc

    }
    
    REGISTER_OP("AllocateOutputOp1").Output("output1: float");
    
    TEST_F(DeviceKernelOpTest, TestAllocateOutputSizeOne) {
      auto my_compute_func = [](void* kernel, TF_OpKernelContext* ctx) {
        // Allocate output
        TF_Status* s = TF_NewStatus();
        int64_t dim = 1;
        size_t tensor_size_bytes = TF_DataTypeSize(TF_FLOAT);
        TF_Tensor* output = TF_AllocateOutput(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 50.4K bytes
    - Viewed (0)
  6. src/vendor/golang.org/x/net/dns/dnsmessage/message.go

    func (p *Parser) AllQuestions() ([]Question, error) {
    	// Multiple questions are valid according to the spec,
    	// but servers don't actually support them. There will
    	// be at most one question here.
    	//
    	// Do not pre-allocate based on info in p.header, since
    	// the data is untrusted.
    	qs := []Question{}
    	for {
    		q, err := p.Question()
    		if err == ErrSectionDone {
    			return qs, nil
    		}
    		if err != nil {
    			return nil, err
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Sat Mar 09 00:09:40 UTC 2024
    - 69K bytes
    - Viewed (0)
  7. cmd/erasure-object.go

    	if err != nil {
    		return ObjectInfo{}, toObjectErr(err, minioMetaBucket, key)
    	}
    
    	// Fetch buffer for I/O, returns from the pool if not allocates a new one and returns.
    	var buffer []byte
    	switch size := data.Size(); {
    	case size == 0:
    		buffer = make([]byte, 1) // Allocate at least a byte to reach EOF
    	case size >= fi.Erasure.BlockSize:
    		buffer = globalBytePoolCap.Load().Get()
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Mon Jun 10 15:51:27 UTC 2024
    - 78.6K bytes
    - Viewed (0)
  8. src/fmt/fmt_test.go

    	}
    	if &b[0] != &got[0] {
    		t.Fatalf("Appendf allocated a new slice")
    	}
    }
    
    func TestAppend(t *testing.T) {
    	b := make([]byte, 100)
    	b = b[:copy(b, hello)]
    	got := Append(b, "world", ", ", 23)
    	if string(got) != appendResult {
    		t.Fatalf("Append returns %q not %q", got, appendResult)
    	}
    	if &b[0] != &got[0] {
    		t.Fatalf("Append allocated a new slice")
    	}
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 04 17:31:55 UTC 2024
    - 58.6K bytes
    - Viewed (0)
  9. pilot/pkg/serviceregistry/serviceentry/controller_test.go

    			},
    			wantServices: []*model.Service{
    				{
    					Hostname:       "*.foo.com",
    					Resolution:     model.ClientSideLB,
    					DefaultAddress: "1.1.1.1",
    				},
    			},
    		},
    		{
    			name: "allocate IP for clientside lb",
    			inServices: []*model.Service{
    				{
    					Hostname:       "foo.com",
    					Resolution:     model.ClientSideLB,
    					DefaultAddress: "0.0.0.0",
    				},
    			},
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Wed May 29 15:31:09 UTC 2024
    - 92.9K bytes
    - Viewed (0)
  10. src/runtime/mbitmap.go

    		}
    		// Handle the case where we have no malloc header.
    		scanSize = span.writeHeapBitsSmall(x, dataSize, typ)
    	} else {
    		if typ.Kind_&abi.KindGCProg != 0 {
    			// Allocate space to unroll the gcprog. This space will consist of
    			// a dummy _type value and the unrolled gcprog. The dummy _type will
    			// refer to the bitmap, and the mspan will refer to the dummy _type.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 00:18:55 UTC 2024
    - 60K bytes
    - Viewed (0)
Back to top