Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 30 of 462 for Allocate (0.15 sec)

  1. platforms/core-runtime/io/src/main/java/org/gradle/internal/io/StreamByteBuffer.java

            CharsetDecoder decoder = charset.newDecoder().onMalformedInput(
                    CodingErrorAction.REPLACE).onUnmappableCharacter(
                    CodingErrorAction.REPLACE);
            CharBuffer charbuffer = CharBuffer.allocate(totalBytesUnread());
            ByteBuffer buf = null;
            boolean wasUnderflow = false;
            ByteBuffer nextBuf = null;
            boolean needsFlush = false;
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Fri Apr 12 08:51:14 UTC 2024
    - 16.6K bytes
    - Viewed (0)
  2. src/runtime/mheap.go

    // to safely deal with potentially invalid pointers, since resolving
    // such pointers may race with a span being allocated.
    type mSpanState uint8
    
    const (
    	mSpanDead   mSpanState = iota
    	mSpanInUse             // allocated for garbage collected heap
    	mSpanManual            // allocated for manual management (e.g., stack allocator)
    )
    
    // mSpanStateNames are the names of the span states, indexed by
    // mSpanState.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 22:31:00 UTC 2024
    - 78K bytes
    - Viewed (0)
  3. src/cmd/compile/internal/ssagen/abi.go

    	// need to allocate stack space, so it should be OK to mark them
    	// as NOSPLIT in these cases. In addition, my assumption is that
    	// functions written in assembly are NOSPLIT in most (but not all)
    	// cases. In the case of an ABIInternal target that has too many
    	// parameters to fit into registers, the wrapper would need to
    	// allocate stack space, but this seems like an unlikely scenario.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 19:57:43 UTC 2024
    - 13.8K bytes
    - Viewed (0)
  4. src/runtime/mklockrank.go

    traceBuf < traceStrings;
    
    # Malloc
    allg,
      allocmR,
      allp, # procresize
      execR, # May grow stack
      execW, # May allocate after BeforeFork
      hchan,
      notifyList,
      reflectOffs,
      timer,
      traceStrings,
      userArenaState
    # Above MALLOC are things that can allocate memory.
    < MALLOC
    # Below MALLOC is the malloc implementation.
    < fin,
      spanSetSpine,
      mspanSpecial,
      traceTypeTab,
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:47:01 UTC 2024
    - 9.1K bytes
    - Viewed (0)
  5. src/runtime/malloc.go

    // persistentChunkSize is the number of bytes we allocate when we grow
    // a persistentAlloc.
    const persistentChunkSize = 256 << 10
    
    // persistentChunks is a list of all the persistent chunks we have
    // allocated. The list is maintained through the first word in the
    // persistent chunk. This is updated atomically.
    var persistentChunks *notInHeap
    
    // Wrapper around sysAlloc that can allocate small chunks.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 59.6K bytes
    - Viewed (0)
  6. pilot/pkg/serviceregistry/serviceentry/store_test.go

    	}
    	if !store.allocateNeeded {
    		t.Errorf("expected allocate needed")
    	}
    	store.allocateNeeded = false
    	store.deleteServices(httpDNSRR.NamespacedName())
    	got = store.getAllServices()
    	if len(got) != 0 {
    		t.Errorf("got unexpected services %v", got)
    	}
    	if store.allocateNeeded {
    		t.Errorf("expected no allocate needed")
    	}
    }
    
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Wed May 29 15:31:09 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  7. tensorflow/c/tf_buffer.cc

      }
      const size_t proto_size = in.ByteSizeLong();
      void* buf = port::Malloc(proto_size);
      if (buf == nullptr) {
        return tensorflow::errors::ResourceExhausted(
            "Failed to allocate memory to serialize message of type '",
            in.GetTypeName(), "' and size ", proto_size);
      }
      if (!in.SerializeWithCachedSizesToArray(static_cast<uint8_t*>(buf))) {
        port::Free(buf);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 21:57:32 UTC 2024
    - 3K bytes
    - Viewed (0)
  8. src/runtime/map_fast64.go

    				continue
    			}
    			insertb = b
    			inserti = i
    			goto done
    		}
    		ovf := b.overflow(t)
    		if ovf == nil {
    			break
    		}
    		b = ovf
    	}
    
    	// Did not find mapping for key. Allocate new cell & add entry.
    
    	// If we hit the max load factor or we have too many overflow buckets,
    	// and we're not already in the middle of growing, start growing.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:17:26 UTC 2024
    - 14.1K bytes
    - Viewed (0)
  9. src/runtime/map_fast32.go

    				continue
    			}
    			inserti = i
    			insertb = b
    			goto done
    		}
    		ovf := b.overflow(t)
    		if ovf == nil {
    			break
    		}
    		b = ovf
    	}
    
    	// Did not find mapping for key. Allocate new cell & add entry.
    
    	// If we hit the max load factor or we have too many overflow buckets,
    	// and we're not already in the middle of growing, start growing.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:17:26 UTC 2024
    - 13.9K bytes
    - Viewed (0)
  10. pkg/registry/core/service/ipallocator/cidrallocator.go

    	if err != nil {
    		return err
    	}
    	return allocator.AllocateService(service, ip)
    }
    
    func (c *MetaAllocator) Allocate(ip net.IP) error {
    	allocator, err := c.getAllocator(ip)
    	if err != nil {
    		return err
    	}
    	return allocator.Allocate(ip)
    }
    
    func (c *MetaAllocator) AllocateNextService(service *api.Service) (net.IP, error) {
    	c.muTree.Lock()
    	defer c.muTree.Unlock()
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Sat May 04 18:33:12 UTC 2024
    - 13.2K bytes
    - Viewed (0)
Back to top