Search Options

Results per page
Sort
Preferred Languages
Advance

Results 61 - 70 of 87 for chunkOf (0.17 sec)

  1. src/runtime/mgcsweep.go

    		// and place it on the ready list. Don't add it back to any sweep lists.
    		systemstack(func() {
    			// It's the arena code's responsibility to get the chunk on the quarantine
    			// list by the time all references to the chunk are gone.
    			if s.list != &mheap_.userArena.quarantineList {
    				throw("user arena span is on the wrong list")
    			}
    			lock(&mheap_.lock)
    			mheap_.userArena.quarantineList.remove(s)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:52:18 UTC 2024
    - 32.9K bytes
    - Viewed (0)
  2. src/crypto/tls/tls_test.go

    	clientConfig.MaxVersion = version
    
    	buf := make([]byte, bufsize)
    	chunks := int(math.Ceil(float64(totalBytes) / float64(len(buf))))
    	for i := 0; i < N; i++ {
    		conn, err := Dial("tcp", ln.Addr().String(), clientConfig)
    		if err != nil {
    			b.Fatal(err)
    		}
    		for j := 0; j < chunks; j++ {
    			_, err := conn.Write(buf)
    			if err != nil {
    				b.Fatal(err)
    			}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 03:10:12 UTC 2024
    - 60.5K bytes
    - Viewed (0)
  3. staging/src/k8s.io/apiserver/pkg/storage/cacher/cacher.go

    	"k8s.io/component-base/tracing"
    	"k8s.io/klog/v2"
    	"k8s.io/utils/clock"
    	"k8s.io/utils/ptr"
    )
    
    var (
    	emptyFunc = func(bool) {}
    )
    
    const (
    	// storageWatchListPageSize is the cacher's request chunk size of
    	// initial and resync watch lists to storage.
    	storageWatchListPageSize = int64(10000)
    	// defaultBookmarkFrequency defines how frequently watch bookmarks should be send
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Jun 12 10:12:02 UTC 2024
    - 51.8K bytes
    - Viewed (0)
  4. pkg/kubelet/server/server.go

    		return
    	}
    	fw := flushwriter.Wrap(response.ResponseWriter)
    	response.Header().Set("Transfer-Encoding", "chunked")
    	if err := s.host.GetKubeletContainerLogs(ctx, kubecontainer.GetPodFullName(pod), containerName, logOptions, fw, fw); err != nil {
    		response.WriteError(http.StatusBadRequest, err)
    		return
    	}
    }
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Tue Jun 04 06:25:43 UTC 2024
    - 40.1K bytes
    - Viewed (0)
  5. src/net/net.go

    }
    
    func releaseThread() {
    	<-threadLimit
    }
    
    // buffersWriter is the interface implemented by Conns that support a
    // "writev"-like batch write optimization.
    // writeBuffers should fully consume and write all chunks from the
    // provided Buffers, else it should report a non-nil error.
    type buffersWriter interface {
    	writeBuffers(*Buffers) (int64, error)
    }
    
    // Buffers contains zero or more runs of bytes to write.
    //
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 26.8K bytes
    - Viewed (0)
  6. src/net/http/fs.go

    	ctypes, haveType := w.Header()["Content-Type"]
    	var ctype string
    	if !haveType {
    		ctype = mime.TypeByExtension(filepath.Ext(name))
    		if ctype == "" {
    			// read a chunk to decide between utf-8 text and binary
    			var buf [sniffLen]byte
    			n, _ := io.ReadFull(content, buf[:])
    			ctype = DetectContentType(buf[:n])
    			_, err := content.Seek(0, io.SeekStart) // rewind to output whole file
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 09 17:06:47 UTC 2024
    - 31.1K bytes
    - Viewed (0)
  7. src/runtime/pprof/pprof.go

    		return fmt.Errorf("cpu profiling already in use")
    	}
    	cpu.profiling = true
    	runtime.SetCPUProfileRate(hz)
    	go profileWriter(w)
    	return nil
    }
    
    // readProfile, provided by the runtime, returns the next chunk of
    // binary CPU profiling stack trace data, blocking until data is available.
    // If profiling is turned off and all the profile data accumulated while it was
    // on has been returned, readProfile returns eof=true.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:52:17 UTC 2024
    - 30.6K bytes
    - Viewed (0)
  8. cmd/storage-rest-server.go

    		// Windows can lock up with this optimization, so we fall back to regular copy.
    		sr, ok := rc.(*sendFileReader)
    		if ok {
    			// Sendfile sends in 4MiB chunks per sendfile syscall which is more than enough
    			// for most setups.
    			_, err = rf.ReadFrom(sr.Reader)
    			if !xnet.IsNetworkOrHostDown(err, true) { // do not need to log disconnected clients
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Mon Jun 10 15:51:27 UTC 2024
    - 44.8K bytes
    - Viewed (0)
  9. src/cmd/go/go_test.go

    		if isAlpineLinux() || runtime.Compiler == "gccgo" {
    			canRace = false
    		}
    	}
    
    	if n, limited := base.NetLimit(); limited && n > 0 {
    		// Split the network limit into chunks, so that each parallel script can
    		// have one chunk. We want to run as many parallel scripts as possible, but
    		// also want to give each script as high a limit as possible.
    		// We arbitrarily split by sqrt(n) to try to balance those two goals.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:00:11 UTC 2024
    - 81.1K bytes
    - Viewed (0)
  10. src/runtime/trace.go

    	}
    	for i, str := range traceGoStopReasonStrings[:] {
    		trace.goStopReasons[gen%2][i] = traceArg(trace.stringTab[gen%2].put(gen, str))
    	}
    }
    
    // ReadTrace returns the next chunk of binary tracing data, blocking until data
    // is available. If tracing is turned off and all the data accumulated while it
    // was on has been returned, ReadTrace returns nil. The caller must copy the
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 21:17:41 UTC 2024
    - 37.1K bytes
    - Viewed (0)
Back to top