Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 61 for pool (0.02 sec)

  1. cmd/erasure-server-pool.go

    	for i, pool := range z.serverPools {
    		pool := pool
    		eg.Go(func() error {
    			dObjectsByPool[i], dErrsByPool[i] = pool.DeleteObjects(ctx, bucket, objects, opts)
    			return nil
    		}, i)
    	}
    	eg.Wait() // wait to check all the pools.
    
    	for i := range dobjects {
    		// Iterate over pools
    		for pool := range z.serverPools {
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 89.1K bytes
    - Viewed (0)
  2. docs/metrics/prometheus/grafana/minio-dashboard.json

              "expr": "minio_cluster_health_status{job=~\"$scrape_jobs\"}",
              "fullMetaSearch": false,
              "includeNullMetadata": true,
              "interval": "",
              "legendFormat": "Pool: {{pool}} Set: {{set}}",
              "range": true,
              "refId": "A",
              "useBackend": false
            }
          ],
          "title": "Cluster Health Status",
          "type": "stat"
        },
        {
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Mon Aug 04 01:46:49 UTC 2025
    - 93.1K bytes
    - Viewed (0)
  3. cmd/batch-handlers.go

    	}
    	jpool.ResizeWorkers(workers)
    
    	randomWait := func() time.Duration {
    		// randomWait depends on the number of nodes to avoid triggering resume and cleanups at the same time.
    		return time.Duration(rand.Float64() * float64(time.Duration(globalEndpoints.NEndpoints())*time.Hour))
    	}
    
    	go func() {
    		jpool.resume(randomWait)
    		jpool.cleanupReports(randomWait)
    	}()
    
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 63.5K bytes
    - Viewed (0)
  4. cmd/erasure-object.go

    	}
    
    	toEncode := io.Reader(data)
    	if data.Size() >= bigFileThreshold {
    		// We use 2 buffers, so we always have a full buffer of input.
    		pool := globalBytePoolCap.Load()
    		bufA := pool.Get()
    		bufB := pool.Get()
    		defer pool.Put(bufA)
    		defer pool.Put(bufB)
    		ra, err := readahead.NewReaderBuffer(data, [][]byte{bufA[:fi.Erasure.BlockSize], bufB[:fi.Erasure.BlockSize]})
    		if err == nil {
    			toEncode = ra
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Sun Sep 07 16:13:09 UTC 2025
    - 80.4K bytes
    - Viewed (0)
  5. docs/changelogs/changelog_3x.md

     *  **There is no longer a global singleton connection pool.** In OkHttp 2.x,
        all `OkHttpClient` instances shared a common connection pool by default.
        In OkHttp 3.x, each new `OkHttpClient` gets its own private connection pool.
        Applications should avoid creating many connection pools as doing so
        prevents connection reuse. Each connection pool holds its own set of
    Registered: Fri Sep 05 11:42:10 UTC 2025
    - Last Modified: Sun Feb 06 14:55:54 UTC 2022
    - 50.8K bytes
    - Viewed (0)
  6. okhttp/src/commonJvmAndroid/kotlin/okhttp3/OkHttpClient.kt

     * your HTTP calls. This is because each client holds its own connection pool and thread pools.
     * Reusing connections and threads reduces latency and saves memory. Conversely, creating a client
     * for each request wastes resources on idle pools.
     *
     * Use `new OkHttpClient()` to create a shared instance with the default settings:
     *
     * ```java
     * // The singleton HTTP client.
    Registered: Fri Sep 05 11:42:10 UTC 2025
    - Last Modified: Mon May 05 16:01:00 UTC 2025
    - 51.7K bytes
    - Viewed (0)
  7. okhttp/src/jvmTest/kotlin/okhttp3/EventListenerTest.kt

        assertThat(dnsEnd.inetAddressList.size).isEqualTo(1)
      }
    
      @Test
      fun noDnsLookupOnPooledConnection() {
        server.enqueue(MockResponse())
        server.enqueue(MockResponse())
    
        // Seed the pool.
        val call1 =
          client.newCall(
            Request
              .Builder()
              .url(server.url("/"))
              .build(),
          )
        val response1 = call1.execute()
    Registered: Fri Sep 05 11:42:10 UTC 2025
    - Last Modified: Fri Jun 20 11:46:46 UTC 2025
    - 60.4K bytes
    - Viewed (0)
  8. cmd/api-errors.go

    	},
    	ErrAdminRebalanceAlreadyStarted: {
    		Code:           "XMinioAdminRebalanceAlreadyStarted",
    		Description:    "Pool rebalance is already started",
    		HTTPStatusCode: http.StatusConflict,
    	},
    	ErrAdminRebalanceNotStarted: {
    		Code:           "XMinioAdminRebalanceNotStarted",
    		Description:    "Pool rebalance is not started",
    		HTTPStatusCode: http.StatusNotFound,
    	},
    	ErrMaximumExpires: {
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Wed Apr 16 07:34:24 UTC 2025
    - 93K bytes
    - Viewed (1)
  9. cmd/xl-storage-format-v2.go

    var metaDataPool = bpool.Pool[[]byte]{New: func() []byte { return make([]byte, 0, metaDataReadDefault) }}
    
    // metaDataPoolGet will return a byte slice with capacity at least metaDataReadDefault.
    // It will be length 0.
    func metaDataPoolGet() []byte {
    	return metaDataPool.Get()[:0]
    }
    
    // metaDataPoolPut will put an unused small buffer back into the pool.
    func metaDataPoolPut(buf []byte) {
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 65.6K bytes
    - Viewed (1)
  10. cmd/test-utils_test.go

    	if err != nil {
    		removeRoots(fsDirs)
    		return nil, nil, err
    	}
    
    	// Wait up to 10 seconds for disks to come online.
    	pools := obj.(*erasureServerPools)
    	t := time.Now()
    	for _, pool := range pools.serverPools {
    		for _, sets := range pool.erasureDisks {
    			for _, s := range sets {
    				if !s.IsLocal() {
    					for !s.IsOnline() {
    						time.Sleep(100 * time.Millisecond)
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 77K bytes
    - Viewed (0)
Back to top