Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 23 for NewSource (0.06 sec)

  1. internal/grid/benchmark_test.go

    			return &testResponse{
    				OrgNum:    req.Num,
    				OrgString: req.String,
    				Embedded:  *req,
    			}, nil
    		}))
    		errFatal(err)
    	}
    	const payloadSize = 512
    	rng := rand.New(rand.NewSource(time.Now().UnixNano()))
    	payload := make([]byte, payloadSize)
    	_, err = rng.Read(payload)
    	errFatal(err)
    
    	// Wait for all to connect
    	// Parallel writes per server.
    	b.Run("bytes", func(b *testing.B) {
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 15.5K bytes
    - Viewed (0)
  2. cmd/erasure.go

    	var wg sync.WaitGroup
    	disks := er.getDisks()
    	infos := make([]DiskInfo, len(disks))
    	r := rand.New(rand.NewSource(time.Now().UnixNano()))
    	for _, i := range r.Perm(len(disks)) {
    		wg.Add(1)
    		go func() {
    			defer wg.Done()
    
    			disk := disks[i]
    			if disk == nil {
    				infos[i].Error = errDiskNotFound.Error()
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sun Sep 28 20:59:21 UTC 2025
    - 16.1K bytes
    - Viewed (0)
  3. internal/ringbuffer/ring_buffer_test.go

    	doSleep := !testing.Short()
    	rb := New(4 << 10).SetBlocking(true)
    
    	// Reader
    	var readErr error
    	var wg sync.WaitGroup
    	wg.Add(1)
    	go func() {
    		readRng := rand.New(rand.NewSource(1))
    		defer wg.Done()
    		defer rb.CloseWithError(readErr)
    		buf := make([]byte, 1024)
    		for {
    			// Read
    			n, err := rb.Read(buf[:readRng.Intn(len(buf))])
    			readBytes += n
    			read.Write(buf[:n])
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 26.7K bytes
    - Viewed (0)
  4. okhttp/src/commonJvmAndroid/kotlin/okhttp3/internal/cache/DiskLruCache.kt

            }
          }
        }
    
        /**
         * Returns an unbuffered input stream to read the last committed value, or null if no value has
         * been committed.
         */
        fun newSource(index: Int): Source? {
          synchronized(this@DiskLruCache) {
            check(!done)
            if (!entry.readable || entry.currentEditor != this || entry.zombie) {
              return null
            }
            return try {
    Registered: Fri Dec 26 11:42:13 UTC 2025
    - Last Modified: Wed May 28 23:28:25 UTC 2025
    - 34.7K bytes
    - Viewed (0)
  5. cmd/local-locker_test.go

    	}
    	if len(l.lockUID) != 0 {
    		t.Fatalf("lockUID len, got %d, want %d + %d", len(l.lockUID), 0, 0)
    	}
    }
    
    func Test_localLocker_expireOldLocksExpire(t *testing.T) {
    	rng := rand.New(rand.NewSource(0))
    	quorum := 0
    	// Numbers of unique locks
    	for _, locks := range []int{100, 1000, 1e6} {
    		if testing.Short() && locks > 100 {
    			continue
    		}
    		t.Run(fmt.Sprintf("%d-locks", locks), func(t *testing.T) {
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 11.8K bytes
    - Viewed (0)
  6. cmd/object-api-utils.go

    	return hosts
    }
    
    // returns an online host (and corresponding port) from a slice of DNS records
    func getHostFromSrv(records []dns.SrvRecord) (host string) {
    	hosts := getHostsSlice(records)
    	rng := rand.New(rand.NewSource(time.Now().UTC().UnixNano()))
    	var d net.Dialer
    	var retry int
    	for retry < len(hosts) {
    		ctx, cancel := context.WithTimeout(GlobalContext, 300*time.Millisecond)
    
    		host = hosts[rng.Intn(len(hosts))]
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Wed Jun 25 15:08:54 UTC 2025
    - 37.3K bytes
    - Viewed (0)
  7. cmd/batch-rotate.go

    		retryAttempts = batchKeyRotateJobDefaultRetries
    	}
    	delay := job.KeyRotate.Flags.Retry.Delay
    	if delay <= 0 {
    		delay = batchKeyRotateJobDefaultRetryDelay
    	}
    
    	rnd := rand.New(rand.NewSource(time.Now().UnixNano()))
    
    	selectObj := func(info FileInfo) (ok bool) {
    		if r.Flags.Filter.OlderThan > 0 && time.Since(info.ModTime) < r.Flags.Filter.OlderThan {
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 14.7K bytes
    - Viewed (0)
  8. cmd/server-main.go

    		return
    	}
    
    	globalTrace.Publish(info)
    }
    
    func initServerConfig(ctx context.Context, newObject ObjectLayer) error {
    	t1 := time.Now()
    
    	r := rand.New(rand.NewSource(time.Now().UnixNano()))
    
    	for {
    		select {
    		case <-ctx.Done():
    			// Retry was canceled successfully.
    			return fmt.Errorf("Initializing sub-systems stopped gracefully %w", ctx.Err())
    		default:
    		}
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Tue May 27 15:18:36 UTC 2025
    - 35.9K bytes
    - Viewed (4)
  9. cmd/tier.go

    		Tiers:       make(map[string]madmin.TierConfig),
    	}
    }
    
    func (config *TierConfigMgr) refreshTierConfig(ctx context.Context, objAPI ObjectLayer) {
    	const tierCfgRefresh = 15 * time.Minute
    	r := rand.New(rand.NewSource(time.Now().UnixNano()))
    	randInterval := func() time.Duration {
    		return time.Duration(r.Float64() * 5 * float64(time.Second))
    	}
    
    	// To avoid all MinIO nodes reading the tier config object at the same
    	// time.
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 15.6K bytes
    - Viewed (0)
  10. internal/rest/client.go

    	if c.HealthCheckFn != nil && atomic.CompareAndSwapInt32(&c.connected, online, offline) {
    		go func() {
    			backOff := exponentialBackoffWait(
    				rand.New(rand.NewSource(time.Now().UnixNano())),
    				200*time.Millisecond,
    				30*time.Second,
    			)
    
    			attempt := uint(0)
    			for {
    				if atomic.LoadInt32(&c.connected) == closed {
    					return
    				}
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sun Sep 28 20:59:21 UTC 2025
    - 14.7K bytes
    - Viewed (0)
Back to top