Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 21 for Peers (0.02 sec)

  1. cmd/site-replication.go

    		})
    	}
    	if err != nil {
    		return errSRServiceAccount(fmt.Errorf("unable to create service account on %s: %v", ourName, err))
    	}
    
    	peers := make(map[string]madmin.PeerInfo, len(arg.Peers))
    	for dID, pi := range arg.Peers {
    		if c.state.Peers != nil {
    			if existingPeer, ok := c.state.Peers[dID]; ok {
    				// retain existing ReplicateILMExpiry of peer if its already set
    				// and incoming arg has it false. it could be default false
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Thu Aug 15 12:04:40 UTC 2024
    - 185.1K bytes
    - Viewed (0)
  2. cmd/notification.go

    		}
    	}
    
    	if pools, ok := objAPI.(*erasureServerPools); ok {
    		pools.StopRebalance()
    	}
    }
    
    // LoadRebalanceMeta notifies all peers to load rebalance.bin from object layer.
    // Note: Only peers participating in rebalance operation, namely the first node
    // in each pool will load rebalance.bin.
    func (sys *NotificationSys) LoadRebalanceMeta(ctx context.Context, startRebalance bool) {
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Mon Sep 09 16:58:30 UTC 2024
    - 46.2K bytes
    - Viewed (0)
  3. cmd/peer-s3-client.go

    func newPeerS3Clients(endpoints EndpointServerPools) (peers []peerS3Client) {
    	nodes := endpoints.GetNodes()
    	peers = make([]peerS3Client, len(nodes))
    	for i, node := range nodes {
    		if node.IsLocal {
    			peers[i] = &localPeerS3Client{node: node}
    		} else {
    			peers[i] = newPeerS3Client(node)
    		}
    		peers[i].SetPools(node.Pools)
    	}
    
    	return peers
    }
    
    // Returns a peer S3 client.
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Tue Aug 13 22:26:05 UTC 2024
    - 15.4K bytes
    - Viewed (0)
  4. cmd/logging.go

    func peersLogIf(ctx context.Context, err error, errKind ...interface{}) {
    	if !errors.Is(err, grid.ErrDisconnected) {
    		logger.LogIf(ctx, "peers", err, errKind...)
    	}
    }
    
    func peersLogAlwaysIf(ctx context.Context, err error, errKind ...interface{}) {
    	if !errors.Is(err, grid.ErrDisconnected) {
    		logger.LogAlwaysIf(ctx, "peers", err, errKind...)
    	}
    }
    
    func peersLogOnceIf(ctx context.Context, err error, id string, errKind ...interface{}) {
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Wed Jul 03 18:49:48 UTC 2024
    - 7.1K bytes
    - Viewed (0)
  5. cmd/listen-notification-handlers.go

    					continue
    				}
    				mergeCh <- append(grid.GetByteBuffer()[:0], buf.Bytes()...)
    			case <-ctx.Done():
    				grid.PutByteBuffer(buf.Bytes())
    				return
    			}
    		}
    	}()
    	peers, _ := newPeerRestClients(globalEndpoints)
    	err := globalHTTPListen.Subscribe(mask, localCh, ctx.Done(), func(ev event.Event) bool {
    		if ev.S3.Bucket.Name != "" && bucketName != "" {
    			if ev.S3.Bucket.Name != bucketName {
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Fri May 24 23:05:23 UTC 2024
    - 6K bytes
    - Viewed (0)
  6. cmd/endpoint.go

    // The LOCAL host will be nil, but the indexes of all hosts should
    // remain consistent across the cluster.
    func (l EndpointServerPools) hostsSorted() []*xnet.Host {
    	peers, localPeer := l.peers()
    	sort.Strings(peers)
    	hosts := make([]*xnet.Host, len(peers))
    	for i, hostStr := range peers {
    		if hostStr == localPeer {
    			continue
    		}
    		host, err := xnet.ParseHost(hostStr)
    		if err != nil {
    			internalLogIf(GlobalContext, err)
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Fri Jun 21 22:22:24 UTC 2024
    - 34.2K bytes
    - Viewed (0)
  7. cmd/tier-handlers.go

    	case storageclass.STANDARD, storageclass.RRS:
    		writeErrorResponseJSON(ctx, w, toAdminAPIErr(ctx, errTierReservedName), r.URL)
    		return
    	}
    
    	// Refresh from the disk in case we had missed notifications about edits from peers.
    	if err := globalTierConfigMgr.Reload(ctx, objAPI); err != nil {
    		writeErrorResponseJSON(ctx, w, toAdminAPIErr(ctx, err), r.URL)
    		return
    	}
    
    	err = globalTierConfigMgr.Add(ctx, cfg, ignoreInUse)
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Thu Sep 12 20:44:05 UTC 2024
    - 7.5K bytes
    - Viewed (0)
  8. cmd/admin-handlers.go

    	// expected in case of read timeout
    	w.Header().Set("Connection", "close")
    
    	setEventStreamHeaders(w)
    
    	logCh := make(chan log.Info, 1000)
    	peers, _ := newPeerRestClients(globalEndpoints)
    	encodedCh := make(chan []byte, 1000+len(peers)*1000)
    	err = globalConsoleSys.Subscribe(logCh, ctx.Done(), node, limitLines, logKind, nil)
    	if err != nil {
    		writeErrorResponseJSON(ctx, w, toAdminAPIErr(ctx, err), r.URL)
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Fri Oct 04 11:32:32 UTC 2024
    - 99.7K bytes
    - Viewed (0)
  9. architecture/ambient/ztunnel-cni-lifecycle.md

    Additionally, eventually consistency means we will often get some final requests trickle in before all clients are aware the Pod is shutting down.
    
    Finally, we SHOULD tell our (HBONE) peers we are shutting down.
    This can help ensure that clients do not hold onto pooled connections for too long.
    Registered: Wed Nov 06 22:53:10 UTC 2024
    - Last Modified: Wed Jul 17 23:10:17 UTC 2024
    - 9.4K bytes
    - Viewed (0)
  10. cmd/admin-handlers_test.go

    	if err != nil {
    		t.Fatal("Failed to initialize a single node Erasure backend for admin handler tests.", err)
    	}
    	defer adminTestBed.TearDown()
    
    	// Initialize admin peers to make admin RPC calls. Note: In a
    	// single node setup, this degenerates to a simple function
    	// call under the hood.
    	globalMinioAddr = "127.0.0.1:9000"
    
    	var wg sync.WaitGroup
    
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Wed Jul 24 10:24:01 UTC 2024
    - 13.9K bytes
    - Viewed (0)
Back to top