- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 258 for Chan (0.26 sec)
-
internal/grid/muxserver.go
if debugPrint { fmt.Println("connected stream mux:", ack.MuxID) } }() // Data inbound to the handler var handlerIn chan []byte if inboundCap > 0 { m.inbound = make(chan []byte, inboundCap) handlerIn = make(chan []byte, 1) go func(inbound chan []byte) { wg.Wait() defer xioutil.SafeClose(handlerIn) m.handleInbound(c, inbound, handlerIn) }(m.inbound) }
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 9.1K bytes - Viewed (0) -
internal/grid/handlers.go
func (h *StreamTypeHandler[Payload, Req, Resp]) RegisterNoPayload(m *Manager, handle func(ctx context.Context, in <-chan Req, out chan<- Resp) *RemoteErr, subroute ...string) error { h.WithPayload = false return h.register(m, func(ctx context.Context, p Payload, in <-chan Req, out chan<- Resp) *RemoteErr { return handle(ctx, in, out) }, subroute...) }
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 26.9K bytes - Viewed (0) -
cmd/peer-rest-client.go
} func (client *peerRESTClient) GetResourceMetrics(ctx context.Context) (<-chan MetricV2, error) { resp, err := getResourceMetricsRPC.Call(ctx, client.gridConn(), grid.NewMSS()) if err != nil { return nil, err } ch := make(chan MetricV2) go func(ch chan<- MetricV2) { defer close(ch) for _, m := range resp.Value() { if m == nil { continue }
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 25.8K bytes - Viewed (0) -
internal/grid/muxclient.go
func (m *muxClient) handleTwowayResponses(responseCh chan<- Response, internalResp <-chan Response) { defer m.parent.deleteMux(false, m.MuxID) defer xioutil.SafeClose(responseCh) for resp := range internalResp { responseCh <- resp m.send(message{Op: OpUnblockSrvMux, MuxID: m.MuxID}) } } func (m *muxClient) handleTwowayRequests(internalResp chan<- Response, requests <-chan []byte) { var errState bool if debugPrint {
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 14.2K bytes - Viewed (0) -
cmd/notification.go
} consolidatedReport.BucketStats[opts] = d } } return consolidatedReport } func (sys *NotificationSys) collectPeerMetrics(ctx context.Context, peerChannels []<-chan MetricV2, g *errgroup.Group) <-chan MetricV2 { ch := make(chan MetricV2) var wg sync.WaitGroup for index, err := range g.Wait() { if err != nil { if sys.peerClients[index] != nil {
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Fri Apr 12 18:13:36 GMT 2024 - 44.5K bytes - Viewed (0) -
cmd/handler-api.go
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/mcontext" ) type apiConfig struct { mu sync.RWMutex requestsDeadline time.Duration requestsPool chan struct{} clusterDeadline time.Duration listQuorum string corsAllowOrigins []string // total drives per erasure set across pools. totalDriveCount int replicationPriority string
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Mon Apr 08 09:22:27 GMT 2024 - 10K bytes - Viewed (0) -
cmd/bucket-replication.go
mrfMU sync.Mutex resyncer *replicationResyncer // workers: workers []chan ReplicationWorkerOperation lrgworkers []chan ReplicationWorkerOperation // mrf: mrfWorkerKillCh chan struct{} mrfReplicaCh chan ReplicationWorkerOperation mrfSaveCh chan MRFReplicateEntry mrfStopCh chan struct{} mrfWorkerSize int }
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Sat Apr 20 09:05:54 GMT 2024 - 112K bytes - Viewed (1) -
internal/grid/connection.go
if handler == nil { return nil, ErrUnknownHandler } var requests chan []byte var responses chan Response if handler.InCapacity > 0 { requests = make(chan []byte, handler.InCapacity) } if handler.OutCapacity > 0 { responses = make(chan Response, handler.OutCapacity) } else { responses = make(chan Response, 1) } cl, err := c.newMuxClient(ctx) if err != nil {
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 42.6K bytes - Viewed (0) -
internal/bpool/bpool.go
type BytePoolCap struct { c chan []byte w int wcap int } // NewBytePoolCap creates a new BytePool bounded to the given maxSize, with new // byte arrays sized based on width. func NewBytePoolCap(maxSize uint64, width int, capwidth int) (bp *BytePoolCap) { if capwidth > 0 && capwidth < 64 { panic("buffer capped with smaller than 64 bytes is not supported") }
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Fri Apr 19 16:44:59 GMT 2024 - 2.6K bytes - Viewed (0) -
cmd/metacache-server-pool.go
saver := z.serverPools[o.pool].sets[o.set] // Disconnect from call above, but cancel on exit. listCtx, cancel := context.WithCancel(GlobalContext) saveCh := make(chan metaCacheEntry, metacacheBlockSize) inCh := make(chan metaCacheEntry, metacacheBlockSize) outCh := make(chan metaCacheEntry, o.Limit) filteredResults := o.gatherResults(ctx, outCh) mc := o.newMetacache()
Go - Registered: Sun Apr 21 19:28:08 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 12.8K bytes - Viewed (0)