- Sort Score
- Num 10 results
- Language All
Results 51 - 60 of 141 for coroutines (0.08 seconds)
-
src/builtin/builtin.go
// panicking goroutine. Executing a call to recover inside a deferred // function (but not any function called by it) stops the panicking sequence // by restoring normal execution and retrieves the error value passed to the // call of panic. If recover is called outside the deferred function it will // not stop a panicking sequence. In this case, or when the goroutine is not // panicking, recover returns nil. //
Created: Tue Apr 07 11:13:11 GMT 2026 - Last Modified: Thu Feb 26 17:14:40 GMT 2026 - 13.2K bytes - Click Count (0) -
cmd/signals.go
} os.Exit(1) } stopProcess := func() bool { shutdownHealMRFWithTimeout() // this can take time sometimes, it needs to be executed // before stopping s3 operations // send signal to various go-routines that they need to quit. cancelGlobalContext() if httpServer := newHTTPServerFn(); httpServer != nil { if err := httpServer.Shutdown(); err != nil && !errors.Is(err, http.ErrServerClosed) {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Wed Sep 04 17:02:39 GMT 2024 - 3.2K bytes - Click Count (0) -
cmd/shared-lock.go
ctx, cancel := context.WithCancel(context.Background()) go func() { select { case <-ctx1.Done(): case <-ctx2.Done(): // The lock acquirer decides to cancel, exit this goroutine case <-ctx.Done(): } cancel() }() return ctx, cancel } func (ld sharedLock) GetLock(ctx context.Context) (context.Context, context.CancelFunc) { l := <-ld.lockContext
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Mon Feb 13 09:26:38 GMT 2023 - 2.3K bytes - Click Count (0) -
src/arena/arena.go
// it is no longer referenced, so it must be kept alive (see runtime.KeepAlive) // until any memory allocated from it is no longer needed. // // An Arena must never be used concurrently by multiple goroutines. type Arena struct { a unsafe.Pointer } // NewArena allocates a new arena. func NewArena() *Arena { return &Arena{a: runtime_arena_newArena()} }
Created: Tue Apr 07 11:13:11 GMT 2026 - Last Modified: Wed Oct 12 20:23:36 GMT 2022 - 4.3K bytes - Click Count (0) -
internal/s3select/progress.go
closedMu sync.Mutex closer io.ReadCloser closed bool } func (pr *progressReader) Read(p []byte) (n int, err error) { // This ensures that Close will block until Read has completed. // This allows another goroutine to close the reader. pr.closedMu.Lock() defer pr.closedMu.Unlock() if pr.closed { return 0, errors.New("progressReader: read after Close") } return pr.processedReader.Read(p) }
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 22 00:33:43 GMT 2024 - 4.3K bytes - Click Count (0) -
internal/dsync/drwmutex_test.go
} // Wait for all parallel RLock()s to succeed. for range numReaders { <-clocked } for range numReaders { cunlock <- true } // Wait for the goroutines to finish. for range numReaders { <-cdone } } // Borrowed from rwmutex_test.go func TestParallelReaders(t *testing.T) { defer runtime.GOMAXPROCS(runtime.GOMAXPROCS(-1))
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 9.4K bytes - Click Count (0) -
cmd/namespace-lock_test.go
// lk2 lk2ch := make(chan struct{}) go func() { defer close(lk2ch) nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 1*time.Millisecond) }() time.Sleep(1 * time.Millisecond) // wait for goroutine to advance; ref=2 // Unlock the 1st lock; ref=1 after this line nsLk.unlock("volume", "path", false) // Taking another lockMapMutex here allows queuing up additional lockers. This should
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 3K bytes - Click Count (0) -
internal/s3select/json/preader.go
const jsonSplitSize = 128 << 10 // startReaders will read the header if needed and spin up a parser // and a number of workers based on GOMAXPROCS. // If an error is returned no goroutines have been started and r.err will have been set. func (r *PReader) startReaders() { r.bufferPool.New = func() []byte { return make([]byte, jsonSplitSize+1024) } // Create queue
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue May 27 15:19:03 GMT 2025 - 6.5K bytes - Click Count (0) -
internal/rest/client.go
sleep := min(unit*time.Duration(1<<attempt), maxSleep) sleep -= time.Duration(r.Float64() * float64(sleep-unit)) return sleep } } func (c *Client) runHealthCheck() bool { // Start goroutine that will attempt to reconnect. // If server is already trying to reconnect this will have no effect. if c.HealthCheckFn != nil && atomic.CompareAndSwapInt32(&c.connected, online, offline) { go func() {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 14.7K bytes - Click Count (0) -
internal/lsync/lrwmutex_test.go
} // Wait for all parallel RLock()s to succeed. for range numReaders { <-clocked } for range numReaders { cunlock <- true } // Wait for the goroutines to finish. for range numReaders { <-cdone } } // Borrowed from rwmutex_test.go func TestParallelReaders(t *testing.T) { defer runtime.GOMAXPROCS(runtime.GOMAXPROCS(-1))
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 7.8K bytes - Click Count (0)