- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 11 for Danish (0.2 sec)
-
cmd/erasure-object.go
if err != nil { return err } diskVersions[index] = resp.Sign dataDirs[index] = resp.OldDataDir return nil }, index) } // Wait for all renames to finish. errs := g.Wait() err := reduceWriteQuorumErrs(ctx, errs, objectOpIgnoredErrs, writeQuorum) if err != nil { dg := errgroup.WithNErrs(len(disks)) for index, nerr := range errs {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Sun May 05 16:56:21 GMT 2024 - 77.2K bytes - Viewed (2) -
internal/logger/target/kafka/kafka.go
// stop the replay go-routine. if h.store != nil { h.storeCtxCancel() } // Set logch to nil and close it. // This will block all Send operations, // and finish the existing ones. // All future ones will be discarded. h.logChMu.Lock() xioutil.SafeClose(h.logCh) h.logCh = nil h.logChMu.Unlock() if h.producer != nil { h.producer.Close()
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Feb 22 06:26:06 GMT 2024 - 10.1K bytes - Viewed (1) -
cmd/format-erasure.go
g.Go(func() error { if formats[index] == nil { return errDiskNotFound } return saveFormatErasure(storageDisks[index], formats[index], "") }, index) } // Wait for the routines to finish. return reduceWriteQuorumErrs(ctx, g.Wait(), nil, len(storageDisks)) } // relinquishes the underlying connection for all storage disks. func closeStorageDisks(storageDisks ...StorageAPI) { var wg sync.WaitGroup
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Fri May 03 15:54:03 GMT 2024 - 23.2K bytes - Viewed (0) -
cmd/admin-heal-ops.go
// time-duration to keep heal sequence state after it // completes. keepHealSeqStateDuration = time.Minute * 10 // nopHeal is a no operating healing action to // wait for the current healing operation to finish nopHeal = "" ) var ( errHealIdleTimeout = fmt.Errorf("healing results were not consumed for too long") errHealStopSignalled = fmt.Errorf("heal stop signaled")
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Thu Apr 04 12:04:40 GMT 2024 - 25.1K bytes - Viewed (1) -
docs/distributed/decom-encrypted-sse-s3.sh
./mc admin decom start myminio/ http://localhost:9000/tmp/xl/{1...10}/disk{0...1} until $(./mc admin decom status myminio/ | grep -q Complete); do echo "waiting for decom to finish..." sleep 1 done kill $pid_1 kill $pid_2 sleep 5 (minio server --address ":9001" http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/removed.log) & pid=$! sleep 30
Shell Script - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Tue Apr 23 17:15:52 GMT 2024 - 4.5K bytes - Viewed (0) -
cmd/erasure-multipart.go
g.Go(func() error { if disks[index] == nil { return errDiskNotFound } return disks[index].RenameFile(ctx, srcBucket, srcEntry, dstBucket, dstEntry) }, index) } // Wait for all renames to finish. errs := g.Wait() // Do not need to undo partial successful operation since those will be cleaned up // in 24hrs via multipart cleaner, never rename() back to `.minio.sys/tmp` as there // is no way to clean them.
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Sun Apr 28 17:53:50 GMT 2024 - 43K bytes - Viewed (0) -
cmd/metacache-set.go
func listPathRaw(ctx context.Context, opts listPathRawOptions) (err error) { disks := opts.disks if len(disks) == 0 { return fmt.Errorf("listPathRaw: 0 drives provided") } // Cancel upstream if we finish before we expect. ctx, cancel := context.WithCancel(ctx) defer cancel() // Keep track of fallback disks var fdMu sync.Mutex fds := opts.fallbackDisks fallback := func(err error) StorageAPI {
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Wed May 01 17:59:08 GMT 2024 - 30.4K bytes - Viewed (0) -
internal/logger/target/http/http.go
atomic.StoreInt32(&h.status, statusClosed) h.storeCtxCancel() // Wait for messages to be sent... h.wg.Wait() // Set logch to nil and close it. // This will block all Send operations, // and finish the existing ones. // All future ones will be discarded. h.logChMu.Lock() xioutil.SafeClose(h.logCh) h.logCh = nil h.logChMu.Unlock()
Go - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Mon Mar 25 16:44:20 GMT 2024 - 14.9K bytes - Viewed (0) -
docs/distributed/decom-compressed-sse-s3.sh
./mc admin decom start myminio/ http://localhost:9000/tmp/xl/{1...10}/disk{0...1} until $(./mc admin decom status myminio/ | grep -q Complete); do echo "waiting for decom to finish..." sleep 1 done kill $pid_1 kill $pid_2 sleep 5 (minio server --address ":9001" http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/removed.log) & pid=$! sleep 30
Shell Script - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Tue Apr 23 17:15:52 GMT 2024 - 4.4K bytes - Viewed (0) -
docs/distributed/decom-encrypted.sh
./mc admin decom start myminio/ http://localhost:9000/tmp/xl/{1...10}/disk{0...1} until $(./mc admin decom status myminio/ | grep -q Complete); do echo "waiting for decom to finish..." sleep 1 done kill $pid_1 kill $pid_2 sleep 5 (minio server --address ":9001" http://localhost:9001/tmp/xl/{11...30}/disk{0...3} 2>&1 >/tmp/removed.log) & pid=$! sleep 30
Shell Script - Registered: Sun May 05 19:28:20 GMT 2024 - Last Modified: Tue Apr 23 17:15:52 GMT 2024 - 4.2K bytes - Viewed (0)