- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 410 for Make (0.01 seconds)
-
internal/grid/grid_test.go
errFatal := func(err error) { t.Helper() if err != nil { t.Fatal(err) } } // We fake a local and remote server. remoteHost := remote.HostName() // 1: Echo serverSent := make(chan struct{}) serverCanceled := make(chan struct{}) register := func(manager *Manager) { errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 36.1K bytes - Click Count (0) -
cmd/peer-s3-client.go
return nil }, idx) } errs := g.Wait() // The list of buckets in a map to avoid duplication resultMap := make(map[string]BucketInfo) for poolIdx := 0; poolIdx < sys.poolsCount; poolIdx++ { perPoolErrs := make([]error, 0, len(sys.peerClients)) for i, client := range sys.peerClients { if slices.Contains(client.GetPools(), poolIdx) {
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 15.6K bytes - Click Count (0) -
cmd/admin-heal-ops.go
Summary: healNotStartedStatus, HealSettings: hs, }, traverseAndHealDoneCh: make(chan error), cancelCtx: cancel, ctx: ctx, scannedItemsMap: make(map[madmin.HealItemType]int64), healedItemsMap: make(map[madmin.HealItemType]int64), healFailedItemsMap: make(map[madmin.HealItemType]int64), } }
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 25.4K bytes - Click Count (0) -
src/cmd/asm/internal/arch/arch.go
return false } func jumpWasm(word string) bool { return word == "JMP" || word == "CALL" || word == "Call" || word == "Br" || word == "BrIf" } func archX86(linkArch *obj.LinkArch) *Arch { register := make(map[string]int16) // Create maps for easy lookup of instruction names etc. for i, s := range x86.Register { register[s] = int16(i + x86.REG_AL) } // Pseudo-registers. register["SB"] = RSB
Created: Tue Dec 30 11:13:12 GMT 2025 - Last Modified: Thu Nov 13 12:17:37 GMT 2025 - 21.7K bytes - Click Count (0) -
cmd/erasure-metadata-utils.go
func shuffleDisksAndPartsMetadataByIndex(disks []StorageAPI, metaArr []FileInfo, fi FileInfo) (shuffledDisks []StorageAPI, shuffledPartsMetadata []FileInfo) { shuffledDisks = make([]StorageAPI, len(disks)) shuffledPartsMetadata = make([]FileInfo, len(disks)) distribution := fi.Erasure.Distribution var inconsistent int for i, meta := range metaArr { if disks[i] == nil {
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 11.7K bytes - Click Count (0) -
cmd/erasure-sets.go
endpointStrings := make([]string, len(endpoints.Endpoints)) for i, endpoint := range endpoints.Endpoints { endpointStrings[i] = endpoint.String() } // Initialize the erasure sets instance. s := &erasureSets{ sets: make([]*erasureObjects, setCount), erasureDisks: make([][]StorageAPI, setCount), erasureLockers: make([][]dsync.NetLocker, setCount),
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 37K bytes - Click Count (1) -
internal/s3select/csv/reader.go
} // Create queue r.queue = make(chan *queueItem, runtime.GOMAXPROCS(0)) r.input = make(chan *queueItem, runtime.GOMAXPROCS(0)) r.readerWg.Add(1) // Start splitter go func() { defer close(r.input) defer close(r.queue) defer r.readerWg.Done() for { q := queueItem{ input: next, dst: make(chan [][]string, 1), err: nextErr, } select {
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue May 27 15:19:03 GMT 2025 - 8.8K bytes - Click Count (0) -
docs/debugging/xl-meta/main.go
return errors.New("cannot combine --data and --combine") } // file / version / file filemap := make(map[string]map[string]string) foundData := make(map[string][]byte) partDataToVerID := make(map[string][2]string) var baseName string // versionID -> combineFiles := make(map[string][]string) decode := func(r io.Reader, file string) ([]byte, error) { file = strings.ReplaceAll(file, ":", "_")
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 40.4K bytes - Click Count (0) -
cmd/peer-s3-server.go
globalLocalDrivesMu.RUnlock() // Initialize sync waitgroup. g := errgroup.WithNErrs(len(localDrives)) // Disk states slices beforeState := make([]string, len(localDrives)) afterState := make([]string, len(localDrives)) // Make a volume entry on all underlying storage disks. for index := range localDrives { g.Go(func() (serr error) { if localDrives[index] == nil {
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 8.2K bytes - Click Count (0) -
cmd/batch-expire.go
} if i > 0 { if wait := globalBatchConfig.ExpirationWait(); wait > 0 { time.Sleep(wait) } } i++ wk.Take() go func(toExpire []expireObjInfo) { defer wk.Give() toExpireAll := make([]expireObjInfo, 0, len(toExpire)) toDel := make([]ObjectToDelete, 0, len(toExpire)) oiCache := newObjInfoCache() for _, exp := range toExpire { if exp.ExpireAll {Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Apr 22 11:16:32 GMT 2025 - 23K bytes - Click Count (0)