- Sort Score
- Num 10 results
- Language All
Results 31 - 40 of 293 for Drivers (0.04 seconds)
-
CHANGELOG/CHANGELOG-1.30.md
- The in-tree cloud provider for vSphere has been deprecated and removed. Users are advised to utilize the external cloud provider and CSI driver available at https://github.com/kubernetes/cloud-provider-vsphere. ([#122937](https://...
Created: Fri Dec 26 09:05:12 GMT 2025 - Last Modified: Wed Jun 18 18:59:10 GMT 2025 - 398.1K bytes - Click Count (0) -
cmd/erasure-decode.go
} return bytesWritten, derr } // Heal reads from readers, reconstruct shards and writes the data to the writers. func (e Erasure) Heal(ctx context.Context, writers []io.Writer, readers []io.ReaderAt, totalLength int64, prefer []bool) (derr error) { if len(writers) != e.parityBlocks+e.dataBlocks { return errInvalidArgument } reader := newParallelReader(readers, e, 0, totalLength)
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Thu Aug 29 01:40:52 GMT 2024 - 9.5K bytes - Click Count (0) -
cmd/sftp-server-driver.go
Klaus Post <******@****.***> 1739205349 -0800
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Mon Feb 10 16:35:49 GMT 2025 - 11.6K bytes - Click Count (0) -
cmd/erasure-multipart.go
} if len(buffer) > int(fi.Erasure.BlockSize) { buffer = buffer[:fi.Erasure.BlockSize] } writers := make([]io.Writer, len(onlineDisks)) for i, disk := range onlineDisks { if disk == nil { continue } writers[i] = newBitrotWriter(disk, bucket, minioMetaTmpBucket, tmpPartPath, erasure.ShardFileSize(data.Size()), DefaultBitrotAlgorithm, erasure.ShardSize()) }
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Oct 24 04:05:31 GMT 2025 - 47.1K bytes - Click Count (0) -
cmd/format-erasure.go
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 23.1K bytes - Click Count (0) -
cmd/erasure-encode.go
} // Encode reads from the reader, erasure-encodes the data and writes to the writers. func (e *Erasure) Encode(ctx context.Context, src io.Reader, writers []io.Writer, buf []byte, quorum int) (total int64, err error) { writer := &multiWriter{ writers: writers, writeQuorum: quorum, errs: make([]error, len(writers)), } for { var blocks [][]byte
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Wed May 15 00:11:04 GMT 2024 - 2.9K bytes - Click Count (0) -
cmd/erasure-object.go
inlineBuffers[i] = bytes.NewBuffer(buf[:0]) defer grid.PutByteBuffer(buf) writers[i] = newStreamingBitrotWriterBuffer(inlineBuffers[i], DefaultBitrotAlgorithm, erasure.ShardSize()) } } n, erasureErr := erasure.Encode(ctx, data, writers, buffer, writeQuorum) closeErrs := closeBitrotWriters(writers) if erasureErr != nil { return ObjectInfo{}, toObjectErr(erasureErr, minioMetaBucket, key) }
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Oct 24 04:05:31 GMT 2025 - 80.4K bytes - Click Count (0) -
api/go1.8.txt
pkg database/sql/driver, type IsolationLevel int pkg database/sql/driver, type NamedValue struct pkg database/sql/driver, type NamedValue struct, Name string pkg database/sql/driver, type NamedValue struct, Ordinal int pkg database/sql/driver, type NamedValue struct, Value Value pkg database/sql/driver, type Pinger interface { Ping }Created: Tue Dec 30 11:13:12 GMT 2025 - Last Modified: Wed Dec 21 05:25:57 GMT 2016 - 16.3K bytes - Click Count (0) -
docs/erasure/storage-class/README.md
on 16 drive MinIO deployment. If you use eight data and eight parity drives, the file space usage will be approximately twice, i.e. 100 MiB file will take 200 MiB space. But, if you use ten data and six parity drives, same 100 MiB file takes around 160 MiB. If you use 14 data and two parity drives, 100 MiB file takes only approximately 114 MiB.
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 5.9K bytes - Click Count (0) -
docs/distributed/README.md
A stand-alone MinIO server would go down if the server hosting the drives goes offline. In contrast, a distributed MinIO setup with _m_ servers and _n_ drives will have your data safe as long as _m/2_ servers or _m*n_/2 or more drives are online. For example, an 16-server distributed setup with 200 drives per node would continue serving files, up to 4 servers can be offline in default configuration i.e around 800 drives down MinIO would continue to read and write objects.
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 8.9K bytes - Click Count (0)