- Sort Score
- Num 10 results
- Language All
Results 41 - 50 of 305 for Drivers (0.07 seconds)
-
api/go1.8.txt
pkg database/sql/driver, type IsolationLevel int pkg database/sql/driver, type NamedValue struct pkg database/sql/driver, type NamedValue struct, Name string pkg database/sql/driver, type NamedValue struct, Ordinal int pkg database/sql/driver, type NamedValue struct, Value Value pkg database/sql/driver, type Pinger interface { Ping }Created: Tue Apr 07 11:13:11 GMT 2026 - Last Modified: Wed Dec 21 05:25:57 GMT 2016 - 16.3K bytes - Click Count (0) -
docs/distributed/README.md
A stand-alone MinIO server would go down if the server hosting the drives goes offline. In contrast, a distributed MinIO setup with _m_ servers and _n_ drives will have your data safe as long as _m/2_ servers or _m*n_/2 or more drives are online. For example, an 16-server distributed setup with 200 drives per node would continue serving files, up to 4 servers can be offline in default configuration i.e around 800 drives down MinIO would continue to read and write objects.
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 8.9K bytes - Click Count (0) -
docs/erasure/storage-class/README.md
on 16 drive MinIO deployment. If you use eight data and eight parity drives, the file space usage will be approximately twice, i.e. 100 MiB file will take 200 MiB space. But, if you use ten data and six parity drives, same 100 MiB file takes around 160 MiB. If you use 14 data and two parity drives, 100 MiB file takes only approximately 114 MiB.
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 5.9K bytes - Click Count (0) -
cmd/erasure-encode_test.go
} writers := make([]io.Writer, len(disks)) for i, disk := range disks { if disk == OfflineDisk { continue } writers[i] = newBitrotWriter(disk, "", "testbucket", "object", erasure.ShardFileSize(int64(len(data[test.offset:]))), test.algorithm, erasure.ShardSize()) } n, err := erasure.Encode(t.Context(), bytes.NewReader(data[test.offset:]), writers, buffer, erasure.dataBlocks+1)
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 11.8K bytes - Click Count (0) -
cmd/peer-s3-server.go
Name: v.Name, Deleted: v.Created, }) } return true }) return buckets, nil } func cloneDrives(drives map[string]StorageAPI) []StorageAPI { copyDrives := make([]StorageAPI, 0, len(drives)) for _, drive := range drives { copyDrives = append(copyDrives, drive) } return copyDrives }
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 8.2K bytes - Click Count (0) -
ci/official/containers/ml_build/rbe_nvidia.packages.txt
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Sep 18 00:19:40 GMT 2025 - 307 bytes - Click Count (0) -
cmd/erasure-decode_test.go
buffer := make([]byte, test.blocksize, 2*test.blocksize) writers := make([]io.Writer, len(disks)) for i, disk := range disks { writers[i] = newBitrotWriter(disk, "", "testbucket", "object", erasure.ShardFileSize(test.data), writeAlgorithm, erasure.ShardSize()) } n, err := erasure.Encode(t.Context(), bytes.NewReader(data), writers, buffer, erasure.dataBlocks+1) closeBitrotWriters(writers) if err != nil {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 21K bytes - Click Count (0) -
tests/go.mod
go 1.24.0 require ( github.com/google/uuid v1.6.0 github.com/jinzhu/now v1.1.5 github.com/lib/pq v1.10.9 github.com/stretchr/testify v1.11.1 gorm.io/driver/gaussdb v0.1.0 gorm.io/driver/mysql v1.6.0 gorm.io/driver/postgres v1.6.0 gorm.io/driver/sqlite v1.6.0 gorm.io/driver/sqlserver v1.6.1 gorm.io/gorm v1.31.0 ) require ( filippo.io/edwards25519 v1.1.0 // indirect
Created: Sun Apr 05 09:35:12 GMT 2026 - Last Modified: Sun Nov 02 14:09:18 GMT 2025 - 1.3K bytes - Click Count (0) -
docs/distributed/DESIGN.md
- We limited the number of drives to 16 for erasure set because, erasure code shards more than 16 can become chatty and do not have any performance advantages. Additionally since 16 drive erasure set gives you tolerance of 8 drives per object by default which is plenty in any practical scenario.
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Wed Feb 26 09:25:50 GMT 2025 - 8K bytes - Click Count (2) -
docs/distributed/SIZING.md
If one or more drives are offline at the start of a PutObject or NewMultipartUpload operation the object will have additional data protection bits added automatically to provide the regular safety for these objects up to 50% of the number of drives. This will allow normal write operations to take place on systems that exceed the write tolerance.
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Aug 15 23:04:20 GMT 2023 - 3.9K bytes - Click Count (0)