- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 71 for Shard (0.04 sec)
-
docs/debugging/xl-meta/main.go
} // Update to new config and add current valid := 0 for shardIdx, shard := range splitFilled[:k] { shardConfig[shardIdx] = shard[offset] valid += int(shard[offset]) if shard[offset] == 0 { shards[shardIdx] = shards[shardIdx][:0] } else {
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Thu Sep 05 11:57:44 UTC 2024 - 40.3K bytes - Viewed (0) -
guava-tests/test/com/google/common/hash/HashingTest.java
Random r = new Random(9); for (int i = 0; i < ITERS; i++) { countRemaps(r.nextLong(), map); } for (int shard = 2; shard <= MAX_SHARDS; shard++) { // Rough: don't exceed 1.2x the expected number of remaps by more than 20 assertTrue(map.get(shard) <= 1.2 * ITERS / shard + 20); } } private void countRemaps(long h, AtomicLongMap<Integer> map) { int last = 0;
Registered: Fri Nov 01 12:43:10 UTC 2024 - Last Modified: Tue Jul 09 17:40:09 UTC 2024 - 26.3K bytes - Viewed (0) -
internal/config/storageclass/storage-class.go
StandardEnv = "MINIO_STORAGE_CLASS_STANDARD" // Optimize storage class environment variable OptimizeEnv = "MINIO_STORAGE_CLASS_OPTIMIZE" // Inline block indicates the size of the shard // that is considered for inlining, remember this // shard value is the value per drive shard it // will vary based on the parity that is configured // for the STANDARD storage_class. // inlining means data and metadata are written
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Thu Sep 12 12:24:04 UTC 2024 - 12.3K bytes - Viewed (0) -
cmd/bitrot-whole.go
"hash" "io" ) // Implementation to calculate bitrot for the whole file. type wholeBitrotWriter struct { disk StorageAPI volume string filePath string shardSize int64 // This is the shard size of the erasure logic hash.Hash // For bitrot hash } func (b *wholeBitrotWriter) Write(p []byte) (int, error) { err := b.disk.AppendFile(context.TODO(), b.volume, b.filePath, p) if err != nil {
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Wed Jan 31 02:11:45 UTC 2024 - 2.7K bytes - Viewed (0) -
cmd/erasure-decode.go
// Reading first time on this disk, hence the buffer needs to be allocated. // Subsequent reads will reuse this buffer. p.buf[bufIdx] = make([]byte, p.shardSize) } // For the last shard, the shardsize might be less than previous shard sizes. // Hence the following statement ensures that the buffer size is reset to the right size. p.buf[bufIdx] = p.buf[bufIdx][:p.shardSize] n, err := rr.ReadAt(p.buf[bufIdx], p.offset)
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Thu Aug 29 01:40:52 UTC 2024 - 9.5K bytes - Viewed (0) -
cmd/erasure-coding.go
fmt.Fprintf(os.Stderr, "%v: error on self-test [d:%d,p:%d]: want %#v, got %#v\n", algo, conf[0], conf[1], a, b) ok = false continue } // Delete first shard and reconstruct... first := encoded[0] encoded[0] = nil failOnErr(e.DecodeDataBlocks(encoded)) if a, b := first, encoded[0]; !bytes.Equal(a, b) {
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Wed Jan 31 02:11:45 UTC 2024 - 8.6K bytes - Viewed (0) -
src/main/config/openapi/openapi-user.yaml
required: false schema: type: string example: en - name: preference in: query description: String to specify a shard for searching required: false schema: type: string example: abc - name: callback in: query description: Callback name for using JSONP
Registered: Thu Oct 31 13:40:30 UTC 2024 - Last Modified: Thu May 09 06:31:27 UTC 2024 - 21.6K bytes - Viewed (0) -
istioctl/pkg/multixds/gather.go
if len(responses) == 0 { return &retval, nil } for _, response := range responses { // Combine all the shards as one, even if that means losing information about // the control plane version from each shard. retval.ControlPlane = response.ControlPlane retval.Resources = append(retval.Resources, response.Resources...) } return &retval, nil } func makeSan(istioNamespace, revision string) string {
Registered: Wed Nov 06 22:53:10 UTC 2024 - Last Modified: Wed Oct 09 16:05:45 UTC 2024 - 13.6K bytes - Viewed (0) -
cmd/xl-storage_test.go
shardSize := int64(1024 * 1024) shard := make([]byte, shardSize) w := newStreamingBitrotWriter(storage, "", volName, fileName, size, algo, shardSize) reader := bytes.NewReader(data) for { // Using io.Copy instead of this loop will not work for us as io.Copy // will use bytes.Reader.WriteTo() which will not do shardSize'ed writes // causing error. n, err := reader.Read(shard) w.Write(shard[:n])
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Wed Aug 14 17:11:51 UTC 2024 - 66.7K bytes - Viewed (0) -
cmd/notification.go
// To avoid these problems we must split the work at scale. With 1000 node // setup becoming a reality we must try to shard the work properly such as // pick 10 nodes that precisely can send those 100 requests the first node // in the 10 node shard would coordinate between other 9 shards to get the // rest of the `99*9` requests. //
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Mon Sep 09 16:58:30 UTC 2024 - 46.2K bytes - Viewed (0)