- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 76 for storageapi (0.2 sec)
-
cmd/erasure-metadata-utils.go
// NewMultipartUpload metadata shuffling. func shuffleDisksAndPartsMetadataByIndex(disks []StorageAPI, metaArr []FileInfo, fi FileInfo) (shuffledDisks []StorageAPI, shuffledPartsMetadata []FileInfo) { shuffledDisks = make([]StorageAPI, len(disks)) shuffledPartsMetadata = make([]FileInfo, len(disks)) distribution := fi.Erasure.Distribution var inconsistent int
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 11.7K bytes - Viewed (0) -
pkg/registry/storage/csinode/storage/storage.go
store := &genericregistry.Store{ NewFunc: func() runtime.Object { return &storageapi.CSINode{} }, NewListFunc: func() runtime.Object { return &storageapi.CSINodeList{} }, DefaultQualifiedResource: storageapi.Resource("csinodes"), SingularQualifiedResource: storageapi.Resource("csinode"), CreateStrategy: csinode.Strategy, UpdateStrategy: csinode.Strategy,
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Nov 18 09:21:13 UTC 2022 - 2.1K bytes - Viewed (0) -
pkg/registry/storage/csistoragecapacity/storage/storage_test.go
etcd3testing "k8s.io/apiserver/pkg/storage/etcd3/testing" storageapi "k8s.io/kubernetes/pkg/apis/storage" _ "k8s.io/kubernetes/pkg/apis/storage/install" "k8s.io/kubernetes/pkg/registry/registrytest" ) func newStorage(t *testing.T) (*REST, *etcd3testing.EtcdTestServer) { etcdStorage, server := registrytest.NewEtcdStorageForResource(t, storageapi.SchemeGroupVersion.WithResource("csistoragecapacities").GroupResource())
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Jul 03 19:54:14 UTC 2020 - 4.4K bytes - Viewed (0) -
pkg/registry/storage/volumeattributesclass/storage/storage_test.go
etcd3testing "k8s.io/apiserver/pkg/storage/etcd3/testing" storageapi "k8s.io/kubernetes/pkg/apis/storage" "k8s.io/kubernetes/pkg/registry/registrytest" ) func newStorage(t *testing.T) (*REST, *etcd3testing.EtcdTestServer) { etcdStorage, server := registrytest.NewEtcdStorageForResource(t, storageapi.SchemeGroupVersion.WithResource("volumeattributesclasses").GroupResource()) restOptions := generic.RESTOptions{
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Oct 31 03:18:56 UTC 2023 - 4.3K bytes - Viewed (0) -
cmd/bitrot-whole.go
} // Returns whole-file bitrot writer. func newWholeBitrotWriter(disk StorageAPI, volume, filePath string, algo BitrotAlgorithm, shardSize int64) io.WriteCloser { return &wholeBitrotWriter{disk, volume, filePath, shardSize, algo.New()} } // Implementation to verify bitrot for the whole file. type wholeBitrotReader struct { disk StorageAPI volume string filePath string
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Wed Jan 31 02:11:45 UTC 2024 - 2.7K bytes - Viewed (0) -
cmd/erasure-healing-common.go
// are set to nil. // - latest (in time) of the maximally occurring modTime(s), which has at least quorum occurrences. func listOnlineDisks(disks []StorageAPI, partsMetadata []FileInfo, errs []error, quorum int) (onlineDisks []StorageAPI, modTime time.Time, etag string) { onlineDisks = make([]StorageAPI, len(disks)) // List all the file commit ids from parts metadata. modTimes := listObjectModtimes(partsMetadata, errs)
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon Jun 10 15:51:27 UTC 2024 - 12.5K bytes - Viewed (0) -
cmd/erasure.go
// OfflineDisk represents an unavailable disk. var OfflineDisk StorageAPI // zero value is nil // erasureObjects - Implements ER object layer. type erasureObjects struct { setDriveCount int defaultParityCount int setIndex int poolIndex int // getDisks returns list of storageAPIs. getDisks func() []StorageAPI // getLockers returns list of remote and local lockers.
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 16K bytes - Viewed (0) -
cmd/storage-rest_test.go
) // Storage REST server, storageRESTReceiver and StorageRESTClient are // inter-dependent, below test functions are sufficient to test all of them. func testStorageAPIDiskInfo(t *testing.T, storage StorageAPI) { testCases := []struct { expectErr bool }{ {true}, } for i, testCase := range testCases { _, err := storage.DiskInfo(context.Background(), DiskInfoOptions{Metrics: true})
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 11.4K bytes - Viewed (0) -
cmd/erasure-common.go
package cmd import ( "context" "fmt" "io" "math/rand" "sync" "time" "github.com/minio/pkg/v3/sync/errgroup" ) func (er erasureObjects) getOnlineDisks() (newDisks []StorageAPI) { disks := er.getDisks() var wg sync.WaitGroup var mu sync.Mutex r := rand.New(rand.NewSource(time.Now().UnixNano())) for _, i := range r.Perm(len(disks)) { i := i wg.Add(1) go func() {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 4.6K bytes - Viewed (0) -
cmd/erasure-object_test.go
t.Fatalf("Test %d: Failed to upload a random object: %v", i+1, err) } // Step 3: Upload the object with some disks offline sets.erasureDisksMu.Lock() xl.getDisks = func() []StorageAPI { disks := make([]StorageAPI, len(origErasureDisks)) copy(disks, origErasureDisks) disks[0] = nil disks[1] = nil return disks } sets.erasureDisksMu.Unlock()
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Tue Jan 30 20:43:25 UTC 2024 - 36.8K bytes - Viewed (0)