- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 453 for dsync (0.03 sec)
-
cmd/lock-rest-server.go
lockRPCRUnlock = newLockHandler(grid.HandlerLockRUnlock) ) func newLockHandler(h grid.HandlerID) *grid.SingleHandler[*dsync.LockArgs, *dsync.LockResp] { return grid.NewSingleHandler[*dsync.LockArgs, *dsync.LockResp](h, func() *dsync.LockArgs { return &dsync.LockArgs{} }, func() *dsync.LockResp { return &dsync.LockResp{} }) } // registerLockRESTHandlers - register lock rest router.
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Jan 31 19:54:34 UTC 2025 - 6.3K bytes - Viewed (0) -
cmd/lock-rest-client.go
r, err := h.Call(ctx, c.connection, args) if err != nil { return false, err } defer h.PutResponse(r) ok = r.Code == dsync.RespOK switch r.Code { case dsync.RespLockConflict, dsync.RespLockNotFound, dsync.RespOK: // no error case dsync.RespLockNotInitialized: err = errLockNotInitialized
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Mon Jul 29 18:10:04 UTC 2024 - 3.3K bytes - Viewed (0) -
cmd/namespace-lock.go
// volume, path and operation ID. func (n *nsLockMap) NewNSLock(lockers func() ([]dsync.NetLocker, string), volume string, paths ...string) RWLocker { sort.Strings(paths) opsID := mustGetUUID() if n.isDistErasure { drwmutex := dsync.NewDRWMutex(&dsync.Dsync{ GetLockers: lockers, Timeouts: dsync.DefaultTimeouts, }, pathsJoinPrefix(volume, paths...)...) return &distLockInstance{drwmutex, opsID}
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sun Sep 29 22:40:36 UTC 2024 - 9.2K bytes - Viewed (0) -
docs/distributed/decom-compressed-sse-s3.sh
./mc rm -r --force myminio/versioned >/dev/null ## mirror again to create another set of version on top ./mc mirror internal myminio/versioned/ --quiet >/dev/null expected_checksum=$(./mc cat internal/dsync/drwmutex.go | md5sum) user_count=$(./mc admin user list myminio/ | wc -l) policy_count=$(./mc admin policy list myminio/ | wc -l) kill $pid
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Mon May 27 19:17:46 UTC 2024 - 4.3K bytes - Viewed (0) -
cmd/local-locker.go
package cmd //go:generate msgp -file=$GOFILE -unexported import ( "context" "fmt" "strconv" "sync" "sync/atomic" "time" "github.com/minio/minio/internal/dsync" ) // Reject new lock requests immediately when this many are queued // for the local lock mutex. // We do not block unlocking or maintenance, but they add to the count.
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 12K bytes - Viewed (0) -
cmd/lock-rest-server-common_test.go
lockRequesterInfo2, } lri := locker.ll.lockMap["name"] // test unknown uid if locker.ll.removeEntry("name", dsync.LockArgs{ Owner: "owner", UID: "unknown-uid", }, &lri) { t.Errorf("Expected %#v, got %#v", false, true) } if !locker.ll.removeEntry("name", dsync.LockArgs{ Owner: "owner", UID: "0123-4567", }, &lri) { t.Errorf("Expected %#v, got %#v", true, false) } else {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Wed Apr 09 14:28:39 UTC 2025 - 3.2K bytes - Viewed (0) -
cmd/lock-rest-client_test.go
} // Attempt all calls. _, err = lkClient.RLock(t.Context(), dsync.LockArgs{}) if err == nil { t.Fatal("Expected for Rlock to fail") } _, err = lkClient.Lock(t.Context(), dsync.LockArgs{}) if err == nil { t.Fatal("Expected for Lock to fail") } _, err = lkClient.RUnlock(t.Context(), dsync.LockArgs{}) if err == nil { t.Fatal("Expected for RUnlock to fail") }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Wed Apr 09 14:28:39 UTC 2025 - 2K bytes - Viewed (0) -
cmd/local-locker_test.go
"math/rand" "testing" "time" "github.com/google/uuid" "github.com/minio/minio/internal/dsync" ) func TestLocalLockerExpire(t *testing.T) { wResources := make([]string, 1000) rResources := make([]string, 1000) quorum := 0 l := newLocker() ctx := t.Context() for i := range wResources { arg := dsync.LockArgs{ UID: mustGetUUID(), Resources: []string{mustGetUUID()}, Source: t.Name(),
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 11.8K bytes - Viewed (0) -
docs/distributed/decom.sh
./mc rm -r --force myminio/versioned >/dev/null ## mirror again to create another set of version on top ./mc mirror internal myminio/versioned/ --quiet >/dev/null expected_checksum=$(./mc cat internal/dsync/drwmutex.go | md5sum) user_count=$(./mc admin user list myminio/ | wc -l) policy_count=$(./mc admin policy list myminio/ | wc -l) ## create a warm tier instance
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Mon May 27 19:17:46 UTC 2024 - 6.4K bytes - Viewed (0) -
docs/distributed/decom-encrypted-kes.sh
exit 1 fi got_checksum=$(./mc cat myminio/versioned/dsync/drwmutex.go | md5sum) if [ "${expected_checksum}" != "${got_checksum}" ]; then echo "BUG: decommission failed on encrypted objects: expected ${expected_checksum} got ${got_checksum}" exit 1 fi got_checksum_1=$(./mc cat myminio/versioned-1/dsync/drwmutex.go | md5sum) if [ "${expected_checksum}" != "${got_checksum_1}" ]; then
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Mon May 27 19:17:46 UTC 2024 - 7.1K bytes - Viewed (1)