- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 93 for lockers (0.06 sec)
-
cmd/namespace-lock.go
} nsLk.ref++ n.lockMap[resource] = nsLk n.lockMapMutex.Unlock() // Locking here will block (until timeout). if readLock { locked = nsLk.GetRLock(ctx, opsID, lockSource, timeout) } else { locked = nsLk.GetLock(ctx, opsID, lockSource, timeout) } if !locked { // We failed to get the lock // Decrement ref count since we failed to get the lock n.lockMapMutex.Lock() n.lockMap[resource].ref--
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sun Sep 29 22:40:36 UTC 2024 - 9.2K bytes - Viewed (0) -
cmd/namespace-lock_test.go
// Unlock the 1st lock; ref=1 after this line nsLk.unlock("volume", "path", false) // Taking another lockMapMutex here allows queuing up additional lockers. This should // not be required but makes reproduction much easier. nsLk.lockMapMutex.Lock() // lk3 blocks. lk3ch := make(chan bool) go func() {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 3K bytes - Viewed (0) -
cmd/lock-rest-client.go
func (c *lockRESTClient) String() string { return c.connection.Remote } func (c *lockRESTClient) call(ctx context.Context, h *grid.SingleHandler[*dsync.LockArgs, *dsync.LockResp], args *dsync.LockArgs) (ok bool, err error) { r, err := h.Call(ctx, c.connection, args) if err != nil { return false, err } defer h.PutResponse(r) ok = r.Code == dsync.RespOK switch r.Code {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Mon Jul 29 18:10:04 UTC 2024 - 3.3K bytes - Viewed (0) -
docs/pt/docs/deployment/server-workers.md
/// info | Informação Se você estiver usando contêineres, por exemplo com Docker ou Kubernetes, falarei mais sobre isso no próximo capítulo: [FastAPI em contêineres - Docker](docker.md){.internal-link target=_blank}.
Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Thu Jan 09 20:41:07 UTC 2025 - 8.5K bytes - Viewed (0) -
cmd/lock-rest-server-common_test.go
TimeLastRefresh: UTCNow().UnixNano(), } locker.ll.lockMap["name"] = []lockRequesterInfo{ lockRequesterInfo1, lockRequesterInfo2, } lri := locker.ll.lockMap["name"] // test unknown uid if locker.ll.removeEntry("name", dsync.LockArgs{ Owner: "owner", UID: "unknown-uid", }, &lri) { t.Errorf("Expected %#v, got %#v", false, true) } if !locker.ll.removeEntry("name", dsync.LockArgs{
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Wed Apr 09 14:28:39 UTC 2025 - 3.2K bytes - Viewed (0) -
cmd/lock-rest-server.go
) func newLockHandler(h grid.HandlerID) *grid.SingleHandler[*dsync.LockArgs, *dsync.LockResp] { return grid.NewSingleHandler[*dsync.LockArgs, *dsync.LockResp](h, func() *dsync.LockArgs { return &dsync.LockArgs{} }, func() *dsync.LockResp { return &dsync.LockResp{} }) } // registerLockRESTHandlers - register lock rest router.
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Jan 31 19:54:34 UTC 2025 - 6.3K bytes - Viewed (0) -
docs/orchestration/docker-compose/docker-compose.yaml
# environment: # MINIO_ROOT_USER: minioadmin # MINIO_ROOT_PASSWORD: minioadmin healthcheck: test: ["CMD", "mc", "ready", "local"] interval: 5s timeout: 5s retries: 5 # starts 4 docker containers running minio server instances. # using nginx reverse proxy, load balancing, you can access # it through port 9000. services: minio1: <<: *minio-common hostname: minio1 volumes:
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sun Sep 07 05:14:10 UTC 2025 - 1.5K bytes - Viewed (0) -
.github/workflows/multipart/docker-compose-site2.yaml
x-minio-common: &minio-common image: quay.io/minio/minio:${RELEASE} command: server http://site2-minio{1...4}/data{1...2} environment: - MINIO_PROMETHEUS_AUTH_TYPE=public - CI=true # starts 4 docker containers running minio server instances. # using nginx reverse proxy, load balancing, you can access # it through port 9000. services: site2-minio1: <<: *minio-common hostname: site2-minio1 volumes:
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sat Sep 30 10:13:56 UTC 2023 - 1.5K bytes - Viewed (0) -
docs/fr/docs/deployment/docker.md
Vous pouvez utiliser <a href="https://www.docker.com/" class="external-link" target="_blank">**Docker**</a> pour le déploiement. Il présente plusieurs avantages comme la sécurité, la réplicabilité, la simplicité de développement, etc. Si vous utilisez Docker, vous pouvez utiliser l'image Docker officielle :
Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Sat Nov 09 16:39:20 UTC 2024 - 7.5K bytes - Viewed (0) -
cmd/callhome.go
// Make sure only 1 callhome is running on the cluster. locker := objAPI.NewNSLock(minioMetaBucket, "callhome/runCallhome.lock") lkctx, err := locker.GetLock(ctx, callhomeLeaderLockTimeout) if err != nil { // lock timedout means some other node is the leader, // cycle back return 'true' return true } ctx = lkctx.Context() defer locker.Unlock(lkctx)
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 5.3K bytes - Viewed (0)