Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 93 for lockers (0.06 sec)

  1. cmd/namespace-lock.go

    	}
    	nsLk.ref++
    	n.lockMap[resource] = nsLk
    	n.lockMapMutex.Unlock()
    
    	// Locking here will block (until timeout).
    	if readLock {
    		locked = nsLk.GetRLock(ctx, opsID, lockSource, timeout)
    	} else {
    		locked = nsLk.GetLock(ctx, opsID, lockSource, timeout)
    	}
    
    	if !locked { // We failed to get the lock
    		// Decrement ref count since we failed to get the lock
    		n.lockMapMutex.Lock()
    		n.lockMap[resource].ref--
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Sun Sep 29 22:40:36 UTC 2024
    - 9.2K bytes
    - Viewed (0)
  2. cmd/namespace-lock_test.go

    		// Unlock the 1st lock; ref=1 after this line
    		nsLk.unlock("volume", "path", false)
    
    		// Taking another lockMapMutex here allows queuing up additional lockers. This should
    		// not be required but makes reproduction much easier.
    		nsLk.lockMapMutex.Lock()
    
    		// lk3 blocks.
    		lk3ch := make(chan bool)
    		go func() {
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 3K bytes
    - Viewed (0)
  3. cmd/lock-rest-client.go

    func (c *lockRESTClient) String() string {
    	return c.connection.Remote
    }
    
    func (c *lockRESTClient) call(ctx context.Context, h *grid.SingleHandler[*dsync.LockArgs, *dsync.LockResp], args *dsync.LockArgs) (ok bool, err error) {
    	r, err := h.Call(ctx, c.connection, args)
    	if err != nil {
    		return false, err
    	}
    	defer h.PutResponse(r)
    	ok = r.Code == dsync.RespOK
    	switch r.Code {
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Mon Jul 29 18:10:04 UTC 2024
    - 3.3K bytes
    - Viewed (0)
  4. docs/pt/docs/deployment/server-workers.md

    /// info | Informação
    
    Se você estiver usando contêineres, por exemplo com Docker ou Kubernetes, falarei mais sobre isso no próximo capítulo: [FastAPI em contêineres - Docker](docker.md){.internal-link target=_blank}.
    
    Registered: Sun Sep 07 07:19:17 UTC 2025
    - Last Modified: Thu Jan 09 20:41:07 UTC 2025
    - 8.5K bytes
    - Viewed (0)
  5. cmd/lock-rest-server-common_test.go

    		TimeLastRefresh: UTCNow().UnixNano(),
    	}
    
    	locker.ll.lockMap["name"] = []lockRequesterInfo{
    		lockRequesterInfo1,
    		lockRequesterInfo2,
    	}
    
    	lri := locker.ll.lockMap["name"]
    
    	// test unknown uid
    	if locker.ll.removeEntry("name", dsync.LockArgs{
    		Owner: "owner",
    		UID:   "unknown-uid",
    	}, &lri) {
    		t.Errorf("Expected %#v, got %#v", false, true)
    	}
    
    	if !locker.ll.removeEntry("name", dsync.LockArgs{
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Wed Apr 09 14:28:39 UTC 2025
    - 3.2K bytes
    - Viewed (0)
  6. cmd/lock-rest-server.go

    )
    
    func newLockHandler(h grid.HandlerID) *grid.SingleHandler[*dsync.LockArgs, *dsync.LockResp] {
    	return grid.NewSingleHandler[*dsync.LockArgs, *dsync.LockResp](h, func() *dsync.LockArgs {
    		return &dsync.LockArgs{}
    	}, func() *dsync.LockResp {
    		return &dsync.LockResp{}
    	})
    }
    
    // registerLockRESTHandlers - register lock rest router.
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Jan 31 19:54:34 UTC 2025
    - 6.3K bytes
    - Viewed (0)
  7. docs/orchestration/docker-compose/docker-compose.yaml

      # environment:
        # MINIO_ROOT_USER: minioadmin
        # MINIO_ROOT_PASSWORD: minioadmin
      healthcheck:
        test: ["CMD", "mc", "ready", "local"]
        interval: 5s
        timeout: 5s
        retries: 5
    
    # starts 4 docker containers running minio server instances.
    # using nginx reverse proxy, load balancing, you can access
    # it through port 9000.
    services:
      minio1:
        <<: *minio-common
        hostname: minio1
        volumes:
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Sun Sep 07 05:14:10 UTC 2025
    - 1.5K bytes
    - Viewed (0)
  8. .github/workflows/multipart/docker-compose-site2.yaml

    x-minio-common: &minio-common
      image: quay.io/minio/minio:${RELEASE}
      command: server http://site2-minio{1...4}/data{1...2}
      environment:
        - MINIO_PROMETHEUS_AUTH_TYPE=public
        - CI=true
    
    # starts 4 docker containers running minio server instances.
    # using nginx reverse proxy, load balancing, you can access
    # it through port 9000.
    services:
      site2-minio1:
        <<: *minio-common
        hostname: site2-minio1
        volumes:
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Sat Sep 30 10:13:56 UTC 2023
    - 1.5K bytes
    - Viewed (0)
  9. docs/fr/docs/deployment/docker.md

    Vous pouvez utiliser <a href="https://www.docker.com/" class="external-link" target="_blank">**Docker**</a> pour le déploiement. Il présente plusieurs avantages comme la sécurité, la réplicabilité, la simplicité de développement, etc.
    
    Si vous utilisez Docker, vous pouvez utiliser l'image Docker officielle :
    
    Registered: Sun Sep 07 07:19:17 UTC 2025
    - Last Modified: Sat Nov 09 16:39:20 UTC 2024
    - 7.5K bytes
    - Viewed (0)
  10. cmd/callhome.go

    	// Make sure only 1 callhome is running on the cluster.
    	locker := objAPI.NewNSLock(minioMetaBucket, "callhome/runCallhome.lock")
    	lkctx, err := locker.GetLock(ctx, callhomeLeaderLockTimeout)
    	if err != nil {
    		// lock timedout means some other node is the leader,
    		// cycle back return 'true'
    		return true
    	}
    
    	ctx = lkctx.Context()
    	defer locker.Unlock(lkctx)
    
    Registered: Sun Sep 07 19:28:11 UTC 2025
    - Last Modified: Fri Aug 29 02:39:48 UTC 2025
    - 5.3K bytes
    - Viewed (0)
Back to top