Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 93 for RLocker (0.06 sec)

  1. internal/lsync/lrwmutex_test.go

    	wl := NewLRWMutex()
    	var rl sync.Locker
    	wlocked := make(chan bool, 1)
    	rlocked := make(chan bool, 1)
    	rl = wl.DRLocker()
    	n := 10
    	go func() {
    		for range n {
    			rl.Lock()
    			rl.Lock()
    			rlocked <- true
    			wl.Lock()
    			wlocked <- true
    		}
    	}()
    	for range n {
    		<-rlocked
    		rl.Unlock()
    		select {
    		case <-wlocked:
    			t.Fatal("RLocker() didn't read-lock it")
    		default:
    		}
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sun Sep 28 20:59:21 UTC 2025
    - 7.8K bytes
    - Viewed (0)
  2. docs/fr/docs/deployment/docker.md

    Vous pouvez utiliser <a href="https://www.docker.com/" class="external-link" target="_blank">**Docker**</a> pour le déploiement. Il présente plusieurs avantages comme la sécurité, la réplicabilité, la simplicité de développement, etc.
    
    Si vous utilisez Docker, vous pouvez utiliser l'image Docker officielle :
    
    Registered: Sun Dec 28 07:19:09 UTC 2025
    - Last Modified: Sat Nov 09 16:39:20 UTC 2024
    - 7.5K bytes
    - Viewed (0)
  3. internal/dsync/locker.go

    // You should have received a copy of the GNU Affero General Public License
    // along with this program.  If not, see <http://www.gnu.org/licenses/>.
    
    package dsync
    
    import "context"
    
    // NetLocker is dsync compatible locker interface.
    type NetLocker interface {
    	// Do read lock for given LockArgs.  It should return
    	// * a boolean to indicate success/failure of the operation
    	// * an error on failure of lock request operation.
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Tue Jan 18 20:44:38 UTC 2022
    - 2.7K bytes
    - Viewed (0)
  4. docker-buildx.sh

    		_build "${each_osarch}"
    	done
    
    	sudo sysctl net.ipv6.conf.all.disable_ipv6=0
    
    	docker buildx build --push --no-cache \
    		--build-arg RELEASE="${release}" \
    		-t "registry.min.dev/community/minio:latest" \
    		-t "registry.min.dev/community/minio:${release}" \
    		--platform=linux/arm64,linux/amd64,linux/ppc64le \
    		-f Dockerfile .
    
    	docker buildx prune -f
    
    	sudo sysctl net.ipv6.conf.all.disable_ipv6=0
    }
    
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sun Oct 19 08:22:05 UTC 2025
    - 1.9K bytes
    - Viewed (0)
  5. .github/workflows/multipart/docker-compose-site2.yaml

    x-minio-common: &minio-common
      image: quay.io/minio/minio:${RELEASE}
      command: server http://site2-minio{1...4}/data{1...2}
      environment:
        - MINIO_PROMETHEUS_AUTH_TYPE=public
        - CI=true
    
    # starts 4 docker containers running minio server instances.
    # using nginx reverse proxy, load balancing, you can access
    # it through port 9000.
    services:
      site2-minio1:
        <<: *minio-common
        hostname: site2-minio1
        volumes:
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sat Sep 30 10:13:56 UTC 2023
    - 1.5K bytes
    - Viewed (0)
  6. docs/orchestration/docker-compose/docker-compose.yaml

      # environment:
        # MINIO_ROOT_USER: minioadmin
        # MINIO_ROOT_PASSWORD: minioadmin
      healthcheck:
        test: ["CMD", "mc", "ready", "local"]
        interval: 5s
        timeout: 5s
        retries: 5
    
    # starts 4 docker containers running minio server instances.
    # using nginx reverse proxy, load balancing, you can access
    # it through port 9000.
    services:
      minio1:
        <<: *minio-common
        hostname: minio1
        volumes:
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sun Sep 07 05:14:10 UTC 2025
    - 1.5K bytes
    - Viewed (0)
  7. docs/resiliency/docker-compose.yaml

      expose:
        - "9000"
        - "9001"
      environment:
        MINIO_CI_CD: 1
      healthcheck:
        test: ["CMD", "mc", "ready", "local"]
        interval: 5s
        timeout: 5s
        retries: 5
    
    # starts 4 docker containers running minio server instances.
    # using nginx reverse proxy, load balancing, you can access
    # it through port 9000.
    services:
      minio1:
        <<: *minio-common
        hostname: minio1
        volumes:
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sat Dec 21 04:24:45 UTC 2024
    - 2.2K bytes
    - Viewed (0)
  8. .github/workflows/multipart/docker-compose-site1.yaml

    x-minio-common: &minio-common
      image: quay.io/minio/minio:${RELEASE}
      command: server http://site1-minio{1...4}/data{1...2}
      environment:
        - MINIO_PROMETHEUS_AUTH_TYPE=public
        - CI=true
    
    # starts 4 docker containers running minio server instances.
    # using nginx reverse proxy, load balancing, you can access
    # it through port 9000.
    services:
      site1-minio1:
        <<: *minio-common
        hostname: site1-minio1
        volumes:
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sat Sep 30 10:13:56 UTC 2023
    - 1.5K bytes
    - Viewed (0)
  9. dockerscripts/docker-entrypoint.sh

    Harshavardhana <******@****.***> 1699046298 -0700
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Fri Nov 03 21:18:18 UTC 2023
    - 675 bytes
    - Viewed (0)
  10. .github/workflows/multipart/migrate.sh

    cd .github/workflows/multipart/
    
    function cleanup() {
    	docker-compose -f docker-compose-site1.yaml rm -s -f || true
    	docker-compose -f docker-compose-site2.yaml rm -s -f || true
    	for volume in $(docker volume ls -q | grep minio); do
    		docker volume rm ${volume} || true
    	done
    
    	docker system prune -f || true
    	docker volume prune -f || true
    	docker volume rm $(docker volume ls -q -f dangling=true) || true
    }
    
    cleanup
    
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Mon May 27 19:17:46 UTC 2024
    - 4.3K bytes
    - Viewed (0)
Back to top