- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 430 for minio1 (0.03 sec)
-
.github/workflows/mint/minio-pools.yaml
minio3: <<: *minio-common hostname: minio3 volumes: - pdata3-1:/pdata1 - pdata3-2:/pdata2 minio4: <<: *minio-common hostname: minio4 volumes: - pdata4-1:/pdata1 - pdata4-2:/pdata2 minio5: <<: *minio-common hostname: minio5 volumes: - pdata5-1:/pdata1 - pdata5-2:/pdata2 minio6: <<: *minio-common hostname: minio6
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Nov 03 21:18:18 UTC 2023 - 2.3K bytes - Viewed (0) -
.github/workflows/mint/nginx-8-node.conf
} upstream console { ip_hash; server minio1:9001; server minio2:9001; server minio3:9001; server minio4:9001; server minio5:9001; server minio6:9001; server minio7:9001; server minio8:9001; } server { listen 9000; listen [::]:9000; server_name localhost;
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Wed May 15 16:52:29 UTC 2024 - 3.4K bytes - Viewed (0) -
docs/site-replication/run-ssec-object-replication.sh
echo "done" # Start MinIO instances echo -n "Starting MinIO instances ..." minio server --certs-dir /tmp/certs --address ":9001" --console-address ":10000" /tmp/minio1/{1...4}/disk{1...4} /tmp/minio1/{5...8}/disk{1...4} >/tmp/minio1_1.log 2>&1 & minio server --certs-dir /tmp/certs --address ":9002" --console-address ":11000" /tmp/minio2/{1...4}/disk{1...4} /tmp/minio2/{5...8}/disk{1...4} >/tmp/minio2_1.log 2>&1 & echo "done"
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue Sep 24 08:03:58 UTC 2024 - 9.9K bytes - Viewed (0) -
docs/orchestration/docker-compose/docker-compose.yaml
retries: 5 # starts 4 docker containers running minio server instances. # using nginx reverse proxy, load balancing, you can access # it through port 9000. services: minio1: <<: *minio-common hostname: minio1 volumes: - data1-1:/data1 - data1-2:/data2 minio2: <<: *minio-common hostname: minio2 volumes: - data2-1:/data1 - data2-2:/data2
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sun Sep 07 05:14:10 UTC 2025 - 1.5K bytes - Viewed (0) -
docs/resiliency/docker-compose.yaml
# using nginx reverse proxy, load balancing, you can access # it through port 9000. services: minio1: <<: *minio-common hostname: minio1 volumes: - data1-1:/data1 - data1-2:/data2 - data1-3:/data3 - data1-4:/data4 - data1-5:/data5 - data1-6:/data6 - data1-7:/data7 - data1-8:/data8 minio2: <<: *minio-common
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sat Dec 21 04:24:45 UTC 2024 - 2.2K bytes - Viewed (0) -
.github/workflows/mint/nginx-4-node.conf
# include /etc/nginx/conf.d/*.conf; upstream minio { server minio1:9000 max_fails=1 fail_timeout=10s; server minio2:9000 max_fails=1 fail_timeout=10s; server minio3:9000 max_fails=1 fail_timeout=10s; } upstream console { ip_hash; server minio1:9001; server minio2:9001; server minio3:9001; server minio4:9001; } server {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Wed May 15 16:52:29 UTC 2024 - 3.1K bytes - Viewed (0) -
cmd/site-replication_test.go
{ []madmin.PeerInfo{ {Endpoint: "minio1:9000", Name: "minio1", DeploymentID: "dep1"}, {Endpoint: "minio2:9000", Name: "minio2", DeploymentID: "dep2"}, {Endpoint: "minio3:9000", Name: "minio3", DeploymentID: "dep3"}, }, set.CreateStringSet("dep1", "dep2", "dep3"), set.CreateStringSet("dep1"), []string{"minio2", "minio3"}, }, // Test2: new site added that is not in replicated setup {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue Jun 20 00:53:08 UTC 2023 - 2.2K bytes - Viewed (0) -
docs/distributed/distributed-from-config-file.sh
EOF done minio server --config /tmp/minio.configfile.1 >/tmp/minio1_1.log 2>&1 & site1_pid=$! minio server --config /tmp/minio.configfile.2 >/tmp/minio2_1.log 2>&1 & site2_pid=$! minio server --config /tmp/minio.configfile.3 >/tmp/minio3_1.log 2>&1 & site3_pid=$! minio server --config /tmp/minio.configfile.4 >/tmp/minio4_1.log 2>&1 & site4_pid=$!
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Jun 28 09:06:49 UTC 2024 - 3.3K bytes - Viewed (0) -
docs/site-replication/run-multi-site-minio-idp.sh
# Update tag on minio2/newbucket when minio1 is down ./mc tag set minio2/newbucket "key=val2" # create a new bucket on minio2. This should replicate to minio1 after it comes online. ./mc mb minio2/newbucket2 # delete bucket2 on minio2. This should replicate to minio1 after it comes online. ./mc rb minio2/bucket2 # Restart minio1 instance
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue Sep 24 08:03:58 UTC 2024 - 12.1K bytes - Viewed (0) -
docs/site-replication/run-sse-kms-object-replication.sh
# shellcheck disable=SC2120 exit_1() { cleanup echo "minio1 ============" cat /tmp/minio1_1.log echo "minio2 ============" cat /tmp/minio2_1.log echo "minio3 ============" cat /tmp/minio3_1.log echo "minio4 ============" cat /tmp/minio4_1.log exit 1 } cleanup() { echo -n "Cleaning up instances of MinIO ..." pkill -9 minio || sudo pkill -9 minio pkill -9 kes || sudo pkill -9 kes rm -rf ${PWD}/keys
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Thu Oct 10 06:49:55 UTC 2024 - 11.5K bytes - Viewed (0)