- Sort Score
- Num 10 results
- Language All
Results 231 - 240 of 296 for clusterid (0.06 seconds)
-
src/main/resources/esclient.xml
<property name="settings"> {"http.cors.enabled":"true", "http.cors.allow-origin":"*", "discovery.type":"single-node", "cluster.allocator.existing_shards_allocator.batch_enabled":"true", <!-- "node.name":"search_engine", "discovery.seed_hosts":"search_engine", "cluster.initial_cluster_manager_nodes":"search_engine", "node.roles":"cluster_manager,data,ingest,ml", -->
Created: Tue Mar 31 13:07:34 GMT 2026 - Last Modified: Sat Mar 28 06:59:19 GMT 2026 - 16.2K bytes - Click Count (0) -
tensorflow/c/eager/c_api_experimental_test.cc
TFE_Context* ctx = TFE_NewContext(opts, status); EXPECT_EQ(TF_GetCode(status), TF_OK) << TF_Message(status); server_def.set_task_index(0); auto cluster = server_def.mutable_cluster(); auto client_job = cluster->add_job(); client_job->set_name("localhost"); int client_port = tensorflow::testing::PickUnusedPortOrDie(); client_job->mutable_tasks()->insert(
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Oct 09 05:56:18 GMT 2025 - 31.5K bytes - Click Count (0) -
cmd/http-tracer.go
op = strings.Replace(op, "(*lockRESTServer)", "lockR", 1) op = strings.Replace(op, "(*stsAPIHandlers)", "sts", 1) op = strings.Replace(op, "(*peerS3Server)", "s3", 1) op = strings.Replace(op, "ClusterCheckHandler", "health.Cluster", 1) op = strings.Replace(op, "ClusterReadCheckHandler", "health.ClusterRead", 1) op = strings.Replace(op, "LivenessCheckHandler", "health.Liveness", 1) op = strings.Replace(op, "ReadinessCheckHandler", "health.Readiness", 1)
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Sep 24 17:13:00 GMT 2024 - 6K bytes - Click Count (0) -
cmd/metacache-manager.go
"github.com/minio/minio/internal/logger" ) // localMetacacheMgr is the *local* manager for this peer. // It should never be used directly since buckets are // distributed deterministically. // Therefore no cluster locks are required. var localMetacacheMgr = &metacacheManager{ buckets: make(map[string]*bucketMetacache), trash: make(map[string]metacache), } type metacacheManager struct { mu sync.RWMutexCreated: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Oct 25 00:44:15 GMT 2022 - 5.2K bytes - Click Count (0) -
build-tools-internal/src/main/java/org/elasticsearch/gradle/internal/InternalDistributionArchiveSetupPlugin.java
* filename, so they must be placed in different directories. * 3. We provide a packed and an unpacked variant of the distribution * - the unpacked variant is used by consumers like test cluster definitions * 4. Having per-distribution sub-projects means we can build them in parallel. */ public class InternalDistributionArchiveSetupPlugin implements InternalPlugin {Created: Wed Apr 08 16:19:15 GMT 2026 - Last Modified: Wed Sep 22 19:01:26 GMT 2021 - 7.9K bytes - Click Count (0) -
src/test/java/jcifs/internal/fscc/FileFsFullSizeInformationTest.java
@Test @DisplayName("Should handle typical Windows NTFS configuration") void shouldHandleTypicalNTFSConfiguration() throws SMBProtocolDecodingException { // Given - typical NTFS: 4KB clusters (8 sectors * 512 bytes) long totalClusters = 26214400L; // 100GB / 4KB long freeClusters = 13107200L; // 50GB / 4KB ByteBuffer buffer = ByteBuffer.allocate(32);Created: Sun Apr 05 00:10:12 GMT 2026 - Last Modified: Thu Aug 14 05:31:44 GMT 2025 - 30.5K bytes - Click Count (0) -
build-tools-internal/build.gradle
id = 'elasticsearch.internal-available-ports' implementationClass = 'org.elasticsearch.gradle.internal.InternalAvailableTcpPortProviderPlugin' } internalClusterTest { id = 'elasticsearch.internal-cluster-test' implementationClass = 'org.elasticsearch.gradle.internal.test.InternalClusterTestPlugin' } internalDistributionArchiveCheck { id = 'elasticsearch.internal-distribution-archive-check'Created: Wed Apr 08 16:19:15 GMT 2026 - Last Modified: Mon Sep 13 18:10:22 GMT 2021 - 11.3K bytes - Click Count (0) -
cmd/metrics-v3-cache.go
loadDataUsage := func(ctx context.Context) (u DataUsageInfo, err error) { objLayer := newObjectLayerFn() if objLayer == nil { return u, err } // Collect cluster level object metrics. u, err = loadDataUsageFromBackend(GlobalContext, objLayer) return u, err } return cachevalue.NewFromFunc(1*time.Minute, cachevalue.Opts{ReturnLastGood: true}, loadDataUsage) }
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 8.1K bytes - Click Count (0) -
docs/kms/README.md
``` In a given setup, there are `n` MinIO instances talking to `m` KES servers but only `1` central KMS. The most simple setup consists of `1` MinIO server or cluster talking to `1` KMS via `1` KES server. The main difference between various MinIO-KMS deployments is the KMS implementation. The following table helps you select the right option for your use case:
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 7.2K bytes - Click Count (0) -
cmd/endpoint.go
type EndpointServerPools []PoolEndpoints // ESCount returns the total number of erasure sets in this cluster func (l EndpointServerPools) ESCount() (count int) { for _, p := range l { count += p.SetCount } return count } // GetNodes returns a sorted list of nodes in this cluster func (l EndpointServerPools) GetNodes() (nodes []Node) { nodesMap := make(map[string]Node) for _, pool := range l {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 34.5K bytes - Click Count (0)