- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 21 for 1st (0.05 seconds)
-
docs/distributed/DECOMMISSION.md
``` λ mc admin decommission status alias/ ┌─────┬─────────────────────────────────┬──────────────────────────────────┬────────┐ │ ID │ Pools │ Capacity │ Status │ │ 1st │ http://minio{1...2}/data{1...4} │ 439 GiB (used) / 561 GiB (total) │ Active │ │ 2nd │ http://minio{3...4}/data{1...4} │ 329 GiB (used) / 421 GiB (total) │ Active │Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Mon Jul 11 14:59:49 GMT 2022 - 8.3K bytes - Click Count (0) -
internal/dsync/dsync_test.go
go func() { time.Sleep(5 * testDrwMutexAcquireTimeout) // fmt.Println("Unlocking dm1") dm1st.Unlock(t.Context()) }() dm2nd.Lock(id, source) // fmt.Printf("2nd lock obtained after 1st lock is released\n") time.Sleep(testDrwMutexRefreshCallTimeout * 2) dm2nd.Unlock(t.Context()) } // Test three locks for same resource, one succeeds, one fails (after timeout)
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 10.8K bytes - Click Count (0) -
internal/dsync/drwmutex_test.go
ctx1, cancel1 := context.WithCancel(t.Context()) if !drwm1.GetRLock(ctx1, cancel1, id, source, Options{Timeout: time.Second}) { panic("Failed to acquire read lock") } // fmt.Println("1st read lock acquired, waiting...") drwm2 := NewDRWMutex(ds, "simplelock") ctx2, cancel2 := context.WithCancel(t.Context()) if !drwm2.GetRLock(ctx2, cancel2, id, source, Options{Timeout: time.Second}) {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 9.4K bytes - Click Count (0) -
compat/maven-resolver-provider/src/main/java/org/apache/maven/repository/internal/SnapshotMetadataGeneratorFactory.java
return new RemoteSnapshotMetadataGenerator(session, request); } @Override public float getPriority() { return 30; // GAV level metadata should be deployed 1st MD }
Created: Sun Apr 05 03:35:12 GMT 2026 - Last Modified: Fri Oct 25 12:31:46 GMT 2024 - 2K bytes - Click Count (0) -
internal/s3select/jstream/decoder_test.go
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 6.6K bytes - Click Count (0) -
internal/lsync/lrwmutex_test.go
panic("Failed to acquire read lock") } // fmt.Println("1st read lock acquired, waiting...") if !lrwm.GetRLock(ctx, "", "object1", time.Second) { panic("Failed to acquire read lock") } // fmt.Println("2nd read lock acquired, waiting...") go func() { time.Sleep(2 * time.Second) lrwm.RUnlock() // fmt.Println("1st read lock released, waiting...") }() go func() {
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 7.8K bytes - Click Count (0) -
.github/workflows/translate.yml
name: Translate on: schedule: - cron: "0 5 1,15 * *" # Run at 05:00 on the 1st and 15th of every month workflow_dispatch: inputs: debug_enabled: description: Run with tmate debugging enabled (https://github.com/marketplace/actions/debugging-with-tmate) required: false default: "false" command: description: Command to run type: choice options:Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Wed Mar 18 15:36:56 GMT 2026 - 4K bytes - Click Count (0) -
cmd/namespace-lock_test.go
go func() { defer close(lk2ch) nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 1*time.Millisecond) }() time.Sleep(1 * time.Millisecond) // wait for goroutine to advance; ref=2 // Unlock the 1st lock; ref=1 after this line nsLk.unlock("volume", "path", false) // Taking another lockMapMutex here allows queuing up additional lockers. This should // not be required but makes reproduction much easier.
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 3K bytes - Click Count (0) -
tensorflow/c/eager/c_api_unified_experimental_test.cc
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Sat Oct 12 05:11:17 GMT 2024 - 39.1K bytes - Click Count (0) -
cmd/erasure-metadata-utils_test.go
distribution := []int{16, 14, 12, 10, 8, 6, 4, 2, 1, 3, 5, 7, 9, 11, 13, 15} shuffledDisks := shuffleDisks(disks, distribution) // From the "distribution" above you can notice that: // 1st data block is in the 9th disk (i.e distribution index 8) // 2nd data block is in the 8th disk (i.e distribution index 7) and so on. if shuffledDisks[0] != disks[8] || shuffledDisks[1] != disks[7] ||Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 7.3K bytes - Click Count (0)